model compression: let me explain in simple terms.
lets x_train (features), y_train (target) training data.
x_train, y_train ------> m1 (example: decision tree) x_train --------> m1 ----> y_pred (predicted y x_train) now
case 1: x_train, y_pred -----------> m2 (example: model not decision tree) x_train ---------------> m2 ----------> y_pred1 case 2: x_train, y_train -----------> m2 (example: model not decision tree) x_train ---------------> m2 ----------> y_pred2 now compute auc score m2.
case 1: auc (y_pred, y_pred1) case 2: auc (y_train, y_pred2) case 1 auc higher case 2 auc. case 1 called model compression. intuition behind it. of course auc calculated probabilities.
the intuition behind result conditional entropy of y_pred given x_train zero. m2 can learn x_train-> y_pred more in second case.
Comments
Post a Comment