evaluation
metrics
- goodness of fit - how well does the learned distribution represent the real distribution?
- define a loss function $\mathcal{L}$
- 0-1 loss: $\vert C-f(X)\vert$ - hard to minimize (combinatorial)
- $L_2$ loss: $[C-f(X)[^2$
- risk = $E_{(x,y)\sim D}[\mathcal L(f(X), y) ]$
- optimal classifiers
- Bayes classifier minimizes 0-1 loss: $\hat{f}(X)=C_i$ if $P(C_i\vert X)=max_f P(f\vert X)$
- KNN minimizes $L_2$ loss: $\hat{f}(X)=E(Y\vert X)$
- classification cost functions
- misclassification error - not differentiable
- Gini index: $\sum_{i != j} p_i q_j$
- cross-entropy: $-\sum_x p(x): log : q(x) $
- accuracy-based
- accuracy = number of correct classifications / total number of test cases
- balanced accuracy = 1/2 (TP/P + TN/N)
- recall - TP/(TP+FN)
- precision - TP/(TP+FP)
- roc curve: TPR = hit rate (recall) vs. FPR (not precision) = FP/N
cv
- cross validation - don’t have enough data for a test set
- properties
- not good when n < complexity of predictor
- because summands are correlated
- assume data units are exchangeable
- can sometimes use this to pick k for k-means
- data is reused
- types
- k-fold - split data into N pieces
- N-1 pieces for fit model, 1 for test
- cycle through all N cases
- average the values we get for testing
- leave one out (LOOCV)
- train on all the data and only test on one
- then cycle through everything
- random split - shuffle and repeat
- one-way CV = prequential analysis - keep testing on next data point, updating model
- ESCV - penalize variance between folds
- k-fold - split data into N pieces
- properties
- regularization path of a regression - plot each coeff v. $\lambda$
- tells you which features get pushed to 0 and when
- for OLS (and maybe other linear models), can compute leave-one-out CV without training separate models
stability
- computational stability
- randomness in the algorithm
- perturbations to models
- generalization stability
- perturbations to data
- sampling methods
- bootstrap - take a sample
- repeatedly sample from observed sample w/ replacement
- bootstrap samples has same size as observed sample
- subsampling
- sample without replacement
- jackknife resampling
- subsample containing all but one of the points
other considerations
- computational cost
- interpretability
- model-selection criteria
- adjusted $R^2_p$ - penalty
- Mallow’s $C_p$
- $AIC_p$
- $BIC_p$
- PRESS