Optimism
Last updated
Last updated
The notation indicates that we observe N new response values at each of the training points .
Extra sample error can be decomposed into in-sample error and out of sample error. In sample error is the sum of training error and optimism.
(Because and are random quantity, we use triple equal sign.) It means has the distribution same as . Let's get the expected op to predict it.
Proof: https://stats.stackexchange.com/questions/88912/optimism-bias-estimates-of-prediction-error
Expected in-sample error. Trace becomes d because of effective number of parameters.
An obvious way to estimate prediction error is to estimate the optimism and then add it to the training error.
To estimate In-sample error we estimate expected In-sample error
The way omega is estimated decide following equations:
AIC(Akaike Information Criterion) uses log-likelihood loss function. 아래의 식은 Expected in-sample error에 관한 식이다.
Proof: http://faculty.washington.edu/yenchic/19A_stat535/Lec7_model.pdf
The idea of AIC is to adjust the empirical risk to be an unbiased estimator of the true risk in a parametric model. Under a likelihood framework, the loss function is the negative log-likelihood function
Logistic regression model(binomial log likelihood)
Gaussian model
The aim is to find alpha that minimizes the value above using our estimated value of test error.
More concisely, in a right side this is conditioned on but we can just use y(X is given input, we only need to consider random quantity y). In several loss functions like squared error and 0-1 satisfies the following equation.
In-sample error estimation = (average optimism) estimation.
With tuning parameter