Degrees of freedom for piecewise Lipschitz estimators (Q1650119)
From MaRDI portal
scientific article
| Language | Label | Description | Also known as |
|---|---|---|---|
| English | Degrees of freedom for piecewise Lipschitz estimators |
scientific article |
Statements
Degrees of freedom for piecewise Lipschitz estimators (English)
0 references
29 June 2018
0 references
The multivariate Gaussian model \(Y\sim N(\mu,\sigma^2I)\) is considered on \(\mathbb{R}^n\) with \(\mu\) the unknown parameter. The degrees of freedom (DF) of an estimator \(\hat{\mu}\) is \(df(\hat{\mu}):=\sum_{i=1}^n \frac{\mathrm{cov}(Y_i,\hat{\mu}(Y)_i)}{\sigma^2}\). If \(\hat{\mu}\) is differentiable in Lebesgue almost all points, Steinś DF of \(\mu\) is \(df_S(\hat{\mu}):=E(\operatorname{div}(\hat{\mu})(Y)).\) If \(\hat{\mu}\) is \textit{almost differentiable} then \(df(\hat{\mu})= df_S(\hat{\mu})\) due to \textit{C. M. Stein} [Ann. Stat. 9, 1135--1151 (1981; Zbl 0476.62035)]. The main result of the paper under review provides a representation of \(df(\hat{\mu})-df_S(\hat{\mu})\) for a range of discontinuous estimators containing a number of regression estimators that include variable selection. For lasso -- Ordinary Least Squares (OLS) this yields an estimate of DF, which in turn enables to estimate the quadratic risk of lasso-OLS. Simulations demonstrate that the risk estimate leads to reliable model selection and that the risk estimate itself has smaller mean squared error than cross-validation estimate. For best subset selection a representation of \(df(\hat{\mu})-df_S(\hat{\mu})\) holds as well, but the situation is more complicated compared with lasso-OLS. However, it is possible to derive an approximation, which is exact for orthogonal designs.
0 references
best subset selection
0 references
lasso-OLS
0 references
degrees of freedom
0 references
Stein's lemma
0 references
0 references