Minimum discrepancy principle strategy for choosing $k$ in $k$-NN regression
From MaRDI portal
Publication:6347398
arXiv2008.08718MaRDI QIDQ6347398
Author name not available (Why is that?)
Publication date: 19 August 2020
Abstract: We present a novel data-driven strategy to choose the hyperparameter in the -NN regression estimator. We treat the problem of choosing the hyperparameter as an iterative procedure (over ) and propose using an easily implemented in practice strategy based on the idea of early stopping and the minimum discrepancy principle. This model selection strategy is proven to be minimax-optimal, under the fixed-design assumption on covariates, over some smoothness function classes, for instance, the Lipschitz functions class on a bounded domain. The novel method often improves statistical performance on artificial and real-world data sets in comparison to other model selection strategies, such as the Hold-out method and 5-fold cross-validation. The novelty of the strategy comes from reducing the computational time of the model selection procedure while preserving the statistical (minimax) optimality of the resulting estimator. More precisely, given a sample of size , assuming that the nearest neighbors are already precomputed, if one should choose among , the strategy reduces the computational time of the generalized cross-validation or Akaike's AIC criteria from to , where is the proposed (minimum discrepancy principle) value of the nearest neighbors. Code for the simulations is provided at https://github.com/YaroslavAveryanov/Minimum-discrepancy-principle-for-choosing-k.
Has companion code repository: https://github.com/YaroslavAveryanov/Minimum-discrepancy-principle-for-choosing-k
This page was built for publication: Minimum discrepancy principle strategy for choosing $k$ in $k$-NN regression
Report a bug (only for logged in users!)Click here to report a bug for this page (MaRDI item Q6347398)