Fast computation of leave-one-out cross-validation for k-NN regression

Kanagawa, Motonobu
Submitted to ArXiV, 8 May 2024

We describe a fast computation method for leave-one-out cross-validation (LOOCV) for k-nearest neighbours (k-NN) regression. We show that, under a tie-breaking condition for nearest neighbours, the LOOCV estimate of the mean square error for k-NN regression is identical to the mean square error of (k + 1)-NN regression evaluated on the training data, multiplied by the scaling factor (k + 1)2/k2 . Therefore, to compute the LOOCV score, one only needs to fit (k + 1)-NN regression only once, and does not need to repeat trainingvalidation of k-NN regression for the number of training data. Numerical experiments confirm the validity of the fast computation method. 


HAL
Type:
Conférence
Date:
2024-05-08
Department:
Data Science
Eurecom Ref:
7722
Copyright:
© EURECOM. Personal use of this material is permitted. The definitive version of this paper was published in Submitted to ArXiV, 8 May 2024 and is available at :
See also:

PERMALINK : https://www.eurecom.fr/publication/7722