摘要
Cross-validation (CV) is a widely adopted approach for selecting the optimal model. However, the computation of empirical cross-validation error (CVE) has high complexity due to multiple times of learner training. In this paper, we develop a novel approximation theory of CVE and present an approximate approach to CV based on the Bouligand influence function (BIF) for kernel-based algorithms. We first represent the BIF and higher order BIFs in Taylor expansions, and approximate CV via the Taylor expansions. We then derive an upper bound of the discrepancy between the original and approximate CV. Furthermore, we provide a novel computing method to calculate the BIF for general distribution, and evaluate BIF criterion for sample distribution to approximate CV. The proposed approximate CV requires training on the full data set only once and is suitable for a wide variety of kernel-based algorithms. Experimental results demonstrate that the proposed approximate CV is sound and effective.
源语言 | 英语 |
---|---|
文章编号 | 8611136 |
页(从-至) | 1083-1096 |
页数 | 14 |
期刊 | IEEE Transactions on Pattern Analysis and Machine Intelligence |
卷 | 42 |
期 | 5 |
DOI | |
出版状态 | 已出版 - 1 5月 2020 |
已对外发布 | 是 |