Processing math: 100%
Skip to yearly menu bar Skip to main content


Poster

Near-Interpolators: Rapid Norm Growth and the Trade-Off between Interpolation and Generalization

Yutong Wang · Rishi Sonthalia · Wei Hu

MR1 & MR2 - Number 57

Abstract: We study the generalization capability of nearly-interpolating linear regressors: β's whose training error τ is positive but small, i.e., below the noise floor. Under a random matrix theoretic assumption on the data distribution and an eigendecay assumption on the data covariance matrix Σ, we demonstrate that any near-interpolator exhibits rapid norm growth: for τ fixed, β has squared 2-norm E[β22]=Ω(nα) where n is the number of samples and α>1 is the exponent of the eigendecay, i.e., λi(Σ)iα.This implies that existing data-independent norm-based bounds are necessarily loose. On the other hand, in the same regime we precisely characterize the asymptotic trade-off between interpolation and generalization. Our characterization reveals thatlarger norm scaling exponents α correspond to worse trade-offs between interpolation and generalization. We verify empirically that a similar phenomenon holds for nearly-interpolating shallow neural networks.

Chat is not available.