Generalization Bounds for High-dimensional M-estimation under Sparsity Constraint

01/20/2020
by   Xiao-Tong Yuan, et al.
0

The ℓ_0-constrained empirical risk minimization (ℓ_0-ERM) is a promising tool for high-dimensional statistical estimation. The existing analysis of ℓ_0-ERM estimator is mostly on parameter estimation and support recovery consistency. From the perspective of statistical learning, another fundamental question is how well the ℓ_0-ERM estimator would perform on unseen samples. The answer to this question is important for understanding the learnability of such a non-convex (and also NP-hard) M-estimator but still relatively under explored. In this paper, we investigate this problem and develop a generalization theory for ℓ_0-ERM. We establish, in both white-box and black-box statistical regimes, a set of generalization gap and excess risk bounds for ℓ_0-ERM to characterize its sparse prediction and optimization capability. Our theory mainly reveals three findings: 1) tighter generalization bounds can be attained by ℓ_0-ERM than those of ℓ_2-ERM if the risk function is (with high probability) restricted strongly convex; 2) tighter uniform generalization bounds can be established for ℓ_0-ERM than the conventional dense ERM; and 3) sparsity level invariant bounds can be established by imposing additional strong-signal conditions to ensure the stability of ℓ_0-ERM. In light of these results, we further provide generalization guarantees for the Iterative Hard Thresholding (IHT) algorithm which serves as one of the most popular greedy pursuit methods for approximately solving ℓ_0-ERM. Numerical evidence is provided to confirm our theoretical predictions when implied to sparsity-constrained linear regression and logistic regression models.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset
Success!
Error Icon An error occurred

Sign in with Google

×

Use your Google Account to sign in to DeepAI

×

Consider DeepAI Pro