Abstract
Constant-specified and exponential concentration inequalities play an essential role in the finite-sample theory of machine learning and high-dimensional statistics area. We obtain sharper and constants-specified concentration inequalities for the sum of independent sub-Weibull random variables, which leads to a mixture of two tails: sub-Gaussian for small deviations and sub-Weibull for large deviations from the mean. These bounds are new and improve existing bounds with sharper constants. In addition, a new sub-Weibull parameter is also proposed, which enables recovering the tight concentration inequality for a random variable (vector). For statistical applications, we give an ℓ2-error of estimated coefficients in negative binomial regressions when the heavy-tailed covariates are sub-Weibull distributed with sparse structures, which is a new result for negative binomial regressions. In applying random matrices, we derive non-asymptotic versions of Bai-Yin’s theorem for sub-Weibull entries with exponential tail bounds. Finally, by demonstrating a sub-Weibull confidence region for a log-truncated Z-estimator without the second-moment condition, we discuss and define the sub-Weibull type robust estimator for independent observations {Xi }ni=1without exponential-moment conditions.
| Original language | English |
|---|---|
| Article number | 2252 |
| Journal | Mathematics |
| Volume | 10 |
| Issue number | 13 |
| DOIs | |
| State | Published - 1 Jul 2022 |
| Externally published | Yes |
Keywords
- constants-specified concentration inequalities
- exponential tail bounds
- heavy-tailed random variables
- lower bounds on the least singular value
- sub-Weibull parameter
Fingerprint
Dive into the research topics of 'Sharper Sub-Weibull Concentrations'. Together they form a unique fingerprint.Cite this
- APA
- Author
- BIBTEX
- Harvard
- Standard
- RIS
- Vancouver