Sharper Sub-Weibull Concentrations

Research output: Contribution to journalArticlepeer-review

Abstract

Constant-specified and exponential concentration inequalities play an essential role in the finite-sample theory of machine learning and high-dimensional statistics area. We obtain sharper and constants-specified concentration inequalities for the sum of independent sub-Weibull random variables, which leads to a mixture of two tails: sub-Gaussian for small deviations and sub-Weibull for large deviations from the mean. These bounds are new and improve existing bounds with sharper constants. In addition, a new sub-Weibull parameter is also proposed, which enables recovering the tight concentration inequality for a random variable (vector). For statistical applications, we give an ℓ2-error of estimated coefficients in negative binomial regressions when the heavy-tailed covariates are sub-Weibull distributed with sparse structures, which is a new result for negative binomial regressions. In applying random matrices, we derive non-asymptotic versions of Bai-Yin’s theorem for sub-Weibull entries with exponential tail bounds. Finally, by demonstrating a sub-Weibull confidence region for a log-truncated Z-estimator without the second-moment condition, we discuss and define the sub-Weibull type robust estimator for independent observations {Xi }ni=1without exponential-moment conditions.

Original languageEnglish
Article number2252
JournalMathematics
Volume10
Issue number13
DOIs
StatePublished - 1 Jul 2022
Externally publishedYes

Keywords

  • constants-specified concentration inequalities
  • exponential tail bounds
  • heavy-tailed random variables
  • lower bounds on the least singular value
  • sub-Weibull parameter

Fingerprint

Dive into the research topics of 'Sharper Sub-Weibull Concentrations'. Together they form a unique fingerprint.

Cite this