Skip to main navigation Skip to search Skip to main content

Non-Asymptotic Guarantees for Robust Statistical Learning under Infinite Variance Assumption

  • Lihu Xu
  • , Fang Yao
  • , Qiuran Yao
  • , Huiming Zhang*
  • *Corresponding author for this work
  • University of Macau
  • Research Institute
  • Peking University

Research output: Contribution to journalArticlepeer-review

Abstract

There has been a surge of interest in developing robust estimators for models with heavy-tailed and bounded variance data in statistics and machine learning, while few works impose unbounded variance. This paper proposes two types of robust estimators, the ridge log-truncated M-estimator and the elastic net log-truncated M-estimator. The first estimator is applied to convex regressions such as quantile regression and generalized linear models, while the other one is applied to high dimensional non-convex learning problems such as regressions via deep neural networks. Simulations and real data analysis demonstrate the robustness of log-truncated estimations over standard estimations.

Original languageEnglish
JournalJournal of Machine Learning Research
Volume24
StatePublished - 2023

Keywords

  • data with infinite variance
  • excess risk bounds
  • robust deep neural network (DNN) regressions
  • robust elastic net regressions
  • robust non-convex regressions
  • robust ridge regressions

Fingerprint

Dive into the research topics of 'Non-Asymptotic Guarantees for Robust Statistical Learning under Infinite Variance Assumption'. Together they form a unique fingerprint.

Cite this