跳到主要导航 跳到搜索 跳到主要内容

Online Continual Adaptation with Active Self-Training

  • Shiji Zhou
  • , Han Zhao
  • , Shanghang Zhang*
  • , Lianzhe Wang
  • , Heng Chang
  • , Zhi Wang
  • , Wenwu Zhu*
  • *此作品的通讯作者

科研成果: 期刊稿件会议文章同行评审

摘要

Models trained with offline data often suffer from continual distribution shifts and expensive labeling in changing environments. This calls for a new online learning paradigm where the learner can continually adapt to changing environments with limited labels. In this paper, we propose a new online setting - Online Active Continual Adaptation, where the learner aims to continually adapt to changing distributions using both unlabeled samples and active queries of limited labels. To this end, we propose Online Self-Adaptive Mirror Descent (OSAMD), which adopts an online teacher-student structure to enable online self-training from unlabeled data, and a margin-based criterion that decides whether to query the labels to track changing distributions. Theoretically, we show that, in the separable case, OSAMD has an O(T2/3) dynamic regret bound under mild assumptions, which is aligned with the Ω(T2/3) lower bound of online learning algorithms with full labels. In the general case, we show a regret bound of O(T2/3 + αT), where α denotes the separability of domains and is usually small. Our theoretical results show that OSAMD can fast adapt to changing environments with active queries. Empirically, we demonstrate that OSAMD achieves favorable regrets under changing environments with limited labels on both simulated and real-world data, which corroborates our theoretical findings.

源语言英语
页(从-至)8852-8883
页数32
期刊Proceedings of Machine Learning Research
151
出版状态已出版 - 2022
已对外发布
活动25th International Conference on Artificial Intelligence and Statistics, AISTATS 2022 - Virtual, Online, 西班牙
期限: 28 3月 202230 3月 2022

指纹

探究 'Online Continual Adaptation with Active Self-Training' 的科研主题。它们共同构成独一无二的指纹。

引用此