Skip to main navigation Skip to search Skip to main content

Multi-Modality Self-Distillation for Weakly Supervised Temporal Action Localization

  • Linjiang Huang
  • , Liang Wang
  • , Hongsheng Li*
  • *Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

As a challenging task of high-level video understanding, Weakly-supervised Temporal Action Localization (WTAL) has attracted increasing attention in recent years. However, due to the weak supervisions of whole-video classification labels, it is challenging to accurately determine action instance boundaries. To address this issue, pseudo-label-based methods [Alwassel <italic>et al.</italic> (2019), Luo <italic>et al.</italic> (2020), and Zhai <italic>et al.</italic> (2020)] were proposed to generate snippet-level pseudo labels from classification results. In spite of the promising performance, these methods hardly take full advantages of multiple modalities, <italic>i.e.</italic>, RGB and optical flow sequences, to generate high quality pseudo labels. Most of them ignored how to mitigate the label noise, which hinders the capability of the network on learning discriminative feature representations. To address these challenges, we propose a Multi-Modality Self-Distillation (MMSD) framework, which contains two single-modal streams and a fused-modal stream to perform multi-modality knowledge distillation and multi-modality self-voting. On the one hand, multi-modality knowledge distillation improves snippet-level classification performance by transferring knowledge between single-modal streams and a fused-modal stream. On the other hand, multi-modality self-voting mitigates the label noise in a modality voting manner according to the reliability and complementarity of the streams. Experimental results on THUMOS14 and ActivityNet1.3 datasets demonstrate the effectiveness of our method and superior performance over state-of-the-art approaches. Our code is available at <uri>https://github.com/LeonHLJ/MMSD</uri>.

Original languageEnglish
Pages (from-to)1504-1519
Number of pages16
JournalIEEE Transactions on Image Processing
Volume31
DOIs
StatePublished - 2022
Externally publishedYes

Keywords

  • Annotations
  • Head
  • Location awareness
  • Noise measurement
  • Reliability
  • Task analysis
  • Training

Fingerprint

Dive into the research topics of 'Multi-Modality Self-Distillation for Weakly Supervised Temporal Action Localization'. Together they form a unique fingerprint.

Cite this