跳到主要导航 跳到搜索 跳到主要内容

Contrastive Self-Supervised Representation Learning for Sensing Signals from the Time-Frequency Perspective

  • Dongxin Liu
  • , Tianshi Wang
  • , Shengzhong Liu
  • , Ruijie Wang
  • , Shuochao Yao
  • , Tarek Abdelzaher

科研成果: 书/报告/会议事项章节会议稿件同行评审

摘要

This paper presents a contrastive self-supervised representation learning framework that is new in being designed specifically for deep learning from frequency domain data. Contrastive self-supervised representation learning trains neural networks using mostly unlabeled data. It is motivated by the need to reduce the labeling burden of deep learning. In this paper, we are specifically interested in applying this approach to physical sensing scenarios, such as those arising in Internet-of-Things (IoT) applications. Deep neural networks have been widely utilized in IoT applications, but the performance of such models largely depends on the availability of large labeled datasets, which in turn entails significant training costs. Motivated by the success of contrastive self-supervised representation learning at substantially reducing the need for labeled data (mostly in areas of computer vision and natural language processing), there is growing interest in customizing the contrastive learning framework to IoT applications. Most existing work in that space approaches the problem from a time-domain perspective. However, IoT applications often measure physical phenomena, where the underlying processes (such as acceleration, vibration, or wireless signal propagation) are fundamentally a function of signal frequencies and thus have sparser and more compact representations in the frequency domain. Recently, this observation motivated the development of Short-Time Fourier Neural Networks (STFNets) that learn directly in the frequency domain, and were shown to offer large performance gains compared to Convolutional Neural Networks (CNNs) when designing supervised learning models for IoT tasks. Hence, in this paper, we introduce an STFNet-based Contrastive Self-supervised representation Learning framework (STF-CSL). STF-CSL takes both time-domain and frequency-domain features into consideration. We build the encoder using STFNet as the fundamental building block. We also apply both time-domain data augmentation and frequency-domain data augmentation during the self-supervised training process. We evaluate the resulting performance of STF-CSL on various human activity recognition tasks. The evaluation results demonstrate that STF-CSL significantly outperforms the time-domain based self-supervised approaches thereby substantially enhancing our ability to train deep neural networks from unlabeled data in IoT contexts.

源语言英语
主期刊名30th International Conference on Computer Communications and Networks, ICCCN 2021
出版商Institute of Electrical and Electronics Engineers Inc.
ISBN(电子版)9780738113302
DOI
出版状态已出版 - 7月 2021
已对外发布
活动30th International Conference on Computer Communications and Networks, ICCCN 2021 - Virtual, Athens, 希腊
期限: 19 7月 202122 7月 2021

出版系列

姓名Proceedings - International Conference on Computer Communications and Networks, ICCCN
2021-July
ISSN(印刷版)1095-2055

会议

会议30th International Conference on Computer Communications and Networks, ICCCN 2021
国家/地区希腊
Virtual, Athens
时期19/07/2122/07/21

指纹

探究 'Contrastive Self-Supervised Representation Learning for Sensing Signals from the Time-Frequency Perspective' 的科研主题。它们共同构成独一无二的指纹。

引用此