@inproceedings{889c56ebd7da4d6d9fe5865d910472aa,
title = "Adversarial Word Dilution as Text Data Augmentation in Low-Resource Regime",
abstract = "Data augmentation is widely used in text classification, especially in the low-resource regime where a few examples for each class are available during training. Despite the success, generating data augmentations as hard positive examples that may increase their effectiveness is under-explored. This paper proposes an Adversarial Word Dilution (AWD) method that can generate hard positive examples as text data augmentations to train the low-resource text classification model efficiently. Our idea of augmenting the text data is to dilute the embedding of strong positive words by weighted mixing with unknown-word embedding, making the augmented inputs hard to be recognized as positive by the classification model. We adversarially learn the dilution weights through a constrained min-max optimization process with the guidance of the labels. Empirical studies on three benchmark datasets show that AWD can generate more effective data augmentations and outperform the state-of-the-art text data augmentation methods. The additional analysis demonstrates that the data augmentations generated by AWD are interpretable and can flexibly extend to new examples without further training.",
author = "Junfan Chen and Richong Zhang and Zheyan Luo and Chunming Hu and Yongyi Mao",
note = "Publisher Copyright: Copyright {\textcopyright} 2023, Association for the Advancement of Artificial Intelligence (www.aaai.org). All rights reserved.; 37th AAAI Conference on Artificial Intelligence, AAAI 2023 ; Conference date: 07-02-2023 Through 14-02-2023",
year = "2023",
month = jun,
day = "27",
doi = "10.1609/aaai.v37i11.26486",
language = "英语",
series = "Proceedings of the 37th AAAI Conference on Artificial Intelligence, AAAI 2023",
publisher = "AAAI press",
pages = "12626--12634",
editor = "Brian Williams and Yiling Chen and Jennifer Neville",
booktitle = "AAAI-23 Technical Tracks 11",
}