Self-Evolution Learning for Mixup: Enhance Data Augmentation on Few-Shot Text Classification Tasks

Text classification tasks often encounter few shot scenarios with limited labeled data, and addressing data scarcity is crucial. Data augmentation with mixup has shown to be effective on various text classification tasks. However, most of the mixup methods do not consider the varying degree of learn...

Full description

Saved in:
Bibliographic Details
Published inarXiv.org
Main Authors Zheng, Haoqi, Zhong, Qihuang, Ding, Liang, Tian, Zhiliang, Niu, Xin, Li, Dongsheng, Tao, Dacheng
Format Paper
LanguageEnglish
Published Ithaca Cornell University Library, arXiv.org 27.11.2023
Subjects
Online AccessGet full text

Cover

Loading…
More Information
Summary:Text classification tasks often encounter few shot scenarios with limited labeled data, and addressing data scarcity is crucial. Data augmentation with mixup has shown to be effective on various text classification tasks. However, most of the mixup methods do not consider the varying degree of learning difficulty in different stages of training and generate new samples with one hot labels, resulting in the model over confidence. In this paper, we propose a self evolution learning (SE) based mixup approach for data augmentation in text classification, which can generate more adaptive and model friendly pesudo samples for the model training. SE focuses on the variation of the model's learning ability. To alleviate the model confidence, we introduce a novel instance specific label smoothing approach, which linearly interpolates the model's output and one hot labels of the original samples to generate new soft for label mixing up. Through experimental analysis, in addition to improving classification accuracy, we demonstrate that SE also enhances the model's generalize ability.
ISSN:2331-8422