Back to Search Start Over

Self-Supervised Contextual Data Augmentation for Natural Language Processing.

Authors :
Park, Dongju
Ahn, Chang Wook
Source :
Symmetry (20738994). Nov2019, Vol. 11 Issue 11, p1393-1393. 1p.
Publication Year :
2019

Abstract

In this paper, we propose a novel data augmentation method with respect to the target context of the data via self-supervised learning. Instead of looking for the exact synonyms of masked words, the proposed method finds words that can replace the original words considering the context. For self-supervised learning, we can employ the masked language model (MLM), which masks a specific word within a sentence and obtains the original word. The MLM learns the context of a sentence through asymmetrical inputs and outputs. However, without using the existing MLM, we propose a label-masked language model (LMLM) that can include label information for the mask tokens used in the MLM to effectively use the MLM in data with label information. The augmentation method performs self-supervised learning using LMLM and then implements data augmentation through the trained model. We demonstrate that our proposed method improves the classification accuracy of recurrent neural networks and convolutional neural network-based classifiers through several experiments for text classification benchmark datasets, including the Stanford Sentiment Treebank-5 (SST5), the Stanford Sentiment Treebank-2 (SST2), the subjectivity (Subj), the Multi-Perspective Question Answering (MPQA), the Movie Reviews (MR), and the Text Retrieval Conference (TREC) datasets. In addition, since the proposed method does not use external data, it can eliminate the time spent collecting external data, or pre-training using external data. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
20738994
Volume :
11
Issue :
11
Database :
Academic Search Index
Journal :
Symmetry (20738994)
Publication Type :
Academic Journal
Accession number :
139863298
Full Text :
https://doi.org/10.3390/sym11111393