Back to Search Start Over

An Empirical Evaluation on Word Embeddings Across Reading Comprehension

Authors :
Xiaolin Gui
Dong Liao
Yingjie Gu
Yi Shen
Source :
ICAIT
Publication Year :
2019
Publisher :
IEEE, 2019.

Abstract

Word embeddings which is real-valued word representation able to capture lexical semantics plays a crucial role in machine reading comprehension tasks because the first step is the embedding of the question and the passage. One of the frequently-used models is Word2Vec, but several fashionable competitors have been proposed in recent years, including GloVe and Fasttext. However the question which word embedding model really performs best and is most suitable across different reading comprehension tasks remains unanswered to this date. In this paper we performed the first extrinsic empirical evaluation of three word embeddings across four types of tasks: Multiple Choice, Cloze, Answer Extraction and Conversation. The experiments showed that GloVe and Fasttext have their own strengths in different type of tasks: The accuracy of Multiple Choice task improves significantly when leveraging GloVe, Fasttext is a bit more suitable for Answer Extraction task, and GloVe performs similarly with Fasttext in Cloze task and Conversation task. Finally, we find that Word2Vec has been demoded compared with GloVe and Fasttext in all tasks.

Details

Database :
OpenAIRE
Journal :
2019 IEEE 11th International Conference on Advanced Infocomm Technology (ICAIT)
Accession number :
edsair.doi...........e9bc5bbf58d78fa56504824f6bd50e35
Full Text :
https://doi.org/10.1109/icait.2019.8935932