1. Collaborative Learning for Answer Selection in Question Answering
- Author
-
Pengfei Zhang, Honghui Chen, Xiaoyan Kui, and Taihua Shao
- Subjects
General Computer Science ,Computer science ,collaborative learning ,02 engineering and technology ,computer.software_genre ,Convolutional neural network ,Knowledge extraction ,Answer selection ,0502 economics and business ,0202 electrical engineering, electronic engineering, information engineering ,Question answering ,Selection (linguistics) ,General Materials Science ,natural language processing ,Artificial neural network ,business.industry ,Deep learning ,05 social sciences ,General Engineering ,deep learning ,Collaborative learning ,question answering ,Task analysis ,Embedding ,050211 marketing ,020201 artificial intelligence & image processing ,lcsh:Electrical engineering. Electronics. Nuclear engineering ,Artificial intelligence ,business ,lcsh:TK1-9971 ,computer ,Sentence ,Natural language processing - Abstract
Answer selection is an essential step in a question answering (QA) system. Traditional methods for this task mainly focus on developing linguistic features that are limited in practice. With the great success of deep learning method in distributed text representation, deep learning-based answer selection approaches have been well investigated, which mainly employ only one neural network, i.e., convolutional neural network (CNN) or long short term memory (LSTM), leading to failures in extracting some rich sentence features. Thus, in this paper, we propose a collaborative learning-based answer selection model (QA-CL), where we deploy a parallel training architecture to collaboratively learn the initial word vector matrix of the sentence by CNN and bidirectional LSTM (BiLSTM) at the same time. In addition, we extend our model by incorporating the sentence embedding generated by the QA-CL model into a joint distributed sentence representation using a strong unsupervised baseline weight removal (WR), i.e., the QA-CLWR model. We evaluate our proposals on a popular QA dataset, InsuranceQA. The experimental results indicate that our proposed answer selection methods can produce a better performance compared with several strong baselines. Finally, we investigate the models’ performance with respect to different question types and find that question types with a medium number of questions have a better and more stable performance than those types with too large or too small number of questions.
- Published
- 2019
- Full Text
- View/download PDF