Inner Attention Based bi-LSTMs with Indexing for non-Factoid Question Answering

No Thumbnail Available

Date

2019

Authors

Sharma, A.
Harithas, C.

Journal Title

Journal ISSN

Volume Title

Publisher

Abstract

In this paper, we focussed on non-factoid question answering problem using a bidirectional LSTM with an inner attention mechanism and indexing for better accuracy. Non factoid QA is an important task and can be significantly applied in constructing useful knowledge bases and extracting valuable information. The advantage of using Deep Learning frameworks in solving these kind of problems is that it does not require any feature engineering and other linguistic tools. The proposed approach is to extend a LSTM (Long Short Term Memory) model in two directions, one with a Convolutional layer and other with an inner attention mechanism, proposed by Bingning Wang, et al., to the LSTMs, to generate answer representations in accordance with the question. On top of this Deep Learning model we used an information retrieval model based on indexing to generate answers and improve the accuracy. The proposed methodology showed an improvement in accuracy over the referred model and respective baselines and also with respect to the answer lengths used. The models are tested with two non factoid QA data sets: TREC-QA and InsuranceQA. � 2018 IEEE.

Description

Keywords

Citation

Proceedings - 17th IEEE International Conference on Machine Learning and Applications, ICMLA 2018, 2019, Vol., , pp.1-7

Endorsement

Review

Supplemented By

Referenced By