Please use this identifier to cite or link to this item: https://idr.nitk.ac.in/jspui/handle/123456789/8318
Title: Inner Attention Based bi-LSTMs with Indexing for non-Factoid Question Answering
Authors: Sharma, A.
Harithas, C.
Issue Date: 2019
Citation: Proceedings - 17th IEEE International Conference on Machine Learning and Applications, ICMLA 2018, 2019, Vol., , pp.1-7
Abstract: In this paper, we focussed on non-factoid question answering problem using a bidirectional LSTM with an inner attention mechanism and indexing for better accuracy. Non factoid QA is an important task and can be significantly applied in constructing useful knowledge bases and extracting valuable information. The advantage of using Deep Learning frameworks in solving these kind of problems is that it does not require any feature engineering and other linguistic tools. The proposed approach is to extend a LSTM (Long Short Term Memory) model in two directions, one with a Convolutional layer and other with an inner attention mechanism, proposed by Bingning Wang, et al., to the LSTMs, to generate answer representations in accordance with the question. On top of this Deep Learning model we used an information retrieval model based on indexing to generate answers and improve the accuracy. The proposed methodology showed an improvement in accuracy over the referred model and respective baselines and also with respect to the answer lengths used. The models are tested with two non factoid QA data sets: TREC-QA and InsuranceQA. � 2018 IEEE.
URI: http://idr.nitk.ac.in/jspui/handle/123456789/8318
Appears in Collections:2. Conference Papers

Files in This Item:
There are no files associated with this item.


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.