Role of Activation Functions and Order of Input Sequences in Question Answering

dc.contributor.authorChenna, Keshava, B.S.
dc.contributor.authorSumukha, P.K.
dc.contributor.authorChandrasekaran, K.
dc.contributor.authorUsha, D.
dc.date.accessioned2020-03-30T09:45:52Z
dc.date.available2020-03-30T09:45:52Z
dc.date.issued2020
dc.description.abstractThis paper describes a solution for the Question Answering problem in Natural Language Processing using LSTMs. We perform an analysis on the effect of choice of activation functions in the final layer of LSTM cell on the accuracy. Facebook Research�s bAbI dataset is used for our experiments. We also propose an alternative solution, which exploits the language structure and order of words in the English language, i.e. reversing the order of paragraph will introduce many short-term dependencies between the textual data and the initial tokens of a question. This method improves the accuracy in more than half of the tasks by more than 30% over the current state of the art. Our contributions in this paper are improving the accuracy of most of the Q&A tasks by reversing the order of words in the query and the story sections. Also, we have provided a comparison of different activation functions and their respective accuracies with respect to all the 20 different NLP tasks. � 2020, Springer Nature Singapore Pte Ltd.en_US
dc.identifier.citationAdvances in Intelligent Systems and Computing, 2020, Vol.1016, , pp.377-390en_US
dc.identifier.urihttps://idr.nitk.ac.in/handle/123456789/6565
dc.titleRole of Activation Functions and Order of Input Sequences in Question Answeringen_US
dc.typeBook chapteren_US

Files