A survey on recurrent neural network architectures for sequential learning

dc.contributor.authorShiva Prakash B.
dc.contributor.authorSanjeev K.V.
dc.contributor.authorPrakash R.
dc.contributor.authorChandrasekaran K.
dc.date.accessioned2020-03-31T14:15:19Z
dc.date.available2020-03-31T14:15:19Z
dc.date.issued2019
dc.description.abstractThe expanding textual information and significance of examining the substance has started a colossal research in the field of synopsis. Text summarization is the process of conveying the gist of a text with a minimized representation. The requirement for automation of the procedure is at its apex with exponential burst of information because of digitization. Text captioning comes under the branch of abstractive summarization which captures the gist of the article in a few words. In this paper, we present an approach to text captioning using recurrent neural networks which comprise of an encoder–decoder model. The key challenges dealt here was to figure out the ideal input required to produce the desired output. The model performs better when the input is fed with the summary as compared to the original article itself. The recurrent neural network model with LSTM results has been effective in transcribing a caption for the textual data. © Springer Nature Singapore Pte Ltd. 2019en_US
dc.identifier.citationAdvances in Intelligent Systems and Computing, 2019, Vol.817, pp.57-66en_US
dc.identifier.uri10.1007/978-981-13-1595-4_5
dc.identifier.urihttps://idr.nitk.ac.in/handle/123456789/13726
dc.titleA survey on recurrent neural network architectures for sequential learningen_US
dc.typeBook Chapteren_US

Files

Collections