A survey on recurrent neural network architectures for sequential learning

dc.contributor.authorPrakash, B.
dc.contributor.authorSanjeev, K.V.
dc.contributor.authorPrakash, R.
dc.contributor.authorChandrasekaran, K.
dc.date.accessioned2026-02-08T16:50:38Z
dc.date.issued2019
dc.description.abstractThe expanding textual information and significance of examining the substance has started a colossal research in the field of synopsis. Text summarization is the process of conveying the gist of a text with a minimized representation. The requirement for automation of the procedure is at its apex with exponential burst of information because of digitization. Text captioning comes under the branch of abstractive summarization which captures the gist of the article in a few words. In this paper, we present an approach to text captioning using recurrent neural networks which comprise of an encoder–decoder model. The key challenges dealt here was to figure out the ideal input required to produce the desired output. The model performs better when the input is fed with the summary as compared to the original article itself. The recurrent neural network model with LSTM results has been effective in transcribing a caption for the textual data. © Springer Nature Singapore Pte Ltd. 2019
dc.identifier.citationAdvances in Intelligent Systems and Computing, 2019, Vol.817, , p. 57-66
dc.identifier.isbn9783319604855
dc.identifier.isbn9783319276427
dc.identifier.isbn9783319419343
dc.identifier.isbn9783319232034
dc.identifier.isbn9783319938844
dc.identifier.isbn9783642330414
dc.identifier.isbn9783319262833
dc.identifier.isbn9788132220084
dc.identifier.isbn9783642375019
dc.identifier.isbn9783030026820
dc.identifier.issn21945357
dc.identifier.urihttps://doi.org/10.1007/s40515-025-00633-5
dc.identifier.urihttps://idr.nitk.ac.in/handle/123456789/33915
dc.publisherSpringer Verlag service@springer.de
dc.titleA survey on recurrent neural network architectures for sequential learning

Files

Collections