Sequential Memory Modelling for Video Captioning

dc.contributor.authorPuttaraja, P.
dc.contributor.authorNayaka, C.
dc.contributor.authorManikesh, M.
dc.contributor.authorSharma, N.
dc.contributor.authorAnand Kumar, A.M.
dc.date.accessioned2026-02-06T06:35:20Z
dc.date.issued2022
dc.description.abstractIn recent years, the automatic generation of natural language descriptions of video has focused on deep learning research and natural voice processing. Video understanding has multiple applications such as video search and indexing, but video subtitles are a correct sophisticated topic for complex and diverse types of video content. However, the understanding between video and natural language sets remains an open issue to better understand the video and create multiple methods to create a set automatically. The deep learning method has a major focus on the direction of video processing with performance and high-speed computing capabilities. This polling discusses an encoder-decoder network end-in-frame based on a deep learning approach to generate caption. In this paper we will describe the model, dataset and parameters used to evaluate the model. © 2022 IEEE.
dc.identifier.citationINDICON 2022 - 2022 IEEE 19th India Council International Conference, 2022, Vol., , p. -
dc.identifier.urihttps://doi.org/10.1109/INDICON56171.2022.10039829
dc.identifier.urihttps://idr.nitk.ac.in/handle/123456789/29790
dc.publisherInstitute of Electrical and Electronics Engineers Inc.
dc.subjectDeep learning
dc.subjectEncoder-Decoder Model
dc.subjectLSTM
dc.subjectNLP - Natural Language Processing
dc.titleSequential Memory Modelling for Video Captioning

Files