Comparison of Image Encoder Architectures for Image Captioning
No Thumbnail Available
Date
2021
Authors
Journal Title
Journal ISSN
Volume Title
Publisher
Institute of Electrical and Electronics Engineers Inc.
Abstract
Image captioning is a fascinating and challenging task combining two interesting fields in computer science - Natural Language Processing and Computer Vision. Several approaches have been tried for this task. Many of these approaches are based on a combination of Convolutional Neural Networks(CNN's) as encoders and Recurrent Neural Networks(RNN's) as decoders. This paper compares two CNN-based encoders - (Vector Geometry Group) VGG16 and InceptionV3, which are used as encoders for encoding the input image features. The proposed research work uses the popular Flickr8k dataset and also compares among two different loss functions used with each of the above architectures- Categorical Cross Entropy loss function and Kullback-Leibler Divergence. Our main goal is to study the effect of the image encoder and the loss function on the image captioning task while keeping all other parameters the same. © 2021 IEEE.
Description
Keywords
(Vector Geometry Group) VGG16, Convolutional Neural Networks(CNN), Encoder-decoder framework, Image captioning, InceptionV3
Citation
Proceedings - 5th International Conference on Computing Methodologies and Communication, ICCMC 2021, 2021, Vol., , p. 740-744
