A Dual Prediction Network for Image Captioning

conference paper
General captioning practice involves a single forward prediction, with the aim of predicting the word in the next timestep given the word in the current timestep. In this paper, we present a novel captioning framework, namely Dual Prediction Network (DPN), which is end-to-end trainable and addresses the captioning problem with dual predictions. Specifically, the dual predictions consist of a forward prediction to generate the next word from the current input word, as well as a backward prediction to reconstruct the input word using the predicted word. DPN has two appealing properties: 1) By introducing an extra supervision signal on the prediction, DPN can better capture the interplay between the input and the target; 2) Utilizing the reconstructed input, DPN can make another new prediction. During the test phase, we average both predictions to formulate the final target sentence. Experimental results on the MS COCO dataset demonstrate that, benefiting from the reconstruction step, both generated predictions in DPN outperform the predictions of methods based on the general captioning practice (single forward prediction), and averaging them can bring a further accuracy boost. Overall, DPN achieves competitive results with state-of-the-art approaches, across multiple evaluation metrics. © 2018 IEEE.
TNO Identifier
862252
ISSN
19457871
ISBN
9781538617373
Publisher
IEEE Computer Society
Article nr.
8486491
Source title
Proceedings - IEEE International Conference on Multimedia and Expo, 2018 IEEE International Conference on Multimedia and Expo, ICME 2018, 23 July 2018 through 27 July 2018
Files
To receive the publication files, please send an e-mail request to TNO Repository.