Connecting Language to Images: A Progressive Attention-Guided Network for Simultaneous Image Captioning and Language Grounding

Authors

  • Lingyun Song Xi'an JiaoTong University
  • Jun Liu Xi'an Jiaotong Univerisity
  • Buyue Qian Xi'an Jiaotong University
  • Yihe Chen University of Toronto

DOI:

https://doi.org/10.1609/aaai.v33i01.33018885

Abstract

Image captioning and visual language grounding are two important tasks for image understanding, but are seldom considered together. In this paper, we propose a Progressive Attention-Guided Network (PAGNet), which simultaneously generates image captions and predicts bounding boxes for caption words. PAGNet mainly has two distinctive properties: i) It can progressively refine the predictive results of image captioning, by updating the attention map with the predicted bounding boxes. ii) It learns bounding boxes of the words using a weakly supervised strategy, which combines the frameworks of Multiple Instance Learning (MIL) and Markov Decision Process (MDP). By using the attention map generated in the captioning process, PAGNet significantly reduces the search space of the MDP. We conduct experiments on benchmark datasets to demonstrate the effectiveness of PAGNet and results show that PAGNet achieves the best performance.

Downloads

Published

2019-07-17

How to Cite

Song, L., Liu, J., Qian, B., & Chen, Y. (2019). Connecting Language to Images: A Progressive Attention-Guided Network for Simultaneous Image Captioning and Language Grounding. Proceedings of the AAAI Conference on Artificial Intelligence, 33(01), 8885-8892. https://doi.org/10.1609/aaai.v33i01.33018885

Issue

Section

AAAI Technical Track: Vision