Cross-modal hybrid feature fusion for image-sentence matching

Xing Xu, Yifan Wang, Yixuan He, Yang Yang, Alan Hanjalic, Heng Tao Shen*

*Corresponding author for this work

Research output: Contribution to journalArticleScientificpeer-review

1 Citation (Scopus)

Abstract

Image-sentence matching is a challenging task in the field of language and vision, which aims at measuring the similarities between images and sentence descriptions. Most existing methods independently map the global features of images and sentences into a common space to calculate the image-sentence similarity. However, the image-sentence similarity obtained by these methods may be coarse as (1) an intermediate common space is introduced to implicitly match the heterogeneous features of images and sentences in a global level, and (2) only the inter-modality relations of images and sentences are captured while the intra-modality relations are ignored. To overcome the limitations, we propose a novel Cross-Modal Hybrid Feature Fusion (CMHF) framework for directly learning the image-sentence similarity by fusing multimodal features with inter- and intra-modality relations incorporated. It can robustly capture the high-level interactions between visual regions in images and words in sentences, where flexible attention mechanisms are utilized to generate effective attention flows within and across the modalities of images and sentences. A structured objective with ranking loss constraint is formed in CMHF to learn the image-sentence similarity based on the fused fine-grained features of different modalities bypassing the usage of intermediate common space. Extensive experiments and comprehensive analysis performed on two widely used datasets - Microsoft COCO and Flickr30K - show the effectiveness of the hybrid feature fusion framework in CMHF, in which the state-of-the-art matching performance is achieved by our proposed CMHF method.

Original languageEnglish
Article number3458281
JournalACM Transactions on Multimedia Computing, Communications and Applications
Volume17
Issue number4
DOIs
Publication statusPublished - 2021

Keywords

  • Attention mechanism
  • Cross-modal retrieval
  • Image-sentence matching
  • Multimodal feature fusion

Fingerprint

Dive into the research topics of 'Cross-modal hybrid feature fusion for image-sentence matching'. Together they form a unique fingerprint.

Cite this