Sorry, you need to enable JavaScript to visit this website.

VISUAL RELATIONSHIP RECOGNITION VIA LANGUAGE AND POSITION GUIDED ATTENTION

Citation Author(s):
Hao Zhou,Chuanping Hu,Chongyang Zhang,Shengyang Shen
Submitted by:
Hao Zhou
Last updated:
9 May 2019 - 6:05am
Document Type:
Poster
Event:
 

Visual relationship recognition, as a challenging task used to distinguish the interactions between object pairs, has received much attention recently. Considering the fact that most visual relationships are semantic concepts defined by human beings, there are many human knowledge, or priors, hidden in them, which haven’t been fully exploited by existing methods. In this work, we propose a novel visual relationship recognition model using language and position guided attention: language and position information are exploited and vectored firstly, and then both of them are used to guide the generation of attention maps. With the guided attention, the hidden human knowledge can be made better use to enhance the selection of spatial and channel features. Experiments on VRD and VGR show that, with language and position guided attention module, our proposed model achieves state-of-the-art performance.

https://ieeexplore.ieee.org/document/8683464

up
0 users have voted: