Sorry, you need to enable JavaScript to visit this website.

Image Generation is MAY All You Need for VQA

Citation Author(s):
Kyungho Kim, Junseo Lee, Jihwa Lee
Submitted by:
Kyungho Kim
Last updated:
19 May 2023 - 1:59am
Document Type:
Poster
Document Year:
2023
Event:
Presenters:
Kyungho Kim
Paper Code:
IVMSP-P9.11
 

Visual Question Answering (VQA) stands to benefit from the boost of increasingly sophisticated Pretrained Language Model (PLM) and Computer Vision-based models. In particular, many language modality studies have been conducted using image captioning or question generation with the knowledge ground of PLM in terms of data augmentation. However, image generation of VQA has been implemented in a limited way to modify only certain parts of the original image in order to control the quality and uncertainty. In this paper, to address this gap, we propose a method that utilizes the diffusion model, pre-trained with various tasks and images, to inject the prior knowledge base into generated images and secure diversity without losing generality about the answer. In addition, we design an effective training strategy by considering the difficulty of questions to address the multiple images per QA pair and to compensate for the weakness of the diffusion model. VQA model trained on our strategy improves significant performance on the dataset that requires factual knowledge without any knowledge information in language modality.

up
0 users have voted: