UniRaG: Unification, Retrieval, and Generation for Multimodal Question Answering With Pre-Trained Language Models

Citation

Lim, Qi Zhi and Lee, Chin Poo and Lim, Kian Ming and Samingan, Ahmad Kamsani (2024) UniRaG: Unification, Retrieval, and Generation for Multimodal Question Answering With Pre-Trained Language Models. IEEE Access, 12. pp. 71505-71519. ISSN 2169-3536

[img] Text
18.pdf - Published Version
Restricted to Repository staff only

Download (1MB)

Abstract

Multimodal Question Answering (MMQA) has emerged as a challenging frontier at the intersection of natural language processing (NLP) and computer vision, demanding the integration of diverse modalities for effective comprehension and response. While pre-trained language models (PLMs) exhibit impressive performance across a range of NLP tasks, the investigation of text-based approaches to address MMQA represents a compelling and promising avenue for further research and advancement in the field. Although recent research has delved into text-based approaches for MMQA, the attained results have been unsatisfactory, which could be attributed to potential information loss during the knowledge transformation processes. In response, a novel three-stage framework named UniRaG is proposed for tackling MMQA, which encompasses unified knowledge representation, context retrieval, and answer generation. At the initial stage, advanced techniques are employed for unified knowledge representation, including LLaVA for image captioning and table linearization for tabular data, facilitating seamless integration of visual and tabular information into textual representation. For context retrieval, a cross-encoder trained on sequence classification is utilized to predict relevance scores for question-document pairs, and a top-k retrieval strategy is employed to retrieve the documents with the highest relevance scores as the contexts for answer generation. Finally, the answer generation stage is facilitated by a text-to-text PLM, Flan-T5-Base, which follows the encoder-decoder architecture with attention mechanisms. During this stage, uniform prefix conditioning is applied to the input text for enhanced adaptability and generalizability. Moreover, contextual diversity training is introduced to improve model robustness by including distractor documents as negative contexts during training. Experimental results on the MultimodalQA dataset demonstrate the superior performance of UniRaG, surpassing the existing state-of-the-art methods across all scenarios with 67.4% EM and 71.3% F1. Overall, UniRaG showcases robustness and reliability in MMQA, heralding significant advancements in multimodal comprehension and question answering research.

Item Type: Article
Uncontrolled Keywords: Computer vision, information retrieval, multimodal question answering, natural language processing, pre-trained language models, unified knowledge representation.
Subjects: Q Science > QA Mathematics > QA71-90 Instruments and machines > QA75.5-76.95 Electronic computers. Computer science
Divisions: Faculty of Information Science and Technology (FIST)
Depositing User: Ms Nurul Iqtiani Ahmad
Date Deposited: 03 Jul 2024 03:13
Last Modified: 03 Jul 2024 03:13
URII: http://shdl.mmu.edu.my/id/eprint/12593

Downloads

Downloads per month over past year

View ItemEdit (login required)