1School of Computer Science and Technology, Harbin Institute of Technology, Shenzhen, China
2School of Computer Science, University of Sydney, Sydney, Australia
3Security Department, Alibaba Group, Hangzhou, China
βCorresponding authors
Multimodal Large Language Models (MLLMs) have recently received substantial interest, which shows their emerging potential as general-purpose models for various vision-language tasks. MLLMs involve significant external knowledge within their parameters; however, it is challenging to continually update these models with the latest knowledge, which involves huge computational costs and poor interpretability. Retrieval augmentation techniques have proven to be effective plugins for both LLMs and MLLMs.
In this study, we propose multimodal adaptive Retrieval-Augmented Bootstrapping Language-Image Pre-training (RA-BLIP), a novel retrieval-augmented framework for various MLLMs. We first leverage the question to instruct the extraction of visual information through interactions with one set of learnable queries, minimizing irrelevant interference and redundancy during retrieval and generation. Besides, we introduce a pre-trained multimodal adaptive fusion module to achieve question text-to-multimodal retrieval and integration of multimodal knowledge by projecting visual and language modalities into a unified semantic space. Furthermore, we present an Adaptive Selection Knowledge Generation (ASKG) strategy to train the generator to autonomously discern the relevance of retrieved knowledge, which realizes excellent denoising performance. Extensive experiments on open multimodal question-answering datasets demonstrate that RA-BLIP achieves significant performance and surpasses the state-of-the-art retrieval-augmented models.
Overview of RA-BLIP: 1. Query-Instructed Visual Extraction: We leverage the input question to guide the extraction of relevant visual features using learnable queries. This drastically reduces visual redundancy and irrelevant interference. 2. Multimodal Adaptive Fusion: A pre-trained fusion module projects both visual and textual modalities into a unified semantic space, enabling seamless text-to-multimodal retrieval. 3. Adaptive Selection Knowledge Generation (ASKG): The generator is trained to autonomously evaluate and filter retrieved knowledge, providing highly effective denoising before the final answer generation.
We recommend setting up the environment using conda:
# Create conda environment
conda create -n rablip python=3.9
conda activate rablip
# Install PyTorch
pip install torch torchvision torchaudio --index-url [https://download.pytorch.org/whl/cu118](https://download.pytorch.org/whl/cu118)
# Install required packages
git clone [https://github.com/YourOrg/RA-BLIP.git](https://github.com/YourOrg/RA-BLIP.git)
cd RA-BLIP
pip install -r requirements.txt
### Data Preparation
Download the pre-training datasets and downstream Multimodal QA datasets (e.g., OK-VQA, A-OKVQA). Organize them in the `./data` directory as follows:
```text
RA-BLIP/
βββ data/
β βββ pretrain/
β βββ okvqa/
β βββ aokvqa/Download the base MLLM weights and the pre-trained RA-BLIP retrieval/fusion modules from our Huggingface repository and place them in the ./checkpoints folder.
To run the Bootstrapping Language-Image Pre-training with retrieval augmentation:
python run_pretrain.py \
--config ./configs/pretrain.yaml \
--output_dir ./output/pretrainExtensive experiments demonstrate that RA-BLIP achieves state-of-the-art performance on open multimodal question-answering datasets, outperforming existing retrieval-augmented baseline models.
Table 1: Performance comparison of RA-BLIP against state-of-the-art retrieval-augmented models.
If you find this work useful for your research, please kindly cite our TMM 2025 paper:
@article{rablip2025,
title={RA-BLIP: Multimodal Adaptive Retrieval-Augmented Bootstrapping Language-Image Pre-training},
author={Ding, Muhe and Ma, Yang and Qin, Pengda and Wu, Jianlong and Li, Yuhong and Nie, Liqiang},
journal={IEEE Transactions on Multimedia (TMM)},
year={2025},
url={[https://ieeexplore.ieee.org/document/10844992](https://ieeexplore.ieee.org/document/10844992)}
}

