luodian commited on
Commit
4a0f69a
·
verified ·
1 Parent(s): 700a515

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +31 -0
README.md ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ ---
4
+ ## MMSearch-R1-7B
5
+
6
+ ### Introduction
7
+ MMSearch-R1-7B is a search-augmented LMM trained with end-to-end reinforcement learning, equipped with the ability to invoke multimodal search tools on demand. The model can dynamically decide whether to perform image or text search based on the question and integrate the retrieved external information into its reasoning process, enabling more accurate answers for knowledge-intensive VQA tasks. For more details on the training process and model evaluation, please refer to the [blog](https://www.lmms-lab.com/posts/mmsearch_r1/) or the [paper](https://arxiv.org/abs/2506.20670).
8
+
9
+ ### Model Details
10
+ - Model name: MMSearch-R1-7B
11
+ - Architecture: Qwen2.5-VL-7B base model, fine-tuned with Reinforcement Learning (GRPO)
12
+ - Model type: Multimodal Large Language Model with Search-Augmentation
13
+ - Languages: English(primary), multilingual(partially)
14
+ - License: Apache license 2.0
15
+ - Paper: [MMSearch-R1: Incentivizing LMMs to Search](https://arxiv.org/abs/2506.20670)
16
+ - Code: [EvolvingLMMs-Lab/multimodal-search-r1](https://github.com/EvolvingLMMs-Lab/multimodal-search-r1)
17
+
18
+ ### Training Details
19
+ - Dataset: [FVQA-train](https://huggingface.co/datasets/lmms-lab/FVQA)
20
+ - RL Framework: [veRL](https://github.com/volcengine/verl)
21
+ - GPUs: 32 * H100
22
+
23
+ ### Citation
24
+ ```
25
+ @article{wu2025mmsearch,
26
+ title={MMSearch-R1: Incentivizing LMMs to Search},
27
+ author={Wu, Jinming and Deng, Zihao and Li, Wei and Liu, Yiding and You, Bo and Li, Bo and Ma, Zejun and Liu, Ziwei},
28
+ journal={arXiv preprint arXiv:2506.20670},
29
+ year={2025}
30
+ }
31
+ ```