AdaptLLM commited on
Commit
51af764
·
verified ·
1 Parent(s): c905096

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -2,7 +2,7 @@
2
  language:
3
  - en
4
  ---
5
- # Adapting Multimodal Large Language Models to Domains via Post-Training
6
 
7
  This project adapts general Multimodal Large Language Models (MLLMs) to specific domains like science and industry to improve their real-world use. It focuses on three main areas:
8
 
@@ -14,7 +14,7 @@ This project adapts general Multimodal Large Language Models (MLLMs) to specific
14
  - Instead of the usual two-step training (image-caption pairs first, then visual tasks), we use a **single-stage training** to handle more tasks for specific domains.
15
 
16
  ### 3. Task Evaluation
17
- - We test our method in important fields like **biomedicine, food, and remote sensing**.
18
  - We train and evaluate MLLMs on domain-specific tasks to show how well they perform.
19
 
20
 
@@ -44,10 +44,10 @@ This project adapts general Multimodal Large Language Models (MLLMs) to specific
44
  ## Citation
45
  If you find our work helpful, please cite us.
46
 
47
- [Adapt MLLM to Domains](https://huggingface.co/papers/2411.19930)
48
  ```bibtex
49
  @article{adamllm,
50
- title={On Domain-Specific Post-Training for Multimodal Large Language Models},
51
  author={Cheng, Daixuan and Huang, Shaohan and Zhu, Ziyu and Zhang, Xintong and Zhao, Wayne Xin and Luan, Zhongzhi and Dai, Bo and Zhang, Zhenliang},
52
  journal={arXiv preprint arXiv:2411.19930},
53
  year={2024}
 
2
  language:
3
  - en
4
  ---
5
+ # Adapting Multimodal Large Language Models to Domains via Post-Training (EMNLP 2025)
6
 
7
  This project adapts general Multimodal Large Language Models (MLLMs) to specific domains like science and industry to improve their real-world use. It focuses on three main areas:
8
 
 
14
  - Instead of the usual two-step training (image-caption pairs first, then visual tasks), we use a **single-stage training** to handle more tasks for specific domains.
15
 
16
  ### 3. Task Evaluation
17
+ - We test our method in high-impact domains like **biomedicine, food, and remote sensing**.
18
  - We train and evaluate MLLMs on domain-specific tasks to show how well they perform.
19
 
20
 
 
44
  ## Citation
45
  If you find our work helpful, please cite us.
46
 
47
+ [Adapt MLLM to Domains](https://huggingface.co/papers/2411.19930) (EMNLP 2025 Findings)
48
  ```bibtex
49
  @article{adamllm,
50
+ title={On Domain-Adaptive Post-Training for Multimodal Large Language Models},
51
  author={Cheng, Daixuan and Huang, Shaohan and Zhu, Ziyu and Zhang, Xintong and Zhao, Wayne Xin and Luan, Zhongzhi and Dai, Bo and Zhang, Zhenliang},
52
  journal={arXiv preprint arXiv:2411.19930},
53
  year={2024}