Update README.md
Browse files
README.md
CHANGED
|
@@ -2,7 +2,7 @@
|
|
| 2 |
language:
|
| 3 |
- en
|
| 4 |
---
|
| 5 |
-
# Adapting Multimodal Large Language Models to Domains via Post-Training
|
| 6 |
|
| 7 |
This project adapts general Multimodal Large Language Models (MLLMs) to specific domains like science and industry to improve their real-world use. It focuses on three main areas:
|
| 8 |
|
|
@@ -14,7 +14,7 @@ This project adapts general Multimodal Large Language Models (MLLMs) to specific
|
|
| 14 |
- Instead of the usual two-step training (image-caption pairs first, then visual tasks), we use a **single-stage training** to handle more tasks for specific domains.
|
| 15 |
|
| 16 |
### 3. Task Evaluation
|
| 17 |
-
- We test our method in
|
| 18 |
- We train and evaluate MLLMs on domain-specific tasks to show how well they perform.
|
| 19 |
|
| 20 |
|
|
@@ -44,10 +44,10 @@ This project adapts general Multimodal Large Language Models (MLLMs) to specific
|
|
| 44 |
## Citation
|
| 45 |
If you find our work helpful, please cite us.
|
| 46 |
|
| 47 |
-
[Adapt MLLM to Domains](https://huggingface.co/papers/2411.19930)
|
| 48 |
```bibtex
|
| 49 |
@article{adamllm,
|
| 50 |
-
title={On Domain-
|
| 51 |
author={Cheng, Daixuan and Huang, Shaohan and Zhu, Ziyu and Zhang, Xintong and Zhao, Wayne Xin and Luan, Zhongzhi and Dai, Bo and Zhang, Zhenliang},
|
| 52 |
journal={arXiv preprint arXiv:2411.19930},
|
| 53 |
year={2024}
|
|
|
|
| 2 |
language:
|
| 3 |
- en
|
| 4 |
---
|
| 5 |
+
# Adapting Multimodal Large Language Models to Domains via Post-Training (EMNLP 2025)
|
| 6 |
|
| 7 |
This project adapts general Multimodal Large Language Models (MLLMs) to specific domains like science and industry to improve their real-world use. It focuses on three main areas:
|
| 8 |
|
|
|
|
| 14 |
- Instead of the usual two-step training (image-caption pairs first, then visual tasks), we use a **single-stage training** to handle more tasks for specific domains.
|
| 15 |
|
| 16 |
### 3. Task Evaluation
|
| 17 |
+
- We test our method in high-impact domains like **biomedicine, food, and remote sensing**.
|
| 18 |
- We train and evaluate MLLMs on domain-specific tasks to show how well they perform.
|
| 19 |
|
| 20 |
|
|
|
|
| 44 |
## Citation
|
| 45 |
If you find our work helpful, please cite us.
|
| 46 |
|
| 47 |
+
[Adapt MLLM to Domains](https://huggingface.co/papers/2411.19930) (EMNLP 2025 Findings)
|
| 48 |
```bibtex
|
| 49 |
@article{adamllm,
|
| 50 |
+
title={On Domain-Adaptive Post-Training for Multimodal Large Language Models},
|
| 51 |
author={Cheng, Daixuan and Huang, Shaohan and Zhu, Ziyu and Zhang, Xintong and Zhao, Wayne Xin and Luan, Zhongzhi and Dai, Bo and Zhang, Zhenliang},
|
| 52 |
journal={arXiv preprint arXiv:2411.19930},
|
| 53 |
year={2024}
|