Abstract
Unified multimodal models suffer from inefficiencies in certain tasks, leading to the proposal of Mixture-of-Experts Adaptation to improve compression and maintain performance.
Large multimodal models have achieved remarkable progress in both understanding and generation. Recent efforts pursue unified multimodal models that integrate heterogeneous components to support both capabilities within a single framework. However, such unification introduces inference inefficiencies, e.g., specific tasks or samples may not require the full knowledge or capacity of the unified model. Yet, a systematic understanding of how these inefficiencies manifest across different components remains limited. In this work, we first conduct a systematic analysis of unified multimodal model components using training-free pruning as a probing methodology, considering both depth pruning and width reduction. Our study reveals that the understanding component exhibits notable compressibility in both understanding and generation tasks, which is more pronounced in the latter. In contrast, the generation components are highly sensitive to compression, with performance deteriorating sharply even under moderate compression ratios. To address this limitation, we propose the Mixture-of-Experts (MoE) Adaptation, inspired by the dynamic activation patterns observed across different samples. This approach partitions the generation module into multiple experts and enables sparse activation to restore generation quality. We validate the effectiveness of sparse activation through expert-frozen tuning and further demonstrate that a fully trainable adaptation delivers additional gains. As a result, the adapted BAGEL model achieves performance comparable to the full model while activating only about half of its parameters. The code is released at https://github.com/Shwai-He/SparseUnifiedModel{this link}.
Community
Large multimodal models have achieved remarkable progress in both
understanding and generation. Recent efforts pursue unified multimodal models
that integrate heterogeneous components to support both capabilities within a
single framework. However, such unification introduces inference
inefficiencies, e.g., specific tasks or samples may not require the full
knowledge or capacity of the unified model. Yet, a systematic understanding of
how these inefficiencies manifest across different components remains limited.
In this work, we first conduct a systematic analysis of unified multimodal
model components using training-free pruning as a probing methodology,
considering both depth pruning and width reduction. Our study reveals that the
understanding component exhibits notable compressibility in both understanding
and generation tasks, which is more pronounced in the latter. In contrast, the
generation components are highly sensitive to compression, with performance
deteriorating sharply even under moderate compression ratios. To address this
limitation, we propose the Mixture-of-Experts (MoE) Adaptation, inspired by the
dynamic activation patterns observed across different samples. This approach
partitions the generation module into multiple experts and enables sparse
activation to restore generation quality. We validate the effectiveness of
sparse activation through expert-frozen tuning and further demonstrate that a
fully trainable adaptation delivers additional gains. As a result, the adapted
BAGEL model achieves performance comparable to the full model while activating
only about half of its parameters. The code is released at
\href{https://github.com/Shwai-He/SparseUnifiedModel}{this link}.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Mixture of Ranks with Degradation-Aware Routing for One-Step Real-World Image Super-Resolution (2025)
- HBridge: H-Shape Bridging of Heterogeneous Experts for Unified Multimodal Understanding and Generation (2025)
- MoDES: Accelerating Mixture-of-Experts Multimodal Large Language Models via Dynamic Expert Skipping (2025)
- Architecture Decoupling Is Not All You Need For Unified Multimodal Model (2025)
- Ming-UniVision: Joint Image Understanding and Generation with a Unified Continuous Tokenizer (2025)
- Can Visual Input Be Compressed? A Visual Token Compression Benchmark for Large Multimodal Models (2025)
- A Comprehensive Study on Visual Token Redundancy for Discrete Diffusion-based Multimodal Large Language Models (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper