CLIP4MC: An RL-Friendly Vision-Language Model for Minecraft Paper • 2303.10571 • Published Mar 19, 2023
Unified Multimodal Understanding via Byte-Pair Visual Encoding Paper • 2506.23639 • Published Jun 30, 2025 • 4
Being-H0: Vision-Language-Action Pretraining from Large-Scale Human Videos Paper • 2507.15597 • Published Jul 21, 2025 • 34
DiG-Flow: Discrepancy-Guided Flow Matching for Robust VLA Models Paper • 2512.01715 • Published Dec 1, 2025 • 9
Spatial-Aware VLA Pretraining through Visual-Physical Alignment from Human Videos Paper • 2512.13080 • Published Dec 15, 2025 • 16
Being-H0.5: Scaling Human-Centric Robot Learning for Cross-Embodiment Generalization Paper • 2601.12993 • Published 2 days ago • 59