-
LLM Pruning and Distillation in Practice: The Minitron Approach
Paper • 2408.11796 • Published • 58 -
TableBench: A Comprehensive and Complex Benchmark for Table Question Answering
Paper • 2408.09174 • Published • 52 -
To Code, or Not To Code? Exploring Impact of Code in Pre-training
Paper • 2408.10914 • Published • 45 -
Open-FinLLMs: Open Multimodal Large Language Models for Financial Applications
Paper • 2408.11878 • Published • 63
Collections
Discover the best community collections!
Collections including paper arxiv:2602.00919
-
InterPrior: Scaling Generative Control for Physics-Based Human-Object Interactions
Paper • 2602.06035 • Published • 23 -
PaperBanana: Automating Academic Illustration for AI Scientists
Paper • 2601.23265 • Published • 192 -
Green-VLA: Staged Vision-Language-Action Model for Generalist Robots
Paper • 2602.00919 • Published • 284 -
Learning Humanoid End-Effector Control for Open-Vocabulary Visual Loco-Manipulation
Paper • 2602.16705 • Published • 26
-
THINKSAFE: Self-Generated Safety Alignment for Reasoning Models
Paper • 2601.23143 • Published • 38 -
PaperBanana: Automating Academic Illustration for AI Scientists
Paper • 2601.23265 • Published • 192 -
Agentic Reasoning for Large Language Models
Paper • 2601.12538 • Published • 195 -
BabyVision: Visual Reasoning Beyond Language
Paper • 2601.06521 • Published • 196
-
World-in-World: World Models in a Closed-Loop World
Paper • 2510.18135 • Published • 77 -
GigaBrain-0: A World Model-Powered Vision-Language-Action Model
Paper • 2510.19430 • Published • 52 -
World Simulation with Video Foundation Models for Physical AI
Paper • 2511.00062 • Published • 44 -
TwinBrainVLA: Unleashing the Potential of Generalist VLMs for Embodied Tasks via Asymmetric Mixture-of-Transformers
Paper • 2601.14133 • Published • 61
-
Gemini Robotics: Bringing AI into the Physical World
Paper • 2503.20020 • Published • 30 -
Magma: A Foundation Model for Multimodal AI Agents
Paper • 2502.13130 • Published • 58 -
LLaVA-Plus: Learning to Use Tools for Creating Multimodal Agents
Paper • 2311.05437 • Published • 51 -
OS-ATLAS: A Foundation Action Model for Generalist GUI Agents
Paper • 2410.23218 • Published • 49
-
FireGNN: Neuro-Symbolic Graph Neural Networks with Trainable Fuzzy Rules for Interpretable Medical Image Classification
Paper • 2509.10510 • Published -
From Pixels to Words -- Towards Native Vision-Language Primitives at Scale
Paper • 2510.14979 • Published • 67 -
Z-Image: An Efficient Image Generation Foundation Model with Single-Stream Diffusion Transformer
Paper • 2511.22699 • Published • 238 -
Self-Supervised Prompt Optimization
Paper • 2502.06855 • Published • 16
-
A Survey on Vision-Language-Action Models: An Action Tokenization Perspective
Paper • 2507.01925 • Published • 39 -
DreamVLA: A Vision-Language-Action Model Dreamed with Comprehensive World Knowledge
Paper • 2507.04447 • Published • 45 -
A Survey on Vision-Language-Action Models for Autonomous Driving
Paper • 2506.24044 • Published • 14 -
EmbRACE-3K: Embodied Reasoning and Action in Complex Environments
Paper • 2507.10548 • Published • 37
-
LLM Pruning and Distillation in Practice: The Minitron Approach
Paper • 2408.11796 • Published • 58 -
TableBench: A Comprehensive and Complex Benchmark for Table Question Answering
Paper • 2408.09174 • Published • 52 -
To Code, or Not To Code? Exploring Impact of Code in Pre-training
Paper • 2408.10914 • Published • 45 -
Open-FinLLMs: Open Multimodal Large Language Models for Financial Applications
Paper • 2408.11878 • Published • 63
-
InterPrior: Scaling Generative Control for Physics-Based Human-Object Interactions
Paper • 2602.06035 • Published • 23 -
PaperBanana: Automating Academic Illustration for AI Scientists
Paper • 2601.23265 • Published • 192 -
Green-VLA: Staged Vision-Language-Action Model for Generalist Robots
Paper • 2602.00919 • Published • 284 -
Learning Humanoid End-Effector Control for Open-Vocabulary Visual Loco-Manipulation
Paper • 2602.16705 • Published • 26
-
THINKSAFE: Self-Generated Safety Alignment for Reasoning Models
Paper • 2601.23143 • Published • 38 -
PaperBanana: Automating Academic Illustration for AI Scientists
Paper • 2601.23265 • Published • 192 -
Agentic Reasoning for Large Language Models
Paper • 2601.12538 • Published • 195 -
BabyVision: Visual Reasoning Beyond Language
Paper • 2601.06521 • Published • 196
-
FireGNN: Neuro-Symbolic Graph Neural Networks with Trainable Fuzzy Rules for Interpretable Medical Image Classification
Paper • 2509.10510 • Published -
From Pixels to Words -- Towards Native Vision-Language Primitives at Scale
Paper • 2510.14979 • Published • 67 -
Z-Image: An Efficient Image Generation Foundation Model with Single-Stream Diffusion Transformer
Paper • 2511.22699 • Published • 238 -
Self-Supervised Prompt Optimization
Paper • 2502.06855 • Published • 16
-
World-in-World: World Models in a Closed-Loop World
Paper • 2510.18135 • Published • 77 -
GigaBrain-0: A World Model-Powered Vision-Language-Action Model
Paper • 2510.19430 • Published • 52 -
World Simulation with Video Foundation Models for Physical AI
Paper • 2511.00062 • Published • 44 -
TwinBrainVLA: Unleashing the Potential of Generalist VLMs for Embodied Tasks via Asymmetric Mixture-of-Transformers
Paper • 2601.14133 • Published • 61
-
A Survey on Vision-Language-Action Models: An Action Tokenization Perspective
Paper • 2507.01925 • Published • 39 -
DreamVLA: A Vision-Language-Action Model Dreamed with Comprehensive World Knowledge
Paper • 2507.04447 • Published • 45 -
A Survey on Vision-Language-Action Models for Autonomous Driving
Paper • 2506.24044 • Published • 14 -
EmbRACE-3K: Embodied Reasoning and Action in Complex Environments
Paper • 2507.10548 • Published • 37
-
Gemini Robotics: Bringing AI into the Physical World
Paper • 2503.20020 • Published • 30 -
Magma: A Foundation Model for Multimodal AI Agents
Paper • 2502.13130 • Published • 58 -
LLaVA-Plus: Learning to Use Tools for Creating Multimodal Agents
Paper • 2311.05437 • Published • 51 -
OS-ATLAS: A Foundation Action Model for Generalist GUI Agents
Paper • 2410.23218 • Published • 49