| rank,name,times_trended,best_rank,avg_rank,median_rank,publish_date,max_upvotes,max_github_stars,arxiv_link | |
| 1,LlamaFactory: Unified Efficient Fine-Tuning of 100+ Language Models,432,2,11.28,11,"Mar 20, 2024",173,63300,https://arxiv.org/abs/2403.13372 | |
| 2,DINOv3,346,1,21.02,15,"Aug 13, 2025",284,8590,https://arxiv.org/abs/2508.10104 | |
| 3,Agent Lightning: Train ANY AI Agents with Reinforcement Learning,325,1,20.18,20,"Aug 5, 2025",120,9090,https://arxiv.org/abs/2508.03680 | |
| 4,WebDancer: Towards Autonomous Information Seeking Agency,398,3,25.85,26,"May 28, 2025",33,17300,https://arxiv.org/abs/2505.22648 | |
| 5,"WebShaper: Agentically Data Synthesizing via Information-Seeking | |
| Formalization",402,2,26.34,26,"Jul 20, 2025",60,17300,https://arxiv.org/abs/2507.15061 | |
| 6,WebWatcher: Breaking New Frontier of Vision-Language Deep Research Agent,393,1,25.86,26,"Aug 7, 2025",139,17300,https://arxiv.org/abs/2508.05748 | |
| 7,WebSailor: Navigating Super-human Reasoning for Web Agent,394,2,26.03,26,"Jul 3, 2025",122,17300,https://arxiv.org/abs/2507.02592 | |
| 8,"InternVL3: Exploring Advanced Training and Test-Time Recipes for | |
| Open-Source Multimodal Models",349,8,23.07,21,"Apr 14, 2025",304,8850,https://arxiv.org/abs/2504.10479 | |
| 9,"AgentScope 1.0: A Developer-Centric Framework for Building Agentic | |
| Applications",290,1,18.33,14,"Aug 22, 2025",53,14100,https://arxiv.org/abs/2508.16279 | |
| 10,Qwen-Image Technical Report,302,1,20.35,19,"Aug 4, 2025",261,6150,https://arxiv.org/abs/2508.02324 | |
| 11,"MinerU2.5: A Decoupled Vision-Language Model for Efficient | |
| High-Resolution Document Parsing",179,2,10.82,9,"Sep 26, 2025",134,49600,https://arxiv.org/abs/2509.22186 | |
| 12,Scaling Agents via Continual Pre-training,196,1,18.28,17,"Sep 16, 2025",115,17300,https://arxiv.org/abs/2509.13310 | |
| 13,"Easy Dataset: A Unified and Extensible Framework for Synthesizing LLM | |
| Fine-Tuning Data from Unstructured Documents",345,13,32.46,32,"Jul 5, 2025",51,12100,https://arxiv.org/abs/2507.04009 | |
| 14,"WebSailor-V2: Bridging the Chasm to Proprietary Agents via Synthetic | |
| Data and Scalable Reinforcement Learning",201,1,19.4,17,"Sep 16, 2025",90,17300,https://arxiv.org/abs/2509.13305 | |
| 15,"ReSum: Unlocking Long-Horizon Search Intelligence via Context | |
| Summarization",198,1,18.98,16,"Sep 16, 2025",78,17300,https://arxiv.org/abs/2509.13313 | |
| 16,"PC-Agent: A Hierarchical Multi-Agent Collaboration Framework for Complex | |
| Task Automation on PC",222,3,22.85,18,"Feb 20, 2025",29,6440,https://arxiv.org/abs/2502.14282 | |
| 17,"WebWeaver: Structuring Web-Scale Evidence with Dynamic Outlines for | |
| Open-Ended Deep Research",194,4,19,16,"Sep 16, 2025",105,17300,https://arxiv.org/abs/2509.13312 | |
| 18,"Look Before You Leap: A GUI-Critic-R1 Model for Pre-Operative Error | |
| Diagnosis in GUI Automation",220,4,22.92,19,"Jun 5, 2025",19,6430,https://arxiv.org/abs/2506.04614 | |
| 19,"PaddleOCR-VL: Boosting Multilingual Document Parsing via a 0.9B | |
| Ultra-Compact Vision-Language Model",124,1,3.35,3,"Oct 16, 2025",98,65400,https://arxiv.org/abs/2510.14528 | |
| 20,Mobile-Agent-v3: Foundamental Agents for GUI Automation,198,3,21.24,17,"Aug 21, 2025",64,6440,https://arxiv.org/abs/2508.15144 | |
| 21,AgentFly: Fine-tuning LLM Agents without Fine-tuning LLMs,144,1,12.94,8,"Aug 22, 2025",153,1740,https://arxiv.org/abs/2508.16153 | |
| 22,FastVLM: Efficient Vision Encoding for Vision Language Models,162,1,17.6,11,"Dec 17, 2024",70,6680,https://arxiv.org/abs/2412.13303 | |
| 23,MIRIX: Multi-Agent Memory System for LLM-Based Agents,272,5,31.7,35,"Jul 10, 2025",79,3380,https://arxiv.org/abs/2507.07957 | |
| 24,"GLM-4.5: Agentic, Reasoning, and Coding (ARC) Foundation Models",243,11,30.1,30,"Aug 8, 2025",185,3000,https://arxiv.org/abs/2508.06471 | |
| 25,Prompt Orchestration Markup Language,153,3,18.99,12,"Aug 19, 2025",49,4560,https://arxiv.org/abs/2508.13948 | |
| 26,OmniFlatten: An End-to-end GPT Model for Seamless Voice Conversation,112,3,9.8,9,"Oct 23, 2024",5,50300,https://arxiv.org/abs/2410.17799 | |
| 27,UI-TARS: Pioneering Automated GUI Interaction with Native Agents,208,4,29.12,28,"Jan 21, 2025",65,7830,https://arxiv.org/abs/2501.12326 | |
| 28,"A Comprehensive Survey of Self-Evolving AI Agents: A New Paradigm | |
| Bridging Foundation Models and Lifelong Agentic Systems",191,5,27.24,29,"Aug 10, 2025",97,1140,https://arxiv.org/abs/2508.07407 | |
| 29,Qwen3 Technical Report,266,2,34.64,34,"May 14, 2025",316,25500,https://arxiv.org/abs/2505.09388 | |
| 30,The Landscape of Agentic Reinforcement Learning for LLMs: A Survey,125,2,17.14,15,"Sep 2, 2025",214,941,https://arxiv.org/abs/2509.02547 | |
| 31,DeepAnalyze: Agentic Large Language Models for Autonomous Data Science,115,2,14.33,14,"Oct 19, 2025",102,2630,https://arxiv.org/abs/2510.16872 | |
| 32,TradingAgents: Multi-Agents LLM Financial Trading Framework,111,6,13.67,13,"Dec 28, 2024",14,25800,https://arxiv.org/abs/2412.20138 | |
| 33,MinerU: An Open-Source Solution for Precise Document Content Extraction,112,3,14.65,16,"Sep 27, 2024",32,49600,https://arxiv.org/abs/2409.18839 | |
| 34,"IndexTTS: An Industrial-Level Controllable and Efficient Zero-Shot | |
| Text-To-Speech System",112,7,14.97,14,"Feb 8, 2025",5,16000,https://arxiv.org/abs/2502.05512 | |
| 35,StableAvatar: Infinite-Length Audio-Driven Avatar Video Generation,137,1,21.63,20,"Aug 11, 2025",27,1020,https://arxiv.org/abs/2508.08248 | |
| 36,"Paper2Agent: Reimagining Research Papers As Interactive and Reliable AI | |
| Agents",135,5,22.01,20,"Sep 8, 2025",41,1660,https://arxiv.org/abs/2509.06917 | |
| 37,PaddleOCR 3.0 Technical Report,100,1,12.08,11,"Jul 8, 2025",17,57600,https://arxiv.org/abs/2507.05595 | |
| 38,"Seeing, Listening, Remembering, and Reasoning: A Multimodal Agent with | |
| Long-Term Memory",121,4,18.92,14,"Aug 13, 2025",54,893,https://arxiv.org/abs/2508.09736 | |
| 39,rStar2-Agent: Agentic Reasoning Technical Report,112,4,16.78,16,"Aug 28, 2025",106,1230,https://arxiv.org/abs/2508.20722 | |
| 40,Mem0: Building Production-Ready AI Agents with Scalable Long-Term Memory,112,6,17.88,15,"Apr 28, 2025",33,43700,https://arxiv.org/abs/2504.19413 | |
| 41,"ScreenCoder: Advancing Visual-to-Code Generation for Front-End | |
| Automation via Modular Multimodal Agents",122,1,21.56,24,"Jul 30, 2025",98,2290,https://arxiv.org/abs/2507.22827 | |
| 42,Step-Audio 2 Technical Report,94,2,13.01,12,"Jul 22, 2025",71,1050,https://arxiv.org/abs/2507.16632 | |
| 43,"USO: Unified Style and Subject-Driven Generation via Disentangled and | |
| Reward Learning",92,2,12.67,6,"Aug 26, 2025",56,1080,https://arxiv.org/abs/2508.18966 | |
| 44,TempFlow-GRPO: When Timing Matters for GRPO in Flow Models,96,4,14.52,10,"Aug 6, 2025",12,782,https://arxiv.org/abs/2508.04324 | |
| 45,RAG-Anything: All-in-One RAG Framework,129,2,25.02,28,"Oct 14, 2025",49,10600,https://arxiv.org/abs/2510.12323 | |
| 46,"ComfyUI-Copilot: An Intelligent Assistant for Automated Workflow | |
| Development",172,6,32.85,36,"Jun 5, 2025",79,3660,https://arxiv.org/abs/2506.05010 | |
| 47,VGGT: Visual Geometry Grounded Transformer,232,12,37.63,39,"Mar 14, 2025",33,11700,https://arxiv.org/abs/2503.11651 | |
| 48,MixGRPO: Unlocking Flow-based GRPO Efficiency with Mixed ODE-SDE,170,12,32.78,34,"Jul 29, 2025",15,1010,https://arxiv.org/abs/2507.21802 | |
| 49,4DNeX: Feed-Forward 4D Generative Modeling Made Easy,178,2,33.96,37,"Aug 18, 2025",61,751,https://arxiv.org/abs/2508.13154 | |
| 50,"FAPO: Flawed-Aware Policy Optimization for Efficient and Reliable | |
| Reasoning",86,3,15.94,11,"Oct 26, 2025",10,17000,https://arxiv.org/abs/2510.22543 | |
| 51,"Agent S2: A Compositional Generalist-Specialist Framework for Computer | |
| Use Agents",145,2,30.46,35,"Apr 1, 2025",26,8450,https://arxiv.org/abs/2504.00906 | |
| 52,The Unreasonable Effectiveness of Scaling Agents for Computer Use,138,2,29.52,34,"Oct 2, 2025",24,8450,https://arxiv.org/abs/2510.02250 | |
| 53,"VLA-Adapter: An Effective Paradigm for Tiny-Scale Vision-Language-Action | |
| Model",131,7,28.43,29,"Sep 11, 2025",235,1590,https://arxiv.org/abs/2509.09372 | |
| 54,A Survey of Reinforcement Learning for Large Reasoning Models,110,4,24.15,24,"Sep 10, 2025",183,1800,https://arxiv.org/abs/2509.08827 | |
| 55,Less is More: Recursive Reasoning with Tiny Networks,90,1,20.33,18,"Oct 6, 2025",483,5670,https://arxiv.org/abs/2510.04871 | |
| 56,LightRAG: Simple and Fast Retrieval-Augmented Generation,92,2,21.25,9,"Oct 8, 2024",20,24900,https://arxiv.org/abs/2410.05779 | |
| 57,PyTorch FSDP: Experiences on Scaling Fully Sharded Data Parallel,104,10,27.97,30,"Apr 21, 2023",4,95500,https://arxiv.org/abs/2304.11277 | |
| 58,Matrix-Game: Interactive World Foundation Model,78,1,20.45,14,"Jun 23, 2025",72,1550,https://arxiv.org/abs/2506.18701 | |
| 59,PyTorch Distributed: Experiences on Accelerating Data Parallel Training,105,9,28.43,31,"Jun 28, 2020",3,95500,https://arxiv.org/abs/2006.15704 | |
| 60,Thyme: Think Beyond Images,82,1,23.55,23,"Aug 15, 2025",78,466,https://arxiv.org/abs/2508.11630 | |
| 61,"RepoMaster: Autonomous Exploration and Understanding of GitHub | |
| Repositories for Complex Task Solving",83,14,23.9,21,"May 27, 2025",2,348,https://arxiv.org/abs/2505.21577 | |
| 62,UI-Venus Technical Report: Building High-performance UI Agents with RFT,74,7,21.26,17,"Aug 14, 2025",41,470,https://arxiv.org/abs/2508.10833 | |
| 63,"ToonComposer: Streamlining Cartoon Production with Generative | |
| Post-Keyframing",65,4,17.37,14,"Aug 14, 2025",50,358,https://arxiv.org/abs/2508.10881 | |
| 64,OpenCUA: Open Foundations for Computer-Use Agents,85,14,25.86,24,"Aug 12, 2025",30,415,https://arxiv.org/abs/2508.09123 | |
| 65,LongSplat: Robust Unposed 3D Gaussian Splatting for Casual Long Videos,59,3,14.98,10,"Aug 19, 2025",59,600,https://arxiv.org/abs/2508.14041 | |
| 66,"Stand-In: A Lightweight and Plug-and-Play Identity Control for Video | |
| Generation",63,2,17.37,9,"Aug 11, 2025",38,512,https://arxiv.org/abs/2508.07901 | |
| 67,Zep: A Temporal Knowledge Graph Architecture for Agent Memory,95,11,28.88,31,"Jan 20, 2025",6,20600,https://arxiv.org/abs/2501.13956 | |
| 68,Paper2Video: Automatic Video Generation from Scientific Papers,55,2,13.04,11,"Oct 6, 2025",115,1790,https://arxiv.org/abs/2510.05096 | |
| 69,"Youtu-GraphRAG: Vertically Unified Agents for Graph Retrieval-Augmented | |
| Complex Reasoning",72,6,22.89,17,"Aug 27, 2025",7,730,https://arxiv.org/abs/2508.19855 | |
| 70,"VeOmni: Scaling Any Modality Model Training with Model-Centric | |
| Distributed Recipe Zoo",66,1,20.94,18,"Aug 4, 2025",17,1180,https://arxiv.org/abs/2508.02317 | |
| 71,"NextStep-1: Toward Autoregressive Image Generation with Continuous | |
| Tokens at Scale",52,1,12.92,5,"Aug 14, 2025",134,496,https://arxiv.org/abs/2508.10711 | |
| 72,Qwen3-Omni Technical Report,72,1,23.96,26,"Sep 22, 2025",128,2690,https://arxiv.org/abs/2509.17765 | |
| 73,3D and 4D World Modeling: A Survey,66,4,22.12,20,"Sep 4, 2025",57,568,https://arxiv.org/abs/2509.07996 | |
| 74,"ScaleCUA: Scaling Open-Source Computer Use Agents with Cross-Platform | |
| Data",78,10,27.24,29,"Sep 18, 2025",105,610,https://arxiv.org/abs/2509.15221 | |
| 75,VibeVoice Technical Report,37,1,1,1,"Aug 26, 2025",118,7980,https://arxiv.org/abs/2508.19205 | |
| 76,"PokeeResearch: Effective Deep Research via Reinforcement Learning from | |
| AI Feedback and Robust Reasoning Scaffold",47,4,12.15,7,"Oct 17, 2025",8,1610,https://arxiv.org/abs/2510.15862 | |
| 77,"MCP-Bench: Benchmarking Tool-Using LLM Agents with Complex Real-World | |
| Tasks via MCP Servers",69,5,25.25,23,"Aug 28, 2025",58,296,https://arxiv.org/abs/2508.20453 | |
| 78,Depth Anything 3: Recovering the Visual Space from Any Views,40,1,6.58,3,"Nov 13, 2025",89,3010,https://arxiv.org/abs/2511.10647 | |
| 79,"MCP-Universe: Benchmarking Large Language Models with Real-World Model | |
| Context Protocol Servers",57,5,20.07,18,"Aug 20, 2025",41,352,https://arxiv.org/abs/2508.14704 | |
| 80,HunyuanImage 3.0 Technical Report,50,1,15.78,7,"Sep 28, 2025",21,2210,https://arxiv.org/abs/2509.23951 | |
| 81,"olmOCR: Unlocking Trillions of Tokens in PDFs with Vision Language | |
| Models",56,3,19.68,20,"Feb 25, 2025",9,16000,https://arxiv.org/abs/2502.18443 | |
| 82,Ovi: Twin Backbone Cross-Modal Fusion for Audio-Video Generation,68,2,25.31,22,"Sep 30, 2025",32,1300,https://arxiv.org/abs/2510.01284 | |
| 83,Pico-Banana-400K: A Large-Scale Dataset for Text-Guided Image Editing,54,1,18.81,9,"Oct 22, 2025",28,1580,https://arxiv.org/abs/2510.19808 | |
| 84,A Survey of Context Engineering for Large Language Models,151,1,39.59,40,"Jul 17, 2025",256,2330,https://arxiv.org/abs/2507.13334 | |
| 85,"HuMo: Human-Centric Video Generation via Collaborative Multi-Modal | |
| Conditioning",46,2,14.24,9,"Sep 10, 2025",120,596,https://arxiv.org/abs/2509.08519 | |
| 86,"EmbodiedOneVision: Interleaved Vision-Text-Action Pretraining for | |
| General Robot Control",52,2,20.06,15,"Aug 28, 2025",76,282,https://arxiv.org/abs/2508.21112 | |
| 87,Waver: Wave Your Way to Lifelike Video Generation,53,5,21.06,22,"Aug 21, 2025",33,499,https://arxiv.org/abs/2508.15761 | |
| 88,"AudioStory: Generating Long-Form Narrative Audio with Large Language | |
| Models",59,16,24.81,21,"Aug 27, 2025",20,264,https://arxiv.org/abs/2508.20088 | |
| 89,Towards a Unified View of Large Language Model Post-Training,46,5,17.63,19,"Sep 4, 2025",61,100,https://arxiv.org/abs/2509.04419 | |
| 90,"InternVL3.5: Advancing Open-Source Multimodal Models in Versatility, | |
| Reasoning, and Efficiency",109,12,37.03,40,"Aug 25, 2025",191,9240,https://arxiv.org/abs/2508.18265 | |
| 91,Transition Models: Rethinking the Generative Learning Objective,47,8,19.38,19,"Sep 4, 2025",28,94,https://arxiv.org/abs/2509.04394 | |
| 92,"AgentGym-RL: Training LLM Agents for Long-Horizon Decision Making | |
| through Multi-Turn Reinforcement Learning",43,1,16.98,8,"Sep 10, 2025",56,362,https://arxiv.org/abs/2509.08755 | |
| 93,"The Dragon Hatchling: The Missing Link between the Transformer and | |
| Models of the Brain",38,1,13.08,1,"Sep 30, 2025",489,3120,https://arxiv.org/abs/2509.26507 | |
| 94,STream3R: Scalable Sequential 3D Reconstruction with Causal Transformer,39,5,14.41,8,"Aug 14, 2025",30,194,https://arxiv.org/abs/2508.10893 | |
| 95,SimpleVLA-RL: Scaling VLA Training via Reinforcement Learning,56,1,25.86,28,"Sep 11, 2025",73,740,https://arxiv.org/abs/2509.09674 | |
| 96,"Enterprise Deep Research: Steerable Multi-Agent Deep Research for | |
| Enterprise Analytics",46,7,21.17,18,"Oct 20, 2025",9,952,https://arxiv.org/abs/2510.17797 | |
| 97,Code2Video: A Code-centric Paradigm for Educational Video Generation,48,2,22.77,19,"Oct 1, 2025",33,1100,https://arxiv.org/abs/2510.01174 | |
| 98,RynnEC: Bringing MLLMs into Embodied World,62,6,29.16,29,"Aug 19, 2025",18,337,https://arxiv.org/abs/2508.14160 | |
| 99,"Kimi Linear: An Expressive, Efficient Attention Architecture",39,1,16.28,4,"Oct 30, 2025",102,1150,https://arxiv.org/abs/2510.26692 | |
| 100,UI-S1: Advancing GUI Automation via Semi-online Reinforcement Learning,67,15,30.96,31,"Sep 15, 2025",47,6440,https://arxiv.org/abs/2509.11543 | |
| 101,"Chain-of-Agents: End-to-End Agent Foundation Models via Multi-Agent | |
| Distillation and Agentic RL",43,5,20.6,19,"Aug 6, 2025",115,333,https://arxiv.org/abs/2508.13167 | |
| 102,"MOSAIC: Multi-Subject Personalized Generation via Correspondence-Aware | |
| Alignment and Disentanglement",33,5,11.58,9,"Sep 2, 2025",11,418,https://arxiv.org/abs/2509.01977 | |
| 103,"A Survey of Scientific Large Language Models: From Data Foundations to | |
| Agent Frontiers",44,7,21.52,16,"Aug 28, 2025",133,312,https://arxiv.org/abs/2508.21148 | |
| 104,MeshCoder: LLM-Powered Structured Mesh Code Generation from Point Clouds,40,4,18.7,16,"Aug 20, 2025",67,374,https://arxiv.org/abs/2508.14879 | |
| 105,From Editor to Dense Geometry Estimator,55,13,28.02,28,"Sep 4, 2025",87,156,https://arxiv.org/abs/2509.04338 | |
| 106,"FlashVSR: Towards Real-Time Diffusion-Based Streaming Video | |
| Super-Resolution",55,7,28.15,29,"Oct 14, 2025",37,910,https://arxiv.org/abs/2510.12747 | |
| 107,"FantasyPortrait: Enhancing Multi-Character Portrait Animation with | |
| Expression-Augmented Diffusion Transformers",55,11,28.22,26,"Jul 17, 2025",24,435,https://arxiv.org/abs/2507.12956 | |
| 108,SpatialLM: Training Large Language Models for Structured Indoor Modeling,50,4,26.46,31,"Jun 9, 2025",49,3900,https://arxiv.org/abs/2506.07491 | |
| 109,Reconstruction Alignment Improves Unified Multimodal Models,34,3,14.94,14,"Sep 8, 2025",38,198,https://arxiv.org/abs/2509.07295 | |
| 110,Transformer Explainer: Interactive Learning of Text-Generative Models,81,21,35.89,35,"Aug 8, 2024",172,5960,https://arxiv.org/abs/2408.04619 | |
| 111,VerlTool: Towards Holistic Agentic Reinforcement Learning with Tool Use,52,7,27.48,28,"Sep 1, 2025",64,499,https://arxiv.org/abs/2509.01055 | |
| 112,"Beyond Ten Turns: Unlocking Long-Horizon Agentic Search with Large-Scale | |
| Asynchronous RL",47,3,25.02,19,"Aug 11, 2025",48,373,https://arxiv.org/abs/2508.07976 | |
| 113,MobiAgent: A Systematic Framework for Customizable Mobile Agents,43,6,22.74,21,"Aug 30, 2025",6,1020,https://arxiv.org/abs/2509.00531 | |
| 114,"AReaL: A Large-Scale Asynchronous Reinforcement Learning System for | |
| Language Reasoning",59,18,30.47,29,"May 30, 2025",28,2660,https://arxiv.org/abs/2505.24298 | |
| 115,SAM 3D: 3Dfy Anything in Images,25,1,2.56,3,"Nov 20, 2025",100,4200,https://arxiv.org/abs/2511.16624 | |
| 116,Continuous Autoregressive Language Models,46,5,24.72,25,"Oct 31, 2025",64,582,https://arxiv.org/abs/2510.27688 | |
| 117,"We-Math 2.0: A Versatile MathBook System for Incentivizing Visual | |
| Mathematical Reasoning",31,1,12.16,3,"Aug 14, 2025",139,138,https://arxiv.org/abs/2508.10433 | |
| 118,Matrix-3D: Omnidirectional Explorable 3D World Generation,45,6,24.27,23,"Aug 11, 2025",67,398,https://arxiv.org/abs/2508.08086 | |
| 119,DeepAgent: A General Reasoning Agent with Scalable Toolsets,43,5,23.09,25,"Oct 24, 2025",93,739,https://arxiv.org/abs/2510.21618 | |
| 120,Logics-Parsing Technical Report,48,4,26.27,28,"Sep 24, 2025",7,619,https://arxiv.org/abs/2509.19760 | |
| 121,"MonkeyOCR: Document Parsing with a Structure-Recognition-Relation | |
| Triplet Paradigm",66,16,33.12,31,"Jun 5, 2025",2,6000,https://arxiv.org/abs/2506.05218 | |
| 122,Emu3.5: Native Multimodal Models are World Learners,35,3,17.94,6,"Oct 30, 2025",103,1210,https://arxiv.org/abs/2510.26583 | |
| 123,Multi-View 3D Point Tracking,33,4,15.97,16,"Aug 28, 2025",20,314,https://arxiv.org/abs/2508.21060 | |
| 124,"Matrix-Game 2.0: An Open-Source, Real-Time, and Streaming Interactive | |
| World Model",41,4,22.83,26,"Aug 18, 2025",22,1550,https://arxiv.org/abs/2508.13009 | |
| 125,"GitTaskBench: A Benchmark for Code Agents Solving Real-World Tasks | |
| Through Code Repository Leveraging",50,13,27.96,23,"Aug 26, 2025",3,183,https://arxiv.org/abs/2508.18993 | |
| 126,"Mini-o3: Scaling Up Reasoning Patterns and Interaction Turns for Visual | |
| Search",39,2,21.49,21,"Sep 9, 2025",59,290,https://arxiv.org/abs/2509.07969 | |
| 127,"GLM-4.1V-Thinking: Towards Versatile Multimodal Reasoning with Scalable | |
| Reinforcement Learning",53,10,29.38,29,"Jul 1, 2025",229,1500,https://arxiv.org/abs/2507.01006 | |
| 128,Diffusion Transformers with Representation Autoencoders,29,1,12.24,7,"Oct 13, 2025",157,1350,https://arxiv.org/abs/2510.11690 | |
| 129,SpatialVID: A Large-Scale Video Dataset with Spatial Annotations,35,8,19.11,16,"Sep 11, 2025",28,331,https://arxiv.org/abs/2509.09676 | |
| 130,Back to Basics: Let Denoising Generative Models Denoise,28,2,11.43,11,"Nov 17, 2025",59,1450,https://arxiv.org/abs/2511.13720 | |
| 131,"ComoRAG: A Cognitive-Inspired Memory-Organized RAG for Stateful Long | |
| Narrative Reasoning",44,8,25.86,28,"Aug 14, 2025",70,192,https://arxiv.org/abs/2508.10419 | |
| 132,"Pixie: Fast and Generalizable Supervised Learning of 3D Physics from | |
| Pixels",34,9,19.24,19,"Aug 20, 2025",34,179,https://arxiv.org/abs/2508.17437 | |
| 133,"In-the-Flow Agentic System Optimization for Effective Planning and Tool | |
| Use",36,2,21.08,22,"Oct 7, 2025",89,851,https://arxiv.org/abs/2510.05592 | |
| 134,"Lyra: Generative 3D Scene Reconstruction via Video Diffusion Model | |
| Self-Distillation",36,10,21.92,15,"Sep 23, 2025",21,458,https://arxiv.org/abs/2509.19296 | |
| 135,"HunyuanWorld 1.0: Generating Immersive, Explorable, and Interactive 3D | |
| Worlds from Words or Pixels",76,22,37.25,38,"Jul 29, 2025",126,2170,https://arxiv.org/abs/2507.21809 | |
| 136,"Concerto: Joint 2D-3D Self-Supervised Learning Emerges Spatial | |
| Representations",22,1,3.5,2,"Oct 27, 2025",172,2610,https://arxiv.org/abs/2510.23607 | |
| 137,LongCat-Video Technical Report,40,6,25.38,28,"Oct 25, 2025",24,1110,https://arxiv.org/abs/2510.22200 | |
| 138,Parallel-R1: Towards Parallel Thinking via Reinforcement Learning,41,5,26.07,27,"Sep 9, 2025",95,151,https://arxiv.org/abs/2509.07980 | |
| 139,"Fish-Speech: Leveraging Large Language Models for Advanced Multilingual | |
| Text-to-Speech Synthesis",28,2,14.79,10,"Nov 2, 2024",11,23900,https://arxiv.org/abs/2411.01156 | |
| 140,LongLive: Real-time Interactive Long Video Generation,29,1,16.55,18,"Sep 26, 2025",174,647,https://arxiv.org/abs/2509.22622 | |
| 141,"Tiny Model, Big Logic: Diversity-Driven Optimization Elicits Large-Model | |
| Reasoning Ability in VibeThinker-1.5B",29,7,16.55,10,"Nov 9, 2025",117,483,https://arxiv.org/abs/2511.06221 | |
| 142,OmniWorld: A Multi-Domain and Multi-Modal Dataset for 4D World Modeling,26,1,13.12,13,"Sep 15, 2025",98,342,https://arxiv.org/abs/2509.12201 | |
| 143,JoyAgent-JDGenie: Technical Report on the GAIA,61,14,35.08,38,"Oct 1, 2025",3,11100,https://arxiv.org/abs/2510.00510 | |
| 144,Ovis2.5 Technical Report,39,6,26.36,27,"Aug 15, 2025",102,1280,https://arxiv.org/abs/2508.11737 | |
| 145,Marco-Voice Technical Report,53,25,32.89,33,"Aug 4, 2025",15,332,https://arxiv.org/abs/2508.02038 | |
| 146,"A.S.E: A Repository-Level Benchmark for Evaluating Security in | |
| AI-Generated Code",62,7,35.66,37,"Aug 25, 2025",340,709,https://arxiv.org/abs/2508.18106 | |
| 147,"Revolutionizing Reinforcement Learning Framework for Diffusion Large | |
| Language Models",25,2,12.96,5,"Sep 8, 2025",51,171,https://arxiv.org/abs/2509.06949 | |
| 148,Intern-S1: A Scientific Multimodal Foundation Model,27,1,15.81,14,"Aug 21, 2025",236,538,https://arxiv.org/abs/2508.15763 | |
| 149,SAM 3: Segment Anything with Concepts,19,1,1.63,2,"Nov 20, 2025",95,4760,https://arxiv.org/abs/2511.16719 | |
| 150,Agent0: Unleashing Self-Evolving Agents from Zero Data via Tool-Integrated Reasoning,24,6,12.17,9,"Nov 20, 2025",97,705,https://arxiv.org/abs/2511.16043 | |
| 151,"Tinker: Diffusion's Gift to 3D--Multi-View Consistent Editing From | |
| Sparse Inputs without Per-Scene Optimization",29,7,19.86,21,"Aug 20, 2025",39,129,https://arxiv.org/abs/2508.14811 | |
| 152,WithAnyone: Towards Controllable and ID Consistent Image Generation,24,1,13.58,9,"Oct 16, 2025",76,415,https://arxiv.org/abs/2510.14975 | |
| 153,ST-Raptor: LLM-Powered Semi-Structured Table Question Answering,32,3,22.97,24,"Aug 25, 2025",6,243,https://arxiv.org/abs/2508.18190 | |
| 154,"Pref-GRPO: Pairwise Preference Reward-based GRPO for Stable | |
| Text-to-Image Reinforcement Learning",26,2,16.62,17,"Aug 28, 2025",85,123,https://arxiv.org/abs/2508.20751 | |
| 155,Arch-Router: Aligning LLM Routing with Human Preferences,60,21,36.28,37,"Jun 19, 2025",17,4170,https://arxiv.org/abs/2506.16655 | |
| 156,GenCompositor: Generative Video Compositing with Diffusion Transformer,44,12,31.52,29,"Sep 2, 2025",24,122,https://arxiv.org/abs/2509.02460 | |
| 157,"Voost: A Unified and Scalable Diffusion Transformer for Bidirectional | |
| Virtual Try-On and Try-Off",37,9,28.11,26,"Aug 6, 2025",56,264,https://arxiv.org/abs/2508.04825 | |
| 158,"FilmAgent: A Multi-Agent Framework for End-to-End Film Automation in | |
| Virtual 3D Spaces",20,1,8.75,3,"Jan 22, 2025",74,1050,https://arxiv.org/abs/2501.12909 | |
| 159,"Story2Board: A Training-Free Approach for Expressive Storyboard | |
| Generation",45,14,32.31,34,"Aug 13, 2025",61,119,https://arxiv.org/abs/2508.09983 | |
| 160,StreamingVLM: Real-Time Understanding for Infinite Video Streams,26,3,18.77,23,"Oct 10, 2025",46,557,https://arxiv.org/abs/2510.09608 | |
| 161,LightMem: Lightweight and Efficient Memory-Augmented Generation,24,2,16.12,13,"Oct 21, 2025",105,295,https://arxiv.org/abs/2510.18866 | |
| 162,"VoxHammer: Training-Free Precise and Coherent 3D Editing in Native 3D | |
| Space",27,3,20.3,23,"Aug 26, 2025",38,130,https://arxiv.org/abs/2508.19247 | |
| 163,RLinf-VLA: A Unified and Efficient Framework for VLA+RL Training,86,25,41.38,42,"Oct 8, 2025",38,1450,https://arxiv.org/abs/2510.06710 | |
| 164,"CommonForms: A Large, Diverse Dataset for Form Field Detection",29,2,22.59,27,"Sep 20, 2025",18,819,https://arxiv.org/abs/2509.16506 | |
| 165,"Thinking with Video: Video Generation as a Promising Multimodal | |
| Reasoning Paradigm",26,2,19.81,10,"Nov 6, 2025",188,182,https://arxiv.org/abs/2511.04570 | |
| 166,"Speed Always Wins: A Survey on Efficient Architectures for Large | |
| Language Models",36,15,28.69,25,"Aug 13, 2025",51,281,https://arxiv.org/abs/2508.09834 | |
| 167,"Scaling Instruction-Based Video Editing with a High-Quality Synthetic | |
| Dataset",23,3,16.39,15,"Oct 17, 2025",49,385,https://arxiv.org/abs/2510.15742 | |
| 168,PhysX-Anything: Simulation-Ready Physical 3D Assets from Single Image,25,2,19.24,15,"Nov 17, 2025",51,593,https://arxiv.org/abs/2511.13648 | |
| 169,"Echo-4o: Harnessing the Power of GPT-4o Synthetic Images for Improved | |
| Image Generation",27,4,21.67,20,"Aug 13, 2025",23,73,https://arxiv.org/abs/2508.09987 | |
| 170,TTT3R: 3D Reconstruction as Test-Time Training,29,3,24.1,28,"Sep 30, 2025",13,396,https://arxiv.org/abs/2509.26645 | |
| 171,"SINQ: Sinkhorn-Normalized Quantization for Calibration-Free | |
| Low-Precision LLM Weights",25,5,19.96,19,"Sep 26, 2025",73,472,https://arxiv.org/abs/2509.22944 | |
| 172,Detect Anything via Next Point Prediction,22,4,16.45,10,"Oct 14, 2025",42,449,https://arxiv.org/abs/2510.12798 | |
| 173,General Agentic Memory Via Deep Research,17,3,7.47,8,"Nov 23, 2025",146,485,https://arxiv.org/abs/2511.18423 | |
| 174,"Rethinking Semantic Segmentation from a Sequence-to-Sequence Perspective | |
| with Transformers",21,2,15.81,10,"Dec 31, 2020",3,1100,https://arxiv.org/abs/2012.15840 | |
| 175,"Killing Two Birds with One Stone:Efficient and Robust Training of Face | |
| Recognition CNNs by Partial FC",92,26,43.16,45,"Mar 28, 2022",3,26700,https://arxiv.org/abs/2203.15565 | |
| 176,Reinforcement Learning in Vision: A Survey,41,25,33.49,33,"Aug 11, 2025",27,184,https://arxiv.org/abs/2508.08189 | |
| 177,HunyuanOCR Technical Report,15,2,3.27,3,"Nov 24, 2025",17,878,https://arxiv.org/abs/2511.19575 | |
| 178,"Collaborating Action by Action: A Multi-agent LLM Framework for Embodied | |
| Reasoning",26,10,23.54,21,"Apr 24, 2025",5,4250,https://arxiv.org/abs/2504.17950 | |
| 179,Training Video Foundation Models with NVIDIA NeMo,87,19,42.82,44,"Mar 17, 2025",7,15700,https://arxiv.org/abs/2503.12964 | |
| 180,"MiroThinker: Pushing the Performance Boundaries of Open-Source Research Agents via Model, Context, and Interactive Scaling",29,4,26.66,30,"Nov 14, 2025",156,1160,https://arxiv.org/abs/2511.11793 | |
| 181,Agent0-VL: Exploring Self-Evolving Agent for Tool-Integrated Vision-Language Reasoning,15,4,5.33,5,"Nov 25, 2025",46,702,https://arxiv.org/abs/2511.19900 | |
| 182,OmniTry: Virtual Try-On Anything without Masks,41,13,34.39,34,"Aug 19, 2025",16,166,https://arxiv.org/abs/2508.13632 | |
| 183,FlashWorld: High-quality 3D Scene Generation within Seconds,18,3,13.17,8,"Oct 15, 2025",66,397,https://arxiv.org/abs/2510.13678 | |
| 184,"Diffusion LLMs Can Do Faster-Than-AR Inference via Discrete Diffusion | |
| Forcing",30,10,28.57,29,"Aug 8, 2025",29,79,https://arxiv.org/abs/2508.09192 | |
| 185,Real-Time Object Detection Meets DINOv3,37,14,32.84,37,"Sep 25, 2025",10,684,https://arxiv.org/abs/2509.20787 | |
| 186,"OmniVinci: Enhancing Architecture and Data for Omni-Modal Understanding | |
| LLM",28,7,27.36,24,"Oct 17, 2025",86,502,https://arxiv.org/abs/2510.15870 | |
| 187,"QeRL: Beyond Efficiency -- Quantization-enhanced Reinforcement Learning | |
| for LLMs",21,1,19.67,20,"Oct 13, 2025",162,350,https://arxiv.org/abs/2510.11696 | |
| 188,TimeGPT-1,21,6,19.86,12,"Oct 5, 2023",7,3470,https://arxiv.org/abs/2310.03589 | |
| 189,"LucidFlux: Caption-Free Universal Image Restoration via a Large-Scale | |
| Diffusion Transformer",37,24,33.35,32,"Sep 26, 2025",21,476,https://arxiv.org/abs/2509.22414 | |
| 190,Ark: An Open-source Python-based Framework for Robot Learning,25,10,25.04,18,"Jun 24, 2025",16,245,https://arxiv.org/abs/2506.21628 | |
| 191,Kwai Keye-VL Technical Report,40,20,34.85,33,"Jul 2, 2025",130,623,https://arxiv.org/abs/2507.01949 | |
| 192,MHR: Momentum Human Rig,24,4,25,28,"Nov 19, 2025",13,419,https://arxiv.org/abs/2511.15586 | |
| 193,P3-SAM: Native 3D Part Segmentation,20,1,19.85,13,"Sep 8, 2025",21,249,https://arxiv.org/abs/2509.06784 | |
| 194,"Spark-TTS: An Efficient LLM-Based Text-to-Speech Model with | |
| Single-Stream Decoupled Speech Tokens",23,14,24.22,22,"Mar 3, 2025",6,10400,https://arxiv.org/abs/2503.01710 | |
| 195,Diffusion Language Models are Super Data Learners,25,16,26.36,19,"Nov 5, 2025",112,259,https://arxiv.org/abs/2511.03276 | |
| 196,"Hulu-Med: A Transparent Generalist Model towards Holistic Medical | |
| Vision-Language Understanding",30,13,30.57,35,"Oct 9, 2025",4,478,https://arxiv.org/abs/2510.08668 | |
| 197,"HoloCine: Holistic Generation of Cinematic Multi-Shot Long Video | |
| Narratives",20,10,20.55,17,"Oct 23, 2025",37,333,https://arxiv.org/abs/2510.20822 | |
| 198,Training-Free Group Relative Policy Optimization,34,25,33.18,32,"Oct 9, 2025",42,3590,https://arxiv.org/abs/2510.08191 | |
| 199,Visual Spatial Tuning,21,3,22.29,18,"Nov 7, 2025",46,124,https://arxiv.org/abs/2511.05491 | |
| 200,Step-Audio-R1 Technical Report,25,10,26.92,32,"Nov 19, 2025",51,316,https://arxiv.org/abs/2511.15848 | |
| 201,WizardCoder: Empowering Code Large Language Models with Evol-Instruct,12,1,1,1,"Jun 14, 2023",30,9470,https://arxiv.org/abs/2306.08568 | |
| 202,Dolphin: Document Image Parsing via Heterogeneous Anchor Prompting,18,8,17.78,15,"May 20, 2025",3,7670,https://arxiv.org/abs/2505.14059 | |
| 203,"FLUX-Reason-6M & PRISM-Bench: A Million-Scale Text-to-Image Reasoning | |
| Dataset and Comprehensive Benchmark",19,3,19.74,15,"Sep 11, 2025",35,74,https://arxiv.org/abs/2509.09680 | |
| 204,MiMo-Embodied: X-Embodied Foundation Model Technical Report,23,9,25.26,28,"Nov 20, 2025",23,260,https://arxiv.org/abs/2511.16518 | |
| 205,Robot Learning: A Tutorial,20,6,21.5,15,"Oct 14, 2025",80,281,https://arxiv.org/abs/2510.12403 | |
| 206,GigaWorld-0: World Models as Data Engine to Empower Embodied AI,15,8,11.93,13,"Nov 25, 2025",28,214,https://arxiv.org/abs/2511.19861 | |
| 207,"Large Language Model Agent: A Survey on Methodology, Applications and | |
| Challenges",82,35,43.88,45,"Mar 27, 2025",82,1710,https://arxiv.org/abs/2503.21460 | |
| 208,Chronos-2: From Univariate to Universal Forecasting,32,23,32.75,31,"Oct 17, 2025",17,4190,https://arxiv.org/abs/2510.15821 | |
| 209,ViSTA-SLAM: Visual SLAM with Symmetric Two-view Association,38,16,35.74,37,"Sep 1, 2025",6,134,https://arxiv.org/abs/2509.01584 | |
| 210,ReCode: Unify Plan and Action for Universal Granularity Control,29,16,31.76,30,"Oct 27, 2025",117,359,https://arxiv.org/abs/2510.23564 | |
| 211,"Skyfall-GS: Synthesizing Immersive 3D Urban Scenes from Satellite | |
| Imagery",37,9,36.08,41,"Oct 17, 2025",45,576,https://arxiv.org/abs/2510.15869 | |
| 212,"OpenVision 2: A Family of Generative Pretrained Visual Encoders for | |
| Multimodal Learning",27,11,30.74,30,"Sep 1, 2025",27,356,https://arxiv.org/abs/2509.01644 | |
| 213,4KAgent: Agentic Any Image to 4K Super-Resolution,21,19,25.19,22,"Jul 9, 2025",104,547,https://arxiv.org/abs/2507.07105 | |
| 214,"ChronoEdit: Towards Temporal Reasoning for Image Editing and World | |
| Simulation",45,23,38.96,42,"Oct 5, 2025",14,578,https://arxiv.org/abs/2510.04290 | |
| 215,X-Part: high fidelity and structure coherent shape decomposition,17,12,19.53,14,"Sep 10, 2025",25,234,https://arxiv.org/abs/2509.08643 | |
| 216,"Huxley-Gödel Machine: Human-Level Coding Agent Development by an | |
| Approximation of the Optimal Self-Improving Machine",21,10,25.71,21,"Oct 24, 2025",18,251,https://arxiv.org/abs/2510.21614 | |
| 217,"Build Your Personalized Research Group: A Multiagent Framework for | |
| Continual and Interactive Science Automation",17,2,19.82,24,"Oct 17, 2025",14,267,https://arxiv.org/abs/2510.15624 | |
| 218,"OpenTSLM: Time-Series Language Models for Reasoning over Multivariate | |
| Medical Text- and Time-Series Data",21,8,26.43,27,"Oct 2, 2025",15,950,https://arxiv.org/abs/2510.02410 | |
| 219,Trace Anything: Representing Any Video in 4D via Trajectory Fields,17,11,21.18,22,"Oct 15, 2025",30,271,https://arxiv.org/abs/2510.13802 | |
| 220,Democratizing AI scientists using ToolUniverse,29,17,33.59,38,"Sep 27, 2025",38,455,https://arxiv.org/abs/2509.23426 | |
| 221,"S^2-Guidance: Stochastic Self Guidance for Training-Free Enhancement of | |
| Diffusion Models",25,15,30.96,23,"Aug 18, 2025",45,134,https://arxiv.org/abs/2508.12880 | |
| 222,UniVerse-1: Unified Audio-Video Generation via Stitching of Experts,15,9,18.07,10,"Sep 7, 2025",14,53,https://arxiv.org/abs/2509.06155 | |
| 223,R-Zero: Self-Evolving Reasoning LLM from Zero Data,39,28,38.41,38,"Aug 7, 2025",123,559,https://arxiv.org/abs/2508.05004 | |
| 224,"On the Generalization of SFT: A Reinforcement Learning Perspective with | |
| Reward Rectification",29,26,34.34,35,"Aug 7, 2025",148,322,https://arxiv.org/abs/2508.05629 | |
| 225,"InternScenes: A Large-scale Simulatable Indoor Scene Dataset with | |
| Realistic Layouts",17,7,22.94,20,"Sep 13, 2025",29,161,https://arxiv.org/abs/2509.10813 | |
| 226,"LightReasoner: Can Small Language Models Teach Large Language Models | |
| Reasoning?",15,9,19.33,10,"Oct 9, 2025",7,362,https://arxiv.org/abs/2510.07962 | |
| 227,The Denario project: Deep knowledge AI agents for scientific discovery,22,15,29.73,26,"Oct 30, 2025",6,372,https://arxiv.org/abs/2510.26887 | |
| 228,"R-4B: Incentivizing General-Purpose Auto-Thinking Capability in MLLMs | |
| via Bi-Mode Annealing and Reinforce Learning",19,21,26.63,23,"Aug 28, 2025",102,84,https://arxiv.org/abs/2508.21113 | |
| 229,Rolling Forcing: Autoregressive Long Video Diffusion in Real Time,15,3,20.13,18,"Sep 29, 2025",20,109,https://arxiv.org/abs/2509.25161 | |
| 230,Robot Learning from a Physical World Model,23,23,31.09,30,"Nov 10, 2025",26,146,https://arxiv.org/abs/2511.07416 | |
| 231,"Genie Envisioner: A Unified World Foundation Platform for Robotic | |
| Manipulation",24,19,31.96,29,"Aug 7, 2025",72,218,https://arxiv.org/abs/2508.05635 | |
| 232,"RLVE: Scaling Up Reinforcement Learning for Language Models with | |
| Adaptive Verifiable Environments",20,18,28.15,23,"Nov 10, 2025",10,119,https://arxiv.org/abs/2511.07317 | |
| 233,"Packing Input Frame Context in Next-Frame Prediction Models for Video | |
| Generation",47,15,41.34,44,"Apr 17, 2025",52,15800,https://arxiv.org/abs/2504.12626 | |
| 234,"M^3FinMeeting: A Multilingual, Multi-Sector, and Multi-Task Financial | |
| Meeting Understanding Evaluation Dataset",22,12,30.41,28,"Jun 3, 2025",3,287,https://arxiv.org/abs/2506.02510 | |
| 235,"Beyond Outlining: Heterogeneous Recursive Planning for Adaptive | |
| Long-form Writing with Language Models",13,9,16.15,12,"Mar 11, 2025",3,736,https://arxiv.org/abs/2503.08275 | |
| 236,"Skywork UniPic: Unified Autoregressive Modeling for Visual Understanding | |
| and Generation",25,18,33,31,"Aug 5, 2025",59,727,https://arxiv.org/abs/2508.03320 | |
| 237,Tree Search for LLM Agent Reinforcement Learning,15,13,21.2,19,"Sep 25, 2025",79,95,https://arxiv.org/abs/2509.21240 | |
| 238,VoXtream: Full-Stream Text-to-Speech with Extremely Low Latency,24,24,32.42,31,"Sep 19, 2025",2,129,https://arxiv.org/abs/2509.15969 | |
| 239,A Survey of Data Agents: Emerging Paradigm or Overstated Hype?,17,11,24.82,19,"Oct 27, 2025",63,202,https://arxiv.org/abs/2510.23587 | |
| 240,SteadyDancer: Harmonized and Coherent Human Image Animation with First-Frame Preservation,15,17,22.07,21,"Nov 24, 2025",38,189,https://arxiv.org/abs/2511.19320 | |
| 241,Latent Collaboration in Multi-Agent Systems,12,12,14.92,14,"Nov 25, 2025",92,240,https://arxiv.org/abs/2511.20639 | |
| 242,Puppeteer: Rig and Animate Your 3D Models,31,27,37.1,37,"Aug 14, 2025",30,178,https://arxiv.org/abs/2508.10898 | |
| 243,GEM: A Gym for Agentic LLMs,21,24,31.29,30,"Oct 1, 2025",79,284,https://arxiv.org/abs/2510.01051 | |
| 244,MedRAX: Medical Reasoning Agent for Chest X-ray,18,9,28,29,"Feb 4, 2025",2,993,https://arxiv.org/abs/2502.02673 | |
| 245,"Omni-Effects: Unified and Spatially-Controllable Visual Effects | |
| Generation",17,14,26.71,28,"Aug 11, 2025",56,122,https://arxiv.org/abs/2508.07981 | |
| 246,FlowRL: Matching Reward Distributions for LLM Reasoning,14,13,21.5,17,"Sep 18, 2025",90,58,https://arxiv.org/abs/2509.15207 | |
| 247,"Uniworld-V2: Reinforce Image Editing with Diffusion Negative-aware | |
| Finetuning and MLLM Implicit Feedback",14,3,21.5,21,"Oct 19, 2025",18,94,https://arxiv.org/abs/2510.16888 | |
| 248,"UMO: Scaling Multi-Identity Consistency for Image Customization via | |
| Matching Reward",19,18,29.37,26,"Sep 8, 2025",29,131,https://arxiv.org/abs/2509.06818 | |
| 249,"From Pixels to Words -- Towards Native Vision-Language Primitives at | |
| Scale",12,8,16.75,13,"Oct 16, 2025",60,161,https://arxiv.org/abs/2510.14979 | |
| 250,LIMI: Less is More for Agency,26,28,35.31,35,"Sep 22, 2025",91,101,https://arxiv.org/abs/2509.17567 | |
| 251,"TaDiCodec: Text-aware Diffusion Speech Tokenizer for Speech Language | |
| Modeling",18,12,28.44,31,"Aug 22, 2025",7,149,https://arxiv.org/abs/2508.16790 | |
| 252,SceneGen: Single-Image 3D Scene Generation in One Feedforward Pass,23,26,33.39,31,"Aug 21, 2025",18,66,https://arxiv.org/abs/2508.15769 | |
| 253,The Markovian Thinker,19,15,29.95,29,"Oct 8, 2025",27,251,https://arxiv.org/abs/2510.06557 | |
| 254,Time-to-Move: Training-Free Motion Controlled Video Generation via Dual-Clock Denoising,21,22,32.05,30,"Nov 9, 2025",50,191,https://arxiv.org/abs/2511.08633 | |
| 255,V-Thinker: Interactive Thinking with Images,16,5,26.31,30,"Nov 6, 2025",92,104,https://arxiv.org/abs/2511.04460 | |
| 256,"UniGenBench++: A Unified Semantic Evaluation Benchmark for Text-to-Image | |
| Generation",12,5,18.17,14,"Oct 21, 2025",64,100,https://arxiv.org/abs/2510.18701 | |
| 257,"MiniCPM-V 4.5: Cooking Efficient MLLMs via Architecture, Data, and | |
| Training Recipe",21,17,32.43,31,"Sep 16, 2025",45,22000,https://arxiv.org/abs/2509.18154 | |
| 258,WorldGrow: Generating Infinite 3D World,17,12,28.12,25,"Oct 24, 2025",40,313,https://arxiv.org/abs/2510.21682 | |
| 259,Video-As-Prompt: Unified Semantic Control for Video Generation,16,11,26.88,22,"Oct 23, 2025",41,217,https://arxiv.org/abs/2510.20888 | |
| 260,"SongBloom: Coherent Song Generation via Interleaved Autoregressive | |
| Sketching and Diffusion Refinement",21,23,32.86,31,"Jun 9, 2025",6,604,https://arxiv.org/abs/2506.07634 | |
| 261,"FireRedTTS-2: Towards Long Conversational Speech Generation for Podcast | |
| and Chatbot",24,21,35.21,35,"Sep 2, 2025",1,1180,https://arxiv.org/abs/2509.02020 | |
| 262,LTX-Video: Realtime Video Latent Diffusion,18,19,30.17,30,"Dec 30, 2024",47,8590,https://arxiv.org/abs/2501.00103 | |
| 263,OpenMMReasoner: Pushing the Frontiers for Multimodal Reasoning with an Open and General Recipe,14,6,24.29,23,"Nov 20, 2025",86,106,https://arxiv.org/abs/2511.16334 | |
| 264,SciReasoner: Laying the Scientific Reasoning Ground Across Disciplines,13,12,23.23,18,"Sep 25, 2025",86,43,https://arxiv.org/abs/2509.21320 | |
| 265,"Uni-MoE-2.0-Omni: Scaling Language-Centric Omnimodal Large Model with Advanced MoE, Training and Data",14,15,25.36,27,"Nov 16, 2025",99,1010,https://arxiv.org/abs/2511.12609 | |
| 266,"Advancing End-to-End Pixel Space Generative Modeling via Self-supervised | |
| Pre-training",14,4,25.43,28,"Oct 14, 2025",104,98,https://arxiv.org/abs/2510.12586 | |
| 267,"IGGT: Instance-Grounded Geometry Transformer for Semantic 3D | |
| Reconstruction",15,16,27.2,23,"Oct 26, 2025",38,156,https://arxiv.org/abs/2510.22706 | |
| 268,"CODA: Coordinating the Cerebrum and Cerebellum for a Dual-Brain Computer | |
| Use Agent with Decoupled Reinforcement Learning",15,5,27.47,30,"Aug 27, 2025",31,25,https://arxiv.org/abs/2508.20096 | |
| 269,"Verbalized Sampling: How to Mitigate Mode Collapse and Unlock LLM | |
| Diversity",20,26,33.45,34,"Oct 1, 2025",15,219,https://arxiv.org/abs/2510.01171 | |
| 270,MMaDA-Parallel: Multimodal Large Diffusion Language Models for Thinking-Aware Editing and Generation,11,7,19.09,21,"Nov 12, 2025",64,248,https://arxiv.org/abs/2511.09611 | |
| 271,Kandinsky 5.0: A Family of Foundation Models for Image and Video Generation,14,13,25.93,22,"Nov 19, 2025",207,471,https://arxiv.org/abs/2511.14993 | |
| 272,MemOS: A Memory OS for AI System,36,13,41.33,45,"Jul 4, 2025",155,3110,https://arxiv.org/abs/2507.03724 | |
| 273,DA^2: Depth Anything in Any Direction,17,8,30.53,37,"Sep 30, 2025",20,120,https://arxiv.org/abs/2509.26618 | |
| 274,"OneReward: Unified Mask-Guided Image Generation via Multi-Task Human | |
| Preference Learning",15,11,28,30,"Aug 28, 2025",13,208,https://arxiv.org/abs/2508.21066 | |
| 275,"Paper2Code: Automating Code Generation from Scientific Papers in Machine | |
| Learning",32,27,40.34,40,"Apr 24, 2025",120,3730,https://arxiv.org/abs/2504.17192 | |
| 276,"ThinkMorph: Emergent Properties in Multimodal Interleaved | |
| Chain-of-Thought Reasoning",17,9,31,28,"Oct 30, 2025",78,87,https://arxiv.org/abs/2510.27492 | |
| 277,"The Well: a Large-Scale Collection of Diverse Physics Simulations for | |
| Machine Learning",12,13,23,14,"Nov 30, 2024",20,1110,https://arxiv.org/abs/2412.00568 | |
| 278,Versatile Framework for Song Generation with Prompt-based Control,44,34,43.5,45,"Apr 27, 2025",6,201,https://arxiv.org/abs/2504.19062 | |
| 279,Durian: Dual Reference-guided Portrait Animation with Attribute Transfer,15,16,29.13,22,"Sep 4, 2025",4,14,https://arxiv.org/abs/2509.04434 | |
| 280,Self-Forcing++: Towards Minute-Scale High-Quality Video Generation,19,22,34,30,"Oct 2, 2025",86,145,https://arxiv.org/abs/2510.02283 | |
| 281,OmniSVG: A Unified Scalable Vector Graphics Generation Model,16,16,30.88,28,"Apr 8, 2025",180,2100,https://arxiv.org/abs/2504.06263 | |
| 282,"SimpleTIR: End-to-End Reinforcement Learning for Multi-Turn | |
| Tool-Integrated Reasoning",22,19,36.55,43,"Sep 2, 2025",76,250,https://arxiv.org/abs/2509.02479 | |
| 283,VGGT-X: When VGGT Meets Dense Novel View Synthesis,13,15,26.69,25,"Sep 29, 2025",15,69,https://arxiv.org/abs/2509.25191 | |
| 284,"VCode: a Multimodal Coding Benchmark with SVG as Symbolic Visual | |
| Representation",11,5,22.45,18,"Nov 4, 2025",95,88,https://arxiv.org/abs/2511.02778 | |
| 285,"Thinking with Camera: A Unified Multimodal Model for Camera-Centric | |
| Understanding and Generation",13,4,26.92,29,"Oct 9, 2025",115,200,https://arxiv.org/abs/2510.08673 | |
| 286,Latent Diffusion Model without Variational Autoencoder,14,13,28.71,24,"Oct 17, 2025",39,156,https://arxiv.org/abs/2510.15301 | |
| 287,"Evolution Strategies at Scale: LLM Fine-Tuning Beyond Reinforcement | |
| Learning",10,8,20.5,19,"Sep 29, 2025",4,164,https://arxiv.org/abs/2509.24372 | |
| 288,"VideoFrom3D: 3D Scene Video Generation via Complementary Image and Video | |
| Diffusion Models",18,18,34.39,36,"Sep 22, 2025",25,83,https://arxiv.org/abs/2509.17985 | |
| 289,"VolSplat: Rethinking Feed-Forward 3D Gaussian Splatting with | |
| Voxel-Aligned Prediction",14,16,29.86,24,"Sep 23, 2025",22,65,https://arxiv.org/abs/2509.19297 | |
| 290,MolmoAct: Action Reasoning Models that can Reason in Space,22,28,37.68,37,"Aug 11, 2025",42,163,https://arxiv.org/abs/2508.07917 | |
| 291,"Ming-UniAudio: Speech LLM for Joint Understanding, Generation and | |
| Editing with Unified Representation",14,9,30.21,28,"Oct 26, 2025",9,370,https://arxiv.org/abs/2511.05516 | |
| 292,MAPO: Mixed Advantage Policy Optimization,12,12,27.08,23,"Sep 23, 2025",25,34,https://arxiv.org/abs/2509.18849 | |
| 293,DepthLM: Metric Depth From Vision Language Models,18,26,35.06,32,"Sep 29, 2025",5,130,https://arxiv.org/abs/2509.25413 | |
| 294,"REINFORCE++: A Simple and Efficient Approach for Aligning Large Language | |
| Models",41,26,44.05,45,"Jan 4, 2025",102,8130,https://arxiv.org/abs/2501.03262 | |
| 295,DeepPHY: Benchmarking Agentic VLMs on Physical Reasoning,26,25,40.15,42,"Aug 7, 2025",62,143,https://arxiv.org/abs/2508.05405 | |
| 296,"Lumina-DiMOO: An Omni Diffusion Large Language Model for Multi-Modal | |
| Generation and Understanding",13,18,29.38,30,"Oct 7, 2025",48,834,https://arxiv.org/abs/2510.06308 | |
| 297,DR Tulu: Reinforcement Learning with Evolving Rubrics for Deep Research,16,24,33.62,29,"Nov 24, 2025",48,404,https://arxiv.org/abs/2511.19399 | |
| 298,"IMAGGarment-1: Fine-Grained Garment Generation for Controllable Fashion | |
| Design",14,21,31.64,31,"Apr 17, 2025",1,263,https://arxiv.org/abs/2504.13176 | |
| 299,Interleaving Reasoning for Better Text-to-Image Generation,11,16,26.64,17,"Sep 8, 2025",12,24,https://arxiv.org/abs/2509.06945 | |
| 300,ARE: Scaling Up Agent Environments and Evaluations,17,24,35.24,34,"Sep 21, 2025",29,257,https://arxiv.org/abs/2509.17158 | |
| 301,GeoVista: Web-Augmented Agentic Visual Reasoning for Geolocalization,11,9,27.09,28,"Nov 19, 2025",88,170,https://arxiv.org/abs/2511.15705 | |
| 302,"Scrub It Out! Erasing Sensitive Memorization in Code Language Models via | |
| Machine Unlearning",10,16,24.8,19,"Sep 17, 2025",18,24,https://arxiv.org/abs/2509.13755 | |
| 303,NaTex: Seamless Texture Generation as Latent Color Diffusion,9,10,21.89,21,"Nov 20, 2025",15,85,https://arxiv.org/abs/2511.16317 | |
| 304,"StealthAttack: Robust 3D Gaussian Splatting Poisoning via Density-Guided | |
| Illusions",9,2,22.11,23,"Oct 2, 2025",55,51,https://arxiv.org/abs/2510.02314 | |
| 305,Human-Agent Collaborative Paper-to-Page Crafting for Under $0.1,9,13,22.56,19,"Oct 22, 2025",64,98,https://arxiv.org/abs/2510.19600 | |
| 306,"F1: A Vision-Language-Action Model Bridging Understanding and Generation | |
| to Actions",10,15,25.8,23,"Sep 8, 2025",26,64,https://arxiv.org/abs/2509.06951 | |
| 307,ROSE: Remove Objects with Side Effects in Videos,12,14,30.67,33,"Aug 26, 2025",4,32,https://arxiv.org/abs/2508.18633 | |
| 308,"VideoCanvas: Unified Video Completion from Arbitrary Spatiotemporal | |
| Patches via In-Context Conditioning",7,2,16.43,16,"Oct 9, 2025",59,50,https://arxiv.org/abs/2510.08555 | |
| 309,"EditScore: Unlocking Online RL for Image Editing via High-Fidelity | |
| Reward Modeling",9,16,24.33,21,"Sep 28, 2025",26,60,https://arxiv.org/abs/2509.23909 | |
| 310,Reinforced Visual Perception with Tools,9,18,24.44,22,"Sep 1, 2025",27,28,https://arxiv.org/abs/2509.01656 | |
| 311,Is Diversity All You Need for Scalable Robotic Manipulation?,18,30,37.78,37,"Jul 8, 2025",20,2460,https://arxiv.org/abs/2507.06219 | |
| 312,RLFR: Extending Reinforcement Learning for LLMs with Flow Environment,7,5,17,14,"Oct 11, 2025",32,34,https://arxiv.org/abs/2510.10201 | |
| 313,"BAPO: Stabilizing Off-Policy Reinforcement Learning for LLMs via | |
| Balanced Policy Optimization with Adaptive Clipping",9,15,24.56,23,"Oct 21, 2025",77,62,https://arxiv.org/abs/2510.18927 | |
| 314,"GenoMAS: A Multi-Agent Framework for Scientific Discovery via | |
| Code-Driven Gene Expression Analysis",11,25,29.45,26,"Jul 28, 2025",3,93,https://arxiv.org/abs/2507.21035 | |
| 315,"Latent Zoning Network: A Unified Principle for Generative Modeling, | |
| Representation Learning, and Classification",9,12,24.89,19,"Sep 19, 2025",43,39,https://arxiv.org/abs/2509.15591 | |
| 316,"Grasp Any Region: Towards Precise, Contextual Pixel Understanding for | |
| Multimodal LLMs",10,13,27.6,27,"Oct 21, 2025",33,57,https://arxiv.org/abs/2510.18876 | |
| 317,"FantasyTalking2: Timestep-Layer Adaptive Preference Optimization for | |
| Audio-Driven Portrait Animation",13,22,33.15,31,"Aug 15, 2025",8,21,https://arxiv.org/abs/2508.11255 | |
| 318,"OpenVision: A Fully-Open, Cost-Effective Family of Advanced Vision | |
| Encoders for Multimodal Learning",17,26,37.47,39,"May 7, 2025",28,351,https://arxiv.org/abs/2505.04601 | |
| 319,"Video-LMM Post-Training: A Deep Dive into Video Reasoning with Large | |
| Multimodal Models",10,10,28.2,27,"Oct 6, 2025",42,89,https://arxiv.org/abs/2510.05034 | |
| 320,"AWorld: Dynamic Multi-Agent System with Stable Maneuvering for Robust | |
| GAIA Problem Solving",31,34,43.71,45,"Aug 13, 2025",32,694,https://arxiv.org/abs/2508.09889 | |
| 321,"Video-Thinker: Sparking ""Thinking with Videos"" via Reinforcement | |
| Learning",10,13,28.5,22,"Oct 27, 2025",81,95,https://arxiv.org/abs/2510.23473 | |
| 322,Cook and Clean Together: Teaching Embodied Agents for Parallel Task Execution,11,19,30.64,28,"Nov 24, 2025",7,270,https://arxiv.org/abs/2511.19430 | |
| 323,"BEAVR: Bimanual, multi-Embodiment, Accessible, Virtual Reality | |
| Teleoperation System for Robots",11,19,30.82,31,"Aug 13, 2025",0,50,https://arxiv.org/abs/2508.09606 | |
| 324,"NaViL: Rethinking Scaling Properties of Native Multimodal Large Language | |
| Models under Data Constraints",8,19,23.38,23,"Oct 9, 2025",17,71,https://arxiv.org/abs/2510.08565 | |
| 325,Human3R: Everyone Everywhere All at Once,9,22,26.44,26,"Oct 7, 2025",8,304,https://arxiv.org/abs/2510.06219 | |
| 326,"BRIDGE - Building Reinforcement-Learning Depth-to-Image Data Generation | |
| Engine for Monocular Depth Estimation",17,33,38.06,37,"Sep 29, 2025",13,106,https://arxiv.org/abs/2509.25077 | |
| 327,"THOR: Tool-Integrated Hierarchical Optimization via RL for Mathematical | |
| Reasoning",10,16,29.1,28,"Sep 17, 2025",11,17,https://arxiv.org/abs/2509.13761 | |
| 328,"T2I-ReasonBench: Benchmarking Reasoning-Informed Text-to-Image | |
| Generation",8,11,23.75,17,"Aug 24, 2025",25,24,https://arxiv.org/abs/2508.17472 | |
| 329,SPATIALGEN: Layout-guided 3D Indoor Scene Generation,10,15,29.5,30,"Sep 18, 2025",22,255,https://arxiv.org/abs/2509.14981 | |
| 330,G^2VLM: Geometry Grounded Vision Language Model with Unified 3D Reconstruction and Spatial Reasoning,10,22,29.5,27,"Nov 26, 2025",8,151,https://arxiv.org/abs/2511.21688 | |
| 331,"CogVLA: Cognition-Aligned Vision-Language-Action Model via | |
| Instruction-Driven Routing & Sparsification",18,31,39.22,39,"Aug 28, 2025",8,56,https://arxiv.org/abs/2508.21046 | |
| 332,DyPE: Dynamic Position Extrapolation for Ultra High Resolution Diffusion,14,26,35.86,31,"Oct 23, 2025",33,248,https://arxiv.org/abs/2510.20766 | |
| 333,Hunyuan-MT Technical Report,14,26,36,35,"Sep 5, 2025",13,517,https://arxiv.org/abs/2509.05209 | |
| 334,CrossOver: 3D Scene Cross-Modal Alignment,8,4,24.75,22,"Feb 20, 2025",2,204,https://arxiv.org/abs/2502.15011 | |
| 335,FullPart: Generating each 3D Part at Full Resolution,8,15,25.12,22,"Oct 30, 2025",4,57,https://arxiv.org/abs/2510.26140 | |
| 336,"Beyond Pass@1: Self-Play with Variational Problem Synthesis Sustains | |
| RLVR",8,14,25.25,21,"Aug 19, 2025",109,26,https://arxiv.org/abs/2508.14029 | |
| 337,RegionE: Adaptive Region-Aware Generation for Efficient Image Editing,9,14,28.22,24,"Oct 29, 2025",24,46,https://arxiv.org/abs/2510.25590 | |
| 338,pi-Flow: Policy-Based Few-Step Generation via Imitation Distillation,13,26,35.38,33,"Oct 16, 2025",7,97,https://arxiv.org/abs/2510.14974 | |
| 339,Reinforcement Learning Foundations for Deep Research Systems: A Survey,8,15,25.88,24,"Sep 8, 2025",25,10,https://arxiv.org/abs/2509.06733 | |
| 340,Efficient Part-level 3D Object Generation via Dual Volume Packing,13,24,35.62,36,"Jun 11, 2025",8,701,https://arxiv.org/abs/2506.09980 | |
| 341,3D Gaussian Splatting for Real-Time Radiance Field Rendering,41,41,46.12,46,"Aug 8, 2023",192,19600,https://arxiv.org/abs/2308.04079 | |
| 342,A Style is Worth One Code: Unlocking Code-to-Style Image Generation with Discrete Style Space,9,19,29.67,33,"Nov 13, 2025",53,122,https://arxiv.org/abs/2511.10555 | |
| 343,"A Vision-Language-Action-Critic Model for Robotic Real-World | |
| Reinforcement Learning",11,21,33.64,33,"Sep 19, 2025",16,145,https://arxiv.org/abs/2509.15937 | |
| 344,"LivePortrait: Efficient Portrait Animation with Stitching and | |
| Retargeting Control",29,39,44.48,44,"Jul 3, 2024",3,16900,https://arxiv.org/abs/2407.03168 | |
| 345,"MMR1: Enhancing Multimodal Reasoning with Variance-Aware Sampling and | |
| Open Resources",11,24,33.91,31,"Sep 25, 2025",90,190,https://arxiv.org/abs/2509.21268 | |
| 346,"Time Is a Feature: Exploiting Temporal Dynamics in Diffusion Language | |
| Models",13,25,36.62,34,"Aug 12, 2025",31,38,https://arxiv.org/abs/2508.09138 | |
| 347,Yume: An Interactive World Generation Model,4,3,5.25,4,"Jul 23, 2025",69,185,https://arxiv.org/abs/2507.17744 | |
| 348,Fast-dLLM v2: Efficient Block-Diffusion LLM,7,7,25,22,"Sep 30, 2025",40,537,https://arxiv.org/abs/2509.26328 | |
| 349,"DeepScientist: Advancing Frontier-Pushing Scientific Findings | |
| Progressively",11,26,34.64,32,"Sep 30, 2025",16,119,https://arxiv.org/abs/2509.26603 | |
| 350,GigaBrain-0: A World Model-Powered Vision-Language-Action Model,13,30,37.15,36,"Oct 22, 2025",46,218,https://arxiv.org/abs/2510.19430 | |
| 351,Visual Jigsaw Post-Training Improves MLLMs,8,20,28.62,26,"Sep 29, 2025",34,29,https://arxiv.org/abs/2509.25190 | |
| 352,GenExam: A Multidisciplinary Text-to-Image Exam,8,17,28.75,28,"Sep 17, 2025",16,17,https://arxiv.org/abs/2509.14232 | |
| 353,"CapRL: Stimulating Dense Image Caption Capabilities via Reinforcement | |
| Learning",8,11,29.12,28,"Sep 26, 2025",30,62,https://arxiv.org/abs/2509.22647 | |
| 354,"InternSVG: Towards Unified SVG Tasks with Multimodal Large Language | |
| Models",6,10,21.83,25,"Oct 13, 2025",31,54,https://arxiv.org/abs/2510.11341 | |
| 355,"UniMoE-Audio: Unified Speech and Music Generation with Dynamic-Capacity | |
| MoE",12,32,36.5,35,"Oct 15, 2025",62,1010,https://arxiv.org/abs/2510.13344 | |
| 356,"VITA-E: Natural Embodied Interaction with Concurrent Seeing, Hearing, | |
| Speaking, and Acting",8,20,29.5,25,"Oct 21, 2025",41,120,https://arxiv.org/abs/2510.21817 | |
| 357,EdgeTAM: On-Device Track Anything Model,16,31,40.38,40,"Jan 13, 2025",1,757,https://arxiv.org/abs/2501.07256 | |
| 358,"Spatial-SSRL: Enhancing Spatial Understanding via Self-Supervised | |
| Reinforcement Learning",11,25,36.09,34,"Oct 31, 2025",25,54,https://arxiv.org/abs/2510.27606 | |
| 359,Self-Rewarding Vision-Language Model via Reasoning Decomposition,16,33,40.88,42,"Aug 27, 2025",77,79,https://arxiv.org/abs/2508.19652 | |
| 360,Uni-MoE: Scaling Unified Multimodal LLMs with Mixture of Experts,11,32,36.27,35,"May 18, 2024",19,1010,https://arxiv.org/abs/2405.11273 | |
| 361,"U-Bench: A Comprehensive Understanding of U-Net through 100-Variant | |
| Benchmarking",8,21,31,28,"Oct 8, 2025",3,65,https://arxiv.org/abs/2510.07041 | |
| 362,VChain: Chain-of-Visual-Thought for Reasoning in Video Generation,9,25,33.33,29,"Oct 6, 2025",34,60,https://arxiv.org/abs/2510.05094 | |
| 363,"EasySteer: A Unified Framework for High-Performance and Extensible LLM | |
| Steering",9,23,33.56,30,"Sep 29, 2025",25,49,https://arxiv.org/abs/2509.25175 | |
| 364,LongCodeZip: Compress Long Context for Code Language Models,8,23,31.38,30,"Oct 1, 2025",88,63,https://arxiv.org/abs/2510.00446 | |
| 365,"ARTDECO: Towards Efficient and High-Fidelity On-the-Fly 3D | |
| Reconstruction with Structured Scene Representation",9,25,33.56,32,"Oct 9, 2025",30,73,https://arxiv.org/abs/2510.08551 | |
| 366,DeCo: Frequency-Decoupled Pixel Diffusion for End-to-End Image Generation,10,25,35.4,33,"Nov 24, 2025",58,75,https://arxiv.org/abs/2511.19365 | |
| 367,π^3: Scalable Permutation-Equivariant Visual Geometry Learning,6,10,25.17,17,"Jul 17, 2025",64,1100,https://arxiv.org/abs/2507.13347 | |
| 368,LoopTool: Closing the Data-Training Loop for Robust LLM Tool Calls,9,21,33.78,39,"Nov 12, 2025",15,20,https://arxiv.org/abs/2511.09148 | |
| 369,Explain Before You Answer: A Survey on Compositional Visual Reasoning,19,35,42.89,44,"Aug 24, 2025",4,279,https://arxiv.org/abs/2508.17298 | |
| 370,IMG: Calibrating Diffusion Models via Implicit Multimodal Guidance,7,16,29.14,21,"Sep 30, 2025",16,25,https://arxiv.org/abs/2509.26231 | |
| 371,ObjectClear: Complete Object Removal via Object-Effect Attention,4,8,13,11,"May 28, 2025",1,328,https://arxiv.org/abs/2505.22636 | |
| 372,"LLaVA-OneVision-1.5: Fully Open Framework for Democratized Multimodal | |
| Training",16,33,41.5,41,"Sep 28, 2025",40,508,https://arxiv.org/abs/2509.23661 | |
| 373,"Stable Video Infinity: Infinite-Length Video Generation with Error | |
| Recycling",19,34,43.05,45,"Oct 10, 2025",14,388,https://arxiv.org/abs/2510.09212 | |
| 374,SAC: Neural Speech Codec with Semantic-Acoustic Dual-Stream Quantization,7,16,29.57,31,"Oct 19, 2025",0,60,https://arxiv.org/abs/2510.16841 | |
| 375,ReasonRank: Empowering Passage Ranking with Strong Reasoning Ability,13,28,39.54,40,"Aug 9, 2025",109,101,https://arxiv.org/abs/2508.07050 | |
| 376,DiT360: High-Fidelity Panoramic Image Generation via Hybrid Training,9,21,34.56,38,"Oct 13, 2025",29,105,https://arxiv.org/abs/2510.11712 | |
| 377,AnyUp: Universal Feature Upsampling,11,31,37.73,37,"Oct 14, 2025",10,268,https://arxiv.org/abs/2510.12764 | |
| 378,"Mixture-of-Recursions: Learning Dynamic Recursive Depths for Adaptive | |
| Token-Level Computation",4,11,15,13,"Jul 14, 2025",60,311,https://arxiv.org/abs/2507.10524 | |
| 379,Performance Prediction for Large Systems via Text-to-Text Regression,15,29,41.47,45,"Jun 26, 2025",6,255,https://arxiv.org/abs/2506.21718 | |
| 380,Sequential Diffusion Language Models,7,22,30.71,29,"Sep 28, 2025",29,29,https://arxiv.org/abs/2509.24007 | |
| 381,TiViBench: Benchmarking Think-in-Video Reasoning for Video Generative Models,5,17,22.6,23,"Nov 17, 2025",40,50,https://arxiv.org/abs/2511.13704 | |
| 382,Gradio: Hassle-Free Sharing and Testing of ML Models in the Wild,10,32,36.9,35,"Jun 6, 2019",1,40700,https://arxiv.org/abs/1906.02569 | |
| 383,"Elevating 3D Models: High-Quality Texture and Geometry Refinement from a | |
| Low-Quality Model",4,9,17.25,15,"Jul 15, 2025",11,106,https://arxiv.org/abs/2507.11465 | |
| 384,"Patch-as-Decodable-Token: Towards Unified Multi-Modal Vision Tasks in | |
| MLLMs",7,23,31.86,29,"Oct 2, 2025",12,183,https://arxiv.org/abs/2510.01954 | |
| 385,"More Thought, Less Accuracy? On the Dual Nature of Reasoning in | |
| Vision-Language Models",9,26,36.33,35,"Sep 30, 2025",56,44,https://arxiv.org/abs/2509.25848 | |
| 386,"Open-o3 Video: Grounded Video Reasoning with Explicit Spatio-Temporal | |
| Evidence",9,22,36.44,34,"Oct 23, 2025",46,63,https://arxiv.org/abs/2510.20579 | |
| 387,AutoPR: Let's Automate Your Academic Promotion!,5,8,25.2,26,"Oct 10, 2025",43,36,https://arxiv.org/abs/2510.09558 | |
| 388,"VMem: Consistent Interactive Video Scene Generation with Surfel-Indexed | |
| View Memory",4,18,19,18,"Jun 23, 2025",22,255,https://arxiv.org/abs/2506.18903 | |
| 389,RLP: Reinforcement as a Pretraining Objective,9,25,36.78,34,"Sep 26, 2025",32,149,https://arxiv.org/abs/2510.01265 | |
| 390,"Gated Attention for Large Language Models: Non-linearity, Sparsity, and | |
| Attention-Sink-Free",10,31,38.2,36,"May 10, 2025",6,240,https://arxiv.org/abs/2505.06708 | |
| 391,"CMPhysBench: A Benchmark for Evaluating Large Language Models in | |
| Condensed Matter Physics",5,13,25.6,19,"Aug 25, 2025",45,16,https://arxiv.org/abs/2508.18124 | |
| 392,Spotlight on Token Perception for Multimodal Reinforcement Learning,6,10,29.83,27,"Oct 10, 2025",31,26,https://arxiv.org/abs/2510.09285 | |
| 393,Video-as-Answer: Predict and Generate Next Video Event with Joint-GRPO,6,23,29.83,29,"Nov 20, 2025",29,44,https://arxiv.org/abs/2511.16669 | |
| 394,"SyGra: A Unified Graph-Based Framework for Scalable Generation, Quality | |
| Tagging, and Management of Synthetic Data",6,16,30,28,"Aug 21, 2025",6,16,https://arxiv.org/abs/2508.15432 | |
| 395,"Uni-cot: Towards Unified Chain-of-Thought Reasoning Across Text and | |
| Vision",7,24,33.29,32,"Aug 7, 2025",0,136,https://arxiv.org/abs/2508.05606 | |
| 396,SRUM: Fine-Grained Self-Rewarding for Unified Multimodal Models,9,15,37.22,44,"Oct 14, 2025",17,51,https://arxiv.org/abs/2510.12784 | |
| 397,"JanusCoder: Towards a Foundational Visual-Programmatic Interface for | |
| Code Intelligence",7,19,33.29,38,"Oct 27, 2025",90,54,https://arxiv.org/abs/2510.23538 | |
| 398,"SearchInstruct: Enhancing Domain Adaptation via Retrieval-Based | |
| Instruction Dataset Creation",5,18,26.4,26,"Sep 12, 2025",9,8,https://arxiv.org/abs/2509.10708 | |
| 399,RynnVLA-002: A Unified Vision-Language-Action and World Model,11,29,39.82,41,"Nov 21, 2025",24,669,https://arxiv.org/abs/2511.17502 | |
| 400,BitNet Distillation,13,35,41.62,41,"Oct 15, 2025",44,24300,https://arxiv.org/abs/2510.13998 | |
| 401,UniREditBench: A Unified Reasoning-based Image Editing Benchmark,8,28,35.88,32,"Nov 3, 2025",35,37,https://arxiv.org/abs/2511.01295 | |
| 402,Artificial Hippocampus Networks for Efficient Long-Context Modeling,6,24,31,28,"Oct 8, 2025",22,67,https://arxiv.org/abs/2510.07318 | |
| 403,Agentic Entropy-Balanced Policy Optimization,8,30,36.12,33,"Oct 16, 2025",90,694,https://arxiv.org/abs/2510.14545 | |
| 404,"Pass@k Training for Adaptively Balancing Exploration and Exploitation of | |
| Large Reasoning Models",21,40,45.43,45,"Aug 14, 2025",21,47,https://arxiv.org/abs/2508.10751 | |
| 405,"URSA: Understanding and Verifying Chain-of-thought Reasoning in | |
| Multimodal Mathematics",6,26,31.83,32,"Jan 8, 2025",53,117,https://arxiv.org/abs/2501.04686 | |
| 406,"MCPMark: A Benchmark for Stress-Testing Realistic and Comprehensive MCP | |
| Use",8,32,36.62,35,"Sep 28, 2025",145,201,https://arxiv.org/abs/2509.24002 | |
| 407,"AMFT: Aligning LLM Reasoners by Meta-Learning the Optimal | |
| Imitation-Exploration Balance",16,31,43.88,45,"Aug 9, 2025",2,163,https://arxiv.org/abs/2508.06944 | |
| 408,"Diffuman4D: 4D Consistent Human View Synthesis from Sparse-View Videos | |
| with Spatio-Temporal Diffusion Models",4,14,22.75,25,"Jul 17, 2025",50,246,https://arxiv.org/abs/2507.13344 | |
| 409,"Cache-to-Cache: Direct Semantic Communication Between Large Language | |
| Models",10,35,39.8,40,"Oct 3, 2025",95,239,https://arxiv.org/abs/2510.03215 | |
| 410,"SageAttention3: Microscaling FP4 Attention for Inference and An | |
| Exploration of 8-Bit Training",12,26,41.75,43,"May 16, 2025",76,2450,https://arxiv.org/abs/2505.11594 | |
| 411,NANO3D: A Training-Free Approach for Efficient 3D Editing Without Masks,5,22,28.8,25,"Oct 16, 2025",52,47,https://arxiv.org/abs/2510.15019 | |
| 412,"MeshSplat: Generalizable Sparse-View Surface Reconstruction via Gaussian | |
| Splatting",6,22,32.67,28,"Aug 25, 2025",4,18,https://arxiv.org/abs/2508.17811 | |
| 413,"Drawing2CAD: Sequence-to-Sequence Learning for CAD Generation from | |
| Vector Drawings",8,22,37.38,41,"Aug 26, 2025",2,50,https://arxiv.org/abs/2508.18733 | |
| 414,OceanGym: A Benchmark Environment for Underwater Embodied Agents,5,22,29.2,28,"Sep 30, 2025",30,28,https://arxiv.org/abs/2509.26536 | |
| 415,GR00T N1: An Open Foundation Model for Generalist Humanoid Robots,13,27,42.77,47,"Mar 18, 2025",4,4800,https://arxiv.org/abs/2503.14734 | |
| 416,TexVerse: A Universe of 3D Objects with High-Resolution Textures,12,37,42.08,40,"Aug 14, 2025",13,180,https://arxiv.org/abs/2508.10868 | |
| 417,"GSM8K-V: Can Vision Language Models Solve Grade School Math Word | |
| Problems in Visual Contexts",6,24,33.17,31,"Sep 29, 2025",25,21,https://arxiv.org/abs/2509.25160 | |
| 418,"TOUCAN: Synthesizing 1.5M Tool-Agentic Data from Real-World MCP | |
| Environments",9,30,39.11,41,"Oct 1, 2025",16,60,https://arxiv.org/abs/2510.01179 | |
| 419,"CoDiEmb: A Collaborative yet Distinct Framework for Unified | |
| Representation Learning in Information Retrieval and Semantic Textual | |
| Similarity",10,36,40.3,39,"Aug 15, 2025",3,125,https://arxiv.org/abs/2508.11442 | |
| 420,3D-R1: Enhancing Reasoning in 3D VLMs for Unified Scene Understanding,10,27,40.4,42,"Jul 31, 2025",15,307,https://arxiv.org/abs/2507.23478 | |
| 421,"VLA-RFT: Vision-Language-Action Reinforcement Fine-tuning with Verified | |
| Rewards in World Simulators",5,20,30,25,"Oct 1, 2025",55,28,https://arxiv.org/abs/2510.00406 | |
| 422,"SeC: Advancing Complex Video Object Segmentation via Progressive Concept | |
| Construction",3,12,16.67,19,"Jul 21, 2025",34,98,https://arxiv.org/abs/2507.15852 | |
| 423,Aligning Multimodal LLM with Human Preference: A Survey,20,31,45.95,48,"Mar 18, 2025",26,16300,https://arxiv.org/abs/2503.14504 | |
| 424,Skywork R1V: Pioneering Multimodal Reasoning with Chain-of-Thought,8,23,38.5,42,"Apr 8, 2025",86,3080,https://arxiv.org/abs/2504.05599 | |
| 425,"HierSearch: A Hierarchical Enterprise Deep Search Framework Integrating | |
| Local and Web Searches",8,33,38.5,37,"Aug 11, 2025",26,30,https://arxiv.org/abs/2508.08088 | |
| 426,SIM-CoT: Supervised Implicit Chain-of-Thought,8,33,38.5,34,"Sep 24, 2025",33,46,https://arxiv.org/abs/2509.20317 | |
| 427,"RewardMap: Tackling Sparse Rewards in Fine-grained Visual Reasoning via | |
| Multi-Stage Reinforcement Learning",5,18,31.2,29,"Oct 2, 2025",15,22,https://arxiv.org/abs/2510.02240 | |
| 428,Generating Physically Stable and Buildable LEGO Designs from Text,7,27,37,41,"May 8, 2025",28,1430,https://arxiv.org/abs/2505.05469 | |
| 429,"Sparse VideoGen2: Accelerate Video Generation with Sparse Attention via | |
| Semantic-Aware Permutation",8,35,39.12,38,"May 24, 2025",42,472,https://arxiv.org/abs/2505.18875 | |
| 430,"Sparse VideoGen: Accelerating Video Diffusion Transformers with | |
| Spatial-Temporal Sparsity",8,34,39.12,38,"Feb 3, 2025",3,472,https://arxiv.org/abs/2502.01776 | |
| 431,"Vision-Zero: Scalable VLM Self-Improvement via Strategic Gamified | |
| Self-Play",9,29,40.44,44,"Sep 29, 2025",120,40,https://arxiv.org/abs/2509.25541 | |
| 432,"OmniVideoBench: Towards Audio-Visual Understanding Evaluation for Omni | |
| MLLMs",6,23,35.17,33,"Oct 12, 2025",40,25,https://arxiv.org/abs/2510.10689 | |
| 433,"E^2Rank: Your Text Embedding can Also be an Effective | |
| and Efficient Listwise Reranker",5,29,32,33,"Oct 26, 2025",29,22,https://arxiv.org/abs/2510.22733 | |
| 434,"LIBERO-Plus: In-depth Robustness Analysis of Vision-Language-Action | |
| Models",9,30,40.56,41,"Oct 15, 2025",40,46,https://arxiv.org/abs/2510.13626 | |
| 435,Drax: Speech Recognition with Discrete Flow Matching,4,21,27.75,21,"Oct 5, 2025",23,18,https://arxiv.org/abs/2510.04162 | |
| 436,"OS-Sentinel: Towards Safety-Enhanced Mobile GUI Agents via Hybrid | |
| Validation in Realistic Workflows",9,29,40.67,45,"Oct 28, 2025",69,31,https://arxiv.org/abs/2510.24411 | |
| 437,Symbolic Graphics Programming with Large Language Models,10,33,42,44,"Sep 5, 2025",37,18,https://arxiv.org/abs/2509.05208 | |
| 438,SparseD: Sparse Attention for Diffusion Language Models,6,30,36,34,"Sep 28, 2025",26,33,https://arxiv.org/abs/2509.24014 | |
| 439,"Language Model Council: Benchmarking Foundation Models on Highly | |
| Subjective Tasks by Consensus",14,35,44.57,46,"Jun 12, 2024",6,180,https://arxiv.org/abs/2406.08598 | |
| 440,"MM-HELIX: Boosting Multimodal Long-Chain Reflective Reasoning with | |
| Holistic Platform and Adaptive Hybrid Policy Optimization",7,33,38.43,37,"Oct 9, 2025",98,55,https://arxiv.org/abs/2510.08540 | |
| 441,PICABench: How Far Are We from Physically Realistic Image Editing?,3,12,21.67,15,"Oct 20, 2025",57,13,https://arxiv.org/abs/2510.17681 | |
| 442,"The Tool Decathlon: Benchmarking Language Agents for Diverse, Realistic, | |
| and Long-Horizon Task Execution",10,38,42.4,42,"Oct 29, 2025",42,99,https://arxiv.org/abs/2510.25726 | |
| 443,Next Visual Granularity Generation,7,33,38.86,35,"Aug 18, 2025",37,7,https://arxiv.org/abs/2508.12811 | |
| 444,Droplet3D: Commonsense Priors from Videos Facilitate 3D Generation,8,35,40.5,39,"Aug 28, 2025",62,18,https://arxiv.org/abs/2508.20470 | |
| 445,OmniGen2: Exploration to Advanced Multimodal Generation,10,32,42.7,42,"Jun 23, 2025",75,3860,https://arxiv.org/abs/2506.18871 | |
| 446,"JavisDiT: Joint Audio-Video Diffusion Transformer with Hierarchical | |
| Spatio-Temporal Prior Synchronization",16,35,45.88,47,"Mar 30, 2025",57,265,https://arxiv.org/abs/2503.23377 | |
| 447,"UP2You: Fast Reconstruction of Yourself from Unconstrained Photo | |
| Collections",7,22,39.29,45,"Sep 29, 2025",8,161,https://arxiv.org/abs/2509.24817 | |
| 448,"LLaSO: A Foundational Framework for Reproducible Research in Large | |
| Language and Speech Model",7,28,39.43,36,"Aug 21, 2025",4,79,https://arxiv.org/abs/2508.15418 | |
| 449,"MUG-V 10B: High-efficiency Training Pipeline for Large Video Generation | |
| Models",3,18,24,24,"Oct 20, 2025",9,59,https://arxiv.org/abs/2510.17519 | |
| 450,"PartCrafter: Structured 3D Mesh Generation via Compositional Latent | |
| Diffusion Transformers",7,23,39.71,40,"Jun 5, 2025",79,2180,https://arxiv.org/abs/2506.05573 | |
| 451,Discrete Diffusion in Large Language and Multimodal Models: A Survey,13,36,44.92,46,"Jun 16, 2025",43,267,https://arxiv.org/abs/2506.13759 | |
| 452,"PosterGen: Aesthetic-Aware Paper-to-Poster Generation via Multi-Agent | |
| LLMs",11,35,43.91,43,"Aug 24, 2025",15,105,https://arxiv.org/abs/2508.17188 | |
| 453,"Implicit Actor Critic Coupling via a Supervised Learning Framework for | |
| RLVR",4,28,31.75,32,"Sep 2, 2025",19,18,https://arxiv.org/abs/2509.02522 | |
| 454,UniME-V2: MLLM-as-a-Judge for Universal Multimodal Embedding Learning,6,27,38.17,39,"Oct 15, 2025",11,26,https://arxiv.org/abs/2510.13515 | |
| 455,Part-X-MLLM: Part-aware 3D Multimodal Large Language Model,3,25,25.33,25,"Nov 17, 2025",65,49,https://arxiv.org/abs/2511.13647 | |
| 456,Scaling RL to Long Videos,3,21,25.67,28,"Jul 10, 2025",141,538,https://arxiv.org/abs/2507.07966 | |
| 457,"X-VLA: Soft-Prompted Transformer as Scalable Cross-Embodiment | |
| Vision-Language-Action Model",7,32,40.14,42,"Oct 11, 2025",12,56,https://arxiv.org/abs/2510.10274 | |
| 458,"Promptomatix: An Automatic Prompt Optimization Framework for Large | |
| Language Models",3,19,26.33,30,"Jul 17, 2025",13,61,https://arxiv.org/abs/2507.14241 | |
| 459,Optimized Minimal 4D Gaussian Splatting,5,30,36.4,35,"Oct 4, 2025",3,29,https://arxiv.org/abs/2510.03857 | |
| 460,Defeating the Training-Inference Mismatch via FP16,9,33,42.89,47,"Oct 30, 2025",23,112,https://arxiv.org/abs/2510.26788 | |
| 461,"Teaching Pretrained Language Models to Think Deeper with Retrofitted | |
| Recurrence",7,36,40.57,36,"Nov 10, 2025",11,23,https://arxiv.org/abs/2511.07384 | |
| 462,RynnVLA-001: Using Human Demonstrations to Improve Robot Manipulation,7,32,40.71,41,"Sep 18, 2025",20,208,https://arxiv.org/abs/2509.15212 | |
| 463,MediaPipe: A Framework for Building Perception Pipelines,11,39,44.45,45,"Jun 14, 2019",1,32100,https://arxiv.org/abs/1906.08172 | |
| 464,PPTAgent: Generating and Evaluating Presentations Beyond Text-to-Slides,2,15,15,15,"Jan 7, 2025",22,2430,https://arxiv.org/abs/2501.03936 | |
| 465,WideSearch: Benchmarking Agentic Broad Info-Seeking,7,37,40.86,41,"Aug 11, 2025",96,62,https://arxiv.org/abs/2508.07999 | |
| 466,"Efficient Multi-modal Large Language Models via Progressive Consistency | |
| Distillation",4,23,33.25,33,"Oct 1, 2025",37,13,https://arxiv.org/abs/2510.00515 | |
| 467,UnSAMv2: Self-Supervised Learning Enables Segment Anything at Any Granularity,3,27,28,28,"Nov 17, 2025",4,34,https://arxiv.org/abs/2511.13714 | |
| 468,"SonicMaster: Towards Controllable All-in-One Music Restoration and | |
| Mastering",7,32,41.29,39,"Aug 5, 2025",1,74,https://arxiv.org/abs/2508.03448 | |
| 469,DINO-Foresight: Looking into the Future with DINO,7,39,41.29,41,"Dec 16, 2024",1,107,https://arxiv.org/abs/2412.11673 | |
| 470,"Euclid's Gift: Enhancing Spatial Perception and Reasoning in | |
| Vision-Language Models via Geometric Surrogate Tasks",4,29,34.25,31,"Sep 29, 2025",15,13,https://arxiv.org/abs/2509.24473 | |
| 471,CoDA: Coding LM via Diffusion Adaptation,5,34,37.6,37,"Sep 27, 2025",25,21,https://arxiv.org/abs/2510.03270 | |
| 472,"MiniMax-M1: Scaling Test-Time Compute Efficiently with Lightning | |
| Attention",8,34,42.75,44,"Jun 16, 2025",265,2890,https://arxiv.org/abs/2506.13585 | |
| 473,"Hyperspherical Latents Improve Continuous-Token Autoregressive | |
| Generation",6,30,40,42,"Sep 29, 2025",6,43,https://arxiv.org/abs/2509.24335 | |
| 474,VLA-0: Building State-of-the-Art VLAs with Zero Modification,6,33,40,39,"Oct 15, 2025",8,103,https://arxiv.org/abs/2510.13054 | |
| 475,"Self Forcing: Bridging the Train-Test Gap in Autoregressive Video | |
| Diffusion",17,41,47.18,49,"Jun 9, 2025",27,2420,https://arxiv.org/abs/2506.08009 | |
| 476,"RealUnify: Do Unified Models Truly Benefit from Unification? A | |
| Comprehensive Benchmark",5,29,38,38,"Sep 29, 2025",41,14,https://arxiv.org/abs/2509.24897 | |
| 477,Uniform Discrete Diffusion with Metric Path for Video Generation,5,24,38,41,"Oct 28, 2025",39,49,https://arxiv.org/abs/2510.24717 | |
| 478,"InternVLA-M1: A Spatially Guided Vision-Language-Action Framework for | |
| Generalist Robot Policy",7,35,42,44,"Oct 15, 2025",13,167,https://arxiv.org/abs/2510.13778 | |
| 479,Vidi: Large Multimodal Models for Video Understanding and Editing,4,27,35.5,37,"Apr 22, 2025",14,247,https://arxiv.org/abs/2504.15681 | |
| 480,UniVid: Unifying Vision Tasks with Pre-trained Video Generation Models,5,23,38.8,48,"Sep 26, 2025",11,25,https://arxiv.org/abs/2509.21760 | |
| 481,Sparser Block-Sparse Attention via Token Permutation,3,22,31,26,"Oct 24, 2025",22,19,https://arxiv.org/abs/2510.21270 | |
| 482,DeepEyesV2: Toward Agentic Multimodal Model,8,39,43.5,45,"Nov 7, 2025",35,963,https://arxiv.org/abs/2511.05271 | |
| 483,Rethinking Reward Models for Multi-Domain Test-Time Scaling,3,28,31.33,33,"Oct 1, 2025",23,14,https://arxiv.org/abs/2510.00492 | |
| 484,Depth Anything with Any Prior,6,37,41.17,40,"May 15, 2025",12,389,https://arxiv.org/abs/2505.10565 | |
| 485,AU-Harness: An Open-Source Toolkit for Holistic Evaluation of Audio LLMs,4,30,36.5,33,"Sep 9, 2025",18,35,https://arxiv.org/abs/2509.08031 | |
| 486,NORA-1.5: A Vision-Language-Action Model Trained using World Model- and Action-based Preference Rewards,3,31,31.67,32,"Nov 18, 2025",7,26,https://arxiv.org/abs/2511.14659 | |
| 487,"Learning to Optimize Multi-Objective Alignment Through Dynamic Reward | |
| Weighting",4,35,37.5,35,"Sep 14, 2025",8,4,https://arxiv.org/abs/2509.11452 | |
| 488,UQ: Assessing Language Models on Unsolved Questions,5,27,40.4,45,"Aug 25, 2025",11,15,https://arxiv.org/abs/2508.17580 | |
| 489,"MAS-Bench: A Unified Benchmark for Shortcut-Augmented Hybrid Mobile GUI | |
| Agents",4,27,37.75,40,"Sep 8, 2025",2,8,https://arxiv.org/abs/2509.06477 | |
| 490,"ObjFiller-3D: Consistent Multi-view 3D Inpainting via Video Diffusion | |
| Models",4,25,38,42,"Aug 25, 2025",5,25,https://arxiv.org/abs/2508.18271 | |
| 491,"RenderFormer: Transformer-based Neural Rendering of Triangle Meshes with | |
| Global Illumination",5,34,40.6,39,"May 28, 2025",37,765,https://arxiv.org/abs/2505.21925 | |
| 492,Upsample Anything: A Simple and Hard to Beat Baseline for Feature Upsampling,9,38,45.22,46,"Nov 20, 2025",6,105,https://arxiv.org/abs/2511.16301 | |
| 493,Foundations of Large Language Models,5,31,40.8,40,"Jan 16, 2025",11,264,https://arxiv.org/abs/2501.09223 | |
| 494,DanceGRPO: Unleashing GRPO on Visual Generation,11,37,46.45,47,"May 12, 2025",32,942,https://arxiv.org/abs/2505.07818 | |
| 495,SPARK: Synergistic Policy And Reward Co-Evolving Framework,3,24,34.33,30,"Sep 26, 2025",15,17,https://arxiv.org/abs/2509.22624 | |
| 496,WEAVE: Unleashing and Benchmarking the In-context Interleaved Comprehension and Generation,4,30,38.5,38,"Nov 14, 2025",42,25,https://arxiv.org/abs/2511.11434 | |
| 497,"Eliciting Fine-Tuned Transformer Capabilities via Inference-Time | |
| Techniques",7,34,44,48,"Jun 9, 2025",8,2910,https://arxiv.org/abs/2506.08060 | |
| 498,SpatialTrackerV2: 3D Point Tracking Made Easy,3,32,35.33,32,"Jul 16, 2025",14,658,https://arxiv.org/abs/2507.12462 | |
| 499,Reasoning with Sampling: Your Base Model is Smarter Than You Think,3,27,35.33,33,"Oct 16, 2025",35,229,https://arxiv.org/abs/2510.14901 | |
| 500,"DocETL: Agentic Query Rewriting and Evaluation for Complex Document | |
| Processing",5,37,41.6,43,"Oct 16, 2024",1,3180,https://arxiv.org/abs/2410.12189 | |
| 501,Nav-R1: Reasoning and Navigation in Embodied Scenes,6,39,43.33,43,"Sep 13, 2025",4,21,https://arxiv.org/abs/2509.10884 | |
| 502,"EPO: Entropy-regularized Policy Optimization for LLM Agents | |
| Reinforcement Learning",4,23,39.5,44,"Sep 26, 2025",113,20,https://arxiv.org/abs/2509.22576 | |
| 503,"PUSA V1.0: Surpassing Wan-I2V with $500 Training Cost by Vectorized | |
| Timestep Adaptation",4,28,39.75,42,"Jul 22, 2025",9,545,https://arxiv.org/abs/2507.16116 | |
| 504,"Efficient Multi-turn RL for GUI Agents via Decoupled Training and | |
| Adaptive Data Curation",3,32,36,35,"Sep 28, 2025",7,9,https://arxiv.org/abs/2509.23866 | |
| 505,OBS-Diff: Accurate Pruning For Diffusion Models in One-Shot,2,26,28.5,28,"Oct 8, 2025",14,24,https://arxiv.org/abs/2510.06751 | |
| 506,GUI-G^2: Gaussian Reward Modeling for GUI Grounding,2,24,29,29,"Jul 21, 2025",118,138,https://arxiv.org/abs/2507.15846 | |
| 507,GIR-Bench: Versatile Benchmark for Generating Images with Reasoning,3,29,37,39,"Oct 13, 2025",16,23,https://arxiv.org/abs/2510.11026 | |
| 508,Beyond English: Toward Inclusive and Scalable Multilingual Machine Translation with LLMs,7,42,45,46,"Nov 10, 2025",31,92,https://arxiv.org/abs/2511.07003 | |
| 509,"Embodied-R1: Reinforced Embodied Reasoning for General Robotic | |
| Manipulation",4,31,40.75,42,"Aug 19, 2025",12,24,https://arxiv.org/abs/2508.13998 | |
| 510,"SwiftEdit: Lightning Fast Text-Guided Image Editing via One-Step | |
| Diffusion",3,36,37.33,36,"Dec 5, 2024",41,7,https://arxiv.org/abs/2412.04301 | |
| 511,CLEAR: Error Analysis via LLM-as-a-Judge Made Easy,1,11,11,11,"Jul 24, 2025",8,10,https://arxiv.org/abs/2507.18392 | |
| 512,MindSearch: Mimicking Human Minds Elicits Deep AI Searcher,7,43,45.29,46,"Jul 29, 2024",44,6580,https://arxiv.org/abs/2407.20183 | |
| 513,dParallel: Learnable Parallel Decoding for dLLMs,5,40,43,42,"Sep 30, 2025",17,16,https://arxiv.org/abs/2509.26488 | |
| 514,"ReCogDrive: A Reinforced Cognitive Framework for End-to-End Autonomous | |
| Driving",8,39,46.12,47,"Jun 9, 2025",1,183,https://arxiv.org/abs/2506.08052 | |
| 515,AWorld: Orchestrating the Training Recipe for Agentic AI,7,36,45.43,46,"Aug 28, 2025",37,692,https://arxiv.org/abs/2508.20404 | |
| 516,Reverse-Engineered Reasoning for Open-Ended Generation,4,38,41.25,40,"Sep 7, 2025",127,34,https://arxiv.org/abs/2509.06160 | |
| 517,MiniCPM4: Ultra-Efficient LLMs on End Devices,7,41,45.57,47,"Jun 9, 2025",90,8300,https://arxiv.org/abs/2506.07900 | |
| 518,SpaceVista: All-Scale Visual Spatial Reasoning from mm to km,2,30,32,32,"Oct 10, 2025",16,22,https://arxiv.org/abs/2510.09606 | |
| 519,Vlaser: Vision-Language-Action Model with Synergistic Embodied Reasoning,3,30,38.67,41,"Oct 13, 2025",14,15,https://arxiv.org/abs/2510.11027 | |
| 520,"Generating an Image From 1,000 Words: Enhancing Text-to-Image With | |
| Structured Captions",5,38,43.6,41,"Nov 10, 2025",19,248,https://arxiv.org/abs/2511.06876 | |
| 521,"VideoGen-of-Thought: A Collaborative Framework for Multi-Shot Video | |
| Generation",4,38,42,41,"Dec 3, 2024",60,42,https://arxiv.org/abs/2412.02259 | |
| 522,"Spatial Forcing: Implicit Spatial Representation Alignment for | |
| Vision-language-action Model",4,39,42.25,41,"Oct 14, 2025",134,43,https://arxiv.org/abs/2510.12276 | |
| 523,MGM-Omni: Scaling Omni LLMs to Personalized Long-Horizon Speech,5,40,44.2,42,"Sep 29, 2025",11,133,https://arxiv.org/abs/2509.25131 | |
| 524,"Building a Foundational Guardrail for General Agentic Systems via | |
| Synthetic Data",3,33,39.67,42,"Oct 10, 2025",21,26,https://arxiv.org/abs/2510.09781 | |
| 525,Mantis: A Versatile Vision-Language-Action Model with Disentangled Visual Foresight,2,32,34,34,"Nov 20, 2025",10,18,https://arxiv.org/abs/2511.16175 | |
| 526,"Franca: Nested Matryoshka Clustering for Scalable Visual Representation | |
| Learning",4,28,42.75,47,"Jul 18, 2025",27,189,https://arxiv.org/abs/2507.14137 | |
| 527,"How Far are VLMs from Visual Spatial Intelligence? A Benchmark-Driven | |
| Perspective",3,37,40,37,"Sep 23, 2025",22,10,https://arxiv.org/abs/2509.18905 | |
| 528,Regression Language Models for Code,6,40,45.5,47,"Sep 30, 2025",13,257,https://arxiv.org/abs/2509.26476 | |
| 529,"CoIRL-AD: Collaborative-Competitive Imitation-Reinforcement Learning in | |
| Latent World Models for Autonomous Driving",3,36,40,42,"Oct 14, 2025",4,10,https://arxiv.org/abs/2510.12560 | |
| 530,"Scaling Text-Rich Image Understanding via Code-Guided Synthetic | |
| Multimodal Data Generation",2,35,35,35,"Feb 20, 2025",13,103,https://arxiv.org/abs/2502.14846 | |
| 531,"StreamDiffusion: A Pipeline-level Solution for Real-time Interactive | |
| Generation",2,34,35,35,"Dec 19, 2023",73,10400,https://arxiv.org/abs/2312.12491 | |
| 532,"LLMs Can Get ""Brain Rot""!",4,38,43,42,"Oct 15, 2025",19,72,https://arxiv.org/abs/2510.13928 | |
| 533,"Reinforcement Learning Optimization for Large-Scale Learning: An | |
| Efficient and User-Friendly Scaling Library",5,40,44.6,45,"Jun 6, 2025",7,2270,https://arxiv.org/abs/2506.06122 | |
| 534,ConsistEdit: Highly Consistent and Precise Training-free Visual Editing,2,35,35.5,35,"Oct 20, 2025",11,25,https://arxiv.org/abs/2510.17803 | |
| 535,"TalkVid: A Large-Scale Diversified Dataset for Audio-Driven Talking Head | |
| Synthesis",4,34,43.5,45,"Aug 19, 2025",16,72,https://arxiv.org/abs/2508.13618 | |
| 536,Energy-Based Transformers are Scalable Learners and Thinkers,8,44,47.25,47,"Jul 2, 2025",65,463,https://arxiv.org/abs/2507.02092 | |
| 537,"PromptCoT 2.0: Scaling Prompt Synthesis for Large Language Model | |
| Reasoning",3,28,41,47,"Sep 24, 2025",29,93,https://arxiv.org/abs/2509.19894 | |
| 538,Multi-Agent Tool-Integrated Policy Optimization,2,35,36,36,"Oct 6, 2025",19,20,https://arxiv.org/abs/2510.04678 | |
| 539,"VisionThink: Smart and Efficient Vision Language Model via Reinforcement | |
| Learning",3,33,41.33,45,"Jul 17, 2025",69,330,https://arxiv.org/abs/2507.13348 | |
| 540,Rectified Point Flow: Generic Point Cloud Pose Estimation,3,34,41.33,44,"Jun 5, 2025",3,123,https://arxiv.org/abs/2506.05282 | |
| 541,"Equilibrium Matching: Generative Modeling with Implicit Energy-Based | |
| Models",4,41,43.75,44,"Oct 2, 2025",5,77,https://arxiv.org/abs/2510.02300 | |
| 542,A decoder-only foundation model for time-series forecasting,6,44,46.17,46,"Oct 14, 2023",6,7060,https://arxiv.org/abs/2310.10688 | |
| 543,MoDA: Multi-modal Diffusion Architecture for Talking Head Generation,9,46,47.89,48,"Jul 4, 2025",2,140,https://arxiv.org/abs/2507.03256 | |
| 544,"Gradient-Attention Guided Dual-Masking Synergetic Framework for Robust | |
| Text-based Person Retrieval",3,30,41.67,45,"Sep 11, 2025",6,10,https://arxiv.org/abs/2509.09118 | |
| 545,Streaming 4D Visual Geometry Transformer,1,25,25,25,"Jul 15, 2025",10,454,https://arxiv.org/abs/2507.11539 | |
| 546,s3: You Don't Need That Much Data to Train a Search Agent via RL,5,41,45.8,47,"May 20, 2025",18,564,https://arxiv.org/abs/2505.14146 | |
| 547,"MesaTask: Towards Task-Driven Tabletop Scene Generation via 3D Spatial | |
| Reasoning",2,36,38,38,"Sep 26, 2025",27,24,https://arxiv.org/abs/2509.22281 | |
| 548,"Ming-UniVision: Joint Image Understanding and Generation with a Unified | |
| Continuous Tokenizer",3,38,42.33,42,"Oct 8, 2025",63,72,https://arxiv.org/abs/2510.06590 | |
| 549,"iMontage: Unified, Versatile, Highly Dynamic Many-to-many Image Generation",2,34,38,38,"Nov 25, 2025",30,120,https://arxiv.org/abs/2511.20635 | |
| 550,"ThinkSound: Chain-of-Thought Reasoning in Multimodal Large Language | |
| Models for Audio Generation and Editing",2,38,38.5,38,"Jun 26, 2025",7,876,https://arxiv.org/abs/2506.21448 | |
| 551,"Q-Sched: Pushing the Boundaries of Few-Step Diffusion Models with | |
| Quantization-Aware Scheduling",2,30,38.5,38,"Sep 1, 2025",5,6,https://arxiv.org/abs/2509.01624 | |
| 552,"GAS: Improving Discretization of Diffusion ODEs via Generalized | |
| Adversarial Solver",2,30,38.5,38,"Oct 20, 2025",2,10,https://arxiv.org/abs/2510.17699 | |
| 553,MUR: Momentum Uncertainty guided Reasoning for Large Language Models,3,40,43,44,"Jul 20, 2025",36,32,https://arxiv.org/abs/2507.14958 | |
| 554,"Scaling Code-Assisted Chain-of-Thoughts and Instructions for Model | |
| Reasoning",3,39,43,42,"Oct 5, 2025",18,6,https://arxiv.org/abs/2510.04081 | |
| 555,"AniMaker: Automated Multi-Agent Animated Storytelling with MCTS-Driven | |
| Clip Generation",8,46,48.12,48,"Jun 12, 2025",37,173,https://arxiv.org/abs/2506.10540 | |
| 556,"D2E: Scaling Vision-Action Pretraining on Desktop Data for Transfer to | |
| Embodied AI",2,39,39.5,39,"Oct 7, 2025",101,22,https://arxiv.org/abs/2510.05684 | |
| 557,V-ReasonBench: Toward Unified Reasoning Benchmark Suite for Video Generation Models,3,38,43.33,43,"Nov 20, 2025",48,15,https://arxiv.org/abs/2511.16668 | |
| 558,"DetailFlow: 1D Coarse-to-Fine Autoregressive Image Generation via | |
| Next-Detail Prediction",1,29,29,29,"May 27, 2025",16,144,https://arxiv.org/abs/2505.21473 | |
| 559,MM-BrowseComp: A Comprehensive Benchmark for Multimodal Browsing Agents,4,40,45.5,46,"Aug 14, 2025",13,10,https://arxiv.org/abs/2508.13186 | |
| 560,"Mem4Nav: Boosting Vision-and-Language Navigation in Urban Environments | |
| with a Hierarchical Spatial-Cognition Long-Short Memory System",4,41,45.5,46,"Jun 24, 2025",3,89,https://arxiv.org/abs/2506.19433 | |
| 561,"Qwen3 Embedding: Advancing Text Embedding and Reranking Through | |
| Foundation Models",5,44,46.8,46,"Jun 5, 2025",72,1390,https://arxiv.org/abs/2506.05176 | |
| 562,"Part II: ROLL Flash -- Accelerating RLVR and Agentic Training with | |
| Asynchrony",4,44,45.75,45,"Oct 13, 2025",15,2270,https://arxiv.org/abs/2510.11345 | |
| 563,Agent-R1: Training Powerful LLM Agents with End-to-End Reinforcement Learning,4,44,45.75,46,"Nov 18, 2025",15,951,https://arxiv.org/abs/2511.14460 | |
| 564,Deep Researcher with Test-Time Diffusion,3,36,44.33,47,"Jul 21, 2025",63,2910,https://arxiv.org/abs/2507.16075 | |
| 565,RoboOmni: Proactive Robot Manipulation in Omni-modal Context,2,40,41,41,"Oct 27, 2025",52,27,https://arxiv.org/abs/2510.23763 | |
| 566,LiteAttention: A Temporal Sparse Attention for Diffusion Transformers,3,42,44.33,43,"Nov 14, 2025",24,23,https://arxiv.org/abs/2511.11062 | |
| 567,"SEAgent: Self-Evolving Computer Use Agent with Autonomous Learning from | |
| Experience",4,45,46.25,46,"Aug 6, 2025",46,141,https://arxiv.org/abs/2508.04700 | |
| 568,Quantile Advantage Estimation for Entropy-Safe Reasoning,2,35,41.5,41,"Sep 26, 2025",100,8,https://arxiv.org/abs/2509.22611 | |
| 569,"TensorBLEU: Vectorized GPU-based BLEU Score Implementation for | |
| Per-Sentence In-Training Evaluation",4,40,46.25,48,"Oct 7, 2025",7,13,https://arxiv.org/abs/2510.05485 | |
| 570,MATRIX: Mask Track Alignment for Interaction-aware Video Generation,2,37,41.5,41,"Oct 8, 2025",29,22,https://arxiv.org/abs/2510.07310 | |
| 571,"HSCodeComp: A Realistic and Expert-level Benchmark for Deep Search | |
| Agents in Hierarchical Rule Application",4,43,46.25,46,"Oct 22, 2025",26,85,https://arxiv.org/abs/2510.19631 | |
| 572,VACE: All-in-One Video Creation and Editing,3,43,45,44,"Mar 10, 2025",54,3000,https://arxiv.org/abs/2503.07598 | |
| 573,Open Deep Search: Democratizing Search with Open-source Reasoning Agents,4,45,46.5,46,"Mar 26, 2025",48,3620,https://arxiv.org/abs/2503.20201 | |
| 574,TUN3D: Towards Real-World Scene Understanding from Unposed Images,1,33,33,33,"Sep 23, 2025",12,11,https://arxiv.org/abs/2509.21388 | |
| 575,Efficient Guided Generation for Large Language Models,3,41,45,45,"Jul 19, 2023",8,12900,https://arxiv.org/abs/2307.09702 | |
| 576,UltraFlux: Data-Model Co-Design for High-quality Native 4K Text-to-Image Generation across Diverse Aspect Ratios,2,36,42,42,"Nov 22, 2025",34,42,https://arxiv.org/abs/2511.18050 | |
| 577,STARFlow-V: End-to-End Video Generative Modeling with Normalizing Flow,6,45,48,48,"Nov 25, 2025",16,40,https://arxiv.org/abs/2511.20462 | |
| 578,Zebra-CoT: A Dataset for Interleaved Vision Language Reasoning,1,34,34,34,"Jul 22, 2025",28,42,https://arxiv.org/abs/2507.16746 | |
| 579,"From reactive to cognitive: brain-inspired spatial intelligence for | |
| embodied agents",2,42,42.5,42,"Aug 24, 2025",3,18,https://arxiv.org/abs/2508.17198 | |
| 580,"Search-R1: Training LLMs to Reason and Leverage Search Engines with | |
| Reinforcement Learning",4,46,46.75,47,"Mar 12, 2025",35,3130,https://arxiv.org/abs/2503.09516 | |
| 581,Retrieval-Augmented Generation with Hierarchical Knowledge,1,34,34,34,"Mar 13, 2025",2,359,https://arxiv.org/abs/2503.10150 | |
| 582,"REASONING GYM: Reasoning Environments for Reinforcement Learning with | |
| Verifiable Rewards",1,34,34,34,"May 30, 2025",72,1130,https://arxiv.org/abs/2505.24760 | |
| 583,Chem-R: Learning to Reason as a Chemist,2,35,42.5,42,"Oct 19, 2025",45,9,https://arxiv.org/abs/2510.16880 | |
| 584,Hierarchical Budget Policy Optimization for Adaptive Reasoning,1,35,35,35,"Jul 21, 2025",12,14,https://arxiv.org/abs/2507.15844 | |
| 585,Qwen2.5-Omni Technical Report,4,45,47,47,"Mar 26, 2025",165,3510,https://arxiv.org/abs/2503.20215 | |
| 586,Variational Reasoning for Language Models,2,36,43,43,"Sep 26, 2025",57,34,https://arxiv.org/abs/2509.22637 | |
| 587,OpenVoice: Versatile Instant Voice Cloning,6,46,48.33,49,"Dec 3, 2023",3,35300,https://arxiv.org/abs/2312.01479 | |
| 588,"BrowseComp-Plus: A More Fair and Transparent Evaluation Benchmark of | |
| Deep-Research Agent",3,43,46,45,"Aug 8, 2025",35,49,https://arxiv.org/abs/2508.06600 | |
| 589,"CodePlot-CoT: Mathematical Visual Reasoning by Thinking with Code-Driven | |
| Images",3,38,46,50,"Oct 13, 2025",12,22,https://arxiv.org/abs/2510.11718 | |
| 590,Continuous Thought Machines,2,42,43.5,43,"May 8, 2025",12,1490,https://arxiv.org/abs/2505.05522 | |
| 591,MMaDA: Multimodal Large Diffusion Language Models,1,37,37,37,"May 21, 2025",94,1230,https://arxiv.org/abs/2505.15809 | |
| 592,"Cognitive Kernel-Pro: A Framework for Deep Research Agents and Agent | |
| Foundation Models Training",4,46,47.5,47,"Aug 1, 2025",91,388,https://arxiv.org/abs/2508.00414 | |
| 593,"DrugReasoner: Interpretable Drug Approval Prediction with a | |
| Reasoning-augmented Language Model",2,44,44,44,"Aug 26, 2025",10,2,https://arxiv.org/abs/2508.18579 | |
| 594,TikZero: Zero-Shot Text-Guided Graphics Program Synthesis,2,44,44,44,"Mar 14, 2025",3,1530,https://arxiv.org/abs/2503.11509 | |
| 595,VideoNSA: Native Sparse Attention Scales Video Understanding,2,42,44,44,"Oct 2, 2025",8,29,https://arxiv.org/abs/2510.02295 | |
| 596,OmniNWM: Omniscient Driving Navigation World Models,2,43,44,44,"Oct 21, 2025",6,48,https://arxiv.org/abs/2510.18313 | |
| 597,EVTAR: End-to-End Try on with Additional Unpaired Visual Reference,2,43,44,44,"Nov 2, 2025",4,17,https://arxiv.org/abs/2511.00956 | |
| 598,P1: Mastering Physics Olympiads with Reinforcement Learning,2,43,44,44,"Nov 17, 2025",106,45,https://arxiv.org/abs/2511.13612 | |
| 599,nablaNABLA: Neighborhood Adaptive Block-Level Attention,1,38,38,38,"Jul 17, 2025",85,10,https://arxiv.org/abs/2507.13546 | |
| 600,"Evolving Language Models without Labels: Majority Drives Selection, | |
| Novelty Promotes Variation",3,41,46.67,49,"Sep 18, 2025",29,19,https://arxiv.org/abs/2509.15194 | |
| 601,YuE: Scaling Open Foundation Models for Long-Form Music Generation,2,40,44.5,44,"Mar 11, 2025",69,5520,https://arxiv.org/abs/2503.08638 | |
| 602,LightsOut: Diffusion-based Outpainting for Enhanced Lens Flare Removal,3,41,46.67,49,"Oct 17, 2025",20,10,https://arxiv.org/abs/2510.15868 | |
| 603,"ToolScope: An Agentic Framework for Vision-Guided and Long-Horizon Tool | |
| Use",5,47,48.4,49,"Oct 31, 2025",20,15,https://arxiv.org/abs/2510.27363 | |
| 604,Drag-and-Drop LLMs: Zero-Shot Prompt-to-Weights,1,39,39,39,"Jun 19, 2025",122,58,https://arxiv.org/abs/2506.16406 | |
| 605,"DCReg: Decoupled Characterization for Efficient Degenerate LiDAR | |
| Registration",3,41,47,50,"Sep 8, 2025",1,58,https://arxiv.org/abs/2509.06285 | |
| 606,"VoiceAssistant-Eval: Benchmarking AI Assistants across Listening, | |
| Speaking, and Viewing",1,39,39,39,"Sep 26, 2025",19,4,https://arxiv.org/abs/2509.22651 | |
| 607,"Beyond the Exploration-Exploitation Trade-off: A Hidden State Approach | |
| for LLM Reasoning in RLVR",3,46,47,47,"Sep 28, 2025",38,7,https://arxiv.org/abs/2509.23808 | |
| 608,"RAPO++: Cross-Stage Prompt Optimization for Text-to-Video Generation via | |
| Data Alignment and Test-Time Scaling",2,40,45,45,"Oct 23, 2025",11,104,https://arxiv.org/abs/2510.20206 | |
| 609,Scaling Language-Centric Omnimodal Representation Learning,1,40,40,40,"Oct 13, 2025",52,11,https://arxiv.org/abs/2510.11693 | |
| 610,"ImagerySearch: Adaptive Test-Time Search for Video Generation Beyond | |
| Semantic Dependency Constraints",2,43,45.5,45,"Oct 16, 2025",47,47,https://arxiv.org/abs/2510.14847 | |
| 611,"Automatic Synthetic Data and Fine-grained Adaptive Feature Alignment for | |
| Composed Person Retrieval",2,46,46,46,"Nov 25, 2023",1,49,https://arxiv.org/abs/2311.16515 | |
| 612,Character Mixing for Video Generation,2,42,46,46,"Oct 6, 2025",5,52,https://arxiv.org/abs/2510.05093 | |
| 613,"TIR-Bench: A Comprehensive Benchmark for Agentic Thinking-with-Images | |
| Reasoning",2,46,46,46,"Nov 3, 2025",12,14,https://arxiv.org/abs/2511.01833 | |
| 614,"Trinity-RFT: A General-Purpose and Unified Framework for Reinforcement | |
| Fine-Tuning of Large Language Models",3,48,48,48,"May 23, 2025",9,269,https://arxiv.org/abs/2505.17826 | |
| 615,Trainable Dynamic Mask Sparse Attention,3,46,48,49,"Aug 4, 2025",17,366,https://arxiv.org/abs/2508.02124 | |
| 616,"FlashAdventure: A Benchmark for GUI Agents Solving Full Story Arcs in | |
| Diverse Adventure Games",2,44,46.5,46,"Sep 1, 2025",8,7,https://arxiv.org/abs/2509.01052 | |
| 617,HoloScene: Simulation-Ready Interactive 3D Worlds from a Single Video,1,42,42,42,"Oct 7, 2025",5,5,https://arxiv.org/abs/2510.05560 | |
| 618,"SIU3R: Simultaneous Scene Understanding and 3D Reconstruction Beyond | |
| Feature Alignment",2,46,46.5,46,"Jul 3, 2025",2,80,https://arxiv.org/abs/2507.02705 | |
| 619,"Learning on the Job: Test-Time Curricula for Targeted Reinforcement | |
| Learning",1,43,43,43,"Oct 6, 2025",1,3,https://arxiv.org/abs/2510.04786 | |
| 620,"Discrete Diffusion Models with MLLMs for Unified Medical Multimodal | |
| Generation",1,43,43,43,"Oct 7, 2025",4,4,https://arxiv.org/abs/2510.06131 | |
| 621,Reasoning in Space via Grounding in the World,3,45,48.33,50,"Oct 15, 2025",13,16,https://arxiv.org/abs/2510.13800 | |
| 622,First Frame Is the Place to Go for Video Content Customization,2,45,47,47,"Nov 19, 2025",45,36,https://arxiv.org/abs/2511.15700 | |
| 623,Thinking-while-Generating: Interleaving Textual Reasoning throughout Visual Generation,2,44,47,47,"Nov 20, 2025",12,43,https://arxiv.org/abs/2511.16671 | |
| 624,Computer-Use Agents as Judges for Generative User Interface,2,45,47,47,"Nov 19, 2025",48,24,https://arxiv.org/abs/2511.15567 | |
| 625,"Graph2Eval: Automatic Multimodal Task Generation for Agents via | |
| Knowledge Graphs",1,44,44,44,"Oct 1, 2025",1,5,https://arxiv.org/abs/2510.00507 | |
| 626,"MathCanvas: Intrinsic Visual Chain-of-Thought for Multimodal | |
| Mathematical Reasoning",2,47,47.5,47,"Oct 16, 2025",11,5,https://arxiv.org/abs/2510.14958 | |
| 627,Grounding Computer Use Agents on Human Demonstrations,4,47,49.25,50,"Nov 10, 2025",18,15,https://arxiv.org/abs/2511.07332 | |
| 628,"Feedback-Driven Tool-Use Improvements in Large Language Models via | |
| Automated Build Environments",2,46,48,48,"Aug 12, 2025",12,17,https://arxiv.org/abs/2508.08791 | |
| 629,Agentic Reinforced Policy Optimization,4,48,49.5,50,"Jul 26, 2025",154,691,https://arxiv.org/abs/2507.19849 | |
| 630,"SAM2Act: Integrating Visual Foundation Model with A Memory Architecture | |
| for Robotic Manipulation",3,48,49,49,"Jan 30, 2025",2,131,https://arxiv.org/abs/2501.18564 | |
| 631,Generative AI for Autonomous Driving: Frontiers and Opportunities,2,48,48,48,"May 13, 2025",1,176,https://arxiv.org/abs/2505.08854 | |
| 632,CHARM: Control-point-based 3D Anime Hairstyle Auto-Regressive Modeling,2,46,48,48,"Sep 25, 2025",11,9,https://arxiv.org/abs/2509.21114 | |
| 633,DeepPrune: Parallel Scaling without Inter-trace Redundancy,1,45,45,45,"Oct 9, 2025",21,9,https://arxiv.org/abs/2510.08483 | |
| 634,UI2Code^N: A Visual Language Model for Test-Time Scalable Interactive UI-to-Code Generation,2,47,48,48,"Nov 11, 2025",28,18,https://arxiv.org/abs/2511.08195 | |
| 635,WorldVLA: Towards Autoregressive Action World Model,1,45,45,45,"Jun 26, 2025",40,669,https://arxiv.org/abs/2506.21539 | |
| 636,"Gaze into the Heart: A Multi-View Video Dataset for rPPG and Health | |
| Biomarkers Estimation",1,46,46,46,"Aug 25, 2025",11,3,https://arxiv.org/abs/2508.17924 | |
| 637,Uncertainty-Aware Testing-Time Optimization for 3D Human Pose Estimation,1,46,46,46,"Feb 4, 2024",1,5,https://arxiv.org/abs/2402.02339 | |
| 638,"AssetOpsBench: Benchmarking AI Agents for Task Automation in Industrial | |
| Asset Operations and Maintenance",1,46,46,46,"Jun 4, 2025",14,252,https://arxiv.org/abs/2506.03828 | |
| 639,"Entropy Regularizing Activation: Boosting Continuous Control, Large | |
| Language Models, and Image Classification with Activation as Entropy | |
| Constraints",1,46,46,46,"Oct 9, 2025",5,7,https://arxiv.org/abs/2510.08549 | |
| 640,"UniPixel: Unified Object Referring and Segmentation for Pixel-Level | |
| Visual Reasoning",1,47,47,47,"Sep 22, 2025",3,40,https://arxiv.org/abs/2509.18094 | |
| 641,"GeoSVR: Taming Sparse Voxels for Geometrically Accurate Surface | |
| Reconstruction",1,47,47,47,"Sep 22, 2025",2,65,https://arxiv.org/abs/2509.18090 | |
| 642,PIPer: On-Device Environment Setup via Online Reinforcement Learning,1,47,47,47,"Sep 29, 2025",23,4,https://arxiv.org/abs/2509.25455 | |
| 643,"Factuality Matters: When Image Generation and Editing Meet Structured | |
| Visuals",1,47,47,47,"Oct 6, 2025",12,4,https://arxiv.org/abs/2510.05091 | |
| 644,"BIRD-INTERACT: Re-imagining Text-to-SQL Evaluation for Large Language | |
| Models via Lens of Dynamic Interactions",1,47,47,47,"Oct 6, 2025",4,260,https://arxiv.org/abs/2510.05318 | |
| 645,Instant4D: 4D Gaussian Splatting in Minutes,1,47,47,47,"Oct 1, 2025",5,58,https://arxiv.org/abs/2510.01119 | |
| 646,"ASTRA: Autonomous Spatial-Temporal Red-teaming for AI Software | |
| Assistants",2,49,49.5,49,"Aug 5, 2025",8,37,https://arxiv.org/abs/2508.03936 | |
| 647,"InfiGUI-G1: Advancing GUI Grounding with Adaptive Exploration Policy | |
| Optimization",2,49,49.5,49,"Aug 7, 2025",25,56,https://arxiv.org/abs/2508.05731 | |
| 648,"NER Retriever: Zero-Shot Named Entity Retrieval with Type-Aware | |
| Embeddings",3,50,50,50,"Sep 4, 2025",19,19,https://arxiv.org/abs/2509.04011 | |
| 649,Where LLM Agents Fail and How They can Learn From Failures,1,48,48,48,"Sep 29, 2025",11,8,https://arxiv.org/abs/2509.25370 | |
| 650,"MarS: a Financial Market Simulation Engine Powered by Generative | |
| Foundation Model",1,48,48,48,"Sep 4, 2024",1,1560,https://arxiv.org/abs/2409.07486 | |
| 651,"Audio-visual Controlled Video Diffusion with Masked Selective State | |
| Spaces Modeling for Natural Talking Head Generation",1,49,49,49,"Apr 3, 2025",49,344,https://arxiv.org/abs/2504.02542 | |
| 652,"SynParaSpeech: Automated Synthesis of Paralinguistic Datasets for Speech | |
| Generation and Understanding",1,49,49,49,"Sep 18, 2025",1,37,https://arxiv.org/abs/2509.14946 | |
| 653,"Beyond Log Likelihood: Probability-Based Objectives for Supervised | |
| Fine-Tuning across the Model Capability Continuum",1,49,49,49,"Oct 1, 2025",7,3,https://arxiv.org/abs/2510.00526 | |
| 654,"SLA: Beyond Sparsity in Diffusion Transformers via Fine-Tunable | |
| Sparse-Linear Attention",2,50,50,50,"Sep 28, 2025",107,38,https://arxiv.org/abs/2509.24006 | |
| 655,Scaling Large-Language-Model-based Multi-Agent Collaboration,2,50,50,50,"Jun 11, 2024",3,27600,https://arxiv.org/abs/2406.07155 | |
| 656,A Survey of Vibe Coding with Large Language Models,1,49,49,49,"Oct 14, 2025",21,7,https://arxiv.org/abs/2510.12399 | |
| 657,Gaussian Splatting with Discretized SDF for Relightable Assets,1,50,50,50,"Jul 21, 2025",19,61,https://arxiv.org/abs/2507.15629 | |
| 658,Flow-GRPO: Training Flow Matching Models via Online RL,1,50,50,50,"May 8, 2025",83,1120,https://arxiv.org/abs/2505.05470 | |
| 659,One-Minute Video Generation with Test-Time Training,1,50,50,50,"Apr 7, 2025",109,2090,https://arxiv.org/abs/2504.05298 | |
| 660,LongCat-Flash-Thinking Technical Report,1,50,50,50,"Sep 23, 2025",2,195,https://arxiv.org/abs/2509.18883 | |
| 661,"Efficient Audio-Visual Speech Separation with Discrete Lip Semantics and | |
| Multi-Scale Global-Local Attention",1,50,50,50,"Sep 28, 2025",13,22,https://arxiv.org/abs/2509.23610 | |
| 662,EVODiff: Entropy-aware Variance Optimized Diffusion Inference,1,50,50,50,"Sep 30, 2025",1,9,https://arxiv.org/abs/2509.26096 | |
| 663,"GLiNER2: An Efficient Multi-Task Information Extraction System with | |
| Schema-Driven Interface",1,50,50,50,"Jul 24, 2025",28,186,https://arxiv.org/abs/2507.18546 | |