diff --git "a/huggingface-top-papers.csv" "b/huggingface-top-papers.csv" --- "a/huggingface-top-papers.csv" +++ "b/huggingface-top-papers.csv" @@ -1,971 +1,971 @@ rank,name,times_trended,best_rank,avg_rank,median_rank,publish_date,max_upvotes,max_github_stars,arxiv_link 1,LlamaFactory: Unified Efficient Fine-Tuning of 100+ Language Models,432,2,11.28,11,"Mar 20, 2024",173,63300,https://arxiv.org/abs/2403.13372 -2,"PaddleOCR-VL: Boosting Multilingual Document Parsing via a 0.9B - Ultra-Compact Vision-Language Model",124,1,3.35,3,"Oct 16, 2025",98,65400,https://arxiv.org/abs/2510.14528 -3,"MinerU2.5: A Decoupled Vision-Language Model for Efficient - High-Resolution Document Parsing",179,2,10.82,9,"Sep 26, 2025",134,49600,https://arxiv.org/abs/2509.22186 -4,"AgentScope 1.0: A Developer-Centric Framework for Building Agentic +2,DINOv3,346,1,21.02,15,"Aug 13, 2025",284,8590,https://arxiv.org/abs/2508.10104 +3,Agent Lightning: Train ANY AI Agents with Reinforcement Learning,325,1,20.18,20,"Aug 5, 2025",120,9090,https://arxiv.org/abs/2508.03680 +4,WebDancer: Towards Autonomous Information Seeking Agency,398,3,25.85,26,"May 28, 2025",33,17300,https://arxiv.org/abs/2505.22648 +5,"WebShaper: Agentically Data Synthesizing via Information-Seeking + Formalization",402,2,26.34,26,"Jul 20, 2025",60,17300,https://arxiv.org/abs/2507.15061 +6,WebWatcher: Breaking New Frontier of Vision-Language Deep Research Agent,393,1,25.86,26,"Aug 7, 2025",139,17300,https://arxiv.org/abs/2508.05748 +7,WebSailor: Navigating Super-human Reasoning for Web Agent,394,2,26.03,26,"Jul 3, 2025",122,17300,https://arxiv.org/abs/2507.02592 +8,"InternVL3: Exploring Advanced Training and Test-Time Recipes for + Open-Source Multimodal Models",349,8,23.07,21,"Apr 14, 2025",304,8850,https://arxiv.org/abs/2504.10479 +9,"AgentScope 1.0: A Developer-Centric Framework for Building Agentic Applications",290,1,18.33,14,"Aug 22, 2025",53,14100,https://arxiv.org/abs/2508.16279 -5,Agent Lightning: Train ANY AI Agents with Reinforcement Learning,325,1,20.18,20,"Aug 5, 2025",120,9090,https://arxiv.org/abs/2508.03680 -6,AgentFly: Fine-tuning LLM Agents without Fine-tuning LLMs,144,1,12.94,8,"Aug 22, 2025",153,1740,https://arxiv.org/abs/2508.16153 -7,OmniFlatten: An End-to-end GPT Model for Seamless Voice Conversation,112,3,9.8,9,"Oct 23, 2024",5,50300,https://arxiv.org/abs/2410.17799 -8,DINOv3,346,1,21.02,15,"Aug 13, 2025",284,8590,https://arxiv.org/abs/2508.10104 -9,Qwen-Image Technical Report,302,1,20.35,19,"Aug 4, 2025",261,6150,https://arxiv.org/abs/2508.02324 -10,Scaling Agents via Continual Pre-training,196,1,18.28,17,"Sep 16, 2025",115,17300,https://arxiv.org/abs/2509.13310 -11,PaddleOCR 3.0 Technical Report,100,1,12.08,11,"Jul 8, 2025",17,57600,https://arxiv.org/abs/2507.05595 -12,"ReSum: Unlocking Long-Horizon Search Intelligence via Context +10,Qwen-Image Technical Report,302,1,20.35,19,"Aug 4, 2025",261,6150,https://arxiv.org/abs/2508.02324 +11,"MinerU2.5: A Decoupled Vision-Language Model for Efficient + High-Resolution Document Parsing",179,2,10.82,9,"Sep 26, 2025",134,49600,https://arxiv.org/abs/2509.22186 +12,Scaling Agents via Continual Pre-training,196,1,18.28,17,"Sep 16, 2025",115,17300,https://arxiv.org/abs/2509.13310 +13,"Easy Dataset: A Unified and Extensible Framework for Synthesizing LLM + Fine-Tuning Data from Unstructured Documents",345,13,32.46,32,"Jul 5, 2025",51,12100,https://arxiv.org/abs/2507.04009 +14,"WebSailor-V2: Bridging the Chasm to Proprietary Agents via Synthetic + Data and Scalable Reinforcement Learning",201,1,19.4,17,"Sep 16, 2025",90,17300,https://arxiv.org/abs/2509.13305 +15,"ReSum: Unlocking Long-Horizon Search Intelligence via Context Summarization",198,1,18.98,16,"Sep 16, 2025",78,17300,https://arxiv.org/abs/2509.13313 -13,TradingAgents: Multi-Agents LLM Financial Trading Framework,111,6,13.67,13,"Dec 28, 2024",14,25800,https://arxiv.org/abs/2412.20138 -14,FastVLM: Efficient Vision Encoding for Vision Language Models,162,1,17.6,11,"Dec 17, 2024",70,6680,https://arxiv.org/abs/2412.13303 -15,"WebWeaver: Structuring Web-Scale Evidence with Dynamic Outlines for +16,"PC-Agent: A Hierarchical Multi-Agent Collaboration Framework for Complex + Task Automation on PC",222,3,22.85,18,"Feb 20, 2025",29,6440,https://arxiv.org/abs/2502.14282 +17,"WebWeaver: Structuring Web-Scale Evidence with Dynamic Outlines for Open-Ended Deep Research",194,4,19,16,"Sep 16, 2025",105,17300,https://arxiv.org/abs/2509.13312 -16,DeepAnalyze: Agentic Large Language Models for Autonomous Data Science,115,2,14.33,14,"Oct 19, 2025",102,2630,https://arxiv.org/abs/2510.16872 -17,"WebSailor-V2: Bridging the Chasm to Proprietary Agents via Synthetic - Data and Scalable Reinforcement Learning",201,1,19.4,17,"Sep 16, 2025",90,17300,https://arxiv.org/abs/2509.13305 -18,MinerU: An Open-Source Solution for Precise Document Content Extraction,112,3,14.65,16,"Sep 27, 2024",32,49600,https://arxiv.org/abs/2409.18839 -19,"IndexTTS: An Industrial-Level Controllable and Efficient Zero-Shot +18,"Look Before You Leap: A GUI-Critic-R1 Model for Pre-Operative Error + Diagnosis in GUI Automation",220,4,22.92,19,"Jun 5, 2025",19,6430,https://arxiv.org/abs/2506.04614 +19,"PaddleOCR-VL: Boosting Multilingual Document Parsing via a 0.9B + Ultra-Compact Vision-Language Model",124,1,3.35,3,"Oct 16, 2025",98,65400,https://arxiv.org/abs/2510.14528 +20,Mobile-Agent-v3: Foundamental Agents for GUI Automation,198,3,21.24,17,"Aug 21, 2025",64,6440,https://arxiv.org/abs/2508.15144 +21,AgentFly: Fine-tuning LLM Agents without Fine-tuning LLMs,144,1,12.94,8,"Aug 22, 2025",153,1740,https://arxiv.org/abs/2508.16153 +22,FastVLM: Efficient Vision Encoding for Vision Language Models,162,1,17.6,11,"Dec 17, 2024",70,6680,https://arxiv.org/abs/2412.13303 +23,MIRIX: Multi-Agent Memory System for LLM-Based Agents,272,5,31.7,35,"Jul 10, 2025",79,3380,https://arxiv.org/abs/2507.07957 +24,"GLM-4.5: Agentic, Reasoning, and Coding (ARC) Foundation Models",243,11,30.1,30,"Aug 8, 2025",185,3000,https://arxiv.org/abs/2508.06471 +25,Prompt Orchestration Markup Language,153,3,18.99,12,"Aug 19, 2025",49,4560,https://arxiv.org/abs/2508.13948 +26,OmniFlatten: An End-to-end GPT Model for Seamless Voice Conversation,112,3,9.8,9,"Oct 23, 2024",5,50300,https://arxiv.org/abs/2410.17799 +27,UI-TARS: Pioneering Automated GUI Interaction with Native Agents,208,4,29.12,28,"Jan 21, 2025",65,7830,https://arxiv.org/abs/2501.12326 +28,"A Comprehensive Survey of Self-Evolving AI Agents: A New Paradigm + Bridging Foundation Models and Lifelong Agentic Systems",191,5,27.24,29,"Aug 10, 2025",97,1140,https://arxiv.org/abs/2508.07407 +29,Qwen3 Technical Report,266,2,34.64,34,"May 14, 2025",316,25500,https://arxiv.org/abs/2505.09388 +30,The Landscape of Agentic Reinforcement Learning for LLMs: A Survey,125,2,17.14,15,"Sep 2, 2025",214,941,https://arxiv.org/abs/2509.02547 +31,DeepAnalyze: Agentic Large Language Models for Autonomous Data Science,115,2,14.33,14,"Oct 19, 2025",102,2630,https://arxiv.org/abs/2510.16872 +32,TradingAgents: Multi-Agents LLM Financial Trading Framework,111,6,13.67,13,"Dec 28, 2024",14,25800,https://arxiv.org/abs/2412.20138 +33,MinerU: An Open-Source Solution for Precise Document Content Extraction,112,3,14.65,16,"Sep 27, 2024",32,49600,https://arxiv.org/abs/2409.18839 +34,"IndexTTS: An Industrial-Level Controllable and Efficient Zero-Shot Text-To-Speech System",112,7,14.97,14,"Feb 8, 2025",5,16000,https://arxiv.org/abs/2502.05512 -20,"USO: Unified Style and Subject-Driven Generation via Disentangled and +35,StableAvatar: Infinite-Length Audio-Driven Avatar Video Generation,137,1,21.63,20,"Aug 11, 2025",27,1020,https://arxiv.org/abs/2508.08248 +36,"Paper2Agent: Reimagining Research Papers As Interactive and Reliable AI + Agents",135,5,22.01,20,"Sep 8, 2025",41,1660,https://arxiv.org/abs/2509.06917 +37,PaddleOCR 3.0 Technical Report,100,1,12.08,11,"Jul 8, 2025",17,57600,https://arxiv.org/abs/2507.05595 +38,"Seeing, Listening, Remembering, and Reasoning: A Multimodal Agent with + Long-Term Memory",121,4,18.92,14,"Aug 13, 2025",54,893,https://arxiv.org/abs/2508.09736 +39,rStar2-Agent: Agentic Reasoning Technical Report,112,4,16.78,16,"Aug 28, 2025",106,1230,https://arxiv.org/abs/2508.20722 +40,Mem0: Building Production-Ready AI Agents with Scalable Long-Term Memory,112,6,17.88,15,"Apr 28, 2025",33,43700,https://arxiv.org/abs/2504.19413 +41,"ScreenCoder: Advancing Visual-to-Code Generation for Front-End + Automation via Modular Multimodal Agents",122,1,21.56,24,"Jul 30, 2025",98,2290,https://arxiv.org/abs/2507.22827 +42,Step-Audio 2 Technical Report,94,2,13.01,12,"Jul 22, 2025",71,1050,https://arxiv.org/abs/2507.16632 +43,"USO: Unified Style and Subject-Driven Generation via Disentangled and Reward Learning",92,2,12.67,6,"Aug 26, 2025",56,1080,https://arxiv.org/abs/2508.18966 -21,Step-Audio 2 Technical Report,94,2,13.01,12,"Jul 22, 2025",71,1050,https://arxiv.org/abs/2507.16632 -22,The Landscape of Agentic Reinforcement Learning for LLMs: A Survey,125,2,17.14,15,"Sep 2, 2025",214,941,https://arxiv.org/abs/2509.02547 -23,TempFlow-GRPO: When Timing Matters for GRPO in Flow Models,96,4,14.52,10,"Aug 6, 2025",12,782,https://arxiv.org/abs/2508.04324 -24,Prompt Orchestration Markup Language,153,3,18.99,12,"Aug 19, 2025",49,4560,https://arxiv.org/abs/2508.13948 -25,rStar2-Agent: Agentic Reasoning Technical Report,112,4,16.78,16,"Aug 28, 2025",106,1230,https://arxiv.org/abs/2508.20722 -26,"InternVL3: Exploring Advanced Training and Test-Time Recipes for - Open-Source Multimodal Models",349,8,23.07,21,"Apr 14, 2025",304,8850,https://arxiv.org/abs/2504.10479 -27,Mobile-Agent-v3: Foundamental Agents for GUI Automation,198,3,21.24,17,"Aug 21, 2025",64,6440,https://arxiv.org/abs/2508.15144 -28,VibeVoice Technical Report,37,1,1,1,"Aug 26, 2025",118,7980,https://arxiv.org/abs/2508.19205 -29,Mem0: Building Production-Ready AI Agents with Scalable Long-Term Memory,112,6,17.88,15,"Apr 28, 2025",33,43700,https://arxiv.org/abs/2504.19413 -30,"FAPO: Flawed-Aware Policy Optimization for Efficient and Reliable +44,TempFlow-GRPO: When Timing Matters for GRPO in Flow Models,96,4,14.52,10,"Aug 6, 2025",12,782,https://arxiv.org/abs/2508.04324 +45,RAG-Anything: All-in-One RAG Framework,129,2,25.02,28,"Oct 14, 2025",49,10600,https://arxiv.org/abs/2510.12323 +46,"ComfyUI-Copilot: An Intelligent Assistant for Automated Workflow + Development",172,6,32.85,36,"Jun 5, 2025",79,3660,https://arxiv.org/abs/2506.05010 +47,VGGT: Visual Geometry Grounded Transformer,232,12,37.63,39,"Mar 14, 2025",33,11700,https://arxiv.org/abs/2503.11651 +48,MixGRPO: Unlocking Flow-based GRPO Efficiency with Mixed ODE-SDE,170,12,32.78,34,"Jul 29, 2025",15,1010,https://arxiv.org/abs/2507.21802 +49,4DNeX: Feed-Forward 4D Generative Modeling Made Easy,178,2,33.96,37,"Aug 18, 2025",61,751,https://arxiv.org/abs/2508.13154 +50,"FAPO: Flawed-Aware Policy Optimization for Efficient and Reliable Reasoning",86,3,15.94,11,"Oct 26, 2025",10,17000,https://arxiv.org/abs/2510.22543 -31,"Seeing, Listening, Remembering, and Reasoning: A Multimodal Agent with - Long-Term Memory",121,4,18.92,14,"Aug 13, 2025",54,893,https://arxiv.org/abs/2508.09736 -32,Depth Anything 3: Recovering the Visual Space from Any Views,40,1,6.58,3,"Nov 13, 2025",89,3010,https://arxiv.org/abs/2511.10647 -33,Paper2Video: Automatic Video Generation from Scientific Papers,55,2,13.04,11,"Oct 6, 2025",115,1790,https://arxiv.org/abs/2510.05096 -34,"PC-Agent: A Hierarchical Multi-Agent Collaboration Framework for Complex - Task Automation on PC",222,3,22.85,18,"Feb 20, 2025",29,6440,https://arxiv.org/abs/2502.14282 -35,"NextStep-1: Toward Autoregressive Image Generation with Continuous - Tokens at Scale",52,1,12.92,5,"Aug 14, 2025",134,496,https://arxiv.org/abs/2508.10711 -36,"Look Before You Leap: A GUI-Critic-R1 Model for Pre-Operative Error - Diagnosis in GUI Automation",220,4,22.92,19,"Jun 5, 2025",19,6430,https://arxiv.org/abs/2506.04614 -37,"PokeeResearch: Effective Deep Research via Reinforcement Learning from - AI Feedback and Robust Reasoning Scaffold",47,4,12.15,7,"Oct 17, 2025",8,1610,https://arxiv.org/abs/2510.15862 -38,LongSplat: Robust Unposed 3D Gaussian Splatting for Casual Long Videos,59,3,14.98,10,"Aug 19, 2025",59,600,https://arxiv.org/abs/2508.14041 -39,StableAvatar: Infinite-Length Audio-Driven Avatar Video Generation,137,1,21.63,20,"Aug 11, 2025",27,1020,https://arxiv.org/abs/2508.08248 -40,SAM 3D: 3Dfy Anything in Images,25,1,2.56,3,"Nov 20, 2025",100,4200,https://arxiv.org/abs/2511.16624 -41,"ToonComposer: Streamlining Cartoon Production with Generative +51,"Agent S2: A Compositional Generalist-Specialist Framework for Computer + Use Agents",145,2,30.46,35,"Apr 1, 2025",26,8450,https://arxiv.org/abs/2504.00906 +52,The Unreasonable Effectiveness of Scaling Agents for Computer Use,138,2,29.52,34,"Oct 2, 2025",24,8450,https://arxiv.org/abs/2510.02250 +53,"VLA-Adapter: An Effective Paradigm for Tiny-Scale Vision-Language-Action + Model",131,7,28.43,29,"Sep 11, 2025",235,1590,https://arxiv.org/abs/2509.09372 +54,A Survey of Reinforcement Learning for Large Reasoning Models,110,4,24.15,24,"Sep 10, 2025",183,1800,https://arxiv.org/abs/2509.08827 +55,Less is More: Recursive Reasoning with Tiny Networks,90,1,20.33,18,"Oct 6, 2025",483,5670,https://arxiv.org/abs/2510.04871 +56,LightRAG: Simple and Fast Retrieval-Augmented Generation,92,2,21.25,9,"Oct 8, 2024",20,24900,https://arxiv.org/abs/2410.05779 +57,PyTorch FSDP: Experiences on Scaling Fully Sharded Data Parallel,104,10,27.97,30,"Apr 21, 2023",4,95500,https://arxiv.org/abs/2304.11277 +58,Matrix-Game: Interactive World Foundation Model,78,1,20.45,14,"Jun 23, 2025",72,1550,https://arxiv.org/abs/2506.18701 +59,PyTorch Distributed: Experiences on Accelerating Data Parallel Training,105,9,28.43,31,"Jun 28, 2020",3,95500,https://arxiv.org/abs/2006.15704 +60,Thyme: Think Beyond Images,82,1,23.55,23,"Aug 15, 2025",78,466,https://arxiv.org/abs/2508.11630 +61,"RepoMaster: Autonomous Exploration and Understanding of GitHub + Repositories for Complex Task Solving",83,14,23.9,21,"May 27, 2025",2,348,https://arxiv.org/abs/2505.21577 +62,UI-Venus Technical Report: Building High-performance UI Agents with RFT,74,7,21.26,17,"Aug 14, 2025",41,470,https://arxiv.org/abs/2508.10833 +63,"ToonComposer: Streamlining Cartoon Production with Generative Post-Keyframing",65,4,17.37,14,"Aug 14, 2025",50,358,https://arxiv.org/abs/2508.10881 -42,"Stand-In: A Lightweight and Plug-and-Play Identity Control for Video +64,OpenCUA: Open Foundations for Computer-Use Agents,85,14,25.86,24,"Aug 12, 2025",30,415,https://arxiv.org/abs/2508.09123 +65,LongSplat: Robust Unposed 3D Gaussian Splatting for Casual Long Videos,59,3,14.98,10,"Aug 19, 2025",59,600,https://arxiv.org/abs/2508.14041 +66,"Stand-In: A Lightweight and Plug-and-Play Identity Control for Video Generation",63,2,17.37,9,"Aug 11, 2025",38,512,https://arxiv.org/abs/2508.07901 -43,"ScreenCoder: Advancing Visual-to-Code Generation for Front-End - Automation via Modular Multimodal Agents",122,1,21.56,24,"Jul 30, 2025",98,2290,https://arxiv.org/abs/2507.22827 -44,"HuMo: Human-Centric Video Generation via Collaborative Multi-Modal +67,Zep: A Temporal Knowledge Graph Architecture for Agent Memory,95,11,28.88,31,"Jan 20, 2025",6,20600,https://arxiv.org/abs/2501.13956 +68,Paper2Video: Automatic Video Generation from Scientific Papers,55,2,13.04,11,"Oct 6, 2025",115,1790,https://arxiv.org/abs/2510.05096 +69,"Youtu-GraphRAG: Vertically Unified Agents for Graph Retrieval-Augmented + Complex Reasoning",72,6,22.89,17,"Aug 27, 2025",7,730,https://arxiv.org/abs/2508.19855 +70,"VeOmni: Scaling Any Modality Model Training with Model-Centric + Distributed Recipe Zoo",66,1,20.94,18,"Aug 4, 2025",17,1180,https://arxiv.org/abs/2508.02317 +71,"NextStep-1: Toward Autoregressive Image Generation with Continuous + Tokens at Scale",52,1,12.92,5,"Aug 14, 2025",134,496,https://arxiv.org/abs/2508.10711 +72,Qwen3-Omni Technical Report,72,1,23.96,26,"Sep 22, 2025",128,2690,https://arxiv.org/abs/2509.17765 +73,3D and 4D World Modeling: A Survey,66,4,22.12,20,"Sep 4, 2025",57,568,https://arxiv.org/abs/2509.07996 +74,"ScaleCUA: Scaling Open-Source Computer Use Agents with Cross-Platform + Data",78,10,27.24,29,"Sep 18, 2025",105,610,https://arxiv.org/abs/2509.15221 +75,VibeVoice Technical Report,37,1,1,1,"Aug 26, 2025",118,7980,https://arxiv.org/abs/2508.19205 +76,"PokeeResearch: Effective Deep Research via Reinforcement Learning from + AI Feedback and Robust Reasoning Scaffold",47,4,12.15,7,"Oct 17, 2025",8,1610,https://arxiv.org/abs/2510.15862 +77,"MCP-Bench: Benchmarking Tool-Using LLM Agents with Complex Real-World + Tasks via MCP Servers",69,5,25.25,23,"Aug 28, 2025",58,296,https://arxiv.org/abs/2508.20453 +78,Depth Anything 3: Recovering the Visual Space from Any Views,40,1,6.58,3,"Nov 13, 2025",89,3010,https://arxiv.org/abs/2511.10647 +79,"MCP-Universe: Benchmarking Large Language Models with Real-World Model + Context Protocol Servers",57,5,20.07,18,"Aug 20, 2025",41,352,https://arxiv.org/abs/2508.14704 +80,HunyuanImage 3.0 Technical Report,50,1,15.78,7,"Sep 28, 2025",21,2210,https://arxiv.org/abs/2509.23951 +81,"olmOCR: Unlocking Trillions of Tokens in PDFs with Vision Language + Models",56,3,19.68,20,"Feb 25, 2025",9,16000,https://arxiv.org/abs/2502.18443 +82,Ovi: Twin Backbone Cross-Modal Fusion for Audio-Video Generation,68,2,25.31,22,"Sep 30, 2025",32,1300,https://arxiv.org/abs/2510.01284 +83,Pico-Banana-400K: A Large-Scale Dataset for Text-Guided Image Editing,54,1,18.81,9,"Oct 22, 2025",28,1580,https://arxiv.org/abs/2510.19808 +84,A Survey of Context Engineering for Large Language Models,151,1,39.59,40,"Jul 17, 2025",256,2330,https://arxiv.org/abs/2507.13334 +85,"HuMo: Human-Centric Video Generation via Collaborative Multi-Modal Conditioning",46,2,14.24,9,"Sep 10, 2025",120,596,https://arxiv.org/abs/2509.08519 -45,"Paper2Agent: Reimagining Research Papers As Interactive and Reliable AI - Agents",135,5,22.01,20,"Sep 8, 2025",41,1660,https://arxiv.org/abs/2509.06917 -46,HunyuanImage 3.0 Technical Report,50,1,15.78,7,"Sep 28, 2025",21,2210,https://arxiv.org/abs/2509.23951 -47,Less is More: Recursive Reasoning with Tiny Networks,90,1,20.33,18,"Oct 6, 2025",483,5670,https://arxiv.org/abs/2510.04871 -48,"Concerto: Joint 2D-3D Self-Supervised Learning Emerges Spatial - Representations",22,1,3.5,2,"Oct 27, 2025",172,2610,https://arxiv.org/abs/2510.23607 -49,"The Dragon Hatchling: The Missing Link between the Transformer and +86,"EmbodiedOneVision: Interleaved Vision-Text-Action Pretraining for + General Robot Control",52,2,20.06,15,"Aug 28, 2025",76,282,https://arxiv.org/abs/2508.21112 +87,Waver: Wave Your Way to Lifelike Video Generation,53,5,21.06,22,"Aug 21, 2025",33,499,https://arxiv.org/abs/2508.15761 +88,"AudioStory: Generating Long-Form Narrative Audio with Large Language + Models",59,16,24.81,21,"Aug 27, 2025",20,264,https://arxiv.org/abs/2508.20088 +89,Towards a Unified View of Large Language Model Post-Training,46,5,17.63,19,"Sep 4, 2025",61,100,https://arxiv.org/abs/2509.04419 +90,"InternVL3.5: Advancing Open-Source Multimodal Models in Versatility, + Reasoning, and Efficiency",109,12,37.03,40,"Aug 25, 2025",191,9240,https://arxiv.org/abs/2508.18265 +91,Transition Models: Rethinking the Generative Learning Objective,47,8,19.38,19,"Sep 4, 2025",28,94,https://arxiv.org/abs/2509.04394 +92,"AgentGym-RL: Training LLM Agents for Long-Horizon Decision Making + through Multi-Turn Reinforcement Learning",43,1,16.98,8,"Sep 10, 2025",56,362,https://arxiv.org/abs/2509.08755 +93,"The Dragon Hatchling: The Missing Link between the Transformer and Models of the Brain",38,1,13.08,1,"Sep 30, 2025",489,3120,https://arxiv.org/abs/2509.26507 -50,"MOSAIC: Multi-Subject Personalized Generation via Correspondence-Aware +94,STream3R: Scalable Sequential 3D Reconstruction with Causal Transformer,39,5,14.41,8,"Aug 14, 2025",30,194,https://arxiv.org/abs/2508.10893 +95,SimpleVLA-RL: Scaling VLA Training via Reinforcement Learning,56,1,25.86,28,"Sep 11, 2025",73,740,https://arxiv.org/abs/2509.09674 +96,"Enterprise Deep Research: Steerable Multi-Agent Deep Research for + Enterprise Analytics",46,7,21.17,18,"Oct 20, 2025",9,952,https://arxiv.org/abs/2510.17797 +97,Code2Video: A Code-centric Paradigm for Educational Video Generation,48,2,22.77,19,"Oct 1, 2025",33,1100,https://arxiv.org/abs/2510.01174 +98,RynnEC: Bringing MLLMs into Embodied World,62,6,29.16,29,"Aug 19, 2025",18,337,https://arxiv.org/abs/2508.14160 +99,"Kimi Linear: An Expressive, Efficient Attention Architecture",39,1,16.28,4,"Oct 30, 2025",102,1150,https://arxiv.org/abs/2510.26692 +100,UI-S1: Advancing GUI Automation via Semi-online Reinforcement Learning,67,15,30.96,31,"Sep 15, 2025",47,6440,https://arxiv.org/abs/2509.11543 +101,"Chain-of-Agents: End-to-End Agent Foundation Models via Multi-Agent + Distillation and Agentic RL",43,5,20.6,19,"Aug 6, 2025",115,333,https://arxiv.org/abs/2508.13167 +102,"MOSAIC: Multi-Subject Personalized Generation via Correspondence-Aware Alignment and Disentanglement",33,5,11.58,9,"Sep 2, 2025",11,418,https://arxiv.org/abs/2509.01977 -51,SAM 3: Segment Anything with Concepts,19,1,1.63,2,"Nov 20, 2025",95,4760,https://arxiv.org/abs/2511.16719 -52,STream3R: Scalable Sequential 3D Reconstruction with Causal Transformer,39,5,14.41,8,"Aug 14, 2025",30,194,https://arxiv.org/abs/2508.10893 -53,LightRAG: Simple and Fast Retrieval-Augmented Generation,92,2,21.25,9,"Oct 8, 2024",20,24900,https://arxiv.org/abs/2410.05779 -54,Matrix-Game: Interactive World Foundation Model,78,1,20.45,14,"Jun 23, 2025",72,1550,https://arxiv.org/abs/2506.18701 -55,"We-Math 2.0: A Versatile MathBook System for Incentivizing Visual +103,"A Survey of Scientific Large Language Models: From Data Foundations to + Agent Frontiers",44,7,21.52,16,"Aug 28, 2025",133,312,https://arxiv.org/abs/2508.21148 +104,MeshCoder: LLM-Powered Structured Mesh Code Generation from Point Clouds,40,4,18.7,16,"Aug 20, 2025",67,374,https://arxiv.org/abs/2508.14879 +105,From Editor to Dense Geometry Estimator,55,13,28.02,28,"Sep 4, 2025",87,156,https://arxiv.org/abs/2509.04338 +106,"FlashVSR: Towards Real-Time Diffusion-Based Streaming Video + Super-Resolution",55,7,28.15,29,"Oct 14, 2025",37,910,https://arxiv.org/abs/2510.12747 +107,"FantasyPortrait: Enhancing Multi-Character Portrait Animation with + Expression-Augmented Diffusion Transformers",55,11,28.22,26,"Jul 17, 2025",24,435,https://arxiv.org/abs/2507.12956 +108,SpatialLM: Training Large Language Models for Structured Indoor Modeling,50,4,26.46,31,"Jun 9, 2025",49,3900,https://arxiv.org/abs/2506.07491 +109,Reconstruction Alignment Improves Unified Multimodal Models,34,3,14.94,14,"Sep 8, 2025",38,198,https://arxiv.org/abs/2509.07295 +110,Transformer Explainer: Interactive Learning of Text-Generative Models,81,21,35.89,35,"Aug 8, 2024",172,5960,https://arxiv.org/abs/2408.04619 +111,VerlTool: Towards Holistic Agentic Reinforcement Learning with Tool Use,52,7,27.48,28,"Sep 1, 2025",64,499,https://arxiv.org/abs/2509.01055 +112,"Beyond Ten Turns: Unlocking Long-Horizon Agentic Search with Large-Scale + Asynchronous RL",47,3,25.02,19,"Aug 11, 2025",48,373,https://arxiv.org/abs/2508.07976 +113,MobiAgent: A Systematic Framework for Customizable Mobile Agents,43,6,22.74,21,"Aug 30, 2025",6,1020,https://arxiv.org/abs/2509.00531 +114,"AReaL: A Large-Scale Asynchronous Reinforcement Learning System for + Language Reasoning",59,18,30.47,29,"May 30, 2025",28,2660,https://arxiv.org/abs/2505.24298 +115,SAM 3D: 3Dfy Anything in Images,25,1,2.56,3,"Nov 20, 2025",100,4200,https://arxiv.org/abs/2511.16624 +116,Continuous Autoregressive Language Models,46,5,24.72,25,"Oct 31, 2025",64,582,https://arxiv.org/abs/2510.27688 +117,"We-Math 2.0: A Versatile MathBook System for Incentivizing Visual Mathematical Reasoning",31,1,12.16,3,"Aug 14, 2025",139,138,https://arxiv.org/abs/2508.10433 -56,Back to Basics: Let Denoising Generative Models Denoise,28,2,11.43,11,"Nov 17, 2025",59,1450,https://arxiv.org/abs/2511.13720 -57,Diffusion Transformers with Representation Autoencoders,29,1,12.24,7,"Oct 13, 2025",157,1350,https://arxiv.org/abs/2510.11690 -58,"AgentGym-RL: Training LLM Agents for Long-Horizon Decision Making - through Multi-Turn Reinforcement Learning",43,1,16.98,8,"Sep 10, 2025",56,362,https://arxiv.org/abs/2509.08755 -59,Pico-Banana-400K: A Large-Scale Dataset for Text-Guided Image Editing,54,1,18.81,9,"Oct 22, 2025",28,1580,https://arxiv.org/abs/2510.19808 -60,Towards a Unified View of Large Language Model Post-Training,46,5,17.63,19,"Sep 4, 2025",61,100,https://arxiv.org/abs/2509.04419 -61,"Kimi Linear: An Expressive, Efficient Attention Architecture",39,1,16.28,4,"Oct 30, 2025",102,1150,https://arxiv.org/abs/2510.26692 -62,Reconstruction Alignment Improves Unified Multimodal Models,34,3,14.94,14,"Sep 8, 2025",38,198,https://arxiv.org/abs/2509.07295 -63,"olmOCR: Unlocking Trillions of Tokens in PDFs with Vision Language - Models",56,3,19.68,20,"Feb 25, 2025",9,16000,https://arxiv.org/abs/2502.18443 -64,UI-Venus Technical Report: Building High-performance UI Agents with RFT,74,7,21.26,17,"Aug 14, 2025",41,470,https://arxiv.org/abs/2508.10833 -65,"FilmAgent: A Multi-Agent Framework for End-to-End Film Automation in - Virtual 3D Spaces",20,1,8.75,3,"Jan 22, 2025",74,1050,https://arxiv.org/abs/2501.12909 -66,HunyuanOCR Technical Report,15,2,3.27,3,"Nov 24, 2025",17,878,https://arxiv.org/abs/2511.19575 -67,"MCP-Universe: Benchmarking Large Language Models with Real-World Model - Context Protocol Servers",57,5,20.07,18,"Aug 20, 2025",41,352,https://arxiv.org/abs/2508.14704 -68,OmniWorld: A Multi-Domain and Multi-Modal Dataset for 4D World Modeling,26,1,13.12,13,"Sep 15, 2025",98,342,https://arxiv.org/abs/2509.12201 -69,"VeOmni: Scaling Any Modality Model Training with Model-Centric - Distributed Recipe Zoo",66,1,20.94,18,"Aug 4, 2025",17,1180,https://arxiv.org/abs/2508.02317 -70,Agent0: Unleashing Self-Evolving Agents from Zero Data via Tool-Integrated Reasoning,24,6,12.17,9,"Nov 20, 2025",97,705,https://arxiv.org/abs/2511.16043 -71,Multi-View 3D Point Tracking,33,4,15.97,16,"Aug 28, 2025",20,314,https://arxiv.org/abs/2508.21060 -72,"Revolutionizing Reinforcement Learning Framework for Diffusion Large - Language Models",25,2,12.96,5,"Sep 8, 2025",51,171,https://arxiv.org/abs/2509.06949 -73,General Agentic Memory Via Deep Research,17,3,7.47,8,"Nov 23, 2025",146,485,https://arxiv.org/abs/2511.18423 -74,"Fish-Speech: Leveraging Large Language Models for Advanced Multilingual +118,Matrix-3D: Omnidirectional Explorable 3D World Generation,45,6,24.27,23,"Aug 11, 2025",67,398,https://arxiv.org/abs/2508.08086 +119,DeepAgent: A General Reasoning Agent with Scalable Toolsets,43,5,23.09,25,"Oct 24, 2025",93,739,https://arxiv.org/abs/2510.21618 +120,Logics-Parsing Technical Report,48,4,26.27,28,"Sep 24, 2025",7,619,https://arxiv.org/abs/2509.19760 +121,"MonkeyOCR: Document Parsing with a Structure-Recognition-Relation + Triplet Paradigm",66,16,33.12,31,"Jun 5, 2025",2,6000,https://arxiv.org/abs/2506.05218 +122,Emu3.5: Native Multimodal Models are World Learners,35,3,17.94,6,"Oct 30, 2025",103,1210,https://arxiv.org/abs/2510.26583 +123,Multi-View 3D Point Tracking,33,4,15.97,16,"Aug 28, 2025",20,314,https://arxiv.org/abs/2508.21060 +124,"Matrix-Game 2.0: An Open-Source, Real-Time, and Streaming Interactive + World Model",41,4,22.83,26,"Aug 18, 2025",22,1550,https://arxiv.org/abs/2508.13009 +125,"GitTaskBench: A Benchmark for Code Agents Solving Real-World Tasks + Through Code Repository Leveraging",50,13,27.96,23,"Aug 26, 2025",3,183,https://arxiv.org/abs/2508.18993 +126,"Mini-o3: Scaling Up Reasoning Patterns and Interaction Turns for Visual + Search",39,2,21.49,21,"Sep 9, 2025",59,290,https://arxiv.org/abs/2509.07969 +127,"GLM-4.1V-Thinking: Towards Versatile Multimodal Reasoning with Scalable + Reinforcement Learning",53,10,29.38,29,"Jul 1, 2025",229,1500,https://arxiv.org/abs/2507.01006 +128,Diffusion Transformers with Representation Autoencoders,29,1,12.24,7,"Oct 13, 2025",157,1350,https://arxiv.org/abs/2510.11690 +129,SpatialVID: A Large-Scale Video Dataset with Spatial Annotations,35,8,19.11,16,"Sep 11, 2025",28,331,https://arxiv.org/abs/2509.09676 +130,Back to Basics: Let Denoising Generative Models Denoise,28,2,11.43,11,"Nov 17, 2025",59,1450,https://arxiv.org/abs/2511.13720 +131,"ComoRAG: A Cognitive-Inspired Memory-Organized RAG for Stateful Long + Narrative Reasoning",44,8,25.86,28,"Aug 14, 2025",70,192,https://arxiv.org/abs/2508.10419 +132,"Pixie: Fast and Generalizable Supervised Learning of 3D Physics from + Pixels",34,9,19.24,19,"Aug 20, 2025",34,179,https://arxiv.org/abs/2508.17437 +133,"In-the-Flow Agentic System Optimization for Effective Planning and Tool + Use",36,2,21.08,22,"Oct 7, 2025",89,851,https://arxiv.org/abs/2510.05592 +134,"Lyra: Generative 3D Scene Reconstruction via Video Diffusion Model + Self-Distillation",36,10,21.92,15,"Sep 23, 2025",21,458,https://arxiv.org/abs/2509.19296 +135,"HunyuanWorld 1.0: Generating Immersive, Explorable, and Interactive 3D + Worlds from Words or Pixels",76,22,37.25,38,"Jul 29, 2025",126,2170,https://arxiv.org/abs/2507.21809 +136,"Concerto: Joint 2D-3D Self-Supervised Learning Emerges Spatial + Representations",22,1,3.5,2,"Oct 27, 2025",172,2610,https://arxiv.org/abs/2510.23607 +137,LongCat-Video Technical Report,40,6,25.38,28,"Oct 25, 2025",24,1110,https://arxiv.org/abs/2510.22200 +138,Parallel-R1: Towards Parallel Thinking via Reinforcement Learning,41,5,26.07,27,"Sep 9, 2025",95,151,https://arxiv.org/abs/2509.07980 +139,"Fish-Speech: Leveraging Large Language Models for Advanced Multilingual Text-to-Speech Synthesis",28,2,14.79,10,"Nov 2, 2024",11,23900,https://arxiv.org/abs/2411.01156 -75,Transition Models: Rethinking the Generative Learning Objective,47,8,19.38,19,"Sep 4, 2025",28,94,https://arxiv.org/abs/2509.04394 -76,Agent0-VL: Exploring Self-Evolving Agent for Tool-Integrated Vision-Language Reasoning,15,4,5.33,5,"Nov 25, 2025",46,702,https://arxiv.org/abs/2511.19900 -77,"EmbodiedOneVision: Interleaved Vision-Text-Action Pretraining for - General Robot Control",52,2,20.06,15,"Aug 28, 2025",76,282,https://arxiv.org/abs/2508.21112 -78,WizardCoder: Empowering Code Large Language Models with Evol-Instruct,12,1,1,1,"Jun 14, 2023",30,9470,https://arxiv.org/abs/2306.08568 -79,WithAnyone: Towards Controllable and ID Consistent Image Generation,24,1,13.58,9,"Oct 16, 2025",76,415,https://arxiv.org/abs/2510.14975 -80,MeshCoder: LLM-Powered Structured Mesh Code Generation from Point Clouds,40,4,18.7,16,"Aug 20, 2025",67,374,https://arxiv.org/abs/2508.14879 -81,Emu3.5: Native Multimodal Models are World Learners,35,3,17.94,6,"Oct 30, 2025",103,1210,https://arxiv.org/abs/2510.26583 -82,Intern-S1: A Scientific Multimodal Foundation Model,27,1,15.81,14,"Aug 21, 2025",236,538,https://arxiv.org/abs/2508.15763 -83,LongLive: Real-time Interactive Long Video Generation,29,1,16.55,18,"Sep 26, 2025",174,647,https://arxiv.org/abs/2509.22622 -84,"Tiny Model, Big Logic: Diversity-Driven Optimization Elicits Large-Model +140,LongLive: Real-time Interactive Long Video Generation,29,1,16.55,18,"Sep 26, 2025",174,647,https://arxiv.org/abs/2509.22622 +141,"Tiny Model, Big Logic: Diversity-Driven Optimization Elicits Large-Model Reasoning Ability in VibeThinker-1.5B",29,7,16.55,10,"Nov 9, 2025",117,483,https://arxiv.org/abs/2511.06221 -85,Waver: Wave Your Way to Lifelike Video Generation,53,5,21.06,22,"Aug 21, 2025",33,499,https://arxiv.org/abs/2508.15761 -86,3D and 4D World Modeling: A Survey,66,4,22.12,20,"Sep 4, 2025",57,568,https://arxiv.org/abs/2509.07996 -87,"Pref-GRPO: Pairwise Preference Reward-based GRPO for Stable - Text-to-Image Reinforcement Learning",26,2,16.62,17,"Aug 28, 2025",85,123,https://arxiv.org/abs/2508.20751 -88,SpatialVID: A Large-Scale Video Dataset with Spatial Annotations,35,8,19.11,16,"Sep 11, 2025",28,331,https://arxiv.org/abs/2509.09676 -89,LightMem: Lightweight and Efficient Memory-Augmented Generation,24,2,16.12,13,"Oct 21, 2025",105,295,https://arxiv.org/abs/2510.18866 -90,"Chain-of-Agents: End-to-End Agent Foundation Models via Multi-Agent - Distillation and Agentic RL",43,5,20.6,19,"Aug 6, 2025",115,333,https://arxiv.org/abs/2508.13167 -91,FlashWorld: High-quality 3D Scene Generation within Seconds,18,3,13.17,8,"Oct 15, 2025",66,397,https://arxiv.org/abs/2510.13678 -92,"Pixie: Fast and Generalizable Supervised Learning of 3D Physics from - Pixels",34,9,19.24,19,"Aug 20, 2025",34,179,https://arxiv.org/abs/2508.17437 -93,"Youtu-GraphRAG: Vertically Unified Agents for Graph Retrieval-Augmented - Complex Reasoning",72,6,22.89,17,"Aug 27, 2025",7,730,https://arxiv.org/abs/2508.19855 -94,"Enterprise Deep Research: Steerable Multi-Agent Deep Research for - Enterprise Analytics",46,7,21.17,18,"Oct 20, 2025",9,952,https://arxiv.org/abs/2510.17797 -95,"Scaling Instruction-Based Video Editing with a High-Quality Synthetic - Dataset",23,3,16.39,15,"Oct 17, 2025",49,385,https://arxiv.org/abs/2510.15742 -96,"Rethinking Semantic Segmentation from a Sequence-to-Sequence Perspective - with Transformers",21,2,15.81,10,"Dec 31, 2020",3,1100,https://arxiv.org/abs/2012.15840 -97,GigaWorld-0: World Models as Data Engine to Empower Embodied AI,15,8,11.93,13,"Nov 25, 2025",28,214,https://arxiv.org/abs/2511.19861 -98,Detect Anything via Next Point Prediction,22,4,16.45,10,"Oct 14, 2025",42,449,https://arxiv.org/abs/2510.12798 -99,A Survey of Reinforcement Learning for Large Reasoning Models,110,4,24.15,24,"Sep 10, 2025",183,1800,https://arxiv.org/abs/2509.08827 -100,Thyme: Think Beyond Images,82,1,23.55,23,"Aug 15, 2025",78,466,https://arxiv.org/abs/2508.11630 -101,"A Survey of Scientific Large Language Models: From Data Foundations to - Agent Frontiers",44,7,21.52,16,"Aug 28, 2025",133,312,https://arxiv.org/abs/2508.21148 -102,StreamingVLM: Real-Time Understanding for Infinite Video Streams,26,3,18.77,23,"Oct 10, 2025",46,557,https://arxiv.org/abs/2510.09608 -103,"Tinker: Diffusion's Gift to 3D--Multi-View Consistent Editing From +142,OmniWorld: A Multi-Domain and Multi-Modal Dataset for 4D World Modeling,26,1,13.12,13,"Sep 15, 2025",98,342,https://arxiv.org/abs/2509.12201 +143,JoyAgent-JDGenie: Technical Report on the GAIA,61,14,35.08,38,"Oct 1, 2025",3,11100,https://arxiv.org/abs/2510.00510 +144,Ovis2.5 Technical Report,39,6,26.36,27,"Aug 15, 2025",102,1280,https://arxiv.org/abs/2508.11737 +145,Marco-Voice Technical Report,53,25,32.89,33,"Aug 4, 2025",15,332,https://arxiv.org/abs/2508.02038 +146,"A.S.E: A Repository-Level Benchmark for Evaluating Security in + AI-Generated Code",62,7,35.66,37,"Aug 25, 2025",340,709,https://arxiv.org/abs/2508.18106 +147,"Revolutionizing Reinforcement Learning Framework for Diffusion Large + Language Models",25,2,12.96,5,"Sep 8, 2025",51,171,https://arxiv.org/abs/2509.06949 +148,Intern-S1: A Scientific Multimodal Foundation Model,27,1,15.81,14,"Aug 21, 2025",236,538,https://arxiv.org/abs/2508.15763 +149,SAM 3: Segment Anything with Concepts,19,1,1.63,2,"Nov 20, 2025",95,4760,https://arxiv.org/abs/2511.16719 +150,Agent0: Unleashing Self-Evolving Agents from Zero Data via Tool-Integrated Reasoning,24,6,12.17,9,"Nov 20, 2025",97,705,https://arxiv.org/abs/2511.16043 +151,"Tinker: Diffusion's Gift to 3D--Multi-View Consistent Editing From Sparse Inputs without Per-Scene Optimization",29,7,19.86,21,"Aug 20, 2025",39,129,https://arxiv.org/abs/2508.14811 -104,"In-the-Flow Agentic System Optimization for Effective Planning and Tool - Use",36,2,21.08,22,"Oct 7, 2025",89,851,https://arxiv.org/abs/2510.05592 -105,"Mini-o3: Scaling Up Reasoning Patterns and Interaction Turns for Visual - Search",39,2,21.49,21,"Sep 9, 2025",59,290,https://arxiv.org/abs/2509.07969 -106,"RepoMaster: Autonomous Exploration and Understanding of GitHub - Repositories for Complex Task Solving",83,14,23.9,21,"May 27, 2025",2,348,https://arxiv.org/abs/2505.21577 -107,PhysX-Anything: Simulation-Ready Physical 3D Assets from Single Image,25,2,19.24,15,"Nov 17, 2025",51,593,https://arxiv.org/abs/2511.13648 -108,"Thinking with Video: Video Generation as a Promising Multimodal - Reasoning Paradigm",26,2,19.81,10,"Nov 6, 2025",188,182,https://arxiv.org/abs/2511.04570 -109,Code2Video: A Code-centric Paradigm for Educational Video Generation,48,2,22.77,19,"Oct 1, 2025",33,1100,https://arxiv.org/abs/2510.01174 -110,"VoxHammer: Training-Free Precise and Coherent 3D Editing in Native 3D +152,WithAnyone: Towards Controllable and ID Consistent Image Generation,24,1,13.58,9,"Oct 16, 2025",76,415,https://arxiv.org/abs/2510.14975 +153,ST-Raptor: LLM-Powered Semi-Structured Table Question Answering,32,3,22.97,24,"Aug 25, 2025",6,243,https://arxiv.org/abs/2508.18190 +154,"Pref-GRPO: Pairwise Preference Reward-based GRPO for Stable + Text-to-Image Reinforcement Learning",26,2,16.62,17,"Aug 28, 2025",85,123,https://arxiv.org/abs/2508.20751 +155,Arch-Router: Aligning LLM Routing with Human Preferences,60,21,36.28,37,"Jun 19, 2025",17,4170,https://arxiv.org/abs/2506.16655 +156,GenCompositor: Generative Video Compositing with Diffusion Transformer,44,12,31.52,29,"Sep 2, 2025",24,122,https://arxiv.org/abs/2509.02460 +157,"Voost: A Unified and Scalable Diffusion Transformer for Bidirectional + Virtual Try-On and Try-Off",37,9,28.11,26,"Aug 6, 2025",56,264,https://arxiv.org/abs/2508.04825 +158,"FilmAgent: A Multi-Agent Framework for End-to-End Film Automation in + Virtual 3D Spaces",20,1,8.75,3,"Jan 22, 2025",74,1050,https://arxiv.org/abs/2501.12909 +159,"Story2Board: A Training-Free Approach for Expressive Storyboard + Generation",45,14,32.31,34,"Aug 13, 2025",61,119,https://arxiv.org/abs/2508.09983 +160,StreamingVLM: Real-Time Understanding for Infinite Video Streams,26,3,18.77,23,"Oct 10, 2025",46,557,https://arxiv.org/abs/2510.09608 +161,LightMem: Lightweight and Efficient Memory-Augmented Generation,24,2,16.12,13,"Oct 21, 2025",105,295,https://arxiv.org/abs/2510.18866 +162,"VoxHammer: Training-Free Precise and Coherent 3D Editing in Native 3D Space",27,3,20.3,23,"Aug 26, 2025",38,130,https://arxiv.org/abs/2508.19247 -111,"SINQ: Sinkhorn-Normalized Quantization for Calibration-Free +163,RLinf-VLA: A Unified and Efficient Framework for VLA+RL Training,86,25,41.38,42,"Oct 8, 2025",38,1450,https://arxiv.org/abs/2510.06710 +164,"CommonForms: A Large, Diverse Dataset for Form Field Detection",29,2,22.59,27,"Sep 20, 2025",18,819,https://arxiv.org/abs/2509.16506 +165,"Thinking with Video: Video Generation as a Promising Multimodal + Reasoning Paradigm",26,2,19.81,10,"Nov 6, 2025",188,182,https://arxiv.org/abs/2511.04570 +166,"Speed Always Wins: A Survey on Efficient Architectures for Large + Language Models",36,15,28.69,25,"Aug 13, 2025",51,281,https://arxiv.org/abs/2508.09834 +167,"Scaling Instruction-Based Video Editing with a High-Quality Synthetic + Dataset",23,3,16.39,15,"Oct 17, 2025",49,385,https://arxiv.org/abs/2510.15742 +168,PhysX-Anything: Simulation-Ready Physical 3D Assets from Single Image,25,2,19.24,15,"Nov 17, 2025",51,593,https://arxiv.org/abs/2511.13648 +169,"Echo-4o: Harnessing the Power of GPT-4o Synthetic Images for Improved + Image Generation",27,4,21.67,20,"Aug 13, 2025",23,73,https://arxiv.org/abs/2508.09987 +170,TTT3R: 3D Reconstruction as Test-Time Training,29,3,24.1,28,"Sep 30, 2025",13,396,https://arxiv.org/abs/2509.26645 +171,"SINQ: Sinkhorn-Normalized Quantization for Calibration-Free Low-Precision LLM Weights",25,5,19.96,19,"Sep 26, 2025",73,472,https://arxiv.org/abs/2509.22944 -112,Dolphin: Document Image Parsing via Heterogeneous Anchor Prompting,18,8,17.78,15,"May 20, 2025",3,7670,https://arxiv.org/abs/2505.14059 -113,Qwen3-Omni Technical Report,72,1,23.96,26,"Sep 22, 2025",128,2690,https://arxiv.org/abs/2509.17765 -114,"Lyra: Generative 3D Scene Reconstruction via Video Diffusion Model - Self-Distillation",36,10,21.92,15,"Sep 23, 2025",21,458,https://arxiv.org/abs/2509.19296 -115,MobiAgent: A Systematic Framework for Customizable Mobile Agents,43,6,22.74,21,"Aug 30, 2025",6,1020,https://arxiv.org/abs/2509.00531 -116,"QeRL: Beyond Efficiency -- Quantization-enhanced Reinforcement Learning +172,Detect Anything via Next Point Prediction,22,4,16.45,10,"Oct 14, 2025",42,449,https://arxiv.org/abs/2510.12798 +173,General Agentic Memory Via Deep Research,17,3,7.47,8,"Nov 23, 2025",146,485,https://arxiv.org/abs/2511.18423 +174,"Rethinking Semantic Segmentation from a Sequence-to-Sequence Perspective + with Transformers",21,2,15.81,10,"Dec 31, 2020",3,1100,https://arxiv.org/abs/2012.15840 +175,"Killing Two Birds with One Stone:Efficient and Robust Training of Face + Recognition CNNs by Partial FC",92,26,43.16,45,"Mar 28, 2022",3,26700,https://arxiv.org/abs/2203.15565 +176,Reinforcement Learning in Vision: A Survey,41,25,33.49,33,"Aug 11, 2025",27,184,https://arxiv.org/abs/2508.08189 +177,HunyuanOCR Technical Report,15,2,3.27,3,"Nov 24, 2025",17,878,https://arxiv.org/abs/2511.19575 +178,"Collaborating Action by Action: A Multi-agent LLM Framework for Embodied + Reasoning",26,10,23.54,21,"Apr 24, 2025",5,4250,https://arxiv.org/abs/2504.17950 +179,Training Video Foundation Models with NVIDIA NeMo,87,19,42.82,44,"Mar 17, 2025",7,15700,https://arxiv.org/abs/2503.12964 +180,"MiroThinker: Pushing the Performance Boundaries of Open-Source Research Agents via Model, Context, and Interactive Scaling",29,4,26.66,30,"Nov 14, 2025",156,1160,https://arxiv.org/abs/2511.11793 +181,Agent0-VL: Exploring Self-Evolving Agent for Tool-Integrated Vision-Language Reasoning,15,4,5.33,5,"Nov 25, 2025",46,702,https://arxiv.org/abs/2511.19900 +182,OmniTry: Virtual Try-On Anything without Masks,41,13,34.39,34,"Aug 19, 2025",16,166,https://arxiv.org/abs/2508.13632 +183,FlashWorld: High-quality 3D Scene Generation within Seconds,18,3,13.17,8,"Oct 15, 2025",66,397,https://arxiv.org/abs/2510.13678 +184,"Diffusion LLMs Can Do Faster-Than-AR Inference via Discrete Diffusion + Forcing",30,10,28.57,29,"Aug 8, 2025",29,79,https://arxiv.org/abs/2508.09192 +185,Real-Time Object Detection Meets DINOv3,37,14,32.84,37,"Sep 25, 2025",10,684,https://arxiv.org/abs/2509.20787 +186,"OmniVinci: Enhancing Architecture and Data for Omni-Modal Understanding + LLM",28,7,27.36,24,"Oct 17, 2025",86,502,https://arxiv.org/abs/2510.15870 +187,"QeRL: Beyond Efficiency -- Quantization-enhanced Reinforcement Learning for LLMs",21,1,19.67,20,"Oct 13, 2025",162,350,https://arxiv.org/abs/2510.11696 -117,Latent Collaboration in Multi-Agent Systems,12,12,14.92,14,"Nov 25, 2025",92,240,https://arxiv.org/abs/2511.20639 -118,"Matrix-Game 2.0: An Open-Source, Real-Time, and Streaming Interactive - World Model",41,4,22.83,26,"Aug 18, 2025",22,1550,https://arxiv.org/abs/2508.13009 -119,TimeGPT-1,21,6,19.86,12,"Oct 5, 2023",7,3470,https://arxiv.org/abs/2310.03589 -120,"Beyond Outlining: Heterogeneous Recursive Planning for Adaptive - Long-form Writing with Language Models",13,9,16.15,12,"Mar 11, 2025",3,736,https://arxiv.org/abs/2503.08275 -121,RAG-Anything: All-in-One RAG Framework,129,2,25.02,28,"Oct 14, 2025",49,10600,https://arxiv.org/abs/2510.12323 -122,DeepAgent: A General Reasoning Agent with Scalable Toolsets,43,5,23.09,25,"Oct 24, 2025",93,739,https://arxiv.org/abs/2510.21618 -123,P3-SAM: Native 3D Part Segmentation,20,1,19.85,13,"Sep 8, 2025",21,249,https://arxiv.org/abs/2509.06784 -124,UniVerse-1: Unified Audio-Video Generation via Stitching of Experts,15,9,18.07,10,"Sep 7, 2025",14,53,https://arxiv.org/abs/2509.06155 -125,"FLUX-Reason-6M & PRISM-Bench: A Million-Scale Text-to-Image Reasoning - Dataset and Comprehensive Benchmark",19,3,19.74,15,"Sep 11, 2025",35,74,https://arxiv.org/abs/2509.09680 -126,"Echo-4o: Harnessing the Power of GPT-4o Synthetic Images for Improved - Image Generation",27,4,21.67,20,"Aug 13, 2025",23,73,https://arxiv.org/abs/2508.09987 -127,"From Pixels to Words -- Towards Native Vision-Language Primitives at - Scale",12,8,16.75,13,"Oct 16, 2025",60,161,https://arxiv.org/abs/2510.14979 -128,X-Part: high fidelity and structure coherent shape decomposition,17,12,19.53,14,"Sep 10, 2025",25,234,https://arxiv.org/abs/2509.08643 -129,"HoloCine: Holistic Generation of Cinematic Multi-Shot Long Video +188,TimeGPT-1,21,6,19.86,12,"Oct 5, 2023",7,3470,https://arxiv.org/abs/2310.03589 +189,"LucidFlux: Caption-Free Universal Image Restoration via a Large-Scale + Diffusion Transformer",37,24,33.35,32,"Sep 26, 2025",21,476,https://arxiv.org/abs/2509.22414 +190,Ark: An Open-source Python-based Framework for Robot Learning,25,10,25.04,18,"Jun 24, 2025",16,245,https://arxiv.org/abs/2506.21628 +191,Kwai Keye-VL Technical Report,40,20,34.85,33,"Jul 2, 2025",130,623,https://arxiv.org/abs/2507.01949 +192,MHR: Momentum Human Rig,24,4,25,28,"Nov 19, 2025",13,419,https://arxiv.org/abs/2511.15586 +193,P3-SAM: Native 3D Part Segmentation,20,1,19.85,13,"Sep 8, 2025",21,249,https://arxiv.org/abs/2509.06784 +194,"Spark-TTS: An Efficient LLM-Based Text-to-Speech Model with + Single-Stream Decoupled Speech Tokens",23,14,24.22,22,"Mar 3, 2025",6,10400,https://arxiv.org/abs/2503.01710 +195,Diffusion Language Models are Super Data Learners,25,16,26.36,19,"Nov 5, 2025",112,259,https://arxiv.org/abs/2511.03276 +196,"Hulu-Med: A Transparent Generalist Model towards Holistic Medical + Vision-Language Understanding",30,13,30.57,35,"Oct 9, 2025",4,478,https://arxiv.org/abs/2510.08668 +197,"HoloCine: Holistic Generation of Cinematic Multi-Shot Long Video Narratives",20,10,20.55,17,"Oct 23, 2025",37,333,https://arxiv.org/abs/2510.20822 -130,"Build Your Personalized Research Group: A Multiagent Framework for +198,Training-Free Group Relative Policy Optimization,34,25,33.18,32,"Oct 9, 2025",42,3590,https://arxiv.org/abs/2510.08191 +199,Visual Spatial Tuning,21,3,22.29,18,"Nov 7, 2025",46,124,https://arxiv.org/abs/2511.05491 +200,Step-Audio-R1 Technical Report,25,10,26.92,32,"Nov 19, 2025",51,316,https://arxiv.org/abs/2511.15848 +201,WizardCoder: Empowering Code Large Language Models with Evol-Instruct,12,1,1,1,"Jun 14, 2023",30,9470,https://arxiv.org/abs/2306.08568 +202,Dolphin: Document Image Parsing via Heterogeneous Anchor Prompting,18,8,17.78,15,"May 20, 2025",3,7670,https://arxiv.org/abs/2505.14059 +203,"FLUX-Reason-6M & PRISM-Bench: A Million-Scale Text-to-Image Reasoning + Dataset and Comprehensive Benchmark",19,3,19.74,15,"Sep 11, 2025",35,74,https://arxiv.org/abs/2509.09680 +204,MiMo-Embodied: X-Embodied Foundation Model Technical Report,23,9,25.26,28,"Nov 20, 2025",23,260,https://arxiv.org/abs/2511.16518 +205,Robot Learning: A Tutorial,20,6,21.5,15,"Oct 14, 2025",80,281,https://arxiv.org/abs/2510.12403 +206,GigaWorld-0: World Models as Data Engine to Empower Embodied AI,15,8,11.93,13,"Nov 25, 2025",28,214,https://arxiv.org/abs/2511.19861 +207,"Large Language Model Agent: A Survey on Methodology, Applications and + Challenges",82,35,43.88,45,"Mar 27, 2025",82,1710,https://arxiv.org/abs/2503.21460 +208,Chronos-2: From Univariate to Universal Forecasting,32,23,32.75,31,"Oct 17, 2025",17,4190,https://arxiv.org/abs/2510.15821 +209,ViSTA-SLAM: Visual SLAM with Symmetric Two-view Association,38,16,35.74,37,"Sep 1, 2025",6,134,https://arxiv.org/abs/2509.01584 +210,ReCode: Unify Plan and Action for Universal Granularity Control,29,16,31.76,30,"Oct 27, 2025",117,359,https://arxiv.org/abs/2510.23564 +211,"Skyfall-GS: Synthesizing Immersive 3D Urban Scenes from Satellite + Imagery",37,9,36.08,41,"Oct 17, 2025",45,576,https://arxiv.org/abs/2510.15869 +212,"OpenVision 2: A Family of Generative Pretrained Visual Encoders for + Multimodal Learning",27,11,30.74,30,"Sep 1, 2025",27,356,https://arxiv.org/abs/2509.01644 +213,4KAgent: Agentic Any Image to 4K Super-Resolution,21,19,25.19,22,"Jul 9, 2025",104,547,https://arxiv.org/abs/2507.07105 +214,"ChronoEdit: Towards Temporal Reasoning for Image Editing and World + Simulation",45,23,38.96,42,"Oct 5, 2025",14,578,https://arxiv.org/abs/2510.04290 +215,X-Part: high fidelity and structure coherent shape decomposition,17,12,19.53,14,"Sep 10, 2025",25,234,https://arxiv.org/abs/2509.08643 +216,"Huxley-Gödel Machine: Human-Level Coding Agent Development by an + Approximation of the Optimal Self-Improving Machine",21,10,25.71,21,"Oct 24, 2025",18,251,https://arxiv.org/abs/2510.21614 +217,"Build Your Personalized Research Group: A Multiagent Framework for Continual and Interactive Science Automation",17,2,19.82,24,"Oct 17, 2025",14,267,https://arxiv.org/abs/2510.15624 -131,"LightReasoner: Can Small Language Models Teach Large Language Models +218,"OpenTSLM: Time-Series Language Models for Reasoning over Multivariate + Medical Text- and Time-Series Data",21,8,26.43,27,"Oct 2, 2025",15,950,https://arxiv.org/abs/2510.02410 +219,Trace Anything: Representing Any Video in 4D via Trajectory Fields,17,11,21.18,22,"Oct 15, 2025",30,271,https://arxiv.org/abs/2510.13802 +220,Democratizing AI scientists using ToolUniverse,29,17,33.59,38,"Sep 27, 2025",38,455,https://arxiv.org/abs/2509.23426 +221,"S^2-Guidance: Stochastic Self Guidance for Training-Free Enhancement of + Diffusion Models",25,15,30.96,23,"Aug 18, 2025",45,134,https://arxiv.org/abs/2508.12880 +222,UniVerse-1: Unified Audio-Video Generation via Stitching of Experts,15,9,18.07,10,"Sep 7, 2025",14,53,https://arxiv.org/abs/2509.06155 +223,R-Zero: Self-Evolving Reasoning LLM from Zero Data,39,28,38.41,38,"Aug 7, 2025",123,559,https://arxiv.org/abs/2508.05004 +224,"On the Generalization of SFT: A Reinforcement Learning Perspective with + Reward Rectification",29,26,34.34,35,"Aug 7, 2025",148,322,https://arxiv.org/abs/2508.05629 +225,"InternScenes: A Large-scale Simulatable Indoor Scene Dataset with + Realistic Layouts",17,7,22.94,20,"Sep 13, 2025",29,161,https://arxiv.org/abs/2509.10813 +226,"LightReasoner: Can Small Language Models Teach Large Language Models Reasoning?",15,9,19.33,10,"Oct 9, 2025",7,362,https://arxiv.org/abs/2510.07962 -132,"CommonForms: A Large, Diverse Dataset for Form Field Detection",29,2,22.59,27,"Sep 20, 2025",18,819,https://arxiv.org/abs/2509.16506 -133,ST-Raptor: LLM-Powered Semi-Structured Table Question Answering,32,3,22.97,24,"Aug 25, 2025",6,243,https://arxiv.org/abs/2508.18190 -134,"UniGenBench++: A Unified Semantic Evaluation Benchmark for Text-to-Image - Generation",12,5,18.17,14,"Oct 21, 2025",64,100,https://arxiv.org/abs/2510.18701 -135,Robot Learning: A Tutorial,20,6,21.5,15,"Oct 14, 2025",80,281,https://arxiv.org/abs/2510.12403 -136,Rolling Forcing: Autoregressive Long Video Diffusion in Real Time,15,3,20.13,18,"Sep 29, 2025",20,109,https://arxiv.org/abs/2509.25161 -137,Yume: An Interactive World Generation Model,4,3,5.25,4,"Jul 23, 2025",69,185,https://arxiv.org/abs/2507.17744 -138,Trace Anything: Representing Any Video in 4D via Trajectory Fields,17,11,21.18,22,"Oct 15, 2025",30,271,https://arxiv.org/abs/2510.13802 -139,Matrix-3D: Omnidirectional Explorable 3D World Generation,45,6,24.27,23,"Aug 11, 2025",67,398,https://arxiv.org/abs/2508.08086 -140,Visual Spatial Tuning,21,3,22.29,18,"Nov 7, 2025",46,124,https://arxiv.org/abs/2511.05491 -141,MMaDA-Parallel: Multimodal Large Diffusion Language Models for Thinking-Aware Editing and Generation,11,7,19.09,21,"Nov 12, 2025",64,248,https://arxiv.org/abs/2511.09611 -142,Tree Search for LLM Agent Reinforcement Learning,15,13,21.2,19,"Sep 25, 2025",79,95,https://arxiv.org/abs/2509.21240 -143,"AudioStory: Generating Long-Form Narrative Audio with Large Language - Models",59,16,24.81,21,"Aug 27, 2025",20,264,https://arxiv.org/abs/2508.20088 -144,"VideoCanvas: Unified Video Completion from Arbitrary Spatiotemporal - Patches via In-Context Conditioning",7,2,16.43,16,"Oct 9, 2025",59,50,https://arxiv.org/abs/2510.08555 -145,"Collaborating Action by Action: A Multi-agent LLM Framework for Embodied - Reasoning",26,10,23.54,21,"Apr 24, 2025",5,4250,https://arxiv.org/abs/2504.17950 -146,FlowRL: Matching Reward Distributions for LLM Reasoning,14,13,21.5,17,"Sep 18, 2025",90,58,https://arxiv.org/abs/2509.15207 -147,RLFR: Extending Reinforcement Learning for LLMs with Flow Environment,7,5,17,14,"Oct 11, 2025",32,34,https://arxiv.org/abs/2510.10201 -148,"Uniworld-V2: Reinforce Image Editing with Diffusion Negative-aware +227,The Denario project: Deep knowledge AI agents for scientific discovery,22,15,29.73,26,"Oct 30, 2025",6,372,https://arxiv.org/abs/2510.26887 +228,"R-4B: Incentivizing General-Purpose Auto-Thinking Capability in MLLMs + via Bi-Mode Annealing and Reinforce Learning",19,21,26.63,23,"Aug 28, 2025",102,84,https://arxiv.org/abs/2508.21113 +229,Rolling Forcing: Autoregressive Long Video Diffusion in Real Time,15,3,20.13,18,"Sep 29, 2025",20,109,https://arxiv.org/abs/2509.25161 +230,Robot Learning from a Physical World Model,23,23,31.09,30,"Nov 10, 2025",26,146,https://arxiv.org/abs/2511.07416 +231,"Genie Envisioner: A Unified World Foundation Platform for Robotic + Manipulation",24,19,31.96,29,"Aug 7, 2025",72,218,https://arxiv.org/abs/2508.05635 +232,"RLVE: Scaling Up Reinforcement Learning for Language Models with + Adaptive Verifiable Environments",20,18,28.15,23,"Nov 10, 2025",10,119,https://arxiv.org/abs/2511.07317 +233,"Packing Input Frame Context in Next-Frame Prediction Models for Video + Generation",47,15,41.34,44,"Apr 17, 2025",52,15800,https://arxiv.org/abs/2504.12626 +234,"M^3FinMeeting: A Multilingual, Multi-Sector, and Multi-Task Financial + Meeting Understanding Evaluation Dataset",22,12,30.41,28,"Jun 3, 2025",3,287,https://arxiv.org/abs/2506.02510 +235,"Beyond Outlining: Heterogeneous Recursive Planning for Adaptive + Long-form Writing with Language Models",13,9,16.15,12,"Mar 11, 2025",3,736,https://arxiv.org/abs/2503.08275 +236,"Skywork UniPic: Unified Autoregressive Modeling for Visual Understanding + and Generation",25,18,33,31,"Aug 5, 2025",59,727,https://arxiv.org/abs/2508.03320 +237,Tree Search for LLM Agent Reinforcement Learning,15,13,21.2,19,"Sep 25, 2025",79,95,https://arxiv.org/abs/2509.21240 +238,VoXtream: Full-Stream Text-to-Speech with Extremely Low Latency,24,24,32.42,31,"Sep 19, 2025",2,129,https://arxiv.org/abs/2509.15969 +239,A Survey of Data Agents: Emerging Paradigm or Overstated Hype?,17,11,24.82,19,"Oct 27, 2025",63,202,https://arxiv.org/abs/2510.23587 +240,SteadyDancer: Harmonized and Coherent Human Image Animation with First-Frame Preservation,15,17,22.07,21,"Nov 24, 2025",38,189,https://arxiv.org/abs/2511.19320 +241,Latent Collaboration in Multi-Agent Systems,12,12,14.92,14,"Nov 25, 2025",92,240,https://arxiv.org/abs/2511.20639 +242,Puppeteer: Rig and Animate Your 3D Models,31,27,37.1,37,"Aug 14, 2025",30,178,https://arxiv.org/abs/2508.10898 +243,GEM: A Gym for Agentic LLMs,21,24,31.29,30,"Oct 1, 2025",79,284,https://arxiv.org/abs/2510.01051 +244,MedRAX: Medical Reasoning Agent for Chest X-ray,18,9,28,29,"Feb 4, 2025",2,993,https://arxiv.org/abs/2502.02673 +245,"Omni-Effects: Unified and Spatially-Controllable Visual Effects + Generation",17,14,26.71,28,"Aug 11, 2025",56,122,https://arxiv.org/abs/2508.07981 +246,FlowRL: Matching Reward Distributions for LLM Reasoning,14,13,21.5,17,"Sep 18, 2025",90,58,https://arxiv.org/abs/2509.15207 +247,"Uniworld-V2: Reinforce Image Editing with Diffusion Negative-aware Finetuning and MLLM Implicit Feedback",14,3,21.5,21,"Oct 19, 2025",18,94,https://arxiv.org/abs/2510.16888 -149,WebDancer: Towards Autonomous Information Seeking Agency,398,3,25.85,26,"May 28, 2025",33,17300,https://arxiv.org/abs/2505.22648 -150,Continuous Autoregressive Language Models,46,5,24.72,25,"Oct 31, 2025",64,582,https://arxiv.org/abs/2510.27688 -151,SteadyDancer: Harmonized and Coherent Human Image Animation with First-Frame Preservation,15,17,22.07,21,"Nov 24, 2025",38,189,https://arxiv.org/abs/2511.19320 -152,WebWatcher: Breaking New Frontier of Vision-Language Deep Research Agent,393,1,25.86,26,"Aug 7, 2025",139,17300,https://arxiv.org/abs/2508.05748 -153,TTT3R: 3D Reconstruction as Test-Time Training,29,3,24.1,28,"Sep 30, 2025",13,396,https://arxiv.org/abs/2509.26645 -154,"Evolution Strategies at Scale: LLM Fine-Tuning Beyond Reinforcement - Learning",10,8,20.5,19,"Sep 29, 2025",4,164,https://arxiv.org/abs/2509.24372 -155,ObjectClear: Complete Object Removal via Object-Effect Attention,4,8,13,11,"May 28, 2025",1,328,https://arxiv.org/abs/2505.22636 -156,"MCP-Bench: Benchmarking Tool-Using LLM Agents with Complex Real-World - Tasks via MCP Servers",69,5,25.25,23,"Aug 28, 2025",58,296,https://arxiv.org/abs/2508.20453 -157,"InternScenes: A Large-scale Simulatable Indoor Scene Dataset with - Realistic Layouts",17,7,22.94,20,"Sep 13, 2025",29,161,https://arxiv.org/abs/2509.10813 -158,Ovi: Twin Backbone Cross-Modal Fusion for Audio-Video Generation,68,2,25.31,22,"Sep 30, 2025",32,1300,https://arxiv.org/abs/2510.01284 -159,"Beyond Ten Turns: Unlocking Long-Horizon Agentic Search with Large-Scale - Asynchronous RL",47,3,25.02,19,"Aug 11, 2025",48,373,https://arxiv.org/abs/2508.07976 -160,"Mixture-of-Recursions: Learning Dynamic Recursive Depths for Adaptive - Token-Level Computation",4,11,15,13,"Jul 14, 2025",60,311,https://arxiv.org/abs/2507.10524 -161,"Spark-TTS: An Efficient LLM-Based Text-to-Speech Model with - Single-Stream Decoupled Speech Tokens",23,14,24.22,22,"Mar 3, 2025",6,10400,https://arxiv.org/abs/2503.01710 -162,"VCode: a Multimodal Coding Benchmark with SVG as Symbolic Visual - Representation",11,5,22.45,18,"Nov 4, 2025",95,88,https://arxiv.org/abs/2511.02778 -163,NaTex: Seamless Texture Generation as Latent Color Diffusion,9,10,21.89,21,"Nov 20, 2025",15,85,https://arxiv.org/abs/2511.16317 -164,"The Well: a Large-Scale Collection of Diverse Physics Simulations for +248,"UMO: Scaling Multi-Identity Consistency for Image Customization via + Matching Reward",19,18,29.37,26,"Sep 8, 2025",29,131,https://arxiv.org/abs/2509.06818 +249,"From Pixels to Words -- Towards Native Vision-Language Primitives at + Scale",12,8,16.75,13,"Oct 16, 2025",60,161,https://arxiv.org/abs/2510.14979 +250,LIMI: Less is More for Agency,26,28,35.31,35,"Sep 22, 2025",91,101,https://arxiv.org/abs/2509.17567 +251,"TaDiCodec: Text-aware Diffusion Speech Tokenizer for Speech Language + Modeling",18,12,28.44,31,"Aug 22, 2025",7,149,https://arxiv.org/abs/2508.16790 +252,SceneGen: Single-Image 3D Scene Generation in One Feedforward Pass,23,26,33.39,31,"Aug 21, 2025",18,66,https://arxiv.org/abs/2508.15769 +253,The Markovian Thinker,19,15,29.95,29,"Oct 8, 2025",27,251,https://arxiv.org/abs/2510.06557 +254,Time-to-Move: Training-Free Motion Controlled Video Generation via Dual-Clock Denoising,21,22,32.05,30,"Nov 9, 2025",50,191,https://arxiv.org/abs/2511.08633 +255,V-Thinker: Interactive Thinking with Images,16,5,26.31,30,"Nov 6, 2025",92,104,https://arxiv.org/abs/2511.04460 +256,"UniGenBench++: A Unified Semantic Evaluation Benchmark for Text-to-Image + Generation",12,5,18.17,14,"Oct 21, 2025",64,100,https://arxiv.org/abs/2510.18701 +257,"MiniCPM-V 4.5: Cooking Efficient MLLMs via Architecture, Data, and + Training Recipe",21,17,32.43,31,"Sep 16, 2025",45,22000,https://arxiv.org/abs/2509.18154 +258,WorldGrow: Generating Infinite 3D World,17,12,28.12,25,"Oct 24, 2025",40,313,https://arxiv.org/abs/2510.21682 +259,Video-As-Prompt: Unified Semantic Control for Video Generation,16,11,26.88,22,"Oct 23, 2025",41,217,https://arxiv.org/abs/2510.20888 +260,"SongBloom: Coherent Song Generation via Interleaved Autoregressive + Sketching and Diffusion Refinement",21,23,32.86,31,"Jun 9, 2025",6,604,https://arxiv.org/abs/2506.07634 +261,"FireRedTTS-2: Towards Long Conversational Speech Generation for Podcast + and Chatbot",24,21,35.21,35,"Sep 2, 2025",1,1180,https://arxiv.org/abs/2509.02020 +262,LTX-Video: Realtime Video Latent Diffusion,18,19,30.17,30,"Dec 30, 2024",47,8590,https://arxiv.org/abs/2501.00103 +263,OpenMMReasoner: Pushing the Frontiers for Multimodal Reasoning with an Open and General Recipe,14,6,24.29,23,"Nov 20, 2025",86,106,https://arxiv.org/abs/2511.16334 +264,SciReasoner: Laying the Scientific Reasoning Ground Across Disciplines,13,12,23.23,18,"Sep 25, 2025",86,43,https://arxiv.org/abs/2509.21320 +265,"Uni-MoE-2.0-Omni: Scaling Language-Centric Omnimodal Large Model with Advanced MoE, Training and Data",14,15,25.36,27,"Nov 16, 2025",99,1010,https://arxiv.org/abs/2511.12609 +266,"Advancing End-to-End Pixel Space Generative Modeling via Self-supervised + Pre-training",14,4,25.43,28,"Oct 14, 2025",104,98,https://arxiv.org/abs/2510.12586 +267,"IGGT: Instance-Grounded Geometry Transformer for Semantic 3D + Reconstruction",15,16,27.2,23,"Oct 26, 2025",38,156,https://arxiv.org/abs/2510.22706 +268,"CODA: Coordinating the Cerebrum and Cerebellum for a Dual-Brain Computer + Use Agent with Decoupled Reinforcement Learning",15,5,27.47,30,"Aug 27, 2025",31,25,https://arxiv.org/abs/2508.20096 +269,"Verbalized Sampling: How to Mitigate Mode Collapse and Unlock LLM + Diversity",20,26,33.45,34,"Oct 1, 2025",15,219,https://arxiv.org/abs/2510.01171 +270,MMaDA-Parallel: Multimodal Large Diffusion Language Models for Thinking-Aware Editing and Generation,11,7,19.09,21,"Nov 12, 2025",64,248,https://arxiv.org/abs/2511.09611 +271,Kandinsky 5.0: A Family of Foundation Models for Image and Video Generation,14,13,25.93,22,"Nov 19, 2025",207,471,https://arxiv.org/abs/2511.14993 +272,MemOS: A Memory OS for AI System,36,13,41.33,45,"Jul 4, 2025",155,3110,https://arxiv.org/abs/2507.03724 +273,DA^2: Depth Anything in Any Direction,17,8,30.53,37,"Sep 30, 2025",20,120,https://arxiv.org/abs/2509.26618 +274,"OneReward: Unified Mask-Guided Image Generation via Multi-Task Human + Preference Learning",15,11,28,30,"Aug 28, 2025",13,208,https://arxiv.org/abs/2508.21066 +275,"Paper2Code: Automating Code Generation from Scientific Papers in Machine + Learning",32,27,40.34,40,"Apr 24, 2025",120,3730,https://arxiv.org/abs/2504.17192 +276,"ThinkMorph: Emergent Properties in Multimodal Interleaved + Chain-of-Thought Reasoning",17,9,31,28,"Oct 30, 2025",78,87,https://arxiv.org/abs/2510.27492 +277,"The Well: a Large-Scale Collection of Diverse Physics Simulations for Machine Learning",12,13,23,14,"Nov 30, 2024",20,1110,https://arxiv.org/abs/2412.00568 -165,SciReasoner: Laying the Scientific Reasoning Ground Across Disciplines,13,12,23.23,18,"Sep 25, 2025",86,43,https://arxiv.org/abs/2509.21320 -166,"Elevating 3D Models: High-Quality Texture and Geometry Refinement from a - Low-Quality Model",4,9,17.25,15,"Jul 15, 2025",11,106,https://arxiv.org/abs/2507.11465 -167,"StealthAttack: Robust 3D Gaussian Splatting Poisoning via Density-Guided +278,Versatile Framework for Song Generation with Prompt-based Control,44,34,43.5,45,"Apr 27, 2025",6,201,https://arxiv.org/abs/2504.19062 +279,Durian: Dual Reference-guided Portrait Animation with Attribute Transfer,15,16,29.13,22,"Sep 4, 2025",4,14,https://arxiv.org/abs/2509.04434 +280,Self-Forcing++: Towards Minute-Scale High-Quality Video Generation,19,22,34,30,"Oct 2, 2025",86,145,https://arxiv.org/abs/2510.02283 +281,OmniSVG: A Unified Scalable Vector Graphics Generation Model,16,16,30.88,28,"Apr 8, 2025",180,2100,https://arxiv.org/abs/2504.06263 +282,"SimpleTIR: End-to-End Reinforcement Learning for Multi-Turn + Tool-Integrated Reasoning",22,19,36.55,43,"Sep 2, 2025",76,250,https://arxiv.org/abs/2509.02479 +283,VGGT-X: When VGGT Meets Dense Novel View Synthesis,13,15,26.69,25,"Sep 29, 2025",15,69,https://arxiv.org/abs/2509.25191 +284,"VCode: a Multimodal Coding Benchmark with SVG as Symbolic Visual + Representation",11,5,22.45,18,"Nov 4, 2025",95,88,https://arxiv.org/abs/2511.02778 +285,"Thinking with Camera: A Unified Multimodal Model for Camera-Centric + Understanding and Generation",13,4,26.92,29,"Oct 9, 2025",115,200,https://arxiv.org/abs/2510.08673 +286,Latent Diffusion Model without Variational Autoencoder,14,13,28.71,24,"Oct 17, 2025",39,156,https://arxiv.org/abs/2510.15301 +287,"Evolution Strategies at Scale: LLM Fine-Tuning Beyond Reinforcement + Learning",10,8,20.5,19,"Sep 29, 2025",4,164,https://arxiv.org/abs/2509.24372 +288,"VideoFrom3D: 3D Scene Video Generation via Complementary Image and Video + Diffusion Models",18,18,34.39,36,"Sep 22, 2025",25,83,https://arxiv.org/abs/2509.17985 +289,"VolSplat: Rethinking Feed-Forward 3D Gaussian Splatting with + Voxel-Aligned Prediction",14,16,29.86,24,"Sep 23, 2025",22,65,https://arxiv.org/abs/2509.19297 +290,MolmoAct: Action Reasoning Models that can Reason in Space,22,28,37.68,37,"Aug 11, 2025",42,163,https://arxiv.org/abs/2508.07917 +291,"Ming-UniAudio: Speech LLM for Joint Understanding, Generation and + Editing with Unified Representation",14,9,30.21,28,"Oct 26, 2025",9,370,https://arxiv.org/abs/2511.05516 +292,MAPO: Mixed Advantage Policy Optimization,12,12,27.08,23,"Sep 23, 2025",25,34,https://arxiv.org/abs/2509.18849 +293,DepthLM: Metric Depth From Vision Language Models,18,26,35.06,32,"Sep 29, 2025",5,130,https://arxiv.org/abs/2509.25413 +294,"REINFORCE++: A Simple and Efficient Approach for Aligning Large Language + Models",41,26,44.05,45,"Jan 4, 2025",102,8130,https://arxiv.org/abs/2501.03262 +295,DeepPHY: Benchmarking Agentic VLMs on Physical Reasoning,26,25,40.15,42,"Aug 7, 2025",62,143,https://arxiv.org/abs/2508.05405 +296,"Lumina-DiMOO: An Omni Diffusion Large Language Model for Multi-Modal + Generation and Understanding",13,18,29.38,30,"Oct 7, 2025",48,834,https://arxiv.org/abs/2510.06308 +297,DR Tulu: Reinforcement Learning with Evolving Rubrics for Deep Research,16,24,33.62,29,"Nov 24, 2025",48,404,https://arxiv.org/abs/2511.19399 +298,"IMAGGarment-1: Fine-Grained Garment Generation for Controllable Fashion + Design",14,21,31.64,31,"Apr 17, 2025",1,263,https://arxiv.org/abs/2504.13176 +299,Interleaving Reasoning for Better Text-to-Image Generation,11,16,26.64,17,"Sep 8, 2025",12,24,https://arxiv.org/abs/2509.06945 +300,ARE: Scaling Up Agent Environments and Evaluations,17,24,35.24,34,"Sep 21, 2025",29,257,https://arxiv.org/abs/2509.17158 +301,GeoVista: Web-Augmented Agentic Visual Reasoning for Geolocalization,11,9,27.09,28,"Nov 19, 2025",88,170,https://arxiv.org/abs/2511.15705 +302,"Scrub It Out! Erasing Sensitive Memorization in Code Language Models via + Machine Unlearning",10,16,24.8,19,"Sep 17, 2025",18,24,https://arxiv.org/abs/2509.13755 +303,NaTex: Seamless Texture Generation as Latent Color Diffusion,9,10,21.89,21,"Nov 20, 2025",15,85,https://arxiv.org/abs/2511.16317 +304,"StealthAttack: Robust 3D Gaussian Splatting Poisoning via Density-Guided Illusions",9,2,22.11,23,"Oct 2, 2025",55,51,https://arxiv.org/abs/2510.02314 -168,Human-Agent Collaborative Paper-to-Page Crafting for Under $0.1,9,13,22.56,19,"Oct 22, 2025",64,98,https://arxiv.org/abs/2510.19600 -169,"VMem: Consistent Interactive Video Scene Generation with Surfel-Indexed - View Memory",4,18,19,18,"Jun 23, 2025",22,255,https://arxiv.org/abs/2506.18903 -170,"SeC: Advancing Complex Video Object Segmentation via Progressive Concept - Construction",3,12,16.67,19,"Jul 21, 2025",34,98,https://arxiv.org/abs/2507.15852 -171,"InternSVG: Towards Unified SVG Tasks with Multimodal Large Language - Models",6,10,21.83,25,"Oct 13, 2025",31,54,https://arxiv.org/abs/2510.11341 -172,LongCat-Video Technical Report,40,6,25.38,28,"Oct 25, 2025",24,1110,https://arxiv.org/abs/2510.22200 -173,Ark: An Open-source Python-based Framework for Robot Learning,25,10,25.04,18,"Jun 24, 2025",16,245,https://arxiv.org/abs/2506.21628 -174,MHR: Momentum Human Rig,24,4,25,28,"Nov 19, 2025",13,419,https://arxiv.org/abs/2511.15586 -175,OpenMMReasoner: Pushing the Frontiers for Multimodal Reasoning with an Open and General Recipe,14,6,24.29,23,"Nov 20, 2025",86,106,https://arxiv.org/abs/2511.16334 -176,PPTAgent: Generating and Evaluating Presentations Beyond Text-to-Slides,2,15,15,15,"Jan 7, 2025",22,2430,https://arxiv.org/abs/2501.03936 -177,"NaViL: Rethinking Scaling Properties of Native Multimodal Large Language - Models under Data Constraints",8,19,23.38,23,"Oct 9, 2025",17,71,https://arxiv.org/abs/2510.08565 -178,A Survey of Data Agents: Emerging Paradigm or Overstated Hype?,17,11,24.82,19,"Oct 27, 2025",63,202,https://arxiv.org/abs/2510.23587 -179,"T2I-ReasonBench: Benchmarking Reasoning-Informed Text-to-Image - Generation",8,11,23.75,17,"Aug 24, 2025",25,24,https://arxiv.org/abs/2508.17472 -180,4KAgent: Agentic Any Image to 4K Super-Resolution,21,19,25.19,22,"Jul 9, 2025",104,547,https://arxiv.org/abs/2507.07105 -181,TiViBench: Benchmarking Think-in-Video Reasoning for Video Generative Models,5,17,22.6,23,"Nov 17, 2025",40,50,https://arxiv.org/abs/2511.13704 -182,MiMo-Embodied: X-Embodied Foundation Model Technical Report,23,9,25.26,28,"Nov 20, 2025",23,260,https://arxiv.org/abs/2511.16518 -183,CLEAR: Error Analysis via LLM-as-a-Judge Made Easy,1,11,11,11,"Jul 24, 2025",8,10,https://arxiv.org/abs/2507.18392 -184,"EditScore: Unlocking Online RL for Image Editing via High-Fidelity +305,Human-Agent Collaborative Paper-to-Page Crafting for Under $0.1,9,13,22.56,19,"Oct 22, 2025",64,98,https://arxiv.org/abs/2510.19600 +306,"F1: A Vision-Language-Action Model Bridging Understanding and Generation + to Actions",10,15,25.8,23,"Sep 8, 2025",26,64,https://arxiv.org/abs/2509.06951 +307,ROSE: Remove Objects with Side Effects in Videos,12,14,30.67,33,"Aug 26, 2025",4,32,https://arxiv.org/abs/2508.18633 +308,"VideoCanvas: Unified Video Completion from Arbitrary Spatiotemporal + Patches via In-Context Conditioning",7,2,16.43,16,"Oct 9, 2025",59,50,https://arxiv.org/abs/2510.08555 +309,"EditScore: Unlocking Online RL for Image Editing via High-Fidelity Reward Modeling",9,16,24.33,21,"Sep 28, 2025",26,60,https://arxiv.org/abs/2509.23909 -185,Reinforced Visual Perception with Tools,9,18,24.44,22,"Sep 1, 2025",27,28,https://arxiv.org/abs/2509.01656 -186,"Diffuman4D: 4D Consistent Human View Synthesis from Sparse-View Videos - with Spatio-Temporal Diffusion Models",4,14,22.75,25,"Jul 17, 2025",50,246,https://arxiv.org/abs/2507.13344 -187,PICABench: How Far Are We from Physically Realistic Image Editing?,3,12,21.67,15,"Oct 20, 2025",57,13,https://arxiv.org/abs/2510.17681 -188,"BAPO: Stabilizing Off-Policy Reinforcement Learning for LLMs via +310,Reinforced Visual Perception with Tools,9,18,24.44,22,"Sep 1, 2025",27,28,https://arxiv.org/abs/2509.01656 +311,Is Diversity All You Need for Scalable Robotic Manipulation?,18,30,37.78,37,"Jul 8, 2025",20,2460,https://arxiv.org/abs/2507.06219 +312,RLFR: Extending Reinforcement Learning for LLMs with Flow Environment,7,5,17,14,"Oct 11, 2025",32,34,https://arxiv.org/abs/2510.10201 +313,"BAPO: Stabilizing Off-Policy Reinforcement Learning for LLMs via Balanced Policy Optimization with Adaptive Clipping",9,15,24.56,23,"Oct 21, 2025",77,62,https://arxiv.org/abs/2510.18927 -189,OpenCUA: Open Foundations for Computer-Use Agents,85,14,25.86,24,"Aug 12, 2025",30,415,https://arxiv.org/abs/2508.09123 -190,"Scrub It Out! Erasing Sensitive Memorization in Code Language Models via - Machine Unlearning",10,16,24.8,19,"Sep 17, 2025",18,24,https://arxiv.org/abs/2509.13755 -191,"Latent Zoning Network: A Unified Principle for Generative Modeling, +314,"GenoMAS: A Multi-Agent Framework for Scientific Discovery via + Code-Driven Gene Expression Analysis",11,25,29.45,26,"Jul 28, 2025",3,93,https://arxiv.org/abs/2507.21035 +315,"Latent Zoning Network: A Unified Principle for Generative Modeling, Representation Learning, and Classification",9,12,24.89,19,"Sep 19, 2025",43,39,https://arxiv.org/abs/2509.15591 -192,CrossOver: 3D Scene Cross-Modal Alignment,8,4,24.75,22,"Feb 20, 2025",2,204,https://arxiv.org/abs/2502.15011 -193,"Uni-MoE-2.0-Omni: Scaling Language-Centric Omnimodal Large Model with Advanced MoE, Training and Data",14,15,25.36,27,"Nov 16, 2025",99,1010,https://arxiv.org/abs/2511.12609 -194,SimpleVLA-RL: Scaling VLA Training via Reinforcement Learning,56,1,25.86,28,"Sep 11, 2025",73,740,https://arxiv.org/abs/2509.09674 -195,"Advancing End-to-End Pixel Space Generative Modeling via Self-supervised - Pre-training",14,4,25.43,28,"Oct 14, 2025",104,98,https://arxiv.org/abs/2510.12586 -196,Fast-dLLM v2: Efficient Block-Diffusion LLM,7,7,25,22,"Sep 30, 2025",40,537,https://arxiv.org/abs/2509.26328 -197,FullPart: Generating each 3D Part at Full Resolution,8,15,25.12,22,"Oct 30, 2025",4,57,https://arxiv.org/abs/2510.26140 -198,"ComoRAG: A Cognitive-Inspired Memory-Organized RAG for Stateful Long - Narrative Reasoning",44,8,25.86,28,"Aug 14, 2025",70,192,https://arxiv.org/abs/2508.10419 -199,"Beyond Pass@1: Self-Play with Variational Problem Synthesis Sustains - RLVR",8,14,25.25,21,"Aug 19, 2025",109,26,https://arxiv.org/abs/2508.14029 -200,"MUG-V 10B: High-efficiency Training Pipeline for Large Video Generation - Models",3,18,24,24,"Oct 20, 2025",9,59,https://arxiv.org/abs/2510.17519 -201,"Huxley-Gödel Machine: Human-Level Coding Agent Development by an - Approximation of the Optimal Self-Improving Machine",21,10,25.71,21,"Oct 24, 2025",18,251,https://arxiv.org/abs/2510.21614 -202,π^3: Scalable Permutation-Equivariant Visual Geometry Learning,6,10,25.17,17,"Jul 17, 2025",64,1100,https://arxiv.org/abs/2507.13347 -203,AutoPR: Let's Automate Your Academic Promotion!,5,8,25.2,26,"Oct 10, 2025",43,36,https://arxiv.org/abs/2510.09558 -204,"CMPhysBench: A Benchmark for Evaluating Large Language Models in - Condensed Matter Physics",5,13,25.6,19,"Aug 25, 2025",45,16,https://arxiv.org/abs/2508.18124 -205,"F1: A Vision-Language-Action Model Bridging Understanding and Generation - to Actions",10,15,25.8,23,"Sep 8, 2025",26,64,https://arxiv.org/abs/2509.06951 -206,Part-X-MLLM: Part-aware 3D Multimodal Large Language Model,3,25,25.33,25,"Nov 17, 2025",65,49,https://arxiv.org/abs/2511.13647 -207,Scaling RL to Long Videos,3,21,25.67,28,"Jul 10, 2025",141,538,https://arxiv.org/abs/2507.07966 -208,Streaming 4D Visual Geometry Transformer,1,25,25,25,"Jul 15, 2025",10,454,https://arxiv.org/abs/2507.11539 -209,Reinforcement Learning Foundations for Deep Research Systems: A Survey,8,15,25.88,24,"Sep 8, 2025",25,10,https://arxiv.org/abs/2509.06733 -210,Kandinsky 5.0: A Family of Foundation Models for Image and Video Generation,14,13,25.93,22,"Nov 19, 2025",207,471,https://arxiv.org/abs/2511.14993 -211,"Promptomatix: An Automatic Prompt Optimization Framework for Large - Language Models",3,19,26.33,30,"Jul 17, 2025",13,61,https://arxiv.org/abs/2507.14241 -212,"SearchInstruct: Enhancing Domain Adaptation via Retrieval-Based - Instruction Dataset Creation",5,18,26.4,26,"Sep 12, 2025",9,8,https://arxiv.org/abs/2509.10708 -213,"DetailFlow: 1D Coarse-to-Fine Autoregressive Image Generation via - Next-Detail Prediction",1,29,29,29,"May 27, 2025",16,144,https://arxiv.org/abs/2505.21473 -214,Parallel-R1: Towards Parallel Thinking via Reinforcement Learning,41,5,26.07,27,"Sep 9, 2025",95,151,https://arxiv.org/abs/2509.07980 -215,Human3R: Everyone Everywhere All at Once,9,22,26.44,26,"Oct 7, 2025",8,304,https://arxiv.org/abs/2510.06219 -216,OBS-Diff: Accurate Pruning For Diffusion Models in One-Shot,2,26,28.5,28,"Oct 8, 2025",14,24,https://arxiv.org/abs/2510.06751 -217,V-Thinker: Interactive Thinking with Images,16,5,26.31,30,"Nov 6, 2025",92,104,https://arxiv.org/abs/2511.04460 -218,GUI-G^2: Gaussian Reward Modeling for GUI Grounding,2,24,29,29,"Jul 21, 2025",118,138,https://arxiv.org/abs/2507.15846 -219,UnSAMv2: Self-Supervised Learning Enables Segment Anything at Any Granularity,3,27,28,28,"Nov 17, 2025",4,34,https://arxiv.org/abs/2511.13714 -220,Interleaving Reasoning for Better Text-to-Image Generation,11,16,26.64,17,"Sep 8, 2025",12,24,https://arxiv.org/abs/2509.06945 -221,TUN3D: Towards Real-World Scene Understanding from Unposed Images,1,33,33,33,"Sep 23, 2025",12,11,https://arxiv.org/abs/2509.21388 -222,Drax: Speech Recognition with Discrete Flow Matching,4,21,27.75,21,"Oct 5, 2025",23,18,https://arxiv.org/abs/2510.04162 -223,Zebra-CoT: A Dataset for Interleaved Vision Language Reasoning,1,34,34,34,"Jul 22, 2025",28,42,https://arxiv.org/abs/2507.16746 -224,Retrieval-Augmented Generation with Hierarchical Knowledge,1,34,34,34,"Mar 13, 2025",2,359,https://arxiv.org/abs/2503.10150 -225,"REASONING GYM: Reasoning Environments for Reinforcement Learning with - Verifiable Rewards",1,34,34,34,"May 30, 2025",72,1130,https://arxiv.org/abs/2505.24760 -226,Hierarchical Budget Policy Optimization for Adaptive Reasoning,1,35,35,35,"Jul 21, 2025",12,14,https://arxiv.org/abs/2507.15844 -227,VGGT-X: When VGGT Meets Dense Novel View Synthesis,13,15,26.69,25,"Sep 29, 2025",15,69,https://arxiv.org/abs/2509.25191 -228,"OpenTSLM: Time-Series Language Models for Reasoning over Multivariate - Medical Text- and Time-Series Data",21,8,26.43,27,"Oct 2, 2025",15,950,https://arxiv.org/abs/2510.02410 -229,Diffusion Language Models are Super Data Learners,25,16,26.36,19,"Nov 5, 2025",112,259,https://arxiv.org/abs/2511.03276 -230,WebSailor: Navigating Super-human Reasoning for Web Agent,394,2,26.03,26,"Jul 3, 2025",122,17300,https://arxiv.org/abs/2507.02592 -231,MMaDA: Multimodal Large Diffusion Language Models,1,37,37,37,"May 21, 2025",94,1230,https://arxiv.org/abs/2505.15809 -232,nablaNABLA: Neighborhood Adaptive Block-Level Attention,1,38,38,38,"Jul 17, 2025",85,10,https://arxiv.org/abs/2507.13546 -233,"Omni-Effects: Unified and Spatially-Controllable Visual Effects - Generation",17,14,26.71,28,"Aug 11, 2025",56,122,https://arxiv.org/abs/2508.07981 -234,"R-4B: Incentivizing General-Purpose Auto-Thinking Capability in MLLMs - via Bi-Mode Annealing and Reinforce Learning",19,21,26.63,23,"Aug 28, 2025",102,84,https://arxiv.org/abs/2508.21113 -235,"Thinking with Camera: A Unified Multimodal Model for Camera-Centric - Understanding and Generation",13,4,26.92,29,"Oct 9, 2025",115,200,https://arxiv.org/abs/2510.08673 -236,SpaceVista: All-Scale Visual Spatial Reasoning from mm to km,2,30,32,32,"Oct 10, 2025",16,22,https://arxiv.org/abs/2510.09606 -237,GeoVista: Web-Augmented Agentic Visual Reasoning for Geolocalization,11,9,27.09,28,"Nov 19, 2025",88,170,https://arxiv.org/abs/2511.15705 -238,Drag-and-Drop LLMs: Zero-Shot Prompt-to-Weights,1,39,39,39,"Jun 19, 2025",122,58,https://arxiv.org/abs/2506.16406 -239,MAPO: Mixed Advantage Policy Optimization,12,12,27.08,23,"Sep 23, 2025",25,34,https://arxiv.org/abs/2509.18849 -240,Logics-Parsing Technical Report,48,4,26.27,28,"Sep 24, 2025",7,619,https://arxiv.org/abs/2509.19760 -241,"VoiceAssistant-Eval: Benchmarking AI Assistants across Listening, - Speaking, and Viewing",1,39,39,39,"Sep 26, 2025",19,4,https://arxiv.org/abs/2509.22651 -242,Ovis2.5 Technical Report,39,6,26.36,27,"Aug 15, 2025",102,1280,https://arxiv.org/abs/2508.11737 -243,Scaling Language-Centric Omnimodal Representation Learning,1,40,40,40,"Oct 13, 2025",52,11,https://arxiv.org/abs/2510.11693 -244,NANO3D: A Training-Free Approach for Efficient 3D Editing Without Masks,5,22,28.8,25,"Oct 16, 2025",52,47,https://arxiv.org/abs/2510.15019 -245,Video-As-Prompt: Unified Semantic Control for Video Generation,16,11,26.88,22,"Oct 23, 2025",41,217,https://arxiv.org/abs/2510.20888 -246,Sparser Block-Sparse Attention via Token Permutation,3,22,31,26,"Oct 24, 2025",22,19,https://arxiv.org/abs/2510.21270 -247,OceanGym: A Benchmark Environment for Underwater Embodied Agents,5,22,29.2,28,"Sep 30, 2025",30,28,https://arxiv.org/abs/2509.26536 -248,Rethinking Reward Models for Multi-Domain Test-Time Scaling,3,28,31.33,33,"Oct 1, 2025",23,14,https://arxiv.org/abs/2510.00492 -249,HoloScene: Simulation-Ready Interactive 3D Worlds from a Single Video,1,42,42,42,"Oct 7, 2025",5,5,https://arxiv.org/abs/2510.05560 -250,"Grasp Any Region: Towards Precise, Contextual Pixel Understanding for +316,"Grasp Any Region: Towards Precise, Contextual Pixel Understanding for Multimodal LLMs",10,13,27.6,27,"Oct 21, 2025",33,57,https://arxiv.org/abs/2510.18876 -251,Mantis: A Versatile Vision-Language-Action Model with Disentangled Visual Foresight,2,32,34,34,"Nov 20, 2025",10,18,https://arxiv.org/abs/2511.16175 -252,"Learning on the Job: Test-Time Curricula for Targeted Reinforcement - Learning",1,43,43,43,"Oct 6, 2025",1,3,https://arxiv.org/abs/2510.04786 -253,"Discrete Diffusion Models with MLLMs for Unified Medical Multimodal - Generation",1,43,43,43,"Oct 7, 2025",4,4,https://arxiv.org/abs/2510.06131 -254,NORA-1.5: A Vision-Language-Action Model Trained using World Model- and Action-based Preference Rewards,3,31,31.67,32,"Nov 18, 2025",7,26,https://arxiv.org/abs/2511.14659 -255,"Scaling Text-Rich Image Understanding via Code-Guided Synthetic - Multimodal Data Generation",2,35,35,35,"Feb 20, 2025",13,103,https://arxiv.org/abs/2502.14846 -256,"StreamDiffusion: A Pipeline-level Solution for Real-time Interactive - Generation",2,34,35,35,"Dec 19, 2023",73,10400,https://arxiv.org/abs/2312.12491 -257,"Graph2Eval: Automatic Multimodal Task Generation for Agents via - Knowledge Graphs",1,44,44,44,"Oct 1, 2025",1,5,https://arxiv.org/abs/2510.00507 -258,"IGGT: Instance-Grounded Geometry Transformer for Semantic 3D - Reconstruction",15,16,27.2,23,"Oct 26, 2025",38,156,https://arxiv.org/abs/2510.22706 -259,DeepPrune: Parallel Scaling without Inter-trace Redundancy,1,45,45,45,"Oct 9, 2025",21,9,https://arxiv.org/abs/2510.08483 -260,ConsistEdit: Highly Consistent and Precise Training-free Visual Editing,2,35,35.5,35,"Oct 20, 2025",11,25,https://arxiv.org/abs/2510.17803 -261,"MiroThinker: Pushing the Performance Boundaries of Open-Source Research Agents via Model, Context, and Interactive Scaling",29,4,26.66,30,"Nov 14, 2025",156,1160,https://arxiv.org/abs/2511.11793 -262,WorldVLA: Towards Autoregressive Action World Model,1,45,45,45,"Jun 26, 2025",40,669,https://arxiv.org/abs/2506.21539 -263,"Gaze into the Heart: A Multi-View Video Dataset for rPPG and Health - Biomarkers Estimation",1,46,46,46,"Aug 25, 2025",11,3,https://arxiv.org/abs/2508.17924 -264,Uncertainty-Aware Testing-Time Optimization for 3D Human Pose Estimation,1,46,46,46,"Feb 4, 2024",1,5,https://arxiv.org/abs/2402.02339 -265,"VLA-RFT: Vision-Language-Action Reinforcement Fine-tuning with Verified - Rewards in World Simulators",5,20,30,25,"Oct 1, 2025",55,28,https://arxiv.org/abs/2510.00406 -266,"AssetOpsBench: Benchmarking AI Agents for Task Automation in Industrial - Asset Operations and Maintenance",1,46,46,46,"Jun 4, 2025",14,252,https://arxiv.org/abs/2506.03828 -267,Multi-Agent Tool-Integrated Policy Optimization,2,35,36,36,"Oct 6, 2025",19,20,https://arxiv.org/abs/2510.04678 -268,"Entropy Regularizing Activation: Boosting Continuous Control, Large - Language Models, and Image Classification with Activation as Entropy - Constraints",1,46,46,46,"Oct 9, 2025",5,7,https://arxiv.org/abs/2510.08549 -269,RegionE: Adaptive Region-Aware Generation for Efficient Image Editing,9,14,28.22,24,"Oct 29, 2025",24,46,https://arxiv.org/abs/2510.25590 -270,"UniPixel: Unified Object Referring and Segmentation for Pixel-Level - Visual Reasoning",1,47,47,47,"Sep 22, 2025",3,40,https://arxiv.org/abs/2509.18094 -271,"GeoSVR: Taming Sparse Voxels for Geometrically Accurate Surface - Reconstruction",1,47,47,47,"Sep 22, 2025",2,65,https://arxiv.org/abs/2509.18090 -272,Visual Jigsaw Post-Training Improves MLLMs,8,20,28.62,26,"Sep 29, 2025",34,29,https://arxiv.org/abs/2509.25190 -273,PIPer: On-Device Environment Setup via Online Reinforcement Learning,1,47,47,47,"Sep 29, 2025",23,4,https://arxiv.org/abs/2509.25455 -274,"Factuality Matters: When Image Generation and Editing Meet Structured - Visuals",1,47,47,47,"Oct 6, 2025",12,4,https://arxiv.org/abs/2510.05091 -275,"BIRD-INTERACT: Re-imagining Text-to-SQL Evaluation for Large Language - Models via Lens of Dynamic Interactions",1,47,47,47,"Oct 6, 2025",4,260,https://arxiv.org/abs/2510.05318 -276,Instant4D: 4D Gaussian Splatting in Minutes,1,47,47,47,"Oct 1, 2025",5,58,https://arxiv.org/abs/2510.01119 -277,"CODA: Coordinating the Cerebrum and Cerebellum for a Dual-Brain Computer - Use Agent with Decoupled Reinforcement Learning",15,5,27.47,30,"Aug 27, 2025",31,25,https://arxiv.org/abs/2508.20096 -278,GenExam: A Multidisciplinary Text-to-Image Exam,8,17,28.75,28,"Sep 17, 2025",16,17,https://arxiv.org/abs/2509.14232 -279,Where LLM Agents Fail and How They can Learn From Failures,1,48,48,48,"Sep 29, 2025",11,8,https://arxiv.org/abs/2509.25370 -280,IMG: Calibrating Diffusion Models via Implicit Multimodal Guidance,7,16,29.14,21,"Sep 30, 2025",16,25,https://arxiv.org/abs/2509.26231 -281,"Video-LMM Post-Training: A Deep Dive into Video Reasoning with Large +317,"FantasyTalking2: Timestep-Layer Adaptive Preference Optimization for + Audio-Driven Portrait Animation",13,22,33.15,31,"Aug 15, 2025",8,21,https://arxiv.org/abs/2508.11255 +318,"OpenVision: A Fully-Open, Cost-Effective Family of Advanced Vision + Encoders for Multimodal Learning",17,26,37.47,39,"May 7, 2025",28,351,https://arxiv.org/abs/2505.04601 +319,"Video-LMM Post-Training: A Deep Dive into Video Reasoning with Large Multimodal Models",10,10,28.2,27,"Oct 6, 2025",42,89,https://arxiv.org/abs/2510.05034 -282,"MarS: a Financial Market Simulation Engine Powered by Generative - Foundation Model",1,48,48,48,"Sep 4, 2024",1,1560,https://arxiv.org/abs/2409.07486 -283,SpatialLM: Training Large Language Models for Structured Indoor Modeling,50,4,26.46,31,"Jun 9, 2025",49,3900,https://arxiv.org/abs/2506.07491 -284,"Audio-visual Controlled Video Diffusion with Masked Selective State - Spaces Modeling for Natural Talking Head Generation",1,49,49,49,"Apr 3, 2025",49,344,https://arxiv.org/abs/2504.02542 -285,"Implicit Actor Critic Coupling via a Supervised Learning Framework for - RLVR",4,28,31.75,32,"Sep 2, 2025",19,18,https://arxiv.org/abs/2509.02522 -286,"SynParaSpeech: Automated Synthesis of Paralinguistic Datasets for Speech - Generation and Understanding",1,49,49,49,"Sep 18, 2025",1,37,https://arxiv.org/abs/2509.14946 -287,"Beyond Log Likelihood: Probability-Based Objectives for Supervised - Fine-Tuning across the Model Capability Continuum",1,49,49,49,"Oct 1, 2025",7,3,https://arxiv.org/abs/2510.00526 -288,Spotlight on Token Perception for Multimodal Reinforcement Learning,6,10,29.83,27,"Oct 10, 2025",31,26,https://arxiv.org/abs/2510.09285 -289,A Survey of Vibe Coding with Large Language Models,1,49,49,49,"Oct 14, 2025",21,7,https://arxiv.org/abs/2510.12399 -290,Step-Audio-R1 Technical Report,25,10,26.92,32,"Nov 19, 2025",51,316,https://arxiv.org/abs/2511.15848 -291,Video-as-Answer: Predict and Generate Next Video Event with Joint-GRPO,6,23,29.83,29,"Nov 20, 2025",29,44,https://arxiv.org/abs/2511.16669 -292,Gaussian Splatting with Discretized SDF for Relightable Assets,1,50,50,50,"Jul 21, 2025",19,61,https://arxiv.org/abs/2507.15629 -293,Flow-GRPO: Training Flow Matching Models via Online RL,1,50,50,50,"May 8, 2025",83,1120,https://arxiv.org/abs/2505.05470 -294,One-Minute Video Generation with Test-Time Training,1,50,50,50,"Apr 7, 2025",109,2090,https://arxiv.org/abs/2504.05298 -295,"SyGra: A Unified Graph-Based Framework for Scalable Generation, Quality - Tagging, and Management of Synthetic Data",6,16,30,28,"Aug 21, 2025",6,16,https://arxiv.org/abs/2508.15432 -296,LongCat-Flash-Thinking Technical Report,1,50,50,50,"Sep 23, 2025",2,195,https://arxiv.org/abs/2509.18883 -297,"MesaTask: Towards Task-Driven Tabletop Scene Generation via 3D Spatial - Reasoning",2,36,38,38,"Sep 26, 2025",27,24,https://arxiv.org/abs/2509.22281 -298,"Efficient Audio-Visual Speech Separation with Discrete Lip Semantics and - Multi-Scale Global-Local Attention",1,50,50,50,"Sep 28, 2025",13,22,https://arxiv.org/abs/2509.23610 -299,EVODiff: Entropy-aware Variance Optimized Diffusion Inference,1,50,50,50,"Sep 30, 2025",1,9,https://arxiv.org/abs/2509.26096 -300,"GLiNER2: An Efficient Multi-Task Information Extraction System with - Schema-Driven Interface",1,50,50,50,"Jul 24, 2025",28,186,https://arxiv.org/abs/2507.18546 -301,"iMontage: Unified, Versatile, Highly Dynamic Many-to-many Image Generation",2,34,38,38,"Nov 25, 2025",30,120,https://arxiv.org/abs/2511.20635 -302,"ThinkSound: Chain-of-Thought Reasoning in Multimodal Large Language - Models for Audio Generation and Editing",2,38,38.5,38,"Jun 26, 2025",7,876,https://arxiv.org/abs/2506.21448 -303,"Q-Sched: Pushing the Boundaries of Few-Step Diffusion Models with - Quantization-Aware Scheduling",2,30,38.5,38,"Sep 1, 2025",5,6,https://arxiv.org/abs/2509.01624 -304,"CapRL: Stimulating Dense Image Caption Capabilities via Reinforcement - Learning",8,11,29.12,28,"Sep 26, 2025",30,62,https://arxiv.org/abs/2509.22647 -305,SPARK: Synergistic Policy And Reward Co-Evolving Framework,3,24,34.33,30,"Sep 26, 2025",15,17,https://arxiv.org/abs/2509.22624 -306,"GAS: Improving Discretization of Diffusion ODEs via Generalized - Adversarial Solver",2,30,38.5,38,"Oct 20, 2025",2,10,https://arxiv.org/abs/2510.17699 -307,SAC: Neural Speech Codec with Semantic-Acoustic Dual-Stream Quantization,7,16,29.57,31,"Oct 19, 2025",0,60,https://arxiv.org/abs/2510.16841 -308,"Video-Thinker: Sparking ""Thinking with Videos"" via Reinforcement +320,"AWorld: Dynamic Multi-Agent System with Stable Maneuvering for Robust + GAIA Problem Solving",31,34,43.71,45,"Aug 13, 2025",32,694,https://arxiv.org/abs/2508.09889 +321,"Video-Thinker: Sparking ""Thinking with Videos"" via Reinforcement Learning",10,13,28.5,22,"Oct 27, 2025",81,95,https://arxiv.org/abs/2510.23473 -309,"RewardMap: Tackling Sparse Rewards in Fine-grained Visual Reasoning via - Multi-Stage Reinforcement Learning",5,18,31.2,29,"Oct 2, 2025",15,22,https://arxiv.org/abs/2510.02240 -310,"D2E: Scaling Vision-Action Pretraining on Desktop Data for Transfer to - Embodied AI",2,39,39.5,39,"Oct 7, 2025",101,22,https://arxiv.org/abs/2510.05684 -311,SpatialTrackerV2: 3D Point Tracking Made Easy,3,32,35.33,32,"Jul 16, 2025",14,658,https://arxiv.org/abs/2507.12462 -312,Reasoning with Sampling: Your Base Model is Smarter Than You Think,3,27,35.33,33,"Oct 16, 2025",35,229,https://arxiv.org/abs/2510.14901 -313,"VITA-E: Natural Embodied Interaction with Concurrent Seeing, Hearing, - Speaking, and Acting",8,20,29.5,25,"Oct 21, 2025",41,120,https://arxiv.org/abs/2510.21817 -314,"Efficient Multi-modal Large Language Models via Progressive Consistency - Distillation",4,23,33.25,33,"Oct 1, 2025",37,13,https://arxiv.org/abs/2510.00515 -315,"OneReward: Unified Mask-Guided Image Generation via Multi-Task Human - Preference Learning",15,11,28,30,"Aug 28, 2025",13,208,https://arxiv.org/abs/2508.21066 -316,"Efficient Multi-turn RL for GUI Agents via Decoupled Training and - Adaptive Data Curation",3,32,36,35,"Sep 28, 2025",7,9,https://arxiv.org/abs/2509.23866 -317,Artificial Hippocampus Networks for Efficient Long-Context Modeling,6,24,31,28,"Oct 8, 2025",22,67,https://arxiv.org/abs/2510.07318 -318,"E^2Rank: Your Text Embedding can Also be an Effective - and Efficient Listwise Reranker",5,29,32,33,"Oct 26, 2025",29,22,https://arxiv.org/abs/2510.22733 -319,RoboOmni: Proactive Robot Manipulation in Omni-modal Context,2,40,41,41,"Oct 27, 2025",52,27,https://arxiv.org/abs/2510.23763 -320,"THOR: Tool-Integrated Hierarchical Optimization via RL for Mathematical +322,Cook and Clean Together: Teaching Embodied Agents for Parallel Task Execution,11,19,30.64,28,"Nov 24, 2025",7,270,https://arxiv.org/abs/2511.19430 +323,"BEAVR: Bimanual, multi-Embodiment, Accessible, Virtual Reality + Teleoperation System for Robots",11,19,30.82,31,"Aug 13, 2025",0,50,https://arxiv.org/abs/2508.09606 +324,"NaViL: Rethinking Scaling Properties of Native Multimodal Large Language + Models under Data Constraints",8,19,23.38,23,"Oct 9, 2025",17,71,https://arxiv.org/abs/2510.08565 +325,Human3R: Everyone Everywhere All at Once,9,22,26.44,26,"Oct 7, 2025",8,304,https://arxiv.org/abs/2510.06219 +326,"BRIDGE - Building Reinforcement-Learning Depth-to-Image Data Generation + Engine for Monocular Depth Estimation",17,33,38.06,37,"Sep 29, 2025",13,106,https://arxiv.org/abs/2509.25077 +327,"THOR: Tool-Integrated Hierarchical Optimization via RL for Mathematical Reasoning",10,16,29.1,28,"Sep 17, 2025",11,17,https://arxiv.org/abs/2509.13761 -321,Quantile Advantage Estimation for Entropy-Safe Reasoning,2,35,41.5,41,"Sep 26, 2025",100,8,https://arxiv.org/abs/2509.22611 -322,MATRIX: Mask Track Alignment for Interaction-aware Video Generation,2,37,41.5,41,"Oct 8, 2025",29,22,https://arxiv.org/abs/2510.07310 -323,UltraFlux: Data-Model Co-Design for High-quality Native 4K Text-to-Image Generation across Diverse Aspect Ratios,2,36,42,42,"Nov 22, 2025",34,42,https://arxiv.org/abs/2511.18050 -324,"From reactive to cognitive: brain-inspired spatial intelligence for - embodied agents",2,42,42.5,42,"Aug 24, 2025",3,18,https://arxiv.org/abs/2508.17198 -325,Sequential Diffusion Language Models,7,22,30.71,29,"Sep 28, 2025",29,29,https://arxiv.org/abs/2509.24007 -326,"Euclid's Gift: Enhancing Spatial Perception and Reasoning in - Vision-Language Models via Geometric Surrogate Tasks",4,29,34.25,31,"Sep 29, 2025",15,13,https://arxiv.org/abs/2509.24473 -327,GIR-Bench: Versatile Benchmark for Generating Images with Reasoning,3,29,37,39,"Oct 13, 2025",16,23,https://arxiv.org/abs/2510.11026 -328,Chem-R: Learning to Reason as a Chemist,2,35,42.5,42,"Oct 19, 2025",45,9,https://arxiv.org/abs/2510.16880 -329,A Style is Worth One Code: Unlocking Code-to-Style Image Generation with Discrete Style Space,9,19,29.67,33,"Nov 13, 2025",53,122,https://arxiv.org/abs/2511.10555 -330,"SwiftEdit: Lightning Fast Text-Guided Image Editing via One-Step - Diffusion",3,36,37.33,36,"Dec 5, 2024",41,7,https://arxiv.org/abs/2412.04301 -331,Variational Reasoning for Language Models,2,36,43,43,"Sep 26, 2025",57,34,https://arxiv.org/abs/2509.22637 -332,SPATIALGEN: Layout-guided 3D Indoor Scene Generation,10,15,29.5,30,"Sep 18, 2025",22,255,https://arxiv.org/abs/2509.14981 -333,"URSA: Understanding and Verifying Chain-of-thought Reasoning in - Multimodal Mathematics",6,26,31.83,32,"Jan 8, 2025",53,117,https://arxiv.org/abs/2501.04686 -334,Continuous Thought Machines,2,42,43.5,43,"May 8, 2025",12,1490,https://arxiv.org/abs/2505.05522 -335,G^2VLM: Geometry Grounded Vision Language Model with Unified 3D Reconstruction and Spatial Reasoning,10,22,29.5,27,"Nov 26, 2025",8,151,https://arxiv.org/abs/2511.21688 -336,"DrugReasoner: Interpretable Drug Approval Prediction with a - Reasoning-augmented Language Model",2,44,44,44,"Aug 26, 2025",10,2,https://arxiv.org/abs/2508.18579 -337,TikZero: Zero-Shot Text-Guided Graphics Program Synthesis,2,44,44,44,"Mar 14, 2025",3,1530,https://arxiv.org/abs/2503.11509 -338,VideoNSA: Native Sparse Attention Scales Video Understanding,2,42,44,44,"Oct 2, 2025",8,29,https://arxiv.org/abs/2510.02295 -339,OmniNWM: Omniscient Driving Navigation World Models,2,43,44,44,"Oct 21, 2025",6,48,https://arxiv.org/abs/2510.18313 -340,WorldGrow: Generating Infinite 3D World,17,12,28.12,25,"Oct 24, 2025",40,313,https://arxiv.org/abs/2510.21682 -341,MedRAX: Medical Reasoning Agent for Chest X-ray,18,9,28,29,"Feb 4, 2025",2,993,https://arxiv.org/abs/2502.02673 -342,EVTAR: End-to-End Try on with Additional Unpaired Visual Reference,2,43,44,44,"Nov 2, 2025",4,17,https://arxiv.org/abs/2511.00956 -343,P1: Mastering Physics Olympiads with Reinforcement Learning,2,43,44,44,"Nov 17, 2025",106,45,https://arxiv.org/abs/2511.13612 -344,YuE: Scaling Open Foundation Models for Long-Form Music Generation,2,40,44.5,44,"Mar 11, 2025",69,5520,https://arxiv.org/abs/2503.08638 -345,"GenoMAS: A Multi-Agent Framework for Scientific Discovery via - Code-Driven Gene Expression Analysis",11,25,29.45,26,"Jul 28, 2025",3,93,https://arxiv.org/abs/2507.21035 -346,Vlaser: Vision-Language-Action Model with Synergistic Embodied Reasoning,3,30,38.67,41,"Oct 13, 2025",14,15,https://arxiv.org/abs/2510.11027 -347,"OmniVinci: Enhancing Architecture and Data for Omni-Modal Understanding - LLM",28,7,27.36,24,"Oct 17, 2025",86,502,https://arxiv.org/abs/2510.15870 -348,Latent Diffusion Model without Variational Autoencoder,14,13,28.71,24,"Oct 17, 2025",39,156,https://arxiv.org/abs/2510.15301 -349,"RAPO++: Cross-Stage Prompt Optimization for Text-to-Video Generation via - Data Alignment and Test-Time Scaling",2,40,45,45,"Oct 23, 2025",11,104,https://arxiv.org/abs/2510.20206 -350,Vidi: Large Multimodal Models for Video Understanding and Editing,4,27,35.5,37,"Apr 22, 2025",14,247,https://arxiv.org/abs/2504.15681 -351,"ImagerySearch: Adaptive Test-Time Search for Video Generation Beyond - Semantic Dependency Constraints",2,43,45.5,45,"Oct 16, 2025",47,47,https://arxiv.org/abs/2510.14847 -352,"MeshSplat: Generalizable Sparse-View Surface Reconstruction via Gaussian - Splatting",6,22,32.67,28,"Aug 25, 2025",4,18,https://arxiv.org/abs/2508.17811 -353,"Automatic Synthetic Data and Fine-grained Adaptive Feature Alignment for - Composed Person Retrieval",2,46,46,46,"Nov 25, 2023",1,49,https://arxiv.org/abs/2311.16515 -354,Character Mixing for Video Generation,2,42,46,46,"Oct 6, 2025",5,52,https://arxiv.org/abs/2510.05093 -355,"U-Bench: A Comprehensive Understanding of U-Net through 100-Variant +328,"T2I-ReasonBench: Benchmarking Reasoning-Informed Text-to-Image + Generation",8,11,23.75,17,"Aug 24, 2025",25,24,https://arxiv.org/abs/2508.17472 +329,SPATIALGEN: Layout-guided 3D Indoor Scene Generation,10,15,29.5,30,"Sep 18, 2025",22,255,https://arxiv.org/abs/2509.14981 +330,G^2VLM: Geometry Grounded Vision Language Model with Unified 3D Reconstruction and Spatial Reasoning,10,22,29.5,27,"Nov 26, 2025",8,151,https://arxiv.org/abs/2511.21688 +331,"CogVLA: Cognition-Aligned Vision-Language-Action Model via + Instruction-Driven Routing & Sparsification",18,31,39.22,39,"Aug 28, 2025",8,56,https://arxiv.org/abs/2508.21046 +332,DyPE: Dynamic Position Extrapolation for Ultra High Resolution Diffusion,14,26,35.86,31,"Oct 23, 2025",33,248,https://arxiv.org/abs/2510.20766 +333,Hunyuan-MT Technical Report,14,26,36,35,"Sep 5, 2025",13,517,https://arxiv.org/abs/2509.05209 +334,CrossOver: 3D Scene Cross-Modal Alignment,8,4,24.75,22,"Feb 20, 2025",2,204,https://arxiv.org/abs/2502.15011 +335,FullPart: Generating each 3D Part at Full Resolution,8,15,25.12,22,"Oct 30, 2025",4,57,https://arxiv.org/abs/2510.26140 +336,"Beyond Pass@1: Self-Play with Variational Problem Synthesis Sustains + RLVR",8,14,25.25,21,"Aug 19, 2025",109,26,https://arxiv.org/abs/2508.14029 +337,RegionE: Adaptive Region-Aware Generation for Efficient Image Editing,9,14,28.22,24,"Oct 29, 2025",24,46,https://arxiv.org/abs/2510.25590 +338,pi-Flow: Policy-Based Few-Step Generation via Imitation Distillation,13,26,35.38,33,"Oct 16, 2025",7,97,https://arxiv.org/abs/2510.14974 +339,Reinforcement Learning Foundations for Deep Research Systems: A Survey,8,15,25.88,24,"Sep 8, 2025",25,10,https://arxiv.org/abs/2509.06733 +340,Efficient Part-level 3D Object Generation via Dual Volume Packing,13,24,35.62,36,"Jun 11, 2025",8,701,https://arxiv.org/abs/2506.09980 +341,3D Gaussian Splatting for Real-Time Radiance Field Rendering,41,41,46.12,46,"Aug 8, 2023",192,19600,https://arxiv.org/abs/2308.04079 +342,A Style is Worth One Code: Unlocking Code-to-Style Image Generation with Discrete Style Space,9,19,29.67,33,"Nov 13, 2025",53,122,https://arxiv.org/abs/2511.10555 +343,"A Vision-Language-Action-Critic Model for Robotic Real-World + Reinforcement Learning",11,21,33.64,33,"Sep 19, 2025",16,145,https://arxiv.org/abs/2509.15937 +344,"LivePortrait: Efficient Portrait Animation with Stitching and + Retargeting Control",29,39,44.48,44,"Jul 3, 2024",3,16900,https://arxiv.org/abs/2407.03168 +345,"MMR1: Enhancing Multimodal Reasoning with Variance-Aware Sampling and + Open Resources",11,24,33.91,31,"Sep 25, 2025",90,190,https://arxiv.org/abs/2509.21268 +346,"Time Is a Feature: Exploiting Temporal Dynamics in Diffusion Language + Models",13,25,36.62,34,"Aug 12, 2025",31,38,https://arxiv.org/abs/2508.09138 +347,Yume: An Interactive World Generation Model,4,3,5.25,4,"Jul 23, 2025",69,185,https://arxiv.org/abs/2507.17744 +348,Fast-dLLM v2: Efficient Block-Diffusion LLM,7,7,25,22,"Sep 30, 2025",40,537,https://arxiv.org/abs/2509.26328 +349,"DeepScientist: Advancing Frontier-Pushing Scientific Findings + Progressively",11,26,34.64,32,"Sep 30, 2025",16,119,https://arxiv.org/abs/2509.26603 +350,GigaBrain-0: A World Model-Powered Vision-Language-Action Model,13,30,37.15,36,"Oct 22, 2025",46,218,https://arxiv.org/abs/2510.19430 +351,Visual Jigsaw Post-Training Improves MLLMs,8,20,28.62,26,"Sep 29, 2025",34,29,https://arxiv.org/abs/2509.25190 +352,GenExam: A Multidisciplinary Text-to-Image Exam,8,17,28.75,28,"Sep 17, 2025",16,17,https://arxiv.org/abs/2509.14232 +353,"CapRL: Stimulating Dense Image Caption Capabilities via Reinforcement + Learning",8,11,29.12,28,"Sep 26, 2025",30,62,https://arxiv.org/abs/2509.22647 +354,"InternSVG: Towards Unified SVG Tasks with Multimodal Large Language + Models",6,10,21.83,25,"Oct 13, 2025",31,54,https://arxiv.org/abs/2510.11341 +355,"UniMoE-Audio: Unified Speech and Music Generation with Dynamic-Capacity + MoE",12,32,36.5,35,"Oct 15, 2025",62,1010,https://arxiv.org/abs/2510.13344 +356,"VITA-E: Natural Embodied Interaction with Concurrent Seeing, Hearing, + Speaking, and Acting",8,20,29.5,25,"Oct 21, 2025",41,120,https://arxiv.org/abs/2510.21817 +357,EdgeTAM: On-Device Track Anything Model,16,31,40.38,40,"Jan 13, 2025",1,757,https://arxiv.org/abs/2501.07256 +358,"Spatial-SSRL: Enhancing Spatial Understanding via Self-Supervised + Reinforcement Learning",11,25,36.09,34,"Oct 31, 2025",25,54,https://arxiv.org/abs/2510.27606 +359,Self-Rewarding Vision-Language Model via Reasoning Decomposition,16,33,40.88,42,"Aug 27, 2025",77,79,https://arxiv.org/abs/2508.19652 +360,Uni-MoE: Scaling Unified Multimodal LLMs with Mixture of Experts,11,32,36.27,35,"May 18, 2024",19,1010,https://arxiv.org/abs/2405.11273 +361,"U-Bench: A Comprehensive Understanding of U-Net through 100-Variant Benchmarking",8,21,31,28,"Oct 8, 2025",3,65,https://arxiv.org/abs/2510.07041 -356,"TIR-Bench: A Comprehensive Benchmark for Agentic Thinking-with-Images - Reasoning",2,46,46,46,"Nov 3, 2025",12,14,https://arxiv.org/abs/2511.01833 -357,"FlashAdventure: A Benchmark for GUI Agents Solving Full Story Arcs in - Diverse Adventure Games",2,44,46.5,46,"Sep 1, 2025",8,7,https://arxiv.org/abs/2509.01052 -358,"SIU3R: Simultaneous Scene Understanding and 3D Reconstruction Beyond - Feature Alignment",2,46,46.5,46,"Jul 3, 2025",2,80,https://arxiv.org/abs/2507.02705 -359,"Building a Foundational Guardrail for General Agentic Systems via - Synthetic Data",3,33,39.67,42,"Oct 10, 2025",21,26,https://arxiv.org/abs/2510.09781 -360,"Patch-as-Decodable-Token: Towards Unified Multi-Modal Vision Tasks in +362,VChain: Chain-of-Visual-Thought for Reasoning in Video Generation,9,25,33.33,29,"Oct 6, 2025",34,60,https://arxiv.org/abs/2510.05094 +363,"EasySteer: A Unified Framework for High-Performance and Extensible LLM + Steering",9,23,33.56,30,"Sep 29, 2025",25,49,https://arxiv.org/abs/2509.25175 +364,LongCodeZip: Compress Long Context for Code Language Models,8,23,31.38,30,"Oct 1, 2025",88,63,https://arxiv.org/abs/2510.00446 +365,"ARTDECO: Towards Efficient and High-Fidelity On-the-Fly 3D + Reconstruction with Structured Scene Representation",9,25,33.56,32,"Oct 9, 2025",30,73,https://arxiv.org/abs/2510.08551 +366,DeCo: Frequency-Decoupled Pixel Diffusion for End-to-End Image Generation,10,25,35.4,33,"Nov 24, 2025",58,75,https://arxiv.org/abs/2511.19365 +367,π^3: Scalable Permutation-Equivariant Visual Geometry Learning,6,10,25.17,17,"Jul 17, 2025",64,1100,https://arxiv.org/abs/2507.13347 +368,LoopTool: Closing the Data-Training Loop for Robust LLM Tool Calls,9,21,33.78,39,"Nov 12, 2025",15,20,https://arxiv.org/abs/2511.09148 +369,Explain Before You Answer: A Survey on Compositional Visual Reasoning,19,35,42.89,44,"Aug 24, 2025",4,279,https://arxiv.org/abs/2508.17298 +370,IMG: Calibrating Diffusion Models via Implicit Multimodal Guidance,7,16,29.14,21,"Sep 30, 2025",16,25,https://arxiv.org/abs/2509.26231 +371,ObjectClear: Complete Object Removal via Object-Effect Attention,4,8,13,11,"May 28, 2025",1,328,https://arxiv.org/abs/2505.22636 +372,"LLaVA-OneVision-1.5: Fully Open Framework for Democratized Multimodal + Training",16,33,41.5,41,"Sep 28, 2025",40,508,https://arxiv.org/abs/2509.23661 +373,"Stable Video Infinity: Infinite-Length Video Generation with Error + Recycling",19,34,43.05,45,"Oct 10, 2025",14,388,https://arxiv.org/abs/2510.09212 +374,SAC: Neural Speech Codec with Semantic-Acoustic Dual-Stream Quantization,7,16,29.57,31,"Oct 19, 2025",0,60,https://arxiv.org/abs/2510.16841 +375,ReasonRank: Empowering Passage Ranking with Strong Reasoning Ability,13,28,39.54,40,"Aug 9, 2025",109,101,https://arxiv.org/abs/2508.07050 +376,DiT360: High-Fidelity Panoramic Image Generation via Hybrid Training,9,21,34.56,38,"Oct 13, 2025",29,105,https://arxiv.org/abs/2510.11712 +377,AnyUp: Universal Feature Upsampling,11,31,37.73,37,"Oct 14, 2025",10,268,https://arxiv.org/abs/2510.12764 +378,"Mixture-of-Recursions: Learning Dynamic Recursive Depths for Adaptive + Token-Level Computation",4,11,15,13,"Jul 14, 2025",60,311,https://arxiv.org/abs/2507.10524 +379,Performance Prediction for Large Systems via Text-to-Text Regression,15,29,41.47,45,"Jun 26, 2025",6,255,https://arxiv.org/abs/2506.21718 +380,Sequential Diffusion Language Models,7,22,30.71,29,"Sep 28, 2025",29,29,https://arxiv.org/abs/2509.24007 +381,TiViBench: Benchmarking Think-in-Video Reasoning for Video Generative Models,5,17,22.6,23,"Nov 17, 2025",40,50,https://arxiv.org/abs/2511.13704 +382,Gradio: Hassle-Free Sharing and Testing of ML Models in the Wild,10,32,36.9,35,"Jun 6, 2019",1,40700,https://arxiv.org/abs/1906.02569 +383,"Elevating 3D Models: High-Quality Texture and Geometry Refinement from a + Low-Quality Model",4,9,17.25,15,"Jul 15, 2025",11,106,https://arxiv.org/abs/2507.11465 +384,"Patch-as-Decodable-Token: Towards Unified Multi-Modal Vision Tasks in MLLMs",7,23,31.86,29,"Oct 2, 2025",12,183,https://arxiv.org/abs/2510.01954 -361,AU-Harness: An Open-Source Toolkit for Holistic Evaluation of Audio LLMs,4,30,36.5,33,"Sep 9, 2025",18,35,https://arxiv.org/abs/2509.08031 -362,"How Far are VLMs from Visual Spatial Intelligence? A Benchmark-Driven - Perspective",3,37,40,37,"Sep 23, 2025",22,10,https://arxiv.org/abs/2509.18905 -363,"CoIRL-AD: Collaborative-Competitive Imitation-Reinforcement Learning in - Latent World Models for Autonomous Driving",3,36,40,42,"Oct 14, 2025",4,10,https://arxiv.org/abs/2510.12560 -364,First Frame Is the Place to Go for Video Content Customization,2,45,47,47,"Nov 19, 2025",45,36,https://arxiv.org/abs/2511.15700 -365,Thinking-while-Generating: Interleaving Textual Reasoning throughout Visual Generation,2,44,47,47,"Nov 20, 2025",12,43,https://arxiv.org/abs/2511.16671 -366,Computer-Use Agents as Judges for Generative User Interface,2,45,47,47,"Nov 19, 2025",48,24,https://arxiv.org/abs/2511.15567 -367,"GSM8K-V: Can Vision Language Models Solve Grade School Math Word +385,"More Thought, Less Accuracy? On the Dual Nature of Reasoning in + Vision-Language Models",9,26,36.33,35,"Sep 30, 2025",56,44,https://arxiv.org/abs/2509.25848 +386,"Open-o3 Video: Grounded Video Reasoning with Explicit Spatio-Temporal + Evidence",9,22,36.44,34,"Oct 23, 2025",46,63,https://arxiv.org/abs/2510.20579 +387,AutoPR: Let's Automate Your Academic Promotion!,5,8,25.2,26,"Oct 10, 2025",43,36,https://arxiv.org/abs/2510.09558 +388,"VMem: Consistent Interactive Video Scene Generation with Surfel-Indexed + View Memory",4,18,19,18,"Jun 23, 2025",22,255,https://arxiv.org/abs/2506.18903 +389,RLP: Reinforcement as a Pretraining Objective,9,25,36.78,34,"Sep 26, 2025",32,149,https://arxiv.org/abs/2510.01265 +390,"Gated Attention for Large Language Models: Non-linearity, Sparsity, and + Attention-Sink-Free",10,31,38.2,36,"May 10, 2025",6,240,https://arxiv.org/abs/2505.06708 +391,"CMPhysBench: A Benchmark for Evaluating Large Language Models in + Condensed Matter Physics",5,13,25.6,19,"Aug 25, 2025",45,16,https://arxiv.org/abs/2508.18124 +392,Spotlight on Token Perception for Multimodal Reinforcement Learning,6,10,29.83,27,"Oct 10, 2025",31,26,https://arxiv.org/abs/2510.09285 +393,Video-as-Answer: Predict and Generate Next Video Event with Joint-GRPO,6,23,29.83,29,"Nov 20, 2025",29,44,https://arxiv.org/abs/2511.16669 +394,"SyGra: A Unified Graph-Based Framework for Scalable Generation, Quality + Tagging, and Management of Synthetic Data",6,16,30,28,"Aug 21, 2025",6,16,https://arxiv.org/abs/2508.15432 +395,"Uni-cot: Towards Unified Chain-of-Thought Reasoning Across Text and + Vision",7,24,33.29,32,"Aug 7, 2025",0,136,https://arxiv.org/abs/2508.05606 +396,SRUM: Fine-Grained Self-Rewarding for Unified Multimodal Models,9,15,37.22,44,"Oct 14, 2025",17,51,https://arxiv.org/abs/2510.12784 +397,"JanusCoder: Towards a Foundational Visual-Programmatic Interface for + Code Intelligence",7,19,33.29,38,"Oct 27, 2025",90,54,https://arxiv.org/abs/2510.23538 +398,"SearchInstruct: Enhancing Domain Adaptation via Retrieval-Based + Instruction Dataset Creation",5,18,26.4,26,"Sep 12, 2025",9,8,https://arxiv.org/abs/2509.10708 +399,RynnVLA-002: A Unified Vision-Language-Action and World Model,11,29,39.82,41,"Nov 21, 2025",24,669,https://arxiv.org/abs/2511.17502 +400,BitNet Distillation,13,35,41.62,41,"Oct 15, 2025",44,24300,https://arxiv.org/abs/2510.13998 +401,UniREditBench: A Unified Reasoning-based Image Editing Benchmark,8,28,35.88,32,"Nov 3, 2025",35,37,https://arxiv.org/abs/2511.01295 +402,Artificial Hippocampus Networks for Efficient Long-Context Modeling,6,24,31,28,"Oct 8, 2025",22,67,https://arxiv.org/abs/2510.07318 +403,Agentic Entropy-Balanced Policy Optimization,8,30,36.12,33,"Oct 16, 2025",90,694,https://arxiv.org/abs/2510.14545 +404,"Pass@k Training for Adaptively Balancing Exploration and Exploitation of + Large Reasoning Models",21,40,45.43,45,"Aug 14, 2025",21,47,https://arxiv.org/abs/2508.10751 +405,"URSA: Understanding and Verifying Chain-of-thought Reasoning in + Multimodal Mathematics",6,26,31.83,32,"Jan 8, 2025",53,117,https://arxiv.org/abs/2501.04686 +406,"MCPMark: A Benchmark for Stress-Testing Realistic and Comprehensive MCP + Use",8,32,36.62,35,"Sep 28, 2025",145,201,https://arxiv.org/abs/2509.24002 +407,"AMFT: Aligning LLM Reasoners by Meta-Learning the Optimal + Imitation-Exploration Balance",16,31,43.88,45,"Aug 9, 2025",2,163,https://arxiv.org/abs/2508.06944 +408,"Diffuman4D: 4D Consistent Human View Synthesis from Sparse-View Videos + with Spatio-Temporal Diffusion Models",4,14,22.75,25,"Jul 17, 2025",50,246,https://arxiv.org/abs/2507.13344 +409,"Cache-to-Cache: Direct Semantic Communication Between Large Language + Models",10,35,39.8,40,"Oct 3, 2025",95,239,https://arxiv.org/abs/2510.03215 +410,"SageAttention3: Microscaling FP4 Attention for Inference and An + Exploration of 8-Bit Training",12,26,41.75,43,"May 16, 2025",76,2450,https://arxiv.org/abs/2505.11594 +411,NANO3D: A Training-Free Approach for Efficient 3D Editing Without Masks,5,22,28.8,25,"Oct 16, 2025",52,47,https://arxiv.org/abs/2510.15019 +412,"MeshSplat: Generalizable Sparse-View Surface Reconstruction via Gaussian + Splatting",6,22,32.67,28,"Aug 25, 2025",4,18,https://arxiv.org/abs/2508.17811 +413,"Drawing2CAD: Sequence-to-Sequence Learning for CAD Generation from + Vector Drawings",8,22,37.38,41,"Aug 26, 2025",2,50,https://arxiv.org/abs/2508.18733 +414,OceanGym: A Benchmark Environment for Underwater Embodied Agents,5,22,29.2,28,"Sep 30, 2025",30,28,https://arxiv.org/abs/2509.26536 +415,GR00T N1: An Open Foundation Model for Generalist Humanoid Robots,13,27,42.77,47,"Mar 18, 2025",4,4800,https://arxiv.org/abs/2503.14734 +416,TexVerse: A Universe of 3D Objects with High-Resolution Textures,12,37,42.08,40,"Aug 14, 2025",13,180,https://arxiv.org/abs/2508.10868 +417,"GSM8K-V: Can Vision Language Models Solve Grade School Math Word Problems in Visual Contexts",6,24,33.17,31,"Sep 29, 2025",25,21,https://arxiv.org/abs/2509.25160 -368,LongCodeZip: Compress Long Context for Code Language Models,8,23,31.38,30,"Oct 1, 2025",88,63,https://arxiv.org/abs/2510.00446 -369,"MathCanvas: Intrinsic Visual Chain-of-Thought for Multimodal - Mathematical Reasoning",2,47,47.5,47,"Oct 16, 2025",11,5,https://arxiv.org/abs/2510.14958 -370,"RLVE: Scaling Up Reinforcement Learning for Language Models with - Adaptive Verifiable Environments",20,18,28.15,23,"Nov 10, 2025",10,119,https://arxiv.org/abs/2511.07317 -371,"Feedback-Driven Tool-Use Improvements in Large Language Models via - Automated Build Environments",2,46,48,48,"Aug 12, 2025",12,17,https://arxiv.org/abs/2508.08791 -372,Generative AI for Autonomous Driving: Frontiers and Opportunities,2,48,48,48,"May 13, 2025",1,176,https://arxiv.org/abs/2505.08854 -373,"TaDiCodec: Text-aware Diffusion Speech Tokenizer for Speech Language - Modeling",18,12,28.44,31,"Aug 22, 2025",7,149,https://arxiv.org/abs/2508.16790 -374,CHARM: Control-point-based 3D Anime Hairstyle Auto-Regressive Modeling,2,46,48,48,"Sep 25, 2025",11,9,https://arxiv.org/abs/2509.21114 -375,"Lumina-DiMOO: An Omni Diffusion Large Language Model for Multi-Modal - Generation and Understanding",13,18,29.38,30,"Oct 7, 2025",48,834,https://arxiv.org/abs/2510.06308 -376,UI2Code^N: A Visual Language Model for Test-Time Scalable Interactive UI-to-Code Generation,2,47,48,48,"Nov 11, 2025",28,18,https://arxiv.org/abs/2511.08195 -377,"PromptCoT 2.0: Scaling Prompt Synthesis for Large Language Model - Reasoning",3,28,41,47,"Sep 24, 2025",29,93,https://arxiv.org/abs/2509.19894 -378,"VisionThink: Smart and Efficient Vision Language Model via Reinforcement - Learning",3,33,41.33,45,"Jul 17, 2025",69,330,https://arxiv.org/abs/2507.13348 -379,"Learning to Optimize Multi-Objective Alignment Through Dynamic Reward +418,"TOUCAN: Synthesizing 1.5M Tool-Agentic Data from Real-World MCP + Environments",9,30,39.11,41,"Oct 1, 2025",16,60,https://arxiv.org/abs/2510.01179 +419,"CoDiEmb: A Collaborative yet Distinct Framework for Unified + Representation Learning in Information Retrieval and Semantic Textual + Similarity",10,36,40.3,39,"Aug 15, 2025",3,125,https://arxiv.org/abs/2508.11442 +420,3D-R1: Enhancing Reasoning in 3D VLMs for Unified Scene Understanding,10,27,40.4,42,"Jul 31, 2025",15,307,https://arxiv.org/abs/2507.23478 +421,"VLA-RFT: Vision-Language-Action Reinforcement Fine-tuning with Verified + Rewards in World Simulators",5,20,30,25,"Oct 1, 2025",55,28,https://arxiv.org/abs/2510.00406 +422,"SeC: Advancing Complex Video Object Segmentation via Progressive Concept + Construction",3,12,16.67,19,"Jul 21, 2025",34,98,https://arxiv.org/abs/2507.15852 +423,Aligning Multimodal LLM with Human Preference: A Survey,20,31,45.95,48,"Mar 18, 2025",26,16300,https://arxiv.org/abs/2503.14504 +424,Skywork R1V: Pioneering Multimodal Reasoning with Chain-of-Thought,8,23,38.5,42,"Apr 8, 2025",86,3080,https://arxiv.org/abs/2504.05599 +425,"HierSearch: A Hierarchical Enterprise Deep Search Framework Integrating + Local and Web Searches",8,33,38.5,37,"Aug 11, 2025",26,30,https://arxiv.org/abs/2508.08088 +426,SIM-CoT: Supervised Implicit Chain-of-Thought,8,33,38.5,34,"Sep 24, 2025",33,46,https://arxiv.org/abs/2509.20317 +427,"RewardMap: Tackling Sparse Rewards in Fine-grained Visual Reasoning via + Multi-Stage Reinforcement Learning",5,18,31.2,29,"Oct 2, 2025",15,22,https://arxiv.org/abs/2510.02240 +428,Generating Physically Stable and Buildable LEGO Designs from Text,7,27,37,41,"May 8, 2025",28,1430,https://arxiv.org/abs/2505.05469 +429,"Sparse VideoGen2: Accelerate Video Generation with Sparse Attention via + Semantic-Aware Permutation",8,35,39.12,38,"May 24, 2025",42,472,https://arxiv.org/abs/2505.18875 +430,"Sparse VideoGen: Accelerating Video Diffusion Transformers with + Spatial-Temporal Sparsity",8,34,39.12,38,"Feb 3, 2025",3,472,https://arxiv.org/abs/2502.01776 +431,"Vision-Zero: Scalable VLM Self-Improvement via Strategic Gamified + Self-Play",9,29,40.44,44,"Sep 29, 2025",120,40,https://arxiv.org/abs/2509.25541 +432,"OmniVideoBench: Towards Audio-Visual Understanding Evaluation for Omni + MLLMs",6,23,35.17,33,"Oct 12, 2025",40,25,https://arxiv.org/abs/2510.10689 +433,"E^2Rank: Your Text Embedding can Also be an Effective + and Efficient Listwise Reranker",5,29,32,33,"Oct 26, 2025",29,22,https://arxiv.org/abs/2510.22733 +434,"LIBERO-Plus: In-depth Robustness Analysis of Vision-Language-Action + Models",9,30,40.56,41,"Oct 15, 2025",40,46,https://arxiv.org/abs/2510.13626 +435,Drax: Speech Recognition with Discrete Flow Matching,4,21,27.75,21,"Oct 5, 2025",23,18,https://arxiv.org/abs/2510.04162 +436,"OS-Sentinel: Towards Safety-Enhanced Mobile GUI Agents via Hybrid + Validation in Realistic Workflows",9,29,40.67,45,"Oct 28, 2025",69,31,https://arxiv.org/abs/2510.24411 +437,Symbolic Graphics Programming with Large Language Models,10,33,42,44,"Sep 5, 2025",37,18,https://arxiv.org/abs/2509.05208 +438,SparseD: Sparse Attention for Diffusion Language Models,6,30,36,34,"Sep 28, 2025",26,33,https://arxiv.org/abs/2509.24014 +439,"Language Model Council: Benchmarking Foundation Models on Highly + Subjective Tasks by Consensus",14,35,44.57,46,"Jun 12, 2024",6,180,https://arxiv.org/abs/2406.08598 +440,"MM-HELIX: Boosting Multimodal Long-Chain Reflective Reasoning with + Holistic Platform and Adaptive Hybrid Policy Optimization",7,33,38.43,37,"Oct 9, 2025",98,55,https://arxiv.org/abs/2510.08540 +441,PICABench: How Far Are We from Physically Realistic Image Editing?,3,12,21.67,15,"Oct 20, 2025",57,13,https://arxiv.org/abs/2510.17681 +442,"The Tool Decathlon: Benchmarking Language Agents for Diverse, Realistic, + and Long-Horizon Task Execution",10,38,42.4,42,"Oct 29, 2025",42,99,https://arxiv.org/abs/2510.25726 +443,Next Visual Granularity Generation,7,33,38.86,35,"Aug 18, 2025",37,7,https://arxiv.org/abs/2508.12811 +444,Droplet3D: Commonsense Priors from Videos Facilitate 3D Generation,8,35,40.5,39,"Aug 28, 2025",62,18,https://arxiv.org/abs/2508.20470 +445,OmniGen2: Exploration to Advanced Multimodal Generation,10,32,42.7,42,"Jun 23, 2025",75,3860,https://arxiv.org/abs/2506.18871 +446,"JavisDiT: Joint Audio-Video Diffusion Transformer with Hierarchical + Spatio-Temporal Prior Synchronization",16,35,45.88,47,"Mar 30, 2025",57,265,https://arxiv.org/abs/2503.23377 +447,"UP2You: Fast Reconstruction of Yourself from Unconstrained Photo + Collections",7,22,39.29,45,"Sep 29, 2025",8,161,https://arxiv.org/abs/2509.24817 +448,"LLaSO: A Foundational Framework for Reproducible Research in Large + Language and Speech Model",7,28,39.43,36,"Aug 21, 2025",4,79,https://arxiv.org/abs/2508.15418 +449,"MUG-V 10B: High-efficiency Training Pipeline for Large Video Generation + Models",3,18,24,24,"Oct 20, 2025",9,59,https://arxiv.org/abs/2510.17519 +450,"PartCrafter: Structured 3D Mesh Generation via Compositional Latent + Diffusion Transformers",7,23,39.71,40,"Jun 5, 2025",79,2180,https://arxiv.org/abs/2506.05573 +451,Discrete Diffusion in Large Language and Multimodal Models: A Survey,13,36,44.92,46,"Jun 16, 2025",43,267,https://arxiv.org/abs/2506.13759 +452,"PosterGen: Aesthetic-Aware Paper-to-Poster Generation via Multi-Agent + LLMs",11,35,43.91,43,"Aug 24, 2025",15,105,https://arxiv.org/abs/2508.17188 +453,"Implicit Actor Critic Coupling via a Supervised Learning Framework for + RLVR",4,28,31.75,32,"Sep 2, 2025",19,18,https://arxiv.org/abs/2509.02522 +454,UniME-V2: MLLM-as-a-Judge for Universal Multimodal Embedding Learning,6,27,38.17,39,"Oct 15, 2025",11,26,https://arxiv.org/abs/2510.13515 +455,Part-X-MLLM: Part-aware 3D Multimodal Large Language Model,3,25,25.33,25,"Nov 17, 2025",65,49,https://arxiv.org/abs/2511.13647 +456,Scaling RL to Long Videos,3,21,25.67,28,"Jul 10, 2025",141,538,https://arxiv.org/abs/2507.07966 +457,"X-VLA: Soft-Prompted Transformer as Scalable Cross-Embodiment + Vision-Language-Action Model",7,32,40.14,42,"Oct 11, 2025",12,56,https://arxiv.org/abs/2510.10274 +458,"Promptomatix: An Automatic Prompt Optimization Framework for Large + Language Models",3,19,26.33,30,"Jul 17, 2025",13,61,https://arxiv.org/abs/2507.14241 +459,Optimized Minimal 4D Gaussian Splatting,5,30,36.4,35,"Oct 4, 2025",3,29,https://arxiv.org/abs/2510.03857 +460,Defeating the Training-Inference Mismatch via FP16,9,33,42.89,47,"Oct 30, 2025",23,112,https://arxiv.org/abs/2510.26788 +461,"Teaching Pretrained Language Models to Think Deeper with Retrofitted + Recurrence",7,36,40.57,36,"Nov 10, 2025",11,23,https://arxiv.org/abs/2511.07384 +462,RynnVLA-001: Using Human Demonstrations to Improve Robot Manipulation,7,32,40.71,41,"Sep 18, 2025",20,208,https://arxiv.org/abs/2509.15212 +463,MediaPipe: A Framework for Building Perception Pipelines,11,39,44.45,45,"Jun 14, 2019",1,32100,https://arxiv.org/abs/1906.08172 +464,PPTAgent: Generating and Evaluating Presentations Beyond Text-to-Slides,2,15,15,15,"Jan 7, 2025",22,2430,https://arxiv.org/abs/2501.03936 +465,WideSearch: Benchmarking Agentic Broad Info-Seeking,7,37,40.86,41,"Aug 11, 2025",96,62,https://arxiv.org/abs/2508.07999 +466,"Efficient Multi-modal Large Language Models via Progressive Consistency + Distillation",4,23,33.25,33,"Oct 1, 2025",37,13,https://arxiv.org/abs/2510.00515 +467,UnSAMv2: Self-Supervised Learning Enables Segment Anything at Any Granularity,3,27,28,28,"Nov 17, 2025",4,34,https://arxiv.org/abs/2511.13714 +468,"SonicMaster: Towards Controllable All-in-One Music Restoration and + Mastering",7,32,41.29,39,"Aug 5, 2025",1,74,https://arxiv.org/abs/2508.03448 +469,DINO-Foresight: Looking into the Future with DINO,7,39,41.29,41,"Dec 16, 2024",1,107,https://arxiv.org/abs/2412.11673 +470,"Euclid's Gift: Enhancing Spatial Perception and Reasoning in + Vision-Language Models via Geometric Surrogate Tasks",4,29,34.25,31,"Sep 29, 2025",15,13,https://arxiv.org/abs/2509.24473 +471,CoDA: Coding LM via Diffusion Adaptation,5,34,37.6,37,"Sep 27, 2025",25,21,https://arxiv.org/abs/2510.03270 +472,"MiniMax-M1: Scaling Test-Time Compute Efficiently with Lightning + Attention",8,34,42.75,44,"Jun 16, 2025",265,2890,https://arxiv.org/abs/2506.13585 +473,"Hyperspherical Latents Improve Continuous-Token Autoregressive + Generation",6,30,40,42,"Sep 29, 2025",6,43,https://arxiv.org/abs/2509.24335 +474,VLA-0: Building State-of-the-Art VLAs with Zero Modification,6,33,40,39,"Oct 15, 2025",8,103,https://arxiv.org/abs/2510.13054 +475,"Self Forcing: Bridging the Train-Test Gap in Autoregressive Video + Diffusion",17,41,47.18,49,"Jun 9, 2025",27,2420,https://arxiv.org/abs/2506.08009 +476,"RealUnify: Do Unified Models Truly Benefit from Unification? A + Comprehensive Benchmark",5,29,38,38,"Sep 29, 2025",41,14,https://arxiv.org/abs/2509.24897 +477,Uniform Discrete Diffusion with Metric Path for Video Generation,5,24,38,41,"Oct 28, 2025",39,49,https://arxiv.org/abs/2510.24717 +478,"InternVLA-M1: A Spatially Guided Vision-Language-Action Framework for + Generalist Robot Policy",7,35,42,44,"Oct 15, 2025",13,167,https://arxiv.org/abs/2510.13778 +479,Vidi: Large Multimodal Models for Video Understanding and Editing,4,27,35.5,37,"Apr 22, 2025",14,247,https://arxiv.org/abs/2504.15681 +480,UniVid: Unifying Vision Tasks with Pre-trained Video Generation Models,5,23,38.8,48,"Sep 26, 2025",11,25,https://arxiv.org/abs/2509.21760 +481,Sparser Block-Sparse Attention via Token Permutation,3,22,31,26,"Oct 24, 2025",22,19,https://arxiv.org/abs/2510.21270 +482,DeepEyesV2: Toward Agentic Multimodal Model,8,39,43.5,45,"Nov 7, 2025",35,963,https://arxiv.org/abs/2511.05271 +483,Rethinking Reward Models for Multi-Domain Test-Time Scaling,3,28,31.33,33,"Oct 1, 2025",23,14,https://arxiv.org/abs/2510.00492 +484,Depth Anything with Any Prior,6,37,41.17,40,"May 15, 2025",12,389,https://arxiv.org/abs/2505.10565 +485,AU-Harness: An Open-Source Toolkit for Holistic Evaluation of Audio LLMs,4,30,36.5,33,"Sep 9, 2025",18,35,https://arxiv.org/abs/2509.08031 +486,NORA-1.5: A Vision-Language-Action Model Trained using World Model- and Action-based Preference Rewards,3,31,31.67,32,"Nov 18, 2025",7,26,https://arxiv.org/abs/2511.14659 +487,"Learning to Optimize Multi-Objective Alignment Through Dynamic Reward Weighting",4,35,37.5,35,"Sep 14, 2025",8,4,https://arxiv.org/abs/2509.11452 -380,Rectified Point Flow: Generic Point Cloud Pose Estimation,3,34,41.33,44,"Jun 5, 2025",3,123,https://arxiv.org/abs/2506.05282 -381,"ASTRA: Autonomous Spatial-Temporal Red-teaming for AI Software - Assistants",2,49,49.5,49,"Aug 5, 2025",8,37,https://arxiv.org/abs/2508.03936 -382,"InfiGUI-G1: Advancing GUI Grounding with Adaptive Exploration Policy - Optimization",2,49,49.5,49,"Aug 7, 2025",25,56,https://arxiv.org/abs/2508.05731 -383,Durian: Dual Reference-guided Portrait Animation with Attribute Transfer,15,16,29.13,22,"Sep 4, 2025",4,14,https://arxiv.org/abs/2509.04434 -384,"MAS-Bench: A Unified Benchmark for Shortcut-Augmented Hybrid Mobile GUI +488,UQ: Assessing Language Models on Unsolved Questions,5,27,40.4,45,"Aug 25, 2025",11,15,https://arxiv.org/abs/2508.17580 +489,"MAS-Bench: A Unified Benchmark for Shortcut-Augmented Hybrid Mobile GUI Agents",4,27,37.75,40,"Sep 8, 2025",2,8,https://arxiv.org/abs/2509.06477 -385,"Gradient-Attention Guided Dual-Masking Synergetic Framework for Robust - Text-based Person Retrieval",3,30,41.67,45,"Sep 11, 2025",6,10,https://arxiv.org/abs/2509.09118 -386,"ObjFiller-3D: Consistent Multi-view 3D Inpainting via Video Diffusion +490,"ObjFiller-3D: Consistent Multi-view 3D Inpainting via Video Diffusion Models",4,25,38,42,"Aug 25, 2025",5,25,https://arxiv.org/abs/2508.18271 -387,"SLA: Beyond Sparsity in Diffusion Transformers via Fine-Tunable - Sparse-Linear Attention",2,50,50,50,"Sep 28, 2025",107,38,https://arxiv.org/abs/2509.24006 -388,Scaling Large-Language-Model-based Multi-Agent Collaboration,2,50,50,50,"Jun 11, 2024",3,27600,https://arxiv.org/abs/2406.07155 -389,"Ming-UniVision: Joint Image Understanding and Generation with a Unified - Continuous Tokenizer",3,38,42.33,42,"Oct 8, 2025",63,72,https://arxiv.org/abs/2510.06590 -390,WEAVE: Unleashing and Benchmarking the In-context Interleaved Comprehension and Generation,4,30,38.5,38,"Nov 14, 2025",42,25,https://arxiv.org/abs/2511.11434 -391,MUR: Momentum Uncertainty guided Reasoning for Large Language Models,3,40,43,44,"Jul 20, 2025",36,32,https://arxiv.org/abs/2507.14958 -392,"Uni-cot: Towards Unified Chain-of-Thought Reasoning Across Text and - Vision",7,24,33.29,32,"Aug 7, 2025",0,136,https://arxiv.org/abs/2508.05606 -393,"Scaling Code-Assisted Chain-of-Thoughts and Instructions for Model - Reasoning",3,39,43,42,"Oct 5, 2025",18,6,https://arxiv.org/abs/2510.04081 -394,"JanusCoder: Towards a Foundational Visual-Programmatic Interface for - Code Intelligence",7,19,33.29,38,"Oct 27, 2025",90,54,https://arxiv.org/abs/2510.23538 -395,Cook and Clean Together: Teaching Embodied Agents for Parallel Task Execution,11,19,30.64,28,"Nov 24, 2025",7,270,https://arxiv.org/abs/2511.19430 -396,Optimized Minimal 4D Gaussian Splatting,5,30,36.4,35,"Oct 4, 2025",3,29,https://arxiv.org/abs/2510.03857 -397,V-ReasonBench: Toward Unified Reasoning Benchmark Suite for Video Generation Models,3,38,43.33,43,"Nov 20, 2025",48,15,https://arxiv.org/abs/2511.16668 -398,"BEAVR: Bimanual, multi-Embodiment, Accessible, Virtual Reality - Teleoperation System for Robots",11,19,30.82,31,"Aug 13, 2025",0,50,https://arxiv.org/abs/2508.09606 -399,"VolSplat: Rethinking Feed-Forward 3D Gaussian Splatting with - Voxel-Aligned Prediction",14,16,29.86,24,"Sep 23, 2025",22,65,https://arxiv.org/abs/2509.19297 -400,"EPO: Entropy-regularized Policy Optimization for LLM Agents +491,"RenderFormer: Transformer-based Neural Rendering of Triangle Meshes with + Global Illumination",5,34,40.6,39,"May 28, 2025",37,765,https://arxiv.org/abs/2505.21925 +492,Upsample Anything: A Simple and Hard to Beat Baseline for Feature Upsampling,9,38,45.22,46,"Nov 20, 2025",6,105,https://arxiv.org/abs/2511.16301 +493,Foundations of Large Language Models,5,31,40.8,40,"Jan 16, 2025",11,264,https://arxiv.org/abs/2501.09223 +494,DanceGRPO: Unleashing GRPO on Visual Generation,11,37,46.45,47,"May 12, 2025",32,942,https://arxiv.org/abs/2505.07818 +495,SPARK: Synergistic Policy And Reward Co-Evolving Framework,3,24,34.33,30,"Sep 26, 2025",15,17,https://arxiv.org/abs/2509.22624 +496,WEAVE: Unleashing and Benchmarking the In-context Interleaved Comprehension and Generation,4,30,38.5,38,"Nov 14, 2025",42,25,https://arxiv.org/abs/2511.11434 +497,"Eliciting Fine-Tuned Transformer Capabilities via Inference-Time + Techniques",7,34,44,48,"Jun 9, 2025",8,2910,https://arxiv.org/abs/2506.08060 +498,SpatialTrackerV2: 3D Point Tracking Made Easy,3,32,35.33,32,"Jul 16, 2025",14,658,https://arxiv.org/abs/2507.12462 +499,Reasoning with Sampling: Your Base Model is Smarter Than You Think,3,27,35.33,33,"Oct 16, 2025",35,229,https://arxiv.org/abs/2510.14901 +500,"DocETL: Agentic Query Rewriting and Evaluation for Complex Document + Processing",5,37,41.6,43,"Oct 16, 2024",1,3180,https://arxiv.org/abs/2410.12189 +501,Nav-R1: Reasoning and Navigation in Embodied Scenes,6,39,43.33,43,"Sep 13, 2025",4,21,https://arxiv.org/abs/2509.10884 +502,"EPO: Entropy-regularized Policy Optimization for LLM Agents Reinforcement Learning",4,23,39.5,44,"Sep 26, 2025",113,20,https://arxiv.org/abs/2509.22576 -401,"PUSA V1.0: Surpassing Wan-I2V with $500 Training Cost by Vectorized +503,"PUSA V1.0: Surpassing Wan-I2V with $500 Training Cost by Vectorized Timestep Adaptation",4,28,39.75,42,"Jul 22, 2025",9,545,https://arxiv.org/abs/2507.16116 -402,Deep Researcher with Test-Time Diffusion,3,36,44.33,47,"Jul 21, 2025",63,2910,https://arxiv.org/abs/2507.16075 -403,"OmniVideoBench: Towards Audio-Visual Understanding Evaluation for Omni - MLLMs",6,23,35.17,33,"Oct 12, 2025",40,25,https://arxiv.org/abs/2510.10689 -404,LiteAttention: A Temporal Sparse Attention for Diffusion Transformers,3,42,44.33,43,"Nov 14, 2025",24,23,https://arxiv.org/abs/2511.11062 -405,ROSE: Remove Objects with Side Effects in Videos,12,14,30.67,33,"Aug 26, 2025",4,32,https://arxiv.org/abs/2508.18633 -406,VACE: All-in-One Video Creation and Editing,3,43,45,44,"Mar 10, 2025",54,3000,https://arxiv.org/abs/2503.07598 -407,Efficient Guided Generation for Large Language Models,3,41,45,45,"Jul 19, 2023",8,12900,https://arxiv.org/abs/2307.09702 -408,CoDA: Coding LM via Diffusion Adaptation,5,34,37.6,37,"Sep 27, 2025",25,21,https://arxiv.org/abs/2510.03270 -409,"Embodied-R1: Reinforced Embodied Reasoning for General Robotic +504,"Efficient Multi-turn RL for GUI Agents via Decoupled Training and + Adaptive Data Curation",3,32,36,35,"Sep 28, 2025",7,9,https://arxiv.org/abs/2509.23866 +505,OBS-Diff: Accurate Pruning For Diffusion Models in One-Shot,2,26,28.5,28,"Oct 8, 2025",14,24,https://arxiv.org/abs/2510.06751 +506,GUI-G^2: Gaussian Reward Modeling for GUI Grounding,2,24,29,29,"Jul 21, 2025",118,138,https://arxiv.org/abs/2507.15846 +507,GIR-Bench: Versatile Benchmark for Generating Images with Reasoning,3,29,37,39,"Oct 13, 2025",16,23,https://arxiv.org/abs/2510.11026 +508,Beyond English: Toward Inclusive and Scalable Multilingual Machine Translation with LLMs,7,42,45,46,"Nov 10, 2025",31,92,https://arxiv.org/abs/2511.07003 +509,"Embodied-R1: Reinforced Embodied Reasoning for General Robotic Manipulation",4,31,40.75,42,"Aug 19, 2025",12,24,https://arxiv.org/abs/2508.13998 -410,"Ming-UniAudio: Speech LLM for Joint Understanding, Generation and - Editing with Unified Representation",14,9,30.21,28,"Oct 26, 2025",9,370,https://arxiv.org/abs/2511.05516 -411,"BrowseComp-Plus: A More Fair and Transparent Evaluation Benchmark of - Deep-Research Agent",3,43,46,45,"Aug 8, 2025",35,49,https://arxiv.org/abs/2508.06600 -412,"RealUnify: Do Unified Models Truly Benefit from Unification? A - Comprehensive Benchmark",5,29,38,38,"Sep 29, 2025",41,14,https://arxiv.org/abs/2509.24897 -413,SparseD: Sparse Attention for Diffusion Language Models,6,30,36,34,"Sep 28, 2025",26,33,https://arxiv.org/abs/2509.24014 -414,"CodePlot-CoT: Mathematical Visual Reasoning by Thinking with Code-Driven - Images",3,38,46,50,"Oct 13, 2025",12,22,https://arxiv.org/abs/2510.11718 -415,Uniform Discrete Diffusion with Metric Path for Video Generation,5,24,38,41,"Oct 28, 2025",39,49,https://arxiv.org/abs/2510.24717 -416,Reverse-Engineered Reasoning for Open-Ended Generation,4,38,41.25,40,"Sep 7, 2025",127,34,https://arxiv.org/abs/2509.06160 -417,"Evolving Language Models without Labels: Majority Drives Selection, - Novelty Promotes Variation",3,41,46.67,49,"Sep 18, 2025",29,19,https://arxiv.org/abs/2509.15194 -418,LightsOut: Diffusion-based Outpainting for Enhanced Lens Flare Removal,3,41,46.67,49,"Oct 17, 2025",20,10,https://arxiv.org/abs/2510.15868 -419,"DCReg: Decoupled Characterization for Efficient Degenerate LiDAR - Registration",3,41,47,50,"Sep 8, 2025",1,58,https://arxiv.org/abs/2509.06285 -420,"Beyond the Exploration-Exploitation Trade-off: A Hidden State Approach - for LLM Reasoning in RLVR",3,46,47,47,"Sep 28, 2025",38,7,https://arxiv.org/abs/2509.23808 -421,"UMO: Scaling Multi-Identity Consistency for Image Customization via - Matching Reward",19,18,29.37,26,"Sep 8, 2025",29,131,https://arxiv.org/abs/2509.06818 -422,"VideoGen-of-Thought: A Collaborative Framework for Multi-Shot Video +510,"SwiftEdit: Lightning Fast Text-Guided Image Editing via One-Step + Diffusion",3,36,37.33,36,"Dec 5, 2024",41,7,https://arxiv.org/abs/2412.04301 +511,CLEAR: Error Analysis via LLM-as-a-Judge Made Easy,1,11,11,11,"Jul 24, 2025",8,10,https://arxiv.org/abs/2507.18392 +512,MindSearch: Mimicking Human Minds Elicits Deep AI Searcher,7,43,45.29,46,"Jul 29, 2024",44,6580,https://arxiv.org/abs/2407.20183 +513,dParallel: Learnable Parallel Decoding for dLLMs,5,40,43,42,"Sep 30, 2025",17,16,https://arxiv.org/abs/2509.26488 +514,"ReCogDrive: A Reinforced Cognitive Framework for End-to-End Autonomous + Driving",8,39,46.12,47,"Jun 9, 2025",1,183,https://arxiv.org/abs/2506.08052 +515,AWorld: Orchestrating the Training Recipe for Agentic AI,7,36,45.43,46,"Aug 28, 2025",37,692,https://arxiv.org/abs/2508.20404 +516,Reverse-Engineered Reasoning for Open-Ended Generation,4,38,41.25,40,"Sep 7, 2025",127,34,https://arxiv.org/abs/2509.06160 +517,MiniCPM4: Ultra-Efficient LLMs on End Devices,7,41,45.57,47,"Jun 9, 2025",90,8300,https://arxiv.org/abs/2506.07900 +518,SpaceVista: All-Scale Visual Spatial Reasoning from mm to km,2,30,32,32,"Oct 10, 2025",16,22,https://arxiv.org/abs/2510.09606 +519,Vlaser: Vision-Language-Action Model with Synergistic Embodied Reasoning,3,30,38.67,41,"Oct 13, 2025",14,15,https://arxiv.org/abs/2510.11027 +520,"Generating an Image From 1,000 Words: Enhancing Text-to-Image With + Structured Captions",5,38,43.6,41,"Nov 10, 2025",19,248,https://arxiv.org/abs/2511.06876 +521,"VideoGen-of-Thought: A Collaborative Framework for Multi-Shot Video Generation",4,38,42,41,"Dec 3, 2024",60,42,https://arxiv.org/abs/2412.02259 -423,UniVid: Unifying Vision Tasks with Pre-trained Video Generation Models,5,23,38.8,48,"Sep 26, 2025",11,25,https://arxiv.org/abs/2509.21760 -424,"Spatial Forcing: Implicit Spatial Representation Alignment for +522,"Spatial Forcing: Implicit Spatial Representation Alignment for Vision-language-action Model",4,39,42.25,41,"Oct 14, 2025",134,43,https://arxiv.org/abs/2510.12276 -425,"Trinity-RFT: A General-Purpose and Unified Framework for Reinforcement - Fine-Tuning of Large Language Models",3,48,48,48,"May 23, 2025",9,269,https://arxiv.org/abs/2505.17826 -426,Trainable Dynamic Mask Sparse Attention,3,46,48,49,"Aug 4, 2025",17,366,https://arxiv.org/abs/2508.02124 -427,VChain: Chain-of-Visual-Thought for Reasoning in Video Generation,9,25,33.33,29,"Oct 6, 2025",34,60,https://arxiv.org/abs/2510.05094 -428,"Franca: Nested Matryoshka Clustering for Scalable Visual Representation +523,MGM-Omni: Scaling Omni LLMs to Personalized Long-Horizon Speech,5,40,44.2,42,"Sep 29, 2025",11,133,https://arxiv.org/abs/2509.25131 +524,"Building a Foundational Guardrail for General Agentic Systems via + Synthetic Data",3,33,39.67,42,"Oct 10, 2025",21,26,https://arxiv.org/abs/2510.09781 +525,Mantis: A Versatile Vision-Language-Action Model with Disentangled Visual Foresight,2,32,34,34,"Nov 20, 2025",10,18,https://arxiv.org/abs/2511.16175 +526,"Franca: Nested Matryoshka Clustering for Scalable Visual Representation Learning",4,28,42.75,47,"Jul 18, 2025",27,189,https://arxiv.org/abs/2507.14137 -429,Reasoning in Space via Grounding in the World,3,45,48.33,50,"Oct 15, 2025",13,16,https://arxiv.org/abs/2510.13800 -430,"EasySteer: A Unified Framework for High-Performance and Extensible LLM - Steering",9,23,33.56,30,"Sep 29, 2025",25,49,https://arxiv.org/abs/2509.25175 -431,"ARTDECO: Towards Efficient and High-Fidelity On-the-Fly 3D - Reconstruction with Structured Scene Representation",9,25,33.56,32,"Oct 9, 2025",30,73,https://arxiv.org/abs/2510.08551 -432,"LLMs Can Get ""Brain Rot""!",4,38,43,42,"Oct 15, 2025",19,72,https://arxiv.org/abs/2510.13928 -433,"SAM2Act: Integrating Visual Foundation Model with A Memory Architecture - for Robotic Manipulation",3,48,49,49,"Jan 30, 2025",2,131,https://arxiv.org/abs/2501.18564 -434,"TalkVid: A Large-Scale Diversified Dataset for Audio-Driven Talking Head +527,"How Far are VLMs from Visual Spatial Intelligence? A Benchmark-Driven + Perspective",3,37,40,37,"Sep 23, 2025",22,10,https://arxiv.org/abs/2509.18905 +528,Regression Language Models for Code,6,40,45.5,47,"Sep 30, 2025",13,257,https://arxiv.org/abs/2509.26476 +529,"CoIRL-AD: Collaborative-Competitive Imitation-Reinforcement Learning in + Latent World Models for Autonomous Driving",3,36,40,42,"Oct 14, 2025",4,10,https://arxiv.org/abs/2510.12560 +530,"Scaling Text-Rich Image Understanding via Code-Guided Synthetic + Multimodal Data Generation",2,35,35,35,"Feb 20, 2025",13,103,https://arxiv.org/abs/2502.14846 +531,"StreamDiffusion: A Pipeline-level Solution for Real-time Interactive + Generation",2,34,35,35,"Dec 19, 2023",73,10400,https://arxiv.org/abs/2312.12491 +532,"LLMs Can Get ""Brain Rot""!",4,38,43,42,"Oct 15, 2025",19,72,https://arxiv.org/abs/2510.13928 +533,"Reinforcement Learning Optimization for Large-Scale Learning: An + Efficient and User-Friendly Scaling Library",5,40,44.6,45,"Jun 6, 2025",7,2270,https://arxiv.org/abs/2506.06122 +534,ConsistEdit: Highly Consistent and Precise Training-free Visual Editing,2,35,35.5,35,"Oct 20, 2025",11,25,https://arxiv.org/abs/2510.17803 +535,"TalkVid: A Large-Scale Diversified Dataset for Audio-Driven Talking Head Synthesis",4,34,43.5,45,"Aug 19, 2025",16,72,https://arxiv.org/abs/2508.13618 -435,LoopTool: Closing the Data-Training Loop for Robust LLM Tool Calls,9,21,33.78,39,"Nov 12, 2025",15,20,https://arxiv.org/abs/2511.09148 -436,"Equilibrium Matching: Generative Modeling with Implicit Energy-Based +536,Energy-Based Transformers are Scalable Learners and Thinkers,8,44,47.25,47,"Jul 2, 2025",65,463,https://arxiv.org/abs/2507.02092 +537,"PromptCoT 2.0: Scaling Prompt Synthesis for Large Language Model + Reasoning",3,28,41,47,"Sep 24, 2025",29,93,https://arxiv.org/abs/2509.19894 +538,Multi-Agent Tool-Integrated Policy Optimization,2,35,36,36,"Oct 6, 2025",19,20,https://arxiv.org/abs/2510.04678 +539,"VisionThink: Smart and Efficient Vision Language Model via Reinforcement + Learning",3,33,41.33,45,"Jul 17, 2025",69,330,https://arxiv.org/abs/2507.13348 +540,Rectified Point Flow: Generic Point Cloud Pose Estimation,3,34,41.33,44,"Jun 5, 2025",3,123,https://arxiv.org/abs/2506.05282 +541,"Equilibrium Matching: Generative Modeling with Implicit Energy-Based Models",4,41,43.75,44,"Oct 2, 2025",5,77,https://arxiv.org/abs/2510.02300 -437,UQ: Assessing Language Models on Unsolved Questions,5,27,40.4,45,"Aug 25, 2025",11,15,https://arxiv.org/abs/2508.17580 -438,"NER Retriever: Zero-Shot Named Entity Retrieval with Type-Aware - Embeddings",3,50,50,50,"Sep 4, 2025",19,19,https://arxiv.org/abs/2509.04011 -439,"RenderFormer: Transformer-based Neural Rendering of Triangle Meshes with - Global Illumination",5,34,40.6,39,"May 28, 2025",37,765,https://arxiv.org/abs/2505.21925 -440,UniME-V2: MLLM-as-a-Judge for Universal Multimodal Embedding Learning,6,27,38.17,39,"Oct 15, 2025",11,26,https://arxiv.org/abs/2510.13515 -441,Foundations of Large Language Models,5,31,40.8,40,"Jan 16, 2025",11,264,https://arxiv.org/abs/2501.09223 -442,The Markovian Thinker,19,15,29.95,29,"Oct 8, 2025",27,251,https://arxiv.org/abs/2510.06557 -443,LTX-Video: Realtime Video Latent Diffusion,18,19,30.17,30,"Dec 30, 2024",47,8590,https://arxiv.org/abs/2501.00103 -444,"Diffusion LLMs Can Do Faster-Than-AR Inference via Discrete Diffusion - Forcing",30,10,28.57,29,"Aug 8, 2025",29,79,https://arxiv.org/abs/2508.09192 -445,VerlTool: Towards Holistic Agentic Reinforcement Learning with Tool Use,52,7,27.48,28,"Sep 1, 2025",64,499,https://arxiv.org/abs/2509.01055 -446,DA^2: Depth Anything in Any Direction,17,8,30.53,37,"Sep 30, 2025",20,120,https://arxiv.org/abs/2509.26618 -447,DiT360: High-Fidelity Panoramic Image Generation via Hybrid Training,9,21,34.56,38,"Oct 13, 2025",29,105,https://arxiv.org/abs/2510.11712 -448,Generating Physically Stable and Buildable LEGO Designs from Text,7,27,37,41,"May 8, 2025",28,1430,https://arxiv.org/abs/2505.05469 -449,OmniSVG: A Unified Scalable Vector Graphics Generation Model,16,16,30.88,28,"Apr 8, 2025",180,2100,https://arxiv.org/abs/2504.06263 -450,"Voost: A Unified and Scalable Diffusion Transformer for Bidirectional - Virtual Try-On and Try-Off",37,9,28.11,26,"Aug 6, 2025",56,264,https://arxiv.org/abs/2508.04825 -451,MM-BrowseComp: A Comprehensive Benchmark for Multimodal Browsing Agents,4,40,45.5,46,"Aug 14, 2025",13,10,https://arxiv.org/abs/2508.13186 -452,"Mem4Nav: Boosting Vision-and-Language Navigation in Urban Environments +542,A decoder-only foundation model for time-series forecasting,6,44,46.17,46,"Oct 14, 2023",6,7060,https://arxiv.org/abs/2310.10688 +543,MoDA: Multi-modal Diffusion Architecture for Talking Head Generation,9,46,47.89,48,"Jul 4, 2025",2,140,https://arxiv.org/abs/2507.03256 +544,"Gradient-Attention Guided Dual-Masking Synergetic Framework for Robust + Text-based Person Retrieval",3,30,41.67,45,"Sep 11, 2025",6,10,https://arxiv.org/abs/2509.09118 +545,Streaming 4D Visual Geometry Transformer,1,25,25,25,"Jul 15, 2025",10,454,https://arxiv.org/abs/2507.11539 +546,s3: You Don't Need That Much Data to Train a Search Agent via RL,5,41,45.8,47,"May 20, 2025",18,564,https://arxiv.org/abs/2505.14146 +547,"MesaTask: Towards Task-Driven Tabletop Scene Generation via 3D Spatial + Reasoning",2,36,38,38,"Sep 26, 2025",27,24,https://arxiv.org/abs/2509.22281 +548,"Ming-UniVision: Joint Image Understanding and Generation with a Unified + Continuous Tokenizer",3,38,42.33,42,"Oct 8, 2025",63,72,https://arxiv.org/abs/2510.06590 +549,"iMontage: Unified, Versatile, Highly Dynamic Many-to-many Image Generation",2,34,38,38,"Nov 25, 2025",30,120,https://arxiv.org/abs/2511.20635 +550,"ThinkSound: Chain-of-Thought Reasoning in Multimodal Large Language + Models for Audio Generation and Editing",2,38,38.5,38,"Jun 26, 2025",7,876,https://arxiv.org/abs/2506.21448 +551,"Q-Sched: Pushing the Boundaries of Few-Step Diffusion Models with + Quantization-Aware Scheduling",2,30,38.5,38,"Sep 1, 2025",5,6,https://arxiv.org/abs/2509.01624 +552,"GAS: Improving Discretization of Diffusion ODEs via Generalized + Adversarial Solver",2,30,38.5,38,"Oct 20, 2025",2,10,https://arxiv.org/abs/2510.17699 +553,MUR: Momentum Uncertainty guided Reasoning for Large Language Models,3,40,43,44,"Jul 20, 2025",36,32,https://arxiv.org/abs/2507.14958 +554,"Scaling Code-Assisted Chain-of-Thoughts and Instructions for Model + Reasoning",3,39,43,42,"Oct 5, 2025",18,6,https://arxiv.org/abs/2510.04081 +555,"AniMaker: Automated Multi-Agent Animated Storytelling with MCTS-Driven + Clip Generation",8,46,48.12,48,"Jun 12, 2025",37,173,https://arxiv.org/abs/2506.10540 +556,"D2E: Scaling Vision-Action Pretraining on Desktop Data for Transfer to + Embodied AI",2,39,39.5,39,"Oct 7, 2025",101,22,https://arxiv.org/abs/2510.05684 +557,V-ReasonBench: Toward Unified Reasoning Benchmark Suite for Video Generation Models,3,38,43.33,43,"Nov 20, 2025",48,15,https://arxiv.org/abs/2511.16668 +558,"DetailFlow: 1D Coarse-to-Fine Autoregressive Image Generation via + Next-Detail Prediction",1,29,29,29,"May 27, 2025",16,144,https://arxiv.org/abs/2505.21473 +559,MM-BrowseComp: A Comprehensive Benchmark for Multimodal Browsing Agents,4,40,45.5,46,"Aug 14, 2025",13,10,https://arxiv.org/abs/2508.13186 +560,"Mem4Nav: Boosting Vision-and-Language Navigation in Urban Environments with a Hierarchical Spatial-Cognition Long-Short Memory System",4,41,45.5,46,"Jun 24, 2025",3,89,https://arxiv.org/abs/2506.19433 -453,"DocETL: Agentic Query Rewriting and Evaluation for Complex Document - Processing",5,37,41.6,43,"Oct 16, 2024",1,3180,https://arxiv.org/abs/2410.12189 -454,"IMAGGarment-1: Fine-Grained Garment Generation for Controllable Fashion - Design",14,21,31.64,31,"Apr 17, 2025",1,263,https://arxiv.org/abs/2504.13176 -455,UniREditBench: A Unified Reasoning-based Image Editing Benchmark,8,28,35.88,32,"Nov 3, 2025",35,37,https://arxiv.org/abs/2511.01295 -456,"Part II: ROLL Flash -- Accelerating RLVR and Agentic Training with +561,"Qwen3 Embedding: Advancing Text Embedding and Reranking Through + Foundation Models",5,44,46.8,46,"Jun 5, 2025",72,1390,https://arxiv.org/abs/2506.05176 +562,"Part II: ROLL Flash -- Accelerating RLVR and Agentic Training with Asynchrony",4,44,45.75,45,"Oct 13, 2025",15,2270,https://arxiv.org/abs/2510.11345 -457,Agent-R1: Training Powerful LLM Agents with End-to-End Reinforcement Learning,4,44,45.75,46,"Nov 18, 2025",15,951,https://arxiv.org/abs/2511.14460 -458,"SEAgent: Self-Evolving Computer Use Agent with Autonomous Learning from +563,Agent-R1: Training Powerful LLM Agents with End-to-End Reinforcement Learning,4,44,45.75,46,"Nov 18, 2025",15,951,https://arxiv.org/abs/2511.14460 +564,Deep Researcher with Test-Time Diffusion,3,36,44.33,47,"Jul 21, 2025",63,2910,https://arxiv.org/abs/2507.16075 +565,RoboOmni: Proactive Robot Manipulation in Omni-modal Context,2,40,41,41,"Oct 27, 2025",52,27,https://arxiv.org/abs/2510.23763 +566,LiteAttention: A Temporal Sparse Attention for Diffusion Transformers,3,42,44.33,43,"Nov 14, 2025",24,23,https://arxiv.org/abs/2511.11062 +567,"SEAgent: Self-Evolving Computer Use Agent with Autonomous Learning from Experience",4,45,46.25,46,"Aug 6, 2025",46,141,https://arxiv.org/abs/2508.04700 -459,"TensorBLEU: Vectorized GPU-based BLEU Score Implementation for +568,Quantile Advantage Estimation for Entropy-Safe Reasoning,2,35,41.5,41,"Sep 26, 2025",100,8,https://arxiv.org/abs/2509.22611 +569,"TensorBLEU: Vectorized GPU-based BLEU Score Implementation for Per-Sentence In-Training Evaluation",4,40,46.25,48,"Oct 7, 2025",7,13,https://arxiv.org/abs/2510.05485 -460,Agentic Entropy-Balanced Policy Optimization,8,30,36.12,33,"Oct 16, 2025",90,694,https://arxiv.org/abs/2510.14545 -461,"HSCodeComp: A Realistic and Expert-level Benchmark for Deep Search +570,MATRIX: Mask Track Alignment for Interaction-aware Video Generation,2,37,41.5,41,"Oct 8, 2025",29,22,https://arxiv.org/abs/2510.07310 +571,"HSCodeComp: A Realistic and Expert-level Benchmark for Deep Search Agents in Hierarchical Rule Application",4,43,46.25,46,"Oct 22, 2025",26,85,https://arxiv.org/abs/2510.19631 -462,Open Deep Search: Democratizing Search with Open-source Reasoning Agents,4,45,46.5,46,"Mar 26, 2025",48,3620,https://arxiv.org/abs/2503.20201 -463,The Denario project: Deep knowledge AI agents for scientific discovery,22,15,29.73,26,"Oct 30, 2025",6,372,https://arxiv.org/abs/2510.26887 -464,"Search-R1: Training LLMs to Reason and Leverage Search Engines with +572,VACE: All-in-One Video Creation and Editing,3,43,45,44,"Mar 10, 2025",54,3000,https://arxiv.org/abs/2503.07598 +573,Open Deep Search: Democratizing Search with Open-source Reasoning Agents,4,45,46.5,46,"Mar 26, 2025",48,3620,https://arxiv.org/abs/2503.20201 +574,TUN3D: Towards Real-World Scene Understanding from Unposed Images,1,33,33,33,"Sep 23, 2025",12,11,https://arxiv.org/abs/2509.21388 +575,Efficient Guided Generation for Large Language Models,3,41,45,45,"Jul 19, 2023",8,12900,https://arxiv.org/abs/2307.09702 +576,UltraFlux: Data-Model Co-Design for High-quality Native 4K Text-to-Image Generation across Diverse Aspect Ratios,2,36,42,42,"Nov 22, 2025",34,42,https://arxiv.org/abs/2511.18050 +577,STARFlow-V: End-to-End Video Generative Modeling with Normalizing Flow,6,45,48,48,"Nov 25, 2025",16,40,https://arxiv.org/abs/2511.20462 +578,Zebra-CoT: A Dataset for Interleaved Vision Language Reasoning,1,34,34,34,"Jul 22, 2025",28,42,https://arxiv.org/abs/2507.16746 +579,"From reactive to cognitive: brain-inspired spatial intelligence for + embodied agents",2,42,42.5,42,"Aug 24, 2025",3,18,https://arxiv.org/abs/2508.17198 +580,"Search-R1: Training LLMs to Reason and Leverage Search Engines with Reinforcement Learning",4,46,46.75,47,"Mar 12, 2025",35,3130,https://arxiv.org/abs/2503.09516 -465,Qwen2.5-Omni Technical Report,4,45,47,47,"Mar 26, 2025",165,3510,https://arxiv.org/abs/2503.20215 -466,"A Vision-Language-Action-Critic Model for Robotic Real-World - Reinforcement Learning",11,21,33.64,33,"Sep 19, 2025",16,145,https://arxiv.org/abs/2509.15937 -467,"Hyperspherical Latents Improve Continuous-Token Autoregressive - Generation",6,30,40,42,"Sep 29, 2025",6,43,https://arxiv.org/abs/2509.24335 -468,VLA-0: Building State-of-the-Art VLAs with Zero Modification,6,33,40,39,"Oct 15, 2025",8,103,https://arxiv.org/abs/2510.13054 -469,dParallel: Learnable Parallel Decoding for dLLMs,5,40,43,42,"Sep 30, 2025",17,16,https://arxiv.org/abs/2509.26488 -470,"MCPMark: A Benchmark for Stress-Testing Realistic and Comprehensive MCP - Use",8,32,36.62,35,"Sep 28, 2025",145,201,https://arxiv.org/abs/2509.24002 -471,"ThinkMorph: Emergent Properties in Multimodal Interleaved - Chain-of-Thought Reasoning",17,9,31,28,"Oct 30, 2025",78,87,https://arxiv.org/abs/2510.27492 -472,"Cognitive Kernel-Pro: A Framework for Deep Research Agents and Agent +581,Retrieval-Augmented Generation with Hierarchical Knowledge,1,34,34,34,"Mar 13, 2025",2,359,https://arxiv.org/abs/2503.10150 +582,"REASONING GYM: Reasoning Environments for Reinforcement Learning with + Verifiable Rewards",1,34,34,34,"May 30, 2025",72,1130,https://arxiv.org/abs/2505.24760 +583,Chem-R: Learning to Reason as a Chemist,2,35,42.5,42,"Oct 19, 2025",45,9,https://arxiv.org/abs/2510.16880 +584,Hierarchical Budget Policy Optimization for Adaptive Reasoning,1,35,35,35,"Jul 21, 2025",12,14,https://arxiv.org/abs/2507.15844 +585,Qwen2.5-Omni Technical Report,4,45,47,47,"Mar 26, 2025",165,3510,https://arxiv.org/abs/2503.20215 +586,Variational Reasoning for Language Models,2,36,43,43,"Sep 26, 2025",57,34,https://arxiv.org/abs/2509.22637 +587,OpenVoice: Versatile Instant Voice Cloning,6,46,48.33,49,"Dec 3, 2023",3,35300,https://arxiv.org/abs/2312.01479 +588,"BrowseComp-Plus: A More Fair and Transparent Evaluation Benchmark of + Deep-Research Agent",3,43,46,45,"Aug 8, 2025",35,49,https://arxiv.org/abs/2508.06600 +589,"CodePlot-CoT: Mathematical Visual Reasoning by Thinking with Code-Driven + Images",3,38,46,50,"Oct 13, 2025",12,22,https://arxiv.org/abs/2510.11718 +590,Continuous Thought Machines,2,42,43.5,43,"May 8, 2025",12,1490,https://arxiv.org/abs/2505.05522 +591,MMaDA: Multimodal Large Diffusion Language Models,1,37,37,37,"May 21, 2025",94,1230,https://arxiv.org/abs/2505.15809 +592,"Cognitive Kernel-Pro: A Framework for Deep Research Agents and Agent Foundation Models Training",4,46,47.5,47,"Aug 1, 2025",91,388,https://arxiv.org/abs/2508.00414 -473,"MMR1: Enhancing Multimodal Reasoning with Variance-Aware Sampling and - Open Resources",11,24,33.91,31,"Sep 25, 2025",90,190,https://arxiv.org/abs/2509.21268 -474,"MM-HELIX: Boosting Multimodal Long-Chain Reflective Reasoning with - Holistic Platform and Adaptive Hybrid Policy Optimization",7,33,38.43,37,"Oct 9, 2025",98,55,https://arxiv.org/abs/2510.08540 -475,"Generating an Image From 1,000 Words: Enhancing Text-to-Image With - Structured Captions",5,38,43.6,41,"Nov 10, 2025",19,248,https://arxiv.org/abs/2511.06876 -476,Next Visual Granularity Generation,7,33,38.86,35,"Aug 18, 2025",37,7,https://arxiv.org/abs/2508.12811 -477,"Drawing2CAD: Sequence-to-Sequence Learning for CAD Generation from - Vector Drawings",8,22,37.38,41,"Aug 26, 2025",2,50,https://arxiv.org/abs/2508.18733 -478,MGM-Omni: Scaling Omni LLMs to Personalized Long-Horizon Speech,5,40,44.2,42,"Sep 29, 2025",11,133,https://arxiv.org/abs/2509.25131 -479,Depth Anything with Any Prior,6,37,41.17,40,"May 15, 2025",12,389,https://arxiv.org/abs/2505.10565 -480,"FantasyTalking2: Timestep-Layer Adaptive Preference Optimization for - Audio-Driven Portrait Animation",13,22,33.15,31,"Aug 15, 2025",8,21,https://arxiv.org/abs/2508.11255 -481,"More Thought, Less Accuracy? On the Dual Nature of Reasoning in - Vision-Language Models",9,26,36.33,35,"Sep 30, 2025",56,44,https://arxiv.org/abs/2509.25848 -482,"UP2You: Fast Reconstruction of Yourself from Unconstrained Photo - Collections",7,22,39.29,45,"Sep 29, 2025",8,161,https://arxiv.org/abs/2509.24817 -483,"Reinforcement Learning Optimization for Large-Scale Learning: An - Efficient and User-Friendly Scaling Library",5,40,44.6,45,"Jun 6, 2025",7,2270,https://arxiv.org/abs/2506.06122 -484,Grounding Computer Use Agents on Human Demonstrations,4,47,49.25,50,"Nov 10, 2025",18,15,https://arxiv.org/abs/2511.07332 -485,Agentic Reinforced Policy Optimization,4,48,49.5,50,"Jul 26, 2025",154,691,https://arxiv.org/abs/2507.19849 -486,"LLaSO: A Foundational Framework for Reproducible Research in Large - Language and Speech Model",7,28,39.43,36,"Aug 21, 2025",4,79,https://arxiv.org/abs/2508.15418 -487,"Open-o3 Video: Grounded Video Reasoning with Explicit Spatio-Temporal - Evidence",9,22,36.44,34,"Oct 23, 2025",46,63,https://arxiv.org/abs/2510.20579 -488,DeCo: Frequency-Decoupled Pixel Diffusion for End-to-End Image Generation,10,25,35.4,33,"Nov 24, 2025",58,75,https://arxiv.org/abs/2511.19365 -489,"DeepScientist: Advancing Frontier-Pushing Scientific Findings - Progressively",11,26,34.64,32,"Sep 30, 2025",16,119,https://arxiv.org/abs/2509.26603 -490,"PartCrafter: Structured 3D Mesh Generation via Compositional Latent - Diffusion Transformers",7,23,39.71,40,"Jun 5, 2025",79,2180,https://arxiv.org/abs/2506.05573 -491,"Speed Always Wins: A Survey on Efficient Architectures for Large - Language Models",36,15,28.69,25,"Aug 13, 2025",51,281,https://arxiv.org/abs/2508.09834 -492,"M^3FinMeeting: A Multilingual, Multi-Sector, and Multi-Task Financial - Meeting Understanding Evaluation Dataset",22,12,30.41,28,"Jun 3, 2025",3,287,https://arxiv.org/abs/2506.02510 -493,"ScaleCUA: Scaling Open-Source Computer Use Agents with Cross-Platform - Data",78,10,27.24,29,"Sep 18, 2025",105,610,https://arxiv.org/abs/2509.15221 -494,RLP: Reinforcement as a Pretraining Objective,9,25,36.78,34,"Sep 26, 2025",32,149,https://arxiv.org/abs/2510.01265 -495,"GitTaskBench: A Benchmark for Code Agents Solving Real-World Tasks - Through Code Repository Leveraging",50,13,27.96,23,"Aug 26, 2025",3,183,https://arxiv.org/abs/2508.18993 -496,s3: You Don't Need That Much Data to Train a Search Agent via RL,5,41,45.8,47,"May 20, 2025",18,564,https://arxiv.org/abs/2505.14146 -497,"X-VLA: Soft-Prompted Transformer as Scalable Cross-Embodiment - Vision-Language-Action Model",7,32,40.14,42,"Oct 11, 2025",12,56,https://arxiv.org/abs/2510.10274 -498,Skywork R1V: Pioneering Multimodal Reasoning with Chain-of-Thought,8,23,38.5,42,"Apr 8, 2025",86,3080,https://arxiv.org/abs/2504.05599 -499,"HierSearch: A Hierarchical Enterprise Deep Search Framework Integrating - Local and Web Searches",8,33,38.5,37,"Aug 11, 2025",26,30,https://arxiv.org/abs/2508.08088 -500,SIM-CoT: Supervised Implicit Chain-of-Thought,8,33,38.5,34,"Sep 24, 2025",33,46,https://arxiv.org/abs/2509.20317 -501,SRUM: Fine-Grained Self-Rewarding for Unified Multimodal Models,9,15,37.22,44,"Oct 14, 2025",17,51,https://arxiv.org/abs/2510.12784 -502,"Teaching Pretrained Language Models to Think Deeper with Retrofitted - Recurrence",7,36,40.57,36,"Nov 10, 2025",11,23,https://arxiv.org/abs/2511.07384 -503,RynnVLA-001: Using Human Demonstrations to Improve Robot Manipulation,7,32,40.71,41,"Sep 18, 2025",20,208,https://arxiv.org/abs/2509.15212 -504,WideSearch: Benchmarking Agentic Broad Info-Seeking,7,37,40.86,41,"Aug 11, 2025",96,62,https://arxiv.org/abs/2508.07999 -505,"Qwen3 Embedding: Advancing Text Embedding and Reranking Through - Foundation Models",5,44,46.8,46,"Jun 5, 2025",72,1390,https://arxiv.org/abs/2506.05176 -506,Nav-R1: Reasoning and Navigation in Embodied Scenes,6,39,43.33,43,"Sep 13, 2025",4,21,https://arxiv.org/abs/2509.10884 -507,"Sparse VideoGen2: Accelerate Video Generation with Sparse Attention via - Semantic-Aware Permutation",8,35,39.12,38,"May 24, 2025",42,472,https://arxiv.org/abs/2505.18875 -508,"Sparse VideoGen: Accelerating Video Diffusion Transformers with - Spatial-Temporal Sparsity",8,34,39.12,38,"Feb 3, 2025",3,472,https://arxiv.org/abs/2502.01776 -509,"SonicMaster: Towards Controllable All-in-One Music Restoration and - Mastering",7,32,41.29,39,"Aug 5, 2025",1,74,https://arxiv.org/abs/2508.03448 -510,DINO-Foresight: Looking into the Future with DINO,7,39,41.29,41,"Dec 16, 2024",1,107,https://arxiv.org/abs/2412.11673 -511,Gradio: Hassle-Free Sharing and Testing of ML Models in the Wild,10,32,36.9,35,"Jun 6, 2019",1,40700,https://arxiv.org/abs/1906.02569 -512,From Editor to Dense Geometry Estimator,55,13,28.02,28,"Sep 4, 2025",87,156,https://arxiv.org/abs/2509.04338 -513,GEM: A Gym for Agentic LLMs,21,24,31.29,30,"Oct 1, 2025",79,284,https://arxiv.org/abs/2510.01051 -514,"Spatial-SSRL: Enhancing Spatial Understanding via Self-Supervised - Reinforcement Learning",11,25,36.09,34,"Oct 31, 2025",25,54,https://arxiv.org/abs/2510.27606 -515,"InternVLA-M1: A Spatially Guided Vision-Language-Action Framework for - Generalist Robot Policy",7,35,42,44,"Oct 15, 2025",13,167,https://arxiv.org/abs/2510.13778 -516,"ToolScope: An Agentic Framework for Vision-Guided and Long-Horizon Tool +593,"DrugReasoner: Interpretable Drug Approval Prediction with a + Reasoning-augmented Language Model",2,44,44,44,"Aug 26, 2025",10,2,https://arxiv.org/abs/2508.18579 +594,TikZero: Zero-Shot Text-Guided Graphics Program Synthesis,2,44,44,44,"Mar 14, 2025",3,1530,https://arxiv.org/abs/2503.11509 +595,VideoNSA: Native Sparse Attention Scales Video Understanding,2,42,44,44,"Oct 2, 2025",8,29,https://arxiv.org/abs/2510.02295 +596,OmniNWM: Omniscient Driving Navigation World Models,2,43,44,44,"Oct 21, 2025",6,48,https://arxiv.org/abs/2510.18313 +597,EVTAR: End-to-End Try on with Additional Unpaired Visual Reference,2,43,44,44,"Nov 2, 2025",4,17,https://arxiv.org/abs/2511.00956 +598,P1: Mastering Physics Olympiads with Reinforcement Learning,2,43,44,44,"Nov 17, 2025",106,45,https://arxiv.org/abs/2511.13612 +599,nablaNABLA: Neighborhood Adaptive Block-Level Attention,1,38,38,38,"Jul 17, 2025",85,10,https://arxiv.org/abs/2507.13546 +600,"Evolving Language Models without Labels: Majority Drives Selection, + Novelty Promotes Variation",3,41,46.67,49,"Sep 18, 2025",29,19,https://arxiv.org/abs/2509.15194 +601,YuE: Scaling Open Foundation Models for Long-Form Music Generation,2,40,44.5,44,"Mar 11, 2025",69,5520,https://arxiv.org/abs/2503.08638 +602,LightsOut: Diffusion-based Outpainting for Enhanced Lens Flare Removal,3,41,46.67,49,"Oct 17, 2025",20,10,https://arxiv.org/abs/2510.15868 +603,"ToolScope: An Agentic Framework for Vision-Guided and Long-Horizon Tool Use",5,47,48.4,49,"Oct 31, 2025",20,15,https://arxiv.org/abs/2510.27363 -517,Uni-MoE: Scaling Unified Multimodal LLMs with Mixture of Experts,11,32,36.27,35,"May 18, 2024",19,1010,https://arxiv.org/abs/2405.11273 -518,Droplet3D: Commonsense Priors from Videos Facilitate 3D Generation,8,35,40.5,39,"Aug 28, 2025",62,18,https://arxiv.org/abs/2508.20470 -519,Regression Language Models for Code,6,40,45.5,47,"Sep 30, 2025",13,257,https://arxiv.org/abs/2509.26476 -520,Robot Learning from a Physical World Model,23,23,31.09,30,"Nov 10, 2025",26,146,https://arxiv.org/abs/2511.07416 -521,"TOUCAN: Synthesizing 1.5M Tool-Agentic Data from Real-World MCP - Environments",9,30,39.11,41,"Oct 1, 2025",16,60,https://arxiv.org/abs/2510.01179 -522,"FlashVSR: Towards Real-Time Diffusion-Based Streaming Video - Super-Resolution",55,7,28.15,29,"Oct 14, 2025",37,910,https://arxiv.org/abs/2510.12747 -523,A decoder-only foundation model for time-series forecasting,6,44,46.17,46,"Oct 14, 2023",6,7060,https://arxiv.org/abs/2310.10688 -524,"FantasyPortrait: Enhancing Multi-Character Portrait Animation with - Expression-Augmented Diffusion Transformers",55,11,28.22,26,"Jul 17, 2025",24,435,https://arxiv.org/abs/2507.12956 -525,pi-Flow: Policy-Based Few-Step Generation via Imitation Distillation,13,26,35.38,33,"Oct 16, 2025",7,97,https://arxiv.org/abs/2510.14974 -526,DR Tulu: Reinforcement Learning with Evolving Rubrics for Deep Research,16,24,33.62,29,"Nov 24, 2025",48,404,https://arxiv.org/abs/2511.19399 -527,"Gated Attention for Large Language Models: Non-linearity, Sparsity, and - Attention-Sink-Free",10,31,38.2,36,"May 10, 2025",6,240,https://arxiv.org/abs/2505.06708 -528,"S^2-Guidance: Stochastic Self Guidance for Training-Free Enhancement of - Diffusion Models",25,15,30.96,23,"Aug 18, 2025",45,134,https://arxiv.org/abs/2508.12880 -529,Efficient Part-level 3D Object Generation via Dual Volume Packing,13,24,35.62,36,"Jun 11, 2025",8,701,https://arxiv.org/abs/2506.09980 -530,"Eliciting Fine-Tuned Transformer Capabilities via Inference-Time - Techniques",7,34,44,48,"Jun 9, 2025",8,2910,https://arxiv.org/abs/2506.08060 -531,"UniMoE-Audio: Unified Speech and Music Generation with Dynamic-Capacity - MoE",12,32,36.5,35,"Oct 15, 2025",62,1010,https://arxiv.org/abs/2510.13344 -532,Time-to-Move: Training-Free Motion Controlled Video Generation via Dual-Clock Denoising,21,22,32.05,30,"Nov 9, 2025",50,191,https://arxiv.org/abs/2511.08633 -533,"OpenVision 2: A Family of Generative Pretrained Visual Encoders for - Multimodal Learning",27,11,30.74,30,"Sep 1, 2025",27,356,https://arxiv.org/abs/2509.01644 -534,AnyUp: Universal Feature Upsampling,11,31,37.73,37,"Oct 14, 2025",10,268,https://arxiv.org/abs/2510.12764 -535,"Vision-Zero: Scalable VLM Self-Improvement via Strategic Gamified - Self-Play",9,29,40.44,44,"Sep 29, 2025",120,40,https://arxiv.org/abs/2509.25541 -536,"LIBERO-Plus: In-depth Robustness Analysis of Vision-Language-Action - Models",9,30,40.56,41,"Oct 15, 2025",40,46,https://arxiv.org/abs/2510.13626 -537,"OS-Sentinel: Towards Safety-Enhanced Mobile GUI Agents via Hybrid - Validation in Realistic Workflows",9,29,40.67,45,"Oct 28, 2025",69,31,https://arxiv.org/abs/2510.24411 -538,STARFlow-V: End-to-End Video Generative Modeling with Normalizing Flow,6,45,48,48,"Nov 25, 2025",16,40,https://arxiv.org/abs/2511.20462 -539,Beyond English: Toward Inclusive and Scalable Multilingual Machine Translation with LLMs,7,42,45,46,"Nov 10, 2025",31,92,https://arxiv.org/abs/2511.07003 -540,"MiniMax-M1: Scaling Test-Time Compute Efficiently with Lightning - Attention",8,34,42.75,44,"Jun 16, 2025",265,2890,https://arxiv.org/abs/2506.13585 -541,OpenVoice: Versatile Instant Voice Cloning,6,46,48.33,49,"Dec 3, 2023",3,35300,https://arxiv.org/abs/2312.01479 -542,MindSearch: Mimicking Human Minds Elicits Deep AI Searcher,7,43,45.29,46,"Jul 29, 2024",44,6580,https://arxiv.org/abs/2407.20183 -543,"MiniCPM-V 4.5: Cooking Efficient MLLMs via Architecture, Data, and - Training Recipe",21,17,32.43,31,"Sep 16, 2025",45,22000,https://arxiv.org/abs/2509.18154 -544,AWorld: Orchestrating the Training Recipe for Agentic AI,7,36,45.43,46,"Aug 28, 2025",37,692,https://arxiv.org/abs/2508.20404 -545,MiniCPM4: Ultra-Efficient LLMs on End Devices,7,41,45.57,47,"Jun 9, 2025",90,8300,https://arxiv.org/abs/2506.07900 -546,"Hulu-Med: A Transparent Generalist Model towards Holistic Medical - Vision-Language Understanding",30,13,30.57,35,"Oct 9, 2025",4,478,https://arxiv.org/abs/2510.08668 -547,"WebShaper: Agentically Data Synthesizing via Information-Seeking - Formalization",402,2,26.34,26,"Jul 20, 2025",60,17300,https://arxiv.org/abs/2507.15061 -548,"Time Is a Feature: Exploiting Temporal Dynamics in Diffusion Language - Models",13,25,36.62,34,"Aug 12, 2025",31,38,https://arxiv.org/abs/2508.09138 -549,DyPE: Dynamic Position Extrapolation for Ultra High Resolution Diffusion,14,26,35.86,31,"Oct 23, 2025",33,248,https://arxiv.org/abs/2510.20766 -550,"Cache-to-Cache: Direct Semantic Communication Between Large Language - Models",10,35,39.8,40,"Oct 3, 2025",95,239,https://arxiv.org/abs/2510.03215 -551,Hunyuan-MT Technical Report,14,26,36,35,"Sep 5, 2025",13,517,https://arxiv.org/abs/2509.05209 -552,DeepEyesV2: Toward Agentic Multimodal Model,8,39,43.5,45,"Nov 7, 2025",35,963,https://arxiv.org/abs/2511.05271 -553,"Genie Envisioner: A Unified World Foundation Platform for Robotic - Manipulation",24,19,31.96,29,"Aug 7, 2025",72,218,https://arxiv.org/abs/2508.05635 -554,"CoDiEmb: A Collaborative yet Distinct Framework for Unified - Representation Learning in Information Retrieval and Semantic Textual - Similarity",10,36,40.3,39,"Aug 15, 2025",3,125,https://arxiv.org/abs/2508.11442 -555,3D-R1: Enhancing Reasoning in 3D VLMs for Unified Scene Understanding,10,27,40.4,42,"Jul 31, 2025",15,307,https://arxiv.org/abs/2507.23478 -556,"SongBloom: Coherent Song Generation via Interleaved Autoregressive - Sketching and Diffusion Refinement",21,23,32.86,31,"Jun 9, 2025",6,604,https://arxiv.org/abs/2506.07634 -557,GigaBrain-0: A World Model-Powered Vision-Language-Action Model,13,30,37.15,36,"Oct 22, 2025",46,218,https://arxiv.org/abs/2510.19430 -558,"Verbalized Sampling: How to Mitigate Mode Collapse and Unlock LLM - Diversity",20,26,33.45,34,"Oct 1, 2025",15,219,https://arxiv.org/abs/2510.01171 -559,"VideoFrom3D: 3D Scene Video Generation via Complementary Image and Video - Diffusion Models",18,18,34.39,36,"Sep 22, 2025",25,83,https://arxiv.org/abs/2509.17985 -560,Self-Forcing++: Towards Minute-Scale High-Quality Video Generation,19,22,34,30,"Oct 2, 2025",86,145,https://arxiv.org/abs/2510.02283 -561,Defeating the Training-Inference Mismatch via FP16,9,33,42.89,47,"Oct 30, 2025",23,112,https://arxiv.org/abs/2510.26788 -562,RynnVLA-002: A Unified Vision-Language-Action and World Model,11,29,39.82,41,"Nov 21, 2025",24,669,https://arxiv.org/abs/2511.17502 -563,VoXtream: Full-Stream Text-to-Speech with Extremely Low Latency,24,24,32.42,31,"Sep 19, 2025",2,129,https://arxiv.org/abs/2509.15969 -564,ARE: Scaling Up Agent Environments and Evaluations,17,24,35.24,34,"Sep 21, 2025",29,257,https://arxiv.org/abs/2509.17158 -565,Symbolic Graphics Programming with Large Language Models,10,33,42,44,"Sep 5, 2025",37,18,https://arxiv.org/abs/2509.05208 -566,"ReCogDrive: A Reinforced Cognitive Framework for End-to-End Autonomous - Driving",8,39,46.12,47,"Jun 9, 2025",1,183,https://arxiv.org/abs/2506.08052 -567,DepthLM: Metric Depth From Vision Language Models,18,26,35.06,32,"Sep 29, 2025",5,130,https://arxiv.org/abs/2509.25413 -568,"The Tool Decathlon: Benchmarking Language Agents for Diverse, Realistic, - and Long-Horizon Task Execution",10,38,42.4,42,"Oct 29, 2025",42,99,https://arxiv.org/abs/2510.25726 -569,OmniGen2: Exploration to Advanced Multimodal Generation,10,32,42.7,42,"Jun 23, 2025",75,3860,https://arxiv.org/abs/2506.18871 -570,ReCode: Unify Plan and Action for Universal Granularity Control,29,16,31.76,30,"Oct 27, 2025",117,359,https://arxiv.org/abs/2510.23564 -571,SceneGen: Single-Image 3D Scene Generation in One Feedforward Pass,23,26,33.39,31,"Aug 21, 2025",18,66,https://arxiv.org/abs/2508.15769 -572,Energy-Based Transformers are Scalable Learners and Thinkers,8,44,47.25,47,"Jul 2, 2025",65,463,https://arxiv.org/abs/2507.02092 -573,Upsample Anything: A Simple and Hard to Beat Baseline for Feature Upsampling,9,38,45.22,46,"Nov 20, 2025",6,105,https://arxiv.org/abs/2511.16301 -574,"Skywork UniPic: Unified Autoregressive Modeling for Visual Understanding - and Generation",25,18,33,31,"Aug 5, 2025",59,727,https://arxiv.org/abs/2508.03320 -575,ReasonRank: Empowering Passage Ranking with Strong Reasoning Ability,13,28,39.54,40,"Aug 9, 2025",109,101,https://arxiv.org/abs/2508.07050 -576,"AniMaker: Automated Multi-Agent Animated Storytelling with MCTS-Driven - Clip Generation",8,46,48.12,48,"Jun 12, 2025",37,173,https://arxiv.org/abs/2506.10540 -577,"GLM-4.1V-Thinking: Towards Versatile Multimodal Reasoning with Scalable - Reinforcement Learning",53,10,29.38,29,"Jul 1, 2025",229,1500,https://arxiv.org/abs/2507.01006 -578,"SageAttention3: Microscaling FP4 Attention for Inference and An - Exploration of 8-Bit Training",12,26,41.75,43,"May 16, 2025",76,2450,https://arxiv.org/abs/2505.11594 -579,TexVerse: A Universe of 3D Objects with High-Resolution Textures,12,37,42.08,40,"Aug 14, 2025",13,180,https://arxiv.org/abs/2508.10868 -580,"OpenVision: A Fully-Open, Cost-Effective Family of Advanced Vision - Encoders for Multimodal Learning",17,26,37.47,39,"May 7, 2025",28,351,https://arxiv.org/abs/2505.04601 -581,RynnEC: Bringing MLLMs into Embodied World,62,6,29.16,29,"Aug 19, 2025",18,337,https://arxiv.org/abs/2508.14160 -582,MoDA: Multi-modal Diffusion Architecture for Talking Head Generation,9,46,47.89,48,"Jul 4, 2025",2,140,https://arxiv.org/abs/2507.03256 -583,"PosterGen: Aesthetic-Aware Paper-to-Poster Generation via Multi-Agent - LLMs",11,35,43.91,43,"Aug 24, 2025",15,105,https://arxiv.org/abs/2508.17188 -584,BitNet Distillation,13,35,41.62,41,"Oct 15, 2025",44,24300,https://arxiv.org/abs/2510.13998 -585,MediaPipe: A Framework for Building Perception Pipelines,11,39,44.45,45,"Jun 14, 2019",1,32100,https://arxiv.org/abs/1906.08172 -586,"BRIDGE - Building Reinforcement-Learning Depth-to-Image Data Generation - Engine for Monocular Depth Estimation",17,33,38.06,37,"Sep 29, 2025",13,106,https://arxiv.org/abs/2509.25077 -587,PyTorch FSDP: Experiences on Scaling Fully Sharded Data Parallel,104,10,27.97,30,"Apr 21, 2023",4,95500,https://arxiv.org/abs/2304.11277 -588,Is Diversity All You Need for Scalable Robotic Manipulation?,18,30,37.78,37,"Jul 8, 2025",20,2460,https://arxiv.org/abs/2507.06219 -589,Chronos-2: From Univariate to Universal Forecasting,32,23,32.75,31,"Oct 17, 2025",17,4190,https://arxiv.org/abs/2510.15821 -590,GR00T N1: An Open Foundation Model for Generalist Humanoid Robots,13,27,42.77,47,"Mar 18, 2025",4,4800,https://arxiv.org/abs/2503.14734 -591,Democratizing AI scientists using ToolUniverse,29,17,33.59,38,"Sep 27, 2025",38,455,https://arxiv.org/abs/2509.23426 -592,"FireRedTTS-2: Towards Long Conversational Speech Generation for Podcast - and Chatbot",24,21,35.21,35,"Sep 2, 2025",1,1180,https://arxiv.org/abs/2509.02020 -593,DanceGRPO: Unleashing GRPO on Visual Generation,11,37,46.45,47,"May 12, 2025",32,942,https://arxiv.org/abs/2505.07818 -594,EdgeTAM: On-Device Track Anything Model,16,31,40.38,40,"Jan 13, 2025",1,757,https://arxiv.org/abs/2501.07256 -595,Performance Prediction for Large Systems via Text-to-Text Regression,15,29,41.47,45,"Jun 26, 2025",6,255,https://arxiv.org/abs/2506.21718 -596,"SimpleTIR: End-to-End Reinforcement Learning for Multi-Turn - Tool-Integrated Reasoning",22,19,36.55,43,"Sep 2, 2025",76,250,https://arxiv.org/abs/2509.02479 -597,"A Comprehensive Survey of Self-Evolving AI Agents: A New Paradigm - Bridging Foundation Models and Lifelong Agentic Systems",191,5,27.24,29,"Aug 10, 2025",97,1140,https://arxiv.org/abs/2508.07407 -598,Self-Rewarding Vision-Language Model via Reasoning Decomposition,16,33,40.88,42,"Aug 27, 2025",77,79,https://arxiv.org/abs/2508.19652 -599,"CogVLA: Cognition-Aligned Vision-Language-Action Model via - Instruction-Driven Routing & Sparsification",18,31,39.22,39,"Aug 28, 2025",8,56,https://arxiv.org/abs/2508.21046 -600,"On the Generalization of SFT: A Reinforcement Learning Perspective with - Reward Rectification",29,26,34.34,35,"Aug 7, 2025",148,322,https://arxiv.org/abs/2508.05629 -601,LIMI: Less is More for Agency,26,28,35.31,35,"Sep 22, 2025",91,101,https://arxiv.org/abs/2509.17567 -602,GenCompositor: Generative Video Compositing with Diffusion Transformer,44,12,31.52,29,"Sep 2, 2025",24,122,https://arxiv.org/abs/2509.02460 -603,Training-Free Group Relative Policy Optimization,34,25,33.18,32,"Oct 9, 2025",42,3590,https://arxiv.org/abs/2510.08191 -604,Discrete Diffusion in Large Language and Multimodal Models: A Survey,13,36,44.92,46,"Jun 16, 2025",43,267,https://arxiv.org/abs/2506.13759 -605,"LLaVA-OneVision-1.5: Fully Open Framework for Democratized Multimodal - Training",16,33,41.5,41,"Sep 28, 2025",40,508,https://arxiv.org/abs/2509.23661 -606,Real-Time Object Detection Meets DINOv3,37,14,32.84,37,"Sep 25, 2025",10,684,https://arxiv.org/abs/2509.20787 -607,PyTorch Distributed: Experiences on Accelerating Data Parallel Training,105,9,28.43,31,"Jun 28, 2020",3,95500,https://arxiv.org/abs/2006.15704 -608,MolmoAct: Action Reasoning Models that can Reason in Space,22,28,37.68,37,"Aug 11, 2025",42,163,https://arxiv.org/abs/2508.07917 -609,"Language Model Council: Benchmarking Foundation Models on Highly - Subjective Tasks by Consensus",14,35,44.57,46,"Jun 12, 2024",6,180,https://arxiv.org/abs/2406.08598 -610,"AReaL: A Large-Scale Asynchronous Reinforcement Learning System for - Language Reasoning",59,18,30.47,29,"May 30, 2025",28,2660,https://arxiv.org/abs/2505.24298 -611,"LucidFlux: Caption-Free Universal Image Restoration via a Large-Scale - Diffusion Transformer",37,24,33.35,32,"Sep 26, 2025",21,476,https://arxiv.org/abs/2509.22414 -612,Zep: A Temporal Knowledge Graph Architecture for Agent Memory,95,11,28.88,31,"Jan 20, 2025",6,20600,https://arxiv.org/abs/2501.13956 -613,"Story2Board: A Training-Free Approach for Expressive Storyboard - Generation",45,14,32.31,34,"Aug 13, 2025",61,119,https://arxiv.org/abs/2508.09983 -614,"AMFT: Aligning LLM Reasoners by Meta-Learning the Optimal - Imitation-Exploration Balance",16,31,43.88,45,"Aug 9, 2025",2,163,https://arxiv.org/abs/2508.06944 -615,Reinforcement Learning in Vision: A Survey,41,25,33.49,33,"Aug 11, 2025",27,184,https://arxiv.org/abs/2508.08189 -616,"JavisDiT: Joint Audio-Video Diffusion Transformer with Hierarchical - Spatio-Temporal Prior Synchronization",16,35,45.88,47,"Mar 30, 2025",57,265,https://arxiv.org/abs/2503.23377 -617,"VLA-Adapter: An Effective Paradigm for Tiny-Scale Vision-Language-Action - Model",131,7,28.43,29,"Sep 11, 2025",235,1590,https://arxiv.org/abs/2509.09372 -618,Explain Before You Answer: A Survey on Compositional Visual Reasoning,19,35,42.89,44,"Aug 24, 2025",4,279,https://arxiv.org/abs/2508.17298 -619,"Stable Video Infinity: Infinite-Length Video Generation with Error - Recycling",19,34,43.05,45,"Oct 10, 2025",14,388,https://arxiv.org/abs/2510.09212 -620,UI-S1: Advancing GUI Automation via Semi-online Reinforcement Learning,67,15,30.96,31,"Sep 15, 2025",47,6440,https://arxiv.org/abs/2509.11543 -621,Puppeteer: Rig and Animate Your 3D Models,31,27,37.1,37,"Aug 14, 2025",30,178,https://arxiv.org/abs/2508.10898 -622,OmniTry: Virtual Try-On Anything without Masks,41,13,34.39,34,"Aug 19, 2025",16,166,https://arxiv.org/abs/2508.13632 -623,Kwai Keye-VL Technical Report,40,20,34.85,33,"Jul 2, 2025",130,623,https://arxiv.org/abs/2507.01949 -624,"Self Forcing: Bridging the Train-Test Gap in Autoregressive Video - Diffusion",17,41,47.18,49,"Jun 9, 2025",27,2420,https://arxiv.org/abs/2506.08009 -625,Marco-Voice Technical Report,53,25,32.89,33,"Aug 4, 2025",15,332,https://arxiv.org/abs/2508.02038 -626,DeepPHY: Benchmarking Agentic VLMs on Physical Reasoning,26,25,40.15,42,"Aug 7, 2025",62,143,https://arxiv.org/abs/2508.05405 -627,ViSTA-SLAM: Visual SLAM with Symmetric Two-view Association,38,16,35.74,37,"Sep 1, 2025",6,134,https://arxiv.org/abs/2509.01584 -628,"Skyfall-GS: Synthesizing Immersive 3D Urban Scenes from Satellite - Imagery",37,9,36.08,41,"Oct 17, 2025",45,576,https://arxiv.org/abs/2510.15869 -629,Aligning Multimodal LLM with Human Preference: A Survey,20,31,45.95,48,"Mar 18, 2025",26,16300,https://arxiv.org/abs/2503.14504 -630,"Pass@k Training for Adaptively Balancing Exploration and Exploitation of - Large Reasoning Models",21,40,45.43,45,"Aug 14, 2025",21,47,https://arxiv.org/abs/2508.10751 -631,"Paper2Code: Automating Code Generation from Scientific Papers in Machine - Learning",32,27,40.34,40,"Apr 24, 2025",120,3730,https://arxiv.org/abs/2504.17192 -632,"MonkeyOCR: Document Parsing with a Structure-Recognition-Relation - Triplet Paradigm",66,16,33.12,31,"Jun 5, 2025",2,6000,https://arxiv.org/abs/2506.05218 -633,R-Zero: Self-Evolving Reasoning LLM from Zero Data,39,28,38.41,38,"Aug 7, 2025",123,559,https://arxiv.org/abs/2508.05004 -634,The Unreasonable Effectiveness of Scaling Agents for Computer Use,138,2,29.52,34,"Oct 2, 2025",24,8450,https://arxiv.org/abs/2510.02250 -635,"LivePortrait: Efficient Portrait Animation with Stitching and - Retargeting Control",29,39,44.48,44,"Jul 3, 2024",3,16900,https://arxiv.org/abs/2407.03168 -636,"AWorld: Dynamic Multi-Agent System with Stable Maneuvering for Robust - GAIA Problem Solving",31,34,43.71,45,"Aug 13, 2025",32,694,https://arxiv.org/abs/2508.09889 -637,MemOS: A Memory OS for AI System,36,13,41.33,45,"Jul 4, 2025",155,3110,https://arxiv.org/abs/2507.03724 -638,JoyAgent-JDGenie: Technical Report on the GAIA,61,14,35.08,38,"Oct 1, 2025",3,11100,https://arxiv.org/abs/2510.00510 -639,"ChronoEdit: Towards Temporal Reasoning for Image Editing and World - Simulation",45,23,38.96,42,"Oct 5, 2025",14,578,https://arxiv.org/abs/2510.04290 -640,"A.S.E: A Repository-Level Benchmark for Evaluating Security in - AI-Generated Code",62,7,35.66,37,"Aug 25, 2025",340,709,https://arxiv.org/abs/2508.18106 -641,Arch-Router: Aligning LLM Routing with Human Preferences,60,21,36.28,37,"Jun 19, 2025",17,4170,https://arxiv.org/abs/2506.16655 -642,"Agent S2: A Compositional Generalist-Specialist Framework for Computer - Use Agents",145,2,30.46,35,"Apr 1, 2025",26,8450,https://arxiv.org/abs/2504.00906 -643,UI-TARS: Pioneering Automated GUI Interaction with Native Agents,208,4,29.12,28,"Jan 21, 2025",65,7830,https://arxiv.org/abs/2501.12326 -644,"Packing Input Frame Context in Next-Frame Prediction Models for Video - Generation",47,15,41.34,44,"Apr 17, 2025",52,15800,https://arxiv.org/abs/2504.12626 -645,"REINFORCE++: A Simple and Efficient Approach for Aligning Large Language - Models",41,26,44.05,45,"Jan 4, 2025",102,8130,https://arxiv.org/abs/2501.03262 -646,Versatile Framework for Song Generation with Prompt-based Control,44,34,43.5,45,"Apr 27, 2025",6,201,https://arxiv.org/abs/2504.19062 -647,Transformer Explainer: Interactive Learning of Text-Generative Models,81,21,35.89,35,"Aug 8, 2024",172,5960,https://arxiv.org/abs/2408.04619 -648,3D Gaussian Splatting for Real-Time Radiance Field Rendering,41,41,46.12,46,"Aug 8, 2023",192,19600,https://arxiv.org/abs/2308.04079 -649,"HunyuanWorld 1.0: Generating Immersive, Explorable, and Interactive 3D - Worlds from Words or Pixels",76,22,37.25,38,"Jul 29, 2025",126,2170,https://arxiv.org/abs/2507.21809 -650,"GLM-4.5: Agentic, Reasoning, and Coding (ARC) Foundation Models",243,11,30.1,30,"Aug 8, 2025",185,3000,https://arxiv.org/abs/2508.06471 -651,MixGRPO: Unlocking Flow-based GRPO Efficiency with Mixed ODE-SDE,170,12,32.78,34,"Jul 29, 2025",15,1010,https://arxiv.org/abs/2507.21802 -652,"ComfyUI-Copilot: An Intelligent Assistant for Automated Workflow - Development",172,6,32.85,36,"Jun 5, 2025",79,3660,https://arxiv.org/abs/2506.05010 -653,"InternVL3.5: Advancing Open-Source Multimodal Models in Versatility, - Reasoning, and Efficiency",109,12,37.03,40,"Aug 25, 2025",191,9240,https://arxiv.org/abs/2508.18265 -654,RLinf-VLA: A Unified and Efficient Framework for VLA+RL Training,86,25,41.38,42,"Oct 8, 2025",38,1450,https://arxiv.org/abs/2510.06710 -655,4DNeX: Feed-Forward 4D Generative Modeling Made Easy,178,2,33.96,37,"Aug 18, 2025",61,751,https://arxiv.org/abs/2508.13154 -656,Training Video Foundation Models with NVIDIA NeMo,87,19,42.82,44,"Mar 17, 2025",7,15700,https://arxiv.org/abs/2503.12964 -657,"Large Language Model Agent: A Survey on Methodology, Applications and - Challenges",82,35,43.88,45,"Mar 27, 2025",82,1710,https://arxiv.org/abs/2503.21460 -658,MIRIX: Multi-Agent Memory System for LLM-Based Agents,272,5,31.7,35,"Jul 10, 2025",79,3380,https://arxiv.org/abs/2507.07957 -659,"Killing Two Birds with One Stone:Efficient and Robust Training of Face - Recognition CNNs by Partial FC",92,26,43.16,45,"Mar 28, 2022",3,26700,https://arxiv.org/abs/2203.15565 -660,A Survey of Context Engineering for Large Language Models,151,1,39.59,40,"Jul 17, 2025",256,2330,https://arxiv.org/abs/2507.13334 -661,"Easy Dataset: A Unified and Extensible Framework for Synthesizing LLM - Fine-Tuning Data from Unstructured Documents",345,13,32.46,32,"Jul 5, 2025",51,12100,https://arxiv.org/abs/2507.04009 -662,Qwen3 Technical Report,266,2,34.64,34,"May 14, 2025",316,25500,https://arxiv.org/abs/2505.09388 -663,VGGT: Visual Geometry Grounded Transformer,232,12,37.63,39,"Mar 14, 2025",33,11700,https://arxiv.org/abs/2503.11651 +604,Drag-and-Drop LLMs: Zero-Shot Prompt-to-Weights,1,39,39,39,"Jun 19, 2025",122,58,https://arxiv.org/abs/2506.16406 +605,"DCReg: Decoupled Characterization for Efficient Degenerate LiDAR + Registration",3,41,47,50,"Sep 8, 2025",1,58,https://arxiv.org/abs/2509.06285 +606,"VoiceAssistant-Eval: Benchmarking AI Assistants across Listening, + Speaking, and Viewing",1,39,39,39,"Sep 26, 2025",19,4,https://arxiv.org/abs/2509.22651 +607,"Beyond the Exploration-Exploitation Trade-off: A Hidden State Approach + for LLM Reasoning in RLVR",3,46,47,47,"Sep 28, 2025",38,7,https://arxiv.org/abs/2509.23808 +608,"RAPO++: Cross-Stage Prompt Optimization for Text-to-Video Generation via + Data Alignment and Test-Time Scaling",2,40,45,45,"Oct 23, 2025",11,104,https://arxiv.org/abs/2510.20206 +609,Scaling Language-Centric Omnimodal Representation Learning,1,40,40,40,"Oct 13, 2025",52,11,https://arxiv.org/abs/2510.11693 +610,"ImagerySearch: Adaptive Test-Time Search for Video Generation Beyond + Semantic Dependency Constraints",2,43,45.5,45,"Oct 16, 2025",47,47,https://arxiv.org/abs/2510.14847 +611,"Automatic Synthetic Data and Fine-grained Adaptive Feature Alignment for + Composed Person Retrieval",2,46,46,46,"Nov 25, 2023",1,49,https://arxiv.org/abs/2311.16515 +612,Character Mixing for Video Generation,2,42,46,46,"Oct 6, 2025",5,52,https://arxiv.org/abs/2510.05093 +613,"TIR-Bench: A Comprehensive Benchmark for Agentic Thinking-with-Images + Reasoning",2,46,46,46,"Nov 3, 2025",12,14,https://arxiv.org/abs/2511.01833 +614,"Trinity-RFT: A General-Purpose and Unified Framework for Reinforcement + Fine-Tuning of Large Language Models",3,48,48,48,"May 23, 2025",9,269,https://arxiv.org/abs/2505.17826 +615,Trainable Dynamic Mask Sparse Attention,3,46,48,49,"Aug 4, 2025",17,366,https://arxiv.org/abs/2508.02124 +616,"FlashAdventure: A Benchmark for GUI Agents Solving Full Story Arcs in + Diverse Adventure Games",2,44,46.5,46,"Sep 1, 2025",8,7,https://arxiv.org/abs/2509.01052 +617,HoloScene: Simulation-Ready Interactive 3D Worlds from a Single Video,1,42,42,42,"Oct 7, 2025",5,5,https://arxiv.org/abs/2510.05560 +618,"SIU3R: Simultaneous Scene Understanding and 3D Reconstruction Beyond + Feature Alignment",2,46,46.5,46,"Jul 3, 2025",2,80,https://arxiv.org/abs/2507.02705 +619,"Learning on the Job: Test-Time Curricula for Targeted Reinforcement + Learning",1,43,43,43,"Oct 6, 2025",1,3,https://arxiv.org/abs/2510.04786 +620,"Discrete Diffusion Models with MLLMs for Unified Medical Multimodal + Generation",1,43,43,43,"Oct 7, 2025",4,4,https://arxiv.org/abs/2510.06131 +621,Reasoning in Space via Grounding in the World,3,45,48.33,50,"Oct 15, 2025",13,16,https://arxiv.org/abs/2510.13800 +622,First Frame Is the Place to Go for Video Content Customization,2,45,47,47,"Nov 19, 2025",45,36,https://arxiv.org/abs/2511.15700 +623,Thinking-while-Generating: Interleaving Textual Reasoning throughout Visual Generation,2,44,47,47,"Nov 20, 2025",12,43,https://arxiv.org/abs/2511.16671 +624,Computer-Use Agents as Judges for Generative User Interface,2,45,47,47,"Nov 19, 2025",48,24,https://arxiv.org/abs/2511.15567 +625,"Graph2Eval: Automatic Multimodal Task Generation for Agents via + Knowledge Graphs",1,44,44,44,"Oct 1, 2025",1,5,https://arxiv.org/abs/2510.00507 +626,"MathCanvas: Intrinsic Visual Chain-of-Thought for Multimodal + Mathematical Reasoning",2,47,47.5,47,"Oct 16, 2025",11,5,https://arxiv.org/abs/2510.14958 +627,Grounding Computer Use Agents on Human Demonstrations,4,47,49.25,50,"Nov 10, 2025",18,15,https://arxiv.org/abs/2511.07332 +628,"Feedback-Driven Tool-Use Improvements in Large Language Models via + Automated Build Environments",2,46,48,48,"Aug 12, 2025",12,17,https://arxiv.org/abs/2508.08791 +629,Agentic Reinforced Policy Optimization,4,48,49.5,50,"Jul 26, 2025",154,691,https://arxiv.org/abs/2507.19849 +630,"SAM2Act: Integrating Visual Foundation Model with A Memory Architecture + for Robotic Manipulation",3,48,49,49,"Jan 30, 2025",2,131,https://arxiv.org/abs/2501.18564 +631,Generative AI for Autonomous Driving: Frontiers and Opportunities,2,48,48,48,"May 13, 2025",1,176,https://arxiv.org/abs/2505.08854 +632,CHARM: Control-point-based 3D Anime Hairstyle Auto-Regressive Modeling,2,46,48,48,"Sep 25, 2025",11,9,https://arxiv.org/abs/2509.21114 +633,DeepPrune: Parallel Scaling without Inter-trace Redundancy,1,45,45,45,"Oct 9, 2025",21,9,https://arxiv.org/abs/2510.08483 +634,UI2Code^N: A Visual Language Model for Test-Time Scalable Interactive UI-to-Code Generation,2,47,48,48,"Nov 11, 2025",28,18,https://arxiv.org/abs/2511.08195 +635,WorldVLA: Towards Autoregressive Action World Model,1,45,45,45,"Jun 26, 2025",40,669,https://arxiv.org/abs/2506.21539 +636,"Gaze into the Heart: A Multi-View Video Dataset for rPPG and Health + Biomarkers Estimation",1,46,46,46,"Aug 25, 2025",11,3,https://arxiv.org/abs/2508.17924 +637,Uncertainty-Aware Testing-Time Optimization for 3D Human Pose Estimation,1,46,46,46,"Feb 4, 2024",1,5,https://arxiv.org/abs/2402.02339 +638,"AssetOpsBench: Benchmarking AI Agents for Task Automation in Industrial + Asset Operations and Maintenance",1,46,46,46,"Jun 4, 2025",14,252,https://arxiv.org/abs/2506.03828 +639,"Entropy Regularizing Activation: Boosting Continuous Control, Large + Language Models, and Image Classification with Activation as Entropy + Constraints",1,46,46,46,"Oct 9, 2025",5,7,https://arxiv.org/abs/2510.08549 +640,"UniPixel: Unified Object Referring and Segmentation for Pixel-Level + Visual Reasoning",1,47,47,47,"Sep 22, 2025",3,40,https://arxiv.org/abs/2509.18094 +641,"GeoSVR: Taming Sparse Voxels for Geometrically Accurate Surface + Reconstruction",1,47,47,47,"Sep 22, 2025",2,65,https://arxiv.org/abs/2509.18090 +642,PIPer: On-Device Environment Setup via Online Reinforcement Learning,1,47,47,47,"Sep 29, 2025",23,4,https://arxiv.org/abs/2509.25455 +643,"Factuality Matters: When Image Generation and Editing Meet Structured + Visuals",1,47,47,47,"Oct 6, 2025",12,4,https://arxiv.org/abs/2510.05091 +644,"BIRD-INTERACT: Re-imagining Text-to-SQL Evaluation for Large Language + Models via Lens of Dynamic Interactions",1,47,47,47,"Oct 6, 2025",4,260,https://arxiv.org/abs/2510.05318 +645,Instant4D: 4D Gaussian Splatting in Minutes,1,47,47,47,"Oct 1, 2025",5,58,https://arxiv.org/abs/2510.01119 +646,"ASTRA: Autonomous Spatial-Temporal Red-teaming for AI Software + Assistants",2,49,49.5,49,"Aug 5, 2025",8,37,https://arxiv.org/abs/2508.03936 +647,"InfiGUI-G1: Advancing GUI Grounding with Adaptive Exploration Policy + Optimization",2,49,49.5,49,"Aug 7, 2025",25,56,https://arxiv.org/abs/2508.05731 +648,"NER Retriever: Zero-Shot Named Entity Retrieval with Type-Aware + Embeddings",3,50,50,50,"Sep 4, 2025",19,19,https://arxiv.org/abs/2509.04011 +649,Where LLM Agents Fail and How They can Learn From Failures,1,48,48,48,"Sep 29, 2025",11,8,https://arxiv.org/abs/2509.25370 +650,"MarS: a Financial Market Simulation Engine Powered by Generative + Foundation Model",1,48,48,48,"Sep 4, 2024",1,1560,https://arxiv.org/abs/2409.07486 +651,"Audio-visual Controlled Video Diffusion with Masked Selective State + Spaces Modeling for Natural Talking Head Generation",1,49,49,49,"Apr 3, 2025",49,344,https://arxiv.org/abs/2504.02542 +652,"SynParaSpeech: Automated Synthesis of Paralinguistic Datasets for Speech + Generation and Understanding",1,49,49,49,"Sep 18, 2025",1,37,https://arxiv.org/abs/2509.14946 +653,"Beyond Log Likelihood: Probability-Based Objectives for Supervised + Fine-Tuning across the Model Capability Continuum",1,49,49,49,"Oct 1, 2025",7,3,https://arxiv.org/abs/2510.00526 +654,"SLA: Beyond Sparsity in Diffusion Transformers via Fine-Tunable + Sparse-Linear Attention",2,50,50,50,"Sep 28, 2025",107,38,https://arxiv.org/abs/2509.24006 +655,Scaling Large-Language-Model-based Multi-Agent Collaboration,2,50,50,50,"Jun 11, 2024",3,27600,https://arxiv.org/abs/2406.07155 +656,A Survey of Vibe Coding with Large Language Models,1,49,49,49,"Oct 14, 2025",21,7,https://arxiv.org/abs/2510.12399 +657,Gaussian Splatting with Discretized SDF for Relightable Assets,1,50,50,50,"Jul 21, 2025",19,61,https://arxiv.org/abs/2507.15629 +658,Flow-GRPO: Training Flow Matching Models via Online RL,1,50,50,50,"May 8, 2025",83,1120,https://arxiv.org/abs/2505.05470 +659,One-Minute Video Generation with Test-Time Training,1,50,50,50,"Apr 7, 2025",109,2090,https://arxiv.org/abs/2504.05298 +660,LongCat-Flash-Thinking Technical Report,1,50,50,50,"Sep 23, 2025",2,195,https://arxiv.org/abs/2509.18883 +661,"Efficient Audio-Visual Speech Separation with Discrete Lip Semantics and + Multi-Scale Global-Local Attention",1,50,50,50,"Sep 28, 2025",13,22,https://arxiv.org/abs/2509.23610 +662,EVODiff: Entropy-aware Variance Optimized Diffusion Inference,1,50,50,50,"Sep 30, 2025",1,9,https://arxiv.org/abs/2509.26096 +663,"GLiNER2: An Efficient Multi-Task Information Extraction System with + Schema-Driven Interface",1,50,50,50,"Jul 24, 2025",28,186,https://arxiv.org/abs/2507.18546