Open-o3 Video: Grounded Video Reasoning with Explicit Spatio-Temporal Evidence Paper • 2510.20579 • Published 26 days ago • 54
Grasp Any Region: Towards Precise, Contextual Pixel Understanding for Multimodal LLMs Paper • 2510.18876 • Published 28 days ago • 35
SAILViT: Towards Robust and Generalizable Visual Backbones for MLLMs via Gradual Feature Refinement Paper • 2507.01643 • Published Jul 2 • 1
Traceable Evidence Enhanced Visual Grounded Reasoning: Evaluation and Methodology Paper • 2507.07999 • Published Jul 10 • 49
Traceable Evidence Enhanced Visual Grounded Reasoning: Evaluation and Methodology Paper • 2507.07999 • Published Jul 10 • 49
The Scalability of Simplicity: Empirical Analysis of Vision-Language Learning with a Single Transformer Paper • 2504.10462 • Published Apr 14 • 15
MiniMax-M1: Scaling Test-Time Compute Efficiently with Lightning Attention Paper • 2506.13585 • Published Jun 16 • 271
Seeing the Image: Prioritizing Visual Correlation by Contrastive Alignment Paper • 2405.17871 • Published May 28, 2024 • 1
World to Code: Multi-modal Data Generation via Self-Instructed Compositional Captioning and Filtering Paper • 2409.20424 • Published Sep 30, 2024
The Scalability of Simplicity: Empirical Analysis of Vision-Language Learning with a Single Transformer Paper • 2504.10462 • Published Apr 14 • 15
Unveiling the Tapestry of Consistency in Large Vision-Language Models Paper • 2405.14156 • Published May 23, 2024