Do NOT Think That Much for 2+3=? On the Overthinking of o1-Like LLMs Paper • 2412.21187 • Published 1 day ago • 8 • 2
Slow Perception: Let's Perceive Geometric Figures Step-by-step Paper • 2412.20631 • Published 2 days ago • 7 • 2
TangoFlux: Super Fast and Faithful Text to Audio Generation with Flow Matching and Clap-Ranked Preference Optimization Paper • 2412.21037 • Published 1 day ago • 16 • 4
Next Token Prediction Towards Multimodal Intelligence: A Comprehensive Survey Paper • 2412.18619 • Published 16 days ago • 39 • 2
HuatuoGPT-o1, Towards Medical Complex Reasoning with LLMs Paper • 2412.18925 • Published 7 days ago • 75 • 6
Molar: Multimodal LLMs with Collaborative Filtering Alignment for Enhanced Sequential Recommendation Paper • 2412.18176 • Published 8 days ago • 13 • 2
MMFactory: A Universal Solution Search Engine for Vision-Language Tasks Paper • 2412.18072 • Published 8 days ago • 13 • 2
WavePulse: Real-time Content Analytics of Radio Livestreams Paper • 2412.17998 • Published 8 days ago • 8 • 2
Video-Panda: Parameter-efficient Alignment for Encoder-free Video-Language Models Paper • 2412.18609 • Published 7 days ago • 11 • 2
Mulberry: Empowering MLLM with o1-like Reasoning and Reflection via Collective Monte Carlo Tree Search Paper • 2412.18319 • Published 8 days ago • 28 • 2
DRT-o1: Optimized Deep Reasoning Translation via Long Chain-of-Thought Paper • 2412.17498 • Published 9 days ago • 17 • 4
view post Post 3144 Google drops Gemini 2.0 Flash Thinkinga new experimental model that unlocks stronger reasoning capabilities and shows its thoughts. The model plans (with thoughts visible), can solve complex problems with Flash speeds, and morenow available in anychat, try it out: akhaliq/anychat See translation 🚀 6 6 🔥 4 4 👀 1 1 + Reply