LLMs Can Learn Complex Math from Just One Example: Researchers from University of Washington, Microsoft, and USC Unlock the Power of 1-Shot Reinforcement Learning with Verifiable Reward
Source: MarkTechPost Recent advancements in LLMs such as OpenAI-o1, DeepSeek-R1, and Kimi-1.5 have significantly improved their performance on...

Novel AI model inspired by neural dynamics from the brain
Source: MIT News – Artificial intelligence Researchers from MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL) have developed...

JetBrains Open Sources Mellum: A Developer-Centric Language Model for Code-Related Tasks
Source: MarkTechPost JetBrains has officially open-sourced Mellum, a purpose-built 4-billion-parameter language model tailored for software development tasks. Developed...
Training LLM Agents Just Got More Stable: Researchers Introduce StarPO-S and RAGEN to Tackle Multi-Turn Reasoning and Collapse in Reinforcement Learning
Source: MarkTechPost Large language models (LLMs) face significant challenges when trained as autonomous agents in interactive environments. Unlike...
Xiaomi introduced MiMo-7B: A Compact Language Model that Outperforms Larger Models in Mathematical and Code Reasoning through Rigorous Pre-Training and Reinforcement Learning
Source: MarkTechPost With rising demand for AI systems that can handle tasks involving multi-step logic, mathematical proofs, and...
Building the Internet of Agents: A Technical Dive into AI Agent Protocols and Their Role in Scalable Intelligence Systems
Source: MarkTechPost As large language model (LLM) agents gain traction across enterprise and research ecosystems, a foundational gap...
DeepSeek-AI Released DeepSeek-Prover-V2: An Open-Source Large Language Model Designed for Formal Theorem, Proving through Subgoal Decomposition and Reinforcement Learning
Source: MarkTechPost Formal mathematical reasoning has evolved into a specialized subfield of artificial intelligence that requires strict logical...
Microsoft AI Released Phi-4-Reasoning: A 14B Parameter Open-Weight Reasoning Model that Achieves Strong Performance on Complex Reasoning Tasks
Source: MarkTechPost Despite notable advancements in large language models (LLMs), effective performance on reasoning-intensive tasks—such as mathematical problem...
Meta AI Introduces ReasonIR-8B: A Reasoning-Focused Retriever Optimized for Efficiency and RAG Performance
Source: MarkTechPost Addressing the Challenges in Reasoning-Intensive Retrieval Despite notable progress in retrieval-augmented generation (RAG) systems, retrieving relevant...

Making AI models more trustworthy for high-stakes settings
Source: MIT News – Artificial intelligence The ambiguity in medical imaging can present major challenges for clinicians who...