Xiaomi introduced MiMo-7B: A Compact Language Model that Outperforms Larger Models in Mathematical and Code Reasoning through Rigorous Pre-Training and Reinforcement Learning
Source: MarkTechPost With rising demand for AI systems that can handle tasks involving multi-step logic, mathematical proofs, and...
Building the Internet of Agents: A Technical Dive into AI Agent Protocols and Their Role in Scalable Intelligence Systems
Source: MarkTechPost As large language model (LLM) agents gain traction across enterprise and research ecosystems, a foundational gap...
DeepSeek-AI Released DeepSeek-Prover-V2: An Open-Source Large Language Model Designed for Formal Theorem, Proving through Subgoal Decomposition and Reinforcement Learning
Source: MarkTechPost Formal mathematical reasoning has evolved into a specialized subfield of artificial intelligence that requires strict logical...
Meta AI Introduces First Version of Its Llama 4-Powered AI App: A Standalone AI Assistant to Rival ChatGPT
Source: MarkTechPost Meta has officially entered the standalone AI assistant arena with the launch of its new Meta...
Microsoft AI Released Phi-4-Reasoning: A 14B Parameter Open-Weight Reasoning Model that Achieves Strong Performance on Complex Reasoning Tasks
Source: MarkTechPost Despite notable advancements in large language models (LLMs), effective performance on reasoning-intensive tasks—such as mathematical problem...
Meta AI Introduces ReasonIR-8B: A Reasoning-Focused Retriever Optimized for Efficiency and RAG Performance
Source: MarkTechPost Addressing the Challenges in Reasoning-Intensive Retrieval Despite notable progress in retrieval-augmented generation (RAG) systems, retrieving relevant...
Multimodal AI on Developer GPUs: Alibaba Releases Qwen2.5-Omni-3B with 50% Lower VRAM Usage and Nearly-7B Model Performance
Source: MarkTechPost Multimodal foundation models have shown substantial promise in enabling systems that can reason across text, images,...
Mem0: A Scalable Memory Architecture Enabling Persistent, Structured Recall for Long-Term AI Conversations Across Sessions
Source: MarkTechPost Large language models can generate fluent responses, emulate tone, and even follow complex instructions; however, they...
Exploring the Sparse Frontier: How Researchers from Edinburgh, Cohere, and Meta Are Rethinking Attention Mechanisms for Long-Context LLMs
Source: MarkTechPost Sparse attention is emerging as a compelling approach to improve the ability of Transformer-based LLMs to...
Diagnosing and Self- Correcting LLM Agent Failures: A Technical Deep Dive into τ-Bench Findings with Atla’s EvalToolbox
Source: MarkTechPost Deploying large language model (LLM)-based agents in production settings often reveals critical reliability issues. Accurately identifying...