Showing posts with label Math Reasoning. Show all posts
Showing posts with label Math Reasoning. Show all posts

11.9.25

Parallel-R1: Teaching LLMs to reason from multiple angles—permanently

 Modern large language models (LLMs) often reason sequentially—one thought chain at a time. Parallel thinking, in contrast, involves spawning multiple reasoning paths (or perspectives), then merging the insights. While prompting tricks can induce this behavior at inference, they carry heavy overhead and brittle generalization. Parallel-R1, a new paper by Tencent AI Lab Seattle with collaborators, pioneers a training-time RL framework for instilling parallel thinking as a native reasoning strategy. 


What is Parallel-R1

The key idea: don’t just prompt models to use parallel paths—train them to do so. Parallel-R1 has a progressive curriculum:

  1. Cold start (format learning via SFT) — teach the model the syntax/tags of parallel blocks (e.g. <Parallel>, <Path>...</Path>, <Summary>), using easier math problems (GSM8K) where high-quality parallel traces are easy to generate.

  2. Reinforcement learning (RL) on easy tasks, to explore usage of parallel thinking, with reward that combines correctness + usage of parallel structure. 

  3. RL on more difficult problems (e.g. DAPO, AMC, AIME), so the model generalizes both performance and the parallel thinking style. 

The architecture has two variants: a causal (structure-agnostic) version and a structured version. The structured version modifies the attention mechanism (via path-window masking, separate position encodings) so paths are more isolated during reasoning. But structured variants show trade-offs—good for generalization in some settings, but less robust under distribution shift.


Results & gains

On a battery of math benchmarks (MATH, AMC23, AIME24, AIME25), Parallel-R1 shows consistent improvements:

  • The “Seen” variant (causal) achieves ~48.9% average across benchmarks (Mean@16 / Pass@16, etc.), beating baseline GRPO RL on general math tasks. 

  • In particular, on AIME’25, Parallel-R1 raises accuracy by ~8.4% over a purely sequential RL model trained on the harder tasks directly. 

  • The structured (Unseen) variant also performs well under certain reward schedules; the “alternating ACC/PAR” reward schedule (switching between rewarding correctness and parallel structure periodically) helps balance parallel usage and performance. 

Beyond numerical gains, the authors observe a behavioral shift: early in training, the model heavily uses parallel paths as an exploration tool, branching in many places; as the model becomes stronger, it shifts to using parallel paths more conservatively, mostly for verification near the end of reasoning. This shift correlates with stronger final performance. 


Why this matters

  • Performance & efficiency trade-off: Parallel-R1 shows that training models for parallel thinking can yield higher reasoning ability without ballooning inference cost (since only when needed are parallel paths triggered).

  • Better than imitation: Many earlier works used supervised fine-tuning on synthetic parallel reasoning traces under teacher forcing; but those often over-fit to particular patterns. RL in Parallel-R1 helps models learn to decide when parallel paths help, not just how to mimic them.

  • Scaffolding exploration: The cold-start + easy tasks + alternating reward strategy functions as a scaffold, enabling RL to find a stronger policy space than direct RL on hard tasks.

  • Architecture designs matter: The structured variant shows that attention masking and position encodings can help or hurt depending on how well training data matches deployment tasks.


Limitations & future directions

  • The gains, though significant, still leave room before human-level performance in very hard math tasks.

  • The structured variants can struggle under domain shift; care needed in architectural changes that assume particular path structures.

  • Triggering parallel thinking (using <Parallel> blocks) costs some token and compute overhead, though the model learns to use it more sparsely over time.

  • There’s a balance tension between pushing for parallel structure (which encourages exploration) and maximizing accuracy (which sometimes pushes toward fewer divergences). Reward engineering is delicate.


Bottom line: Parallel-R1 is a breakthrough toward training LLMs that think in parallel, not just deeper. By combining curriculum learning, structured or causal variants, and reinforcement learning with rewards for both correctness and reasoning style, it unlocks better performance on challenging math tasks. As reasoning benchmarks and applications demand both correctness and robustness, methods like this will likely become a standard part of the toolkit.

Paper link: arXiv 2509.07980 (PDF)

10.9.25

TraceRL puts diffusion LLMs on the reasoning map

 Autoregressive (AR) giants have dominated reasoning benchmarks, while diffusion language models (DLMs) were seen as “fast samplers” with limited logic chops. A new paper from Princeton and UChicago argues that’s mostly a training-objective problem—and offers TraceRL, a trajectory-aware reinforcement learning framework that aligns what a DLM learns with how it actually samples. The team also releases code and ready-to-run models under the TraDo banner. 

What’s new

  • Trajectory-aware RL for DLMs. Instead of scoring randomly masked sequences, TraceRL optimizes against the model’s intermediate inference traces, matching the left-to-right / blockwise behavior used at decode time. A diffusion-based value model stabilizes training by reducing variance. Crucially, the method works for full-attention and block-attention DLMs. 

  • Open stack. The release includes a framework to build/train/deploy DLMs across architectures, with KV-cache acceleration, inference engines, SFT + RL recipes for math and code, and links to TraDo-4B/8B checkpoints. 

The receipts

On headline benchmarks (dynamic vs. static sampling shown in the paper), the TraDo models post the strongest DLM numbers to date and overtake AR peers at similar scale on math:

  • TraDo-8B-Instruct: MATH500 78.5, AIME’24 13.3, LCB-V2 25.9—a +6.1% relative lift over Qwen2.5-7B-Instruct and +51.3% over Llama-3.1-8B-Instruct on math reasoning. 

  • TraDo-4B-Instruct: MATH500 75.6, AIME’24 10.3, LCB-V2 18.7, consistently edging 7B AR baselines on math. 

  • TraDo-8B-Thinking (long-CoT): first long chain-of-thought diffusion LLM, hitting MATH500 87.4, AIME’24 35.5, LCB-V2 34.6 with very long answers. 

The authors attribute gains to objective/trajectory alignment and show smoother curves with the value model vs. policy-only RL. They also document a speed/accuracy trade-off: dynamic sampling is faster; static top-1 decoding squeezes out extra points. 

Why it matters

  1. DLMs aren’t just “fast”—they can reason. With the right RL target, parallel generation stacks clear long-form math and coding hurdles previously ceded to AR. 2) Unifies the zoo. One RL recipe spans full-attention and block-diffusion, and even helps enlarge block size for more flexible sampling. 3) Practical path. The open framework + KV-cache tricks make DLM post-training and deployment feel product-ready, not just a lab exercise. 

Setup notes

Math RL uses 8k hard MATH tasks; coding RL uses 6k verified problems from PrimeIntellect. Long-CoT training mixes TraceRL with long-form SFT as a curriculum. 

Bottom line: TraceRL reframes diffusion LLMs as credible reasoners, not just fast generators—and TraDo-8B-Thinking plants the first long-CoT flag on the DLM side of the field. 

Paper link: arXiv 2509.06949 (PDF)

31.5.25

DeepSeek R1-0528: China's Open-Source AI Model Challenges Industry Giants

 Chinese AI startup DeepSeek has unveiled its latest open-source model, R1-0528, marking a significant stride in the global AI landscape. This release underscores China's growing prowess in AI development, offering a model that rivals established giants in both performance and accessibility.

Enhanced Reasoning and Performance

R1-0528 showcases notable improvements in reasoning tasks, particularly in mathematics, programming, and general logic. Benchmark evaluations indicate that the model has achieved impressive scores, nearing the performance levels of leading models like OpenAI's o3 and Google's Gemini 2.5 Pro. Such advancements highlight DeepSeek's commitment to pushing the boundaries of AI capabilities.

Reduced Hallucination Rates

One of the standout features of R1-0528 is its reduced tendency to produce hallucinations—instances where AI models generate incorrect or nonsensical information. By addressing this common challenge, DeepSeek enhances the reliability and trustworthiness of its AI outputs, making it more suitable for real-world applications.

Open-Source Accessibility

Released under the permissive MIT License, R1-0528 allows developers and researchers worldwide to access, modify, and deploy the model without significant restrictions. This open-source approach fosters collaboration and accelerates innovation, enabling a broader community to contribute to and benefit from DeepSeek's advancements.

Considerations on Content Moderation

While R1-0528 offers numerous technical enhancements, it's essential to note observations regarding its content moderation. Tests suggest that the model may exhibit increased censorship, particularly concerning topics deemed sensitive by certain governing bodies. Users should be aware of these nuances when deploying the model in diverse contexts.

Conclusion

DeepSeek's R1-0528 represents a significant milestone in the evolution of open-source AI models. By delivering enhanced reasoning capabilities, reducing hallucinations, and maintaining accessibility through open-source licensing, DeepSeek positions itself as a formidable contender in the AI arena. As the global AI community continues to evolve, contributions like R1-0528 play a pivotal role in shaping the future of artificial intelligence.

30.5.25

DeepSeek R1‑0528: The Open‑Source Challenger That Rivals GPT‑4o and Gemini 2.5 Pro

 Chinese startup DeepSeek has just released R1‑0528, a major update to its flagship reasoning model, positioning it as an affordable yet powerful open‑source alternative to OpenAI’s o3 and Google’s Gemini 2.5 Pro.

The new release, published on Hugging Face under the permissive MIT License, brings a host of enhancements to math, science, business, and coding reasoning—all while reinforcing its competitive edge.



🚀 What’s New in R1‑0528

  • Stronger Reasoning:
    On the AIME 2025 benchmark, accuracy surged from 70% to an impressive 87.5%, thanks to longer reasoning chains (average 23k tokens vs. 12k before). Code generation also jumped, with LiveCodeBench scores rising from 63.5% to 73.3% alongside doubling performance on the challenging “Humanity’s Last Exam.”

  • Developer-Friendly Features:
    R1‑0528 now supports JSON output and function calling, streamlining integration into developer pipelines and automation workflows.

  • New Model Variant:
    A distilled version—R1‑0528‑Qwen3‑8B—brings lightweight performance that's still on par with larger models in open benchmarks like AIME 2024.

🏆 Why This Matters

DeepSeek continues to challenge the perception that high performance requires closed-source models and massive budgets. R1‑0528 delivers competitive strength on par with expensive proprietary systems, but under an MIT license and at significantly lower cost—R1's API even cost just $0.14/1M tokens (peak) with local runtime options detailed on GitHub.

This open-access approach puts serious pressure on dominant U.S. models and fosters global collaboration—developers worldwide can use, modify, and deploy R1‑0528 freely.


🌍 Open-Source Renaissance in AI

Since its initial R1 model launch in January, DeepSeek has quickly become a key player in the global AI landscape. R1‑0528 maintains the open-source ethos and stakes its claim as a champion of community-driven innovation in areas where cost and licensing are bottlenecks.


🗣️ Community Buzz

Feedback from enthusiasts is bullish: voices from Reddit’s LocalLLaMA community noted that “DeepSeek is now almost on par with OpenAI’s o3 High model on LiveCodeBench! Huge win for opensource!”

Analysts also see this release as a strategic “Sputnik moment” that could disrupt AI dominance—similar to earlier 2025 reports on DeepSeek’s initial release.


✅ Final Verdict

DeepSeek R1‑0528 marks a significant milestone in open-source AI: powerful reasoning, developer utility, and community support—all while costing a fraction of proprietary counterparts. As a truly accessible yet competitive model, it nudges the AI ecosystem toward openness and transparency—without sacrificing performance.

27.5.25

NVIDIA Introduces AceReason-Nemotron: Enhancing Math and Code Reasoning through Reinforcement Learning

 NVIDIA has unveiled AceReason-Nemotron, a 14-billion-parameter open-source model designed to enhance mathematical and coding reasoning through large-scale reinforcement learning (RL). This model demonstrates that RL can significantly improve reasoning capabilities in small to mid-sized models, surpassing traditional distillation-based approaches.

Key Features and Innovations

  • Sequential RL Training Strategy: The model undergoes a two-phase RL training process—initially on math-only prompts, followed by code-only prompts. This approach not only boosts performance in respective domains but also ensures minimal degradation across tasks. 

  • Enhanced Benchmark Performance: AceReason-Nemotron-14B achieves notable improvements on various benchmarks:

    • AIME 2025: 67.4% (+17.4%)

    • LiveCodeBench v5: 61.1% (+8%)

    • LiveCodeBench v6: 54.9% (+7%) 

  • Robust Data Curation Pipeline: NVIDIA developed a comprehensive data curation system to collect challenging prompts with verifiable answers, facilitating effective verification-based RL across both math and code domains. 

  • Curriculum Learning and Stability: The training incorporates curriculum learning with progressively increasing response lengths and utilizes on-policy parameter updates to stabilize the RL process. 

Implications for AI Development

AceReason-Nemotron's success illustrates the potential of reinforcement learning in enhancing the reasoning abilities of AI models, particularly in mathematical and coding tasks. By releasing this model under the NVIDIA Open Model License, NVIDIA encourages further research and development in the AI community.

What Claude offers now From Anthropic’s announcements: Creates and edits real files directly in chats or the desktop app: Excel (.xlsx)...