🎓How I Study AIHISA
📖Read
📄Papers📰Blogs🎬Courses
💡Learn
🛤️Paths📚Topics💡Concepts🎴Shorts
🎯Practice
🧩Problems🎯Prompts🧠Review
Search
DiffCoT: Diffusion-styled Chain-of-Thought Reasoning in LLMs | How I Study AI

DiffCoT: Diffusion-styled Chain-of-Thought Reasoning in LLMs

Intermediate
Shidong Cao, Hongzhan Lin, Yuxuan Gu et al.1/7/2026
arXivPDF

Key Summary

  • •DiffCoT treats a model’s step-by-step thinking (Chain-of-Thought) like a messy draft that can be cleaned up over time, not something fixed forever.
  • •It borrows the idea of diffusion (turning noise into clarity) to iteratively revise earlier reasoning steps while still writing the next step one token at a time.
  • •A sliding window lets the model both fix a few recent steps and extend the solution, so it doesn’t get stuck carrying forward an early mistake.
  • •A causal noise schedule adds more ‘shake’ to later steps than earlier ones, protecting the natural order of reasoning.
  • •Training uses preference signals: better partial steps are treated as ‘cleaner’ than worse ones and are learned via DPO so the model prefers better reasoning paths.
  • •Across GSM8K, SVAMP, and MATH (various backbones like Llama3-8B and Qwen3), DiffCoT consistently matches or beats strong baselines.
  • •Ablations show that too-small or too-large windows hurt, and removing causal noise clearly drops accuracy—so these design choices matter.
  • •Under deliberate ‘noisy prefix’ corruption, DiffCoT corrects itself far more often than prior step-wise preference methods.
  • •It fine-tunes standard autoregressive LLMs (keeps token-by-token generation) but adds step-level diffusion-style revision.
  • •Result: more robust, self-correcting reasoning that resists error accumulation.

Why This Research Matters

In real life, we make small early slips when solving problems, and good thinkers catch and fix them as they go. DiffCoT trains language models to do the same: it can revise recent steps while moving forward, so one early wobble doesn’t ruin the whole solution. This makes math help, homework feedback, and planning assistants more trustworthy and less brittle. Because DiffCoT keeps standard token-by-token generation, it plugs into today’s LLMs without rebuilding everything from scratch. Its causal noise schedule respects the natural order of reasoning, which helps in domains where steps must follow a logical timeline. Overall, you get assistants that are steadier under pressure and better at self-correction.

Detailed Explanation

Tap terms for definitions

01Background & Problem Definition

🍞 Hook: Imagine solving a big jigsaw puzzle. If you place a wrong piece early and never look back, every later piece might line up badly and your picture won’t make sense.

🥬 The Concept: Chain-of-Thought (CoT) reasoning is when a language model explains its thinking step by step to solve problems.

  • How it works: 1) Break the problem into small steps. 2) Solve each step in order. 3) Use the steps to reach the final answer. 4) Share the steps as a written chain of thought.
  • Why it matters: Without CoT, models may jump to conclusions; with CoT, they can show their work and often get harder problems right. 🍞 Anchor: When solving “What is 37×(20+5)?”, CoT writes: 20+5=25; 37×25=37×(100/4)=3700/4=925. Answer: 925.

🍞 Hook: You know how you write an essay draft one sentence after another? That’s like how most language models write: they add the next word based on the ones before it.

🥬 The Concept: Autoregressive (AR) modeling means generating text one token at a time using what’s already been written.

  • How it works: 1) Read the existing tokens. 2) Predict the next token. 3) Append it. 4) Repeat.
  • Why it matters: Without AR, the model can’t build thoughts step-by-step or stay consistent with its own past. 🍞 Anchor: When ChatGPT completes “Peanut butter and …”, AR helps it choose “jelly.”

🍞 Hook: Imagine practicing a song only by listening to a perfect recording. When you perform live and make a tiny mistake early, the rest falls apart because you never practiced recovering.

🥬 The Concept: Exposure bias is when a model is trained only on perfect step prefixes but tested on its own possibly-flawed prefixes.

  • How it works: 1) Training shows only correct histories. 2) At test time, the model must use its own outputs, which may contain errors. 3) Early mistakes snowball.
  • Why it matters: Without fixing exposure bias, one wrong step can drag the entire solution off course. 🍞 Anchor: If the model mis-writes “0.25x=600” instead of “0.15x=600,” all later math goes wrong.

🍞 Hook: Think of a coach watching a team. Just yelling “win or lose” at the end isn’t enough; you need feedback on each play within the game to improve.

🥬 The Concept: Preference Optimization (PO) teaches a model to prefer better answers or better steps using comparisons (this vs. that), not just final correctness.

  • How it works: 1) Collect pairs of better/worse responses. 2) Train the model to assign higher probability to better ones. 3) Repeat across many examples.
  • Why it matters: Without PO, the model may learn only from final answers and miss where the reasoning went wrong. 🍞 Anchor: Given two mid-solution steps, “Add 2+2=4” (good) vs. “Add 2+2=5” (bad), PO pushes the model toward 4.

🍞 Hook: Imagine a maze where you sometimes peek ahead to see if your path still makes sense—and if not, you backtrack and fix earlier moves.

🥬 The Concept: Many prior CoT systems move strictly forward with no built-in way to fix earlier steps.

  • How it works: 1) Generate step 1, then step 2, etc. 2) If step 2 was based on a flawed step 1, there’s often no mechanism to revise step 1.
  • Why it matters: Without revising powers, early mistakes cause error accumulation. 🍞 Anchor: If you assume “25% of x + 600” equals “0.25x + 600” (instead of 0.25(x+600)), every later equation breaks.

🍞 Hook: Think about cleaning a foggy window slowly, swipe by swipe, until you can see clearly.

🥬 The Concept: Diffusion models are systems that start with noisy data and learn to remove the noise step by step.

  • How it works: 1) Add noise to clean data repeatedly (forward). 2) Train a model to reverse this, removing noise bit by bit (denoising). 3) Generate by starting from noise and denoising.
  • Why it matters: Without diffusion’s denoising power, it’s hard to fix corrupted information. 🍞 Anchor: In images, diffusion can turn a grainy picture into a sharp one by iteratively cleaning it.

The world before: CoT increased transparency and often improved math problem solving. But strict “only forward” AR decoding, plus teacher-forced training on perfect prefixes, made models brittle: a small slip early could mislead everything downstream. People tried to explore multiple paths (Tree-of-Thought with search) or used external critics to pick better solutions, but these methods can be slow and still don’t truly let the model revise its own earlier steps while it reasons.

The problem: How do we make CoT robust to early slips, so the model can correct itself mid-solution rather than snowball the error?

Failed attempts: 1) Pure AR CoT with teacher forcing—fast but fragile. 2) Heavy search like MCTS—can be effective but costly at inference time. 3) Step-wise preference learning—better, but still tied to local, forward-only training on mostly clean prefixes.

The gap: Missing is a unified way to both generate and revise reasoning steps as one coherent process, keeping AR strengths but adding a principled “cleaning” (denoising) ability.

Real stakes: In everyday math, science homework, coding, and planning, we all make small early mistakes. A smart helper should notice and fix them on the fly. That’s what DiffCoT aims to do—help models stay accurate even when the path gets a little messy.

02Core Idea

🍞 Hook: You know how using an eraser lets you fix earlier sentences while you keep writing the rest of your story?

🥬 The Concept: The key idea is to treat the whole Chain-of-Thought as a draft that can be iteratively cleaned up (denoised) while still writing the next part.

  • How it works: 1) View each reasoning step as having a ‘noise level’ (worse steps = noisier). 2) Keep a sliding window over recent steps. 3) Iteratively refine the window (clean earlier steps) while predicting the next step. 4) Use a causal noise plan so earlier steps get gentler noise than later ones.
  • Why it matters: Without iterative cleanup, early mistakes snowball; with cleanup, the model can self-correct and stay on track. 🍞 Anchor: If your Step 2 was a shaky equation, DiffCoT can revise Step 2 while producing Step 3, guiding the solution back on course.

Three analogies for the same idea:

  1. Drafting an essay: First draft is messy. You reread the last paragraph (sliding window), fix awkward sentences (denoise), then write the next paragraph (AR). Repeat until the essay shines.
  2. Sculpting clay: You shape a section (steps), smooth imperfections (denoise), then add a new piece (next step), always able to go back a little to refine.
  3. Noise-canceling headphones: They continuously listen (context), remove noise (denoise), and let the true signal (correct reasoning) come through as you proceed.

Before vs. After:

  • Before: CoT ran forward only. Early slips often haunted the rest of the solution. Fixes required external critics or expensive search.
  • After: DiffCoT lets the model revise its own recent thinking in-line, reducing exposure bias and error accumulation while keeping familiar token-by-token generation.

🍞 Hook: Imagine ranking homework solutions from best to worst and using that ordering to understand what ‘clean’ vs. ‘noisy’ work looks like.

🥬 The Concept: Step-level diffusion-styled noising ranks candidate step responses by reward (e.g., rollout success rate) and treats higher-ranked ones as ‘cleaner’.

  • How it works: 1) For a step, gather multiple candidate mini-steps. 2) Score them (e.g., via Monte Carlo rollouts). 3) Order them from best (low noise) to worse (high noise). 4) Use these as the forward ‘noising’ ladder for training denoising.
  • Why it matters: Without a graded ‘noise’ notion, you can’t train the model to move systematically from messy to clean reasoning. 🍞 Anchor: If candidates are [excellent, okay, poor], the model learns to transform ‘poor’ toward ‘excellent’ during refinement.

🍞 Hook: Think of a magnifying glass that slides over a few recent sentences so you can polish them before moving on.

🥬 The Concept: The sliding-window mechanism lets the model revise a handful of previous steps (denoise) and then generate the next step (AR) in one loop.

  • How it works: 1) Keep m recent steps in a window. 2) Apply denoising to make them cleaner. 3) Shift window forward and predict the new step. 4) Repeat until done.
  • Why it matters: Without this, you must choose either pure forward generation or full-sequence diffusion; the window balances both. 🍞 Anchor: While solving a word problem, the model refines Steps 2–4 and then writes Step 5, all in one cycle.

🍞 Hook: Picture sprinkling a tiny bit of challenge on early steps but more on the later, foggier ones to match how real reasoning unfolds.

🥬 The Concept: A causal noise schedule perturbs later steps more than earlier ones, matching the natural forward flow of cause → effect in reasoning.

  • How it works: 1) Define noise strength as a function of step index and iteration. 2) Earlier steps get lighter noise; later steps get heavier noise. 3) This preserves causal order while enabling correction.
  • Why it matters: Without causal noise, denoising can jumble temporal order and harm coherence. 🍞 Anchor: When fixing a 5-step solution, Step 1 gets only a light tweak, Step 5 gets a stronger clean-up.

🍞 Hook: Think of judging two mini-explanations and learning to prefer the clearer one.

🥬 The Concept: Direct Preference Optimization (DPO) trains the model to choose cleaner (higher-reward) windowed sequences over noisier ones.

  • How it works: 1) Build ‘win’ (cleaner) and ‘lose’ (noisier) sequences within the window. 2) Apply DPO so the model increases probability of the win. 3) Repeat across training data.
  • Why it matters: Without this preference signal, the model won’t reliably shift its reasoning toward higher-quality steps. 🍞 Anchor: If two partial chains differ only by one shaky equation, DPO nudges the model toward the chain with the correct equation.

Why it works (intuition): Reasoning errors behave like noise sprinkled across steps. If you can estimate which steps are cleaner vs. noisier and give the model a way to roll backward a bit (denoise) while still moving forward (AR), it can steadily steer toward globally consistent, correct solutions.

Building blocks: (1) Reward-ranked candidates per step (noise ladder), (2) Sliding-window denoising + next-step generation, (3) Causal noise schedule for temporal order, (4) Preference learning (DPO) to prefer cleaner windows, (5) An AR backbone so it plugs into existing LLMs.

03Methodology

High-level recipe: Input question → Build step-level noise ladder → Sliding-window denoise-and-generate → Output final chain-of-thought and answer.

Step A: Build diffusion-styled step noise with reward-ranked candidates. 🍞 Hook: Imagine testing a few different mini-steps for the same point in a solution, then lining them up from ‘great’ to ‘not-so-great’. 🥬 The Concept: For each reasoning step, create multiple candidate steps and rank them by their estimated quality; treat the best as low-noise and the worse ones as higher-noise.

  • How it works: 1) Use search (e.g., MCTS) or multiple samples to produce step candidates. 2) For each candidate, run Monte Carlo rollouts (e.g., 8 tries) to see how often it leads to a correct final answer (success rate). 3) Rank by success rate. 4) Store [low-noise → high-noise] versions per step.
  • Why it matters: Without graded candidates, the model cannot learn to transform noisy reasoning into clean reasoning. 🍞 Anchor: If at Step 2 you have candidates with 75%, 50%, and 12% success, they form your clean→noisy ladder.

Step B: Sliding-window denoising integrated with AR generation. 🍞 Hook: Think about polishing the last few sentences you wrote before typing the next one. 🥬 The Concept: Keep a window of the last m steps. During each iteration, refine (denoise) those steps and then predict the next step token-by-token.

  • How it works: 1) Take prompt + current steps. 2) Replace the window’s steps with their noisier/cleaner variants according to the current iteration’s noise level. 3) Train the model to map noisier windowed sequences (‘lose’) toward cleaner ones (‘win’) using DPO. 4) After refinement, generate the next step and slide forward.
  • Why it matters: Without this, the model either never revises (pure AR) or loses causality (full-sequence diffusion). 🍞 Anchor: While solving, the model tidies Steps 3–5 and then writes Step 6, then repeats.

Step C: Causal diffusion noise schedule. 🍞 Hook: You don’t want to shake the beginning of a recipe as much as the end—you wrote those first steps more carefully. 🥬 The Concept: Assign noise strength based on both the iteration and which step it is: earlier steps get lighter noise, later steps heavier.

  • How it works: 1) Define σ_t^k that depends on denoising iteration t and step index k. 2) Within the window, earlier k’s have smaller σ, later k’s larger σ. 3) This matches reasoning’s cause→effect structure.
  • Why it matters: Without this, refinement might scramble the order and weaken logical flow. 🍞 Anchor: In a 5-step window, Steps 1–2 get a gentle clean; Steps 4–5 get stronger cleanup.

Step D: Preference training with DPO using windowed win/lose pairs. 🍞 Hook: If you were choosing between two short paragraphs, you’d learn to favor the clearer one. 🥬 The Concept: Construct pairs that differ in the window: the refined ‘win’ vs. the unrefined or noisier ‘lose,’ conditioned on the same prefix.

  • How it works: 1) Prefix is real context up to k−1, possibly including some noise in-window (to reduce exposure bias). 2) Win = refined window + cleaner next-step candidate. 3) Lose = unrefined window + noisiest next-step. 4) Apply DPO with a reference model to push probability mass toward Win.
  • Why it matters: Without learning these preferences under partially noisy prefixes, models revert to brittle, clean-only training. 🍞 Anchor: Given two versions of steps [2–4], DPO makes the model more likely to produce the version with the correct sub-equation.

Step E: Inference-time reasoning (how it runs when solving new problems). 🍞 Hook: It’s like reading the last paragraph you wrote, fixing small bumps, and then writing the next one until the essay is done. 🥬 The Concept: At test time, the model repeats the same sliding-window refinement while extending the chain, but now guided by its learned preferences.

  • How it works: 1) Start generating steps. 2) Iteratively refine the windowed steps and add the next step. 3) Stop when a final answer is reached.
  • Why it matters: Without running refinement at test time, the training-time benefits won’t fully show up during real use. 🍞 Anchor: On the Mrs. Snyder salary problem: If it writes Step 3 as “0.4x = 0.25x + 600,” it can refine to “0.25(x+600) = 0.4x,” leading correctly to x=1000.

Concrete example (Mrs. Snyder):

  • Problem: Rent/utilities were 40% of old income x. After a $600 raise, rent is 25% of new income. Find old income x.
  • A typical mistaken step: “0.4x = 0.25x + 600” (treats 25% of (x+600) as 0.25x + 600).
  • Window refinement: The model revises to “0.25(x+600)=0.4x” → 0.25x+150=0.4x → 0.15x=150 → x=1000.
  • Outcome: Iterative denoising fixed the earlier equation and saved the solution.

The secret sauce:

  • Two axes at once: temporal (AR steps) and noise (diffusion-style denoising). The sliding window is the bridge. Causal noise keeps time’s arrow intact. DPO preferences ensure the model consistently favors cleaner reasoning paths even under partially noisy prefixes.

Implementation notes (friendly summary):

  • Candidates per step come from search/sampling; quality is estimated via rollouts (how often they lead to the right final answer).
  • Fine-tuning uses LoRA for efficiency and DPO with a reference model; training mixes cleaner and noisier prefixes to reduce exposure bias.
  • Works with standard instruction-tuned LLMs (e.g., Qwen3, Llama3) without redesigning token-level generation.

04Experiments & Results

🍞 Hook: Think of a spelling bee where everyone tries many words. We measure who gets the most right, who improves with practice, and who stays steady even when we toss in a few tricky surprises.

🥬 The Concept: The authors test whether DiffCoT solves math problems more accurately, stays robust when earlier steps are noisy, and which design choices matter.

  • How it works: 1) Benchmarks: GSM8K, SVAMP, and MATH (with levels from easy L1 to hard L5). 2) Backbones: Llama3-8B, Qwen3-8B, Qwen3-4B. 3) Baselines: CoT, ToT, TS-SFT, CPO, Step-DPO, Full-Step-DPO. 4) Metrics: Test accuracy (and correction success under controlled noise).
  • Why it matters: Without rigorous comparisons, we can’t tell if the approach really helps beyond prior methods. 🍞 Anchor: On GSM8K with Llama3-8B, DiffCoT’s 39.6% is like inching from a B- to a B when others sit around 37–39%.

The competition (context):

  • CoT (plain step-by-step), ToT (tree search), TS-SFT (supervised on ToT paths), CPO/Step-DPO/Full-Step-DPO (preference-based step/trajectory alignment). Some are fast but brittle; some are more stable but can falter on certain datasets; ToT can be expensive.

Scoreboard highlights with context:

  • GSM8K: DiffCoT typically edges out or matches strong baselines. Examples: Llama3-8B: 39.6% (DiffCoT) vs 39.3% (TS-SFT) vs 37.2% (CoT). Qwen3-8B: 66.2% (DiffCoT) vs 65.9% (Full-Step-DPO). Qwen3-4B: 65.4% (DiffCoT) vs 64.7% (Full-Step-DPO).
  • SVAMP: DiffCoT is among the top. Qwen3-8B: 85.5% (DiffCoT) vs 85.9% (Full-Step-DPO)—a near tie. Llama3-8B: 50.4% (DiffCoT) vs ~49–50% for others.
  • MATH (L1–L5): DiffCoT is consistently competitive and often better, notably maintaining stability across levels where others wobble.
  • Plain English: When the class average is around a B, DiffCoT often gets a B+ and, importantly, keeps that B+ across many different tests.

Surprising or key findings:

  1. Window size matters (ablation):
  • Too small (size/stride=1) → behaves like pure AR; less ability to fix past steps. Llama3-8B drops ~3.3 pts on GSM8K; Qwen3-4B drops ~2.4.
  • Too large (size/stride=K) → behaves like full diffusion; hurts causal flow. Llama3-8B drops ~9.3 on GSM8K; Qwen3-4B drops ~8.7.
  • Sweet spot: moderate window sizes that balance correction and causality.
  1. Causal noise is critical:
  • Removing it causes clear drops (e.g., −4.1 on GSM8K with Llama3-8B; −3.5 with Qwen3-4B), showing temporal-aware noising really helps preserve coherent reasoning.
  1. Robustness under prefix corruption:
  • When the first half of the chain is randomly perturbed with ‘low-reward but plausible’ steps, DiffCoT’s correction success rate stays much higher than Full-Step-DPO across Llama3-8B, Qwen3-8B, Qwen3-4B, and across multiple noise levels.
  • Plain English: DiffCoT is better at noticing, “Hey, that earlier step got weird,” and steering back to the right path.

Efficiency notes:

  • Fine-tunes standard AR LLMs with LoRA; training/inference on 500 GSM8K samples took ~11 GPU-hours (A100 80GB). Not free, but far lighter than heavy search at inference for every problem.

Bottom line: Across datasets and models, DiffCoT is not just occasionally better—it’s reliably strong and more resilient when the chain contains bumps. The ablations also teach us which knobs to turn: window size and causal noise are must-haves.

05Discussion & Limitations

🍞 Hook: Imagine a great study habit that still needs the right desk, the right lighting, and the right schedule to shine.

🥬 The Concept: Honest assessment of DiffCoT—where it shines, what it needs, when not to use it, and what questions remain.

  • How it works: 1) List limitations. 2) State resource needs. 3) Explain scenarios where it’s not ideal. 4) Share open questions that guide future work.
  • Why it matters: Without a fair look at trade-offs, we can’t responsibly apply or extend the method. 🍞 Anchor: Even a strong basketball strategy may still struggle if the court is too slippery or players are exhausted—context matters.

Limitations:

  • Off-policy data construction: Candidates for steps are gathered with a policy that isn’t exactly the one being trained; this can cause distribution shifts and mild instability as tasks get harder or chains get longer.
  • Breaks strict Markov prefix behavior: Because it refines past steps, generation is no longer “only forward.” While that’s the point, it can increase variance and require more data/iterations for stable convergence.
  • Hyperparameter sensitivity: Window size/stride and noise schedule need tuning; wrong settings hurt.

Required resources:

  • Compute: Rollout-based scoring for candidates (e.g., 8 rollouts per candidate step) isn’t free; training then runs DPO on many windowed win/lose pairs. A few GPUs and hours are typically needed for meaningful results.
  • Data: Needs enough problems to sample candidates, estimate success rates, and cover step variations.

When NOT to use:

  • Very short tasks with almost no multi-step reasoning: The gains from iterative denoising may be minimal.
  • Strictly causal logs needed for audit where no retroactive change is allowed: DiffCoT’s revising nature may conflict with immutable ledgers.
  • Ultra low-latency settings: The extra refinement cycles can add overhead compared to plain greedy decoding.

Open questions:

  • Can we learn the causal noise schedule automatically per domain/problem instead of hand-setting it?
  • How well does DiffCoT transfer to non-math domains (coding, medical reasoning, law) with different error patterns?
  • Can reinforcement learning directly optimize the denoising dynamics of the sliding window for even stronger correction?
  • How large can windows get before benefits flatten or reverse, and can adaptive windows help?

Overall: DiffCoT is a practical, robust step forward, but it still asks for thoughtful tuning and resources, and it opens exciting paths for future learning-to-denoise reasoning.

06Conclusion & Future Work

Three-sentence summary: DiffCoT reframes step-by-step reasoning as an iterative denoising process that lets models revise recent steps while generating the next one. Using a sliding window, a causal noise schedule, and preference learning (DPO), it reduces exposure bias and corrects early mistakes without abandoning standard token-by-token generation. Experiments on GSM8K, SVAMP, and MATH show reliable gains and stronger robustness than prior CoT preference methods.

Main achievement: Unifying autoregressive generation with diffusion-styled revision at the step level—so reasoning can both move forward and clean itself up in one loop.

Future directions: Automatically learn noise schedules; adapt window size on-the-fly; connect denoising dynamics with reinforcement learning; extend to coding, science QA, and planning; and scale to longer, more complex chains.

Why remember this: It’s a mindset shift—treat a model’s thoughts like erasable pencil, not permanent ink. By cleaning while creating, DiffCoT turns fragile, forward-only reasoning into a steadier, self-correcting process that better survives the little bumps we all make when we think.

Practical Applications

  • •Build math tutors that can spot and fix their own mid-solution slips instead of doubling down on a wrong step.
  • •Create coding helpers that revise earlier reasoning about variable states or edge cases while proposing the next code edit.
  • •Design study assistants that refine outlines or proofs as they draft the next section, catching logical gaps early.
  • •Develop planning tools (travel, projects) that can correct earlier assumptions when new constraints appear, while still moving plans forward.
  • •Improve scientific and data-analysis assistants that refine earlier hypotheses or data-cleaning choices as they run subsequent analyses.
  • •Enhance grading or feedback systems that can suggest repaired intermediate steps, not just mark final answers right/wrong.
  • •Build safer reasoning for chain-of-thought agents in healthcare triage or legal research by enabling iterative correction and temporal discipline.
  • •Use DiffCoT-style training to curate better reasoning datasets, ranking and denoising step candidates for future models.
  • •Integrate with lightweight search: use a small amount of rollout scoring to build noise ladders, then rely on DiffCoT during inference.
  • •Deploy in educational apps that show students how to correct a mistaken step and continue productively, modeling good metacognition.
#Chain-of-Thought#Diffusion models#Autoregressive decoding#Exposure bias#Preference Optimization#DPO#Sliding window#Causal noise schedule#Monte Carlo rollouts#MCTS#Error accumulation#Iterative denoising#Mathematical reasoning#Reasoning robustness#Self-correction
Version: 1