🎓How I Study AIHISA
📖Read
📄Papers📰Blogs🎬Courses
💡Learn
🛤️Paths📚Topics💡Concepts🎴Shorts
🎯Practice
🧩Problems🎯Prompts🧠Review
Search
ToolPRMBench: Evaluating and Advancing Process Reward Models for Tool-using Agents | How I Study AI

ToolPRMBench: Evaluating and Advancing Process Reward Models for Tool-using Agents

Intermediate
Dawei Li, Yuguang Yao, Zhen Tan et al.1/18/2026
arXivPDF

Key Summary

  • •ToolPRMBench is a new benchmark that checks, step by step, whether an AI agent using tools picks the right next action.
  • •It turns long agent runs into tiny tests: same history, one correct action, one very-plausible but wrong action, plus tool info.
  • •Two ways make the tiny tests: offline sampling (swap a single step near a known-good path) and online sampling (let the agent run and catch the first real mistake).
  • •A multi-LLM verification team (GPT-5, Claude-4.5-haiku, Gemini-2.5-flash) votes on which action is truly better, cutting label noise; humans spot-check at 96% agreement.
  • •They compare many models and find tool-specialized PRMs, especially with RL (ToolPRM-GRPO), are much stronger than general PRMs or similarly sized open LLMs.
  • •Bigger base models help, but size alone isn’t enough; targeted tool-use training and RL greatly boost robustness and out-of-distribution generalization.
  • •Good PRM scores on ToolPRMBench predict real gains when guiding search (like best-of-n), while bad PRMs can make agents worse.
  • •Synthetic data helps in some environments (GTA) but not all (ToolTalk), so realism of injected mistakes matters.
  • •ToolPRMBench offers a cost-effective way to evaluate and train step-level judges that make tool-using agents more reliable in real tasks.

Why This Research Matters

ToolPRMBench makes tool-using AIs more dependable by checking if they pick the right next step, not just whether they got lucky at the end. This reduces costly mistakes, like copying to the wrong folder, sending an invoice to the wrong account, or querying the wrong database table. The benchmark’s verified, diverse cases help teams train PRMs that actually transfer to real work settings. Because better ToolPRMBench scores predict bigger gains during search-time decision-making, organizations can choose the right PRM for their budget and reliability needs. With lower inference costs than giant API models, specialized PRMs can serve as practical, always-on judges inside production agents. Ultimately, this helps businesses and users trust AI to use tools safely, consistently, and transparently.

Detailed Explanation

Tap terms for definitions

01Background & Problem Definition

🍞 Top Bread (Hook): You know how when you follow a recipe, making a tiny mistake early—like adding salt instead of sugar—can mess up the whole cake? Catching small errors early saves the entire dessert.

🥬 Filling (The Actual Concept — Tool-using Agents):

  • What it is: Tool-using agents are AIs that don’t just talk; they press buttons, call apps (APIs), and use tools like calculators, calendars, or file systems to get real work done.
  • How it works (step by step):
    1. The user gives an instruction, like “Find photos with ‘test’ in their name.”
    2. The agent looks at available tools and picks an action, such as “search the file system.”
    3. The tool returns results (e.g., a list of matching files).
    4. The agent uses that new info to pick the next action (e.g., copy selected files).
    5. This repeats—sometimes for many steps—until the final goal is done.
  • Why it matters: Without tool use, an AI can only talk. With tool use, the AI can actually do things in the world—but small mistakes along the way can snowball into big failures.

🍞 Bottom Bread (Anchor): Imagine a homework helper that can read your calendar and email your teacher. If it picks the wrong calendar or sends the email to the wrong person early on, the whole plan falls apart.

The World Before:

  • AIs were great at writing and chatting, but real tasks (booking, file operations, data lookups) demanded tool use: APIs, databases, or even web browsing. Benchmarks mostly checked final success (Did you get the right answer?), not where things went wrong.
  • Long tasks have many steps. If step 2 goes wrong, step 10 is doomed, but only grading the final answer doesn’t tell you which step failed.

🍞 Top Bread (Hook): Imagine your teacher giving you a sticker after every correct step in a math problem, not only at the end. That helps you stay on track.

🥬 Filling (The Actual Concept — Process Reward Model, PRM):

  • What it is: A PRM is a judge that scores each intermediate step of an agent’s process, telling it which next action looks right.
  • How it works (step by step):
    1. Look at the task history so far.
    2. Compare candidate next actions.
    3. Give higher scores to actions that follow the rules and move toward the goal.
    4. Use these scores to guide search, sample better actions, or prune bad paths.
  • Why it matters: Without step-level feedback, agents wander. With PRMs, agents get gentle nudges at every fork in the road, avoiding early mistakes that wreck the ending.

🍞 Bottom Bread (Anchor): In a file-copy task, a PRM prefers “cd into the right folder, then copy” over “copy using the wrong path,” so the agent doesn’t break the tool’s rules.

The Problem:

  • We lacked a reliable way to test PRMs for tool-using agents. Existing tests were web-only, final-answer focused, or not designed for step-level judging across diverse tools.

Failed Attempts:

  • Outcome-only rewards: Too sparse; they don’t say which step failed.
  • Web-only PRM tests: Too narrow; not all tool use is web browsing.
  • General PRMs (for math or web) applied to tool use: Often misaligned; they miss tool-specific constraints like parameter formats or state changes.

The Gap:

  • We needed a benchmark that: (1) zooms into a single decision step, (2) shows a correct action vs. a very-plausible wrong one, (3) covers diverse tools/APIs, and (4) has low label noise.

Real Stakes:

  • Everyday jobs—copying files safely, booking travel, updating spreadsheets, sending invoices—depend on correct tool use. A tiny mistake (wrong argument name, wrong folder, skipping a required step) can cost time or money.

🍞 Top Bread (Hook): Think of a referee crew in a sports game. One ref can miss a foul, but three refs together catch more mistakes.

🥬 Filling (The Actual Concept — Multi-LLM Verification):

  • What it is: Multiple strong language models independently judge which action is better, then vote.
  • How it works (step by step):
    1. Show each model the same history, actions, and tool description.
    2. Each model votes which action is strictly better.
    3. Keep samples with strong agreement and discard noisy ones.
    4. Human spot-checks confirm quality.
  • Why it matters: Cleaner labels make a fairer, sturdier benchmark; noisy labels make judges learn the wrong lessons.

🍞 Bottom Bread (Anchor): GPT-5, Gemini-2.5-flash, and Claude-4.5-haiku each vote on “Action A vs. Action B.” If all three agree, we trust that pair for testing.

Bottom line: ToolPRMBench was created to fairly, cleanly, and precisely test how well PRMs guide tool-using agents at each step, so real-world tasks become safer and more reliable.

02Core Idea

🍞 Top Bread (Hook): Imagine choosing the next move in a maze. If someone tells you, at each fork, which turn is better, you finish faster and avoid dead ends.

🥬 Filling (The Actual Concept — The Aha!):

  • What it is: Turn long, messy tool-using runs into bite-sized, step-level “which-next-action-is-better?” tests, then verify them with multiple expert models so PRMs can be compared fairly.
  • How it works (step by step):
    1. Start with agent trajectories from diverse tool benchmarks.
    2. Create step-level pairs: same history, one correct action, one plausible-but-wrong action, plus tool metadata.
    3. Build pairs in two ways: offline (local swaps near gold) and online (catch real first errors in free runs).
    4. Use a panel of strong LLMs to verify labels, keeping only reliable cases.
    5. Evaluate PRMs by asking: do they pick the correct action?
  • Why it matters: Without step-level, reliable cases, we can’t tell if a PRM truly helps an agent choose the right next move.

🍞 Bottom Bread (Anchor): The benchmark asks, “Given this file-system history and tool rules, is ‘cd into folder’ or ‘cp with absolute path’ the right next step?” A good PRM picks “cd,” and the benchmark records that win.

Multiple Analogies:

  1. GPS Turn-by-Turn: Instead of grading only whether you reach the destination, ToolPRMBench checks each turn. Did you choose the right turn now?
  2. Cooking Coach: Not just “Was the cake tasty?” but “Did you preheat? Measure flour correctly? Mix before baking?”
  3. Lego Instructions: At each step, ToolPRMBench asks if you picked the correct next brick, not just whether the spaceship looks right at the end.

Before vs. After:

  • Before: PRM evaluation was sparse, web-centered, or final-outcome-based, hiding where agents tripped.
  • After: ToolPRMBench offers diverse tools, clean step-level pairs, and verified labels, exposing strengths and weaknesses of PRMs clearly.

🍞 Top Bread (Hook): You know how trying a few puzzle pieces and choosing the one that clicks speeds you up?

🥬 Filling (The Actual Concept — Reward-Guided Search):

  • What it is: A strategy where the agent considers multiple candidate next steps and uses a PRM to pick the best one.
  • How it works (step by step):
    1. Generate several action candidates.
    2. Score each with the PRM.
    3. Keep high-scorers and prune low ones.
    4. Repeat to build a strong trajectory.
  • Why it matters: It avoids committing early to a bad plan, improving reliability.

🍞 Bottom Bread (Anchor): When answering “What’s the next file command?”, the PRM helps the agent pick “cd then cp” over “cp with wrong path,” preventing a dead end.

Why It Works (intuition):

  • Long tasks need credit assignment: knowing which step helped or hurt.
  • Step-level comparisons sharpen a PRM’s sense of tool rules (right tool, right parameters, right order).
  • Diverse APIs and verified labels prevent overfitting to one environment.

Building Blocks (Sandwiches for new terms):

🍞 Hook: Picture a movie split into frames; focusing on one frame lets you see a small mistake clearly. 🥬 Concept — Offline Sampling:

  • What it is: Create a local mistake near a known-good step to form a clean pair.
  • How it works: (1) Follow the gold path up to step t; (2) sample an alternative action at t; (3) if it differs in meaning from gold, keep the pair; (4) continue gold afterward (no environment change).
  • Why it matters: It isolates single-step errors without the chaos of later steps. 🍞 Anchor: With a correct ‘search files’ step, the alternative might use the wrong filter. The pair tests if the PRM prefers the gold filter.

🍞 Hook: Free play at the arcade lets you see where you really mess up. 🥬 Concept — Online Sampling:

  • What it is: Let the agent run freely and convert its first real mistake into a step-level pair.
  • How it works: (1) Generate a full trajectory; (2) keep failed runs; (3) use an LLM to find the first wrong step and propose a fix; (4) form the pair.
  • Why it matters: Catches realistic, chained errors that happen in the wild. 🍞 Anchor: An agent copies files before changing to the right directory. The pair tests “copy-now” (wrong) vs. “cd-then-copy” (right).

🍞 Hook: Three referees beat one at spotting fouls. 🥬 Concept — Multi-LLM Verification:

  • What it is: A voting team of strong LLMs judges which action is better.
  • How it works: (1) Each model votes; (2) keep strong-agreement pairs; (3) discard noisy ones; (4) human spot-check.
  • Why it matters: Cleaner data → fairer evaluation. 🍞 Anchor: GPT-5, Claude-4.5-haiku, and Gemini-2.5-flash all agree which action is better; that pair enters the test set.

🍞 Hook: Sometimes, learning by doing (with rewards) beats memorizing answers. 🥬 Concept — Reinforcement Learning (RL) with GRPO:

  • What it is: A training method where the PRM policy is nudged to pick the correct action more often by comparing groups of sampled outputs.
  • How it works: (1) Sample several reasoning-and-choice outputs; (2) give reward 1 if the choice matches ground truth, else 0; (3) update the policy using group-relative advantages; (4) repeat.
  • Why it matters: Builds robust decision boundaries and resists spurious patterns. 🍞 Anchor: ToolPRM-GRPO learns to consistently choose “cd-then-copy” across varied folders, not just the training ones.

Put together, ToolPRMBench’s insight is simple but powerful: test the judge at the exact place decisions happen—the next step—and make those tests trustworthy and diverse.

03Methodology

At a high level: Diverse tool tasks → (A) Trajectory sampling (offline + online) → (B) Candidate step-level pairs → (C) Multi-LLM verification and filtering → (D) Verified benchmark → (E) Train PRMs (Base, CoT, GRPO) → (F) Evaluate models by pairwise accuracy.

Step A: Trajectory Sampling 🍞 Hook: Think of pausing a game and trying a different button press at that exact moment. 🥬 Concept — Offline Sampling:

  • What happens: The agent follows the gold history up to step t. At t, we sample an alternative action. We don’t let this alternative change the environment; after t, we resume the gold steps.
  • Why this step exists: It isolates single-step differences, avoiding later-step chaos and making the comparison clean.
  • Example: Gold action uses tool “find” with name_contains = "test". The sampled action tries name_contains = "tes" or the wrong key. The pair tests if a PRM prefers the precise gold. 🍞 Anchor: Like testing whether “turn left now” is better than “turn slightly left” while keeping the rest of the route identical.

🍞 Hook: Free-roaming shows true habits, good or bad. 🥬 Concept — Online Sampling:

  • What happens: Let the agent solve the task freely. Keep failed runs. An annotator LLM finds the first incorrect step and proposes the corrected action; this creates a pair.
  • Why this step exists: Real agents make cascaded mistakes; catching the first wrong turn mirrors reality.
  • Example: The agent tries to copy files to a folder without first cd-ing into the required directory per tool rules. The corrected action is “cd …” 🍞 Anchor: Like rewatching gameplay to find the moment the player first missed a jump, then suggesting the right jump at that frame.

Step B: Construct Candidate Pairs

  • What happens: Each sample is (history at step t, chosen_action, rejected_action, tool_metadata). For offline, chosen is gold; for online, chosen is the LLM-proposed fix.
  • Why this step exists: PRMs need apples-to-apples comparisons: same history, two actions, one must be better.
  • Example with data: History shows user asked to find and then copy files. Tool metadata says cp requires both paths local to current directory. Actions: (A) cp with long absolute paths (violates rule) vs. (B) cd then cp (meets rule).

Step C: Multi-LLM Verification and Filtering 🍞 Hook: Ask three librarians to confirm which book edition you need; if they all agree, you’re safe. 🥬 Concept — Multi-LLM Verification:

  • What happens: GPT-5, Claude-4.5-haiku, and Gemini-2.5-flash independently vote which action is better. Majority wins; unanimous-yes pairs are kept; unanimous-no are dropped; mixed cases may get human checks.
  • Why this step exists: It reduces label noise so PRMs train and eval on trustworthy data.
  • Example: All three judges prefer “cd then cp” over “cp now,” matching the tool’s constraint. A human audit later shows 96% agreement with these labels. 🍞 Anchor: Like a panel grading a science fair project; if all agree on the winner, the decision is solid.

Step D: Finalize ToolPRMBench

  • What happens: Aggregate verified pairs from multiple benchmarks (ToolTalk, GTA, BFCL, ToolSandbox). Keep diversity in tools, errors (wrong tool, wrong params, should-chat vs. should-tool), and trajectory lengths.
  • Why this step exists: Broad coverage ensures PRMs aren’t overfit to one environment or error type.
  • Example: GTA offers general APIs; BFCL stresses function-calling rules; ToolSandbox adds stateful, conversational tool use; ToolTalk focuses on dialogue-grounded tool decisions.

Step E: Train PRMs (three variants) 🍞 Hook: Sometimes choosing is enough; sometimes explaining first helps you choose better; sometimes practice with feedback (rewards) builds strongest instincts. 🥬 Concept — ToolPRM-Base (SFT):

  • What happens: Given (history, two actions, tool meta), predict which action is correct. Train with standard supervised fine-tuning to output Action 1 or Action 2.
  • Why this step exists: Establishes a fast, simple judge.
  • Example: Pick the action that obeys cp’s local-path rule. 🍞 Anchor: Like a quiz where you choose A or B.

🍞 Hook: Explaining your thinking often improves your answer. 🥬 Concept — ToolPRM-CoT (SFT with distilled reasoning):

  • What happens: The model first generates a short rationale (distilled from a stronger teacher like GPT-5-mini) and then picks the action.
  • Why this step exists: Lightweight reasoning can sharpen decisions.
  • Example: “cp requires local paths; since we’re outside the folder, we must cd first; choose action B.” 🍞 Anchor: Like showing your math steps before circling the final answer.

🍞 Hook: Practice with points makes you game-ready. 🥬 Concept — ToolPRM-GRPO (RL):

  • What happens: Sample multiple rationale+choice outputs; give reward 1 if the final choice matches ground truth, else 0; update policy with Group Relative Policy Optimization to prefer better choices.
  • Why this step exists: RL builds robustness and better generalization, especially OOD.
  • Example: Across many folders and tools, it learns to respect preconditions consistently. 🍞 Anchor: Like scrimmages where good plays earn points, shaping better instincts.

Step F: Evaluate Models

  • What happens: Present each verified pair; the model must pick the correct action. Report accuracy overall and by subset; analyze scaling, in/out-of-distribution, cost, and search gains.
  • Why this step exists: Simple, fair metric answers the core question: does your PRM choose the right next step?
  • Example metric meaning: 78% accuracy is like getting an A when many others get a C+ on the same tough test.

Secret Sauce (what’s clever):

  • Dual sampling (offline+online) captures both clean, local contrasts and realistic chained mistakes.
  • Multi-LLM verification slashes label noise, confirmed by 96% human agreement.
  • Step-level, tool-aware pairs generalize across diverse APIs, surfacing what generic PRMs miss (tool constraints, parameter formats, and state transitions).

04Experiments & Results

🍞 Hook: When you try out for a team, coaches don’t just ask your final score—they watch each move to see if you choose the right play.

🥬 Concept — The Test (What they measured):

  • What it is: Given the same history and two candidate actions, can a model pick the correct next step?
  • How it works: For every verified pair in ToolPRMBench, the model chooses Action 1 or Action 2; accuracy is the percent of times it chooses correctly.
  • Why it matters: This directly measures whether a PRM can guide an agent at the exact moment decisions are made. 🍞 Anchor: Like a multiple-choice drill where each question is a single move in a larger game.

Benchmarks and Competitors:

  • Datasets: GTA (general APIs), ToolTalk (conversational tool use), BFCL (function-calling rules), ToolSandbox (stateful, interactive tools).
  • Competitors: API-based LLMs (GPT-5, Claude-4.5-haiku, Gemini-2.5-flash), open-source LLMs (Qwen3, LLaMA-3), general PRMs (math/web), and tool-specific PRMs (ToolPRM-Base/CoT/GRPO).

Scoreboard (with context):

  • API-based LLMs top many subsets. Example: GPT-5 hits very strong accuracy on GTA; Claude-4.5-haiku and Gemini-2.5-flash are also strong. Think “A to A+ grades.”
  • Tool-specific PRMs shine among non-API models. ToolPRM-GRPO averages about 78.6%, which is like an A−, and it often beats larger open LLMs that only get C+/B−.
  • General PRMs for math or web often hover near or below mid-50s on average—like a C—showing weak transfer to tool-specific constraints.
  • Bigger base models help (Qwen3/LLaMA-3 scaling improves accuracy), but size alone doesn’t match tool-specialized PRMs.

Surprising/Important Findings:

  1. RL wins on generalization: ToolPRM-GRPO improves in both in-distribution and out-of-distribution tests, while SFT-only models (Base, CoT) drop sharply OOD. That’s like staying calm even when the playbook changes.
  2. Good PRM scores predict real search gains: In meta-evaluation, models that score higher on ToolPRMBench give bigger boosts in best-of-n search; models below ~50% can hurt performance—like a bad coach calling the wrong plays.
  3. Synthetic data is a mixed bag: Injected mistakes help a lot in GTA (big jumps), but barely help or slightly hurt in ToolTalk. Realism and task match matter.
  4. Cost vs. performance: API LLMs are great but pricey; ToolPRMs deliver strong accuracy at much lower per-call cost, making them practical judges for day-to-day agent runs.

Concrete Results Snapshots:

  • API LLMs: Near the top across subsets; e.g., strong 80–90%+ on GTA/ToolTalk, lower on BFCL (harder, rule-heavy), but still competitive.
  • ToolPRM-GRPO: Best average among non-API models (~78.6%), robust across datasets, and often beats even some API models on certain subsets.
  • Open LLM scaling: Qwen3 from 1.7B → 14B climbs from low-40s to ~63%, showing capacity helps but doesn’t close the gap with specialized PRMs.

Takeaway: If you need a reliable judge to guide tool-using actions at inference time, ToolPRM-GRPO-style training is a sweet spot of accuracy, robustness, and cost. ToolPRMBench’s step-level accuracy not only ranks PRMs but also predicts their real impact on search-time performance.

05Discussion & Limitations

Limitations (be specific):

  • Coverage: ToolPRMBench spans several major tool-use datasets but doesn’t yet include newer MCP-based ecosystems. That limits exposure to certain real-world protocols and dynamic tool registries.
  • Focus: The main metric is intrinsic step-level discrimination, not full end-to-end agent success under very large training or search budgets. Some inference-time scaling strategies weren’t exhaustively tested.
  • Labeling: Multi-LLM voting plus human spot-checks reduce noise (96% agreement), but any automated labeling pipeline can still miss subtle, domain-specific valid alternatives.
  • Pair format: Each test is a binary comparison. Real agents may face more than two plausible actions or need longer lookahead; pairs are a simplification (useful, but not complete).

Required Resources:

  • To reproduce training: 8× H20-class GPUs (or similar) for full-parameter SFT and GRPO; LLaMA-Factory/TRL stacks; access to strong teacher models (for CoT distillation) and API credits for verification.
  • To use the benchmark: Modest compute—models only need to choose between two actions per case; inference cost is low for open-source PRMs.

When NOT to Use:

  • If your agent acts in domains with tools and constraints very unlike those in ToolTalk/GTA/BFCL/ToolSandbox (e.g., robotics control protocols or specialized medical devices), direct transfer may be weak.
  • If you need multi-action or plan-level judgments (ranking 5–10 complex candidates or scoring multi-step plans holistically), pairwise, single-step tests may be too narrow.
  • If your environment requires strict formal verification (e.g., safety-critical), a learned PRM alone is not sufficient without additional guarantees.

Open Questions:

  • How to best integrate MCP-style, rapidly changing tool catalogs without exploding data collection costs?
  • Can we design synthetic error generators that more closely mirror real agent failures across domains (beyond GTA-like wins)?
  • What’s the optimal blend of SFT, CoT distillation, and RL (GRPO or alternatives) for maximal OOD robustness per unit cost?
  • How can we extend from binary pairs to multi-candidate ranking and multi-step plan scoring while keeping labels reliable and affordable?
  • Can test-time compute scaling (e.g., adaptive sample-and-rerank with PRMs) be made cost-efficient enough for always-on enterprise agents?

06Conclusion & Future Work

Three-Sentence Summary:

  • ToolPRMBench turns long tool-using agent runs into verified, step-level, two-action tests so we can fairly measure whether a model picks the right next move.
  • With both offline (local swaps) and online (first real error) sampling plus multi-LLM verification, it creates a clean, diverse benchmark across GTA, ToolTalk, BFCL, and ToolSandbox.
  • Experiments show that bigger models help, but tool-specialized PRMs—especially with RL via GRPO—are the most robust and cost-effective step judges, and their scores predict real search-time gains.

Main Achievement:

  • Establishing the first broad, step-level benchmark for process reward models in diverse tool-using settings, with a reliable verification pipeline that correlates with real-world reward-guided search improvements.

Future Directions:

  • Expand to MCP-based and other evolving tool ecosystems; add multi-candidate and plan-level judging; refine synthetic error generation for better realism; explore more RL variants and test-time compute strategies.

Why Remember This:

  • Because reliable step-level judges are the “compass” that keep tool-using agents on track. ToolPRMBench gives the community a trusted way to build and compare those compasses, so agents can act safely and effectively in the real world.

Practical Applications

  • •Evaluate your in-house PRM by running it on ToolPRMBench and tracking step-level accuracy before deployment.
  • •Plug a high-scoring PRM (e.g., GRPO-trained) into your agent’s best-of-n sampler to safely prune bad actions in real time.
  • •Use the multi-LLM verification idea to clean your own step-level datasets for new tools or domains.
  • •Fine-tune a lightweight PRM on ToolPRMBench subsets that match your environment (e.g., file system ops) to lower inference costs.
  • •Adopt online sampling in your sandbox to capture realistic first-error cases from your agent and convert them into training pairs.
  • •Combine CoT distillation with SFT to teach smaller PRMs to explain and decide better with minimal overhead.
  • •Run a cost–performance analysis (as in the paper) to select an affordable PRM for continuous monitoring in production.
  • •Use ToolPRMBench metrics as a proxy to predict gains from reward-guided search in your own tasks before heavy integration.
  • •Design synthetic error injectors for your stack (start with GTA-like wins) and validate them against ToolPRMBench for realism.
  • •Set up an OOD evaluation split (new tools/tasks) to stress-test PRM robustness before rollout to new product surfaces.
#process reward model#tool-using agents#offline sampling#online sampling#multi-LLM verification#reward-guided search#GRPO#chain-of-thought distillation#step-level evaluation#benchmarking#best-of-n search#tool metadata#trajectory analysis#out-of-distribution generalization
Version: 1