WideSeek-R1: Exploring Width Scaling for Broad Information Seeking via Multi-Agent Reinforcement Learning
Key Summary
- ā¢WideSeek-R1 teaches a small 4B-parameter language model to act like a well-run team: one leader plans, many helpers work in parallel, and everyone learns together with reinforcement learning.
- ā¢Instead of only thinking longer (depth scaling), it thinks wider (width scaling) by adding more subagents that search at the same time with separate notes so they donāt mix up information.
- ā¢A new training recipe (multi-agent RL with shared rewards, group normalization, and dual-level reweighting) lets the leader and helpers improve together without arguing over credit.
- ā¢On the WideSearch benchmark, WideSeek-R1-4B reaches 40.0% item F1, about as good as the giant DeepSeek-R1-671B, using roughly 170Ć fewer parameters.
- ā¢Performance keeps going up as you add more parallel subagents, showing true width scaling, while depth scaling alone quickly plateaus.
- ā¢A 20k-task dataset of broad, table-style queries was auto-built with self-consistency checks to train agents to gather and organize many facts reliably.
- ā¢Ablation studies show you need both a trained leader and trained helpers, and mixing āwideā and ādeepā training data works best.
- ā¢The system stays strong on standard QA tasks, beating several larger multi-agent baselines, so it doesnāt trade away general reasoning.
- ā¢The work highlights a shift from āhow smart is one agent?ā to āhow well does the team organize itself?ā for broad information seeking.
- ā¢This approach can make powerful research assistants more affordable and faster by coordinating many small agents instead of relying on one huge model.
Why This Research Matters
Many real tasks are about breadthāgathering lots of small facts across many itemsāand this system is built for exactly that. By coordinating many small agents with clean notes, it delivers strong, organized tables faster than a single agent can. It shows that you donāt always need a massive model; a well-trained team of smaller ones can compete, cutting costs and making advanced AI more accessible. This helps businesses, journalists, scientists, and students compile reliable datasets with fewer mistakes. It also opens a new research path: training better teams, not just bigger solo models.
Detailed Explanation
Tap terms for definitions01Background & Problem Definition
š Hook: Imagine your class needs to create a giant world atlas in one day. If only one student tries to do everything, theyāll run out of time and start mixing notes. But if the class splits upāsome find capitals, others gather populationsāand one leader organizes the pieces, you finish faster and cleaner.
š„¬ The Concept (Multi-Agent Systems): A multi-agent system is when many AI helpers each do part of a job instead of one AI doing everything. How it works:
- A big task is split into smaller, independent jobs.
- Different agents take different jobs.
- They work at the same time and share results. Why it matters: One agentās memory gets messy (context pollution) and time runs out when tasks are done one-by-one. š Anchor: Making a movie: writers draft the story, artists design scenes, sound editors add musicāall in parallelāso the film ships on time.
š Hook: You know how a video game gives you points for doing well? That makes you try better strategies.
š„¬ The Concept (Reinforcement Learning): Reinforcement learning (RL) teaches an AI by giving it rewards when its choices lead to good outcomes. How it works:
- The AI tries actions.
- It gets a score (reward) based on results.
- It updates its strategy to earn higher rewards next time. Why it matters: Without rewards, the AI doesnāt know if itās improving. š Anchor: Practicing basketball: each swish is a reward, so you repeat what worked.
š Hook: Picture a relay race where each runner learns when to pass the baton and how fast to run to help the entire team win.
š„¬ The Concept (Multi-Agent Reinforcement Learning): MARL is RL for a whole team of AIs that learn to cooperate. How it works:
- Many agents act in the same task.
- The team gets a shared score.
- Everyone updates their behavior to boost the shared score. Why it matters: Without learning as a team, agents can work at cross-purposes and hurt the final result. š Anchor: Soccer practice: strikers, midfielders, and defenders learn patterns that raise the teamās chance of scoring.
The world before: Recent LLM advances focused on ādepth scaling.ā
š Hook: You know how you can sometimes solve a tough math problem by thinking through many careful steps?
š„¬ The Concept (Depth Scaling): Depth scaling means giving one AI more turns and longer reasoning to solve a job. How it works:
- The single agent plans, searches, and thinks over many steps.
- It accumulates notes in one long conversation.
- It slowly builds to an answer. Why it matters: For very deep, single questions, more careful steps can helpābut it also risks long, messy contexts. š Anchor: One detective reading every clue in a giant fileāthorough, but slow and easily overwhelmed.
The problem: Broad information-seeking tasks (like āmake a table with dozens of items and attributesā) are about breadth, not just depth. Two big pains appear:
- Context pollution: one agentās notebook fills with unrelated bits from many subtasks.
- Sequential bottleneck: independent subtasks are done one at a time.
Failed attempts:
- Hand-crafted workflows: designers hard-code who does what, but itās inflexible as tasks or agent counts change.
- Turn-taking systems: agents wait for each other, which ruins parallel speed.
- Sample-more tricks (best-of-N): repeat the same plan many times to pick a good tryāhelps reliability but not breadth, because subtasks arenāt split.
The gap: We need a system that learns to organize a team and truly parallelize work, keeping each subtaskās notes separate while still producing a single, clean table at the end.
š Hook: Think of an orchestra conductor who knows who should play when, and keeps each section on their own sheet music.
š„¬ The Concept (Orchestration): Orchestration is the leaderās skill to assign clear subtasks, time them, and combine the results. How it works:
- Break the big job into independent parts.
- Send each part to a helper with instructions.
- Collect and merge the answers. Why it matters: Without orchestration, helpers duplicate work, miss pieces, or collide. š Anchor: A kitchen head chef assigns appetizers, mains, and desserts to different cooks, then plates the full meal.
Real stakes: This isnāt just about benchmarks. It affects:
- Business research: compiling competitive product tables.
- Journalism: summarizing many sources into clean datasets.
- Science: aggregating related studiesā attributes.
- Education: building resource catalogs.
- Everyday life: planning trips with multiple cities, budgets, and dates.
WideSeek-R1 was built to fill this gap: a learned leader-helper team that scales āwidth,ā not just ādepth,ā to conquer broad info-seeking without drowning in its own notes.
02Core Idea
š Hook: Imagine building a giant LEGO city: one kid plans the neighborhoods, many friends each build a block at the same time, and together they finish faster and cleaner than any single super-builder could.
š„¬ The Concept (Aha!): Train one lead agent and many parallel subagents together (with MARL) so the leader learns to split the job and the helpers learn to search well, all using separate notes and specialized tools. How it works (high level):
- The lead agent breaks the big question into independent subtasks.
- Each subagent searches and summarizes in its own clean mini-notebook (context isolation).
- The lead collects the pieces and writes a structured final table.
- A shared team reward and a special RL recipe teach everyone to cooperate. Why it matters: Without learned team play, parallel workers create chaos or idle; without isolation, notes get tangled; without shared rewards, they donāt improve as a group. š Anchor: A newsroom editor (lead) assigns reporters (subagents) different beats, each files their clean report, and the editor assembles a front-page story.
Three analogies:
- Restaurant kitchen: head chef (lead) assigns stations; line cooks (subs) prepare dishes at once; the pass (final table) is neat and on time.
- Bee colony: a queen (lead) signals tasks; foragers (subs) gather nectar in different fields; honeycomb (table) fills quickly without mixing pollen.
- School project: a team captain (lead) splits topics; teammates (subs) research in separate docs; captain merges into a tidy report.
Before vs After:
- Before: One big agent thinks longer; info gets mixed; tasks queue up; hand-crafted multi-agent flows struggle to scale.
- After: A trained leader delegates; subagents run truly in parallel; each keeps a clean context; the final table is stronger and faster.
Why it works (intuition, not equations):
- Context isolation is like separate notebooks: fewer mix-ups.
- Parallel execution is like many hands making light work.
- A shared backbone model keeps the team speaking the same ālanguage,ā so prompts and summaries mesh.
- Group-normalized rewards and dual-level reweighting stabilize learning so no single chatterbox (a long response or too many subagents) dominates training.
- The dataset forces breadth: many entities, columns, and row counts train the system to cover and organize widely.
Building blocks (each is a mini concept):
š Hook: Picture a smart project manager who only gives out tasks and waits for results.
š„¬ The Concept (Lead-AgentāSubagent Framework): One leader assigns subtasks; many helpers solve them. How it works:
- Leader uses a single tool: call_subagent.
- Creates clear prompts for subagents.
- Waits, then consolidates findings into the answer. Why it matters: Without a leader, helpers overlap or leave gaps. š Anchor: A librarian assigns shelf sections to volunteers and later checks all books are in order.
š Hook: Think of adding more checkout counters at a grocery store.
š„¬ The Concept (Width Scaling): Improve performance by adding more parallel agents instead of only adding more steps to one agent. How it works:
- Keep turn count fixed.
- Increase number of subagents working at once.
- Merge their outputs. Why it matters: Without width, independent subtasks block each other in line. š Anchor: Ten cashiers serve shoppers faster than one cashier taking ten times as long.
š Hook: You know how you donāt bring your math notes into art class?
š„¬ The Concept (Context Isolation): Each subtask has its own clean workspace so irrelevant info doesnāt spill over. How it works:
- Separate contexts per subagent.
- Strip out unnecessary āthinkingā when reporting back.
- Leader only sees final, tidy summaries. Why it matters: Without isolation, errors cascade across subtasks. š Anchor: Separate Google Docs per teammate avoid messy edits.
š Hook: Imagine a scoreboard that only cares about the final team result, not who shouted the most.
š„¬ The Concept (Group Advantage + Dual Reweighting): A training trick that shares one team reward across agents, then balances influence by tokens and by agents. How it works:
- Compute one outcome score per multi-agent rollout.
- Normalize it within a group (fair comparison).
- Reweight by token counts (longer, meaningful turns count) and by number of agents (crowds donāt drown others). Why it matters: Without this, agents could game the system by being overly long or by spawning too many helpers. š Anchor: In a relay, the final time trains the whole team, and coaching attention is fairly split among runners.
Put together, these pieces let WideSeek-R1 turn a broad question into a crisp, structured tableāfaster and with fewer mix-ups than a lone agent trying to juggle everything.
03Methodology
At a high level: Input query ā Lead agent decomposes (call_subagent) ā Subagents search in parallel (search/access) ā Subagents return clean summaries ā Lead agent synthesizes ā Output structured table.
Weāll walk through each step like a recipe, and introduce new ideas with the Sandwich pattern.
- Input and Goal
- What happens: The system receives a broad info-seeking query that asks for a table (e.g., āList Ivy League universities with name, city, and founding yearā).
- Why it exists: The request needs many facts across multiple entities, so breadth and organization matter.
- Example data: Ivy League universities ā rows are schools; columns include name, city, founding year.
- Lead Agent: Decompose and Delegate
š Hook: Think of a coach who writes a play, sends players to their spots, and waits for the pass back.
š„¬ The Concept (Task Decomposition): Breaking a big job into independent subtasks that can be done at the same time. How it works:
- Parse the requested tableās schema (columns) and coverage (row count or full list requirement).
- Identify independent units (e.g., one row per entity), and split them into prompts.
- Create up to N subtasks in one turn via call_subagent. Why it matters: Without decomposition, the leader canāt distribute work or ensure full coverage. š Anchor: For 8 Ivy League schools, create 8 prompts: āFind city and founding year of Harvard,ā ā...of Yale,ā etc.
- The leader only has one tool: call_subagent. This keeps the leaderās context clean and prevents context pollution.
- The leader waits until subagents finish, then either launches another parallel wave or finalizes the answer.
- Subagents: Parallel Information Seeking
š Hook: Picture many reporters going to different libraries at once.
š„¬ The Concept (Parallel Execution): Multiple helpers work simultaneously on separate tasks. How it works:
- Each subagent receives one subtask and has its own context.
- Subagents use tools: search (find snippets + URLs) and access (open a chosen URL and summarize details).
- They iterate: search to discover, access to confirm and extract, then return a concise, structured summary. Why it matters: Without parallel execution, total time balloons and context gets tangled. š Anchor: Ten subagents each research one universityās city and founding year, finishing in one or two turns instead of ten turns.
š Hook: You know how you first Google a topic, then click a promising link to read more closely?
š„¬ The Concept (Tool Use: search + access): Two-step evidence gathering: broad discovery, then deep reading. How it works:
-
search: return candidate URLs and brief snippets.
-
access: fetch and summarize content from a selected URL tailored to the subtask.
-
Summarize findings with citations back to the leader. Why it matters: Without this two-step flow, you either miss coverage (no discovery) or waste time reading random pages (no focus). š Anchor: Search āHarvard founding year,ā then access the official page or a reliable wiki to confirm itās 1636.
-
Returning Results to the Leader
- What happens: Subagents send compact summaries (thinking removed) back to keep the leaderās context small and neat.
- Why it exists: The leader must merge many pieces and avoid drowning in long notes.
- Example: Each subagent returns: School=Harvard, City=Cambridge, MA, FoundingYear=1636, Evidence=URL.
- Leader: Synthesize and Finalize Table
- What happens: The leader checks coverage (all rows present?), consistency (any conflicts?), and formats a single Markdown table.
- Why it exists: To deliver exactly what the user asked forāone clean, complete table.
- Example: 8 rows, 3 columns; if any school is missing, the leader launches another round for just the missing bits.
- Training the Whole Team (MARL)
š Hook: Think of a class graded on a group project, with fair grading so both the presenter and the researchers improve.
š„¬ The Concept (Outcome Reward): One score based on the final answerās quality (e.g., table cell accuracy), plus small bonuses for correct format and using tools. How it works:
- Compare the generated table to ground truth to get an Item F1 score.
- Add rewards for proper Markdown format and at least one tool use; subtract a penalty if responses are too long.
- Share this single rollout reward with all agents in that attempt. Why it matters: Without a clear, verifiable outcome, the team canāt learn what really leads to better answers. š Anchor: If the final tableās cells match well, the whole team gets a higher score; if formatting breaks, the score drops to zero for that run.
š Hook: Imagine normalizing test scores so teams in the same class are compared fairly, not across wildly different conditions.
š„¬ The Concept (Group-Normalized Advantage + Dual Reweighting): Stabilizes training by comparing runs fairly and balancing influence. How it works:
-
Normalize rewards across multiple rollouts in a group (fair baseline).
-
Token-level reweighting: longer, contentful turns contribute proportionally (not drowned by averaging).
-
Agent-level reweighting: average over agents so spawning more helpers only helps if quality improves. Why it matters: Without these, the system might game the reward by being verbose or by spawning excessive subagents. š Anchor: In practice, models that simply write more donāt win; models that write better and coordinate better do.
-
Data to Teach Breadth
- Auto-built 20k-task dataset: ⢠Stage 1: Turn intents into strict, schema-bound queries (columns fixed; 10ā50 rows targeted). ⢠Stage 2: Generate two independent answers and identify unique column(s) for row matching. ⢠Stage 3: Keep only pairs that agree (high cell match) and are hard enough (not tiny or trivial).
- Why it matters: Without broad, table-focused data, the team wouldnāt practice wide coverage and reliable synthesis.
- Example: āTop 20 countries by population (2025) with Rank, Country, Population (millions, 1 decimal).ā
- Practical Controls
- Shared 4B backbone (Qwen3-4B, thinking mode) for both leader and subagents keeps the team fluent.
- Limits: up to 10 parallel subagents/turn; leader up to 10 turns; subagent up to 20 turns.
- Tools: often backed by an offline Wiki2018 KB during training to reduce cost and keep evaluation consistent.
Secret sauce (why this recipe is clever):
- Narrow, clean leader tool (call_subagent) avoids self-mess.
- True parallel subagents with isolated contexts prevent cross-contamination and speed coverage.
- A group-outcome reward plus normalization and dual reweighting keeps training fair, stable, and team-oriented.
- The dataset forces the skills that matter for breadth: consistent schemas, many rows, careful row matching.
04Experiments & Results
The test: WideSearchāa benchmark of 200 broad, table-style tasks (100 English, 100 Chinese). It checks whether the system can gather many attributes across many entities and produce one correct table.
š Hook: Think of grading a big class science fair with three rubrics: every measured number correct (cells), each project fully right (rows), and perfect blue-ribbon projects (entire table).
š„¬ The Concept (Metrics: Item F1, Row F1, Success Rate): How it works:
- Item F1: compares table cellsālike checking each data cell.
- Row F1: compares whole rowsāmust match exactly to count.
- Success Rate: a perfect full-table match. Why it matters: Without multi-level grading, you can miss whether errors are scattered (cells) or whole rows are wrong. š Anchor: If most cells are correct but a few rows are mismatched, Item F1 will look okay, Row F1 will drop, and Success Rate may be zero.
Competition: Single-agent baselines (Qwen3-4B, Search-R1-7B, ASearcher-7B, DeepSeek-R1-671B) and multi-agent baselines (AgentFlow-7B, OWL-8B, MiroFlow-8B, and a Qwen3-4B multi-agent setup). WideSeek-R1-4B is the proposed system.
Scoreboard (with context):
- WideSeek-R1-4B gets 40.0% Item F1 on WideSearchālike getting an A when many others got a C+āand roughly ties the giant DeepSeek-R1-671B while using about 170Ć fewer parameters.
- It beats all 4B and 8B baselines on five of six metrics (Item/Row F1 Avg@4, Max@4, and Success Rate variants).
- Compared to the same backbone used as a basic multi-agent (Qwen3-4B), MARL training adds about +8.8% Item F1āclear proof that learned orchestration and execution matter.
Surprising findings:
- Depth scaling plateaus: letting a single agent take more turns helps at first but quickly saturatesālike studying longer without changing strategy.
- Width scaling shinesābut only if learned: adding more untrained subagents eventually hurts (too much conflicting noise). With WideSeek-R1ās training, performance keeps rising as subagents increase (up to 10 in tests).
- Joint optimization matters: ablations show that training only the leader or only the subagents helps, but training both together is bestāteamwork is learned.
- Data mixing helps: models trained on hybrid (wide + deep) data beat those trained on only-wide or only-deep, showing these skills complement each other.
Standard QA (beyond tables): On seven single-hop and multi-hop QA sets (NQ, TriviaQA, PopQA, 2Wiki, HotpotQA, Bamboogle, MuSiQue), WideSeek-R1-4B maintains strong average performance, surpassing some larger multi-agent systems (like OWL-8B, MiroFlow-8B). This means the system didnāt sacrifice general reasoning to get good at tables.
Behavioral patterns after training:
- More thoughtful, useful tool use: more parallel subagents and a higher fraction of deep-access calls to fetch richer evidence.
- Multi-agent answers use more total turnsāexpected because many helpers are workingābut this turns into better coverage and cleaner final tables.
Bottom line: Width scaling with learned orchestration and parallel execution converts extra compute into real accuracy gains on broad info-seeking, something depth scaling alone couldnāt sustain.
05Discussion & Limitations
Limitations:
- Compute cost: Even with a 4B model, training took about 3,000 H100 GPU-hours. Scaling up or exploring larger backbones would cost more.
- Credit assignment: The training uses one final outcome reward shared by all agents. This is simple and stable, but itās coarse: it canāt always tell whether the leaderās plan or a subagentās execution caused a failure.
- Fixed hierarchy: Training uses a two-layer design (leader ā subagents). Allowing subagents to spawn more subagents (recursion) could help on very complex tasks but destabilized training in early trials.
- Training latency: Synchronous rollouts made training stable but slow; most time was spent waiting for long generations.
- Tool coverage: Results depend on the search/access sources; offline KBs are cheaper but may miss some live facts.
Required resources:
- A cluster with enough GPUs for multi-agent rollouts.
- A search + access tool setup (online APIs or a local wiki index).
- The 20k wide dataset and optionally deep QA data for hybrid training.
When NOT to use:
- Very deep, single-focus reasoning tasks where one strong agent with long chain-of-thought is enough (width adds overhead).
- Ultra-low-latency settings where spinning up multiple subagents per query isnāt acceptable.
- Tasks without clear subtask boundaries (hard to decompose cleanly).
Open questions:
- Can we design role-specific rewards to separate orchestration quality from execution quality?
- Can we train dynamic hierarchies (recursive delegation) without destabilizing MARL?
- Will asynchronous rollouts (no waiting for the slowest subagent) keep stability and speed training?
- How to merge conflicting subagent answers betterāconfidence weighting, source reliability scoring, or learned aggregation heads?
- Whatās the best way to balance number of subagents vs. turns per subagent for a fixed compute budget?
06Conclusion & Future Work
Three-sentence summary: WideSeek-R1 trains a leader and parallel helpers to tackle broad, table-style information seeking, focusing on width (more agents at once) instead of only depth (more turns for one agent). A multi-agent RL recipeāwith group-normalized rewards and dual-level reweightingāteaches the team to decompose tasks, search in parallel with clean contexts, and merge results into one structured table. This delivers 40.0% item F1 on WideSearch with a 4B model, rivaling far larger single-agent systems, and performance keeps improving as more subagents are added.
Main achievement: Proving that learned orchestration plus true parallel execution unlocks real width scalingāturning more agents into more correct answers, not just more noise.
Future directions:
- Role-specific rewards to pinpoint and fix planning vs. execution errors.
- Dynamic, possibly recursive hierarchies for harder tasks.
- Asynchronous training to cut rollout latency.
- Smarter evidence merging (confidence and source reliability) and better safety guardrails.
Why remember this: It marks a shift from asking, āHow deep can one agent think?ā to āHow well can a team coordinate?āāa practical path for faster, cheaper, and more reliable broad information seeking with small-but-many agents acting in concert.
Practical Applications
- ā¢Competitive product analysis: build side-by-side feature/price tables across dozens of brands.
- ā¢Market landscaping: map companies, locations, and founding years for an industry overview.
- ā¢Academic literature reviews: tabulate studies, methods, datasets, and results.
- ā¢Travel planning: compile city-by-city tables of attractions, hours, prices, and transit options.
- ā¢Public data dashboards: gather municipal metrics (schools, parks, services) into unified tables.
- ā¢E-commerce cataloging: enrich product listings with specs, certifications, and availability.
- ā¢News research: summarize multiple sources into a clean fact table for an investigative story.
- ā¢Education resources: build course topic tables with references, examples, and difficulty levels.
- ā¢Healthcare info aggregation: list clinics with services, locations, hours, and contact details.
- ā¢Policy comparison: tabulate laws or regulations across regions with dates and key clauses.