VLS: Steering Pretrained Robot Policies via Vision-Language Models
Key Summary
- âąRobots often learn good hand motions during training but get confused when the scene or the instructions change at test time, even a little bit.
- âąThis paper introduces VisionâLanguage Steering (VLS), which adapts a frozen robot policy on the fly without any retraining.
- âąVLS asks a visionâlanguage model (VLM) to understand the scene and the instruction, pick important 3D keypoints, and write tiny differentiable reward functions that score action proposals.
- âąThese rewards gently push a diffusion or flow-matching policy during denoising, so the sampled actions fit the new scene and task.
- âąTo stay both smart and flexible, VLS mixes three tricks: gradient guidance (precise nudges), FeynmanâKac resampling (keep the good candidates), and repulsive forces (keep candidates diverse).
- âąVLS also runs in closed loop: it adjusts guidance strength automatically and switches task stages with a Schmitt-trigger rule to avoid flip-flopping.
- âąAcross benchmarks, VLS beats prior steering methods, improving success by up to 31% on CALVIN and 13% on LIBERO-PRO.
- âąOn a real Franka robot, VLS keeps working under new object looks, new positions, and new goalsâno policy finetuning needed.
- âąMain tradeoff: better performance needs more sampling, which adds latency; still, itâs practical and training-free.
- âąBig idea: donât reteach the robotâsteer the skills it already has to fit the new situation.
Why This Research Matters
Homes, hospitals, and factories change every day: objects move, new tools appear, and instructions vary. VLS lets a robot handle these changes on the spot by steering what it already knows, instead of sending it back for retraining. This makes robots faster to deploy, safer around people and clutter, and cheaper to maintain. Because rewards are written by a VLM from live camera views and language, the system can flex to new scenes and goals without new datasets. The training-free nature also lowers the barrier for smaller labs and companies to achieve robust behavior. In short, VLS helps turn reliable, adaptable robot help from a research demo into an everyday reality.
Detailed Explanation
Tap terms for definitions01Background & Problem Definition
You know how: Once you learn to place a cup in the middle of a table, you can also place it near the edge or inside a cabinet without learning from scratch. Your hands already know how to move; you just adjust for the new spot. What the world was like before: Robots learned from lots of examples (called imitation learning). This works great when the test looks like the training: same table, same object spot, same instruction. Recent robot policies based on diffusion or flow-matching are especially good at this. But small changes at test timeâlike the target moves a bit, thereâs clutter, or the words in the instruction changeâcan make the robot fail. Itâs not because the robot lacks basic motor skills; itâs because those skills are tied to the training setup. The problem: Out-of-distribution (OOD) situations. The robot faces a pictureâinstruction pair it never saw before. The old way is to retrain or finetune. Thatâs expensive and odd because the needed motions are already in the robotâs headâthe robot just canât pick and shape them for the new scene. Failed attempts: People tried three main things.
- Re-optimize with VLMs at test time: build a new plan by rolling out simulations or searching. Powerful, but slow and not great for real-time control.
- Use critics or dynamics models to guide sampling: helpful, but you end up optimizing toward the critic/model, which can push the robot off what the base policy knows best.
- Filter or select with VLMs or humans: you can pick among candidates, but the feedback is sparse (approve/reject), so itâs sample-inefficient and struggles with fine, continuous corrections. The gap: We need a way to adapt during inference that (1) keeps the base policy frozen, (2) understands open-world scenes and language, and (3) gives smooth, dense nudges to the action samplingâlike a GPS for the robotâs already-learned skills. Real stakes: In homes and factories, objects move, new items appear, and instructions vary. A robot that needs retraining each time is impractical. A robot that can adjust on the fly is safer (avoids bumping into new clutter), faster to deploy (no retraining), and more useful (handles new tasks and layouts). Thatâs what this paper delivers: VisionâLanguage Steering (VLS), a way to steer what the robot already knowsâright when it acts. â New concept 1 â Out-of-Distribution (OOD) đ Hook: Imagine practicing basketball in one gym, then playing in another with different lighting and floor lines. Your moves are fine; the setting changed. đ„Ź The Concept: OOD means the test scene or instruction doesnât match the training examples.
- How it works: 1) The robot sees a new layout or wording; 2) Its policy expects familiar patterns; 3) Small mismatches snowball; 4) Actions miss their mark.
- Why it matters: Without handling OOD, even good policies stumble when anything shifts. đ Anchor: The robot learned âput the block in the center.â Now the center is occupied, and the edge is requestedâit hesitates or collides. â
02Core Idea
The âAha!â in one sentence: Donât reteach the robotâuse a VLM to write small, differentiable rewards that gently steer the frozen policyâs sampling so its actions fit the new scene and task. Three analogies for the same idea:
- GPS for skills: The robot already knows how to drive; VLS is the GPS that says âturn a little left nowâ to match the new road.
- Music conductor: The orchestra (policy) knows all the notes; the conductor (VLS) cues dynamics and timing to suit todayâs hall and audience.
- Editing clay: The base policy molds a shape; VLS adds tiny pushes to refine it into the needed figure for this specific scene. Before vs. After:
- Before: Frozen policies sampled actions guided only by what they learned, so they stuck to training-era spatial assumptions and instructions.
- After: The same policies get live, scene-aware nudges from rewards that a VLM composes, so samples bend toward success in new layouts and goalsâwithout any weight updates. Why it works (intuition):
- Diffusion/flow policies generate actions by âdenoisingâ noise into trajectories. If you can score a partial trajectory and say âbetter if closer to keypoint X and aligned with Y,â you can push the next denoising step in a helpful direction. Do this each step, and the whole trajectory drifts toward satisfying the current scene and instruction.
- VLMs are good at open-world perception and language. If we ask them to reduce the scene to a small set of keypoints and write differentiable scoring rules (rewards), we can compute gradientsâtiny arrows showing how to adjust actions to improve the score.
- Mixing gradient nudges (precise, local) with resampling and diversity (broad, global) balances stability and exploration. Building blocks (with sandwich explanations): â New concept 2 â VisionâLanguage Models (VLMs) đ Hook: You know how you can look at a picture and read a sentence about it and instantly connect the two? đ„Ź The Concept: A VLM understands images and words together.
- How it works: 1) See the scene; 2) Read the instruction; 3) Identify relevant objects and relations; 4) Output structured hints (like keypoints or rules).
- Why it matters: We need open-world understanding to tell the robot what matters right now. đ Anchor: âPut the mug on the green plate.â The VLM finds the mug, the green plate, and their positions. â New concept 3 â Spatial Keypoints (Grounding) đ Hook: When you aim to hang a picture, you mark the wall spot with a small dot. đ„Ź The Concept: Keypoints are a compact set of 3D spots that capture whatâs important for the task.
- How it works: 1) Detect/segment objects; 2) Extract features; 3) Project to 3D; 4) Cluster to pick a few meaningful points.
- Why it matters: Keypoints shrink the problem so rewards can focus on the right geometry. đ Anchor: For âplace mug on plate,â keypoints might be mug handle tip, mug base center, and plate center. â New concept 4 â Reward Functions (Differentiable) đ Hook: Think of a video game score that goes up when youâre closer to the goal. đ„Ź The Concept: A reward function gives a number showing how well an action proposal fits the taskâand is smooth enough to differentiate.
- How it works: 1) VLM decomposes task into stages; 2) Writes small PyTorch functions that softly reward desired distances/alignments; 3) We compute gradients w.r.t. the action.
- Why it matters: Gradients tell us which way to tweak the action during denoising. đ Anchor: Reward = higher when mug base is above plate center and gripper is level. â New concept 5 â Denoising Diffusion đ Hook: Imagine a blurry photo becoming clear a little bit at a time. đ„Ź The Concept: Diffusion starts from noise and gradually âcleansâ it into a sampleâin our case, an action trajectory.
- How it works: 1) Start with random actions; 2) Predict a denoising step; 3) Repeat many times; 4) End with a plausible action sequence.
- Why it matters: If we can nudge each cleaning step, we can shape the final actions. đ Anchor: Each denoising step moves the gripper path a bit closer to the plate center. â New concept 6 â Flow Matching đ Hook: Think of a river guiding leaves downstream along a smooth path. đ„Ź The Concept: Flow matching learns a velocity field that smoothly moves noise to a clean sample.
- How it works: 1) Define a continuous time; 2) Learn the flow that carries samples; 3) Integrate along time to get an action.
- Why it matters: We can add a guidance vector to the flow to steer where the actions go. đ Anchor: The velocity field points the gripper smoothly toward the plate. â New concept 7 â Gradient-Based Steering đ Hook: Like a coach whispering, âa bit more to the left.â đ„Ź The Concept: Use the rewardâs gradient to slightly adjust the denoising/flow step.
- How it works: 1) Score current action proposal; 2) Compute gradient; 3) Add scaled gradient to the modelâs step; 4) Iterate.
- Why it matters: Tiny, continuous nudges produce big, reliable alignment over time. đ Anchor: The gradient says âreduce distance to plate center,â so the next sample step moves that way. â New concept 8 â FeynmanâKac Resampling đ Hook: When baking cookies, you keep the best-shaped ones and redo the bad ones. đ„Ź The Concept: Keep a batch of action proposals, weight them by reward, and resample so good ones multiply.
- How it works: 1) Compute reward-based weights; 2) Probabilistically copy high-reward proposals; 3) Drop low-reward ones; 4) Continue denoising.
- Why it matters: Prevents wasting time on bad directions and boosts promising ones. đ Anchor: Good mug-to-plate trajectories get more copies; awkward ones get replaced. â New concept 9 â Repulsive Forces (Diversity) đ Hook: Friends spread out at a picnic so everyone has room. đ„Ź The Concept: Add a gentle âspread outâ force so multiple proposals donât collapse to the same idea too soon.
- How it works: 1) Measure pairwise closeness; 2) Push candidates apart early; 3) Keep coverage; 4) Later focus with rewards.
- Why it matters: Diversity finds better solutions in tricky scenes. đ Anchor: Some candidates try over-plate-center, others try near-edgeâthen the best wins via rewards. â New concept 10 â Adaptive Guidance Strength đ Hook: Like turning the steering wheel more when youâre far off course and less when youâre nearly straight. đ„Ź The Concept: Automatically scale how strong the gradient nudges are based on recent reward.
- How it works: 1) Compare current reward to a baseline; 2) If worse, increase guidance; 3) If better, ease off; 4) Repeat each chunk.
- Why it matters: Big corrections when needed; gentle touch near success. đ Anchor: If mug drifts off target, guidance ramps up; as it aligns, guidance relaxes. â New concept 11 â Stage Switching (Schmitt Trigger) đ Hook: A light with two click points avoids flickerâyou need a clear push to turn it on or off. đ„Ź The Concept: Use two thresholds (high/low) to decide when to move to the next task stage without wobbling back and forth.
- How it works: 1) If reward > high, advance stage; 2) If between, hold; 3) If < low, reinforce/try again.
- Why it matters: Stable progress across multi-step tasks. đ Anchor: Only when the mug is clearly over the plate do we switch from âmoveâ to âplace.â â
03Methodology
At a high level: Observation + Instruction â Ground to keypoints â VLM writes differentiable stage rewards â Denoising with gradient nudges + diversity + resampling â Closed-loop adjustment and stage switching â Final action chunk. Step-by-step (like a recipe):
- Input and grounding
- What happens: The robot gets an RGB-D image and a sentence (e.g., âPlace the mug on the green plateâ). A VLM identifies relevant objects; SAM segments them; DINO features plus depth turn pixels into a 3D cloud; clustering picks a few task-relevant keypoints (mug base center, plate center, etc.).
- Why it exists: Raw images are messy. Keypoints capture the geometry we need, compactly and robustly.
- Example data: Keypoints P = {pmug_base = (x1,y1,z1), pplate_center = (x2,y2,z2)}.
- Stage-aware reward generation
- What happens: The VLM decomposes the task into stages (e.g., Stage 1: move above mug; Stage 2: grasp; Stage 3: move to plate; Stage 4: place). For each stage s, it outputs a small PyTorch function R_s that smoothly rewards desired relations (e.g., distance to target point, alignment of gripper, soft collision avoidance).
- Why it exists: We need a number to tell how good a partial trajectory is, and a gradient to know which way to improve.
- Example data: R_move_to_plate(a) = softmax(-||gripper_tip(a_T) - pplate_center||) + 0.2*alignment_bonus.
- Diverse proposal initialization with repulsive forces
- What happens: Sample B random action proposals from the diffusion/flow noise. Add a gentle repulsion term so candidates donât pile up early.
- Why it exists: Early diversity explores alternatives (center vs. edge approaches), which helps in clutter or weird layouts.
- Example: With B=16, proposals spread around the plate instead of all aiming dead center.
- Gradient-based refinement during denoising
- What happens: For each denoising (or flow) step, compute â_a R_s and add it to the modelâs predicted update (noise or velocity). Do a few inner updates (like tiny MCMC steps) to smooth out noise.
- Why it exists: These are the tiny âturn-left-a-bitâ nudges that align the trajectory with the current scene and goal.
- Example: If distance to plate center is too big, the gradient points to reduce it; the next step shifts that way.
- Gradient-free resampling via FeynmanâKac
- What happens: Weight each candidate by exp(R_s), resample so high-reward ones are more likely to continue.
- Why it exists: This accelerates convergence and drops poor options without getting stuck on a single idea too soon.
- Example: Two best candidates that skirt clutter cleanly get duplicated; others that aim straight through an obstacle get removed.
- Closed-loop control: adaptive guidance and stage switching
- What happens: After an action chunk is formed, compute its reward. If itâs lower than earlier chunks, increase guidance next time; if higher, relax guidance to let the base policy finesse the motion. Check the Schmitt-trigger thresholds to decide if itâs time to switch stages or retry.
- Why it exists: Keeps control stable and efficient across long tasks under real-world uncertainty (slips, bumps, new object positions).
- Example: Once the mug is right over the plate (reward high), flip to âplaceâ stage and reduce guidance for gentle lowering. Secret sauce (what makes this clever):
- Programmatic, differentiable rewards written by a VLM turn open-world understanding into gradients the robot can use instantly.
- Hybrid steering: precise gradient nudges + global resampling + enforced diversity.
- Closed-loop stage logic with hysteresis avoids oscillations and micromanaging.
- Training-free: The base policy stays frozen; all adaptation is during inference. Concrete walkthrough example:
- Scene: Two plates (red, green) have swapped positions; instruction: âPlace the mug on the green plate.â
- Grounding: VLM finds mug, both plates; keypoints include centers; the green plate is now on the left.
- Rewards: Stage 1 (approach mug): reward increases as gripper tip nears mug base center and aligns with grasp axis. Stage 3 (transfer): reward increases as the carried mug nears green-plate center while keeping a safe height.
- Denoising: Start 16 proposals. Repulsion spreads them. Gradients pull toward green plateâs new spot. FK resampling multiplies clean approaches. Adaptive guidance eases near success. Schmitt trigger flips to âplaceâ only when clearly above the green plate.
- Output: A smooth, OOD-aware action chunk that succeeds without changing any policy weights.
04Experiments & Results
The test: Do robots succeed more often under OOD changes when we steer them with VLS? We measure success rate (did the task finish), episode length (how many steps it took), and inference time (latency). We test in simulation on CALVIN and LIBERO-PRO, and on a real Franka robot. The competition (baselines):
- VLA models: OpenVLA, Ï, Ï-0.5, and a finetuned Ï-0.5 (LeRobot).
- Steering methods: DynaGuide (feature-distance guidance), ITPS (human/VLM-in-the-loop policy steering). The scoreboard (with context):
- CALVIN: VLS lifts success strongly. On movable objects (colored cubes), VLS hits about 94% averageâlike acing a testâversus the base policyâs much lower rate (about 7.4Ă improvement). On articulated parts (door, drawer, button, switch), VLS averages about 87%âroughly a 9.6Ă boost over base. Compared to DynaGuide and ITPS, VLS is ahead by roughly 15â25 percentage points, especially when object positions vary.
- LIBERO-PRO: VLS improves frozen VLA policies under both position and task perturbations. Overall, VLS gives up to a 13% absolute gain compared to strong baselines across suites (Goal, Spatial, 10-Long, Object). Think of that like moving from a solid B to a strong A- when others slip to C in harder versions of the exam.
- Real robot (Franka): In in-distribution setups, VLS improves average success by about 19% over the frozen baseline. Under OOD changes (new plate color, swapped plate positions, or replacing a banana with an unseen mug), the baseline often collapses, while VLS keeps goingâachieving, for example, 40% success even on the hardest unseen-object swap where the baseline fails. Surprising findings:
- Dense gradients are the star: Removing gradient guidance collapses success and lengthens episodes, showing that continuous, differentiable feedback is essential.
- Diversity and FK help efficiency: Without repulsive forces or FK resampling, success drops a bit, but efficiency and stability suffer notably. The combo prevents early mode collapse and speeds convergence.
- Computeâperformance tradeoff: Larger sample batches raise success and shorten episodes but increase inference time. This makes VLS tunable: more compute for tougher scenes, less for easier ones. Why these results matter: They show that training-free, inference-time control can rival or beat methods that depend on learned critics, dynamics rolls, or heavy searchâand it works in the messy real world, not just in clean simulators. â Ablation highlights (like taking parts out of a bike to see what breaks):
- w/o Gradient Guidance: Big performance drop; episodes get long. This is the bikeâs chainâremove it, you donât move.
- w/o FK Resampling: Success modestly down; runtime and stability worse. This is like losing your gear shifterâstill rides, but less smooth.
- w/o Repulsive Forces: Slight success drop; more time to converge. This is losing your bellâyou still ride, but you miss helpful signals early. Scaling behavior:
- Increasing batch size B improves success and reduces steps but raises latency. Pick B based on how fast you must react in your application. Overall takeaway: VLS brings a strong, practical gain on both benchmarks and a real robot, with understandable tradeoffs you can dial for your needs.
05Discussion & Limitations
Limitations (be specific):
- Latency: Keeping a batch of candidates, doing inner refinements, and resampling adds milliseconds to hundreds of milliseconds, depending on batch size and steps. Real-time constraints may cap how big you can go.
- VLM dependence: If the VLM misidentifies objects or writes awkward rewards, guidance can point the wrong way. Guardrails and prompt engineering help, but errors can sneak in.
- Differentiable reward design: Rewards must be smooth and informative. Some goals (like exact contact physics) are hard to capture with simple, differentiable surrogates.
- Very large OOD gaps: If the new task is far beyond the base policyâs skill set (e.g., tool use never seen), steering wonât invent missing skills. Required resources:
- A frozen diffusion or flow-matching robot policy.
- A VLM, segmentation (e.g., SAM), and feature extractor (e.g., DINOv2) with RGB-D input.
- GPU/accelerator for fast denoising with batches and gradients; CPU is possible but slower. When not to use VLS:
- Ultra-low-latency control (tight reflex loops) where any extra milliseconds are unacceptable.
- Tasks demanding fine-grained non-differentiable objectives (e.g., brittle contact dynamics) unless you can approximate them smoothly.
- Situations where the base policy lacks the core motor primitivesâsteering canât add skills it never learned. Open questions:
- Can we auto-check and refine the VLMâs rewards online (e.g., progress-aware prompts, self-consistency checks)?
- Can we learn a tiny, general reward prior that speeds up VLM reward writing while staying training-free for the policy?
- How to combine tactile/force feedback into differentiable rewards for better contact handling?
- Can we compress compute via distillation or caching so small robots can run VLS on-board?
- How does VLS scale to very long-horizon, multi-object tasks with many stage switches, and can we plan stage order automatically?
06Conclusion & Future Work
Three-sentence summary:
- VLS adapts frozen diffusion/flow robot policies at inference time by using a VLM to turn open-world scenes and instructions into differentiable, stage-wise rewards.
- These rewards provide gradients that steer denoising, while diversity and resampling keep exploration healthy, and closed-loop logic stabilizes long tasks.
- Across challenging OOD tests in sim and on a real Franka robot, VLS boosts success significantly without any retraining. Main achievement:
- Showing that programmatic, VLM-authored, differentiable rewards can reliably steer pretrained robot policies at test time, closing much of the OOD gap training used to handle. Future directions:
- Cut latency via better sampling schedules, caching, or distilled guidance; add progress-aware reward generation; fold in richer sensing (e.g., force) for contact phases. Why remember this:
- It flips the script from âretrain for every changeâ to âsteer what you already know,â using a smart bridgeâVLM-written rewardsâbetween open-world understanding and low-level action generation. This makes robust, practical robot deployment far more attainable in the messy, changing real world.
Practical Applications
- âąHome assistants that can still tidy up when furniture is moved or new items appear.
- âąFactory pick-and-place robots that adapt to shifting bins and mixed parts without reprogramming.
- âąWarehouse robots that reroute grasps and placements when shelves are reorganized.
- âąHospital delivery robots that handle new carts, trays, or instructions on the fly.
- âąKitchen robots that place, pour, or fetch items despite clutter and changing counter layouts.
- âąEducation and research platforms that test many tasks quickly without finetuning policies each time.
- âąField robots that cope with weathered, dirty, or novel objects through VLM-informed rewards.
- âąRapid prototyping: engineers swap goals in language and see immediate behavior changes.
- âąAssistive devices that adjust to each userâs environment and preferences via simple instructions.
- âąQuality control arms that adapt inspection paths to new product variants using keyed keypoints.