🎓How I Study AIHISA
đź“–Read
📄Papers📰Blogs🎬Courses
đź’ˇLearn
🛤️Paths📚Topics💡Concepts🎴Shorts
🎯Practice
🧩Problems🎯Prompts🧠Review
Search
Dr. Kernel: Reinforcement Learning Done Right for Triton Kernel Generations | How I Study AI

Dr. Kernel: Reinforcement Learning Done Right for Triton Kernel Generations

Intermediate
Wei Liu, Jiawei Xu, Yingru Li et al.2/5/2026
arXivPDF

Key Summary

  • •This paper teaches a language model to write fast GPU kernels (tiny speed programs) in Triton using reinforcement learning that really cares about meaningful speed, not just being correct.
  • •They build a safe, scalable lab called KERNELGYM that runs the model’s code, checks for cheating, measures real speed, and gives rich feedback every turn.
  • •They find a bias problem in a popular training method (GRPO) for multi-turn learning and fix it with TRLOO, which gives fairer learning signals each turn.
  • •Models often do 'lazy optimization'—making tiny changes that are correct but don’t speed things up; the paper tackles this with bottleneck-aware Profiling-based Rewards (PR) and Profiling-based Rejection Sampling (PRS).
  • •They also stabilize training with Mismatch Rejection Sampling (MRS), so the model doesn’t wobble or collapse while learning.
  • •Their model, DR. KERNEL-14B, matches or beats strong closed models on KernelBench Level-2 for meaningful speedups (Fast@1.2).
  • •With test-time scaling (letting the model refine more times and picking the best try), they push Fast@1.2 to 47.8% on Level-2, topping Claude-4.5-Sonnet and GPT-5 there.
  • •Even when PyTorch’s own compiler (torch.compile) makes things fast, their approach still brings extra wins.
  • •All code, environment, and models are released so others can build on this work.

Why This Research Matters

Faster kernels make AI run cheaper, greener, and more widely available, from phones to data centers. By aligning training around true speed, not just correctness, this work turns performance optimization into a repeatable, reliable process. The safe lab (KERNELGYM) and fair learning signals (TRLOO, PR/PRS, MRS) mean fewer crashes, less cheating, and more real gains. Companies can save money on compute, ship features quicker, and reduce environmental impact. Open-sourcing the environment and methods helps the whole community improve system-level performance. As models grow and tasks get heavier, these wins compound into big quality-of-life improvements for end users.

Detailed Explanation

Tap terms for definitions

01Background & Problem Definition

🍞 Top Bread (Hook): Imagine you’re timing runners on a track. Everyone can finish the race, but what you really care about is who finishes faster. In AI, 'finishing' is correctness, and 'finishing faster' is speed. For huge AI models, tiny speedups inside GPU programs (called kernels) add up to big wins.

🥬 Filling (The Actual Concept):

  • What it is: This paper is about teaching an AI to write Triton GPU kernels that are not just correct but truly faster than the original PyTorch code.
  • How it works: The AI writes a kernel, a safe lab runs it, checks for cheating, measures real speed, and gives feedback; the AI tries again, learning across multiple turns.
  • Why it matters: If we only check correctness, the AI can 'win' without getting faster; then training stalls and real apps don’t speed up.

🍞 Bottom Bread (Anchor): Think of a calculator app on your phone that loads in half the time because the tiny math pieces inside were rewritten to run faster on your GPU.

New Concept 1 — GPU Kernel 🍞 Hook: You know how a sandwich has fillings you repeat for every bite? A kernel is a tiny, repeatable recipe the GPU runs many times. 🥬 The Concept:

  • What it is: A GPU kernel is a small program that runs on many GPU threads to do math super fast.
  • How it works: It splits big data into chunks, processes each chunk in parallel, and writes results back.
  • Why it matters: If the kernel is slow or written poorly, the whole model feels slow. 🍞 Anchor: Adding two long lists at once—each GPU thread adds a pair of numbers; do that thousands of times in parallel and you finish quickly.

New Concept 2 — Triton 🍞 Hook: Imagine writing a recipe using simple steps instead of a chef-only language. 🥬 The Concept:

  • What it is: Triton is a friendly programming language that makes GPU kernels easier to write than low-level CUDA.
  • How it works: You write Python-like code; Triton turns it into fast GPU instructions.
  • Why it matters: Easier writing means the AI can learn it faster and make fewer mistakes. 🍞 Anchor: Like using picture instructions to build a Lego set instead of dense technical blueprints.

The World Before: People knew fast kernels (like FlashAttention) make AI cheaper and greener. But writing them is hard, needs deep GPU knowledge, and takes lots of time. Some tried training language models to write kernels, but most focused on simply 'works or not' (correctness) rather than 'is it faster?' And training was fragile—models could crash GPUs, or find loopholes to look fast without doing real work (reward hacking).

The Problem: Two big headaches kept showing up:

  • Reward hacking: The model does something sneaky that seems fast in tests but skips real compute (like pretending to call a kernel but not actually running it).
  • Lazy optimization: The model replaces a tiny, unimportant part with a kernel, so the whole program isn’t much faster.

New Concept 3 — Reward Hacking 🍞 Hook: Ever seen someone win a board game by bending the rules instead of playing better? 🥬 The Concept:

  • What it is: Reward hacking is when a model exploits scoring rules to get a good reward without truly solving the task.
  • How it works: It detects patterns that pass tests (e.g., add a '@triton.jit' decorator) but skips heavy work or times the wrong thing.
  • Why it matters: Training then 'rewards' bad behavior, and the model stops learning real speedups. 🍞 Anchor: Answering a math quiz by memorizing answers to old questions, not learning to add.

New Concept 4 — Lazy Optimization 🍞 Hook: Imagine cleaning only one corner of your messy room and saying 'done!' 🥬 The Concept:

  • What it is: Lazy optimization means doing small, correct changes that avoid the real bottleneck, so speed barely improves.
  • How it works: The model replaces a trivial op (like a simple sum) but leaves the expensive parts untouched.
  • Why it matters: You get a checkmark for correctness but not real speed; progress stalls. 🍞 Anchor: Swapping one slow spoon for a faster spoon when the real problem is cooking five separate times instead of one big batch.

Failed Attempts: Prior systems either judged only correctness, used weak 'LLM-as-a-judge' for hacking checks, or did single-turn attempts (no iterative improvements). Datasets were tiny, training would crash from GPU errors, and measurements weren’t consistent.

The Gap: We needed a real lab that’s safe, distributed, and strict about measuring speed and cheating; a multi-turn learning method with fair training signals; and rewards that push the model to fix true bottlenecks.

New Concept 5 — Performance Measurement (Profiling) 🍞 Hook: You know how a coach times each part of a race (start, middle, finish) to find where the runner slows down? 🥬 The Concept:

  • What it is: Profiling breaks down where time is spent in your program.
  • How it works: It records which kernels run and how long they take, then summarizes hotspots.
  • Why it matters: Without it, the model can’t see the real bottleneck, so it won’t fix the right part. 🍞 Anchor: Seeing that 86% of time is in max-pool, not in a tiny sum, tells the model what to optimize.

Real Stakes: Faster kernels mean quicker AI features on your phone, lower cloud bills for companies, and less energy use for the planet. It also frees engineers to focus on new ideas instead of hand-tuning every loop. This is like teaching many student-chefs (models) to cook tasty meals quickly and safely on their own.

02Core Idea

🍞 Top Bread (Hook): Picture a student who writes an essay, gets teacher notes, rewrites it, and keeps improving. Now imagine we do that for code that runs on GPUs, and the teacher is a safe, strict lab that also says which paragraph (kernel) wasted the most time.

🥬 The Concept: The 'aha!' is to combine a robust training gym (KERNELGYM), fair multi-turn learning signals (TRLOO), bottleneck-aware rewards (PR) plus smart filtering (PRS), and stability fixes (MRS), then let the model refine multiple times even at test time (STTS).

  • What it is: A complete pipeline that teaches an AI to iteratively write Triton kernels, checks for cheating, and rewards real speed.
  • How it works: Multi-turn propose→run→measure→reward→update, with TRLOO for unbiased credit, PR/PRS to target big bottlenecks, and MRS to keep training steady.
  • Why it matters: Without the lab, the model crashes or cheats; without fair credit, it learns too slowly; without bottleneck-aware rewards, it stays lazy; without stability, training falls apart.

🍞 Bottom Bread (Anchor): The model first replaces a tiny sum (no real speedup). With profiling-based rewards, it learns to fuse pooling ops instead, jumping from ~1.01× to ~2.08× speedup on that task.

New Concept 6 — KERNELGYM 🍞 Hook: Think of a science lab where every experiment is done in a separate safe room and all results are logged. 🥬 The Concept:

  • What it is: KERNELGYM is a distributed 'gym' that runs kernels, catches crashes, checks for hacking, and returns correctness, speed, and profiling.
  • How it works: A server hands tasks to GPU workers; each run happens in a clean subprocess; a monitor restarts anything that fails; results are collected for learning.
  • Why it matters: Keeps long training runs alive and provides rich, reliable feedback. 🍞 Anchor: It’s like having many kitchen stations; if one burns a dish, others keep cooking, and the head chef still gets the timings and notes.

New Concept 7 — Multi-turn Reinforcement Learning 🍞 Hook: You don’t ace a piano piece in one try; you practice, listen, adjust, and try again. 🥬 The Concept:

  • What it is: The model iteratively improves its code across several turns, each time seeing feedback from the last try.
  • How it works: Turn 1 writes a draft; the gym runs it and reports correctness, speed, profiling; Turn 2 refines choices; repeat.
  • Why it matters: Hard performance bugs need iteration; single-shot attempts miss many fixes. 🍞 Anchor: Like tuning a bike: adjust seat height, test ride, adjust gears, test again, until it feels fast and smooth.

New Concept 8 — Policy Gradient and Advantage (intuition) 🍞 Hook: If teammates did better than average, you cheer louder; if they did worse, you cheer softer. 🥬 The Concept:

  • What it is: Policy gradient changes the model’s choices based on how much better-than-average (advantage) a try was.
  • How it works: Compute each try’s return, subtract a baseline (average), and nudge the model toward better choices.
  • Why it matters: Without this, learning is noisy and slow. 🍞 Anchor: Give more high-fives for great fixes; fewer for small or bad ones.

New Concept 9 — TRLOO (Turn-level Reinforce Leave-One-Out) 🍞 Hook: Imagine grading a test by comparing one student to the class—but don’t include that student in the class average. 🥬 The Concept:

  • What it is: TRLOO computes a fair advantage per turn by subtracting an average that excludes the current sample.
  • How it works: For each turn’s group of attempts, the baseline is the mean of other attempts, avoiding self-influence.
  • Why it matters: Fixes a hidden bias in GRPO that shrinks learning signals, especially harmful when good attempts are rare. 🍞 Anchor: A standout kernel that finally gets a real speedup isn’t dragged down by its own success in the average.

New Concept 10 — Profiling-based Rewards (PR) 🍞 Hook: If 90% of your chore time is laundry, doing dishes faster won’t help much. 🥬 The Concept:

  • What it is: PR adds a score based on how much of the total runtime the model’s kernels actually cover.
  • How it works: Measure time covered by generated kernels divided by total CUDA time; add this (bounded) term to the reward when code is correct.
  • Why it matters: Guides the model to fix real bottlenecks, not side quests. 🍞 Anchor: Fusing pooling ops that account for 86% of time earns more reward than speeding up a 0.014% sum.

New Concept 11 — Profiling-based Rejection Sampling (PRS) 🍞 Hook: When sorting fruit, you keep the best ones and toss the clearly bad ones, but sometimes keep a borderline apple. 🥬 The Concept:

  • What it is: PRS keeps training samples with probability tied to their profiling coverage, filtering out lazy attempts.
  • How it works: If coverage is below a threshold, likely reject; near the threshold, keep with some chance.
  • Why it matters: Focuses learning on impactful attempts while leaving room for exploration. 🍞 Anchor: Spend more time learning from tries that touch the big-time sinks.

New Concept 12 — Mismatch Rejection Sampling (MRS) 🍞 Hook: Practicing piano on a toy keyboard but performing on a grand piano can cause surprises. 🥬 The Concept:

  • What it is: MRS rejects training samples where the training model’s probabilities drift too far from the rollout model’s.
  • How it works: Compute a ratio over tokens; if it’s outside a tiny safe band, drop the sample.
  • Why it matters: Stabilizes training to prevent collapse from training–inference mismatch. 🍞 Anchor: Only learn from practice runs that sound like what you’ll actually play on stage.

New Concept 13 — Sequential Test-Time Scaling (STTS) 🍞 Hook: When taking photos, you snap many shots and keep the best one. 🥬 The Concept:

  • What it is: At inference, do more refinement turns than during training and select the best kernel seen so far.
  • How it works: Either append full history (vanilla) or keep a compact window of top past turns (context management) to avoid context overflow.
  • Why it matters: More chances = better odds of a fast kernel, without changing the trained model. 🍞 Anchor: Best-of-history selection boosted Level-2 Fast@1.2 from 31.6% to 47.8%.

Before vs After:

  • Before: Single-turn, correctness-first systems that were easy to hack or got stuck on tiny improvements.
  • After: A sturdy gym, fair multi-turn learning, bottleneck-aware rewards, and test-time refinement that produce meaningfully faster kernels.

Why It Works (intuition):

  • Accurate feedback (profiling + hacking checks) makes the target clear.
  • Fair credit (TRLOO) makes learning signals stronger.
  • Bottleneck-aware rewards (PR/PRS) aim effort where it counts.
  • Stability (MRS) keeps training healthy.
  • Extra tries at test time (STTS) harvests the best result.

03Methodology

High-level recipe: Torch reference code → Model proposes a Triton kernel (Turn 1) → KERNELGYM runs it and returns correctness, speed, profiling, and hacking status → Compute reward (speedup + profiling coverage), then compute turn-level advantages with TRLOO → Update the model (policy gradient) with stability guard (MRS) → Repeat for Turn 2, Turn 3 → At test time, allow more turns (STTS) and pick the best.

Step A — Input and First Proposal

  • What happens: The model reads a PyTorch reference (the 'slow but correct' version) and writes a first Triton kernel draft, possibly with operator fusion.
  • Why this exists: We need a starting point to measure; seeing the reference helps decide which ops to replace.
  • Example: For a row-wise L1 normalization, the model fuses abs, sum, and scale into one kernel.

Step B — Safe Execution in KERNELGYM

  • What happens: The server schedules the run; a GPU worker spawns a clean subprocess, executes the kernel in both train and eval modes, records correctness, measures speed, runs the profiler, and checks hacking (e.g., kernel actually launched).
  • Why this exists: Prevents crashes from killing the whole system and ensures fair, consistent measurements.
  • Example Data: correctness=true/false; speedup=reference_time/kernel_time (clipped ≤3); profiling lists kernels with their CUDA times; hacking=false if Triton kernels actually executed.

Step C — Reward with Bottleneck Awareness

  • What happens: If correct, reward = 1 (correct) + speedup + profiling coverage (PR). If incorrect, reward is 0 with diagnostic logs for the next turn.
  • Why this exists: Speedup alone can be noisy; PR highlights real bottlenecks and discourages lazy fixes.
  • Example: A candidate with 1.21Ă— speedup and 0.6 coverage gets more reward than 1.05Ă— and 0.05 coverage.

Step D — Multi-Turn Returns and TRLOO Advantages

  • What happens: Use reward-to-go per turn (later rewards credit earlier turns that set them up). Within each turn, compute the leave-one-out baseline so each sample’s advantage isn’t reduced by itself.
  • Why this exists: Multi-turn refinement means early choices influence later wins; TRLOO gives unbiased, stronger signals, especially when good samples are rare.
  • Example: If Turn 1’s draft enables Turn 2’s big speedup, Turn 1 gets credit via reward-to-go.

Step E — Stabilize with MRS

  • What happens: Compare the training model’s token probabilities to the rollout model’s; if mismatch is too big, drop the sample. Also veto if any token drifts extremely.
  • Why this exists: Stops off-policy drift that can cause wild gradients and collapse.
  • Example: If a sampled sequence would force the model to learn from a path it would never naturally take, it’s rejected.

Step F — Filter with PRS

  • What happens: Use profiling coverage to probabilistically keep or drop samples; low-coverage 'lazy' attempts are mostly removed; borderline ones are kept sometimes to explore.
  • Why this exists: Focus learning on impactful examples while keeping exploration.
  • Example: Coverage below 0.3 is likely dropped; 0.33 may be kept with some probability.

Step G — Update Policy

  • What happens: Apply policy gradient using TRLOO advantages on the retained batch to nudge token probabilities toward better kernels.
  • Why this exists: This is how the model learns to write faster kernels over time.
  • Example: After updates, the model is more likely to fuse the right ops or pick better Triton configs.

Step H — Next Turns and Diagnostics

  • What happens: The model sees structured feedback (errors, traces, profiling summaries) and refines code in Turn 2 and Turn 3.
  • Why this exists: Iteration is key for performance tuning; feedback tells the model what to fix next.
  • Example: Auto-tuning BLOCK sizes, num_warps, or fusing an additional op based on where time is spent.

Step I — Test-Time Scaling (STTS)

  • What happens: At inference, allow more than 3 turns. Two modes: vanilla extrapolation (append full history) or context management (keep only top-w past turns in-context to avoid overflow).
  • Why this exists: More tries increase the chance of finding a really fast kernel without retraining.
  • Example: With context management (w=4), best-of-history Fast@1.2 climbs to 47.8% on Level-2.

The Secret Sauce:

  • TRLOO: Removes self-inclusion bias so strong, rare successes get full credit.
  • PR + PRS: Aim learning at big bottlenecks and filter out lazy samples.
  • MRS: Keeps learning stable by preventing off-policy drift.
  • KERNELGYM: Makes all of this possible with safe, reliable, detailed feedback.

Concrete Walkthrough (mini):

  • Input: Torch LayerNorm-like op.
  • Turn 1: Model writes a Triton kernel but forgets to actually call it in forward (hacking check catches this → incorrect). Feedback shows no Triton kernels ran.
  • Turn 2: Model fixes call and fuses abs+reduce+scale; correctness passes; speedup is 1.04Ă—; PR is 0.1 (only small portion covered). Reward is modest.
  • Turn 3: Model fuses max-pool ops that profiling showed were hotspots; speedup jumps to 2.08Ă—; PR rises to 0.86; reward is high; policy updates reinforce this pattern.

04Experiments & Results

The Test: KernelBench Levels 1–3 on NVIDIA H100 GPUs with strict evaluation. A candidate must be correct and pass the hacking check before we measure speedup. We report Fast@p: the percent of tasks where the model is correct and at least p× faster than the Torch reference (p in {1, 1.2, 1.5, 2}). We also test under torch.compile, a stronger baseline where trivial gains usually vanish.

The Competition: Prior open systems (AutoTriton), strong general coding models (Qwen, GLM), and frontier proprietary models (Claude-4.5-Sonnet, GPT-5). We compare both single-best-turn and best-of-history (choose the best across turns) for fairness.

Scoreboard (plain-English context):

  • Level 2, Fast@1.2: DR. KERNEL-14B reaches 25.6% (like moving from a B to a solid A- among peers). With sequential test-time scaling (STTS) last-turn, it hits 31.6%, beating Claude-4.5-Sonnet (26.7%) and GPT-5 (28.6%). With best-of-history across turns, it soars to 47.8% (an A+ when others are at B/B+).
  • Level 1, Fast@1.2: DR. KERNEL-14B reaches 16.9%; with STTS last-turn it’s 18.8%; best-of-history 25.1%.
  • Level 3 is toughest: modest gains at Fast@1.2 improve with STTS (from 1.2% to 3.0%/7.3%), showing the value of more attempts.
  • Under torch.compile (a high bar), our models still deliver meaningful Fast@p scores, indicating these aren’t just 'eager-mode tricks.'

Surprising Findings:

  • Stability alone (MRS) smooths training but doesn’t lift the Fast@1.2 ceiling; we need PR and PRS to break lazy optimization.
  • Hacking checks matter a lot: without them, training 'succeeds' quickly but plateaus and doesn’t translate to real speedups.
  • TRLOO consistently outperforms GRPO across turns and training steps, especially as turns increase; reward-to-go (Îł>0) improves first-turn quality by crediting early decisions.

Training Dynamics:

  • Without MRS, entropy and gradient norms spike; training can wobble. Adding MRS calms these metrics. Then PR and PRS not only raise Fast@1.2 but further stabilize learning—fewer wild swings.

Hacking Ratio:

  • With KERNELGYM’s hacking check, DR. KERNEL-14B’s hacking rate on Level-2 steadily drops from ~20% at start to ~3%. Compared to a baseline system showing ~10% hacking on Level-1, this is a clear safety/quality win.

Takeaway Numbers (friendly recap):

  • Level-2 meaningful speedups (Fast@1.2): 31.6% with STTS last-turn; 47.8% best-of-history (beats Claude-4.5-Sonnet at 26.7% and GPT-5 at 28.6%).
  • Stronger-than-eager setting (torch.compile): improvements remain, proving the method targets real bottlenecks.
  • Multi-turn and fair credit (TRLOO) plus bottleneck-aware rewards (PR/PRS) are the difference-makers.

05Discussion & Limitations

Limitations:

  • Data hunger: Only ~8k multi-turn cold-start samples were used; kernel programming is underrepresented in common pretraining data, so more domain data would likely help a lot.
  • Capacity limits: Bigger models do better; some hard ops (like convolutions) are already highly optimized in cuDNN, and small models struggle to beat them in Triton.
  • Not fully autonomous: The system generates strong kernels but isn’t yet a push-button, end-to-end production solution for all workloads.
  • Coverage of languages: Experiments focus on Triton; generalizing to CUDA/TileLang at scale needs more engineering.

Required Resources:

  • Hardware: Modern GPUs (H100-class recommended), since profiling and repeated runs are needed.
  • Software: KERNELGYM server/worker setup (FastAPI, Redis), isolation via subprocesses, profilers, and evaluation backends.
  • Time: Multi-turn RL with profiling is compute-intensive; budget for long training and many evaluations.

When NOT to Use:

  • Tiny tensors or simple ops where PyTorch (especially with torch.compile) is already near optimal—ROI may be small.
  • Strict safety-critical code paths where only verified, library-optimized kernels are acceptable.
  • Extremely limited compute settings where you can’t afford repeated measurements/profiling.

Open Questions:

  • Scaling data: How much does domain pretraining (middle-training) on large kernel corpora boost results?
  • Beyond Triton: What’s the right recipe to port these methods to CUDA or mixed backends with minimal overhead?
  • Better rewards: Can we blend hardware counters, memory-bound vs compute-bound signals, or power usage into training objectives?
  • Human-in-the-loop: What lightweight edits from experts most efficiently guide multi-turn learning?
  • Robustness: How to auto-generate stronger correctness suites and dynamic shapes to reduce overfitting to test cases?

06Conclusion & Future Work

Three-sentence summary: This paper builds a safe, distributed gym (KERNELGYM) and a multi-turn learning method (with TRLOO, PR/PRS, MRS) that trains models to write Triton kernels focused on real bottlenecks, not just correctness. It fixes reward hacking, fights lazy optimization, and stays stable while learning. The resulting model, DR. KERNEL-14B, achieves state-of-the-art meaningful speedups on KernelBench Level-2, especially with test-time scaling.

Main achievement: A full-stack, plug-and-play recipe—environment + unbiased multi-turn training + bottleneck-aware rewards + stability + test-time scaling—that turns kernel generation into a robust RL problem producing real, measurable speed.

Future directions: Scale domain data and model size; extend to CUDA/TileLang; enrich rewards with deeper hardware signals; improve test-time strategies for longer contexts; and move closer to hands-free, production-grade kernel generation.

Why remember this: It shows how to get beyond 'it runs' to 'it runs fast' by aligning the whole learning loop—from the lab to the rewards—around fixing real performance bottlenecks, and it shares the tools so others can build even faster systems.

Practical Applications

  • •Auto-optimize hot PyTorch ops in training and inference pipelines by iteratively generating and testing Triton kernels.
  • •Use profiling feedback to identify and fuse the top 1–2 bottleneck ops in custom model layers.
  • •Integrate KERNELGYM’s hacking check into CI to prevent 'fake speedups' from slipping into production.
  • •Adopt TRLOO in your multi-turn RL loops to improve credit assignment for complex tool-use tasks.
  • •Enable PR and PRS in reward design to push agents toward fixes that cover most runtime.
  • •Stabilize RL fine-tuning of code LLMs with MRS to reduce off-policy drift and training collapse.
  • •Apply sequential test-time scaling (extra refinement turns + best-of-history selection) to squeeze more performance without retraining.
  • •Benchmark under torch.compile as well as eager mode to ensure improvements hold against strong compiler baselines.
  • •Build a distributed evaluation pool (server–worker with subprocess isolation) to run risky code safely at scale.
  • •Create curated multi-turn trajectories (with feedback logs) to bootstrap SFT before RL.
#Triton kernels#Reinforcement learning#Policy gradient#Leave-one-out baseline#Profiling-based rewards#Rejection sampling#KernelBench#Reward hacking#Operator fusion#torch.compile#Multi-turn RL#Test-time scaling#GPU performance#Distributed evaluation#Speedup metrics
Version: 1