šŸŽ“How I Study AIHISA
šŸ“–Read
šŸ“„PapersšŸ“°BlogsšŸŽ¬Courses
šŸ’”Learn
šŸ›¤ļøPathsšŸ“šTopicsšŸ’”ConceptsšŸŽ“Shorts
šŸŽÆPractice
🧩ProblemsšŸŽÆPrompts🧠Review
Search
Trainable Log-linear Sparse Attention for Efficient Diffusion Transformers | How I Study AI

Trainable Log-linear Sparse Attention for Efficient Diffusion Transformers

Beginner
Yifan Zhou, Zeqi Xiao, Tianyi Wei et al.12/18/2025
arXivPDF

Key Summary

  • •This paper introduces Log-linear Sparse Attention (LLSA), a new way for Diffusion Transformers to focus only on the most useful information using a smart, layered search.
  • •LLSA changes the cost of attention from growing like a square (very slow) to growing like N log N (much faster) as sequences get longer.
  • •It does this by picking important pieces in stages from coarse to fine (hierarchical Top-K selection) and then mixing in a few big-picture tokens (Hierarchical KV Enrichment).
  • •A special GPU trick avoids building giant masks by flipping sparse indices directly, keeping training fast and memory light.
  • •On 256Ɨ256 pixel sequences, LLSA speeds up attention inference by about 28Ɨ and DiT training by about 6Ɨ while keeping image quality.
  • •LLSA works well even with a small K (like K = 8), beating older methods that need much larger K (like 20 or 32) to match quality.
  • •It scales pixel-space Diffusion Transformers without using VAEs or patchification, training up to 65,536 tokens on a single H200 GPU.
  • •Across FFHQ and ImageNet-256, LLSA improves FID/Inception Score and throughput compared to prior trainable sparse attention baselines (VSA, SLA).
  • •The method keeps global context by adding a few coarse tokens and giving them the right weight, which prevents the model from missing the big picture.

Why This Research Matters

LLSA lets powerful image and video generators run much faster, which lowers cost and energy use and makes high-quality AI creation more accessible. It helps small labs or startups train models on fewer GPUs, and it speeds up research by shortening iteration cycles. For creative users, faster attention means quicker drafts and more responsive tools for design, animation, and storytelling. In science and medicine, it can enable high-resolution analysis without waiting hours or days. For on-device and edge use, the efficiency gains bring advanced generation closer to phones and portable devices. Overall, LLSA is a step toward long-sequence AI that’s both fast and faithful to global structure.

Detailed Explanation

Tap terms for definitions

01Background & Problem Definition

šŸž Hook: Imagine you’re trying to find your friend in a giant crowd. If you check every single person one by one, it takes forever. But if you first look from a balcony to spot the right area, then zoom in only on that corner, you’ll find your friend much faster.

🄬 The Concept (Self-attention and its cost)

  • What it is: Self-attention lets a model look at every token (like a pixel or word) and decide how much to pay attention to every other token.
  • How it works: For each token (query), it compares with all others (keys) to find who matters, then gathers their information (values) to update itself.
  • Why it matters: Without attention, the model can’t connect far-away parts (like matching eyes in a face). But full attention checks every pair, which is slow for long sequences.

šŸž Anchor: If you have 65,536 pixels (a 256Ɨ256 image), full attention compares every pixel to every other pixel—billions of checks—like searching every seat in a stadium.

šŸž Hook: You know how a librarian might first check which shelf has the right category before opening a specific book? That’s smarter than opening every book.

🄬 The Concept (Sparse attention)

  • What it is: Sparse attention only looks at the most likely important parts instead of checking everything.
  • How it works: It picks a few candidates (Top-K) for each token and ignores the rest.
  • Why it matters: It saves time and memory, letting models handle longer inputs.

šŸž Anchor: When you ask ā€œWhat’s the capital of France?ā€, the model focuses on ā€œcapitalā€ and ā€œFranceā€ instead of reading every word equally.

šŸž Hook: Picture grouping a big puzzle into blocks and only comparing blocks that seem to match.

🄬 The Concept (Block sparse attention + Top-K)

  • What it is: The model groups tokens into blocks, makes a mini-summary per block, and then selects the Top-K most related blocks.
  • How it works: 1) Pool each block into a coarse token, 2) compute scores between coarse tokens, 3) keep only Top-K blocks per query, 4) do attention only on those blocks.
  • Why it matters: It cuts down work—unless the selection step itself becomes too big.

šŸž Anchor: It’s like reading the first sentence of each paragraph to decide which paragraphs to read fully.

šŸž Hook: But here’s the catch—if you only look once from far away, you might miss important details.

🄬 The Problem (Single-level Top-K hits a wall)

  • What it is: Prior methods do selection at a single coarse level, which still needs comparing many blocks to many blocks.
  • How it works: They pool once, compute a smaller but still wide similarity matrix, and pick Top-K from it.
  • Why it matters: As images or videos get larger, that selection still grows quadratically on compressed tokens, and they often must increase K to avoid losing global context—both make it slow again.

šŸž Anchor: It’s like choosing a city to search but never narrowing down to the neighborhood and street—you still walk a lot.

šŸž Hook: Think of using a map app: first you look at the whole city (zoomed out), then the district, then the street, then the house.

🄬 The Gap (Why a hierarchy is needed)

  • What it is: A single zoom level can’t capture both the big picture and fine details efficiently.
  • How it works: A hierarchy moves from coarse to fine, shrinking the search each time.
  • Why it matters: This lets the model touch only a tiny slice of the full comparisons while keeping the big-picture context.

šŸž Anchor: Like finding a restaurant by first choosing the city, then the neighborhood, then the block, and finally the door—fast and precise.

šŸž Hook: Now think about drawing pictures with AI (Diffusion Transformers). They need to look across a whole image (or video) to make coherent results.

🄬 The Concept (Diffusion Transformers and pixel-space)

  • What it is: Diffusion Transformers (DiTs) generate images by repeatedly denoising noisy pixels while using attention to stay consistent.
  • How it works: Starting from noise, they predict a cleaner version step by step, relying on attention to tie distant parts together.
  • Why it matters: Full attention is too heavy for high-resolution images or long videos, especially when operating directly on pixels without compressing them (no VAE or big patches).

šŸž Anchor: If you try to draw a high-res face, the eyes must match even if they’re far apart in the pixel grid—attention helps them ā€œtalk,ā€ but it must be efficient.

šŸž Hook: Imagine if you could keep the whole-world view while only doing extra work where needed.

🄬 The Paper’s Promise (LLSA fills the gap)

  • What it is: Log-linear Sparse Attention (LLSA) makes attention scale roughly like N log N instead of N squared.
  • How it works: It uses hierarchical Top-K selection (coarse-to-fine) and Hierarchical KV Enrichment (add a few coarse, big-picture tokens into attention with proper weights), plus an efficient GPU implementation that never builds big masks.
  • Why it matters: Now we can train pixel-space DiTs on long sequences (like 65,536 tokens) faster, cheaper, and with quality close to (or even exceeding) full attention in some settings.

šŸž Anchor: On 256Ɨ256 images, LLSA sped up attention inference by about 28Ɨ and whole DiT training by about 6Ɨ while keeping image quality competitive.

02Core Idea

šŸž Hook: You know how you find a book in a library? First, you go to the floor, then the aisle, then the shelf, then the exact spot. You don’t scan every book in the whole building.

🄬 The Aha Moment

  • What it is: Do Top-K selection in multiple zoom levels (hierarchically), then mix a few big-picture tokens into the final attention so you keep the global context—this turns attention from O(N²) into about O(N log N) while keeping quality high.
  • How it works: 1) Build a pyramid of tokens by pooling (coarse to fine). 2) At the coarsest level, pick Top-K. 3) At the next level, only compare to candidates hinted by the coarser level, and repeat. 4) During attention, enrich keys/values with a few coarse tokens and weight them by how many fine tokens they summarize. 5) Implement everything with sparse indices instead of huge masks.
  • Why it matters: The selection cost drops to linear in N (with small constants), attention uses only K + a few log-level coarse tokens, and quality holds because coarse tokens carry the big picture.

šŸž Anchor: It’s like searching city → neighborhood → street, then asking a friendly local (coarse token) to confirm you’re in the right place before you knock on the door.

Multiple Analogies (same idea, new angles)

  1. Telescope-to-Microscope: First use a telescope (coarse level) to spot the right galaxy, then switch to a microscope (fine level) to see details. You don’t microscope the whole sky.
  2. Tournament Bracket: Instead of having every team play every team (quadratic), you run playoffs in rounds. A few winners move forward each time.
  3. Grocery Trip: Check your pantry (coarse), write a shortlist (Top-K), then at the store you only visit those aisles. You also bring a map (coarse KV) so you don’t get lost.

Before vs. After

  • Before: Single-level selection needed comparing many coarse blocks to many coarse blocks; K had to grow for longer sequences to keep context; selection cost dominated runtime.
  • After: Hierarchical selection limits comparisons at each level to a few candidates; K can stay small; a few coarse tokens keep global context. Complexity becomes O(N log N), so long sequences stay practical.

Why It Works (intuition without equations)

  • Coarse-to-fine funnels the search: each level narrows the candidate list using hints from the level above. The work across levels adds up like a short geometric series (not exploding with N).
  • Enriched KV keeps context: Adding a few coarse tokens is like having a summary of whole regions. Weighting them by block size lets those summaries ā€œspeak upā€ appropriately.
  • Sparse indices instead of masks: Handling only the pieces you actually use avoids storing and scanning giant, mostly empty grids. That’s why the GPU kernels stay fast and memory-friendly.

Building Blocks (the pieces that make LLSA) šŸž Hook: Think of building a LEGO city—start with big base plates, then districts, then houses, then furniture.

🄬 The Concepts

  1. Hierarchical Compression
  • What it is: Repeatedly pool tokens into coarser levels (like averaging B neighbors each time).
  • How it works: Level l summarizes B^l fine tokens.
  • Why it matters: Gives quick, global overviews that guide selection.
  1. Hierarchical Top-K Selection
  • What it is: Pick Top-K at the coarsest level, then refine only among those candidates at each finer level.
  • How it works: Each level uses the previous level’s indices to avoid full scans.
  • Why it matters: Cuts selection cost from quadratic to linear-ish in N.
  1. Hierarchical KV Enrichment
  • What it is: During final attention, append a small set of coarse K/V tokens selected from higher levels.
  • How it works: This mixes local detail with global summaries.
  • Why it matters: Prevents losing big-picture context when being sparse.
  1. KV Reweighting
  • What it is: Give coarse tokens a weight equal to their block size (B^l).
  • How it works: A token summarizing 16 pixels gets weight 16.
  • Why it matters: Ensures summaries have the right influence.
  1. Sparse Index Transpose (GPU secret sauce)
  • What it is: Flip query-major Top-K indices into key-major form without building giant masks.
  • How it works: Count occurrences per key, prefix-sum to get offsets, fill a flat index list (like CSR→CSC in sparse matrices).
  • Why it matters: Enables efficient backward pass with near-constant throughput over sequence length.

šŸž Anchor: Put together, this is like using a map’s insets (global context) plus a street view (fine detail), guided step-by-step, and carrying only the notes you need—no phone-book-sized binders.

03Methodology

At a high level: Input (Q, K, V) → Hierarchical Compression → Hierarchical Top-K Selection → Sparse Attention with Hierarchical KV Enrichment (and KV Reweighting) → Output O. Under the hood, a GPU-efficient sparse-indices transpose powers the backward pass, and two practical helpers—index reordering for images and noise rescaling for stable diffusion training—round out the recipe.

Step 0: Prerequisites in action šŸž Hook: You know how a chef preps ingredients before cooking? Washing, chopping, and marinating make the final dish turn out right.

🄬 The Concepts

  • Self-attention (what it is/how/why): Compare each query to all keys, turn scores into weights, mix values. Without it, far-away pieces can’t coordinate.
  • Block sparse attention (what/how/why): Group tokens into blocks, skip blocks with low scores. Without blocks, you still face huge pairwise work.

šŸž Anchor: Like splitting a 1000-piece puzzle into 25 small plates and only checking the plates that look like sky for the sky piece.

Step 1: Hierarchical Compression šŸž Hook: First zoom out—like looking at the whole city before finding a street.

🄬 What happens: Starting with the finest level l = 0 (Q(0), K(0), V(0)), repeatedly pool (e.g., mean) in non-overlapping blocks of size B to form level l = 1, then 2, up to L. So Q(l), K(l), V(l) each have N/B^l tokens.

  • Why this step exists: A single coarse view can miss structure; a pyramid of views catches both big patterns and details.
  • Example: If N = 8, B = 2, levels are l = 0 (8 tokens), l = 1 (4 tokens), l = 2 (2 tokens). Each level halves the length by averaging neighbors.

šŸž Anchor: Summaries at each level act like neighborhood, district, and city maps.

Step 2: Hierarchical Top-K Selection šŸž Hook: Now we choose where to zoom in, level by level—like following signs from highway to exit to side street.

🄬 What happens: Start at the coarsest level L. Compute full similarities S(L) between Q(L) and K(L). For each query block, keep Top-K key blocks—this makes index list I(L). Then, for level Lāˆ’1, do not compare against all keys. Instead, gather only the keys indicated by I(L) (expanded to that level), compute similarities, and take Top-K again to get I(Lāˆ’1). Repeat until level 0.

  • Why this step exists: If you compared with all keys at every level, you’d fall back to quadratic work. This funneling keeps only a tiny candidate set at each stage.
  • Example with N = 8, B = 2, K = 1: at l = 2 (2 tokens), pick the best 1 candidate; at l = 1, refine only inside that candidate’s children; at l = 0, you end with a single best fine block.

šŸž Anchor: Like a tournament—only winners advance, so you never schedule every team against every other team.

Step 3: Hierarchical KV Enrichment + KV Reweighting šŸž Hook: Even if you find the right street, it helps to glance at a small city map to stay oriented.

🄬 What happens: For each fine query block, build the attention key/value set by including (a) its fine Top-K neighbors and (b) a few coarse tokens chosen from higher levels via the selection indices. Then weight each coarse token by how many fine tokens it represents (W(l) = B^l). Run FlashAttention on this compact set.

  • Why this step exists: Sparse selection shrinks the receptive field; enriched coarse tokens bring back the global context. Weighting ensures summaries have fair influence.
  • Example: With K = 8 and L = 2, you might use 8 fine blocks plus 8 from level 1 plus 8 from level 2 (numbers illustrative), with coarser ones multiplied by 2 or 4 to reflect their size.

šŸž Anchor: You read a paragraph (fine tokens) but also keep the chapter summary (coarse tokens) in mind, and you value the summary more because it covers more pages.

Step 4: Efficient Sparse Index Transpose (Backward pass) šŸž Hook: Imagine you have a class list sorted by students, but now you need it sorted by clubs they joined.

🄬 What happens: During backward, keys/values need to know which queries attended to them (key-major view). Instead of building a huge binary mask (slow, memory-heavy), LLSA flips the sparse index lists directly using a scan-based algorithm (like CSR→CSC in sparse matrices): count per key, prefix-sum to get offsets, then fill a flat index array.

  • Why this step exists: Mask-based methods accidentally bring back quadratic overhead when sequences are long.
  • Example: If 100 queries each pick K = 8 keys, you store 800 pairs and flip them efficiently, rather than making and scanning a 100Ɨ100 matrix.

šŸž Anchor: It’s like reorganizing sticky notes by topic without rewriting a giant whiteboard.

Step 5: Index Reordering for Images (2D → 1D) šŸž Hook: If you read pixels row-by-row, nearby pixels in 2D can end up far apart in 1D order—like splitting up neighbors.

🄬 What happens: Reorder pixel indices so spatial neighbors become near each other in the 1D sequence (group pixels inside growing 2^i patches). This preserves local continuity so hierarchical pooling respects real image structure.

  • Why this step exists: Without reordering, pooling could mix unrelated pixels, hurting quality.
  • Example: Instead of pure raster scan, group 2Ɨ2, then 4Ɨ4, etc., so pooling at level l summarizes a true spatial patch.

šŸž Anchor: It’s like shelving books by topic rather than by the exact time they arrived.

Step 6: Noise Rescaling and Pretraining (training helpers) šŸž Hook: Baking the same recipe in a bigger pan needs temperature/time adjustments; otherwise it won’t set right.

🄬 What happens: As resolution grows, diffusion needs stronger noise to keep the signal-to-noise ratio consistent. LLSA uses noise rescaling (e.g., scale noise by n/64 for nƗn images) inside a flow-matching scheduler, plus low-resolution pretraining to speed convergence.

  • Why this step exists: Stabilizes learning at high resolution and cuts total training time.
  • Example: Training at 128Ɨ128 or 256Ɨ256 becomes as stable as at 64Ɨ64 when you adjust the noise level appropriately.

šŸž Anchor: It’s like turning up the music volume in a bigger room so it sounds the same.

Secret Sauce (what makes it clever)

  • The hierarchy funnels comparisons so selection is linear in N with small constants.
  • Enrichment + reweighting keeps global context without needing big K.
  • The GPU kernel avoids dense masks entirely, so both forward and backward stay fast.
  • Image-friendly index reordering and noise rescaling make the whole system practical for pixel-space DiTs.

Failure mode without each part

  • No hierarchy: selection turns quadratic again.
  • No enrichment: model may miss global structure, hurting quality unless K grows.
  • No reweighting: coarse tokens under-influence the result.
  • No index transpose: backward pass becomes the bottleneck.
  • No reordering/noise scaling: pixel-space training slows or degrades in quality.

04Experiments & Results

The Test: What did they measure and why?

  • Datasets and tasks: Pixel-space Diffusion Transformers on FFHQ (128Ɨ128, 256Ɨ256, and even 512Ɨ512 in ablations) and PixelFlow ImageNet-256.
  • Metrics: Image quality via FID (lower is better) and, for ImageNet-256, Inception Score (higher is better). Efficiency via throughput (tokens or images per second) and attention/inference speedups.
  • Goal: Prove that LLSA keeps or improves quality while delivering much better speed, especially for long sequences.

The Competition: Who did they compare against?

  • Full Attention (gold standard for quality but slow).
  • Single-level Top-K sparse attention (baseline): typical approach that still suffers from selection cost and needs large K for quality.
  • VSA and SLA: two recent trainable Top-K methods that add extra branches (coarse or linear attention) to compensate for unselected tokens.

The Scoreboard: Results with context

  • Big takeaway: On 256Ɨ256 pixel tokens (65,536 tokens), LLSA accelerates attention inference by about 28.27Ɨ and overall DiT training by about 6.09Ɨ while maintaining generation quality. That’s like finishing your homework in 1 hour instead of nearly 6.
  • FFHQ-128 quality: Full attention FID ā‰ˆ 24.91; LLSA ā‰ˆ 24.37 (better is lower). LLSA matches/exceeds quality while being much faster. Compared to VSA (~26.91) and SLA (~25.73) with larger K, LLSA still wins.
  • FFHQ-256 quality: Full attention ā‰ˆ 38.77; LLSA ā‰ˆ 39.29—very close while being far faster. Against VSA (~40.69) and SLA (~39.98), LLSA is better in FID and faster.
  • ImageNet-256 (PixelFlow highest stage): LLSA achieves better FID (~20.41) and Inception Score (~73.21) than VSA (~23.59, ~64.07) and SLA (~22.58, ~65.31), with higher throughput.
  • Small K works: With K = 8, LLSA beats single-level Top-K even when that baseline uses K = 20 or 32. That’s like winning with a smaller team because your game plan is smarter.

Surprising Findings

  • Block size trade-off: Larger blocks (e.g., B = 64) boost throughput for single-level methods but hurt quality a lot. With LLSA’s hierarchy, you can keep smaller blocks (B = 16) for quality without paying a big selection cost.
  • Backward pass breakthrough: The sparse index transpose achieves nearly constant throughput across sequence lengths, confirming near-linear complexity in practice. In contrast, mask-based backward slows down steadily as N grows.
  • Enrichment levels help: Adding more KV enrichment levels improves quality slightly (with a small speed hit), showing that a little extra global context goes a long way.
  • Scaling to 512Ɨ512: Single-level fails to converge efficiently; LLSA with 2–3 levels maintains practical throughput and sensible quality, matching the predicted O(N log N) scaling.

Contextualizing the numbers

  • ā€œ28Ɨ faster attention inferenceā€ is like turning a 28-minute task into 1 minute. ā€œ6Ɨ faster trainingā€ is like finishing in 1 day what used to take almost a week.
  • Quality close to full attention means you don’t have to trade away sharpness or consistency just to be fast.

Ablations that explain why it works

  • KV Enrichment and Reweighting: Each adds a piece of quality back without big overhead; together, they can even surpass full attention at 128Ɨ128.
  • Hierarchy depth (L): Going from L = 1 to L = 2 unlocks the log-linear benefits; L = 3 refines speed further for very long sequences.
  • Practical helpers: Index reordering improves FID; noise rescaling beats other SNR tricks when scaling resolution; pretraining from lower resolution slashes time to quality.

05Discussion & Limitations

Limitations

  • Hyperparameter tuning: Choosing K, block size B, number of levels L, and enrichment depth Le affects the speed/quality balance. Wrong settings can underperform.
  • Data structure assumptions: Index reordering and block pooling work best when nearby tokens are truly related (e.g., images). For modalities with weak locality, benefits may shrink.
  • Memory vs. speed: Enriching with multi-level KV adds a handful of coarse tokens. It’s still light, but not entirely free; extremely tight memory budgets might require trimming enrichment levels.
  • Training, not just inference: LLSA is trainable and learns good sparse patterns, which is a strength—but also means you can’t always plug it in as a zero-shot speedup without some finetuning.
  • Implementation detail sensitivity: The efficient indices transpose and Triton kernels are key to observed gains; suboptimal implementations could hide theoretical advantages.

Required Resources

  • A modern GPU (e.g., H200/A100-class) to benefit from Triton kernels and fast sparse operations.
  • Reasonable engineering to integrate hierarchical pooling, sparse index management, and the enriched attention path.
  • For pixel-space DiTs: index reordering and noise rescaling integrated into the training loop, plus optional low-res pretraining.

When NOT to Use

  • Very short sequences: Full attention may be simpler and fast enough; LLSA’s setup overhead may not pay off.
  • Tasks needing dense global pairwise reasoning everywhere (e.g., small graphs with all-to-all interactions): sparsity may cut out needed links.
  • Highly non-local modalities with fragile structure: If ā€œneighborsā€ in 1D order rarely match true relationships, hierarchy may misguide selection.

Open Questions

  • Adaptive K and adaptive enrichment: Can the model learn to vary K per token or timestep to push further speed/quality trade-offs?
  • Better compression: Could learned pooling or attention-based coarsening beat simple mean pooling for even higher quality?
  • Cross-modal generalization: How well does LLSA extend to long-form text, audio, and very long video streams without careful reordering?
  • Theoretical guarantees: What bounds can we prove on approximation quality vs. full attention under realistic data assumptions?
  • Combining with other efficiency tricks: How does LLSA stack with low-rank adapters, kernelized attention, or memory tokens?

06Conclusion & Future Work

Three-sentence summary

  • This paper introduces Log-linear Sparse Attention (LLSA), which performs hierarchical Top-K selection and mixes in a few weighted coarse tokens so that attention cost grows like N log N instead of N squared.
  • A GPU-efficient sparse indices transpose keeps both forward and backward fast without building dense masks, enabling big speedups in practice.
  • On pixel-space Diffusion Transformers, LLSA preserves or improves quality versus strong baselines while delivering up to ~28Ɨ faster attention inference and ~6Ɨ faster training.

Main achievement

  • Turning the single-level Top-K paradigm into a hierarchical, trainable, and mask-free pipeline that scales Diffusion Transformers to very long pixel sequences with strong quality retention at surprisingly small K.

Future directions

  • Adaptive K and enrichment depth, learned hierarchical pooling, and broader applications to long-text, audio, and video. Exploring combination with other memory- and compute-saving techniques for even larger scales.

Why remember this

  • LLSA shows that you don’t need to choose between speed and global context: with the right hierarchy and a few well-weighted summaries, you get both. It reframes sparse attention as a practical, scalable default for long-sequence generative transformers, opening the door to higher resolutions, longer videos, and more accessible training on modest hardware.

Practical Applications

  • •Train pixel-space Diffusion Transformers at 128–256 resolutions on a single high-end GPU without VAEs or large patches.
  • •Speed up high-resolution image synthesis in production pipelines, reducing compute cost and latency.
  • •Accelerate long video diffusion models by scaling attention to very long token sequences.
  • •Enable interactive creative tools (storyboarding, concept art, graphic design) with faster preview and iteration.
  • •Improve medical or scientific imaging pipelines that require large fields of view while preserving global context.
  • •Run higher-resolution or longer-context generative models on limited hardware (e.g., edge servers or small clusters).
  • •Enhance remote sensing and satellite image generation/denoising where images are huge and global structure matters.
  • •Use as a drop-in efficient attention layer for pixel-focused diffusion frameworks like PixelFlow at high stages.
  • •Prototype efficient long-sequence research models in Triton without dense mask overhead.
  • •Combine with low-res pretraining to quickly scale models to higher resolutions with stable training.
#Log-linear Sparse Attention#Hierarchical Top-K#Hierarchical KV Enrichment#KV Reweighting#Block Sparse Attention#Diffusion Transformers#Pixel-space DiT#FlashAttention#Sparse index transpose#Index reordering#Image generation#FID#Inception Score#Triton kernels#O(N log N) attention
Version: 1