🎓How I Study AIHISA
📖Read
📄Papers📰Blogs🎬Courses
💡Learn
🛤️Paths📚Topics💡Concepts🎴Shorts
🎯Practice
🧩Problems🎯Prompts🧠Review
Search
FrankenMotion: Part-level Human Motion Generation and Composition | How I Study AI

FrankenMotion: Part-level Human Motion Generation and Composition

Beginner
Chuqiao Li, Xianghui Xie, Yong Cao et al.1/15/2026
arXivPDF

Key Summary

  • •FrankenMotion is a new AI that makes human motion by controlling each body part over time, like a careful puppeteer.
  • •It works because the authors built FrankenStein, a dataset with fine-grained labels for head, arms, legs, spine, and trajectory at each moment.
  • •A language agent called FrankenAgent (using an LLM) reads existing high-level labels and infers detailed body-part actions with timestamps.
  • •The model uses hierarchical conditioning: part-level prompts, action-level segments, and one sequence-level caption all guide the final motion.
  • •Under the hood it is a diffusion model with a transformer that mixes motion signals and text features from CLIP.
  • •FrankenMotion outperforms strong baselines (STMC, UniMotion, DART) in both how well motions match the text and how realistic they look.
  • •It can compose new combinations it never saw during training, like sitting while raising only the left arm.
  • •Human raters confirmed the LLM annotations are highly accurate (about 93% correctness) with strong agreement.
  • •The main limitation is very long sequences; it does not yet generate minute-long motions in one pass.
  • •This work makes character animation, games, VR/AR, and robotics more controllable, precise, and creative.

Why This Research Matters

This research turns vague, whole-body commands into precise body-part choreography, making animated characters feel more alive and responsive. Game designers and filmmakers get fine-tuned control to direct heads, arms, legs, and spines over exact time windows, reducing manual keyframing. In VR/AR and education, creators can script nuanced motions like “hold the rail” or “look left now,” improving clarity and safety. For robotics, part-aware motion control helps machines act more predictably around people and objects. The LLM-assisted labeling pipeline also shows how we can upgrade old datasets into richer ones without massive human effort. As models become better at composition, we can invent new motions on the fly, opening doors for creativity and accessibility. Overall, this brings motion generation closer to real choreography—detailed, flexible, and story-driven.

Detailed Explanation

Tap terms for definitions

01Background & Problem Definition

🍞 Top Bread (Hook): Imagine you are directing a school play. You can tell the whole class, “Everyone dance now!” but sometimes you want to give special instructions like, “Only the head turns left,” or “Right arm waves while walking.” That second kind of control is much harder.

🥬 Filling (The Actual Concept):

  • What it is: Human motion generation from text means a computer turns words like “walk forward and wave your right hand” into a moving 3D person.
  • How it works (before this paper): Most systems understood only whole-scene or whole-action commands, not tiny part-by-part directions over time.
  • Why it matters: Without part-level and time-aware control, characters can look unrealistic, ignore details, or fail to match what you asked.

🍞 Bottom Bread (Anchor): If you say, “Sit down while turning your head left,” old systems might sit but forget to move the head correctly, or move everything at once.

🍞 Top Bread (Hook): You know how a recipe can say “bake for 10 minutes, then add frosting”? Timing matters. So does which part you’re working on—cake vs. frosting.

🥬 Filling (The Actual Concept):

  • What it is: Datasets used to train motion models often had simple labels like “walk” or a single sentence for the whole clip.
  • How it works: These labels didn’t say what each body part did or when, so models learned only fuzzy, big-picture motions.
  • Why it matters: If the data doesn’t say what the head, arms, legs, or spine do at each time, the model can’t learn fine control for those parts.

🍞 Bottom Bread (Anchor): Saying “The person walks” doesn’t tell the AI how the knees bend or how the arms swing, so the motion can look stiff or off.

🍞 Top Bread (Hook): Think of LEGO builds: complex castles come from snapping many small bricks together.

🥬 Filling (The Actual Concept):

  • What it is: Complex human motions can be broken into small atomic elements: which part moves, how, and when.
  • How it works: By teaching a model these atoms and how to combine them, we can assemble brand-new motions from familiar pieces.
  • Why it matters: Without atomic parts, you can’t mix and match—for example, “sit” with a “left arm wave” at the same time.

🍞 Bottom Bread (Anchor): You can combine “turn head left,” “bend knees,” and “raise right arm” into a new, precise motion recipe.

🍞 Top Bread (Hook): You know how a good editor can turn messy notes into a clean schedule?

🥬 Filling (The Actual Concept):

  • What it is: Large Language Models (LLMs) can read a high-level description and infer likely body-part motions and timings.
  • How it works: Given captions like “pick up box, then set it down,” the LLM guesses which parts move (arms bend, spine leans) and when, adding time stamps.
  • Why it matters: We avoid hand-labeling every frame, which would take forever, and still get fine-grained, time-aware labels.

🍞 Bottom Bread (Anchor): From “lift object above your head,” the LLM labels “arms up,” “head looks forward,” “spine extends,” all at the right moments.

🍞 Top Bread (Hook): Imagine trying to direct a dance using only one sentence—you’d miss many details.

🥬 Filling (The Actual Concept):

  • What it is: Before this paper, many models lacked hierarchical control—sequence, action segments, and body parts didn’t talk to each other.
  • How it works: Without layers of control, small instructions get lost, and timing can be off.
  • Why it matters: Games, movies, VR, and robots need precise, coordinated motions that respect tiny details and big goals.

🍞 Bottom Bread (Anchor): In a game, you might want your character to walk, stumble, then recover—while the right hand holds a rail the whole time. That needs layered control.

🍞 Top Bread (Hook): Think of building a city: You need a map (sequence), blocks (actions), and bricks (parts). Missing one makes the plan fall apart.

🥬 Filling (The Actual Concept):

  • What it is: The gap was a lack of a dataset with synchronized, atomic, part-level labels over time—and a model designed to use them.
  • How it works: Previous tries added some part labels but forced all parts to share the same time cuts, or they edited after the fact, which reduces realism.
  • Why it matters: To get reliable, controllable motions, we need both detailed data and a model that composes parts and actions smoothly.

🍞 Bottom Bread (Anchor): Without good part-and-time labels, telling the AI “left arm up while legs stay still” at second 2.5 just won’t land.

🍞 Top Bread (Hook): Picture your favorite animated movie. Every subtle head tilt and finger tap helps tell the story.

🥬 Filling (The Actual Concept):

  • What it is: FrankenMotion and the FrankenStein dataset deliver part-aware, time-aware generation and control.
  • How it works: An LLM (FrankenAgent) constructs detailed annotations; a diffusion-based model learns to follow part, action, and sequence text together.
  • Why it matters: This brings studio-level precision to everyday tools for animators, game devs, AR/VR creators, and robot designers.

🍞 Bottom Bread (Anchor): Now you can say: “Walk forward, get pushed left, then walk right; keep the right hand holding the rail,” and the character does it—smoothly and on time.

02Core Idea

🍞 Top Bread (Hook): You know how a band sounds best when each instrument follows the conductor and the sheet music at the same time?

🥬 Filling (The Actual Concept):

  • What it is: The key idea is to learn atomic, part-level motion pieces and compose them over time using hierarchical text guidance (parts, actions, sequence).
  • How it works: A new dataset (FrankenStein) labels each body part with time-aware prompts; the model (FrankenMotion) reads part, action, and sequence text together and generates coordinated motion via diffusion.
  • Why it matters: This lets you precisely control head, arms, legs, spine, and trajectory—alone or together—while keeping the whole motion realistic and meaningful.

🍞 Bottom Bread (Anchor): You can create a sequence like “stand, climb stairs; right hand holds the rail,” and the model keeps the hand constraint while moving the legs correctly.

🍞 Top Bread (Hook): Imagine dressing a paper doll: you layer shirt (sequence idea), pants (action segments), and accessories (part details) to get the final look.

🥬 Filling (The Actual Concept - FrankenMotion):

  • What it is: FrankenMotion is a diffusion-based motion generator that uses three levels of text at once.
  • How it works: It fuses part-level prompts (per frame), action segments (windows), and one sequence caption (global) into a shared space, then denoises to a clean motion sequence.
  • Why it matters: Without multi-level fusion, either tiny details or big-picture flow gets lost.

🍞 Bottom Bread (Anchor): When asked for “turn around, sit, stand, then walk,” with “right arm stays by side,” FrankenMotion respects both the timeline and the arm constraint.

🍞 Top Bread (Hook): Think of a librarian who can split big books into chapters and then into paragraphs—and then reassemble them in new ways.

🥬 Filling (The Actual Concept - FrankenStein dataset):

  • What it is: FrankenStein is a multi-level annotated dataset (sequence, action, body-part) with fine temporal detail.
  • How it works: An LLM (FrankenAgent) infers body-part labels and timestamps from high-level annotations, marking unknown when unsure to avoid guesses.
  • Why it matters: High-quality, atomic labels teach the model accurate, composable motion elements.

🍞 Bottom Bread (Anchor): From “lift above head then put down,” the dataset adds arms up, spine bend, hand tap timing, making the learning signal rich and precise.

🍞 Top Bread (Hook): You know how building blocks can make castles, rockets, or animals? Same blocks, different combos.

🥬 Filling (The Actual Concept - Part-level control):

  • What it is: Part-level control means you can tell specific body parts what to do at specific times.
  • How it works: The model reads per-frame prompts like “head: look left,” “left arm: rest,” “legs: climb.”
  • Why it matters: Without this, finely tuned motions (like keeping a hand on a rail) fall apart.

🍞 Bottom Bread (Anchor): “Head look left from 2s to 4s while sitting” is now doable and consistent.

🍞 Top Bread (Hook): A coach breaks a routine into drills: warm-up, sprint, cooldown.

🥬 Filling (The Actual Concept - Action-level control):

  • What it is: Action segments split the sequence into named windows like “walk,” “stumble,” “stand.”
  • How it works: Each window adds meaning so parts know how to cooperate during that interval.
  • Why it matters: Without action context, parts might move correctly but tell the wrong overall story.

🍞 Bottom Bread (Anchor): During a “stumble” segment, legs shuffle and arms react, then settle in the next “stand” segment.

🍞 Top Bread (Hook): A movie has an overall plot tying scenes together.

🥬 Filling (The Actual Concept - Sequence-level control):

  • What it is: A single global caption sets the big goal for the whole motion.
  • How it works: The model uses it as a north star so all parts and actions stay on theme.
  • Why it matters: Without it, local motions could be correct but feel disjointed.

🍞 Bottom Bread (Anchor): “Walk to knock, then walk back” keeps the motion’s purpose clear throughout.

🍞 Top Bread (Hook): Picture a three-story cake; each layer supports the next.

🥬 Filling (The Actual Concept - Hierarchical conditioning):

  • What it is: The model mixes part, action, and sequence texts together so they guide the motion jointly.
  • How it works: Text features from all levels are fused with noisy motion, then the diffusion model denoises them into a clean, coordinated result.
  • Why it matters: This prevents conflicts (like parts ignoring the action) and improves realism.

🍞 Bottom Bread (Anchor): The right arm can keep holding the rail (part) during “climb” (action) in a “stair ascent” story (sequence) without drifting.

🍞 Top Bread (Hook): Changing one instrument in a song changes the whole vibe.

🥬 Filling (The Actual Concept - Why it works):

  • What it is: Composing atomic parts under a clear hierarchy lets the model cover tiny details and long-range meaning together.
  • How it works: The dataset provides precise supervision; the diffusion model learns to align text layers with motion at every step.
  • Why it matters: This design beats models that only stitch parts at test time or blur all text into one bag of words.

🍞 Bottom Bread (Anchor): Compared to baselines, FrankenMotion turns “turn around” into a real 180° pivot at the right time, not a vague shuffle.

03Methodology

🍞 Top Bread (Hook): Imagine following a cooking show that shows the dish (final motion), the recipe (text prompts), and the timer (when each step happens).

🥬 Filling (The Actual Concept - High-level pipeline):

  • What it is: Input → Text encoders + motion features → Hierarchical fusion → Diffusion denoising → Output motion.
  • How it works: We take three kinds of text (sequence, action windows, per-frame parts), embed them, fuse them with noisy motion, and denoise to clean motion using a transformer-based diffusion model.
  • Why it matters: Without this layered fusion and denoising, details or timing would get lost.

🍞 Bottom Bread (Anchor): For “walk forward, get pushed left, then walk right; right hand holds rail,” the model blends part, action, and sequence info to generate a smooth, believable motion.

🍞 Top Bread (Hook): Think of a stick figure where each joint has a role.

🥬 Filling (The Actual Concept - SMPL pose representation):

  • What it is: A per-frame vector includes pelvis height and velocities, facing rotation speed, SMPL pose (in a stable 6D form), and joint positions.
  • How it works: These features make motion both expressive and rotation-aware while staying numerically stable.
  • Why it matters: If the pose is poorly represented, the model can wobble or twist unnaturally.

🍞 Bottom Bread (Anchor): When turning around, angular velocity and joint positions change smoothly, so the model learns a crisp 180° pivot.

🍞 Top Bread (Hook): You know how we use different sticky notes for different tasks?

🥬 Filling (The Actual Concept - Multi-granularity control inputs):

  • What it is: Three text inputs: a single sequence caption, windowed action labels, and per-frame part prompts for K body parts.
  • How it works: Users can provide any subset at inference; the model was trained to handle sparse or full control.
  • Why it matters: Without flexible inputs, small edits (“keep left arm still”) would be hard.

🍞 Bottom Bread (Anchor): You can add just “head: look left from 1s–2s” to tweak a generated clip without rewriting everything.

🍞 Top Bread (Hook): Translating from words to understanding is like turning a recipe into flavors.

🥬 Filling (The Actual Concept - Text and motion embeddings with CLIP and PCA):

  • What it is: Text is encoded with CLIP; action and part embeddings are dimension-reduced (PCA) for efficiency and stability.
  • How it works: We align action windows with frames and concatenate part+action features per frame, then fuse with noisy motion through an MLP.
  • Why it matters: If text and motion aren’t in the same “conversation,” the model can’t follow instructions.

🍞 Bottom Bread (Anchor): “Right arm up; legs stationary” becomes a compact vector that the model can use frame by frame.

🍞 Top Bread (Hook): A movie has both scenes and the overall theme.

🥬 Filling (The Actual Concept - Sequence context and timestep tokens):

  • What it is: The sequence caption becomes a global token; the diffusion timestep is another token.
  • How it works: We append these tokens to the per-frame features, creating a (T+2)-token sequence for the transformer.
  • Why it matters: The global theme and the current denoising step guide consistent, stable generation.

🍞 Bottom Bread (Anchor): The “walk to knock, then walk back” theme stays coherent while denoising from noisy to clean motion.

🍞 Top Bread (Hook): Think of cleaning a blurry photo step by step until it’s sharp.

🥬 Filling (The Actual Concept - Diffusion-based motion generation):

  • What it is: Start with noisy motion; a transformer predicts the clean motion at each diffusion step.
  • How it works: The network takes the noisy sequence, timestep, and all text features, and learns to denoise toward the ground-truth motion.
  • Why it matters: Diffusion provides stable, high-quality generation that respects conditioning.

🍞 Bottom Bread (Anchor): From noise, the model reveals “turn, sit, stand” with the right arm behavior intact.

🍞 Top Bread (Hook): Sometimes a schedule makes all the difference.

🥬 Filling (The Actual Concept - Robust training with masking):

  • What it is: Unknown labels are zeroed; we also randomly mask known labels using a Beta-distributed rate.
  • How it works: This teaches the model to be robust when some parts or actions are missing at test time.
  • Why it matters: Without masking, the model might break when given sparse instructions.

🍞 Bottom Bread (Anchor): Even if you only specify “legs: climb stairs,” the model still produces realistic whole-body motion.

🍞 Top Bread (Hook): A good loss function is like a coach giving clear feedback.

🥬 Filling (The Actual Concept - Training objective):

  • What it is: Standard DDPM sample-prediction loss—predict the clean motion from the noisy input.
  • How it works: The model minimizes the difference between predicted clean motion and true motion across timesteps.
  • Why it matters: Without this, denoising wouldn’t converge to realistic results.

🍞 Bottom Bread (Anchor): Over training, “turn around” becomes a reliable, well-timed pivot.

🍞 Top Bread (Hook): Tools and settings matter, like oven temperature for baking.

🥬 Filling (The Actual Concept - Implementation details):

  • What it is: Cosine noise schedule with 100 steps; AdamW optimizer; CLIP ViT-B/32 for text; trained on modern GPUs.
  • How it works: These choices stabilize learning and keep text-motion alignment crisp.
  • Why it matters: A shaky setup can lead to jittery, unrealistic motions.

🍞 Bottom Bread (Anchor): The final clips look smooth and match the prompts, not wobbly or off-beat.

🍞 Top Bread (Hook): Let’s walk through a tiny example.

🥬 Filling (The Actual Concept - Step-by-step mini example):

  • What it is: Prompt: “Stand, then climb stairs; right hand holds rail; head looks left at 2–3s.”
  • How it works:
    1. Encode sequence: “stand then climb stairs.”
    2. Encode actions: [0–2s: stand], [2–5s: climb].
    3. Encode parts per frame: head look-left at 2–3s; right arm hold-rail 0–5s; legs: climb at 2–5s.
    4. Fuse all text with noisy motion; transformer denoises over 100 steps.
    5. Output: a smooth motion with exact timing and part behavior.
  • Why it matters: Shows how all layers cooperate in practice.

🍞 Bottom Bread (Anchor): You get a believable stair climb with steady rail contact and a timely head turn.

04Experiments & Results

🍞 Top Bread (Hook): Imagine a talent show where each performer is judged on doing the right moves and looking natural doing them.

🥬 Filling (The Actual Concept - The Test):

  • What it is: The authors checked two things—semantic correctness (does the motion match the text?) and realism (does it look natural and varied?).
  • How it works: They trained separate evaluation models for parts, actions, and full sequences to score text-motion alignment, plus standard realism scores.
  • Why it matters: A motion that matches the words but looks robotic—or looks nice but ignores instructions—is not good enough.

🍞 Bottom Bread (Anchor): If the text says “turn around,” does the body really pivot 180° at the right time, and does it look human?

🍞 Top Bread (Hook): It’s not fair to race without other runners.

🥬 Filling (The Actual Concept - The Competition):

  • What it is: They compared FrankenMotion to adapted baselines: STMC (stitches parts after the fact), UniMotion (hierarchical but no explicit part control), and DART (autoregressive control).
  • How it works: All baselines were trained on the same data and given similar cues to ensure fairness.
  • Why it matters: Strong benchmarks make the win meaningful.

🍞 Bottom Bread (Anchor): In tricky sequences like “walk, turn, sit, stand, walk,” baselines often missed the precise turn or repeated segments; FrankenMotion didn’t.

🍞 Top Bread (Hook): Scoreboards need clear grades, not just numbers.

🥬 Filling (The Actual Concept - The Scoreboard):

  • What it is: Semantic scores (like R-Precision and M2T) measure how well motions match text; realism uses FID and Diversity.
  • How it works: Higher R-Precision and M2T are better; lower FID is better; healthy Diversity means motions aren’t copies.
  • Why it matters: Together, they tell if the motion is both correct and life-like.

🍞 Bottom Bread (Anchor): FrankenMotion scored higher than baselines on part/action/sequence correctness and achieved top realism (e.g., best or near-best FID), meaning “A+ when others got B’s.”

🍞 Top Bread (Hook): Sometimes the best magic trick is combining old cards in a new way.

🥬 Filling (The Actual Concept - Surprising findings):

  • What it is: Even using only part-level inputs, the model performed strongly—then improved further when adding action and sequence text.
  • How it works: Part atoms carry a lot of signal; higher-level text adds meaning and smooth coordination.
  • Why it matters: It shows the core idea—atomic composition under a hierarchy—really works.

🍞 Bottom Bread (Anchor): The system could generate plausible motions it never saw paired before, like “sit while raising only the left arm,” without breaking realism.

🍞 Top Bread (Hook): Trust but verify.

🥬 Filling (The Actual Concept - Data quality check):

  • What it is: Human experts judged the LLM-generated annotations ~93% correct with high agreement.
  • How it works: Multiple raters scored labels per part/action/sequence; agreement metrics confirmed reliability.
  • Why it matters: Good training labels are the backbone of good models.

🍞 Bottom Bread (Anchor): If the data says “arms up at 1.5–2.6s,” the model learns that timing and uses it faithfully.

05Discussion & Limitations

🍞 Top Bread (Hook): Even great tools have a few knobs that still need tuning.

🥬 Filling (The Actual Concept - Limitations):

  • What it is: The model doesn’t yet generate very long (minute-scale) motions in a single pass.
  • How it works: Current diffusion and transformer context make extremely long sequences challenging.
  • Why it matters: Some stories—like full scenes or long robot tasks—need longer memory.

🍞 Bottom Bread (Anchor): Think of a marathon dance routine; the model currently prefers shorter, well-phrased numbers.

🍞 Top Bread (Hook): You can’t build a treehouse without wood and tools.

🥬 Filling (The Actual Concept - Required resources):

  • What it is: Training uses GPUs, a CLIP text encoder, and preprocessed motion (SMPL) data; the FrankenStein annotations come from an LLM agent.
  • How it works: With standard deep learning hardware and the released code/data, teams can reproduce and extend the system.
  • Why it matters: Clear requirements help others build on this work.

🍞 Bottom Bread (Anchor): A single H100 or A100-class GPU can train evaluation models; the main model trained in under two days.

🍞 Top Bread (Hook): Sometimes a screwdriver isn’t the right tool for a nail.

🥬 Filling (The Actual Concept - When not to use):

  • What it is: If you need minute-long continuous motion, super-precise physics (forces/contacts), or very unusual body morphologies out of distribution, results may degrade.
  • How it works: The model focuses on controllable kinematics guided by text, not full physics simulation or extreme long-horizon planning.
  • Why it matters: Picking the right tool saves time and avoids frustration.

🍞 Bottom Bread (Anchor): For heavy object lifting with realistic strain forces, a physics-based controller might be more appropriate.

🍞 Top Bread (Hook): Questions are the seeds of the next garden.

🥬 Filling (The Actual Concept - Open questions):

  • What it is: Extending to much longer sequences; tighter physics; richer interactions with scenes/objects; and multimodal prompts (audio, video).
  • How it works: Chunked diffusion, memory mechanisms, or hybrid physics-diffusion could help; better LLM reasoning might add even cleaner labels.
  • Why it matters: These steps would make motion generation even more versatile and trustworthy.

🍞 Bottom Bread (Anchor): Imagine saying, “Dance to this song across the whole room, avoid chairs, and keep your left hand on the rail,” for a full minute—future systems could nail it.

06Conclusion & Future Work

🍞 Top Bread (Hook): Think of choreographing a dance where each finger, arm, and step matters, all timed perfectly to the music.

🥬 Filling (The Actual Concept - 3-sentence summary):

  • What it is: FrankenMotion is a diffusion-based model that composes atomic, body-part motions over time using hierarchical text—parts, actions, and sequence.
  • How it works: A new dataset, FrankenStein, created with an LLM agent, provides detailed, time-aligned labels for body parts and actions, enabling precise learning.
  • Why it matters: The system achieves finer control and higher realism than prior methods and can compose new motions it never explicitly saw.

🍞 Bottom Bread (Anchor): Ask for “walk forward, get pushed left, then walk right, with right hand holding the rail,” and the character does exactly that—cleanly and believably.

Main Achievement: Delivering the first framework and dataset to provide atomic, temporally-aware, part-level annotations and a model that truly uses them for multi-level control.

Future Directions: Longer sequences in one pass, tighter physics and contact modeling, deeper scene/interaction grounding, and richer multimodal control.

Why Remember This: It turns motion generation from rough sketches into precise choreography, giving creators and robots fine-grained, reliable control over how bodies move through time.

Practical Applications

  • •Game animation authoring: Script precise limb behaviors (e.g., weapon-hand constraints) while maintaining natural whole-body movement.
  • •Film and TV previsualization: Block complex action sequences with timed part constraints (head turns, hand poses) quickly.
  • •VR training simulations: Ensure safety cues like “keep right hand on the rail” while performing multi-step procedures.
  • •Robotics motion sketching: Prototype human-like motions for social robots with fine-grained part control before deploying controllers.
  • •Sports instruction content: Generate step-by-step demonstrations (e.g., pivot footwork, arm angles) for coaching and tutorials.
  • •Accessibility tools: Create sign-language-inspired or gesture-based avatars with controllable hand and arm motions.
  • •Education and edutainment: Build interactive lessons showing exactly how body parts coordinate in actions (climb, sit, turn).
  • •Motion editing pipelines: Add or tweak a single part instruction (e.g., “head: look left at 2s”) without redoing the whole clip.
  • •AR character overlays: Align avatars to real scenes with specific part constraints that respect obstacles or supports.
  • •Prototype testing for HRI: Simulate human motions with precise part constraints when testing robot perception and planning.
#Human motion generation#Part-level control#Hierarchical conditioning#Diffusion model#Transformer#FrankenMotion#FrankenStein dataset#LLM-based annotation#CLIP embeddings#SMPL#PCA#R-Precision#FID#Temporal control#Compositional motion
Version: 1