Mixture-of-Experts (MoE) models use many small specialist networks (experts) and a router to pick which experts handle each token, but the router isnβt explicitly taught what each expert is good at.
ProPhy is a new two-step method that helps video AIs follow real-world physics, not just make pretty pictures.