This paper teaches AI teams to get better by scoring every move they make, not just the final answer.
Deep search agents can plan and browse the web in many steps, but they often fail because they don’t notice when their own thinking drifts off-track.
Chain-of-Thought (CoT) makes AI think step by step, but it is slow because it writes many tokens one by one.
Diffusion language models (dLLMs) can write text in any order, but common decoding methods still prefer left-to-right, which wastes their superpower.
DIFFA-2 is a new audio AI that listens to speech, sounds, and music and answers questions about them using a diffusion-style language model instead of the usual step-by-step (autoregressive) method.
Large reasoning models got very good at thinking step-by-step, but that sometimes made them too eager to follow harmful instructions.
Golden Goose turns messy internet text into clean multiple-choice puzzles that computers can learn from and get automatic rewards for.
Diffusion language models (dLLMs) generate several tokens at once but usually throw away lots of helpful clues each step—RCD keeps and reuses those clues.
DINO-SAE is a new autoencoder that keeps both the meaning of an image (semantics) and tiny textures (fine details) at the same time.
This paper builds a smart team of AI helpers, called MEnvAgent, that automatically sets up the right computer environments for code projects in many languages.
BatCoder teaches a code model to write both code and its documentation by doing a round trip: from code to docs and back to code.
This paper fixes a hidden mismatch in image generation: tokenizers make tokens without order, but generators need an order to predict the next token well.