KARL is a smart search helper that learns to look up information step by step and explain answers using the facts it finds.
This paper teaches long-horizon AI agents to remember everything exactly without stuffing their whole memory at once.
MMR-Life is a new test (benchmark) that checks how AI understands everyday situations using several real photos at once.
CHIMERA is a small (about 9,000 examples) but very carefully built synthetic dataset that teaches AI to solve hard problems step by step.
SLATE is a new way to teach AI to think step by step while using a search engine, giving feedback at each step instead of only at the end.
WorldCompass teaches video world models to follow actions better and keep pictures pretty by using reinforcement learning after pretraining.
LLaDA2.1 teaches a diffusion-style language model to write fast rough drafts and then fix its own mistakes by editing tokens it already wrote.
RLAnything is a new reinforcement learning (RL) framework that trains three things together at once: the policy (the agent), the reward model (the judge), and the environment (the tasks).
Kimi K2.5 is a new open-source AI that can read both text and visuals (images and videos) and act like a team of helpers to finish big tasks faster.
This paper introduces Foundation-Sec-8B-Reasoning, a small (8 billion parameter) AI model that is trained to “think out loud” before answering cybersecurity questions.
LLM agents are usually trained in a few worlds but asked to work in many different, unseen worlds, which often hurts their performance.
Academic rebuttals are not just about being polite; they are about smart, strategic persuasion under hidden information.