SWE-World lets code-fixing AI agents practice and learn without heavy Docker containers by using smart models that pretend to be the computer and tests.
SWE-Master is a fully open, step-by-step recipe for turning a regular coding model into a strong software-fixing agent that works across many steps, files, and tests.
MeKi is a new way to grow a language model’s knowledge by using storage (ROM) instead of extra heavy calculations (FLOPs).
The paper shows that even if a model is great at predicting when an AI agent will fail, jumping in to “fix” the agent mid-task can still make things worse.
This paper speeds up how AI models read very long texts by carefully choosing which words (tokens) to focus on at each step.
The paper introduces PRIVASIS, a huge, fully synthetic dataset (1.4 million records) filled with realistic-looking private details, but created from scratch so it does not belong to any real person.
FASA is a training-free method that makes large language models faster and lighter on memory by keeping only the most useful past tokens during decoding.
When rewards are rare, a popular training method for language models (GRPO) often stops learning because every try in a group gets the same score, so there is nothing to compare.
The paper solves a big problem in fast image generators: they got quick, but they lost variety and kept making similar pictures.
This paper shows that many hard math and AI problems can be solved with one shared idea called homotopy, where we move from an easy version of a problem to the real one step by step.
Large language models learn better when we spend more practice time on the right questions at the right moments.
LatentMem is a new memory system that helps teams of AI agents remember the right things for their specific jobs without overloading them with text.