ToolSafe is a new way to keep AI agents safe when they use external tools, by checking each action before it runs.
MAXS is a new way for AI agents to think a few steps ahead while using tools like search and code, so they make smarter choices.
Agents often act like tourists without a map: they react to what they see now and miss long-term consequences.
This paper teaches AI to build and improve its own small computer helpers (tools) while solving science problems, instead of relying only on a fixed toolbox made beforehand.
OpenTinker is an open-source system that makes training AI agents with reinforcement learning simple, modular, and reusable.
EnvScaler is an automatic factory that builds many safe, rule-following practice worlds where AI agents can talk to users and call tools, just like real apps.
This paper turns an AI agent’s memory from a flat list of notes into a logic map of events connected by cause-and-time links.
This paper teaches a computer agent to grow a toolbox of skills that are real, runnable programs, not just text ideas.
This paper presents BEDA, a simple way to make chatty AI act strategically by turning what it believes into gentle rules (probabilistic constraints) that guide what it can say.
Youtu-Agent is a build-and-grow factory for AI agents that cuts manual setup and keeps agents improving over time.
GenEnv is a training system where a student AI and a teacher simulator grow together by exchanging tasks and feedback.
This paper builds InternGeometry, a large language model agent that solves Olympiad-level geometry by talking to a math engine, remembering what worked, and trying smart new ideas.