ToolSafe: Enhancing Tool Invocation Safety of LLM-based agents via Proactive Step-level Guardrail and Feedback
IntermediateYutao Mou, Zhangchi Xue et al.Jan 15arXiv
ToolSafe is a new way to keep AI agents safe when they use external tools, by checking each action before it runs.
#step-level safety#tool invocation#LLM agents