Making AI agents robust for using imperfect tools
Tool-augmented LLM agents often rely on tools that are usually correct but can occasionally return plausible, incorrect outputs. In this project, we will mitigate these silent errors through a cost-aware strategy that combines cross-tool consistency checks with targeted verification at high-risk steps.





