Agentic AI
Agentic Metrics
Understand and evaluate the performance of AI agents using Galileo’s agentic metrics.
Agentic metrics help you measure how well your AI agents perform complex, multi-step tasks—especially when those agents need to use tools, make decisions, or interact with external systems. These metrics and helpful for those for anyone building advanced AI assistants, workflow automation, or any system where the AI acts on behalf of a user.
Use agentic metrics when you want to:
- Track whether your agent is making meaningful progress toward its goals.
- Detect and diagnose errors that occur when your agent uses tools or APIs.
- Ensure your agent is choosing the best tools or actions for each situation.
Below is a quick reference table of all agentic performance metrics:
Name | Description | When to Use | Example Use Case |
---|---|---|---|
Tool Error | Detects errors or failures during the execution of tools. | When implementing AI agents that use tools and want to track error rates. | A coding assistant that uses external APIs to run code and must handle and report execution errors appropriately. |
Tool Selection Quality | Evaluates whether the agent selected the most appropriate tools for the task. | When optimizing agent systems for effective tool usage. | A data analysis agent that must choose the right visualization or statistical method based on the data type and user question. |
Action Advancement | Measures how effectively each action advances toward the goal. | When assessing whether an agent is making meaningful progress in multi-step tasks. | A travel planning agent that needs to book flights, hotels, and activities in the correct sequence. |