Skip to product information
1 of 1

XALON Tools™

Evaluation metric example: Check if tool was called

Evaluation metric example: Check if tool was called

Regular price $9.99 USD
Regular price $49.99 USD Sale price $9.99 USD
Sale Sold out
Plan

Say goodbye to guessing which tools your agent actually used!

This automation evaluates your AI workflow in n8n by checking whether the correct tool was triggered during execution. It compares the tools used by your agent with the expected tools from a dataset — giving you a clear performance score for each run.

Perfect for validating AI agent behavior in tool-using workflows like automation agents, copilots, or decision trees.

What it does:

🧪 Runs evaluation tests in parallel using a dedicated trigger

🧠 Captures the full list of tools the agent used during execution

🔍 Compares used tools against expected tools in your test dataset

📊 Sends a performance metric back to n8n so you can track accuracy

💰 Skips evaluation logic during standard runs to save resources

✅ Setup guide & importable automation included

Need help setting it up? We offer full configuration and testing for a one-time fee.

View full details