Formerly known as Azure AI Services or Azure Cognitive Services is a unified collection of prebuilt AI capabilities within the Microsoft Foundry platform
Hello bamb00,
Welcome to Microsoft Q&A and Thank you for reaching out.
In addition to Evaluations and Observability token usage, it’s important to understand that “Foundry Tools” is an umbrella billing category that can include several platform-level capabilities beyond evaluation workflows.
Even if you are not intentionally running benchmarking or monitoring jobs, Foundry Tools charges can also come from:
1.Agent Tool Orchestration Runtime
If you are using Agents in Azure AI Foundry, the platform performs:
Tool planning and routing
Function/tool invocation orchestration
Conversation state management
- Retrieval pipelines
Even when the underlying model inference is billed elsewhere, the orchestration layer and tool execution compute are billed under Foundry Tools.
In agent-heavy workloads, this orchestration cost can exceed direct model inference costs.
2.Knowledge & Retrieval Pipelines
If your project uses Knowledge bases, Vector search / hybrid retrieval, Grounding connections, Fabric / SQL / Search connectors, MCP servers (preview)
Each request may trigger Retrieval calls, Context assembly, Additional model calls
These supporting operations are typically metered under Foundry Tools rather than Foundry Models.
3.Prompt Flow & Development Runs
Beyond formal “Evaluations,” development activities can also drive usage:
Prompt Flow test runs
Batch testing pipelines
Playground experimentation
CI/CD validation flows
Frequent test iterations can generate significant token consumption that appears under Foundry Tools.
4.Tool-Enabled Model Calls vs Direct Deployments
Another important distinction:
If you call a model through a direct Azure OpenAI deployment endpoint, costs appear under model deployment meters.
If you call the same model through Foundry SaaS endpoints, Agents, or evaluation pipelines, usage may appear under Foundry Tools.
So the billing category depends on the access path, not just the model itself.
How to Isolate Non-Evaluation Tool Costs
Since Evaluations/Observability has already been identified in your screenshot, you can further validate:
In Cost Analysis:
- Group by Meter
- Then group by Operation name
Look for non-evaluation meters (e.g., orchestration or tool execution related).
Compare timestamps of spikes with:
- Agent usage
- Prompt flow activity
- Scheduled runs
This will help determine whether the driver is:
- Monitoring workloads
- Agent/tool orchestration
- Development testing
- Or intentional benchmarking
Next Steps
If the footprint is larger than expected:
- Review whether continuous evaluation is enabled
- Check for scheduled or recurring evaluation jobs
- Review agent tool configurations (especially automatic retrieval)
- Reduce dataset sizes for monitoring runs
- Lower sampling rates
- Remove unused schedules or dev projects
Key Takeaway
“Foundry Tools” is not just model tokens, it represents platform-level intelligence, orchestration, and monitoring layers that operate on top of your models.
If it’s your largest cost category, it usually means one of two things:
- You are running evaluation/monitoring workflows at scale
- Or you are heavily using agents, retrieval, or tool-enabled pipelines.
Please refer this, Plan and manage costs for Microsoft Foundry https://learn.microsoft.com/azure/ai-foundry/concepts/manage-costs?view=foundry-classic
I Hope this helps. Do let me know if you have any further queries.
If this answers your query, please do click Accept Answer and Yes for was this answer helpful.
Thank you!