Metric Set
A practical guide to metric set for AI procurement and vendor-risk practitioners.
What This Lesson Covers
Metric Set is a key topic within Performance Monitoring. In this lesson you will learn the underlying procurement and vendor-risk discipline, the contractual or operational lever that gives the buyer control, how to apply the procedures to real AI vendor relationships, and the open questions practitioners are actively working through. By the end you will be able to engage with metric set in real AI procurement and vendor-risk work with confidence.
This lesson belongs to the Ongoing Vendor Management category of the AI Procurement & Vendor Risk track. AI procurement and vendor-risk management sits at the intersection of procurement, third-party risk, security, privacy, legal, and AI engineering. Understanding the ongoing vendor-management practice that keeps the contract alive in everyday operations is what lets you build a vendor relationship that delivers value while limiting downside.
Why It Matters
Monitor AI vendor performance continuously rather than only when there is a problem. Learn the metric set (uptime, latency p50/p95/p99, model quality on a held-out eval set, error rate, drift on canary inputs), instrumentation that does not depend on vendor self-reporting, dashboards that the vendor manager and the business owner share, anomaly detection, and the escalation playbook when metrics breach thresholds.
The reason metric set deserves dedicated attention is that AI vendor markets are immature: vendor terms are slanted toward providers, attestation standards (ISO 42001 in particular) are only just becoming consumable, model deprecation cycles are aggressive, IP litigation is unsettled, and the EU AI Act has just begun applying flow-down obligations to upstream providers. Practitioners who reason from first principles will navigate the next vendor pitch, the next renewal, the next incident, and the next regulatory inquiry far more effectively than those who only have a checklist.
How It Works in Practice
Below is a practical procurement and vendor-risk pattern for metric set. Read through it once, then think about how you would apply it to a real AI vendor in your portfolio.
# Ongoing vendor management pattern
OVM_STEPS = [
'Run operating rhythm (ops + business + executive reviews)',
'Monitor performance with own instrumentation, not vendor self-reporting',
'Refresh attestations on cycle (SOC 2 annually, ISO certs per cycle)',
'Run annual recertification + renewal/exit decision',
'Track and assess vendor-driven changes (model deprecations included)',
'Update risk register and report KRIs to TPRM committee',
]
Step-by-Step Analytical Approach
- Establish the criteria — What is the policy, standard, or contractual requirement that governs this decision (procurement policy, TPRM policy, security baseline, AI governance, AI Act provider/deployer rules, sector regulation)? Document the criteria up front; vendor decisions made without explicit criteria are intuition, not governance.
- Tier the vendor — Map the vendor to the right tier (critical / high / medium / low) so DD depth, contract requirements, and ongoing management intensity are proportionate to risk.
- Plan the evidence — For DD, lay out questionnaires, attestations, independent technical evaluation, and on-site / live assessment as appropriate. For contracts, identify the AI-specific clauses and the negotiation strategy. For ongoing management, define the metric set and the operating rhythm.
- Collect sufficient appropriate evidence — Multiple sources, time-stamped, hash-pinned where applicable, independent of vendor self-reporting. The bar is what a sophisticated reviewer (board, regulator, customer, plaintiff) would expect.
- Form the decision — Compare evidence to criteria; identify residual risks; route for acceptance per matrix; document the audit trail; communicate the decision to vendor and stakeholders.
- Operate the relationship — Onboard, monitor, refresh attestations on cadence, run incident response, manage change, recertify annually, exit cleanly when the time comes.
When This Topic Applies (and When It Does Not)
Metric Set applies when:
- You are evaluating, contracting with, or managing an AI vendor (foundation-model API, AI SaaS, AI infrastructure, AI consulting)
- You are running a TPRM programme that includes AI vendors (regulated sector or otherwise)
- You are responding to a customer or regulator question about AI vendor governance
- You are operating under the EU AI Act and need to flow obligations through the vendor chain
- You are exiting a vendor or planning a vendor substitution
It does not apply (or applies lightly) when:
- The AI is genuinely first-party (built and operated entirely in-house with internal data)
- The vendor relationship is below the procurement-policy threshold for full TPRM (usually a small-dollar exception with no sensitive data)
- The work is research-only with no path to production
Practitioner Checklist
- Are the criteria for this decision explicit, written, and tied to the procurement / TPRM policy?
- Is the vendor risk-tiered, and is the depth of DD / contract / management proportionate to tier?
- Does the contract include AI-specific clauses (data rights, no-train, IP indemnity, SLAs, exit, AI Act flow-down)?
- Are residual risks documented, accepted at the right level, and tracked in the register?
- Is performance monitored from your own instrumentation, not vendor self-reporting?
- Are attestations refreshed on cycle, with gaps surfaced and addressed?
- Is the exit runbook in place before you need it, with parallel-running and data-migration plans?
Disclaimer
This educational content is provided for general informational purposes only. It does not constitute procurement, legal, or professional advice; it does not create a professional engagement; and it should not be relied on for any specific vendor selection, contract negotiation, or risk-acceptance decision. AI vendor markets, contracts, and regulatory obligations vary by jurisdiction and change rapidly. Consult qualified procurement, legal, and risk professionals for advice on your specific situation.
Next Steps
The other lessons in Performance Monitoring build directly on this one. Once you are comfortable with metric set, the natural next step is to combine it with the patterns in the surrounding lessons — that is where doctrinal mastery turns into a working AI procurement and vendor-risk programme. Procurement and vendor-risk are most useful as an integrated discipline covering intake, DD, RFx, contracts, ongoing management, incidents, and exit.
Lilly Tech Systems