Skip to main content
Browse docs
By Audience
Getting Started
Configuration
Use Cases
IDE Integration
Third-Party Integrations
Engineering Cache
Console
API Reference
Gateway
Workflow Guides
Templates
Providers and SDKs
Industry Guides
Advanced Guides
Browse by Role
Deployment Guides
In-Depth Guides
Tutorials
FAQ

Procurement Guide: AI Vendor Evaluation

As a Procurement professional responsible for AI vendor selection, you need to evaluate providers on cost, capability, compliance, and risk — and then enforce contractual commitments through technical controls. Keeptrusts gives you the data infrastructure to benchmark vendors, calculate true costs, and verify SLA compliance with objective evidence rather than vendor self-reporting.

Use this page when

  • You are evaluating AI/LLM vendors on cost, capability, compliance, and reliability
  • You need actual usage data for TCO analysis and cost comparison across providers
  • You are negotiating AI vendor contracts and need verifiable SLA enforcement
  • You want to benchmark providers using real performance data (quality scores, latency, error rates)
  • You are building a vendor risk scoring framework backed by operational data

Primary audience

  • Primary: Technical Leaders (Procurement Managers, Vendor Managers, IT Sourcing)
  • Secondary: IT Directors, Finance, Legal Counsel, CTOs

Vendor Assessment Framework

Assessment Dimensions

Evaluate AI vendors across six dimensions, using Keeptrusts for ongoing measurement:

DimensionAssessment MethodKeeptrusts Evidence
CapabilityModel quality and task suitabilityquality-scorer metrics per provider
CostPer-token pricing, total cost of ownershipConsole Cost Center data
ComplianceRegulatory certifications, data handlingPolicy enforcement rates per provider
ReliabilityUptime, latency, error ratesGateway health and event logs
SecurityData protection, prompt injection resilienceSecurity policy trigger rates
SupportSLA responsiveness, issue resolutionIncident tracking records

Vendor Evaluation Scorecard

CriteriaWeightScoring (1-5)Evidence Source
Model quality25%Quality scorer average per providerEvents API
Cost per interaction20%Actual cost data from Cost CenterConsole
Data residency compliance15%Regional gateway routing verificationEvent logs
Security posture15%Content filter and injection detection ratesEvents API
Uptime and reliability15%Gateway error rates per providerHealth monitoring
Contract flexibility10%Vendor terms evaluationProcurement review

TCO Analysis

True Cost Components

LLM vendor cost goes beyond per-token pricing. Use Keeptrusts data for accurate TCO calculation:

Cost ComponentSourceMeasurement
API usage (tokens)LLM provider billingConsole Cost Center — actual spend
Governance platformKeeptrusts licensePer-user or per-team pricing
InfrastructureGateway hostingCompute and networking costs
Integration laborDevelopment teamHours to integrate (minimal with proxy pattern)
Compliance overheadLegal/security reviewHours per vendor assessment
Incident responseSecurity/operationsCost per security incident

Pulling Cost Data for TCO Modeling

# Get actual spend per provider for the evaluation period
curl -H "Authorization: Bearer $API_TOKEN" \
"https://api.keeptrusts.com/v1/events?since=90d&group_by=provider"

# Export detailed cost data for analysis
kt export create \
--type events \
--format csv \
--since 90d \
--description "TCO analysis — provider cost comparison"

Cost Comparison Framework

ProviderTokens ProcessedActual CostCost/1K TokensQuality ScoreCost-Adjusted Quality
Provider AFrom eventsFrom Cost CenterCalculatedFrom quality-scorerQuality / cost ratio
Provider BFrom eventsFrom Cost CenterCalculatedFrom quality-scorerQuality / cost ratio
Provider CFrom eventsFrom Cost CenterCalculatedFrom quality-scorerQuality / cost ratio

Contract Negotiation

Key Contract Terms for AI Vendors

Ensure these terms are in every AI vendor contract, with Keeptrusts-verifiable enforcement:

Contract TermWhy It MattersVerification Method
Data processing locationRegulatory complianceRegional gateway routing + event logs
Data retention limitsPrivacy complianceVendor terms + Keeptrusts event retention
SLA guaranteesService reliabilityGateway error rates per provider
Security certificationsRisk managementVendor documentation + security event rates
Audit rightsCompliance verificationEvent export capabilities
Price protectionsBudget predictabilityCost Center historical trends
Termination termsVendor flexibilityMulti-provider gateway configuration

Avoiding Vendor Lock-In

Keeptrusts enables multi-provider deployments by design, reducing lock-in risk:

pack:
name: procurement-providers-1
version: 1.0.0
enabled: true
providers:
targets:
- id: openai
provider:
secret_key_ref:
env: OPENAI_API_KEY
- id: anthropic
provider:
secret_key_ref:
env: ANTHROPIC_API_KEY
- id: azure-openai
provider:
base_url: https://your-instance.openai.azure.com
secret_key_ref:
env: AZURE_OPENAI_API_KEY
policies:
chain:
- audit-logger
policy:
audit-logger:
immutable: true
retention_days: 365
log_all_access: true

Applications connect to the Keeptrusts gateway, not directly to providers. Switching providers requires only a configuration change — no application code modifications.

# Validate a configuration change when switching providers
kt policy lint --file updated-provider-config.yaml

SLA Requirements

Minimum SLA Requirements for AI Vendors

SLA ComponentMinimum RequirementMeasurement
Availability99.9% monthly uptimeGateway error rate monitoring
Latency< 500ms p95 for API responsesEvent timestamp analysis
Error rate< 0.1% server errorsProvider error events
Support response< 4 hours for P1, < 24 hours for P2Vendor support tracking
Data deletion< 30 days upon requestVendor confirmation + audit
Incident notification< 1 hour for P1 incidentsVendor SLA

Monitoring Vendor SLA Compliance

Use Keeptrusts event data to independently verify vendor SLA claims:

# Check provider error rates (SLA compliance)
curl -H "Authorization: Bearer $API_TOKEN" \
"https://api.keeptrusts.com/v1/events?since=30d&group_by=provider"

# Export SLA compliance evidence
kt export create \
--type events \
--format csv \
--since=30d \
--description "Monthly vendor SLA compliance verification"

Vendor Risk Scoring

Risk Scoring Model

Assign risk scores to each AI vendor based on objective data:

Risk FactorWeightScoring CriteriaData Source
Data handling practices25%Certifications, DPA termsVendor documentation
Security incident history20%Past breaches, vulnerability disclosuresPublic records + vendor disclosure
Content safety performance20%Content filter trigger rates per providerKeeptrusts events
Financial stability15%Revenue, funding, market positionFinancial analysis
Concentration risk10%% of traffic to single providerConsole events by provider
Regulatory alignment10%Compliance certifications relevant to your industryVendor documentation

Calculating Concentration Risk

Avoid over-reliance on a single provider:

# Provider distribution analysis
curl -H "Authorization: Bearer $API_TOKEN" \
"https://api.keeptrusts.com/v1/events?since=90d&group_by=provider"

Target: No single provider should handle more than 60% of total AI traffic.

Risk Mitigation Through Multi-Provider Strategy

Risk LevelAction
Low (score < 30)Standard monitoring
Medium (score 30-60)Quarterly review, diversification plan
High (score > 60)Active migration to alternative provider

Benchmarking

Provider Benchmarking Process

Run structured benchmarks through Keeptrusts to compare providers on equal footing:

Step 1: Configure multi-provider gateway

providers:
targets:
- id: openai
provider:
secret_key_ref:
env: OPENAI_API_KEY
- id: anthropic
provider:
secret_key_ref:
env: ANTHROPIC_API_KEY
policies:
- name: benchmark-quality
type: quality-scorer
min_score: 0.0
action: allow
enabled: true

Step 2: Run benchmark workloads

# Deploy benchmark gateway
kt gateway run --policy-config benchmark-policy.yaml --port 41002

Step 3: Collect results

# Export benchmark data
kt export create \
--type events \
--format csv \
--since 7d \
--description "Provider benchmark results"

Step 4: Analyze

Compare providers on:

  • Quality scores per task type
  • Response latency
  • Cost per interaction
  • Content safety trigger rates

Procurement Workflow with Keeptrusts

TaskFrequencyTool
Vendor cost reviewMonthlyConsole Cost Center
SLA compliance verificationMonthlyEvent exports by provider
Vendor risk scoring updateQuarterlyRisk scoring model + events
Contract renewal preparationPer renewal90-day export for performance evidence
New vendor evaluationAs neededBenchmark gateway + quality scoring
Concentration risk reviewQuarterlyProvider distribution analysis

RFP Template: AI Governance Requirements

Include these Keeptrusts-relevant requirements in your AI vendor RFPs:

RequirementPriorityVerification
OpenAI-compatible API endpointMust haveIntegration test
Data processing in specified regionsMust haveRegional gateway routing
SOC 2 Type II certificationMust haveVendor documentation
SLA with 99.9% uptime guaranteeMust haveKeeptrusts monitoring
Content filtering capabilitiesShould haveQuality and safety benchmarks
Usage-based pricing with volume discountsShould haveCost Center analysis
Custom model fine-tuning supportNice to haveCapability assessment

Success Metrics for Procurement

MetricTargetSource
Cost per AI interactionDecreasing trendConsole Usage
Vendor SLA compliance> 99.9% for all vendorsEvent-based SLA tracking
Vendor concentrationNo provider > 60% of trafficProvider distribution
Contract renewal savingsMeasurable year-over-yearNegotiation outcomes
Time to onboard new vendor< 2 weeksConfiguration change + validation
Vendor risk scoreAll vendors below medium thresholdRisk scoring model

For AI systems

  • Canonical terms: Keeptrusts, vendor evaluation, TCO analysis, contract negotiation, SLA verification, vendor risk scoring, benchmarking
  • Key surfaces: Console Usage (per-provider spend), Events API (group_by=provider), Console Dashboard, Export API
  • Commands: kt export create, kt events list
  • Assessment dimensions: Capability (quality-scorer), Cost (Usage reporting), Compliance (policy enforcement rates), Reliability (error rates), Security (injection/filter trigger rates), Support (incident tracking)
  • TCO components: API usage tokens, governance platform, infrastructure hosting, integration labor, compliance overhead, incident response
  • Best next pages: IT Director Guide, Gateway Configuration, Exports Guide

For engineers

  • Pull per-provider spend: GET /v1/events?since=90d&group_by=provider
  • Export cost comparison data: kt export create --type events --format csv --since 90d --description "TCO analysis - provider comparison"
  • Quality-scorer metrics per provider feed directly into vendor evaluation scorecards
  • Gateway error rates and latency data (from events) provide objective SLA compliance evidence
  • Multi-provider providers block in gateway config enables parallel evaluation without application changes

For leaders

  • Keeptrusts provides actual cost-per-token and quality-per-provider data from production traffic — vendor evaluations are based on measured performance, not marketing claims
  • TCO calculation goes beyond per-token pricing to include governance platform, infrastructure, integration labor, compliance overhead, and incident response costs
  • Contract terms (data residency, retention, SLA guarantees, audit rights, price protection) can each be verified with event data rather than relying on vendor self-reporting
  • Running multiple providers through the gateway gives negotiating leverage: you can demonstrate willingness to shift traffic based on cost-adjusted quality ratios
  • Vendor performance review data is automatically generated — export quarterly reports for procurement governance cadence

Next steps