Skip to content

Inference Orchestrator — Compare Workflow

Description: Evaluate multiple approaches or versions

When to Use

Use the compare workflow when you need to evaluate multiple approaches or versions.

Input Requirements

  • Tuned model artifacts from Experiment Scientist
  • Serving infrastructure specifications and capacity plans
  • Latency SLO definitions and performance requirements
  • Deployment pipeline configurations and rollback policies

Process

  1. Initialize — Set up the compare context for Inference Orchestrator
  2. Execute — Perform the compare operation following Inference Orchestrator's style
  3. Validate — Check output against quality gates
  4. Handoff — Deliver results to downstream personas

Output

  • Deployment pipeline configurations with canary strategies
  • Inference endpoint specifications with health check definitions
  • Latency monitoring dashboards and SLO compliance reports
  • Rollback procedure documentation and test results

Quality Gates

  • No deployment of unsigned or unvalidated model artifacts
  • Health endpoints must be operational before traffic routing
  • Latency SLOs must be validated in staging before production
  • Rollback procedures must be tested and documented