4bis. AI, models, and inference portability
The CTO saw it first in the weekly dashboard: median inference latency had drifted from 1.4 to 2.8 seconds over three weeks. No alerts. No incident. The provider's status page was green. But two days later, the internal eval suite came back. Contract extraction quality had