24h Latency Trend
When p95 latency rises while p50 remains stable, user-facing risk is increasing even if aggregate status still looks healthy. This is an early signal for timeout and retry pressure.
AI Checker Hub
Mistral API Status Today is an independent reliability page for teams who need clear operational context before shipping traffic to Mistral endpoints. It combines live status signals, rolling uptime windows, latency trend behavior, and recent incident windows in one production-friendly workflow.
When p95 latency rises while p50 remains stable, user-facing risk is increasing even if aggregate status still looks healthy. This is an early signal for timeout and retry pressure.
This page is designed for incident-time decisions. Start with overall state and update time, then inspect latency trend for direction. If degradation is short-lived, tune retry behavior first. If instability persists over multiple windows, switch critical paths to fallback with traffic caps instead of immediate global cutover.
Stable recovery matters as much as fast response. Keep mitigations active until latency and status signals stay normal across consecutive checks. Rapid rollback after a brief improvement often causes repeated user impact in real workloads.
Service is reachable and timing behavior is near baseline. Keep normal routing with standard monitoring.
Traffic succeeds but risk is higher. Reduce retries, add jitter, and protect critical user journeys.
Sustained failures are likely. Activate fallback and preserve core functionality first.
| Symptom | Likely Cause | Immediate Action |
|---|---|---|
| 429 spikes | Rate-limit or quota pressure | Backoff with jitter, smooth concurrency, validate limits |
| Timeout growth | Tail latency increase | Tighten timeout budget, shrink payloads, fallback critical paths |
| 5xx responses | Provider instability window | Use circuit breakers, cap retries, route canary failover |
| 401/403 auth errors | Credential or permission issue | Validate key scope, project mapping, environment settings |
Use it as one signal. Confirm with your own logs and provider updates before large routing changes.
Tail latency is where timeout risk appears first and where most end-user failures are felt.
Most production teams check every 60 to 120 seconds and alert on sustained change, not one-off spikes.
Reduce retry pressure, protect critical paths, and verify if errors are account-specific or broad.
No, but tested fallback policies materially reduce user impact and recovery time.
Yes. Region, request shape, account limits, and auth configuration can still cause app-level failures.
Treat status changes as triggers for controlled actions, not panic reroutes. Good response strategy protects user experience while avoiding unnecessary provider switching.
This sequence reduces retry amplification and lowers the risk of cascading failures across dependencies.
Historical context is where this page creates long-term value. Compare short-window disruptions against 30-day baselines to identify whether reliability risk is increasing, stable, or improving.
Turn each incident into one measurable runbook or threshold improvement to compound reliability gains over time.
Independent status with endpoint and latency detail for OpenAI APIs.
Independent monitor for Gemini uptime, latency, and incident behavior.
Fast diagnosis workflow for active OpenAI incidents.
Real-time outage triage view for Mistral integrations.
Historical incident and uptime perspective for planning and postmortems.
Estimate tradeoffs between resilience patterns and monthly spend.