AI Checker Hub
Status History by Provider
Use live filters to compare provider reliability over 24h, 7d, and 30d windows. This page is designed for teams that need fast trend context before incident response, fallback policy changes, or routing updates.
Provider Snapshot
Top 3 Most Stable Providers (Last 30 Days)
Ranked by highest 30d uptime. Use this as directional planning context, not as the only routing signal.
Window Comparison Table
| Window | Uptime | Interpretation |
|---|
Recent Incident Windows
Decision Examples
If 24h uptime dips but 30d uptime remains stable, treat it as a likely short disruption window. Tighten retries and monitor recovery before major routing changes.
If p95 latency jumps while uptime stays high, user experience can still degrade. Prioritize timeout tuning and selective fallback for interactive paths even if hard failures are limited.
How This Differs From Official Status Pages
Official pages communicate provider-reported events. This page provides independent monitor-based comparison views and consistent cross-provider filters. Use both together for stronger incident decisions.
FAQ
Should I optimize for uptime only?
No. Uptime without latency context can hide real user pain in interactive products.
When should I trigger fallback based on this page?
Use consecutive threshold breaches, not single spikes, and apply traffic caps during transition.
Can rankings change quickly?
Yes. Short windows can shift quickly during incidents; use 7d/30d for policy planning.
Why compare 24h, 7d, and 30d together?
It helps separate short-term noise from persistent reliability trends.
Is this enough for production decisions alone?
No. Combine with application telemetry, user impact, and official provider updates.
How Teams Use This Page in Real Operations
Reliability teams typically use this view in two loops: a rapid response loop during active incidents and a planning loop for weekly threshold tuning. The rapid loop focuses on 24h behavior and live symptom trends. The planning loop uses 7d/30d windows to improve retry rules, fallback thresholds, and provider mix.
Rapid Response Loop
- Check the selected provider snapshot and current state label.
- Inspect p95 movement and incident windows before adjusting traffic.
- Apply small, controlled routing changes instead of all-at-once failover.
Weekly Planning Loop
- Compare 24h vs 30d reliability gap for each critical provider.
- Update one alert threshold and one fallback threshold per cycle.
- Record decisions and link them to incident outcomes for accountability.
Interpreting Conflicting Signals
Conflicting metrics are normal. A provider can show high uptime while users still experience slowness due to elevated tail latency. Another provider might show a brief outage but faster recovery and lower long-term volatility.
- High uptime + high p95: prioritize user-experience routing and timeout controls.
- Low uptime + normal p95: investigate intermittent hard failures and transport stability.
- Stable 30d + noisy 24h: treat as transient unless repeated over multiple days.
- Noisy 30d trend: reassess provider weighting and fallback trigger thresholds.
This page is designed to support decisions, not automate them. The strongest approach is combining these comparisons with your service-level objectives and customer-impact telemetry.
Change Management: Turning Insight Into Safer Releases
Status comparisons are most useful when tied to release governance. Before major launches, teams should review provider trends and decide whether to increase fallback readiness, reduce burst risk, or adjust timeout budgets.
Pre-Release Reliability Checklist
- Validate 24h and 7d trend for your primary provider and backup provider.
- Confirm fallback path supports required model features and context sizes.
- Set temporary alert sensitivity higher during launch windows.
- Define rollback and traffic-cap rules before enabling new workloads.
Teams that include this checklist in release reviews usually detect risk earlier and avoid emergency routing decisions under load.
Status Cluster Links
OpenAI API Status
Independent OpenAI uptime, latency, and incidents.
Is OpenAI Down?
Fast outage diagnosis and immediate actions.
Anthropic API Status
Live Anthropic reliability and trend view.
Is Anthropic Down?
Real-time triage workflow for Anthropic incidents.
Gemini API Status
Independent status and latency trend for Gemini APIs.
Provider Reliability Comparison
Cross-provider methodology and benchmark context.