24h Latency Trend
Latency trend is useful for seeing whether the service is stabilizing or worsening. A flat p50 with rising p95 usually indicates tail-risk growth that can hurt user-facing flows before broad outages appear.
AI Checker Hub
Gemini API Status Today is an independent operations page built for engineering teams that need a practical, data-backed view before routing traffic to Gemini endpoints. This page combines live status, rolling uptime, latency behavior, and incident windows so teams can decide quickly whether to continue normal routing, enable fallback, or tighten retry controls.
Latency trend is useful for seeing whether the service is stabilizing or worsening. A flat p50 with rising p95 usually indicates tail-risk growth that can hurt user-facing flows before broad outages appear.
Teams often overreact to temporary latency spikes and underreact to sustained degradation. This page is designed to prevent both mistakes. If state is operational but p95 is trending up for several intervals, you should proactively protect user paths by tightening timeout budgets and reducing aggressive retries. If degradation persists, failover should be gradual and policy-driven, not a sudden all-traffic switch.
A practical workflow is: validate status, inspect latency behavior, review incident window details, then apply the smallest safe mitigation first. For example, reduce retry count and add jitter before switching provider for every request. This avoids creating avoidable cost spikes while still protecting reliability.
Requests are passing and latency is close to baseline. Keep monitoring but no emergency action needed.
Service still works but error and latency risk are elevated. Reduce retry storms and enable controlled fallback.
High failure probability. Activate tested incident policy, shed non-critical load, and protect SLO paths first.
| Symptom | Likely Cause | Immediate Action |
|---|---|---|
| 429 spikes | Rate-limit or quota pressure | Apply backoff with jitter, smooth concurrency, verify quotas |
| Timeout growth | Tail latency surge | Tune timeout budgets, reduce payload size, fallback critical routes |
| Intermittent 5xx | Provider instability window | Use circuit breakers, cap retries, route canary to backup |
| Auth failures | Key/project config mismatch | Validate key scope, project permissions, environment variables |
Use it as one signal. Combine this monitor with application telemetry, user impact, and provider announcements.
p95 exposes tail behavior where most timeouts and user-visible failures occur during partial degradation.
Most teams use lightweight checks every 60–120 seconds and alert on sustained changes, not single spikes.
No, but a tested fallback policy significantly reduces incident impact and recovery time.
Reduce retry burst, protect critical paths, and confirm whether error rates are localized or broad.
Yes. Incident windows and rolling uptime help you understand if issues are transient or part of recurring patterns.
Reliability planning should use multiple time windows. The 24h view is best for immediate mitigation, while 7d and 30d trends reveal recurring risk patterns and help set realistic fallback thresholds.
This reduces overreaction to short spikes and improves reliability consistency during busy demand periods.
Most disruptions are not full outages; they are performance degradations that first appear in tail latency and then spread to timeout-heavy endpoints. Monitoring these early signals helps teams intervene before users see broad failures.
After each event, update one concrete control: retry cap, fallback trigger, queue policy, or alert threshold. Continuous incremental tuning produces better outcomes than infrequent major policy changes.
Independent status with endpoint and latency detail for OpenAI APIs.
Independent monitor for Anthropic uptime, latency, and incident behavior.
Fast diagnosis workflow for active OpenAI incidents.
Real-time outage triage view for Anthropic integrations.
Historical incident and uptime perspective for planning and postmortems.
Estimate tradeoffs between resilience patterns and monthly spend.