LLM Degradation Triggered by Widespread Internet Disruption – June 12, 2025
Resolved
What happened

On Thursday, June 12, 2025, at approximately 2 p.m. EDT, a massive internet outage erupted across Google Cloud’s infrastructure as well AWS, OpenAI, Anthropic, and more — leading to widespread disruptions in connectivity and service across the internet.

Impact on LLMs

As a downstream consequence of this outage, all major LLM-based services (ChatGPT, Anthropic, Character.AI, etc.) experienced significantly degraded performance. Users reported slow, inconsistent responses, increased error rates, and intermittent unavailability—behaviors commonly seen during external network or storage disruptions.

🔍 Timeline of Events
✅ What We Did
  • Real-time monitoring flagged elevated latency and 5xx errors from all LLM services.

  • Incident triaging confirmed the root cause was an external internet/cloud infrastructure failure—not LLM model degradation.

  • Fallback strategies (rate limiting, cached responses) were enacted where feasible until network stability returned later that afternoon.

Thu, Jun 12, 2025, 09:19 PM
(1 week ago)
·
Affected components

No components marked as affected

Updates

Resolved

What happened

On Thursday, June 12, 2025, at approximately 2 p.m. EDT, a massive internet outage erupted across Google Cloud’s infrastructure as well AWS, OpenAI, Anthropic, and more — leading to widespread disruptions in connectivity and service across the internet.

Impact on LLMs

As a downstream consequence of this outage, all major LLM-based services (ChatGPT, Anthropic, Character.AI, etc.) experienced significantly degraded performance. Users reported slow, inconsistent responses, increased error rates, and intermittent unavailability—behaviors commonly seen during external network or storage disruptions.

🔍 Timeline of Events
✅ What We Did
  • Real-time monitoring flagged elevated latency and 5xx errors from all LLM services.

  • Incident triaging confirmed the root cause was an external internet/cloud infrastructure failure—not LLM model degradation.

  • Fallback strategies (rate limiting, cached responses) were enacted where feasible until network stability returned later that afternoon.

Thu, Jun 12, 2025, 09:19 PM