AWS is working to resolve a power outage affecting the use1-az4 availability zone in its problematic US-EAST-1 region, causing impairments to EC2 instances, EBS volumes, and dependent services.
Amazon Web Services is addressing a significant power outage in the US-EAST-1 region that has created service impairments for customers, continuing a pattern of reliability issues for this particular AWS region.
According to an incident report published on May 7, AWS identified problems in the use1-az4 availability zone of the US-EAST-1 Region. The company stated that "EC2 instances and EBS volumes hosted on impacted hardware are affected by the loss of power during the thermal event." This thermal event has led to cooling issues that AWS is working to resolve, though progress has been slower than originally anticipated.
"We continue to work towards mitigating the increased temperatures to its normal levels," AWS reported in an update, while warning that "Other AWS services that depend on the affected EC2 instances and EBS volumes in this Availability Zone may also experience impairments." By 8:06 PM PDT, Amazon acknowledged making "incremental progress to restore cooling systems" but confirmed that users are "experiencing elevated error rates and latencies for some workflows."
In response to the outage, AWS has shifted traffic away from the stricken availability zone and is advising customers to move their workloads to other availability zones within the same region. However, the company has also noted that customers may experience "longer than usual provisioning times" when attempting this redistribution.
The US-EAST-1 region has developed a reputation as AWS's "problem child," having been the site of major outages that significantly impacted internet services in 2021 and again in October 2025. Despite this history, AWS executives have maintained that the region isn't inherently more fragile than other parts of Amazon's cloud infrastructure. Instead, they explain that US-EAST-1 often operates at larger scales than other regions, placing additional stress on services and potentially making any disruptions more noticeable.
For customers relying on services in the affected availability zone, this incident highlights the importance of implementing proper multi-AZ deployment strategies. AWS's own best practices recommend distributing workloads across multiple availability zones to achieve high availability and fault tolerance. The current situation serves as a real-world example of why these recommendations matter.
The incident also raises questions about the physical infrastructure supporting AWS's cloud services. While cloud providers typically emphasize the redundancy and resilience of their data centers, events like this demonstrate that even major providers can experience unexpected hardware failures that impact service availability.
Customers experiencing issues can check the AWS Service Health Dashboard for real-time updates on service restoration. For guidance on designing resilient applications that can withstand availability zone failures, AWS provides documentation on architecting for high availability.
As AWS continues to work on restoring normal operations in the affected availability zone, this incident adds another chapter to the ongoing story of US-EAST-1's reliability challenges. For organizations with critical workloads in this region, the outage may prompt renewed evaluation of their disaster recovery strategies and consideration of deploying additional resources in alternative regions.

Comments
Please log in or register to join the discussion