On Feb 9, 2026, between 07:12 UTC and 08:05 UTC, Atlassian customers were unable to access Trello. The event was triggered by Trello servers reaching maximum memory limits and a subsequent failure to automatically scale for the traffic. The incident was detected within 5 minutes by automated monitoring systems, engaging Trello teams for resolution. Two parallel mitigation efforts were undertaken to restore access: 1) manual scaling to address concerns with high utilization of hosts, and 2) temporarily limiting traffic from free customers. This intervention put Trello systems in a known good state with a total time to resolution of ~53 minutes.
The overall impact occurred on Feb 9, 2026 between 07:12 AM UTC and 08:05 AM UTC for Trello customers and caused service disruption for all customers, making Trello inaccessible during that time. Access was restored for paid users approximately 43 minutes after the onset, with full service restoration for free users 10 minutes later.
The issue was cause by increased memory requirements during the transition from the weekend to EU business hours, combined with Auto Scaling group scaling rules based on CPU utilization. Memory allocation on our instances outpaced the CPU usage, which caused processes to hit Out-Of-Memory (OOM) errors and crash before reaching CPU usage thresholds that would trigger the auto scaling policies. As a result, Trello went down and users received HTTP 502 errors until the incident was resolved.
We know that outages impact your productivity. While we have a suite of automated infrastructure management in place, this specific issue required manual intervention to restore Trello access to users. To avoid repeating this type of incident, we are prioritizing the following remedial action items:
We apologize to customers whose services were impacted during this incident; we are taking immediate steps to improve the platform’s performance and availability.
Thanks,
Atlassian Customer Support