Trello Status
Operational
Last incident: 2/21/2026
Current Status
Overall StatusOperational
Last IncidentTrello Slow or down for users
Incident Statusresolved
Recent Incidents
Trello Slow or down for users
2/21/2026, 3:32:17 PM
Impact
Trello users are experiencing issues with websocket connections, leading to difficulties in maintaining active sessions. Affected users may notice disruptions in service as connections are intermittently dropping.
Current Status
Efforts are underway to restore normal service. Teams are actively engaged in identifying the best path to full resolution.
Next Steps
The incident team is focusing on pinpointing the root cause and implementing necessary fixes. The next communication will be shared in 30 minutes.
Affected Components:
Trello.com
API
Atlassian Support - Support Portal
Atlassian Support Ticketing
Atlassian Support Knowledge Base
Degraded performance of Trello
2/19/2026, 3:06:25 PM
On February 19, 2026, Trello users may have experienced performance degradation. The issue has now been resolved, and the service is operating normally for all affected customers.
Affected Components:
Trello.com
API
Atlassian Support - Support Portal
Atlassian Support Ticketing
Atlassian Support Knowledge Base
Trello performance is degraded
2/9/2026, 8:20:54 AM
### **Summary**
On Feb 9, 2026, between 07:12 UTC and 08:05 UTC, Atlassian customers were unable to access Trello. The event was triggered by Trello servers reaching maximum memory limits and a subsequent failure to automatically scale for the traffic. The incident was detected within 5 minutes by automated monitoring systems, engaging Trello teams for resolution. Two parallel mitigation efforts were undertaken to restore access: 1\) manual scaling to address concerns with high utilization of hosts, and 2\) temporarily limiting traffic from free customers. This intervention put Trello systems in a known good state with a total time to resolution of ~53 minutes.
### **IMPACT**
The overall impact occurred on Feb 9, 2026 between 07:12 AM UTC and 08:05 AM UTC for Trello customers and caused service disruption for all customers, making Trello inaccessible during that time. Access was restored for paid users approximately 43 minutes after the onset, with full service restoration for free users 10 minutes later.
### **ROOT CAUSE**
The issue was cause by increased memory requirements during the transition from the weekend to EU business hours, combined with Auto Scaling group scaling rules based on CPU utilization. Memory allocation on our instances outpaced the CPU usage, which caused processes to hit Out-Of-Memory \(OOM\) errors and crash before reaching CPU usage thresholds that would trigger the auto scaling policies. As a result, Trello went down and users received HTTP 502 errors until the incident was resolved.
### **REMEDIAL ACTION PLAN & NEXT STEPS**
We know that outages impact your productivity. While we have a suite of automated infrastructure management in place, this specific issue required manual intervention to restore Trello access to users. To avoid repeating this type of incident, we are prioritizing the following remedial action items:
* Pre-scale capacity before EU morning traffic - This change has already been introduced to prevent further incidents while we implement additional safeguards.
* Adjust Trello’s Auto Scaling group settings - This change will ensure that unhealthy hosts will be replaced more rapidly and scaling policies will consider memory usage.
* Refine Trello host memory commitments - This change will decrease the likelihood of memory overcommitment on hosts and associated OOM errors.
* Increase isolation for OOM errors - This change will improve the ability for hosts to recover in the event of a single worker experiencing an OOM error.
We apologize to customers whose services were impacted during this incident; we are taking immediate steps to improve the platform’s performance and availability.
Thanks,
Atlassian Customer Support
Affected Components:
Trello.com
API
Atlassian Support - Support Portal
Atlassian Support Ticketing
Atlassian Support Knowledge Base
Outbound Email, Mobile Push Notifications, and Support Ticket Delivery Impacting All Cloud Products
12/27/2025, 4:42:32 AM
### Summary
On **December 27, 2025**, between **02:48 UTC and 05:20 UTC**, some Atlassian cloud customers experienced failures in sending and receiving emails and mobile notifications. Core Jira and Confluence functionality remained available.
The issue was triggered when **TLS certificates used by Atlassian’s monitoring infrastructure expired**, causing parts of our metrics pipeline to stop accepting traffic. Services responsible for email and mobile notifications had a critical path dependency on monitoring path leading to service disruptions.
All impacted services were fully restored by **05:20 UTC**, around **2.5 hours** after customer impact began.
### IMPACT
During the impact window, customers experienced:
* **Outbound product email failures** \(notifications and other product emails did not send\).
* **Identity and account flow failures** where emails were required \(e.g. sign‑ups, password resets, one‑time‑password / step‑up challenges\).
* **Jira and Confluence mobile push notifications**
* **Customer site activations and some admin policy changes** failing and requiring later reprocessing.
### ROOT CAUSE
The incident was caused by:
1. **Expired TLS certificates** on domains used by our monitoring and metrics infrastructure caused by **misconfigured DNS authorization record** which prevented automatic renewal.
2. **Tight coupling of services to metrics publishing**, which caused them to fail when monitoring endpoints became unavailable, instead of degrading gracefully.
### REMEDIAL ACTIONS PLAN & NEXT STEPS
We recognize that outages like this have a direct impact on customers’ ability to receive important notifications, complete account tasks, and operate their sites.
We are prioritizing the following actions to improve our existing testing, monitoring and certificate management processes:
* **Hardening monitoring and certificate infrastructure**
* We are refining DNS and certificate configuration across our monitoring domains and strengthening proactive checks to detect and address failed renewals and certificate issues well before expiry.
* We are also improving alerting on our monitoring and metrics pipeline.
* **Decoupling monitoring from critical customer flows**
We are updating services such as outbound email, identity, mobile push, provisioning, and admin policy changes so they no longer depend on metrics publishing to operate. If monitoring becomes unavailable, these services will continue to run and degrade gracefully by dropping or buffering metrics instead of failing customer operations.
We apologize to customers impacted during this incident. We are implementing the improvements above to help ensure that similar issues are avoided.
Thanks,
Atlassian Customer Support
Affected Components:
Atlassian Support Ticketing
500 Errors being experienced by Trello users
12/12/2025, 6:40:21 AM
### Summary
On December 12, 2025, between 05:31 UTC and 07:11 UTC, Trello customers experienced errors or slow loading times due to a dependency library upgrade that caused high resource usage. The incident was detected within 1 minute by the automated monitoring system and mitigated by rolling back the version of the dependency library which put Atlassian systems into a known good state. The total time to resolution was 1 hour and 41 minutes.
### **IMPACT**
The issue caused service disruption for all Trello customers on December 12, 2025 between 05:31 UTC and 07:11 UTC. Users experienced degraded service between 05:31 UTC and 06:32 UTC, and most users were unable to load Trello between 06:32 UTC to 07:07 UTC.
### **ROOT CAUSE**
The issue was caused by an upgrade to a dependency library. As a result, there was an increase in system memory consumption causing startups to take longer than expected. When the startup time exceeded the allowed threshold, the worker process was flagged as unresponsive and restarted as an automated recovery action. Because every worker on the instance failed to startup within the expected time, the instance entered a startup loop that required intervention.
### **REMEDIAL ACTIONS PLAN & NEXT STEPS**
We know that outages impact your productivity. While we have several testing and preventive processes in place, we didn’t anticipate that the upgrade would consume more memory, which ultimately caused the startup failure loop and service disruption.
We are prioritizing the following improvement actions designed to avoid this issue in the future:
* Review our automated detection of unresponsive workers and adjust the detection thresholds during initialization.
* Improve monitoring during workers startup and restarts due to unresponsiveness.
* Enhance the stricter standard process for confirming memory/other data processing resource requirements for planned updates.
We apologize to customers whose services were impacted during this incident; we are taking immediate steps to improve the platform’s performance and availability.
Thanks,
Atlassian Customer Support
Affected Components:
Trello.com
API
Atlassian Support - Support Portal
Atlassian Support Ticketing
Atlassian Support Knowledge Base