Last checked: a minute ago
Get notified about any outages, downtime or incidents for Asana and 1800+ other cloud vendors. Monitor 10 companies, for free.
Outage and incident data over the last 30 days for Asana.
Join OutLogger to be notified when any of your vendors or the components you use experience an outage. It's completely free and takes less than 2 minutes!
Sign Up NowOutlogger tracks the status of these components for Xero:
Component | Status |
---|---|
Australia | Active |
API | Active |
App | Active |
Automations and Background Actions | Active |
Mobile | Active |
Webhooks and Event Streams | Active |
EU | Active |
API | Active |
App | Active |
Automations and Background Actions | Active |
Mobile | Active |
Webhooks and Event Streams | Active |
Japan | Active |
API | Active |
App | Active |
Automations and Background Actions | Active |
Mobile | Active |
Webhooks and Event Streams | Active |
US | Active |
API | Active |
App | Active |
Automations and Background Actions | Active |
Mobile | Active |
Webhooks and Event Streams | Active |
View the latest incidents for Asana and check for official updates:
Description: Incident: From 2023-03-15 14:38 UTC to 2023-03-15 15:18 UTC there was a webhook outage for about 40 minutes in the US region. A bad deployment caused reduced capacity and delays in scheduling some asynchronous work, including webhook delivery. Impact: Webhook users experienced a considerable amount of delay, max ~one hour difference between events generated and events delivered. There was no data loss due to this delay and backlog was recovered in ~20 minutes Moving forward: We have added better tooling that would help us investigate this issue faster and have increased the priority of the job responsible for streaming webhook events.
Status: Resolved
Impact: None | Started At: March 15, 2023, 3 p.m.
Description: Incident: From 2023-03-15 14:38 UTC to 2023-03-15 15:18 UTC there was a webhook outage for about 40 minutes in the US region. A bad deployment caused reduced capacity and delays in scheduling some asynchronous work, including webhook delivery. Impact: Webhook users experienced a considerable amount of delay, max ~one hour difference between events generated and events delivered. There was no data loss due to this delay and backlog was recovered in ~20 minutes Moving forward: We have added better tooling that would help us investigate this issue faster and have increased the priority of the job responsible for streaming webhook events.
Status: Resolved
Impact: None | Started At: March 15, 2023, 3 p.m.
Description: **Incident**: Around 2023-02-28 20:31 UTC servers responsible for reactivity started to fail due to unexpected input. After reverting the change triggering this, most recovered between 21:30 and 22:00. In some cases stale data was displayed until caches were cleared, which finished at 23:05. Approximately 1% of users continued to see reactivity failures and stale data until around 2023-03-01 00:04 UTC. **Impact**: While reactivity servers were down, API writes failed and changes were not reflected to other tabs. After recovery of reactivity servers, in some cases stale data was displayed within our applications until the caches were fully cleared. No customer data was lost. **Moving forward**: We are making changes to the application servers which crashed to make them more resilient against unexpected input, and making tooling changes to reduce time to resolution for this class of incident. Architectural changes which are in progress will provide smaller failure domains, which would reduce impact and provide faster resolution for this class of failure. We use the [5 Whys](https://wavelength.asana.com/workstyle-ask-5-whys-to-get-to-the-root-of-any-problem/) approach to identify technical, operational, and organizational changes to reduce the likelihood and severity of incidents. Our metric considers a weighted average of uptime experienced by users at each data center. The number of minutes of downtime shown reflects this weighted average.
Status: Postmortem
Impact: Minor | Started At: Feb. 28, 2023, 9:32 p.m.
Description: **Incident**: Around 2023-02-28 20:31 UTC servers responsible for reactivity started to fail due to unexpected input. After reverting the change triggering this, most recovered between 21:30 and 22:00. In some cases stale data was displayed until caches were cleared, which finished at 23:05. Approximately 1% of users continued to see reactivity failures and stale data until around 2023-03-01 00:04 UTC. **Impact**: While reactivity servers were down, API writes failed and changes were not reflected to other tabs. After recovery of reactivity servers, in some cases stale data was displayed within our applications until the caches were fully cleared. No customer data was lost. **Moving forward**: We are making changes to the application servers which crashed to make them more resilient against unexpected input, and making tooling changes to reduce time to resolution for this class of incident. Architectural changes which are in progress will provide smaller failure domains, which would reduce impact and provide faster resolution for this class of failure. We use the [5 Whys](https://wavelength.asana.com/workstyle-ask-5-whys-to-get-to-the-root-of-any-problem/) approach to identify technical, operational, and organizational changes to reduce the likelihood and severity of incidents. Our metric considers a weighted average of uptime experienced by users at each data center. The number of minutes of downtime shown reflects this weighted average.
Status: Postmortem
Impact: Minor | Started At: Feb. 28, 2023, 9:32 p.m.
Description: **Incident**: Around 15:50 UTC, a system used for caching entered an overload state due to network connection tracking within our hosting environment. Connections from servers to this cache timed out, and request queuing created a cascading failure. Reducing load via application changes allowed recovery, and we then made configuration changes to provide additional capacity. **Impact**: Asana was unavailable in all regions for as long as 16 minutes, with a partial outage for 6 additional minutes. **Moving forward**: We've updated configuration to avoid overload, and are adding monitoring to detect this type of saturation before failure. Our metric considers a weighted average of uptime experienced by users at each data center. The number of minutes of downtime shown reflects this weighted average.
Status: Postmortem
Impact: Critical | Started At: Feb. 6, 2023, 4 p.m.
Join OutLogger to be notified when any of your vendors or the components you use experience an outage or down time. Join for free - no credit card required.