Company Logo

Is there an InfluxDB outage?

InfluxDB status: Systems Active

Last checked: 9 minutes ago

Get notified about any outages, downtime or incidents for InfluxDB and 1800+ other cloud vendors. Monitor 10 companies, for free.

Subscribe for updates

InfluxDB outages and incidents

Outage and incident data over the last 30 days for InfluxDB.

There have been 1 outages or incidents for InfluxDB in the last 30 days.

Severity Breakdown:

Tired of searching for status updates?

Join OutLogger to be notified when any of your vendors or the components you use experience an outage. It's completely free and takes less than 2 minutes!

Sign Up Now

Components and Services Monitored for InfluxDB

Outlogger tracks the status of these components for Xero:

API Queries Active
API Writes Active
Compute Active
Other Active
Persistent Storage Active
Tasks Active
Web UI Active
API Reads Active
API Writes Active
Management API Active
API Queries Active
API Writes Active
Compute Active
Other Active
Persistent Storage Active
Tasks Active
Web UI Active
API Queries Active
API Writes Active
Compute Active
Other Active
Persistent Storage Active
Tasks Active
Web UI Active
API Queries Active
API Writes Active
Compute Active
Other Active
Persistent Storage Active
Tasks Active
Web UI Active
API Queries Active
API Writes Active
Compute Active
Other Active
Persistent Storage Active
Tasks Active
Web UI Active
API Queries Active
API Writes Active
Compute Active
Other Active
Persistent Storage Active
Tasks Active
Web UI Active
API Queries Active
API Writes Active
Compute Active
Other Active
Persistent Storage Active
Tasks Active
Web UI Active
API Queries Active
API Writes Active
Compute Active
Other Active
Persistent Storage Active
Tasks Active
Web UI Active
API Queries Active
API Writes Active
Compute Active
Other Active
Persistent Storage Active
Tasks Active
Web UI Active
Auth0 User Authentication Active
Marketplace integrations Active
Web UI Authentication (Auth0) Active
Component Status
Active
API Queries Active
API Writes Active
Compute Active
Other Active
Persistent Storage Active
Tasks Active
Web UI Active
Active
API Reads Active
API Writes Active
Management API Active
Active
API Queries Active
API Writes Active
Compute Active
Other Active
Persistent Storage Active
Tasks Active
Web UI Active
Active
API Queries Active
API Writes Active
Compute Active
Other Active
Persistent Storage Active
Tasks Active
Web UI Active
Active
API Queries Active
API Writes Active
Compute Active
Other Active
Persistent Storage Active
Tasks Active
Web UI Active
Active
API Queries Active
API Writes Active
Compute Active
Other Active
Persistent Storage Active
Tasks Active
Web UI Active
Active
API Queries Active
API Writes Active
Compute Active
Other Active
Persistent Storage Active
Tasks Active
Web UI Active
Active
API Queries Active
API Writes Active
Compute Active
Other Active
Persistent Storage Active
Tasks Active
Web UI Active
Active
API Queries Active
API Writes Active
Compute Active
Other Active
Persistent Storage Active
Tasks Active
Web UI Active
Active
API Queries Active
API Writes Active
Compute Active
Other Active
Persistent Storage Active
Tasks Active
Web UI Active
Active
Auth0 User Authentication Active
Marketplace integrations Active
Web UI Authentication (Auth0) Active

Latest InfluxDB outages and incidents.

View the latest incidents for InfluxDB and check for official updates:

Updates:

  • Time: May 12, 2023, 6:40 p.m.
    Status: Resolved
    Update: This incident has been resolved.
  • Time: May 12, 2023, 6:16 p.m.
    Status: Monitoring
    Update: We are continuing to monitor for any further issues.
  • Time: May 12, 2023, 12:39 p.m.
    Status: Monitoring
    Update: Read and write durations have recovered in the region, we are continuing to monitor the incident.
  • Time: May 12, 2023, 10:33 a.m.
    Status: Identified
    Update: Our team has identified the issue and continue to work on it.
  • Time: May 12, 2023, 9:08 a.m.
    Status: Investigating
    Update: Elevated read/write times in Azure West Europe and we are investigating the issue further.
  • Time: May 12, 2023, 9 a.m.
    Status: Investigating
    Update: We are currently investigating this issue.

Updates:

  • Time: May 12, 2023, 6:40 p.m.
    Status: Resolved
    Update: This incident has been resolved.
  • Time: May 12, 2023, 6:16 p.m.
    Status: Monitoring
    Update: We are continuing to monitor for any further issues.
  • Time: May 12, 2023, 12:39 p.m.
    Status: Monitoring
    Update: Read and write durations have recovered in the region, we are continuing to monitor the incident.
  • Time: May 12, 2023, 10:33 a.m.
    Status: Identified
    Update: Our team has identified the issue and continue to work on it.
  • Time: May 12, 2023, 9:08 a.m.
    Status: Investigating
    Update: Elevated read/write times in Azure West Europe and we are investigating the issue further.
  • Time: May 12, 2023, 9 a.m.
    Status: Investigating
    Update: We are currently investigating this issue.

Updates:

  • Time: May 16, 2023, 11:21 p.m.
    Status: Postmortem
    Update: # **Incident RCA** RCA - query errors in us-central-1 on May 11, 2023 ## **Summary** One of the customers in this cluster submitted a large number of deletes. By itself, this would not have caused an outage. However, at the same time, one of the storage pods ran out of disk space. We added more disk space, but due to the large number of tombstone files \(created to keep track of deleted measurements\) the pod was very slow to recover, and there was a high rate of query failures until the storage pod recovered. ## **Cause of the Incident** The immediate root cause was that a disk filled up. Under normal circumstances, this is not service-impacting. We get alerted when the disk is close to filling up, and we have a run book in place to add capacity to the storage layer. The pod must be restarted after adjusting the disk size, and when the pod restarted, it was unavailable for a long time, while processing all the tombstone files. ## **Recovery** As soon as we identified that deletes were contributing to the slow recovery, we reached out to the customer that had generated the large volume of deletes, to ask them to stop sending requests to delete measurements, while the cluster was in recovery mode. While we waited to hear back from them, we blocked deletes for all customers on this cluster as a temporary measure. We also manually removed the tombstone files from one replica of the most heavily impacted storage partition, so that it could recover more quickly. This enabled the cluster to return to normal operation and process queries. Meanwhile the other replica of this partition continued to process the backlog of tombstone files so that once it was complete we could restart both replicas and the data will be complete and correct. ## **Timeline** May 11, 2023 18:20 - Alerted that storage pod disk was close to capacity May 11, 2023 18:43 - Added more disk capacity, and restarted the pods. May 11, 2023 19:05 - Pod was very slow to start, so queries started failing. Storage pod became unavailable and queries started failing. Investigations showed that the problem partition was pegged processing the enormous number of deletes, but that it was making progress. We decided to let the process run its course so that the data was correct. Continuous monitoring of the progress gave a predicted outage time of around two hours. During this time writes were being accepted but queries were failing. This was seen as the least bad option. The partition completed its work but then, because of the number of delete requests that had continued to be added, effectively had to start over. May 11, 2023 21:16 - Blocked all deletes to the cluster and continued to monitor the situation. Progress was still being made. We examined the code paths and decided that this was not a software fault as such and so the best course of action was to allow the process to continue to run. We started investigating ideas to restore services faster. May 11, 2023 22:00 - Manually deleted tombstone files on the secondary replica of the impacted storage pods to speed up the recovery process. This means that queries were able to be serviced but that data which should have been deleted was visible. This was decided to be the least bad option at this time. We intended to leave the primary replica to process the remaining deletes expecting this to take many more hours. When this process was complete we would restart the secondary and the replicas would be back in sync with the correct data. May 11, 2023 23:08 - Primary replica of impacted storage pod recovered, query rate back to normal. ## **Future mitigations** 1. We are reducing the rate limit for deletes, to reduce the load on the cluster caused by deletions. 2. We continue to work with customers to find alternative approaches that reduce the need for them to delete measurements.
  • Time: May 16, 2023, 11:20 p.m.
    Status: Resolved
    Update: The incident has been resolved. We are still blocking deletes.
  • Time: May 16, 2023, 11:19 p.m.
    Status: Monitoring
    Update: We are continuing to monitor for any further issues.
  • Time: May 12, 2023, 12:31 a.m.
    Status: Monitoring
    Update: Writes and queries are working. Deletes are temporarily disabled.
  • Time: May 12, 2023, 12:31 a.m.
    Status: Monitoring
    Update: Writes and queries are working. Deletes are temporarily disabled.
  • Time: May 11, 2023, 11:59 p.m.
    Status: Investigating
    Update: We are continuing to investigate this issue.
  • Time: May 11, 2023, 11:59 p.m.
    Status: Investigating
    Update: We are continuing to investigate this issue.
  • Time: May 11, 2023, 11:58 p.m.
    Status: Investigating
    Update: Reads-Writes Issues in US-central-1 are resolved. Deletes are still disabled.
  • Time: May 11, 2023, 11:58 p.m.
    Status: Investigating
    Update: Reads-Writes Issues in US-central-1 are resolved. Deletes are still disabled.
  • Time: May 11, 2023, 9:28 p.m.
    Status: Investigating
    Update: We are continuing to investigate this issue.
  • Time: May 11, 2023, 9:28 p.m.
    Status: Investigating
    Update: We are continuing to investigate this issue.
  • Time: May 11, 2023, 9:27 p.m.
    Status: Investigating
    Update: Queries currently unavailable in GCP us-central-1. Also, deletes are currently disabled.
  • Time: May 11, 2023, 9:27 p.m.
    Status: Investigating
    Update: Queries currently unavailable in GCP us-central-1. Also, deletes are currently disabled.
  • Time: May 11, 2023, 7:44 p.m.
    Status: Investigating
    Update: We are investigating a potential issue in GCP us-central-1.
  • Time: May 11, 2023, 7:44 p.m.
    Status: Investigating
    Update: We are investigating a potential issue in GCP us-central-1.
  • Time: May 11, 2023, 7:42 p.m.
    Status: Investigating
    Update: We know about it and we're investigating.
  • Time: May 11, 2023, 7:42 p.m.
    Status: Investigating
    Update: We know about it and we're investigating.

Updates:

  • Time: April 13, 2023, 6:09 p.m.
    Status: Resolved
    Update: This incident has been resolved.
  • Time: April 13, 2023, 5:30 p.m.
    Status: Monitoring
    Update: We are continuing to monitor for any further issues.
  • Time: April 13, 2023, 5:03 p.m.
    Status: Monitoring
    Update: A fix has been implemented and we are monitoring the results.
  • Time: April 13, 2023, 4:54 p.m.
    Status: Identified
    Update: The issue has been identified and a fix is being implemented.
  • Time: April 13, 2023, 4:37 p.m.
    Status: Investigating
    Update: prod01-us-west-2 AWS isn't accepting writes or queries. We are currently Investigating
  • Time: April 13, 2023, 4:37 p.m.
    Status: Investigating
    Update: prod01-us-west-2 AWS isn't accepting writes or queries. We are currently Investigating

Updates:

  • Time: April 5, 2023, 3:52 p.m.
    Status: Resolved
    Update: This incident has been resolved.
  • Time: April 5, 2023, 2:26 p.m.
    Status: Monitoring
    Update: Query error rates now have returned to normal and we are continuing to monitor.
  • Time: April 5, 2023, 2:19 p.m.
    Status: Investigating
    Update: We are currently investigating this issue.

Check the status of similar companies and alternatives to InfluxDB

Smartsheet
Smartsheet

Systems Active

ESS (Public)
ESS (Public)

Systems Active

ESS (Public)
ESS (Public)

Systems Active

Cloudera
Cloudera

Systems Active

New Relic
New Relic

Systems Active

Boomi
Boomi

Systems Active

AppsFlyer
AppsFlyer

Systems Active

Imperva
Imperva

Systems Active

Bazaarvoice
Bazaarvoice

Issues Detected

Optimizely
Optimizely

Systems Active

Electric
Electric

Systems Active

ABBYY
ABBYY

Systems Active

Frequently Asked Questions - InfluxDB

Is there a InfluxDB outage?
The current status of InfluxDB is: Systems Active
Where can I find the official status page of InfluxDB?
The official status page for InfluxDB is here
How can I get notified if InfluxDB is down or experiencing an outage?
To get notified of any status changes to InfluxDB, simply sign up to OutLogger's free monitoring service. OutLogger checks the official status of InfluxDB every few minutes and will notify you of any changes. You can veiw the status of all your cloud vendors in one dashboard. Sign up here
What does InfluxDB do?
Efficiently store and access time series data in a specialized database designed for speed, available in cloud, on-premises, or edge environments.