Update: We are still closely monitoring our infrastructure. Service has been restored and indexing delay is gone. Currently, filtering is still not being updated. We should be resolving the filter issue fairly shortly.
Posted 6 days ago. Sep 12, 2019 - 17:54 UTC
Our infra team has implemented a fix to restore service. Live tail, indexing, and alerting delays are resolved. We are closely monitoring the state of the infrastructure at this time to prevent any further delays. Currently, filters within the top menu might be delayed.
Posted 7 days ago. Sep 11, 2019 - 06:42 UTC
Unexpected traffic along with scaling issues to handle the increased traffic caused delays with indexing, alerting and live tail. The start time was around 17:00 UTC Sept 9th, 2019. All customers are experiencing the above-said delays. Our infrastructure team and all other stakeholders are working diligently to get the scaling issue resolved as soon as possible and the application back to being fully operational.
The issue has been identified and this status will be updated soon with more information as our teams work on the incident.
Posted 8 days ago. Sep 11, 2019 - 00:53 UTC
We are continuing to work through the ingestion issues. It is likely that you are not getting alert notifications at this time.
Posted 8 days ago. Sep 10, 2019 - 22:08 UTC
The issue has been identified. Indexing and live tail look to be slightly delayed, alerting is still delayed. We will set to monitoring as soon as the delay has disappeared.
Posted 8 days ago. Sep 10, 2019 - 14:41 UTC
We are currently experiencing a delay in ingestion including livetail, alerting, and indexing. Logs are being ingested but could be delayed 1 hour or more. Our infra team is actively working on this issue.
Posted 8 days ago. Sep 10, 2019 - 05:50 UTC
This incident affected: Log Ingestion (Agent/REST API/Code Libraries), Log Ingestion (Heroku), and Log Ingestion (Syslog).