Observe Outage History

Uptime record, past incidents, and downtime history for Observe.

Checking current status...
95.1% uptime over 61 days
99.9% ✗ 99.5% ✗ 99% ✗ 95% ✓

90-Day Trend

Feb 9Apr 10

Monthly Uptime

Month Uptime Days Tracked Days with Issues
April 2026 100% 10 0
March 2026 90.3% 31 3
February 2026 100% 20 0

Uptime is calculated from daily worst-status snapshots. A day with any non-operational status counts as a day with issues.

Daily Status (Last 61 Days)

Feb 9 Today
Operational Degraded Partial Outage Major Outage Maintenance No Data

Incident History

March 2026
2026-03-25 High Rate of Ingest Errors (US West)

Started: Mar 26, 1:25 AM

monitoring
A fix has been implemented and we are monitoring the results.
Mar 26, 1:28 AM
investigating
We’ve identified an issue causing elevated errors for ingest endpoints in US West region. As a result, data ingest will be affected. Our team is actively investigating the root cause and working to restore normal performance as quickly as possible. We’ll provide updates as we learn more. Thank you for your patience!
Mar 26, 1:25 AM
2026-03-20 Intermittent ingest errors and query instability
critical

Started: Mar 21, 6:21 AM

monitoring
A fix has been applied and the system is recovering.
Mar 21, 6:33 AM
identified
The issue has been identified by Snowflake engineering teams and a mitigation is being applied.
Mar 21, 6:23 AM
investigating
We’ve identified an issue causing data ingest to have higher than normal errors and queries to be unreliable in the following regions: US-WEST-2. As a result, some users may experience higher than normal ingest errors and slow queries. We do not anticipate any data loss. This incident is related to an ongoing Snowflake incident - https://status.snowflake.com/. Our team is actively investigating the root cause and working to restore normal performance as quickly as possible. We’ll provi...
Mar 21, 6:21 AM
Unable to access Observe Tenant

Started: Mar 19, 12:20 AM

monitoring
We are aware that some users may be unable to log into Observe. The issue has been mitigated and we're monitoring.
Mar 19, 12:20 AM
Ingest performance degredation in Prod EU cluster
minor

Started: Mar 16, 6:55 PM

monitoring
A large increase in tracing data caused ingestion lag on the prod-eu-1 cluster, affecting multiple customers. Encoder replicas were scaled up and subsequently rolled to resolve unassigned partition workers. Memory pressure was identified as a contributing factor. Ingestion lag peaked at up to 45 minutes for some customers and has since recovered to near-normal levels. The team is continuing to monitor.
Mar 16, 6:55 PM
Service Degradation
minor

Started: Mar 16, 6:44 PM

investigating
A large increase in tracing data caused ingestion lag on the prod-eu-1 cluster. Encoder replicas were scaled up, but some partition workers were not reading data due to unassigned partitions. Encoders were rolled, and partition lag is now decreasing. The team is continuing to monitor recovery.
Mar 16, 6:44 PM
Performance Degradation in GCP
minor

Started: Mar 12, 2:51 PM

monitoring
The issue with elevated warehouse resume times in GCP has been mitigated. Query queueing times have returned to normal levels and resume times have stabilized. A case was raised with the infrastructure provider and warehouse resource management was adjusted to reduce customer impact during the incident. We are continuing to monitor to confirm the issue is fully resolved.
Mar 12, 5:15 PM
monitoring
We are continuing to monitor for any further issues.
Mar 12, 2:52 PM
monitoring
Our internal team identified an issue with warehouse resume times in Prod GCP - causing p99 queueing times of up to 2m (including user queries). The issue is believed to be addressed and we are monitoring on our side.
Mar 12, 2:51 PM