Service Degradation
Incident Lifecycle
Incident Timeline
Investigating
A large increase in tracing data caused ingestion lag on the prod-eu-1 cluster. Encoder replicas were scaled up, but some partition workers were not reading data due to unassigned partitions. Encoders were rolled, and partition lag is now decreasing. The team is continuing to monitor recovery.
Mar 16, 2026 at 6:44 PM UTC
Was your business affected by this Observe outage?
Set up instant alerts for Observe, be the first to know about outages via email, Slack, Teams, or Discord.