On July 9th, a small number of our customers experienced service degradation due to an issue within our cloud platform's database layer. A background process led to an unexpected spike in temporary storage usage which caused one of our database clusters to stop processing requests for a short period. Services were fully restored within 30 minutes.
A scheduled background process consumed temporary database storage at a rate that exceeded expected usage patterns. As a result, the TempDB volume reached full capacity, which prevented the database cluster from responding to incoming queries. This impacted platform responsiveness for affected customers during the incident window. Our internal monitoring systems had already flagged elevated resource usage, and an investigation was underway at the time. However, the issue escalated faster than anticipated, requiring manual intervention to clear the affected process and restore normal operations.
To prevent this kind of issue from recurring, we have taken the following actions:
We sincerely apologise for the inconvenience caused. Platform reliability is one of our highest priorities, and we are continuously working to strengthen the resilience of our systems. This incident has highlighted opportunities to improve early warning and automated protection, and we’re acting decisively to address them.
If you have any questions or need further clarification, please don’t hesitate to reach out to your account representative or our support team.