GUIDEcx - Notice history

All systems operational

100% - uptime

Workato Website - Operational

Workato Email notifications - Operational

Workbot for Teams - Operational

Workbot for Slack - Operational

Recipe runtime for job execution - Operational

Recipe Webhook ingestion - Operational

Recipe API gateway - Operational

Notice history

Feb 2026

No notices reported this month

Jan 2026

Latency Issues with 1.0 Platform
  • Resolved
    Resolved

    RCA – Root Cause Analysis: Latency Issues with 1.0 Platform
    Status: Resolved Date: January 27, 2026
    Incident Window: 9:06 AM – 10:15 AM MST
    Summary
    Between 9:06 AM and 10:15 AM MST, the 1.0 platform experienced elevated latency and intermittent API timeouts. The root cause was identified as storage exhaustion within the Redis cache cluster. This saturation forced the platform to fall back to primary database queries, leading to system congestion and slower response times.
    Timeline

    • 9:06 AM: Incident start; Redis storage reaches capacity.

    • 9:26 AM: System alerts received by IT and Operations teams.

    • 9:33 AM: Formal incident ticket submitted and investigation intensified.

    • 9:50 AM: Root cause confirmed as Redis cluster saturation.

    • 9:55 AM: Cluster upgrade initiated to increase storage capacity.

    • 10:06 AM: Upgrade complete; cache layer fully operational.

    • 10:15 AM: Resolution: API latency returned to baseline.

    Resolution & Prevention
    To resolve the incident, the Redis cluster was vertically scaled to increase available storage. To prevent a recurrence, we are recalibrating our monitoring and alerting thresholds. These updates will ensure that alerts are triggered more appropriately and earlier in the utilization cycle, allowing our team to intervene before the system reaches saturation.

  • Monitoring
    Monitoring

    We have identified the issue and it has been resolved. We'll continue to monitor the issue.

  • Investigating
    Investigating

    We’re aware of an issue affecting the legacy (1.0) platform and are actively investigating. The team is fully engaged and working to resolve this as quickly as possible. We’ll share updates as they become available.

Dec 2025

General latency when navigating application
  • Postmortem
    Postmortem

    Root Cause Analysis – Redis Cache Storage Exhaustion
    Date of incident: 2025-12-01
    Incident window: ~1:10 PM–1:45 PM MST
    Impact: Elevated latency and intermittent API timeouts on the 1.0 platform

    1. Summary

    At 1:10 PM MST, the Redis cache supporting the 1.0 platform reached 100% storage utilization, causing cache write failures and forcing the platform to fall back to slower database queries. This resulted in significant API latency increases, ultimately leading to timeouts for some users. An alert was triggered at 1:17 PM, the issue was identified as Redis storage exhaustion, and the Redis instance was resized. Normal operation resumed at 1:45 PM.

    2. Impact
    Affected system: 1.0 Platform APIs

    User impact:
    Some API requests experienced elevated latency
    A subset of users experienced full timeouts

    Business impact:
    Temporary performance degradation
    Increased system load during fallback operations
    Duration: Approximately 35 minutes

    3. Root Cause
    The Redis cache used to optimize 1.0 platform queries ran out of available storage. Once storage was exhausted:

    • Redis evictions and failures occurred

    • Systems reliant on cached query results began performing full database queries

    • Resulting load contributed to latency increases and some request timeouts

    Contributing factor:
    A faulty Redis storage monitoring alert failed to notify the operations team before the cache reached saturation.

    4. Timeline
    Time (MST) Event

    • 1:10 PM Redis cache reaches 100% storage capacity; platform begins falling back to database queries, increasing latency.

    • 1:17 PM Operations alert triggered due to API timeouts.

    • 1:20–1:40 PM Investigation identifies Redis storage exhaustion; Redis instance is resized.

    • 1:45 PM Resized Redis instance becomes fully operational; traffic and latency return to normal ranges.

    5. Resolution
    Redis instance was resized, increasing available storage capacity. Platform traffic and API latency returned to baseline levels after the upgrade was completed.

    6. Preventative Measures
    Completed
    Fixed the faulty Redis storage monitor, ensuring future alerts will correctly trigger before storage is fully consumed.

  • Resolved
    Resolved

    This incident has been resolved. Thank you for your patience.

  • Monitoring
    Monitoring
    We implemented a fix and are currently monitoring the result.
  • Investigating
    Investigating
    We are currently investigating this incident.

Dec 2025 to Feb 2026

Next