Kustomer Outage History

Kustomer is up right now

There were 6 Kustomer outages since February 12, 2026 totaling 10h 51m of downtime. Each is summarised below — incident details, duration, and resolution information.

Source: https://status.kustomer.com

Minor May 1, 2026

[Gmail]Email Connectivity & Send/Receive Issues - Prod 1 & Prod 2

Detected by Pingoru
May 01, 2026, 01:39 PM UTC
Resolved
May 01, 2026, 02:17 PM UTC
Duration
38m
Affected: Channel - EmailChannel - Email
Timeline · 4 updates
  1. identified May 01, 2026, 01:39 PM UTC

    Kustomer is aware of an event affecting Gmail Connectivity and Send & Receive that may cause emails to not route into your Kustomer Platform and impact your ability to send messages on conversations. Our team is currently working to identify the cause of this issue in an effort to implement a resolution. Please expect additional updates within the next 30 minutes, please reach out to Kustomer Support via Chat for any further questions or updates.

  2. identified May 01, 2026, 01:39 PM UTC

    Kustomer is aware of an event affecting Gmail Connectivity and Send & Receive that may cause emails to not route into your Kustomer Platform and impact your ability to send messages on conversations. Our team is currently working to identify the cause of this issue in an effort to implement a resolution. Please expect additional updates within the next 30 minutes, please reach out to Kustomer Support via Chat for any further questions or updates.

  3. monitoring May 01, 2026, 02:02 PM UTC

    Kustomer has released a fix for the event affecting Gmail Connectivity and Send & Receive functionality that was causing emails to not forward as expected into your Kustomer Platform and impacted sending functionality on conversations. Our team is currently monitoring the released fix to ensure the issue is resolved. Please expect additional updates within the next 30 minutes, please reach out to Kustomer Support via Chat for any further questions or updates.

  4. resolved May 01, 2026, 02:17 PM UTC

    Kustomer has resolved an event affecting Gmail Connectivity and Send & Receive functionality in Prod 1 and Prod 2. After careful monitoring, our team has determined that all affected areas are now fully restored. Please reach out to Kustomer support at [email protected] if you have additional questions or concerns.

Read the full incident report →

Minor April 24, 2026

[DRAFTS] Internal API Errors (Prod1)

Detected by Pingoru
Apr 24, 2026, 08:03 PM UTC
Resolved
Apr 24, 2026, 09:47 PM UTC
Duration
1h 44m
Affected: Channel - ChatChannel - Email
Timeline · 6 updates
  1. monitoring Apr 24, 2026, 08:03 PM UTC

    Kustomer has identified an event affecting Prod1 that may cause internal API errors and latency. Our team is currently working to implement a resolution. Please expect further updates within the next 30 minutes, and reach out to Kustomer support at [email protected] if you have additional questions or concerns.

  2. monitoring Apr 24, 2026, 08:32 PM UTC

    Kustomer has identified an event affecting Prod1 that may cause internal API errors and latency. Our team is currently working to implement a resolution. Please expect further updates within the next 30 minutes, and reach out to Kustomer support at [email protected] if you have additional questions or concerns.

  3. monitoring Apr 24, 2026, 09:02 PM UTC

    Kustomer has identified an event affecting Prod1 that may cause internal API errors and latency. Our team is currently working to implement a resolution. Please expect further updates within the next 30 minutes, and reach out to Kustomer support at [email protected] if you have additional questions or concerns.

  4. monitoring Apr 24, 2026, 09:31 PM UTC

    Kustomer has implemented an update to address an event affecting Prod1 that caused internal API errors and latency. Our team is currently monitoring this update to ensure the issue is fully resolved. Please expect further updates within the next 30 minutes, and reach out to Kustomer support at [email protected] if you have additional questions or concerns.

  5. resolved Apr 24, 2026, 09:47 PM UTC

    Kustomer has resolved an event affecting Prod1 that caused API errors and latency. To resolve this issue, our team has released an update. After careful monitoring, our team has determined that all affected areas are now fully restored. Please reach out to Kustomer support at [email protected] if you have additional questions or concerns.

  6. postmortem Apr 30, 2026, 06:21 PM UTC

    ## Summary On April 24, 2026, customers in our prod1 environment experienced a service event that caused elevated errors and latency in messaging-related workflows. The broad cross-customer impact was limited to approximately 41 minutes, from 3:26 PM ET to 4:07 PM ET. During that window, some customers saw failed or delayed messaging operations. The immediate platform impact was resolved the same day, and overall system health returned to normal. We then completed follow-up mitigation to stop the underlying event source and reduce the risk of recurrence. ## Impact * Customers in prod1 experienced elevated API errors and latency in messaging-related workflows. * The broad cross-customer impact lasted about 41 minutes. * A subset of messaging workflows failed or were delayed during that period. ## Timeline * **~3:25 PM ET** — A newly enabled automation began processing a large backlog of historical conversations for one tenant. * **3:26 PM ET** — Elevated errors and latency began affecting shared messaging workflows in prod1. * **4:01 PM ET** — We published a status update for the production issue. * **4:07 PM ET** — Broad cross-customer impact ended as the affected services stabilized. * **~5:17 PM ET** — We disabled the triggering automation configuration for the affected tenant. * **~5:23 PM ET** — The remaining retry activity stopped. ## Root cause The event was triggered when a newly enabled automation for one tenant processed a much larger set of eligible conversations than intended. That sudden volume overloaded a shared downstream service and caused elevated errors and timeouts in dependent workflows. The incident was amplified by missing safeguards in how this automation handled backlog volume and retries. In particular, the system did not sufficiently limit the number of conversations processed at once or prevent the same failed work from being retried too aggressively. ## Resolution We restored platform stability during the incident by allowing the affected services to recover under increased capacity, then disabled the triggering automation configuration and cleared the remaining retry backlog. System health is currently normal. ## Preventative actions We are treating the following preventative actions as a priority bug effort. These actions are expected to be resolved by the end of May in accordance with our SLOs: * Prevent newly enabled automation settings from processing large historical backlogs unintentionally. * Add stronger batch limits and tenant-level throttling for this workflow. * Reduce retry amplification by improving how failed work is tracked and re-queued. * Improve error handling so rate-limit conditions are classified correctly and handled with the right retry behavior.

Read the full incident report →

Minor April 14, 2026

[Knowledge Base] [Custom Domains returning 500 error] [POD(s) AFFECTED]

Detected by Pingoru
Apr 14, 2026, 02:44 PM UTC
Resolved
Apr 14, 2026, 03:01 PM UTC
Duration
16m
Affected: Knowledge baseKnowledge baseKnowledge base
Timeline · 3 updates
  1. investigating Apr 14, 2026, 02:44 PM UTC

    Kustomer is aware of an event affecting Knowledge bases and forms that may be displaying a 500 error code and preventing access to these urls. Our team is currently working to identify the cause of this issue in an effort to implement a resolution. Please expect additional updates within the next 30 minutes, please reach out to Kustomer Support via kustomer.com for any further questions or updates.

  2. resolved Apr 14, 2026, 03:01 PM UTC

    Kustomer has resolved an event affecting Knowledge Bases and forms that caused a 500 error and preventing access to these pages. To resolve this issue, our team has performed a roll back in this area. After careful monitoring, our team has determined that all affected areas are now fully restored. Please reach out to Kustomer Support via kustomer.com if you have additional questions or concerns.

  3. postmortem Apr 24, 2026, 05:14 PM UTC

    ## **Summary** ‌ On April 14, 2026, customers experienced errors accessing the Kustomer Knowledge Base, with all KB pages returning 500 errors. The issue was caused by a dependency upgrade in a recent deployment that introduced a TLS certificate mismatch in internal service communication. Customer impact began at 9:51 AM ET when the deployment reached the first production environment, and was fully resolved by 10:38 AM ET — a ~47 minute impact window. Engineers identified the root cause and completed a full rollback across all production environments within 12 minutes of the initial alert. ‌ ## **Root Cause** ‌ A recent deployment to the KB service included an upgrade to an internal library that contained a known issue with TLS hostname verification. This caused internal service-to-service requests to fail, resulting in 500 errors for all KB requests. The affected library version had previously been identified as problematic in a staging environment, but the fix had not been fully applied across all services before this deployment reached production. ## **Timeline** ‌ **Apr 14, 2026** ‌ 9:51 AM — Deployment reached the first production environment; customers began experiencing 500 errors when accessing the Knowledge Base ‌ 10:26 AM — Automated alerting fired; incident response began ‌ 10:31 AM — Engineers identified a recent deployment as the likely cause and began investigating rollback options ‌ 10:35 AM — Root cause confirmed as a problematic internal library version; rollback initiated across all production environments ‌ 10:37 AM — Rollback completed on prod1; KB restored for affected customers ‌ 10:38–10:39 AM — Rollback completed across remaining production environments ‌ 10:45 AM — Full KB functionality confirmed restored for all customers ‌ 12:09 PM — Corrected fix deployed to the Knowledge Base service; remediation completed across all other affected services ‌ ## **Lessons/Improvements** * Implementing stricter controls to prevent pre-release or beta library versions from being deployed to production * Improving the process for tracking and completing cross-service remediation work when an issue is identified in one service, to ensure all affected services are addressed * Enhancing our pre-production validation process to improve detection of this class of issue before it reaches production environments

Read the full incident report →

Minor March 19, 2026

[AIC/AIR] AIC/AIR May Not Respond - Prod 1

Detected by Pingoru
Mar 19, 2026, 03:14 PM UTC
Resolved
Mar 19, 2026, 06:29 PM UTC
Duration
3h 14m
Affected: Channel - Chat
Timeline · 8 updates
  1. investigating Mar 19, 2026, 02:42 PM UTC

    Kustomer is aware of an event affecting AI for Customers and Reps that have resulted in them not responding. Our team is currently working to identify the cause for the issue in order to implement a resolution. Please expect additional updates within the next 30 minutes, please reach out to Kustomer Support via kustomer.com for any further questions or updates.

  2. identified Mar 19, 2026, 03:44 PM UTC

    Kustomer has identified an event in AIC/AIR that may cause unresponsive functionality. Our team is still continuing to work on implementing a resolution. Please expect additional updates within the next 30 minutes, please reach out to Kustomer Support via Kustomer.com for any further questions or updates.

  3. identified Mar 19, 2026, 04:16 PM UTC

    Kustomer has identified an event that is causing unresponsiveness in AIC/AIR. Our team is working directly with our database provider in order to implement a resolution for this event. Please expect additional updates within the next 30 minutes, please reach out to Kustomer Support via Kustomer.com for any further questions or updates.

  4. identified Mar 19, 2026, 04:58 PM UTC

    Kustomer has identified an event affecting PROD 1 that may cause failing Kustomer AI-Agent features. Our team is currently working to implement a resolution. Please expect further updates within the next 30 minutes, and reach out to Kustomer support at [email protected] if you have additional questions or concerns.

  5. identified Mar 19, 2026, 05:34 PM UTC

    Kustomer has identified an event affecting PROD 1 that may result in failures with Kustomer AI Agent features. Our team is still actively working to implement a resolution. Please expect further updates within the next 30 minutes, and feel free to reach out to Kustomer Support at [email protected] if you have any additional questions or concerns.

  6. monitoring Mar 19, 2026, 05:52 PM UTC

    Kustomer has implemented an update to address an event affecting Prod 1 that caused failures with Kustomer's AI features. Our team is currently monitoring this update to ensure the issue is fully resolved. Please expect further updates within the next 30 minutes, and reach out to Kustomer support at [email protected] if you have additional questions or concerns.

  7. resolved Mar 19, 2026, 06:29 PM UTC

    Kustomer has resolved an event affecting Prod 1 that caused failures with Kustomer's AI features. After careful monitoring, our team has determined that all affected areas are now fully restored. Please reach out to Kustomer support at [email protected] if you have additional questions or concerns.

  8. postmortem Mar 27, 2026, 03:38 PM UTC

    ## Summary On March 19, 2026, some Kustomer AI features were unavailable for customers hosted in one production environment \(Prod1\). The disruption affected AI-powered responses and related AI workflows for 41 customers in our Prod 1 instance. Other production environments were not affected. Service was fully restored within 4 hours, and the platform is operating normally. ## Root cause The incident was caused by a production database index change that was introduced during a service deployment. That change caused database performance to degrade significantly for a high-traffic dataset used by our AI services. As performance deteriorated, dependent services were unable to start or process requests normally, which led to broad disruption across affected AI features. Recovery was prolonged by duplicate records created during the incident window, which complicated restoration of normal database constraints and required additional remediation before services could be brought back cleanly. ## Timeline All times below are in EDT \(UTC-4\) on March 19, 2026. * **~10:27 AM** — A production deployment introduced a database index change that degraded performance for AI-related services. * **~10:37 AM** — Customer impact was confirmed and incident response began. * **~10:42 AM** — We published a public status update and began active mitigation. * **~11:37 AM** — We identified the primary cause and focused recovery on database stability and service restoration. * **~12:10 PM** — After scaling database capacity and reducing load, service recovery began. * **~1:43 PM** — AI services began recovering and queued work started draining. * **~2:19 PM** — Backlogged work had been processed and core functionality was restored. * **Later that afternoon** — Follow-up cleanup was completed and normal safeguards were re-applied. ## Lessons and improvements We take this incident seriously and are making changes to reduce the likelihood of recurrence. * We are tightening how database schema and index changes are deployed in production, including stronger pre-deployment validation and safer rollout sequencing. * We are removing index-management behavior from service startup paths where it can create unnecessary risk during deployment. * We are improving monitoring and alerting so service health failures and database stress are detected earlier. * We are refining incident response procedures, access readiness, and operational runbooks to speed mitigation during future incidents. * We are reviewing capacity and resilience safeguards for this part of the platform to better handle abnormal database load.

Read the full incident report →

Minor February 20, 2026

[Voice] [Trouble accepting Inbound Calls] [PROD 1]

Detected by Pingoru
Feb 20, 2026, 10:01 PM UTC
Resolved
Feb 21, 2026, 12:01 AM UTC
Duration
2h
Affected: Kustomer Voice
Timeline · 7 updates
  1. monitoring Feb 20, 2026, 10:01 PM UTC

    Kustomer has implemented an update to address an event affecting inbound Voice calls in PROD 1 & 2 that caused inbound calls to drop before getting answered by an agent. Our team is currently monitoring this update to ensure the issue is fully resolved. Please expect further updates within the next 30 minutes, and reach out to Kustomer support at Email or Chat if you have additional questions or concerns.

  2. investigating Feb 20, 2026, 10:21 PM UTC

    Kustomer is aware of an event affecting Voice conversations that may cause inbound calls to not be accepted properly in the system. Our team is currently working to identify the cause of this issue in an effort to implement a resolution. Please expect additional updates within the next 30 minutes, please reach out to Kustomer Support via Chat or Email for any further questions or updates.

  3. investigating Feb 20, 2026, 10:53 PM UTC

    Kustomer is continuing to investigate an event impacting Voice conversations that may cause inbound calls to not be accepted properly within the system. Our engineering team is actively working to identify the root cause and implement a resolution as quickly as possible. We will provide another update within the next 30 minutes or sooner as more information becomes available. If you have any urgent questions or need assistance, please contact Kustomer Support via Chat or Email.

  4. identified Feb 20, 2026, 11:08 PM UTC

    Kustomer is aware of an issue affecting inbound call acceptance in our PROD1 environment where inbound calls were not properly being accepted in the system. Our team has identified the cause of the issue within the code and are actively working on a fix to restore system behavior back to normal. Please expect additional updates within the next 30 minutes, and reach out to Kustomer Support at Email or Chat Channels if you have additional questions or concerns.

  5. monitoring Feb 20, 2026, 11:33 PM UTC

    Kustomer has implemented an update to address an event affecting Voice conversations in PROD 1 that caused inbound calls to be dropped when trying to accept calls. Our team is currently monitoring this update to ensure the issue is fully resolved. If your agents are continuing to experience this issue please have them refresh their browser to apply the fix. Please expect further updates within the next 30 minutes, and reach out to Kustomer support at Email or Channel if you have additional questions or concerns.

  6. resolved Feb 21, 2026, 12:01 AM UTC

    Kustomer has resolved an incident affecting Voice in PROD 1 that caused inbound calls to drop after being initially accepted. To remediate the issue, our team rolled back a recent code change to a previously stable version, which successfully resolved the error. After thorough monitoring, we have confirmed that all impacted services are fully restored. If any agents continue to experience issues, please have them refresh their browser to ensure the fix is applied. If you have any additional questions or concerns, please reach out to Kustomer Support via Email or Chat.

  7. postmortem Mar 02, 2026, 04:53 PM UTC

    ## Summary On February 20, 2026, a frontend change intended to improve performance unintentionally disrupted call setup in our Voice experience. As a result, agents could see an incoming call and click accept/decline, but audio would not connect and calls would time out. We rolled back the change and then deployed an additional fix to make startup more reliable and prevent recurrence. ## What happened When the Voice widget starts, it needs to receive a small set of configuration settings \(including feature-flag values\) from the main app before it can route and connect calls correctly. A pre-existing timing edge case meant that, in some cases, that initial configuration could be sent before the Voice widget was fully ready to receive it. Historically, the widget would receive the same configuration again shortly afterward, which masked a state issue. The performance change reduced those repeated configuration sends \(which is normally a good optimization\). But because the widget sometimes missed the first configuration during startup, it could start with incomplete settings and fall back to legacy call-routing behavior. In that fallback mode, calls could appear answerable in the UI but fail to fully connect audio. ## Root cause A performance optimization changed how/when configuration settings were delivered during Voice widget startup. Combined with a pre-existing startup timing edge case, some sessions did not receive the required configuration in time and fell back to legacy call-routing behavior, preventing audio from being bridged correctly. ## Timeline * 3:04 PM ET: Frontend change deployed * 4:38 PM ET: Reports received that calls could not be answered \(accept/decline visible, no audio\) * 4:44 PM ET: Initial rollback attempt of an unrelated change did not resolve the issue * 4:50 PM ET: Incident communication initiated via status page * 6:24 PM ET: Frontend change reverted/rolled back * 8:47 PM ET: Reports continued for some agents * 10:31 PM ET: Identified that affected sessions could persist without a browser refresh due to cached UI state; refresh/new session restored correct behavior * Saturday 9:00 AM ET: Confirmed reports that calls were successfully connecting ## Lessons / improvements * Harden Voice widget startup so required configuration can’t be missed during initialization. * Add automated smoke/E2E coverage for core Voice call flows \(answer/connect audio\) to detect regressions before production. * Improve safeguards and monitoring around call-connect failures to reduce time-to-detect and time-to-mitigate.

Read the full incident report →

Minor February 12, 2026

Reported Third Party Event - OpenAI

Detected by Pingoru
Feb 12, 2026, 03:11 PM UTC
Resolved
Feb 12, 2026, 06:07 PM UTC
Duration
2h 56m
Affected: OpenAI
Timeline · 2 updates
  1. monitoring Feb 12, 2026, 03:11 PM UTC

    Kustomer is aware of an event reported by one of our third party vendors affecting Prod1 that may cause AIC and AIR failures within the platform. Our team is monitoring the incident, and working with the vendor where possible to resolve the issue. OpenAI's status page for this incident can be found here: https://status.openai.com/incidents/01KH94NGSXNH9H4WBPXB3RFZWX Please expect further updates within the next 3 hours, and reach out to Kustomer support at [email protected] if you have additional questions or concerns.

  2. resolved Feb 12, 2026, 06:07 PM UTC

    OpenAI has confirmed that all services are fully recovered. After careful monitoring, our team has determined that all affected areas are now fully restored. Please reach out to Kustomer support at [email protected] if you have additional questions or concerns.

Read the full incident report →

Looking to track Kustomer downtime and outages?

Pingoru polls Kustomer's status page every 5 minutes and alerts you the moment it reports an issue — before your customers do.

  • Real-time alerts when Kustomer reports an incident
  • Email, Slack, Discord, Microsoft Teams, and webhook notifications
  • Track Kustomer alongside 5,000+ providers in one dashboard
  • Component-level filtering
  • Notification groups + maintenance calendar
Start monitoring Kustomer for free

5 free monitors · No credit card required