Vapi Docs recovered
Timeline · 1 update
- resolved Apr 20, 2026, 10:22 PM UTC
Vapi Docs recovered
There were 19 Is Vapi Down Right Now? Live Vapi Status & Outages | IsDown outages since February 12, 2026 totaling 361h 8m of downtime. Each is summarised below — incident details, duration, and resolution information.
Vapi Docs recovered
Our documentation provider is experiencing an outage. We’re in touch with the team and will provide communication when services are back up
We are investigating reports of degraded performance affecting a subset of calls. Our team is actively working to determine the root cause and will provide updates as we learn more.
We have now resolved this incident. Between ~7:30AM PT and ~9:30AM PT, a new query pattern caused database slowness that increased API latency and led to dropped inbound calls with certain providers. We mitigated by rolling back the deployment and are following up with a deeper review of the query change.
Vapi Numbers Inbound recovered
We are currently experiencing degradation in our SIP infrastructure, resulting in in-call failures including call transfers, increased latency, and other related issues. Our team is actively investigating and working to resolve the problem. We will provide updates as more information becomes available.
The issue has been resolved. After applying the remediation, we monitored the affected systems and confirmed stable operation. All services are functioning normally.
We are currently observing elevated error rates affecting calls that use the Soniox transcriber. Impacted calls may terminate unexpectedly with the ended reason call.in-progress.error-vapifault-soniox-transcriber-failed. While we work to resolve this, we recommend switching to an alternative transcriber or configuring a transcriber fallback plan to ensure call continuity. You can set up fallbacks by following the guide here: https://docs.vapi.ai/customization/transcriber-fallback-plan We are actively monitoring the situation and will provide updates as more information becomes available.
We observed an elevated rate of API errors from 1:28pm to 1:39pm PDT. The errors have since resolved. We are closely monitoring API performance and investigating the root cause.
We're seeing elevated call failures on weekly, and the team is actively looking into it.
Incident Report, March 19, 2026 Impact: A service disruption affected inbound and outbound call reliability on the Daily and Weekly channel. Some calls failed to connect with transport-never-connected, worker-not-available, worker-died, and deepgram-transcriber-failed end reasons. Timeline (all times PDT): 12:20 PM We detected elevated call failure rates on the Weekly production cluster. 12:22 PM We published a status page incident and began investigating. 12:25 PM We identified the trigger as an unanticipated surge in call volume that exceeded our provisioned cluster capacity and downstream rate limits with a model provider. 12:30 PM We applied traffic controls and began working with the model provider to increase capacity. Call failures began declining. 1:40 PM Call success rates returned to normal and held stable. First incident window closed. ~4:00 PM A separate traffic spike re-triggered infrastructure constraints, leading to elevated failures. We began investigating immediately. 4:00 PM to 4:40 PM We rebalanced traffic and migrated affected workloads to dedicated infrastructure to restore headroom on shared clusters. 4:50 PM All mitigations took effect. Call success rates returned to normal. 4:50 PM to 8:10 PM We continued active monitoring. No further failures observed. 8:10 PM Second incident window closed. Immediate Action Items: Improve workload isolation and per-account capacity guardrails to prevent resource contention from cascading across the platform. Note: A full root cause analysis is underway and will be available upon request. We sincerely apologize for the disruption and thank you for your patience.
We are aware of elevated call failures rates on the weekly cluster with worker-not-available ended reason, and deepgram-transcriber-unavailable in both daily and weekly channels. Our team is actively investigating the issue.
Resolved — The issue causing degraded performance has been identified and mitigated as of 13:45 PDT. All services are operating normally. We will continue to monitor and provide an update if needed.
SIP Inbound recovered
High failure rate in connecting calls and API in the daily channel. Team is working on resolving it.
Resolved — The issue causing degraded performance has been identified and mitigated as of 8:22 AM. All services are operating normally. We will continue to monitor and provide an update if needed.
We're noticing a small percentage of calls using GPT 5.2 fail with an internal error during inference from OpenAI's side. We've reached out to the team and are closely monitoring the situation. In the meantime, we recommend switching to another model as we're seeing the degradation only on 5.2 currently.
The issue has been resolved we're not seeing any further degradation
We're noticing some call degradation primarily on the Daily channel. We are monitoring the situation and will update as we know more.
The issue has been resolved and all systems are operational.
We're noticing issues with calls using the Vapi voice "Emma". If you are using this voice, we recommend switching to another voice while this is resolved, and add voice fallbacks to prevent complete failures - https://docs.vapi.ai/voice-fallback-plan No other voices are currently impacted. We will update the status page as we know more.
The issue has been resolved and the voice is currently usable. We still recommend setting up fallbacks for any voices for the future to avoid call drops - https://docs.vapi.ai/voice-fallback-plan
The issue has been resolved and the voice is currently usable. We still recommend setting up fallbacks for any voices for the future to avoid call drops - https://docs.vapi.ai/voice-fallback-plan
We are seeing calls being dropped on the Daily channel with ended reason "call.in-progress.error-vapifault-worker-died". The team is looking into it and will update here.
We have not seen the issue since ~3:50pm today. We have determined the root cause and are rolling out a fix to improve stability in the daily channel. Incident Report — Daily Channel Call Failures (February 25, 2026) Impact Between 7:27 AM – 3:59 PM PST, approximately 19,527 calls failed on the Daily channel due to call worker failures. All Daily channel users were impacted. The Weekly channel was not affected. Timeline (all times in PST) 7:27 AM — Degraded call reliability detected on the Daily channel. Status page updated and investigation begins immediately. 8:33 AM — Issue escalated. Team recommends affected customers switch to the Weekly channel while investigation continues. Status page updated. 9:04 AM — Team begins proactive outreach to guide affected customers to the Weekly channel. 10:55 AM — Additional call failures observed on Daily after a brief period of stability. Investigation continues. 11:00 AM — Rolled back previous deployment. Did not observe any significant improvement. 1:30 PM — Continued investigating the issue. 3:59 PM — No further issues observed. 6:00 PM — Released a fix to improve stability in the Daily channel. Incident resolved. What Went Well The issue was detected and acknowledged quickly. A dedicated incident response was organized promptly to focus investigation. Teams were notified early and guided affected customers to switch to the Weekly channel. Action Items Isolate background operations from call handling Strengthen deployment validation Improve resilience under load Expand monitoring and alerting Note This report is intended as a summary of the incident timeline, impact, and immediate action items. A deeper root cause analysis is available upon request. This issue impacted the Daily channel only. Customers desiring increased stability (at the cost of delayed access to features) can switch to the Weekly channel by navigating to Organization Settings on the Vapi Dashboard and changing the Channel to "weekly".
We have not seen the issue since ~3:50pm today. We have determined the root cause and are rolling out a fix to improve stability in the `daily` channel. # Incident Report — Daily Channel Call Failures (February 25, 2026) ## Impact Between 7:27 AM – 3:59 PM PST, approximately 19,527 calls failed on the Daily channel due to call worker failures. All Daily channel users were impacted. The Weekly channel was not affected. ## Timeline (all times in PST) **7:27 AM** — Degraded call reliability detected on the Daily channel. Status page updated and investigation begins immediately. **8:33 AM** — Issue escalated. Team recommends affected customers switch to the Weekly channel while investigation continues. Status page updated. **9:04 AM** — Team begins proactive outreach to guide affected customers to the Weekly channel. **10:55 AM** — Additional call failures observed on Daily after a brief period of stability. Investigation continues. **11:00 AM** — Rolled back previous deployment. Did not observe any significant improvement. **1:30 PM** — Continued investigating the issue. **3:59 PM** — No further issues observed. **6:00 PM** — Released a fix to improve stability in the Daily channel. Incident resolved. ## What Went Well - The issue was detected and acknowledged quickly. - A dedicated incident response was organized promptly to focus investigation. - Teams were notified early and guided affected customers to switch to the Weekly channel. ## Action Items - Isolate background operations from call handling - Strengthen deployment validation - Improve resilience under load - Expand monitoring and alerting ## Note This report is intended as a summary of the incident timeline, impact, and immediate action items. A deeper root cause analysis is available upon request. This issue impacted the Daily channel only. Customers desiring increased stability (at the cost of delayed access to features) can switch to the Weekly channel by navigating to **Organization Settings** on the Vapi Dashboard and changing the Channel to **"weekly"**.
We are seeing calls degraded on Weekly channel. The team is looking into the issue and will share updates here.
Incident report: Impact: A service disruption affected call reliability on the Weekly channel. Some calls ended unexpectedly with worker-not-available or worker-died end reasons. Timeline (all times PT): 8:07 AM - We detected a burst of call failures across the platform. 8:16 AM - Automated monitoring alert fired. We acknowledged and began investigating. 8:42 AM - We scoped the impact across affected accounts. 8:47 AM - The issue self-resolved. We identified the root cause as resource contention in our call processing infrastructure during a traffic spike. 9:18 AM - We completed an initial root cause analysis and identified an underlying bottleneck in our call queue infrastructure. 11:13 AM - A related issue resurfaced due to cascading effects from the earlier contention. We began investigating immediately. 11:25 AM - We published a status page to notify customers. 11:38 AM - We confirmed the root cause as CPU contention between infrastructure components. 11:39 AM - We applied a mitigation. Call queue metrics began recovering. 11:45 AM - We updated the status page with the identified cause and fix. 11:46 AM - Error rates began declining. We continued active monitoring. 1:11 PM - We declared resolution on the status page. ~1:35 PM - A brief secondary spike occurred during an infrastructure resource adjustment. We responded immediately. 3:21 PM - All systems fully stabilized. Action Items Enforce resource limits across processing components and improve infrastructure isolation for critical call processing. Note A full root cause analysis is underway and will be available upon request. We sincerely apologize for the disruption and thank you for your patience.
Incident report: **Impact:** A service disruption affected call reliability on the Weekly channel. Some calls ended unexpectedly with `worker-not-available` or `worker-died` end reasons. **Timeline (all times PT):** - **8:07 AM** - We detected a burst of call failures across the platform. - **8:16 AM** - Automated monitoring alert fired. We acknowledged and began investigating. - **8:42 AM** - We scoped the impact across affected accounts. - **8:47 AM** - The issue self-resolved. We identified the root cause as resource contention in our call processing infrastructure during a traffic spike. - **9:18 AM** - We completed an initial root cause analysis and identified an underlying bottleneck in our call queue infrastructure. - **11:13 AM** - A related issue resurfaced due to cascading effects from the earlier contention. We began investigating immediately. - **11:25 AM** - We published a status page to notify customers. - **11:38 AM** - We confirmed the root cause as CPU contention between infrastructure components. - **11:39 AM** - We applied a mitigation. Call queue metrics began recovering. - **11:45 AM** - We updated the status page with the identified cause and fix. - **11:46 AM** - Error rates began declining. We continued active monitoring. - **1:11 PM** - We declared resolution on the status page. - **~1:35 PM** - A brief secondary spike occurred during an infrastructure resource adjustment. We responded immediately. - **3:21 PM** - All systems fully stabilized. **Action Items** Enforce resource limits across processing components and improve infrastructure isolation for critical call processing. **Note** A full root cause analysis is underway and will be available upon request. We sincerely apologize for the disruption and thank you for your patience.
Our auth service provider is reporting a degradation specifically in India. Some customers in that region will see issues with login. See our providers status page for live updates: https://status.supabase.com/incidents/xmgq69x4brfk.
Our provider has confirmed the issue is specific to accessing projects and seems to not include authentication. They are still resolving the issue from their end, but we are marking the issue as resolved. If customers continue to see issues with authentication, please reach out to [email protected]
We are seeing decreased success rate in calls on the daily channel. The team is investigating and we will post updates here. In the meantime, we highly recommend switching to weekly channel to mitigate service disruption.
The issue is resolved as of 10:05am PST. Incident Report Impact Between 9:10–10:05 AM, 37,806 calls were dropped due to call worker failures. All Daily users were impacted. Timeline (all times in PST) 9:02 AM — On-call engineer notices pods crashing in the Daily cluster. 9:11 AM — Black box probe alert fires; acknowledged by on-call engineer, triggering investigation. 9:27 AM — Issue escalates to the point of impacting all calls on Daily. 9:30 AM — Status page created to inform users of impact and request they switch to the Weekly channel. 9:34 AM — Incident team assembles. 9:37 AM — Rollback to previous deployment is initiated. Due to a large backlog of unprocessed jobs, rollback is delayed waiting for an excessive number of pods to become ready. 10:05 AM — Forceful cutover is initiated and service is restored. What Went Well Monitoring detected the issue before it became widespread. On-call engineer assembled the incident team quickly. Action Items Improve emergency rollback procedure to bypass or relax pod readiness checks during incidents, enabling faster cutover. Continue ongoing observability improvements to reduce MTTD. Note A full root cause analysis is underway and available upon request. This report is intended as a summary of the incident timeline, impact, and immediate action items. Note that this issue impacted the Daily cluster only. Customers desiring increased stability (at the cost of delayed access to features) should switch to the Weekly channel by navigating to Organization Settings on the Vapi Dashboard and changing the Channel to "weekly".
The issue is resolved as of 10:05am PST. ## Incident Report ### Impact Between 9:10–10:05 AM, 37,806 calls were dropped due to call worker failures. All Daily users were impacted. ### Timeline (all times in PST) 9:02 AM — On-call engineer notices pods crashing in the Daily cluster. 9:11 AM — Black box probe alert fires; acknowledged by on-call engineer, triggering investigation. 9:27 AM — Issue escalates to the point of impacting all calls on Daily. 9:30 AM — Status page created to inform users of impact and request they switch to the Weekly channel. 9:34 AM — Incident team assembles. 9:37 AM — Rollback to previous deployment is initiated. Due to a large backlog of unprocessed jobs, rollback is delayed waiting for an excessive number of pods to become ready. 10:05 AM — Forceful cutover is initiated and service is restored. ### What Went Well Monitoring detected the issue before it became widespread. On-call engineer assembled the incident team quickly. ### Action Items Improve emergency rollback procedure to bypass or relax pod readiness checks during incidents, enabling faster cutover. Continue ongoing observability improvements to reduce MTTD. ### Note A full root cause analysis is underway and available upon request. This report is intended as a summary of the incident timeline, impact, and immediate action items. Note that this issue impacted the Daily cluster only. Customers desiring increased stability (at the cost of delayed access to features) should switch to the Weekly channel by navigating to Organization Settings on the Vapi Dashboard and changing the Channel to "weekly".
email authentication for the vapi dashboard was experiencing issues. users were unable to sign in using their email credentials.
Email authentication for the Vapi dashboard has been restored. Users should now be able to sign in normally using their email credentials.
Pingoru polls Is Vapi Down Right Now? Live Vapi Status & Outages | IsDown's status page every 5 minutes and alerts you the moment it reports an issue — before your customers do.
5 free monitors · No credit card required