voice-agents
Voice agents represent the frontier of AI interaction - humans speaking naturally with AI systems.
Best use case
voice-agents is best used when you need a repeatable AI agent workflow instead of a one-off prompt. It is especially useful for teams working in multi. You are a voice AI architect who has shipped production voice agents handling millions of calls. You understand the physics of latency - every component adds milliseconds, and the sum determines whether conversations feel natural or awkward.
Voice agents represent the frontier of AI interaction - humans speaking naturally with AI systems.
Users should expect a more consistent workflow output, faster repeated execution, and less time spent rewriting prompts from scratch.
Practical example
Example input
Use the "voice-agents" skill to help with this workflow task. Context: You are a voice AI architect who has shipped production voice agents handling millions of calls. You understand the physics of latency - every component adds milliseconds, and the sum determines whether conversations feel natural or awkward.
Example output
A structured workflow result with clearer steps, more consistent formatting, and an output that is easier to reuse in the next run.
When to use this skill
- Use this skill when you want a reusable workflow rather than writing the same prompt again and again.
When not to use this skill
- Do not use this when you only need a one-off answer and do not need a reusable workflow.
- Do not use it if you cannot install or maintain the related files, repository context, or supporting tools.
Installation
Claude Code / Cursor / Codex
Manual Installation
- Download SKILL.md from GitHub
- Place it in
.claude/skills/voice-agents/SKILL.mdinside your project - Restart your AI agent — it will auto-discover the skill
How voice-agents Compares
| Feature / Agent | voice-agents | Standard Approach |
|---|---|---|
| Platform Support | Not specified | Limited / Varies |
| Context Awareness | High | Baseline |
| Installation Complexity | Unknown | N/A |
Frequently Asked Questions
What does this skill do?
Voice agents represent the frontier of AI interaction - humans speaking naturally with AI systems.
Where can I find the source code?
You can find the source code on GitHub using the link provided at the top of the page.
Related Guides
Best AI Skills for Claude
Explore the best AI skills for Claude and Claude Code across coding, research, workflow automation, documentation, and agent operations.
AI Agents for Marketing
Discover AI agents for marketing workflows, from SEO and content production to campaign research, outreach, and analytics.
AI Agents for Startups
Explore AI agent skills for startup validation, product research, growth experiments, documentation, and fast execution with small teams.
SKILL.md Source
# Voice Agents
Voice agents represent the frontier of AI interaction - humans speaking
naturally with AI systems. The challenge isn't just speech recognition
and synthesis, it's achieving natural conversation flow with sub-800ms
latency while handling interruptions, background noise, and emotional
nuance.
This skill covers two architectures: speech-to-speech (OpenAI Realtime API,
lowest latency, most natural) and pipeline (STT→LLM→TTS, more control,
easier to debug). Key insight: latency is the constraint. Humans expect
responses in 500ms. Every millisecond matters.
84% of organizations are increasing voice AI budgets in 2025. This is the
year voice agents go mainstream.
## Principles
- Latency is the constraint - target <800ms end-to-end
- Jitter (variance) matters as much as absolute latency
- VAD quality determines conversation flow
- Interruption handling makes or breaks the experience
- Start with focused MVP, iterate based on real conversations
- Combine best-in-class components (Deepgram STT + ElevenLabs TTS)
## Capabilities
- voice-agents
- speech-to-speech
- speech-to-text
- text-to-speech
- conversational-ai
- voice-activity-detection
- turn-taking
- barge-in-detection
- voice-interfaces
## Scope
- phone-system-integration → backend
- audio-processing-dsp → audio-specialist
- music-generation → audio-specialist
- accessibility-compliance → accessibility-specialist
## Tooling
### Speech_to_speech
- OpenAI Realtime API - When: Lowest latency, most natural conversation Note: gpt-4o-realtime-preview, native voice, sub-500ms
- Pipecat - When: Open-source voice orchestration Note: Daily-backed, enterprise-grade, modular
### Speech_to_text
- OpenAI Whisper - When: Highest accuracy, multilingual Note: gpt-4o-transcribe for best results
- Deepgram Nova-3 - When: Production workloads, 54% lower WER Note: 150-184ms TTFT, 90%+ accuracy on noisy audio
- AssemblyAI - When: Real-time streaming, speaker diarization Note: Good accuracy-latency balance
### Text_to_speech
- ElevenLabs - When: Most natural voice, emotional control Note: Flash model 75ms latency, V3 for expression
- OpenAI TTS - When: Integrated with OpenAI stack Note: gpt-4o-mini-tts, 13 voices, streaming
- Deepgram Aura-2 - When: Cost-effective production TTS Note: 40% cheaper than ElevenLabs, 184ms TTFB
### Frameworks
- Pipecat - When: Open-source voice agent orchestration Note: Silero VAD, SmartTurn, interruption handling
- Vapi - When: Managed voice agent platform Note: No infrastructure management
- Retell AI - When: Low-latency voice agents Note: Best context preservation on interruption
## Patterns
### Speech-to-Speech Architecture
Direct audio-to-audio processing for lowest latency
**When to use**: Maximum naturalness, emotional preservation, real-time conversation
# SPEECH-TO-SPEECH ARCHITECTURE:
"""
[User Audio] → [S2S Model] → [Agent Audio]
Advantages:
- Lowest latency (sub-500ms)
- Preserves emotion, emphasis, accents
- Most natural conversation flow
Disadvantages:
- Less control over responses
- Harder to debug/audit
- Can't easily modify what's said
"""
## OpenAI Realtime API
"""
import { RealtimeClient } from '@openai/realtime-api-beta';
const client = new RealtimeClient({
apiKey: process.env.OPENAI_API_KEY,
});
// Configure for voice conversation
client.updateSession({
modalities: ['text', 'audio'],
voice: 'alloy',
input_audio_format: 'pcm16',
output_audio_format: 'pcm16',
instructions: `You are a helpful customer service agent.
Be concise and friendly. If you don't know something,
say so rather than making things up.`,
turn_detection: {
type: 'server_vad', // or 'semantic_vad'
threshold: 0.5,
prefix_padding_ms: 300,
silence_duration_ms: 500,
},
});
// Handle audio streams
client.on('conversation.item.input_audio_transcription', (event) => {
console.log('User said:', event.transcript);
});
client.on('response.audio.delta', (event) => {
// Stream audio to speaker
audioPlayer.write(Buffer.from(event.delta, 'base64'));
});
// Send user audio
client.appendInputAudio(audioBuffer);
"""
## Use Cases:
- Real-time customer support
- Voice assistants
- Interactive voice response (IVR)
- Live language translation
### Pipeline Architecture
Separate STT → LLM → TTS for maximum control
**When to use**: Need to know/control exactly what's said, debugging, compliance
# PIPELINE ARCHITECTURE:
"""
[Audio] → [STT] → [Text] → [LLM] → [Text] → [TTS] → [Audio]
Advantages:
- Full control at each step
- Can log/audit all text
- Easier to debug
- Mix best-in-class components
Disadvantages:
- Higher latency (700-1200ms typical)
- Loses some emotion/nuance
- More components to manage
"""
## Production Pipeline Example
"""
import { Deepgram } from '@deepgram/sdk';
import { ElevenLabsClient } from 'elevenlabs';
import OpenAI from 'openai';
// Initialize clients
const deepgram = new Deepgram(process.env.DEEPGRAM_API_KEY);
const elevenlabs = new ElevenLabsClient();
const openai = new OpenAI();
async function processVoiceInput(audioStream) {
// 1. Speech-to-Text (Deepgram Nova-3)
const transcription = await deepgram.transcription.live({
model: 'nova-3',
punctuate: true,
endpointing: 300, // ms of silence before end
});
transcription.on('transcript', async (data) => {
if (data.is_final && data.speech_final) {
const userText = data.channel.alternatives[0].transcript;
console.log('User:', userText);
// 2. LLM Processing
const completion = await openai.chat.completions.create({
model: 'gpt-4o-mini',
messages: [
{ role: 'system', content: 'You are a concise voice assistant.' },
{ role: 'user', content: userText }
],
max_tokens: 150, // Keep responses short for voice
});
const agentText = completion.choices[0].message.content;
console.log('Agent:', agentText);
// 3. Text-to-Speech (ElevenLabs)
const audioStream = await elevenlabs.textToSpeech.stream({
voice_id: 'voice_id_here',
text: agentText,
model_id: 'eleven_flash_v2_5', // Lowest latency
});
// Stream to user
playAudioStream(audioStream);
}
});
// Pipe audio to transcription
audioStream.pipe(transcription);
}
"""
## Optimization Tips:
- Start TTS while LLM still generating (streaming)
- Pre-compute first response segment during user speech
- Use Flash/turbo models for latency
### Voice Activity Detection Pattern
Detect when user starts/stops speaking
**When to use**: All voice agents need VAD for turn-taking
# VOICE ACTIVITY DETECTION (VAD):
"""
VAD Types:
1. Energy-based: Simple, fast, noise-sensitive
2. Model-based: Silero VAD, more accurate
3. Semantic VAD: Understands meaning, best for conversation
"""
## Silero VAD (Popular Open Source)
"""
import { SileroVAD } from '@pipecat-ai/silero-vad';
const vad = new SileroVAD({
threshold: 0.5, // Speech probability threshold
min_speech_duration: 250, // ms before speech confirmed
min_silence_duration: 500, // ms of silence = end of turn
});
vad.on('speech_start', () => {
console.log('User started speaking');
// Stop any playing TTS (barge-in)
audioPlayer.stop();
});
vad.on('speech_end', () => {
console.log('User finished speaking');
// Trigger response generation
processTranscript();
});
// Feed audio to VAD
audioStream.on('data', (chunk) => {
vad.process(chunk);
});
"""
## OpenAI Semantic VAD
"""
// In Realtime API session config
client.updateSession({
turn_detection: {
type: 'semantic_vad', // Uses meaning, not just silence
// Model waits longer after "ummm..."
// Responds faster after "Yes, that's correct."
},
});
"""
## Barge-In Handling
"""
// When user interrupts:
function handleBargeIn() {
// 1. Stop TTS immediately
audioPlayer.stop();
// 2. Cancel pending LLM generation
llmController.abort();
// 3. Reset state
conversationState.checkpoint();
// 4. Listen to new input
startListening();
}
// VAD triggers barge-in
vad.on('speech_start', () => {
if (audioPlayer.isPlaying) {
handleBargeIn();
}
});
"""
### Latency Optimization Pattern
Achieving <800ms end-to-end response time
**When to use**: Production voice agents
# LATENCY OPTIMIZATION:
"""
Target Metrics:
- End-to-end: <800ms (ideal: <500ms)
- Time-to-First-Token (TTFT): <300ms
- Barge-in response: <200ms
- Jitter variance: <100ms std dev
"""
## Pipeline Latency Breakdown
"""
Typical breakdown:
- VAD processing: 50-100ms
- STT first result: 150-200ms
- LLM TTFT: 100-300ms
- TTS TTFA: 75-200ms
- Audio buffering: 50-100ms
Total: 425-900ms
"""
## Optimization Strategies
### 1. Streaming Everything
"""
// Stream STT results as they come
stt.on('partial_transcript', (text) => {
// Start processing before final transcript
llmPreprocessor.prepare(text);
});
// Stream LLM output to TTS
const llmStream = await openai.chat.completions.create({
stream: true,
// ...
});
for await (const chunk of llmStream) {
tts.appendText(chunk.choices[0].delta.content);
}
"""
### 2. Pre-computation
"""
// While user is speaking, predict and prepare
stt.on('partial_transcript', async (text) => {
// Pre-fetch relevant context
const context = await retrieveContext(text);
// Pre-compute likely first sentence
const firstSentence = await generateOpener(context);
});
"""
### 3. Use Low-Latency Models
"""
// STT: Deepgram Nova-3 (150ms TTFT)
// LLM: gpt-4o-mini (fastest GPT-4 class)
// TTS: ElevenLabs Flash (75ms) or Deepgram Aura-2 (184ms)
"""
### 4. Edge Deployment
"""
// Run inference closer to user
// - Cloud regions near user
// - Edge computing for VAD/STT
// - WebSocket over HTTP for lower overhead
"""
### Conversation Design Pattern
Designing natural voice conversations
**When to use**: Building voice UX
# CONVERSATION DESIGN:
## Voice-First Principles
"""
Voice is different from text:
- No undo button - say it right the first time
- Linear - user can't scroll back
- Ephemeral - easy to miss information
- Emotional - tone matters as much as words
"""
## Response Design
"""
# Keep responses short (10-20 seconds max)
# Front-load the answer
# Use signposting for lists
Bad: "I found several options. The first is... second is..."
Good: "I found 3 options. Want me to go through them?"
# Confirm understanding
Bad: "I'll transfer $500 to John."
Good: "So that's $500 to John Smith. Should I proceed?"
"""
## Prompting for Voice
"""
system_prompt = '''
You are a voice assistant. Follow these rules:
1. Be concise - keep responses under 30 words
2. Use natural speech - contractions, casual language
3. Never use formatting (bullets, numbers in lists)
4. Spell out numbers and abbreviations
5. End with a question to keep conversation flowing
6. If unclear, ask for clarification
7. Never say "I'm an AI" unless asked
Good: "Got it. I'll set that reminder for three pm. Anything else?"
Bad: "I have set a reminder for 3:00 PM. Is there anything else I can assist you with today?"
'''
"""
## Error Recovery
"""
// Handle recognition errors gracefully
const errorResponses = {
no_speech: "I didn't catch that. Could you say it again?",
unclear: "Sorry, I'm not sure I understood. You said [repeat]. Is that right?",
timeout: "Still there? I'm here when you're ready.",
};
// Always offer human fallback for complex issues
if (confidenceScore < 0.6) {
response = "I want to make sure I get this right. Would you like to speak with a human agent?";
}
"""
## Sharp Edges
### Response Latency Exceeds 800ms
Severity: CRITICAL
Situation: Building a voice agent pipeline
Symptoms:
Conversations feel awkward. Users repeat themselves. "Are you
there?" questions. Users hang up or give up. Low satisfaction
scores despite correct answers.
Why this breaks:
In human conversation, responses typically arrive within 500ms.
Anything over 800ms feels like the agent is slow or confused.
Users lose confidence and patience. Every component adds latency:
VAD (100ms) + STT (200ms) + LLM (300ms) + TTS (200ms) = 800ms.
Recommended fix:
# Measure and budget latency for each component:
## Target latencies:
- VAD processing: <100ms
- STT time-to-first-token: <200ms
- LLM time-to-first-token: <300ms
- TTS time-to-first-audio: <150ms
- Total end-to-end: <800ms
## Optimization strategies:
1. Use low-latency models:
- STT: Deepgram Nova-3 (150ms) vs Whisper (500ms+)
- TTS: ElevenLabs Flash (75ms) vs standard (200ms+)
- LLM: gpt-4o-mini streaming
2. Stream everything:
- Don't wait for full STT transcript
- Stream LLM output to TTS
- Start audio playback before TTS finishes
3. Pre-compute:
- While user speaks, prepare context
- Generate opening phrase in parallel
4. Edge deployment:
- Run VAD/STT at edge
- Use nearest cloud region
## Measure continuously:
Log timestamps at each stage, track P50/P95 latency
### Response Time Variance Disrupts Rhythm
Severity: HIGH
Situation: Voice agent with inconsistent response times
Symptoms:
Conversations feel unpredictable. User doesn't know when to speak.
Sometimes agent responds immediately, sometimes after long pause.
Users talk over agent. Agent talks over users.
Why this breaks:
Jitter (variance in response time) disrupts conversational rhythm
more than absolute latency. Consistent 800ms feels better than
alternating 400ms and 1200ms. Users can't adapt to unpredictable
timing.
Recommended fix:
# Target jitter metrics:
- Standard deviation: <100ms
- P95-P50 gap: <200ms
## Reduce jitter sources:
1. Consistent model loading:
- Keep models warm
- Pre-load on connection start
2. Buffer audio output:
- Small buffer (50-100ms) smooths playback
- Don't start playing until buffer filled
3. Handle LLM variance:
- gpt-4o-mini more consistent than larger models
- Set max_tokens to limit long responses
4. Monitor and alert:
- Track response time distribution
- Alert on jitter spikes
## Implementation:
const MIN_RESPONSE_TIME = 400; // ms
async function respondWithConsistentTiming(text) {
const startTime = Date.now();
const audio = await generateSpeech(text);
const elapsed = Date.now() - startTime;
if (elapsed < MIN_RESPONSE_TIME) {
await delay(MIN_RESPONSE_TIME - elapsed);
}
playAudio(audio);
}
### Using Silence Duration for Turn Detection
Severity: HIGH
Situation: Detecting when user finishes speaking
Symptoms:
Agent interrupts user mid-thought. Or waits too long after user
finishes. "Let me think..." triggers premature response. Short
answers have awkward pause before response.
Why this breaks:
Simple silence detection (e.g., "end turn after 500ms silence")
doesn't understand conversation. Humans pause mid-sentence.
"Yes." needs fast response, "Well, let me think about that..."
needs patience. Fixed timeout fits neither.
Recommended fix:
# Use semantic VAD:
## OpenAI Semantic VAD:
client.updateSession({
turn_detection: {
type: 'semantic_vad',
// Waits longer after "umm..."
// Responds faster after "Yes, that's correct."
},
});
## Pipecat SmartTurn:
const pipeline = new Pipeline({
vad: new SileroVAD(),
turnDetection: new SmartTurn(),
});
// SmartTurn considers:
// - Speech content (complete sentence?)
// - Prosody (falling intonation?)
// - Context (question asked?)
## Fallback: Adaptive silence threshold:
function calculateSilenceThreshold(transcript) {
const endsWithComplete = transcript.match(/[.!?]$/);
const hasFillers = transcript.match(/um|uh|like|well/i);
if (endsWithComplete && !hasFillers) {
return 300; // Fast response
} else if (hasFillers) {
return 1500; // Wait for continuation
}
return 700; // Default
}
### Agent Doesn't Stop When User Interrupts
Severity: HIGH
Situation: User tries to interrupt agent mid-sentence
Symptoms:
Agent talks over user. User has to wait for agent to finish.
Frustrating experience. Users give up and abandon call.
"STOP! STOP!" doesn't work.
Why this breaks:
Without barge-in handling, the TTS plays to completion regardless
of user input. This violates basic conversational norms - in human
conversation, we stop when interrupted.
Recommended fix:
# Implement barge-in detection:
## Basic barge-in:
vad.on('speech_start', () => {
if (ttsPlayer.isPlaying) {
// 1. Stop audio immediately
ttsPlayer.stop();
// 2. Cancel pending TTS generation
ttsController.abort();
// 3. Checkpoint conversation state
conversationState.save();
// 4. Listen to new input
startTranscription();
}
});
## Advanced: Distinguish interruption types:
vad.on('speech_start', async () => {
if (!ttsPlayer.isPlaying) return;
// Wait 200ms to get first words
await delay(200);
const firstWords = getTranscriptSoFar();
if (isBackchannel(firstWords)) {
// "uh-huh", "yeah" - don't interrupt
return;
}
if (isClarification(firstWords)) {
// "What?", "Sorry?" - repeat last sentence
repeatLastSentence();
} else {
// Real interruption - stop and listen
handleFullInterruption();
}
});
## Response time target:
- Barge-in response: <200ms
- User should feel heard immediately
### Generating Text-Length Responses for Voice
Severity: MEDIUM
Situation: Prompting LLM for voice agent responses
Symptoms:
Agent rambles. Users lose track of information. "Can you repeat
that?" requests. Users interrupt to ask for shorter version.
Low comprehension of conveyed information.
Why this breaks:
Text can be scanned and re-read. Voice is linear and ephemeral.
A 3-paragraph response that works in chat is overwhelming in voice.
Users can only hold ~7 items in working memory.
Recommended fix:
# Constrain response length in prompts:
system_prompt = '''
You are a voice assistant. Keep responses UNDER 30 WORDS.
For complex information, break into chunks and confirm
understanding between each.
Instead of: "Here are the three options. First, you could...
Second... Third..."
Say: "I found 3 options. Want me to go through them?"
Never list more than 3 items without pausing for confirmation.
'''
## Enforce at generation:
const response = await openai.chat.completions.create({
max_tokens: 100, // Hard limit
// ...
});
## Chunking pattern:
if (information.length > 3) {
response = `I have ${information.length} items. Let's go through them one at a time. First: ${information[0]}. Ready for the next?`;
}
## Progressive disclosure:
"I found your account. Want the balance, recent transactions, or something else?"
// Don't dump all info at once
### Using Bullets/Numbers/Markdown in Voice
Severity: MEDIUM
Situation: Formatting LLM output for voice
Symptoms:
"First bullet point: item one" read aloud. Numbers read as "one
two three" instead of "one, two, three." Markdown artifacts in
speech. Robotic, unnatural delivery.
Why this breaks:
TTS models read what they're given. Text formatting intended for
visual display sounds robotic when read aloud. Users can't "see"
structure in audio.
Recommended fix:
# Prompt for spoken format:
system_prompt = '''
Format responses for SPOKEN delivery:
- No bullet points, numbered lists, or markdown
- Spell out numbers: "twenty-three" not "23"
- Spell out abbreviations: "United States" not "US"
- Use verbal signposting: "There are three things. First..."
- Never use asterisks, dashes, or special characters
'''
## Post-processing:
function prepareForSpeech(text) {
return text
// Remove markdown
.replace(/[*_#`]/g, '')
// Convert numbers
.replace(/\d+/g, numToWords)
// Expand abbreviations
.replace(/\betc\b/gi, 'et cetera')
.replace(/\be\.g\./gi, 'for example')
// Add pauses
.replace(/\. /g, '... ')
.replace(/, /g, '... ');
}
## SSML for precise control:
<speak>
The total is <say-as interpret-as="currency">$49.99</say-as>.
<break time="500ms"/>
Want to proceed?
</speak>
### VAD/STT Fails in Noisy Environments
Severity: MEDIUM
Situation: Users in cars, cafes, outdoors
Symptoms:
"I didn't catch that" frequently. Background noise triggers
false starts. Fan/AC causes continuous listening. Car engine
noise confuses STT.
Why this breaks:
Default VAD thresholds work for quiet environments. Real-world
usage includes background noise that triggers false positives
or masks speech, causing false negatives.
Recommended fix:
# Implement noise handling:
## 1. Noise reduction in STT:
const transcription = await deepgram.transcription.live({
model: 'nova-3',
noise_reduction: true,
// or
smart_format: true,
});
## 2. Adaptive VAD threshold:
// Measure ambient noise level
const ambientLevel = measureAmbientNoise(5000); // 5 sec sample
vad.setThreshold(ambientLevel * 1.5); // Above ambient
## 3. Confidence filtering:
stt.on('transcript', (data) => {
if (data.confidence < 0.7) {
// Low confidence - probably noise
askForRepeat();
return;
}
processTranscript(data.transcript);
});
## 4. Echo cancellation:
// Prevent agent's voice from being transcribed
const echoCanceller = new EchoCanceller();
echoCanceller.reference(ttsOutput);
const cleanedAudio = echoCanceller.process(userAudio);
### STT Produces Incorrect or Hallucinated Text
Severity: MEDIUM
Situation: Processing unclear or accented speech
Symptoms:
Agent responds to something user didn't say. Names consistently
wrong. Technical terms misheard. "I said X, not Y" frustration.
Why this breaks:
STT models can hallucinate, especially on proper nouns, technical
terms, or accented speech. These errors propagate through the
pipeline and produce nonsensical responses.
Recommended fix:
# Mitigate STT errors:
## 1. Use keywords/biasing:
const transcription = await deepgram.transcription.live({
keywords: ['Acme Corp', 'ProductName', 'John Smith'],
keyword_boost: 'high',
});
## 2. Confirmation for critical info:
if (containsNameOrNumber(transcript)) {
response = `I heard "${name}". Is that correct?`;
}
## 3. Confidence-based fallback:
if (confidence < 0.8) {
response = `I think you said "${transcript}". Did I get that right?`;
}
## 4. Multiple hypothesis handling:
// Some STT APIs return n-best list
const alternatives = transcription.alternatives;
if (alternatives[0].confidence - alternatives[1].confidence < 0.1) {
// Ambiguous - ask for clarification
}
## 5. Error correction patterns:
promptPattern = `
User may correct previous mistakes. If they say "no, I said X"
or "not Y, Z", update your understanding accordingly.
`;
## Validation Checks
### Missing Latency Measurement
Severity: ERROR
Voice agents must track latency at each stage
Message: Voice pipeline without latency tracking. Add timestamps at each stage to measure performance.
### Using Batch STT Instead of Streaming
Severity: WARNING
Streaming STT reduces latency significantly
Message: Using batch transcription. Consider streaming for lower latency in voice agents.
### TTS Without Streaming Output
Severity: WARNING
Streaming TTS reduces time to first audio
Message: TTS without streaming. Stream audio to reduce time to first audio.
### Hardcoded VAD Silence Threshold
Severity: WARNING
Fixed silence thresholds don't adapt to conversation
Message: Fixed silence threshold. Consider semantic VAD or adaptive thresholds for better turn-taking.
### Missing Barge-In Handling
Severity: WARNING
Voice agents should stop when user interrupts
Message: VAD without barge-in handling. Stop TTS when user starts speaking.
### Voice Prompt Without Length Constraints
Severity: WARNING
Voice prompts should constrain response length
Message: Voice prompt without length constraints. Add 'Keep responses under 30 words' to system prompt.
### Markdown Formatting Sent to TTS
Severity: WARNING
Markdown will be read literally by TTS
Message: Check for markdown in TTS input. Strip formatting before sending to TTS.
### STT Without Error Handling
Severity: WARNING
STT can fail or return low confidence
Message: STT without error handling. Check confidence scores and handle failures.
### WebSocket Without Reconnection
Severity: WARNING
Realtime APIs need reconnection handling
Message: Realtime connection without reconnection logic. Handle disconnects gracefully.
### Missing Noise Handling
Severity: INFO
Real-world audio includes background noise
Message: Consider adding noise handling for real-world audio quality.
## Collaboration
### Delegation Triggers
- user needs phone/telephony integration -> backend (Twilio, Vonage, SIP integration)
- user needs LLM optimization -> llm-architect (Model selection, prompting, fine-tuning)
- user needs tools for voice agent -> agent-tool-builder (Tool design for voice context)
- user needs multi-agent voice system -> multi-agent-orchestration (Voice agents working together)
- user needs accessibility compliance -> accessibility-specialist (Voice interface accessibility)
## Related Skills
Works well with: `agent-tool-builder`, `multi-agent-orchestration`, `llm-architect`, `backend`
## When to Use
- User mentions or implies: voice agent
- User mentions or implies: speech to text
- User mentions or implies: text to speech
- User mentions or implies: whisper
- User mentions or implies: elevenlabs
- User mentions or implies: deepgram
- User mentions or implies: realtime api
- User mentions or implies: voice assistant
- User mentions or implies: voice ai
- User mentions or implies: conversational ai
- User mentions or implies: tts
- User mentions or implies: stt
- User mentions or implies: asr
## Limitations
- Use this skill only when the task clearly matches the scope described above.
- Do not treat the output as a substitute for environment-specific validation, testing, or expert review.
- Stop and ask for clarification if required inputs, permissions, safety boundaries, or success criteria are missing.Related Skills
m365-agents-py
Microsoft 365 Agents SDK for Python. Build multichannel agents for Teams/M365/Copilot Studio with aiohttp hosting, AgentApplication routing, streaming responses, and MSAL-based auth.
m365-agents-dotnet
Microsoft 365 Agents SDK for .NET. Build multichannel agents for Teams/M365/Copilot Studio with ASP.NET Core hosting, AgentApplication routing, and MSAL-based auth.
voice-ai-engine-development
Build real-time conversational AI voice engines using async worker pipelines, streaming transcription, LLM agents, and TTS synthesis with interrupt handling and multi-provider support
voice-ai-development
Expert in building voice AI applications - from real-time voice agents to voice-enabled apps. Covers OpenAI Realtime API, Vapi for voice agents, Deepgram for transcription, ElevenLabs for synthesis, LiveKit for real-time infrastructure, and WebRTC fundamentals.
parallel-agents
Multi-agent orchestration patterns. Use when multiple independent tasks can run with different domain expertise or when comprehensive analysis requires multiple perspectives.
azure-ai-agents-persistent-java
Azure AI Agents Persistent SDK for Java. Low-level SDK for creating and managing AI agents with threads, messages, runs, and tools.
azure-ai-agents-persistent-dotnet
Azure AI Agents Persistent SDK for .NET. Low-level SDK for creating and managing AI agents with threads, messages, runs, and tools.
autonomous-agents
Autonomous agents are AI systems that can independently decompose goals, plan actions, execute tools, and self-correct without constant human guidance. The challenge isn't making them capable - it's making them reliable. Every extra decision multiplies failure probability.
ai-agents-architect
Expert in designing and building autonomous AI agents. Masters tool use, memory systems, planning strategies, and multi-agent orchestration.
agents-v2-py
Build container-based Foundry Agents with Azure AI Projects SDK (ImageBasedHostedAgentDefinition). Use when creating hosted agents with custom container images in Azure AI Foundry.
agents-md
This skill should be used when the user asks to "create AGENTS.md", "update AGENTS.md", "maintain agent docs", "set up CLAUDE.md", or needs to keep agent instructions concise. Enforces research-backed best practices for minimal, high-signal agent documentation.
nextjs-best-practices
Next.js App Router principles. Server Components, data fetching, routing patterns.