Docs
Models
Mix and match voices, transcription, and language models for every agent. All models are invocable from the API and dashboard.
Voice (text-to-speech)
Natural-sounding voices optimized for conversational latency. Clone your own or pick from the library.
Model ID
Languages
Latency
Best for
simba-natural-v2Languages: 70+
70+
Latency: < 300 ms
< 300 ms
General-purpose conversational agents
simba-expressive-v2Languages: 32
32
Latency: < 350 ms
< 350 ms
High-emotion outbound, marketing, entertainment
simba-fast-v1Languages: 29
29
Latency: < 180 ms
< 180 ms
Low-latency streaming, edge deployments
simba-clone-v2Languages: 70+
70+
Latency: < 320 ms
< 320 ms
Custom voice clones from 30 seconds of audio
Speech-to-text
Streaming and batch transcription tuned for telephony and noisy environments.
Model ID
Languages
Latency
Best for
simba-transcribe-v1Languages: 92
92
Latency: streaming
streaming
Real-time conversation transcripts
simba-transcribe-batch-v1Languages: 92
92
Latency: batch
batch
Post-call analytics, bulk processing
Language (LLM) — bring your own
SIMBA routes to the LLM of your choice. Use hosted providers or connect your own.
Model ID
Languages
Latency
Best for
gpt-4oLanguages: any
any
Latency: ~ 400 ms
~ 400 ms
Highest reasoning, premium pricing
gpt-4o-miniLanguages: any
any
Latency: ~ 220 ms
~ 220 ms
Balanced latency and cost
claude-sonnet-4-6Languages: any
any
Latency: ~ 300 ms
~ 300 ms
Complex instructions, long context
claude-haiku-4-5Languages: any
any
Latency: ~ 180 ms
~ 180 ms
High-throughput, low-latency workflows
gemini-2.5-flashLanguages: any
any
Latency: ~ 200 ms
~ 200 ms
Cost-sensitive production traffic
self-hostedLanguages: any
any
Latency: varies
varies
Compliance or data-residency requirements
Pick a model from the API
await simba.agents.create({
name: "Support agent",
conversationConfig: {
agent: { prompt: { prompt: "..." }, llm: "gpt-4o-mini" },
tts: { voice_id: "simba-natural-v2" },
asr: { model: "simba-transcribe-v1" },
},
});See the API reference for the full config schema.