Deepgram Flux now available in Telnyx Voice AI

6, Nov 2025

Telnyx now hosts Deepgram’s Flux model inside its global network, enabling sub-second conversational speech recognition. Running Flux at the edge removes public cloud hops, reducing latency for real-time AI voice agents.

What’s new

  • Flux integration: Use Deepgram’s Conversational Speech Recognition model within Telnyx Voice AI.
  • Edge deployment: Model hosted in Telnyx PoPs alongside telephony and GPU inference layers.
  • Low-latency processing: Average end-of-turn detection in ~260 ms.
  • Unified pipeline: Handles both transcription and turn detection—no separate VAD setup.
  • API access: Available via Mission Control and Telnyx AI Assistants API.

Why it matters

  • Reduces speech-to-response latency by 100–300 ms compared to cloud setups.
  • Improves barge-in detection for more natural conversations.
  • Simplifies configuration by consolidating transcription and endpoint logic.
  • Increases reliability by keeping audio within Telnyx’s private backbone.

Example use cases

  • Real-time customer support agents with human-like timing.
  • Automated appointment scheduling and call routing.
  • Interactive voice assistants for logistics or healthcare.

Getting started

  1. Log in to Mission Control → AI → Transcription Models.
  2. Select Deepgram Flux or set via API:
"transcription": {
  "model": "deepgram/flux",
  "language": "en",
  "settings": {
    "eot_threshold": 0.7,
    "eot_timeout_ms": 5000
  }
}
  1. Test live conversations through AI Assistants or the Voice AI API.

Learn more in our developer documentation or contact your Telnyx team.

Deepgram Flux Added to Telnyx Voice AI for Sub-Second Speech