Skip to main content

Deepgram STT integration guide

How to use the Deepgram STT plugin for LiveKit Agents.

Available in
Python
|
Node.js

Overview

Deepgram provides advanced speech recognition technology and AI-driven audio processing solutions. Customizable speech models allow you to fine tune transcription performance for your specific use case. With LiveKit's Deepgram integration and the Agents framework, you can build AI agents that provide high-accuracy transcriptions.

Quick reference

This section includes a basic usage example and some reference material. For links to more detailed documentation, see Additional resources.

Installation

Install the plugin from PyPI:

pip install "livekit-agents[deepgram]~=1.2"

Authentication

The Deepgram plugin requires a Deepgram API key.

Set DEEPGRAM_API_KEY in your .env file.

Usage

Use Deepgram STT in an AgentSession or as a standalone transcription service. For example, you can use this STT in the Voice AI quickstart.

from livekit.plugins import deepgram
session = AgentSession(
stt = deepgram.STT(
model="nova-3",
),
# ... llm, tts, etc.
)

Parameters

This section describes the available parameters. See the plugin reference for more details.

modelstringOptionalDefault: nova-3

The Deepgram model to use for speech recognition.

languagestringOptionalDefault: en-US

The language code for recognition.

endpointing_msintOptionalDefault: 25

Time in milliseconds of silence to consider end of speech. Set to 0 to disable.

keytermslist[string]OptionalDefault: []

List of key terms to improve recognition accuracy. Supported by Nova-3 models.

Additional resources

The following resources provide more information about using Deepgram with LiveKit Agents.