Overview
Deepgram provides advanced speech recognition technology and AI-driven audio processing solutions. Customizable speech models allow you to fine tune transcription performance for your specific use case. With LiveKit's Deepgram integration and the Agents framework, you can build AI agents that provide high-accuracy transcriptions.
Quick reference
This section includes a basic usage example and some reference material. For links to more detailed documentation, see Additional resources.
Installation
Install the plugin from PyPI:
pip install "livekit-agents[deepgram]~=1.2"
pnpm add @livekit/agents-plugin-deepgram@1.x
Authentication
The Deepgram plugin requires a Deepgram API key.
Set DEEPGRAM_API_KEY
in your .env
file.
Usage
Use Deepgram STT in an AgentSession
or as a standalone transcription service. For example, you can use this STT in the Voice AI quickstart.
from livekit.plugins import deepgramsession = AgentSession(stt = deepgram.STT(model="nova-3",),# ... llm, tts, etc.)
import * as deepgram from '@livekit/agents-plugin-deepgram';const session = new voice.AgentSession({stt: new deepgram.STT(model: "nova-3"),// ... llm, tts, etc.});
Parameters
This section describes the available parameters. See the plugin reference links in the Additional resources section for more details.
The Deepgram model to use for speech recognition.
The language code for recognition.
Time in milliseconds of silence to consider end of speech. Set to 0 to disable.
List of key terms to improve recognition accuracy. Supported by Nova-3 models.
Additional resources
The following resources provide more information about using Deepgram with LiveKit Agents.