Overview
Deepgram provides advanced speech recognition technology and AI-driven audio processing solutions. Customizable speech models allow you to fine tune transcription performance for your specific use case. With LiveKit's Deepgram integration and the Agents framework, you can build AI agents that provide high-accuracy transcriptions.
If you're looking to build an AI voice assistant with Deepgram, check out our Voice Agent Quickstart guide and use the Deepgram STT and/or TTS module as demonstrated below.
Quick reference
Environment variables
DEEPGRAM_API_KEY=<your-deepgram-api-key>
STT
LiveKit's Deepgram integration provides a speech-to-text (STT) interface that can be used as the first stage in a VoicePipelineAgent
or as a standalone transcription service. For a complete reference of all available parameters, see the plugin reference for Python or Node.
Usage
from livekit.plugins.deepgram import sttdeepgram_stt = deepgram.stt.STT(model="nova-2-general",interim_results=True,smart_format=True,punctuate=True,filler_words=True,profanity_filter=False,keywords=[("LiveKit", 1.5)],language="en-US",)
Parameters
ID of the model to use for inference. To learn more, see supported models.
Enable perliminary results before the final transcription is available.
Enable smart formatting to improve the readability of transcriptions.
Enable punctuation in transcriptions.
Enable filler words to improve turn detection.
Replace recognized profanity with asterisks in transcriptions.
A list of keywords and intensifiers to boost or suppress in transcriptions. Positive values boost; negative values suppress.
TTS
LiveKit's Deepgram integration also provides a text-to-speech (TTS) interface. This can be used in a VoicePipelineAgent
or as a standalone speech generator. For a complete reference of all available parameters, see the plugin reference.
Usage
from livekit.plugins.deepgram import ttsdeepgram_tts = tts.TTS(model="aura-asteria-en",)
Parameters
ID of the model to use for generation. To learn more, see supported models.