Overview
This plugin allows you to use OVHCloud AI Endpoints as an STT provider for your voice agents.
Quick reference
This section includes a basic usage example and some reference material. For links to more detailed documentation, see Additional resources.
Installation
Install the OpenAI plugin to add OVHCloud AI Endpoints support:
uv add "livekit-agents[openai]~=1.3"
pnpm add @livekit/agents-plugin-openai@1.x
Authentication
The OVHCloud AI Endpoints plugin requires an API key. You can generate one by creating a new Public Cloud project, then navigate to AI Endpoints > API key.
Set OVHCLOUD_API_KEY in your .env file.
Usage
Use OVHCloud AI Endpoints STT in your AgentSession or as a standalone transcription service. For example, you can use this STT in the Voice AI quickstart.
from livekit.plugins import openaisession = AgentSession(stt=openai.STT.with_ovhcloud(model="whisper-large-v3-turbo",language="en",),# ... tts, llm, vad, turn_detection, etc.)
import * as openai from '@livekit/agents-plugin-openai';const session = new voice.AgentSession({stt: new openai.STT.withOVHcloud(model: "whisper-large-v3-turbo"),// ... tts, llm, vad, turn_detection, etc.});
Parameters
This section describes some of the available parameters. See the plugin reference links in the Additional resources section for a complete list of all available parameters.
stringOptionalDefault: whisper-large-v3-turboName of the STT model to use. For help with model selection, see the OVHCloud AI Endpoints STT documentation.
stringOptionalDefault: enLanguage of the input audio in ISO-639-1 format.
stringOptionalPrompt to guide the model's style or specify how to spell unfamiliar words. 224 tokens max.
Additional resources
The following resources provide more information about using OVHCloud AI Endpoints with LiveKit Agents.