Skip to main content

OVHCloud STT plugin guide

How to use the OVHCloud STT plugin for LiveKit Agents.

Available in
Python
|
Node.js

Overview

This plugin allows you to use OVHCloud AI Endpoints as an STT provider for your voice agents.

Quick reference

This section includes a basic usage example and some reference material. For links to more detailed documentation, see Additional resources.

Installation

Install the OpenAI plugin to add OVHCloud AI Endpoints support:

uv add "livekit-agents[openai]~=1.3"
pnpm add @livekit/agents-plugin-openai@1.x

Authentication

The OVHCloud AI Endpoints plugin requires an API key. You can generate one by creating a new Public Cloud project, then navigate to AI Endpoints > API key.

Set OVHCLOUD_API_KEY in your .env file.

Usage

Use OVHCloud AI Endpoints STT in your AgentSession or as a standalone transcription service. For example, you can use this STT in the Voice AI quickstart.

from livekit.plugins import openai
session = AgentSession(
stt=openai.STT.with_ovhcloud(
model="whisper-large-v3-turbo",
language="en",
),
# ... tts, llm, vad, turn_detection, etc.
)
import * as openai from '@livekit/agents-plugin-openai';
const session = new voice.AgentSession({
stt: new openai.STT.withOVHcloud(
model: "whisper-large-v3-turbo"
),
// ... tts, llm, vad, turn_detection, etc.
});

Parameters

This section describes some of the available parameters. See the plugin reference links in the Additional resources section for a complete list of all available parameters.

modelstringOptionalDefault: whisper-large-v3-turbo

Name of the STT model to use. For help with model selection, see the OVHCloud AI Endpoints STT documentation.

languagestringOptionalDefault: en

Language of the input audio in ISO-639-1 format.

promptstringOptional

Prompt to guide the model's style or specify how to spell unfamiliar words. 224 tokens max.

Additional resources

The following resources provide more information about using OVHCloud AI Endpoints with LiveKit Agents.