Skip to main content
Telnyx hosts models with an OpenAI-compatible API. No concurrency limits. Use the standard OpenAI plugin with the .with_telnyx() helper:
from livekit.plugins import openai

llm = openai.LLM.with_telnyx(model="moonshotai/Kimi-K2.6")

About .with_telnyx()

This is a built-in static method on openai.LLM in the official livekit-plugins-openai package, maintained by LiveKit — not a Telnyx package or fork. It works the same way as the other OpenAI-compatible helpers in that package (.with_azure(), .with_fireworks(), etc.): it sets base_url to Telnyx’s OpenAI-compatible inference endpoint (https://api.telnyx.com/v2/ai/openai) and reads your TELNYX_API_KEY from the environment. You don’t need any additional packages beyond livekit-plugins-openai.

Hosted models

These run on Telnyx infrastructure — no external API key needed, just your TELNYX_API_KEY:
ModelDescription
moonshotai/Kimi-K2.6Moonshot AI — voice AI, with thinking disabled (Recommended)
zai-org/GLM-5.1-FP8Zhipu AI — most efficient reasoning, function calling
MiniMaxAI/MiniMax-M2.7MiniMax — cheapest, high intelligence

Proprietary models (BYOK)

For models like GPT-4o or Claude, Telnyx proxies the request using your own API key. Add your provider key in the Telnyx Portal under Inference settings.
# Proprietary model via BYOK (bring your own key)
llm = openai.LLM.with_telnyx(model="openai/gpt-4o-mini")
Full models list →