cookbook/livekit_agent_sdk/README.md
Simple example showing how to use LiveKit's xAI realtime plugin with LiteLLM as a proxy. This lets you switch between xAI, OpenAI, and Azure realtime APIs without changing your code.
pip install livekit-agents[xai] websockets
# With xAI
export XAI_API_KEY="your-xai-key"
litellm --config config.yaml --port 4000
python main.py
Type your message and get a voice response from Grok!
Set these environment variables if needed:
export LITELLM_PROXY_URL="http://localhost:4000"
export LITELLM_API_KEY="sk-1234"
export LITELLM_MODEL="grok-voice-agent"
Or use the defaults - connects to http://localhost:4000 by default.
Create a config.yaml with your realtime models:
model_list:
- model_name: grok-voice-agent
litellm_params:
model: xai/grok-2-vision-1212
api_key: os.environ/XAI_API_KEY
model_info:
mode: realtime
- model_name: openai-voice-agent
litellm_params:
model: gpt-4o-realtime-preview
api_key: os.environ/OPENAI_API_KEY
model_info:
mode: realtime
general_settings:
master_key: sk-1234
Then start: litellm --config config.yaml --port 4000
LiveKit's xAI plugin connects through LiteLLM proxy by setting base_url:
from livekit.plugins import xai
model = xai.realtime.RealtimeModel(
voice="ara",
api_key="sk-1234", # LiteLLM proxy key
base_url="http://localhost:4000", # Point to LiteLLM
)
Just change the model in your config - no code changes needed:
xAI Grok:
model: xai/grok-2-vision-1212
OpenAI:
model: gpt-4o-realtime-preview
Azure OpenAI:
model: azure/gpt-4o-realtime-preview
api_base: https://your-endpoint.openai.azure.com/