docs/examples/llm/nebius.ipynb
<a href="https://colab.research.google.com/github/run-llama/llama_index/blob/main/docs/examples/llm/nebius.ipynb" target="_parent"></a>
This notebook demonstrates how to use LLMs from Nebius AI Studio with LlamaIndex. Nebius AI Studio implements all state-of-the-art LLMs available for commercial use.
First, let's install LlamaIndex and dependencies of Nebius AI Studio.
%pip install llama-index-llms-nebius llama-index
Upload your Nebius AI Studio key from system variables below or simply insert it. You can get it by registering for free at Nebius AI Studio and issuing the key at API Keys section."
import os
NEBIUS_API_KEY = os.getenv("NEBIUS_API_KEY") # NEBIUS_API_KEY = ""
from llama_index.llms.nebius import NebiusLLM
llm = NebiusLLM(
api_key=NEBIUS_API_KEY, model="meta-llama/Llama-3.3-70B-Instruct-fast"
)
complete with a promptresponse = llm.complete("Amsterdam is the capital of ")
print(response)
chat with a list of messagesfrom llama_index.core.llms import ChatMessage
messages = [
ChatMessage(role="system", content="You are a helpful AI assistant."),
ChatMessage(
role="user",
content="Answer briefly: who is Wall-e?",
),
]
response = llm.chat(messages)
print(response)
stream_complete endpointresponse = llm.stream_complete("Amsterdam is the capital of ")
for r in response:
print(r.delta, end="")
stream_chat with a list of messagesfrom llama_index.core.llms import ChatMessage
messages = [
ChatMessage(role="system", content="You are a helpful AI assistant."),
ChatMessage(
role="user",
content="Answer briefly: who is Wall-e?",
),
]
response = llm.stream_chat(messages)
for r in response:
print(r.delta, end="")