llama-index-integrations/llms/llama-index-llms-asi/README.md
This package contains the LlamaIndex integration with ASI-1 Mini, a powerful language model designed for various natural language processing tasks.
ASI-1 Mini is the world's first Web3-native Large Language Model (LLM) developed by Fetch.ai Inc., a founding member of the Artificial Superintelligence Alliance. Unlike general-purpose LLMs, ASI-1 Mini is specifically designed and optimized for supporting complex agentic workflows.
With ASI-1 Mini, you can leverage these powerful capabilities:
Want to learn more about ASI? Visit the ASI website or Fetch.ai for more information!
pip install llama-index-llms-asi
Here's an example of how to use the ASI integration with LlamaIndex:
from llama_index.llms.asi import ASI
# Initialize the ASI LLM
llm = ASI(model="asi1-mini", api_key="your_api_key")
# Generate text
response = llm.complete("Tell me about artificial intelligence.")
print(response)
# Chat completion
from llama_index.core.llms import ChatMessage, MessageRole
messages = [
ChatMessage(
role=MessageRole.SYSTEM, content="You are a helpful AI assistant."
),
ChatMessage(
role=MessageRole.USER, content="Tell me about artificial intelligence."
),
]
response = llm.chat(messages)
print(response)
The ASI integration has different streaming implementations for completion and chat:
Streaming Completion: ASI doesn't support streaming for completions (returns 404 error). Our implementation uses a fallback mechanism that returns the complete response as a single chunk.
Streaming Chat: ASI supports streaming for chat, but with a unique format that includes:
thought and init_thought that contain intermediate reasoningOur enhanced implementation:
# Streaming completion (falls back to regular completion)
for chunk in llm.stream_complete("Tell me about artificial intelligence."):
print(chunk.text, end="", flush=True)
# Streaming chat (handles ASI's unique streaming format)
for chunk in llm.stream_chat(messages):
if hasattr(chunk, "delta") and chunk.delta.strip():
print(chunk.delta, end="", flush=True)
The ASI integration also supports async operations:
# Async completion
response = await llm.acomplete("Tell me about artificial intelligence.")
print(response)
# Async chat
response = await llm.achat(messages)
print(response)
# Async streaming completion (falls back to regular completion)
async for chunk in await llm.astream_complete(
"Tell me about artificial intelligence."
):
print(chunk.text, end="", flush=True)
# Async streaming chat (handles ASI's unique streaming format)
# Note: This follows the same pattern as OpenAI's implementation
stream = await llm.astream_chat(messages)
async for chunk in stream:
print(chunk.delta, end="", flush=True)
You need an API key to use ASI's API. You can provide it in two ways:
ASI(api_key="your_api_key")export ASI_API_KEY="your_api_key"Currently, this integration supports the following models:
asi1-mini: A powerful language model for various natural language processing tasks.ASI is designed to be a drop-in replacement for OpenAI in LlamaIndex applications. Developers can easily switch from OpenAI to ASI by simply changing the LLM initialization:
# From this (OpenAI)
from llama_index.llms.openai import OpenAI
llm = OpenAI(api_key=openai_api_key, model="gpt-3.5-turbo")
# To this (ASI)
from llama_index.llms.asi import ASI
llm = ASI(api_key=asi_api_key, model="asi1-mini")
The rest of your code remains unchanged, including:
llm.complete()llm.chat()for chunk in llm.stream_chat()stream = await llm.astream_chat() followed by async for chunk in streamASI works seamlessly with other LlamaIndex components, such as using OpenAI embeddings for vector search while using ASI as the LLM for generating responses.
Here's a complete example showing how to use ASI as a drop-in replacement for OpenAI in a RAG application:
from llama_index.llms.asi import ASI
from llama_index.embeddings.openai import OpenAIEmbedding
from llama_index.core import VectorStoreIndex, SimpleDirectoryReader
from llama_index.core.node_parser import SentenceSplitter
# Load documents
documents = SimpleDirectoryReader("./data").load_data()
# Create parser
parser = SentenceSplitter(chunk_size=1024)
# Create embeddings model (still using OpenAI for embeddings)
embedding_model = OpenAIEmbedding(model="text-embedding-3-small")
# Create ASI LLM (instead of OpenAI)
llm = ASI(model="asi1-mini", api_key="your_asi_api_key")
# Create index
nodes = parser.get_nodes_from_documents(documents)
index = VectorStoreIndex(nodes, embed_model=embedding_model)
# Create query engine with ASI as the LLM
query_engine = index.as_query_engine(llm=llm)
# Query with streaming
response = query_engine.query(
"What information is in these documents?", streaming=True
)
# Process streaming response
for token in response.response_gen:
print(token, end="", flush=True)
This example demonstrates how you can use ASI as the LLM while still using OpenAI for embeddings, showing the flexibility of the integration.
To create a development environment, install poetry then run:
poetry install --with dev
To test the integration, first enter the poetry venv:
poetry shell
Then tests can be run with make
make test
Integration tests will be skipped unless an API key is provided. API keys can be obtained from the Fetch.ai team. Once created, store the API key in an environment variable and run tests
export ASI_API_KEY=<your key here>
make test
Linting and code formatting can be executed with make.
make format
make lint