Back to Langextract

Ollama Examples

examples/ollama/README.md

1.3.01.7 KB
Original Source

Ollama Examples

This directory contains examples for using LangExtract with Ollama for local LLM inference.

For setup instructions and documentation, see the main README's Ollama section.

Quick Reference

Option 1: Run locally

bash
# Install and start Ollama
ollama pull gemma2:2b
ollama serve  # Keep this running in a separate terminal

# Run the demo
python demo_ollama.py

Option 2: Run with Docker

bash
# Runs both Ollama and the demo in containers
docker-compose up

Files

  • demo_ollama.py - Comprehensive extraction examples demonstrating Ollama on README examples
  • docker-compose.yml - Production-ready Docker setup with health checks
  • Dockerfile - Container definition for LangExtract

Configuration Options

Timeout Settings

For slower models or large prompts, you may need to increase the timeout (default: 120 seconds):

python
import langextract as lx

result = lx.extract(
    text_or_documents=input_text,
    prompt_description=prompt,
    examples=examples,
    model_id="llama3.1:70b",  # Larger model may need more time
    timeout=300,  # 5 minutes
    model_url="http://localhost:11434",
)

Or using ModelConfig:

python
config = lx.factory.ModelConfig(
    model_id="llama3.1:70b",
    provider_kwargs={
        "model_url": "http://localhost:11434",
        "timeout": 300,  # 5 minutes
    }
)

Model License

Ollama models come with their own licenses. For example:

Please review the license for any model you use.