Back to Copilotkit

CopilotKit <> PydanticAI AG-UI Canvas Starter

examples/canvas/pydantic-ai/README.md

1.57.011.1 KB
Original Source

CopilotKit <> PydanticAI AG-UI Canvas Starter

This is a starter template for building AI-powered canvas applications using PydanticAI and CopilotKit. It provides a modern Next.js application with an integrated PydanticAI agent that manages a visual canvas of interactive cards with real-time AI synchronization.

https://github.com/user-attachments/assets/2a4ec718-b83b-4968-9cbe-7c1fe082e958

🚀 Key Features

  • Visual Canvas Interface: Drag-free canvas displaying cards in a responsive grid layout
  • Four Card Types:
    • Project: Includes text fields, dropdown, date picker, and checklist
    • Entity: Features text fields, dropdown, and multi-select tags
    • Note: Simple rich text content area
    • Chart: Visual metrics with percentage-based bar charts
  • Real-time AI Sync: Bidirectional synchronization between the AI agent and UI canvas
  • Multi-step Planning: AI can create and execute plans with visual progress tracking
  • Human-in-the-Loop (HITL): Intelligent interrupts for clarification when needed
  • JSON View: Toggle between visual canvas and raw JSON state
  • Responsive Design: Optimized for both desktop (sidebar chat) and mobile (popup chat)

Prerequisites

  • Node.js 18+
  • Python 3.8+
  • OpenAI API Key (for the PydanticAI agent)
  • Any of the following package managers:

Note: This repository ignores lock files (package-lock.json, yarn.lock, pnpm-lock.yaml, bun.lockb) to avoid conflicts between different package managers. Each developer should generate their own lock file using their preferred package manager. After that, make sure to delete it from the .gitignore.

Getting Started

  1. Install dependencies using your preferred package manager:
bash
# Using pnpm (recommended)
pnpm install

# Using npm
npm install

# Using yarn
yarn install

# Using bun
bun install
  1. Install Python dependencies for the PydanticAI agent:
bash
# Using pnpm
pnpm install:agent

# Using npm
npm run install:agent

# Using yarn
yarn install:agent

# Using bun
bun run install:agent

Note: This will automatically setup a .venv (virtual environment) inside the agent directory.

To activate the virtual environment manually, you can run:

bash
source agent/.venv/bin/activate
  1. Set up your OpenAI API key:
bash
export OPENAI_API_KEY="your-openai-api-key-here"
  1. Start the development server:
bash
# Using pnpm
pnpm dev

# Using npm
npm run dev

# Using yarn
yarn dev

# Using bun
bun run dev

This will start both the UI and agent servers concurrently.

Getting Started with the Canvas

Once the application is running, you can:

  1. Create Cards: Use the "New Item" button or ask the AI to create cards

    • "Create a new project"
    • "Add an entity and a note"
    • "Create a chart with sample metrics"
  2. Edit Cards: Click on any field to edit directly, or ask the AI

    • "Set the project field1 to 'Q1 Planning'"
    • "Add a checklist item 'Review budget'"
    • "Update the chart metrics"
  3. Execute Plans: Give the AI multi-step instructions

    • "Create 3 projects with different priorities and add 2 checklist items to each"
    • The AI will create a plan and execute it step by step with visual progress
  4. View JSON: Toggle between the visual canvas and JSON view using the button at the bottom

Available Scripts

The following scripts can also be run using your preferred package manager:

  • dev - Starts both UI and agent servers in development mode
  • dev:debug - Starts development servers with debug logging enabled
  • dev:ui - Starts only the Next.js UI server
  • dev:agent - Starts only the PydanticAI agent server
  • build - Builds the Next.js application for production
  • start - Starts the production server
  • lint - Runs ESLint for code linting
  • install:agent - Installs Python dependencies for the agent

Architecture Overview

mermaid
graph TB
    subgraph "Frontend (Next.js)"
        UI[Canvas UI
page.tsx]
        Actions[Frontend Actions
useCopilotAction]
        State[State Management
useCoAgent]
        Chat[CopilotChat]
    end

    subgraph "Backend (Python)"
        Agent[PydanticAI Agent
agent.py]
        Tools[Backend Tools
- set_plan
- update_plan_progress
- complete_plan]
        AgentState[Canvas State
StateDeps]
        Model[LLM
GPT-4o]
    end

    subgraph "Communication"
        Runtime[CopilotKit Runtime
:8000]
    end

    UI <--> State
    State <--> Runtime
    Chat <--> Runtime
    Actions --> Runtime
    Runtime <--> Agent
    Agent --> Tools
    Agent --> AgentState
    Agent --> Model

    style UI fill:#e1f5fe
    style Agent fill:#fff3e0
    style Runtime fill:#f3e5f5

    click UI "https://github.com/CopilotKit/CopilotKit/blob/main/examples/canvas/pydantic-ai/src/app/page.tsx"
    click Agent "https://github.com/CopilotKit/CopilotKit/blob/main/examples/canvas/pydantic-ai/agent/agent.py"

Frontend (Next.js + CopilotKit)

The main UI component is in src/app/page.tsx. It includes:

  • Canvas Management: Visual grid of cards with create, read, update, and delete operations
  • State Synchronization: Uses useCoAgent hook for real-time state sync with the agent
  • Frontend Actions: Exposed as tools to the AI agent via useCopilotAction
  • Plan Visualization: Shows multi-step plan execution with progress indicators
  • HITL (Tool-based): Uses useCopilotAction with renderAndWaitForResponse for disambiguation prompts (e.g., choosing an item or card type)

Backend (PydanticAI Agent)

The agent logic is in agent/agent.py. It features:

  • State Management: Uses StateDeps[CanvasState] for typed state management
  • Tool Integration: Backend tools decorated with @agent.tool for planning and state updates
  • Strict Grounding: Enforces data consistency by always using shared state as truth
  • Dynamic Instructions: Uses @agent.instructions to provide context-aware guidance
  • AG-UI Integration: Converts to FastAPI app with agent.to_ag_ui() for seamless integration
  • Type Safety: Leverages Pydantic models for all data structures

Card Field Schema

Each card type has specific fields defined in the agent:

  • Project: field1 (text), field2 (select), field3 (date), field4 (checklist)
  • Entity: field1 (text), field2 (select), field3 (tags), field3_options (available tags)
  • Note: field1 (textarea content)
  • Chart: field1 (array of metrics with label and value 0-100)

Data Flow

mermaid
sequenceDiagram
    participant User
    participant UI as Canvas UI
    participant CK as CopilotKit
    participant Agent as PydanticAI Agent
    participant Tools

    User->>UI: Interact with canvas
    UI->>CK: Update state via useCoAgent
    CK->>Agent: Send state + message
    Agent->>Agent: Process with GPT-4o
    Agent->>Tools: Execute tools
    Tools-->>Agent: Return results
    Agent->>CK: Return updated state
    CK->>UI: Sync state changes
    UI->>User: Display updates

    Note over Agent: Maintains ground truth
    Note over UI,CK: Real-time bidirectional sync

Customization Guide

Adding New Card Types

  1. Define the data schema in src/lib/canvas/types.ts
  2. Add the card type to the CardType union
  3. Create rendering logic in src/components/canvas/CardRenderer.tsx
  4. Update the agent's Pydantic models in agent/agent.py
  5. Add corresponding frontend actions in src/app/page.tsx

Modifying Existing Cards

  • Field definitions are in the agent's Pydantic models (e.g., ProjectData, EntityData)
  • UI components are in CardRenderer.tsx
  • Frontend actions follow the pattern: set[Type]Field[Number]

Extending the Agent

PydanticAI makes it easy to extend the agent with:

  • New Tools: Add functions decorated with @agent.tool
  • Custom Instructions: Modify the @agent.instructions function
  • State Extensions: Add fields to the CanvasState model
  • Type Safety: All changes benefit from Pydantic's type validation

Styling

  • Global styles: src/app/globals.css
  • Component styles use Tailwind CSS with shadcn/ui components
  • Theme colors can be modified via CSS custom properties

📚 Documentation

Contributing

Feel free to submit issues and enhancement requests! This starter is designed to be easily extensible.

License

This project is licensed under the MIT License - see the LICENSE file for details.

Troubleshooting

Agent Connection Issues

If you see "I'm having trouble connecting to my tools", make sure:

  1. The PydanticAI agent is running on port 8000 (check terminal output)
  2. Your OpenAI API key is set correctly as an environment variable
  3. Both servers started successfully (UI and agent)

Port Already in Use

If you see "[Errno 48] Address already in use":

  1. The agent might still be running from a previous session
  2. Kill the process using the port: lsof -ti:8000 | xargs kill -9
  3. For the UI port: lsof -ti:3000 | xargs kill -9

State Synchronization Issues

If the canvas and AI seem out of sync:

  1. Check the browser console for errors
  2. Ensure all frontend actions are properly registered
  3. Verify the agent is using the latest shared state (not cached values)

Python Dependencies

If you encounter Python import errors:

bash
cd agent
pip install -r requirements.txt

Virtual Environment Issues

If the virtual environment is not activated properly:

bash
cd agent
source .venv/bin/activate  # On macOS/Linux
# or
.venv\Scripts\activate  # On Windows

Dependency Conflicts

If issues persist, recreate the virtual environment:

bash
cd agent
rm -rf .venv
python -m venv .venv
.venv/bin/pip install --upgrade pip
.venv/bin/pip install -r requirements.txt

[!IMPORTANT] Some features are still under active development and may not yet work as expected. If you encounter a problem using this template, please report an issue to this repository.