Back to Copilotkit

Showcase Platform

showcase/README.md

1.57.016.8 KB
Original Source

Showcase Platform

Per-framework demos of CopilotKit (LangGraph, CrewAI, Mastra, Claude Agent SDK, etc.). Each package is a Next.js frontend + agent backend bundled in a Docker image. Railway deploys those images from main on push. This README is the from-scratch setup for running the same stack locally.

Layout

showcase/
  bin/showcase                  # unified CLI — run showcase/bin/showcase <command> for help
  integrations/<slug>/          # one per framework (17 total) — Dockerfile, src/app/demos/*/, src/agents/ or equivalent
  shell/                        # hub: home page, /matrix, canonical /integrations/[slug]/[demo]/{preview,code}
  shell-dashboard/              # internal-only feature × integration grid (port 3002)
  shared/
    feature-registry.json       # canonical features + categories (feeds the grid rows)
    constraints.yaml            # allowlist for which demos a package can expose
    local-ports.json            # deterministic host ports per package for local Docker runs
    python/ typescript/tools/   # shared agent utility code; CI stages these into each build context
  scripts/
    dev-local.sh                # low-level Docker Compose wrapper (prefer bin/showcase)
    cli/                        # command modules for bin/showcase
    generate-registry.ts        # builds shell/src/data/registry.json from all manifest.yaml
    bundle-demo-content.ts      # bundles per-demo source + README into shell/src/data/demo-content.json
  docker-compose.local.yml      # one service per package; ports from local-ports.json; env from .env
  .env.example                  # commit template — copy to .env and fill in

Generated data files

The shell apps consume JSON data files that are generated at build time by scripts in scripts/. These files are gitignored — every build path (Docker, CI, npm run build, npm run dev) regenerates them automatically.

FileGeneratorShell appsWhat it does
registry.jsongenerate-registry.tsshell, shell-docs, shell-dojo, shell-dashboardIntegration manifest — scans integrations/*/manifest.yaml, builds the full catalog with metadata, feature flags, categories
demo-content.jsonbundle-demo-content.tsshell, shell-docs, shell-dojoBundled source code from every demo directory — powers the Code tab, Snippet components, dojo cell viewer
constraints.jsongenerate-registry.tsshellFilter facets for the integration explorer (categories, frameworks, features)
search-index.jsongenerate-search-index.tsshell, shell-docsCmd-K search entries — scans MDX docs, AG-UI content, and registry data
starter-content.jsonbundle-starter-content.tsshellStarter template source bundles for the "Get Started" code viewer
docs-status.jsonprobe-docs.tsshell-dashboardPer-feature docs reachability — HTTP HEAD on og_docs_url, file-exists check on shell-docs MDX

Each generator writes to the src/data/ directory of every shell app that consumes it. Shell apps are independent — no shell cross-imports another shell's data directory.

If you add a new shell app that needs registry or demo data, add its output path to the relevant generator script in scripts/.

Prerequisites

  • macOS or Linux
  • Homebrew
  • Docker engine. Any of: Docker Desktop, Colima (recommended, no GUI / no sign-in), or OrbStack.
  • Node 22+ and npm (for shell / shell-dashboard dev servers — they're not in the compose)

Colima install (one time)

sh
brew install colima docker docker-buildx docker-compose

# Tell the docker CLI where its plugins live
mkdir -p ~/.docker
cat > ~/.docker/config.json <<'JSON'
{
  "cliPluginsExtraDirs": ["/opt/homebrew/lib/docker/cli-plugins"]
}
JSON

# Start the engine (adjust resources to taste; needed for building 17 images)
colima start --cpu 4 --memory 8 --disk 60

# Verify
docker compose version

Colima auto-starts with brew services start colima if you want it on login.

API keys

One .env file feeds every container. Not committed.

sh
cp showcase/.env.example showcase/.env
# Edit showcase/.env and fill in:
#   OPENAI_API_KEY=<required>
#   ANTHROPIC_API_KEY=<optional; needed for Claude Agent SDK demos and a few others>
#   LANGSMITH_API_KEY=<optional; enables LangSmith tracing for LangGraph demos>

Only OPENAI_API_KEY is strictly required. Missing optional keys fail gracefully (per-package).

Local CLI — parity testing without Railway

The showcase harness includes a CLI that runs the same probe drivers (liveness, D4, D5) used in CI/Railway against local Docker Compose containers. This lets you debug Dn failures, iterate on new demos, or run background parity checks without pushing to a branch.

Quick start

sh
# runs from any directory — all paths are resolved relative to the script itself

# 1. Make sure Docker is running (Colima, Docker Desktop, or OrbStack)

# 2. Start infra + one integration
./showcase/bin/showcase up langgraph-python

# 3. Run D5 probes
./showcase/bin/showcase test langgraph-python --d5

# 4. Run with verbose output
./showcase/bin/showcase test langgraph-python --d5 --verbose

# 5. Tear down when done
./showcase/bin/showcase down

Commands

CommandDescription
test <slug>Run probes against a running service
up [slugs...]Start infra (aimock, pocketbase, dashboard) + named packages. No args = infra only
down [slugs...]Stop services. No args = stop everything
build [slugs...]Build Docker images
psShow running services
portsPrint slug to host port mapping
logs <slug>Follow container logs (supports --grep, --since, -n, --no-follow)

Debugging commands (see DEBUGGING.md for full details):

CommandDescription
aimock-rebuildRebuild aimock from local source
recreate <slug>Force-recreate a service (picks up new image)
fixtures validateValidate fixture JSON files for common errors
doctorCheck local environment and stack health
diff-logs <slug>Show log delta for a time window

Test options

OptionDescription
--d5Run D5 (subagents/tool-rendering/agentic-chat) probes only
--d6Run D6 probes only
--verboseVerbose test output
--cycleOn failure, auto-dump aimock logs from the test window
--timeout <ms>Test timeout in milliseconds (default: 30000)

--d5 and --d6 are mutually exclusive. When neither is given, all tests run.

How it works

The CLI reuses the same probe drivers that the showcase-harness service runs on Railway. The difference is infrastructure: Railway probes hit Railway-deployed containers; the CLI probes hit local Docker Compose containers. Same assertions, same aimock fixtures, same Playwright flows.

┌───────────────┐     ┌──────────────────┐     ┌─────────────────────┐
│  CLI          │────▸│  Probe Drivers   │────▸│  Docker Compose     │
│  bin/showcase │     │  d5/d6           │     │  containers         │
└───────────────┘     └──────────────────┘     │  (aimock + integs)  │
                                               └─────────────────────┘

Use cases

  1. Debugging Dn failures — reproduce a CI/Railway failure locally with --d5 --verbose --cycle to see what aimock matched
  2. Building new demosup the integration, iterate on code, test <slug> --d5 to verify, repeat
  3. Background automation — run test <slug> --d5 as a pre-push sanity check

Local Docker workflow

bin/showcase is the unified CLI for local development. It wraps Docker Compose and adds debugging commands.

sh
# from the repo root

# inspect
showcase/bin/showcase ports              # slug → host port
showcase/bin/showcase ps                 # what's running

# build + start
showcase/bin/showcase build langgraph-python
showcase/bin/showcase up langgraph-python

# start everything (17 containers, heavy)
showcase/bin/showcase up

# follow logs (with optional grep filtering)
showcase/bin/showcase logs langgraph-python
showcase/bin/showcase logs aimock --grep "fixture|match"

# stop
showcase/bin/showcase down langgraph-python
showcase/bin/showcase down               # all

Each container exposes port 10000 internally and is mapped to the host port in shared/local-ports.json. The image and entrypoint are the same ones Railway runs.

For debugging workflows (aimock rebuild cycles, fixture validation, probe testing, diagnostics), see DEBUGGING.md.

Note: scripts/dev-local.sh still works for the core commands (up, down, build, logs, ps, ports). bin/showcase wraps the same logic and adds the debugging commands.

Hooking the local containers into the shell

The shell app's /preview route iframes integration.backend_url (Railway) by default. Set SHOWCASE_LOCAL=1 when running shell to swap in the localhost ports from local-ports.json instead — per-slug, falling back to Railway for anything you don't have running.

sh
cd showcase/shell
npm install                     # once
SHOWCASE_LOCAL=1 npm run dev    # now /preview iframes http://localhost:<port>/demos/...

In production the env var is unset → Railway URLs, unchanged.

shell-dashboard — feature × integration matrix

Internal overview of which packages support which features, linking to the canonical /preview and /code routes on shell. Lives at http://localhost:3002 and reads the same registry.json shell does.

sh
cd showcase/shell-dashboard
npm install
npm run dev

Column ordering lives in shell-dashboard/src/lib/sort-order.ts — internal to this app, not part of the public registry.

Iterating on a demo

  1. Edit the demo in integrations/<slug>/src/app/demos/<demo-id>/page.tsx (and the backend under src/agents/ if applicable).
  2. Rebundle so /code in shell reflects the edit: cd showcase && npx tsx scripts/bundle-demo-content.ts.
  3. If you changed manifest.yaml or added a feature to shared/feature-registry.json: npx tsx scripts/generate-registry.ts.
  4. Rebuild + restart the container: showcase/bin/showcase up <slug>.
  5. The grid in shell-dashboard and /preview in shell now show the new state.

Debugging

For the full debugging playbook — including aimock rebuild cycles, fixture validation, probe testing, container diagnostics, and common gotchas — see DEBUGGING.md.

Quick start:

sh
showcase/bin/showcase doctor              # check your stack
showcase/bin/showcase test mastra --d5    # run D5 probes
showcase/bin/showcase aimock-rebuild      # rebuild aimock from local source

Relationship to Railway

  • Dockerfile, entrypoint.sh, and build context (shared_python/, shared_typescript/) are shared between local and Railway.
  • .github/workflows/showcase_deploy.yml builds each image on push to main and pushes it to Railway. Per-PR deploys are opt-in via gh workflow run showcase_deploy.yml -r <branch> -f service=<slug>.
  • The only real differences at runtime are env var values and the URL. If something works locally in Docker, it works on Railway (and vice versa).

Updating the Dashboard

The dashboard at showcase.copilotkit.ai reads two data sources:

  1. Static catalog.json — generated at build time by pnpm generate-registry. Contains the full 38-feature x 17-integration cell matrix with status (wired / stub / unshipped), parity tiers, and feature categories. Changes require a generator run + commit.
  2. Live PocketBase probe results — streamed via SSE. Probes discover demo routes automatically and update the dashboard in real time. No manual intervention needed for probe data.

Known limitation — PocketBase fetch cap: useLiveStatus.ts fetches status records with a hard INITIAL_CAP (currently 2000). PocketBase returns records in rowid (creation) order. If the total record count exceeds the cap, later-created dimensions (e.g. e2e:<slug>/<featureId> per-cell records from the 6-hourly e2e-demos probe) get silently truncated, causing the dashboard to show D2 instead of D4 across the board. If new probe types are added and the dashboard regresses to D2, raise INITIAL_CAP in shell-dashboard/src/hooks/useLiveStatus.ts. The correct long-term fix is dimension-scoped fetching or sort=-updated so the cap never silently drops functional records.

Key invariants:

  • Parity tiers are never manually set. They are computed by comparing each integration's wired feature set against the reference integration's.
  • The reference integration is auto-detected as the integration with the most wired features (ties broken alphabetically). No reference: true flag exists.
  • catalog.json is gitignored — the generator emits it into the shell apps' src/data/ directories, which are already in .gitignore.
  • The stub status means: feature declared in manifest, demo entry exists, but no route field. Today only langgraph-python/cli-start qualifies.

SOP 1: Wire a new demo on an existing integration

  1. Edit showcase/integrations/<slug>/manifest.yaml — add the feature to features[] and a corresponding demos[] entry with a route.
  2. Run pnpm generate-registry — updates registry.json AND catalog.json. The cell flips from unshipped to wired. Parity tiers auto-recompute.
  3. Commit the manifest + both generated files. PR, merge.
  4. CI rebuilds the package image + dashboard image. Railway auto-deploys both.
  5. Ops probes discover the new demo route and begin probing. Dashboard updates live via PocketBase SSE — no further action needed.

SOP 2: Code fix on an existing demo (no manifest change)

  1. Edit code under showcase/integrations/<slug>/src/....
  2. PR, merge. No generator run needed (manifest unchanged).
  3. CI rebuilds the package image. Railway auto-deploys.
  4. Probes re-probe on the next tick. If the fix turns a red cell green, the dashboard updates live. Zero manual steps beyond the normal PR workflow.

SOP 3: Add a brand-new integration

  1. Create showcase/integrations/<new-slug>/manifest.yaml with features[] + demos[].
  2. Add {"slug": "<new-slug>", "name": "<Display Name>"} to showcase/shared/packages.json.
  3. Provision a Railway service (manual: railway service create or Dashboard UI).
  4. Run pnpm generate-registry — catalog gains 38 new cells (mostly unshipped, some wired). Parity tier computed automatically.
  5. Commit, PR, merge. CI + Railway deploy. Probes discover the new service automatically via the Railway discovery filter.

SOP 4: Reference migration (move the reference integration)

  1. No manual flag needed — the generator auto-detects the reference as the integration with the most wired features (ties broken alphabetically).
  2. If you want a different integration to be reference, wire more features on it until it leads the count.
  3. Run pnpm generate-registry — all parity tiers recompute automatically.
  4. Commit, PR, merge.