Back to Copilotkit

QA: BYOC json-render — LangGraph (Python)

showcase/integrations/langgraph-python/qa/byoc-json-render.md

1.57.02.3 KB
Original Source

QA: BYOC json-render — LangGraph (Python)

Prerequisites

  • Demo deployed at /demos/byoc-json-render.
  • Railway backend healthy (showcase-langgraph-python-production.up.railway.app).
  • OPENAI_API_KEY and LANGGRAPH_DEPLOYMENT_URL configured on the Next.js app.
  • @json-render/core + @json-render/react present in package.json (pinned to 0.18.0).
  • byoc_json_render graph registered in langgraph.json.

Test Steps

1. Page load

  • Navigate to /demos/byoc-json-render.
  • Chat composer is visible.
  • Three suggestion pills appear with titles: "Sales dashboard", "Revenue by category", "Expense trend".
  • No console errors.

2. Sales dashboard suggestion

  • Click the "Sales dashboard" suggestion.
  • Within 60 seconds, a data-testid="json-render-root" wrapper appears in the assistant bubble.
  • A data-testid="metric-card" renders inside the wrapper.
  • A chart (data-testid="bar-chart" or data-testid="pie-chart") renders inside the wrapper.
  • No raw JSON text is shown once rendering finishes — the streaming JSON is replaced by components.

3. Revenue by category

  • Click the "Revenue by category" suggestion.
  • Within 60 seconds, a data-testid="pie-chart" renders with multiple category slices + legend.

4. Expense trend

  • Click the "Expense trend" suggestion.
  • Within 60 seconds, a data-testid="bar-chart" renders with month labels.

5. Free-form prompt

  • Type "Show me a metric for quarterly revenue" and send.
  • Verify at least one metric-card renders; no console errors.

6. Multi-turn

  • After a previous render is visible, send a follow-up prompt ("Now break that down by region").
  • A new assistant message appears with a new json-render rendering — prior renders stay in the transcript.

7. Malformed output handling

  • If the agent ever replies with non-JSON text (force it by asking "tell me a joke"), the chat falls back to rendering that raw text via the default assistant bubble. No crash, no stuck spinner.

Expected Results

  • Suggestion renders land within 60 seconds. Budget is slightly higher than Wave 4a because a JSON { root, elements } spec is more verbose than hashbrown's token stream.
  • No uncaught errors in the console.
  • Streaming falls back to plain text until the JSON parses, then swaps to rendered components.