Back to Opik

Observability for Cloudflare Workers AI with Opik

apps/opik-documentation/documentation/fern/docs/tracing/integrations/cloudflare-workers-ai.mdx

2.0.24-52622.5 KB
Original Source

Opik provides seamless integration with Cloudflare Workers AI allowing you to trace, monitor, and debug your Cloudflare Workers AI applications.

Features

  • Comprehensive Tracing: Automatically trace Cloudflare Workers AI requests
  • Hierarchical Visualization: View your Cloudflare Workers AI requests as structured traces with parent-child relationships
  • Custom Tagging: Add custom tags to organize and filter your traces

Installation

Option 1: Using npm

bash
npm install opik

Option 2: Using yarn

bash
yarn add opik

Requirements

  • Node.js ≥ 18
  • Opik SDK

Enable Node.js compatibility mode

The Opik SDK requires Node.js compatibility mode to be enabled in Cloudflare Workers AI. You can enable it by adding the following to your wrangler.toml file:

json
{
  "compatibility_flags": [
    "nodejs_compat"
  ],
  "compatibility_date": "2024-09-23"
}

See the Cloudflare documentation for more details.

Basic Usage

Using with LLM Calls

Here is an example of how to use the Opik SDK with Cloudflare Workers AI:

typescript
/**
 * Welcome to Cloudflare Workers! This is your first worker.
 *
 * - Run `npm run dev` in your terminal to start a development server
 * - Open a browser tab at http://localhost:8787/ to see your worker in action
 * - Run `npm run deploy` to publish your worker
 *
 * Bind resources to your worker in `wrangler.jsonc`. After adding bindings, a type definition for the
 * `Env` object can be regenerated with `npm run cf-typegen`.
 *
 * Learn more at https://developers.cloudflare.com/workers/
 */
import { Opik } from 'opik';

const client = new Opik({
	apiKey: '<API-KEY>',
	apiUrl: 'https://www.comet.com/opik/api',
	projectName: 'demo-cloudflare-workers-ai',
	workspaceName: '<YOUR-WORKSPACE>',
});

export default {
	async fetch(request, env) {
		const input = 'What is the origin of the phrase Hello, World';
		const response = await env.AI.run('@cf/meta/llama-3.1-8b-instruct', {
			prompt: input,
		});

		const someTrace = client.trace({
			name: `Trace`,
			input: {
				prompt: input,
			},
			output: {
				response: response,
			},
		});

		const someSpan = someTrace.span({
			name: `Span`,
			type: 'llm',
			input: {
				prompt: input,
			},
			output: {
				response: response,
			},
		});
		someSpan.end();

		someTrace.end();

		await client.flush();

		return new Response(JSON.stringify(response));
	},
} satisfies ExportedHandler<Env>;