Back to Ai

AI SDK - Open Responses Provider

packages/open-responses/README.md

2.1.101.6 KB
Original Source

AI SDK - Open Responses Provider

The Open Responses provider for the AI SDK contains language model support for Open Responses compatible APIs.

Setup

The Open Responses provider is available in the @ai-sdk/open-responses module. You can install it with

bash
npm i @ai-sdk/open-responses

Skill for Coding Agents

If you use coding agents such as Claude Code or Cursor, we highly recommend adding the AI SDK skill to your repository:

shell
npx skills add vercel/ai

Provider Instance

Create an Open Responses provider instance using createOpenResponses:

ts
import { createOpenResponses } from '@ai-sdk/open-responses';

const openResponses = createOpenResponses({
  name: 'aProvider',
  url: 'http://localhost:1234/v1/responses',
});

You can use this instance to access models served by any Open Responses compatible endpoint.

Example

ts
import { createOpenResponses } from '@ai-sdk/open-responses';
import { generateText } from 'ai';

const openResponses = createOpenResponses({
  name: 'aProvider',
  url: 'http://localhost:1234/v1/responses',
});

const { text } = await generateText({
  model: openResponses('mistralai/ministral-3-14b-reasoning'),
  prompt: 'Invent a new holiday and describe its traditions.',
  maxOutputTokens: 100,
});

Documentation

Please check out the Open Responses provider documentation for more information.