Back to Ai

AI SDK - Cerebras Provider

packages/cerebras/README.md

2.1.101.5 KB
Original Source

AI SDK - Cerebras Provider

The Cerebras provider for the AI SDK contains language model support for Cerebras, offering high-speed AI model inference powered by Cerebras Wafer-Scale Engines and CS-3 systems.

Setup

The Cerebras provider is available in the @ai-sdk/cerebras module. You can install it with

bash
npm i @ai-sdk/cerebras

Skill for Coding Agents

If you use coding agents such as Claude Code or Cursor, we highly recommend adding the AI SDK skill to your repository:

shell
npx skills add vercel/ai

Provider Instance

You can import the default provider instance cerebras from @ai-sdk/cerebras:

ts
import { cerebras } from '@ai-sdk/cerebras';

Available Models

Cerebras offers a variety of high-performance language models: https://inference-docs.cerebras.ai/models/overview

Example

ts
import { cerebras } from '@ai-sdk/cerebras';
import { generateText } from 'ai';

const { text } = await generateText({
  model: cerebras('llama-3.3-70b'),
  prompt: 'Write a JavaScript function that sorts a list:',
});

Documentation

For more information about Cerebras' high-speed inference capabilities and API documentation, please visit:

Note: Due to high demand in the early launch phase, context windows are temporarily limited to 8192 tokens in the Free Tier.