Cerebras Provider

The Cerebras provider offers access to powerful language models through the Cerebras API, including their high-speed inference capabilities powered by Wafer-Scale Engines and CS-3 systems.

API keys can be obtained from the Cerebras Platform.

Setup

The Cerebras provider is available via the @ai-sdk/cerebras module. You can install it with:

pnpm
npm
yarn
pnpm add @ai-sdk/cerebras

Provider Instance

You can import the default provider instance cerebras from @ai-sdk/cerebras:

import { cerebras } from '@ai-sdk/cerebras';

For custom configuration, you can import createCerebras and create a provider instance with your settings:

import { createCerebras } from '@ai-sdk/cerebras';
const cerebras = createCerebras({
apiKey: process.env.CEREBRAS_API_KEY ?? '',
});

You can use the following optional settings to customize the Cerebras provider instance:

  • baseURL string

    Use a different URL prefix for API calls. The default prefix is https://api.cerebras.ai/v1.

  • apiKey string

    API key that is being sent using the Authorization header. It defaults to the CEREBRAS_API_KEY environment variable.

  • headers Record<string,string>

    Custom headers to include in the requests.

  • fetch (input: RequestInfo, init?: RequestInit) => Promise<Response>

    Custom fetch implementation.

Language Models

You can create language models using a provider instance:

import { cerebras } from '@ai-sdk/cerebras';
import { generateText } from 'ai';
const { text } = await generateText({
model: cerebras('llama3.1-8b'),
prompt: 'Write a vegetarian lasagna recipe for 4 people.',
});

Cerebras language models can be used in the streamText and streamUI functions (see AI SDK Core and AI SDK RSC).

Model Capabilities

ModelImage InputObject GenerationTool UsageTool Streaming
llama3.1-8b
llama3.1-70b
llama3.3-70b

Please see the Cerebras docs for more details about the available models. Note that context windows are temporarily limited to 8192 tokens in the Free Tier.