Cerebras Provider
The Cerebras provider offers access to powerful language models through the Cerebras API, including their high-speed inference capabilities powered by Wafer-Scale Engines and CS-3 systems.
API keys can be obtained from the Cerebras Platform.
Setup
The Cerebras provider is available via the @ai-sdk/cerebras
module. You can install it with:
pnpm add @ai-sdk/cerebras
Provider Instance
You can import the default provider instance cerebras
from @ai-sdk/cerebras
:
import { cerebras } from '@ai-sdk/cerebras';
For custom configuration, you can import createCerebras
and create a provider instance with your settings:
import { createCerebras } from '@ai-sdk/cerebras';
const cerebras = createCerebras({ apiKey: process.env.CEREBRAS_API_KEY ?? '',});
You can use the following optional settings to customize the Cerebras provider instance:
-
baseURL string
Use a different URL prefix for API calls. The default prefix is
https://api.cerebras.ai/v1
. -
apiKey string
API key that is being sent using the
Authorization
header. It defaults to theCEREBRAS_API_KEY
environment variable. -
headers Record<string,string>
Custom headers to include in the requests.
-
fetch (input: RequestInfo, init?: RequestInit) => Promise<Response>
Custom fetch implementation.
Language Models
You can create language models using a provider instance:
import { cerebras } from '@ai-sdk/cerebras';import { generateText } from 'ai';
const { text } = await generateText({ model: cerebras('llama3.1-8b'), prompt: 'Write a vegetarian lasagna recipe for 4 people.',});
Cerebras language models can be used in the streamText
and streamUI
functions (see AI SDK Core and AI SDK RSC).
Model Capabilities
Model | Image Input | Object Generation | Tool Usage | Tool Streaming |
---|---|---|---|---|
llama3.1-8b | ||||
llama3.1-70b | ||||
llama3.3-70b |
Please see the Cerebras docs for more details about the available models. Note that context windows are temporarily limited to 8192 tokens in the Free Tier.