Prompts
Prompts are instructions that you give a large language model (LLM) to tell it what to do. It's like when you ask someone for directions; the clearer your question, the better the directions you'll get.
Many LLM providers offer complex interfaces for specifying prompts. They involve different roles and message types. While these interfaces are powerful, they can be hard to use and understand.
In order to simplify prompting, the AI SDK support text, message, and system prompts.
Text Prompts
Text prompts are strings. They are ideal for simple generation use cases, e.g. repeatedly generating content for variants of the same prompt text.
You can set text prompts using the prompt
property made available by AI SDK functions like streamText
or generateObject
.
You can structure the text in any way and inject variables, e.g. using a template literal.
const result = await generateText({ model: yourModel, prompt: 'Invent a new holiday and describe its traditions.',});
You can also use template literals to provide dynamic data to your prompt.
const result = await generateText({ model: yourModel, prompt: `I am planning a trip to ${destination} for ${lengthOfStay} days. ` + `Please suggest the best tourist activities for me to do.`,});
System Prompts
System prompts are the initial set of instructions given to models that help guide and constrain the models' behaviors and responses.
You can set system prompts using the system
property.
System prompts work with both the prompt
and the messages
properties.
const result = await generateText({ model: yourModel, system: `You help planning travel itineraries. ` + `Respond to the users' request with a list ` + `of the best stops to make in their destination.`, prompt: `I am planning a trip to ${destination} for ${lengthOfStay} days. ` + `Please suggest the best tourist activities for me to do.`,});
When you use a message prompt, you can also use system messages instead of a system prompt.
Message Prompts
A message prompt is an array of user, assistant, and tool messages.
They are great for chat interfaces and more complex, multi-modal prompts.
You can use the messages
property to set message prompts.
Each message has a role
and a content
property. The content can either be text (for user and assistant messages), or an array of relevant parts (data) for that message type.
const result = await streamUI({ model: yourModel, messages: [ { role: 'user', content: 'Hi!' }, { role: 'assistant', content: 'Hello, how can I help?' }, { role: 'user', content: 'Where can I buy the best Currywurst in Berlin?' }, ],});
Instead of sending a text in the content
property, you can send an array of parts that includes a mix of text and other content parts.
Not all language models support all message and content types. For example, some models might not be capable of handling multi-modal inputs or tool messages. Learn more about the capabilities of select models.
User Messages
Text Parts
Text content is the most common type of content. It is a string that is passed to the model.
If you only need to send text content in a message, the content
property can be a string,
but you can also use the parts
property to send multiple parts of content.
const result = await generateText({ model: yourModel, messages: [ { role: 'user', content: [ { type: 'text', text: 'Where can I buy the best Currywurst in Berlin?', }, ], }, ],});
Image Parts
User messages can include image parts. An image can be one of the following:
- base64-encoded image:
string
with base-64 encoded content- data URL
string
, e.g.data:image/png;base64,...
- binary image:
ArrayBuffer
Uint8Array
Buffer
- URL:
- http(s) URL
string
, e.g.https://example.com/image.png
URL
object, e.g.new URL('https://example.com/image.png')
- http(s) URL
Example: Binary image (Buffer)
const result = await generateText({ model, messages: [ { role: 'user', content: [ { type: 'text', text: 'Describe the image in detail.' }, { type: 'image', image: fs.readFileSync('./data/comic-cat.png'), }, ], }, ],});
Example: Base-64 encoded image (string)
const result = await generateText({ model: yourModel, messages: [ { role: 'user', content: [ { type: 'text', text: 'Describe the image in detail.' }, { type: 'image', image: fs.readFileSync('./data/comic-cat.png').toString('base64'), }, ], }, ],});
Example: Image URL (string)
const result = await generateText({ model: yourModel, messages: [ { role: 'user', content: [ { type: 'text', text: 'Describe the image in detail.' }, { type: 'image', image: 'https://github.com/vercel/ai/blob/main/examples/ai-core/data/comic-cat.png?raw=true', }, ], }, ],});
File Parts
Only a few providers and models currently support file parts: Google
Generative AI, Google
Vertex AI,
OpenAI (for wav
and mp3
audio with
gpt-4o-audio-preview
), Anthropic
(for pdf
).
User messages can include file parts. A file can be one of the following:
- base64-encoded file:
string
with base-64 encoded content- data URL
string
, e.g.data:image/png;base64,...
- binary data:
ArrayBuffer
Uint8Array
Buffer
- URL:
- http(s) URL
string
, e.g.https://example.com/some.pdf
URL
object, e.g.new URL('https://example.com/some.pdf')
- http(s) URL
You need to specify the MIME type of the file you are sending.
Example: PDF file from Buffer
import { google } from '@ai-sdk/google';import { generateText } from 'ai';
const result = await generateText({ model: google('gemini-1.5-flash'), messages: [ { role: 'user', content: [ { type: 'text', text: 'What is the file about?' }, { type: 'file', mimeType: 'application/pdf', data: fs.readFileSync('./data/example.pdf'), }, ], }, ],});
Example: mp3 audio file from Buffer
import { openai } from '@ai-sdk/openai';import { generateText } from 'ai';
const result = await generateText({ model: openai('gpt-4o-audio-preview'), messages: [ { role: 'user', content: [ { type: 'text', text: 'What is the audio saying?' }, { type: 'file', mimeType: 'audio/mpeg', data: fs.readFileSync('./data/galileo.mp3'), }, ], }, ],});
Assistant Messages
Assistant messages are messages that have a role of assistant
.
They are typically previous responses from the assistant and can contain text and tool call parts.
Example: Assistant message with text
const result = await generateText({ model: yourModel, messages: [ { role: 'user', content: 'Hi!' }, { role: 'assistant', content: 'Hello, how can I help?' }, ],});
Example: Assistant message with tool call
const result = await generateText({ model: yourModel, messages: [ { role: 'user', content: 'How many calories are in this block of cheese?' }, { type: 'tool-call', toolCallId: '12345', toolName: 'get-nutrition-data', args: { cheese: 'Roquefort' }, }, ],});
Tool messages
Tools (also known as function calling) are programs that you can provide an LLM to extend it's built-in functionality. This can be anything from calling an external API to calling functions within your UI. Learn more about Tools in the next section.
For models that support tool calls, assistant messages can contain tool call parts, and tool messages can contain tool result parts. A single assistant message can call multiple tools, and a single tool message can contain multiple tool results.
const result = await generateText({ model: yourModel, messages: [ { role: 'user', content: [ { type: 'text', text: 'How many calories are in this block of cheese?', }, { type: 'image', image: fs.readFileSync('./data/roquefort.jpg') }, ], }, { role: 'assistant', content: [ { type: 'tool-call', toolCallId: '12345', toolName: 'get-nutrition-data', args: { cheese: 'Roquefort' }, }, // there could be more tool calls here (parallel calling) ], }, { role: 'tool', content: [ { type: 'tool-result', toolCallId: '12345', // needs to match the tool call id toolName: 'get-nutrition-data', result: { name: 'Cheese, roquefort', calories: 369, fat: 31, protein: 22, }, }, // there could be more tool results here (parallel calling) ], }, ],});
Multi-modal Tool Results
Multi-part tool results are experimental and only supported by Anthropic.
Tool results can be multi-part and multi-modal, e.g. a text and an image.
You can use the experimental_content
property on tool parts to specify multi-part tool results.
const result = await generateText({ model: yourModel, messages: [ // ... { role: 'tool', content: [ { type: 'tool-result', toolCallId: '12345', // needs to match the tool call id toolName: 'get-nutrition-data', // for models that do not support multi-part tool results, // you can include a regular result part: result: { name: 'Cheese, roquefort', calories: 369, fat: 31, protein: 22, }, // for models that support multi-part tool results, // you can include a multi-part content part: content: [ { type: 'text', text: 'Here is an image of the nutrition data for the cheese:', }, { type: 'image', data: fs.readFileSync('./data/roquefort-nutrition-data.png'), mimeType: 'image/png', }, ], }, ], }, ],});
System Messages
System messages are messages that are sent to the model before the user messages to guide the assistant's behavior.
You can alternatively use the system
property.
const result = await generateText({ model: yourModel, messages: [ { role: 'system', content: 'You help planning travel itineraries.' }, { role: 'user', content: 'I am planning a trip to Berlin for 3 days. Please suggest the best tourist activities for me to do.', }, ],});