Handling Loading State
AI SDK RSC is currently experimental. We recommend using AI SDK UI for production. For guidance on migrating from RSC to UI, see our migration guide.
Given that responses from language models can often take a while to complete, it's crucial to be able to show loading state to users. This provides visual feedback that the system is working on their request and helps maintain a positive user experience.
There are three approaches you can take to handle loading state with the AI SDK RSC:
- Managing loading state similar to how you would in a traditional Next.js application. This involves setting a loading state variable in the client and updating it when the response is received.
- Streaming loading state from the server to the client. This approach allows you to track loading state on a more granular level and provide more detailed feedback to the user.
- Streaming loading component from the server to the client. This approach allows you to stream a React Server Component to the client while awaiting the model's response.
Handling Loading State on the Client
Client
Let's create a simple Next.js page that will call the generateResponse
function when the form is submitted. The function will take in the user's prompt (input
) and then generate a response (response
). To handle the loading state, use the loading
state variable. When the form is submitted, set loading
to true
, and when the response is received, set it back to false
. While the response is being streamed, the input field will be disabled.
'use client';
import { useState } from 'react';import { generateResponse } from './actions';import { readStreamableValue } from 'ai/rsc';
// Force the page to be dynamic and allow streaming responses up to 30 secondsexport const maxDuration = 30;
export default function Home() { const [input, setInput] = useState<string>(''); const [generation, setGeneration] = useState<string>(''); const [loading, setLoading] = useState<boolean>(false);
return ( <div> <div>{generation}</div> <form onSubmit={async e => { e.preventDefault(); setLoading(true); const response = await generateResponse(input);
let textContent = '';
for await (const delta of readStreamableValue(response)) { textContent = `${textContent}${delta}`; setGeneration(textContent); } setInput(''); setLoading(false); }} > <input type="text" value={input} disabled={loading} className="disabled:opacity-50" onChange={event => { setInput(event.target.value); }} /> <button>Send Message</button> </form> </div> );}
Server
Now let's implement the generateResponse
function. Use the streamText
function to generate a response to the input.
'use server';
import { streamText } from 'ai';import { openai } from '@ai-sdk/openai';import { createStreamableValue } from 'ai/rsc';
export async function generateResponse(prompt: string) { const stream = createStreamableValue();
(async () => { const { textStream } = streamText({ model: openai('gpt-4o'), prompt, });
for await (const text of textStream) { stream.update(text); }
stream.done(); })();
return stream.value;}
Streaming Loading State from the Server
If you are looking to track loading state on a more granular level, you can create a new streamable value to store a custom variable and then read this on the frontend. Let's update the example to create a new streamable value for tracking loading state:
Server
'use server';
import { streamText } from 'ai';import { openai } from '@ai-sdk/openai';import { createStreamableValue } from 'ai/rsc';
export async function generateResponse(prompt: string) { const stream = createStreamableValue(); const loadingState = createStreamableValue({ loading: true });
(async () => { const { textStream } = streamText({ model: openai('gpt-4o'), prompt, });
for await (const text of textStream) { stream.update(text); }
stream.done(); loadingState.done({ loading: false }); })();
return { response: stream.value, loadingState: loadingState.value };}
Client
'use client';
import { useState } from 'react';import { generateResponse } from './actions';import { readStreamableValue } from 'ai/rsc';
// Force the page to be dynamic and allow streaming responses up to 30 secondsexport const maxDuration = 30;
export default function Home() { const [input, setInput] = useState<string>(''); const [generation, setGeneration] = useState<string>(''); const [loading, setLoading] = useState<boolean>(false);
return ( <div> <div>{generation}</div> <form onSubmit={async e => { e.preventDefault(); setLoading(true); const { response, loadingState } = await generateResponse(input);
let textContent = '';
for await (const responseDelta of readStreamableValue(response)) { textContent = `${textContent}${responseDelta}`; setGeneration(textContent); } for await (const loadingDelta of readStreamableValue(loadingState)) { if (loadingDelta) { setLoading(loadingDelta.loading); } } setInput(''); setLoading(false); }} > <input type="text" value={input} disabled={loading} className="disabled:opacity-50" onChange={event => { setInput(event.target.value); }} /> <button>Send Message</button> </form> </div> );}
This allows you to provide more detailed feedback about the generation process to your users.
Streaming Loading Components with streamUI
If you are using the streamUI
function, you can stream the loading state to the client in the form of a React component. streamUI
supports the usage of JavaScript generator functions , which allow you to yield some value (in this case a React component) while some other blocking work completes.
Server
'use server';
import { openai } from '@ai-sdk/openai';import { streamUI } from 'ai/rsc';
export async function generateResponse(prompt: string) { const result = await streamUI({ model: openai('gpt-4o'), prompt, text: async function* ({ content }) { yield <div>loading...</div>; return <div>{content}</div>; }, });
return result.value;}
Remember to update the file from .ts
to .tsx
because you are defining a
React component in the streamUI
function.
Client
'use client';
import { useState } from 'react';import { generateResponse } from './actions';import { readStreamableValue } from 'ai/rsc';
// Force the page to be dynamic and allow streaming responses up to 30 secondsexport const maxDuration = 30;
export default function Home() { const [input, setInput] = useState<string>(''); const [generation, setGeneration] = useState<React.ReactNode>();
return ( <div> <div>{generation}</div> <form onSubmit={async e => { e.preventDefault(); const result = await generateResponse(input); setGeneration(result); setInput(''); }} > <input type="text" value={input} onChange={event => { setInput(event.target.value); }} /> <button>Send Message</button> </form> </div> );}