# Build Agents on Cloudflare URL: https://developers.cloudflare.com/agents/ import { CardGrid, Description, Feature, LinkButton, LinkTitleCard, PackageManagers, Plan, RelatedProduct, Render, TabItem, Tabs, TypeScriptExample, } from "~/components"; The Agents SDK enables you to build and deploy AI-powered agents that can autonomously perform tasks, communicate with clients in real time, call AI models, persist state, schedule tasks, run asynchronous workflows, browse the web, query data from your database, support human-in-the-loop interactions, and [a lot more](/agents/api-reference/). ### Ship your first Agent To use the Agent starter template and create your first Agent with the Agents SDK: ```sh # install it npm create cloudflare@latest agents-starter -- --template=cloudflare/agents-starter # and deploy it npx wrangler@latest deploy ``` Head to the guide on [building a chat agent](/agents/getting-started/build-a-chat-agent) to learn how the starter project is built and how to use it as a foundation for your own agents. If you're already building on [Workers](/workers/), you can install the `agents` package directly into an existing project: ```sh npm i agents ``` And then define your first Agent by creating a class that extends the `Agent` class: <TypeScriptExample> ```ts import { Agent, AgentNamespace } from 'agents'; export class MyAgent extends Agent { // Define methods on the Agent: // https://developers.cloudflare.com/agents/api-reference/agents-api/ // // Every Agent has built in state via this.setState and this.sql // Built-in scheduling via this.schedule // Agents support WebSockets, HTTP requests, state synchronization and // can run for seconds, minutes or hours: as long as the tasks need. } ``` </TypeScriptExample> Dive into the [Agent SDK reference](/agents/api-reference/agents-api/) to learn more about how to use the Agents SDK package and defining an `Agent`. ### Why build agents on Cloudflare? We built the Agents SDK with a few things in mind: - **Batteries (state) included**: Agents come with [built-in state management](/agents/api-reference/store-and-sync-state/), with the ability to automatically sync state between an Agent and clients, trigger events on state changes, and read+write to each Agent's SQL database. - **Communicative**: You can connect to an Agent via [WebSockets](/agents/api-reference/websockets/) and stream updates back to client in real-time. Handle a long-running response from a reasoning model, the results of an [asynchronous workflow](/agents/api-reference/run-workflows/), or build a chat app that builds on the `useAgent` hook included in the Agents SDK. - **Extensible**: Agents are code. Use the [AI models](/agents/api-reference/using-ai-models/) you want, bring-your-own headless browser service, pull data from your database hosted in another cloud, add your own methods to your Agent and call them. Agents built with Agents SDK can be deployed directly to Cloudflare and run on top of [Durable Objects](/durable-objects/) — which you can think of as stateful micro-servers that can scale to tens of millions — and are able to run wherever they need to. Run your Agents close to a user for low-latency interactivity, close to your data for throughput, and/or anywhere in between. --- ### Build on the Cloudflare Platform <RelatedProduct header="Workers" href="/workers/" product="workers"> Build serverless applications and deploy instantly across the globe for exceptional performance, reliability, and scale. </RelatedProduct> <RelatedProduct header="AI Gateway" href="/ai-gateway/" product="ai-gateway"> Observe and control your AI applications with caching, rate limiting, request retries, model fallback, and more. </RelatedProduct> <RelatedProduct header="Vectorize" href="/vectorize/" product="vectorize"> Build full-stack AI applications with Vectorize, Cloudflare’s vector database. Adding Vectorize enables you to perform tasks such as semantic search, recommendations, anomaly detection or can be used to provide context and memory to an LLM. </RelatedProduct> <RelatedProduct header="Workers AI" href="/workers-ai/" product="workers-ai"> Run machine learning models, powered by serverless GPUs, on Cloudflare's global network. </RelatedProduct> <RelatedProduct header="Workflows" href="/workflows/" product="workflows"> Build stateful agents that guarantee executions, including automatic retries, persistent state that runs for minutes, hours, days, or weeks. </RelatedProduct> --- # Agents API URL: https://developers.cloudflare.com/agents/api-reference/agents-api/ import { MetaInfo, Render, Type, TypeScriptExample, WranglerConfig } from "~/components"; This page provides an overview of the Agent SDK API, including the `Agent` class, methods and properties built-in to the Agents SDK. The Agents SDK exposes two main APIs: * The server-side `Agent` class. An Agent encapsulates all of the logic for an Agent, including how clients can connect to it, how it stores state, the methods it exposes, how to call AI models, and any error handling. * The client-side `AgentClient` class, which allows you to connect to an Agent instance from a client-side application. The client APIs also include React hooks, including `useAgent` and `useAgentChat`, and allow you to automatically synchronize state between each unique Agent (running server-side) and your client applications. :::note Agents require [Cloudflare Durable Objects](/durable-objects/), see [Configuration](/agents/getting-started/testing-your-agent/#add-the-agent-configuration) to learn how to add the required bindings to your project. ::: You can also find more specific usage examples for each API in the [Agents API Reference](/agents/api-reference/). <TypeScriptExample> ```ts import { Agent } from "agents"; class MyAgent extends Agent { // Define methods on the Agent } export default MyAgent; ``` </TypeScriptExample> An Agent can have many (millions of) instances: each instance is a separate micro-server that runs independently of the others. This allows Agents to scale horizontally: an Agent can be associated with a single user, or many thousands of users, depending on the agent you're building. Instances of an Agent are addressed by a unique identifier: that identifier (ID) can be the user ID, an email address, GitHub username, a flight ticket number, an invoice ID, or any other identifier that helps to uniquely identify the instance and for whom it is acting on behalf of. <Render file="unique-agents" /> ### Agent class API Writing an Agent requires you to define a class that extends the `Agent` class from the Agents SDK package. An Agent encapsulates all of the logic for an Agent, including how clients can connect to it, how it stores state, the methods it exposes, and any error handling. You can also define your own methods on an Agent: it's technically valid to publish an Agent only has your own methods exposed, and create/get Agents directly from a Worker. Your own methods can access the Agent's environment variables and bindings on `this.env`, state on `this.setState`, and call other methods on the Agent via `this.yourMethodName`. ```ts import { Agent } from "agents"; interface Env { // Define environment variables & bindings here } // Pass the Env as a TypeScript type argument // Any services connected to your Agent or Worker as Bindings // are then available on this.env.<BINDING_NAME> // The core class for creating Agents that can maintain state, orchestrate // complex AI workflows, schedule tasks, and interact with users and other // Agents. class MyAgent extends Agent<Env, State> { // Optional initial state definition initialState = { counter: 0, messages: [], lastUpdated: null }; // Called when a new Agent instance starts or wakes from hibernation async onStart() { console.log('Agent started with state:', this.state); } // Handle HTTP requests coming to this Agent instance // Returns a Response object async onRequest(request: Request): Promise<Response> { return new Response("Hello from Agent!"); } // Called when a WebSocket connection is established // Access the original request via ctx.request for auth etc. async onConnect(connection: Connection, ctx: ConnectionContext) { // Connections are automatically accepted by the SDK. // You can also explicitly close a connection here with connection.close() // Access the Request on ctx.request to inspect headers, cookies and the URL } // Called for each message received on a WebSocket connection // Message can be string, ArrayBuffer, or ArrayBufferView async onMessage(connection: Connection, message: WSMessage) { // Handle incoming messages connection.send("Received your message"); } // Handle WebSocket connection errors async onError(connection: Connection, error: unknown): Promise<void> { console.error(`Connection error:`, error); } // Handle WebSocket connection close events async onClose(connection: Connection, code: number, reason: string, wasClean: boolean): Promise<void> { console.log(`Connection closed: ${code} - ${reason}`); } // Called when the Agent's state is updated from any source // source can be "server" or a client Connection onStateUpdate(state: State, source: "server" | Connection) { console.log("State updated:", state, "Source:", source); } // You can define your own custom methods to be called by requests, // WebSocket messages, or scheduled tasks async customProcessingMethod(data: any) { // Process data, update state, schedule tasks, etc. this.setState({ ...this.state, lastUpdated: new Date() }); } } ``` <TypeScriptExample> ```ts // Basic Agent implementation with custom methods import { Agent } from "agents"; interface MyState { counter: number; lastUpdated: Date | null; } class MyAgent extends Agent<Env, MyState> { initialState = { counter: 0, lastUpdated: null }; async onRequest(request: Request) { if (request.method === "POST") { await this.incrementCounter(); return new Response(JSON.stringify(this.state), { headers: { "Content-Type": "application/json" } }); } return new Response(JSON.stringify(this.state), { headers: { "Content-Type": "application/json" } }); } async incrementCounter() { this.setState({ counter: this.state.counter + 1, lastUpdated: new Date() }); } } ``` </TypeScriptExample> ### WebSocket API The WebSocket API allows you to accept and manage WebSocket connections made to an Agent. #### Connection Represents a WebSocket connection to an Agent. ```ts // WebSocket connection interface interface Connection<State = unknown> { // Unique ID for this connection id: string; // Client-specific state attached to this connection state: State; // Update the connection's state setState(state: State): void; // Accept an incoming WebSocket connection accept(): void; // Close the WebSocket connection with optional code and reason close(code?: number, reason?: string): void; // Send a message to the client // Can be string, ArrayBuffer, or ArrayBufferView send(message: string | ArrayBuffer | ArrayBufferView): void; } ``` <TypeScriptExample> ```ts // Example of handling WebSocket messages export class YourAgent extends Agent { async onMessage(connection: Connection, message: WSMessage) { if (typeof message === 'string') { try { // Parse JSON message const data = JSON.parse(message); if (data.type === 'update') { // Update connection-specific state connection.setState({ ...connection.state, lastActive: Date.now() }); // Update global Agent state this.setState({ ...this.state, connections: this.state.connections + 1 }); // Send response back to this client only connection.send(JSON.stringify({ type: 'updated', status: 'success' })); } } catch (e) { connection.send(JSON.stringify({ error: 'Invalid message format' })); } } } } ``` </TypeScriptExample> #### WSMessage Types of messages that can be received from a WebSocket. ```ts // Types of messages that can be received from WebSockets type WSMessage = string | ArrayBuffer | ArrayBufferView; ``` #### ConnectionContext Context information for a WebSocket connection. ```ts // Context available during WebSocket connection interface ConnectionContext { // The original HTTP request that initiated the WebSocket connection request: Request; } ``` ### State synchronization API :::note To learn more about how to manage state within an Agent, refer to the documentation on [managing and syncing state](/agents/api-reference/store-and-sync-state/). ::: #### State Methods and types for managing Agent state. ```ts // State management in the Agent class class Agent<Env, State = unknown> { // Initial state that will be set if no state exists yet initialState: State = {} as unknown as State; // Current state of the Agent, persisted across restarts get state(): State; // Update the Agent's state // Persists to storage and notifies all connected clients setState(state: State): void; // Called when state is updated from any source // Override to react to state changes onStateUpdate(state: State, source: "server" | Connection): void; } ``` <TypeScriptExample> ```ts // Example of state management in an Agent interface ChatState { messages: Array<{ sender: string; text: string; timestamp: number }>; participants: string[]; settings: { allowAnonymous: boolean; maxHistoryLength: number; }; } interface Env { // Your bindings and environment variables } // Inside your Agent class export class YourAgent extends Agent<Env, ChatState> { async addMessage(sender: string, text: string) { // Update state with new message this.setState({ ...this.state, messages: [ ...this.state.messages, { sender, text, timestamp: Date.now() } ].slice(-this.state.settings.maxHistoryLength) // Maintain max history }); // The onStateUpdate method will automatically be called // and all connected clients will receive the update } // Override onStateUpdate to add custom behavior when state changes onStateUpdate(state: ChatState, source: "server" | Connection) { console.log(`State updated by ${source === "server" ? "server" : "client"}`); // You could trigger additional actions based on state changes if (state.messages.length > 0) { const lastMessage = state.messages[state.messages.length - 1]; if (lastMessage.text.includes('@everyone')) { this.notifyAllParticipants(lastMessage); } } } } ``` </TypeScriptExample> ### Scheduling API #### Scheduling tasks Schedule tasks to run at a specified time in the future. ```ts // Scheduling API for running tasks in the future class Agent<Env, State = unknown> { // Schedule a task to run in the future // when: seconds from now, specific Date, or cron expression // callback: method name on the Agent to call // payload: data to pass to the callback // Returns a Schedule object with the task ID async schedule<T = any>( when: Date | string | number, callback: keyof this, payload?: T ): Promise<Schedule<T>>; // Get a scheduled task by ID // Returns undefined if the task doesn't exist async getSchedule<T = any>(id: string): Promise<Schedule<T> | undefined>; // Get all scheduled tasks matching the criteria // Returns an array of Schedule objects getSchedules<T = any>(criteria?: { description?: string; id?: string; type?: "scheduled" | "delayed" | "cron"; timeRange?: { start?: Date; end?: Date }; }): Schedule<T>[]; // Cancel a scheduled task by ID // Returns true if the task was cancelled, false otherwise async cancelSchedule(id: string): Promise<boolean>; } ``` <TypeScriptExample> ```ts // Example of scheduling in an Agent interface ReminderData { userId: string; message: string; channel: string; } export class YourAgent extends Agent { // Schedule a one-time reminder in 2 hours async scheduleReminder(userId: string, message: string) { const twoHoursFromNow = new Date(Date.now() + 2 * 60 * 60 * 1000); const schedule = await this.schedule<ReminderData>( twoHoursFromNow, 'sendReminder', { userId, message, channel: 'email' } ); console.log(`Scheduled reminder with ID: ${schedule.id}`); return schedule.id; } // Schedule a recurring daily task using cron async scheduleDailyReport() { // Run at 08:00 AM every day const schedule = await this.schedule( '0 8 * * *', // Cron expression: minute hour day month weekday 'generateDailyReport', { reportType: 'daily-summary' } ); console.log(`Scheduled daily report with ID: ${schedule.id}`); return schedule.id; } // Method that will be called when the scheduled task runs async sendReminder(data: ReminderData) { console.log(`Sending reminder to ${data.userId}: ${data.message}`); // Add code to send the actual notification } } ``` </TypeScriptExample> #### Schedule object Represents a scheduled task. ```ts // Represents a scheduled task type Schedule<T = any> = { // Unique identifier for the schedule id: string; // Name of the method to be called callback: string; // Data to be passed to the callback payload: T; } & ( | { // One-time execution at a specific time type: "scheduled"; // Timestamp when the task should execute time: number; } | { // Delayed execution after a certain time type: "delayed"; // Timestamp when the task should execute time: number; // Number of seconds to delay execution delayInSeconds: number; } | { // Recurring execution based on cron expression type: "cron"; // Timestamp for the next execution time: number; // Cron expression defining the schedule cron: string; } ); ``` <TypeScriptExample> ```ts export class YourAgent extends Agent { // Example of managing scheduled tasks async viewAndManageSchedules() { // Get all scheduled tasks const allSchedules = this.getSchedules(); console.log(`Total scheduled tasks: ${allSchedules.length}`); // Get tasks scheduled for a specific time range const upcomingSchedules = this.getSchedules({ timeRange: { start: new Date(), end: new Date(Date.now() + 24 * 60 * 60 * 1000) // Next 24 hours } }); // Get a specific task by ID const taskId = "task-123"; const specificTask = await this.getSchedule(taskId); if (specificTask) { console.log(`Found task: ${specificTask.callback} at ${new Date(specificTask.time)}`); // Cancel a scheduled task const cancelled = await this.cancelSchedule(taskId); console.log(`Task cancelled: ${cancelled}`); } } } ``` </TypeScriptExample> ### SQL API Each Agent instance has an embedded SQLite database that can be accessed using the `this.sql` method within any method on your `Agent` class. #### SQL queries Execute SQL queries against the Agent's built-in SQLite database using the `this.sql` method within any method on your `Agent` class. ```ts // SQL query API for the Agent's embedded database class Agent<Env, State = unknown> { // Execute a SQL query with tagged template literals // Returns an array of rows matching the query sql<T = Record<string, string | number | boolean | null>>( strings: TemplateStringsArray, ...values: (string | number | boolean | null)[] ): T[]; } ``` <TypeScriptExample> ```ts // Example of using SQL in an Agent interface User { id: string; name: string; email: string; created_at: number; } export class YourAgent extends Agent { async setupDatabase() { // Create a table if it doesn't exist this.sql` CREATE TABLE IF NOT EXISTS users ( id TEXT PRIMARY KEY, name TEXT NOT NULL, email TEXT UNIQUE, created_at INTEGER ) `; } async createUser(id: string, name: string, email: string) { // Insert a new user this.sql` INSERT INTO users (id, name, email, created_at) VALUES (${id}, ${name}, ${email}, ${Date.now()}) `; } async getUserById(id: string): Promise<User | null> { // Query a user by ID const users = this.sql<User>` SELECT * FROM users WHERE id = ${id} `; return users.length ? users[0] : null; } async searchUsers(term: string): Promise<User[]> { // Search users with a wildcard return this.sql<User>` SELECT * FROM users WHERE name LIKE ${'%' + term + '%'} OR email LIKE ${'%' + term + '%'} ORDER BY created_at DESC `; } } ``` </TypeScriptExample> :::note Visit the [state management API documentation](/agents/api-reference/store-and-sync-state/) within the Agents SDK, including the native `state` APIs and the built-in `this.sql` API for storing and querying data within your Agents. ::: ### Client API The Agents SDK provides a set of client APIs for interacting with Agents from client-side JavaScript code, including: * React hooks, including `useAgent` and `useAgentChat`, for connecting to Agents from client applications. * Client-side [state syncing](/agents/api-reference/store-and-sync-state/) that allows you to subscribe to state updates between the Agent and any connected client(s) when calling `this.setState` within your Agent's code. * The ability to call remote methods (Remote Procedure Calls; RPC) on the Agent from client-side JavaScript code using the `@callable` method decorator. #### AgentClient Client for connecting to an Agent from the browser. ```ts import { AgentClient } from "agents/client"; // Options for creating an AgentClient type AgentClientOptions = Omit<PartySocketOptions, "party" | "room"> & { // Name of the agent to connect to (class name in kebab-case) agent: string; // Name of the specific Agent instance (optional, defaults to "default") name?: string; // Other WebSocket options like host, protocol, etc. }; // WebSocket client for connecting to an Agent class AgentClient extends PartySocket { static fetch(opts: PartyFetchOptions): Promise<Response>; constructor(opts: AgentClientOptions); } ``` <TypeScriptExample> ```ts // Example of using AgentClient in the browser import { AgentClient } from "agents/client"; // Connect to an Agent instance const socket = new AgentClient({ agent: "chat-agent", // Name of your Agent class in kebab-case name: "support-room-123", // Specific instance name // Optional event handlers onOpen: () => { console.log("Connected to agent"); // Send an initial message socket.send(JSON.stringify({ type: "join", user: "user123" })); }, onMessage: (event) => { // Handle incoming messages const data = JSON.parse(event.data); console.log("Received:", data); if (data.type === "state_update") { // Update local UI with new state updateUI(data.state); } }, onClose: () => console.log("Disconnected from agent") }); // Send messages to the Agent function sendMessage(text) { socket.send(JSON.stringify({ type: "message", text, timestamp: Date.now() })); } ``` </TypeScriptExample> #### agentFetch Make an HTTP request to an Agent. ```ts import { agentFetch } from "agents/client"; // Options for the agentFetch function type AgentClientFetchOptions = Omit<PartyFetchOptions, "party" | "room"> & { // Name of the agent to connect to agent: string; // Name of the specific Agent instance (optional) name?: string; }; // Make an HTTP request to an Agent function agentFetch( opts: AgentClientFetchOptions, init?: RequestInit ): Promise<Response>; ``` <TypeScriptExample> ```ts // Example of using agentFetch in the browser import { agentFetch } from "agents/client"; // Function to get data from an Agent async function fetchAgentData() { try { const response = await agentFetch( { agent: "task-manager", name: "user-123-tasks" }, { method: "GET", headers: { "Authorization": `Bearer ${userToken}` } } ); if (!response.ok) { throw new Error(`Error: ${response.status}`); } const data = await response.json(); return data; } catch (error) { console.error("Failed to fetch from agent:", error); } } ``` </TypeScriptExample> ### React API The Agents SDK provides a React API for simplifying connection and routing to Agents from front-end frameworks, including React Router (Remix), Next.js, and Astro. #### useAgent React hook for connecting to an Agent. ```ts import { useAgent } from "agents/react"; // Options for the useAgent hook type UseAgentOptions<State = unknown> = Omit< Parameters<typeof usePartySocket>[0], "party" | "room" > & { // Name of the agent to connect to agent: string; // Name of the specific Agent instance (optional) name?: string; // Called when the Agent's state is updated onStateUpdate?: (state: State, source: "server" | "client") => void; }; // React hook for connecting to an Agent // Returns a WebSocket connection with setState method function useAgent<State = unknown>( options: UseAgentOptions<State> ): PartySocket & { // Update the Agent's state setState: (state: State) => void }; ``` ### Chat Agent The Agents SDK exposes an `AIChatAgent` class that extends the `Agent` class and exposes an `onChatMessage` method that simplifies building interactive chat agents. You can combine this with the `useAgentChat` React hook from the `agents/ai-react` package to manage chat state and messages between a user and your Agent(s). #### AIChatAgent Extension of the `Agent` class with built-in chat capabilities. ```ts import { AIChatAgent } from "agents/ai-chat-agent"; import { Message, StreamTextOnFinishCallback, ToolSet } from "ai"; // Base class for chat-specific agents class AIChatAgent<Env = unknown, State = unknown> extends Agent<Env, State> { // Array of chat messages for the current conversation messages: Message[]; // Handle incoming chat messages and generate a response // onFinish is called when the response is complete async onChatMessage( onFinish: StreamTextOnFinishCallback<ToolSet> ): Promise<Response | undefined>; // Persist messages within the Agent's local storage. async saveMessages(messages: Message[]): Promise<void>; } ``` <TypeScriptExample> ```ts // Example of extending AIChatAgent import { AIChatAgent } from "agents/ai-chat-agent"; import { Message } from "ai"; interface Env { AI: any; // Your AI binding } class CustomerSupportAgent extends AIChatAgent<Env> { // Override the onChatMessage method to customize behavior async onChatMessage(onFinish) { // Access the AI models using environment bindings const { openai } = this.env.AI; // Get the current conversation history const chatHistory = this.messages; // Generate a system prompt based on knowledge base const systemPrompt = await this.generateSystemPrompt(); // Generate a response stream const stream = await openai.chat({ model: "gpt-4o", messages: [ { role: "system", content: systemPrompt }, ...chatHistory ], stream: true }); // Return the streaming response return new Response(stream, { headers: { "Content-Type": "text/event-stream" } }); } // Helper method to generate a system prompt async generateSystemPrompt() { // Query knowledge base or use static prompt return `You are a helpful customer support agent. Respond to customer inquiries based on the following guidelines: - Be friendly and professional - If you don't know an answer, say so - Current company policies: ...`; } } ``` </TypeScriptExample> ### Chat Agent React API #### useAgentChat React hook for building AI chat interfaces using an Agent. ```ts import { useAgentChat } from "agents/ai-react"; import { useAgent } from "agents/react"; import type { Message } from "ai"; // Options for the useAgentChat hook type UseAgentChatOptions = Omit< Parameters<typeof useChat>[0] & { // Agent connection from useAgent agent: ReturnType<typeof useAgent>; }, "fetch" >; // React hook for building AI chat interfaces using an Agent function useAgentChat(options: UseAgentChatOptions): { // Current chat messages messages: Message[]; // Set messages and synchronize with the Agent setMessages: (messages: Message[]) => void; // Clear chat history on both client and Agent clearHistory: () => void; // Append a new message to the conversation append: (message: Message, chatRequestOptions?: any) => Promise<string | null | undefined>; // Reload the last user message reload: (chatRequestOptions?: any) => Promise<string | null | undefined>; // Stop the AI response generation stop: () => void; // Current input text input: string; // Set the input text setInput: React.Dispatch<React.SetStateAction<string>>; // Handle input changes handleInputChange: (e: React.ChangeEvent<HTMLInputElement | HTMLTextAreaElement>) => void; // Submit the current input handleSubmit: (event?: { preventDefault?: () => void }, chatRequestOptions?: any) => void; // Additional metadata metadata?: Object; // Whether a response is currently being generated isLoading: boolean; // Current status of the chat status: "submitted" | "streaming" | "ready" | "error"; // Tool data from the AI response data?: any[]; // Set tool data setData: (data: any[] | undefined | ((data: any[] | undefined) => any[] | undefined)) => void; // Unique ID for the chat id: string; // Add a tool result for a specific tool call addToolResult: ({ toolCallId, result }: { toolCallId: string; result: any }) => void; // Current error if any error: Error | undefined; }; ``` <TypeScriptExample> ```tsx // Example of using useAgentChat in a React component import { useAgentChat } from "agents/ai-react"; import { useAgent } from "agents/react"; import { useState } from "react"; function ChatInterface() { // Connect to the chat agent const agentConnection = useAgent({ agent: "customer-support", name: "session-12345" }); // Use the useAgentChat hook with the agent connection const { messages, input, handleInputChange, handleSubmit, isLoading, error, clearHistory } = useAgentChat({ agent: agentConnection, initialMessages: [ { role: "system", content: "You're chatting with our AI assistant." }, { role: "assistant", content: "Hello! How can I help you today?" } ] }); return ( <div className="chat-container"> <div className="message-history"> {messages.map((message, i) => ( <div key={i} className={`message ${message.role}`}> {message.role === 'user' ? '👤' : '🤖'} {message.content} </div> ))} {isLoading && <div className="loading">AI is typing...</div>} {error && <div className="error">Error: {error.message}</div>} </div> <form onSubmit={handleSubmit} className="message-input"> <input value={input} onChange={handleInputChange} placeholder="Type your message..." disabled={isLoading} /> <button type="submit" disabled={isLoading || !input.trim()}> Send </button> <button type="button" onClick={clearHistory}> Clear Chat </button> </form> </div> ); } ``` </TypeScriptExample> ### Next steps * [Build a chat Agent](/agents/getting-started/build-a-chat-agent/) using the Agents SDK and deploy it to Workers. * Learn more [using WebSockets](/agents/api-reference/websockets/) to build interactive Agents and stream data back from your Agent. * [Orchestrate asynchronous workflows](/agents/api-reference/run-workflows) from your Agent by combining the Agents SDK and [Workflows](/workflows). --- # Browse the web URL: https://developers.cloudflare.com/agents/api-reference/browse-the-web/ import { MetaInfo, Render, Type, TypeScriptExample, WranglerConfig } from "~/components"; Agents can browse the web using the [Browser Rendering](/browser-rendering/) API or your preferred headless browser service. ### Browser Rendering API The [Browser Rendering](/browser-rendering/) allows you to spin up headless browser instances, render web pages, and interact with websites through your Agent. You can define a method that uses Puppeteer to pull the content of a web page, parse the DOM, and extract relevant information by calling the OpenAI model: <TypeScriptExample> ```ts interface Env { BROWSER: Fetcher; } export class MyAgent extends Agent<Env> { async browse(browserInstance: Fetcher, urls: string[]) { let responses = []; for (const url of urls) { const browser = await puppeteer.launch(browserInstance); const page = await browser.newPage(); await page.goto(url); await page.waitForSelector('body'); const bodyContent = await page.$eval('body', (element) => element.innerHTML); const client = new OpenAI({ apiKey: this.env.OPENAI_API_KEY, }); let resp = await client.chat.completions.create({ model: this.env.MODEL, messages: [ { role: 'user', content: `Return a JSON object with the product names, prices and URLs with the following format: { "name": "Product Name", "price": "Price", "url": "URL" } from the website content below. <content>${bodyContent}</content>`, }, ], response_format: { type: 'json_object', }, }); responses.push(resp); await browser.close(); } return responses; } } ``` </TypeScriptExample> You'll also need to add install the `@cloudflare/puppeteer` package and add the following to the wrangler configuration of your Agent: ```sh npm install @cloudflare/puppeteer --save-dev ``` <WranglerConfig> ```jsonc { // ... "browser": { "binding": "MYBROWSER" } // ... } ``` </WranglerConfig> ### Browserbase You can also use [Browserbase](https://docs.browserbase.com/integrations/cloudflare/typescript) by using the Browserbase API directly from within your Agent. Once you have your [Browserbase API key](https://docs.browserbase.com/integrations/cloudflare/typescript), you can add it to your Agent by creating a [secret](/workers/configuration/secrets/): ```sh cd your-agent-project-folder npx wrangler@latest secret put BROWSERBASE_API_KEY ``` ```sh output Enter a secret value: ****** Creating the secret for the Worker "agents-example" Success! Uploaded secret BROWSERBASE_API_KEY ``` Install the `@cloudflare/puppeteer` package and use it from within your Agent to call the Browserbase API: ```sh npm install @cloudflare/puppeteer ``` <TypeScriptExample> ```ts interface Env { BROWSERBASE_API_KEY: string; } export class MyAgent extends Agent<Env> { constructor(env: Env) { super(env); } } ``` </TypeScriptExample> --- # Calling Agents URL: https://developers.cloudflare.com/agents/api-reference/calling-agents/ import { MetaInfo, Render, Type, TypeScriptExample, WranglerConfig } from "~/components"; Learn how to call your Agents from Workers, including how to create Agents on-the-fly, address them, and route requests to specific instances of an Agent. ### Calling your Agent Agents are created on-the-fly and can serve multiple requests concurrently. Each Agent instance is isolated from other instances, can maintain its own state, and has a unique address. <Render file="unique-agents" /> You can create and run an instance of an Agent directly from a Worker using either: * The `routeAgentRequest` helper: this will automatically map requests to an individual Agent based on the `/agents/:agent/:name` URL pattern. The value of `:agent` will be the name of your Agent class converted to `kebab-case`, and the value of `:name` will be the name of the Agent instance you want to create or retrieve. * `getAgentByName`, which will create a new Agent instance if none exists by that name, or retrieve a handle to an existing instance. See the usage patterns in the following example: <TypeScriptExample> ```ts import { Agent, AgentNamespace, getAgentByName, routeAgentRequest } from 'agents'; interface Env { // Define your Agent on the environment here // Passing your Agent class as a TypeScript type parameter allows you to call // methods defined on your Agent. MyAgent: AgentNamespace<MyAgent>; } export default { async fetch(request, env, ctx): Promise<Response> { // Routed addressing // Automatically routes HTTP requests and/or WebSocket connections to /agents/:agent/:name // Best for: connecting React apps directly to Agents using useAgent from agents/react return (await routeAgentRequest(request, env)) || Response.json({ msg: 'no agent here' }, { status: 404 }); // Named addressing // Best for: convenience method for creating or retrieving an agent by name/ID. // Bringing your own routing, middleware and/or plugging into an existing // application or framework. let namedAgent = getAgentByName<Env, MyAgent>(env.MyAgent, 'my-unique-agent-id'); // Pass the incoming request straight to your Agent let namedResp = (await namedAgent).fetch(request); return namedResp }, } satisfies ExportedHandler<Env>; export class MyAgent extends Agent<Env> { // Your Agent implementation goes here } ``` </TypeScriptExample> #### Calling methods directly When using `getAgentByName`, you can pass both requests (including WebSocket) connections and call methods defined directly on the Agent itself using the native [JavaScript RPC](/workers/runtime-apis/rpc/) (JSRPC) API. For example, once you have a handle (or "stub") to an unique instance of your Agent, you can call methods on it: <TypeScriptExample> ```ts import { Agent, AgentNamespace, getAgentByName } from 'agents'; interface Env { // Define your Agent on the environment here // Passing your Agent class as a TypeScript type parameter allows you to call // methods defined on your Agent. MyAgent: AgentNamespace<MyAgent>; } interface UserHistory { history: string[]; lastUpdated: Date; } export default { async fetch(request, env, ctx): Promise<Response> { let namedAgent = getAgentByName<Env, MyAgent>(env.MyAgent, 'my-unique-agent-id'); // Call methods directly on the Agent, and pass native JavaScript objects let chatResponse = namedAgent.chat('Hello!'); // No need to serialize/deserialize it from a HTTP request or WebSocket // message and back again let agentState = getState() // agentState is of type UserHistory return namedResp }, } satisfies ExportedHandler<Env>; export class MyAgent extends Agent<Env, UserHistory> { // Your Agent implementation goes here async chat(prompt: string) { // call your favorite LLM return "result" } async getState() { // Return the Agent's state directly return this.state; } // Other methods as you see fit! } ``` </TypeScriptExample> When using TypeScript, ensure you pass your Agent class as a TypeScript type parameter to the AgentNamespace type so that types are correctly inferred: ```ts interface Env { // Passing your Agent class as a TypeScript type parameter allows you to call // methods defined on your Agent. MyAgent: AgentNamespace<CodeReviewAgent>; } export class CodeReviewAgent extends Agent<Env, AgentState> { // Agent methods here } ``` ### Naming your Agents When creating names for your Agents, think about what the Agent represents. A unique user? A team or company? A room or channel for collaboration? A consistent approach to naming allows you to: * direct incoming requests directly to the right Agent * deterministically route new requests back to that Agent, no matter where the client is in the world. * avoid having to rely on centralized session storage or external services for state management, since each Agent instance can maintain its own state. For a given Agent definition (or 'namespace' in the code below), there can be millions (or tens of millions) of instances of that Agent, each handling their own requests, making calls to LLMs, and maintaining their own state. For example, you might have an Agent for every user using your new AI-based code editor. In that case, you'd want to create Agents based on the user ID from your system, which would then allow that Agent to handle all requests for that user. It also ensures that [state within the Agent](/agents/api-reference/store-and-sync-state/), including chat history, language preferences, model configuration and other context can associated specifically with that user, making it easier to manage state. The example below shows how to create a unique agent Agent for each `userId` in a request: <TypeScriptExample> ```ts import { Agent, AgentNamespace, getAgentByName, routeAgentRequest } from 'agents'; interface Env { MyAgent: AgentNamespace<MyAgent>; } export default { async fetch(request, env, ctx): Promise<Response> { let userId = new URL(request.url).searchParams.get('userId') || 'anonymous'; // Use an identifier that allows you to route to requests, WebSockets or call methods on the Agent // You can also put authentication logic here - e.g. to only create or retrieve Agents for known users. let namedAgent = getAgentByName<Env, MyAgent>(env.MyAgent, 'my-unique-agent-id'); return (await namedAgent).fetch(request); }, } satisfies ExportedHandler<Env>; export class MyAgent extends Agent<Env> { // You can access the name of the agent via this.name in any method within // the Agent async onStartup() { console.log(`agent ${this.name} ready!`)} } ``` </TypeScriptExample> Replace `userId` with `teamName`, `channel`, `companyName` as fits your Agents goals - and/or configure authentication to ensure Agents are only created for known, authenticated users. ### Authenticating Agents When building and deploying Agents using the Agents SDK, you will often want to authenticate clients before passing requests to an Agent in order to restrict who the Agent will call, authorize specific users for specific Agents, and/or to limit who can access administrative or debug APIs exposed by an Agent. As best practices: * Handle authentication in your Workers code, before you invoke your Agent. * Use the built-in hooks when using the `routeAgentRequest` helper - `onBeforeConnect` and `onBeforeRequest` * Use your preferred router (such as Hono) and authentication middleware or provider to apply custom authentication schemes before calling an Agent using other methods. The `routeAgentRequest` helper documented earlier in this guide exposes two useful hooks (`onBeforeConnect`, `onBeforeRequest`) that allow you to apply custom logic before creating or retrieving an Agent: <TypeScriptExample> ```ts import { Agent, AgentNamespace, routeAgentRequest } from 'agents'; interface Env { MyAgent: AgentNamespace<MyAgent>; } export default { async fetch(request, env, ctx): Promise<Response> { // Use the onBeforeConnect and onBeforeRequest hooks to authenticate clients // or run logic before handling a HTTP request or WebSocket. return ( (await routeAgentRequest(request, env, { // Run logic before a WebSocket client connects onBeforeConnect: (request) => { // Your code/auth code here // You can return a Response here - e.g. a HTTP 403 Not Authorized - // which will stop further request processing and will NOT invoke the // Agent. // return Response.json({"error": "not authorized"}, { status: 403 }) }, // Run logic before a HTTP client clients onBeforeRequest: (request) => { // Your code/auth code here // Returning nothing will result in the call to the Agent continuing }, // Prepend a prefix for how your Agents are named here prefix: 'name-prefix-here', })) || Response.json({ msg: 'no agent here' }, { status: 404 }) ); }, } satisfies ExportedHandler<Env>; ``` </TypeScriptExample> If you are using `getAgentByName` or the underlying Durable Objects routing API, you should authenticate incoming requests or WebSocket connections before calling `getAgentByName`. For example, if you are using [Hono](https://hono.dev/), you can authenticate in the middleware before calling an Agent and passing a request (or a WebSocket connection) to it: <TypeScriptExample> ```ts import { Agent, AgentNamespace, getAgentByName } from 'agents'; import { Hono } from 'hono'; const app = new Hono<{ Bindings: Env }>(); app.use('/code-review/*', async (c, next) => { // Perform auth here // e.g. validate a Bearer token, a JWT, use your preferred auth library // return Response.json({ msg: 'unauthorized' }, { status: 401 }); await next(); // continue on if valid }); app.get('/code-review/:id', async (c) => { const id = c.req.param('teamId'); if (!id) return Response.json({ msg: 'missing id' }, { status: 400 }); // Call the Agent, creating it with the name/identifier from the ":id" segment // of our URL const agent = await getAgentByName<Env, MyAgent>(c.env.MyAgent, id); // Pass the request to our Agent instance return await agent.fetch(c.req.raw); }); ``` </TypeScriptExample> This ensures we only create Agents for authenticated users, and allows you to validate whether Agent names conform to your preferred naming scheme before instances are created. ### Next steps * Review the [API documentation](/agents/api-reference/agents-api/) for the Agents class to learn how to define * [Build a chat Agent](/agents/getting-started/build-a-chat-agent/) using the Agents SDK and deploy it to Workers. * Learn more [using WebSockets](/agents/api-reference/websockets/) to build interactive Agents and stream data back from your Agent. * [Orchestrate asynchronous workflows](/agents/api-reference/run-workflows) from your Agent by combining the Agents SDK and [Workflows](/workflows). --- # Configuration URL: https://developers.cloudflare.com/agents/api-reference/configuration/ import { MetaInfo, Render, Type, WranglerConfig } from "~/components"; An Agent is configured like any other Cloudflare Workers project, and uses [a wrangler configuration](/workers/wrangler/configuration/) file to define where your code is and what services (bindings) it will use. ### Project structure The typical file structure for an Agent project created from `npm create cloudflare@latest agents-starter -- --template cloudflare/agents-starter` follows: ```sh . |-- package-lock.json |-- package.json |-- public | `-- index.html |-- src | `-- index.ts // your Agent definition |-- test | |-- index.spec.ts // your tests | `-- tsconfig.json |-- tsconfig.json |-- vitest.config.mts |-- worker-configuration.d.ts `-- wrangler.jsonc // your Workers & Agent configuration ``` ### Example configuration Below is a minimal `wrangler.jsonc` file that defines the configuration for an Agent, including the entry point, `durable_object` namespace, and code `migrations`: <WranglerConfig> ```jsonc { "$schema": "node_modules/wrangler/config-schema.json", "name": "agents-example", "main": "src/index.ts", "compatibility_date": "2025-02-23", "compatibility_flags": ["nodejs_compat"], "durable_objects": { "bindings": [ { // Required: "name": "MyAgent", // How your Agent is called from your Worker "class_name": "MyAgent", // Must match the class name of the Agent in your code // Optional: set this if the Agent is defined in another Worker script "script_name": "the-other-worker" }, ], }, "migrations": [ { "tag": "v1", // Mandatory for the Agent to store state "new_sqlite_classes": ["MyAgent"], }, ], "observability": { "enabled": true, }, } ``` </WranglerConfig> The configuration includes: - A `main` field that points to the entry point of your Agent, which is typically a TypeScript (or JavaScript) file. - A `durable_objects` field that defines the [Durable Object namespace](/durable-objects/reference/glossary/) that your Agents will run within. - A `migrations` field that defines the code migrations that your Agent will use. This field is mandatory and must contain at least one migration. The `new_sqlite_classes` field is mandatory for the Agent to store state. Agents must define these fields in their `wrangler.jsonc` (or `wrangler.toml`) config file. --- # HTTP and Server-Sent Events URL: https://developers.cloudflare.com/agents/api-reference/http-sse/ import { MetaInfo, Render, Type, TypeScriptExample, WranglerConfig } from "~/components"; The Agents SDK allows you to handle HTTP requests and has native support for [Server-Sent Events](https://developer.mozilla.org/en-US/docs/Web/API/Server-sent_events) (SSE). This allows you build applications that can push data to clients and avoid buffering. ### Handling HTTP requests Agents can handle HTTP requests using the `onRequest` method, which is called whenever an HTTP request is received by the Agent instance. The method takes a `Request` object as a parameter and returns a `Response` object. <TypeScriptExample> ```ts class MyAgent extends Agent<Env, State> { // Handle HTTP requests coming to this Agent instance // Returns a Response object async onRequest(request: Request) { return new Response("Hello from Agent!"); } async callAIModel(prompt: string) { // Implement AI model call here } } ``` </TypeScriptExample> Review the [Agents API reference](/agents/api-reference/agents-api/) to learn more about the `Agent` class and its methods. ### Implementing Server-Sent Events The Agents SDK support Server-Sent Events directly: you can use SSE to stream data back to the client over a long running connection. This avoids buffering large responses, which can both make your Agent feel slow, and forces you to buffer the entire response in memory. When an Agent is deployed to Cloudflare Workers, there is no effective limit on the total time it takes to stream the response back: large AI model responses that take several minutes to reason and then respond will not be prematurely terminated. Note that this does not mean the client can't potentially disconnect during the streaming process: you can account for this by either [writing to the Agent's stateful storage](/agents/api-reference/store-and-sync-state/) and/or [using WebSockets](/agents/api-reference/websockets/). Because you can always [route to the same Agent](/agents/api-reference/calling-agents/), you do not need to use a centralized session store to pick back up where you left off when a client disconnects. The following example uses the AI SDK to generate text and stream it back to the client. It will automatically stream the response back to the client as the model generates it: <TypeScriptExample> ```ts import { Agent, AgentNamespace, getAgentByName, routeAgentRequest } from 'agents'; import { streamText } from 'ai'; import { createOpenAI, openai } from '@ai-sdk/openai'; interface Env { MyAgent: AgentNamespace<MyAgent>; OPENAI_API_KEY: string; } export class MyAgent extends Agent<Env> { async onRequest(request: Request) { // Test it via: // curl -d '{"prompt": "Write me a Cloudflare Worker"}' <url> let data = await request.json<{ prompt: string }>(); let stream = await this.callAIModel(data.prompt); // This uses Server-Sent Events (SSE) return stream.toTextStreamResponse({ headers: { 'Content-Type': 'text/x-unknown', 'content-encoding': 'identity', 'transfer-encoding': 'chunked', }, }); } async callAIModel(prompt: string) { const openai = createOpenAI({ apiKey: this.env.OPENAI_API_KEY, }); return streamText({ model: openai('gpt-4o'), prompt: prompt, }); } } export default { async fetch(request: Request, env: Env) { let agentId = new URL(request.url).searchParams.get('agent-id') || ''; const agent = await getAgentByName<Env, MyAgent>(env.MyAgent, agentId); return agent.fetch(request); }, }; ``` </TypeScriptExample> ### WebSockets vs. Server-Sent Events Both WebSockets and Server-Sent Events (SSE) enable real-time communication between clients and Agents. Agents built on the Agents SDK can expose both WebSocket and SSE endpoints directly. * WebSockets provide full-duplex communication, allowing data to flow in both directions simultaneously. SSE only supports server-to-client communication, requiring additional HTTP requests if the client needs to send data back. * WebSockets establish a single persistent connection that stays open for the duration of the session. SSE, being built on HTTP, may experience more overhead due to reconnection attempts and header transmission with each reconnection, especially when there is a lot of client-server communication. * While SSE works well for simple streaming scenarios, WebSockets are better suited for applications requiring minutes or hours of connection time, as they maintain a more stable connection with built-in ping/pong mechanisms to keep connections alive. * WebSockets use their own protocol (ws:// or wss://), separating them from HTTP after the initial handshake. This separation allows WebSockets to better handle binary data transmission and implement custom subprotocols for specialized use cases. If you're unsure of which is better for your use-case, we recommend WebSockets. The [WebSockets API documentation](/agents/api-reference/websockets/) provides detailed information on how to use WebSockets with the Agents SDK. ### Next steps * Review the [API documentation](/agents/api-reference/agents-api/) for the Agents class to learn how to define them. * [Build a chat Agent](/agents/getting-started/build-a-chat-agent/) using the Agents SDK and deploy it to Workers. * Learn more [using WebSockets](/agents/api-reference/websockets/) to build interactive Agents and stream data back from your Agent. * [Orchestrate asynchronous workflows](/agents/api-reference/run-workflows) from your Agent by combining the Agents SDK and [Workflows](/workflows). --- # API Reference URL: https://developers.cloudflare.com/agents/api-reference/ import { DirectoryListing } from "~/components" Learn more about what Agents can do, the `Agent` class, and the APIs that Agents expose: <DirectoryListing /> --- # Retrieval Augmented Generation URL: https://developers.cloudflare.com/agents/api-reference/rag/ import { MetaInfo, Render, Type, TypeScriptExample, WranglerConfig } from "~/components"; Agents can use Retrieval Augmented Generation (RAG) to retrieve relevant information and use it augment [calls to AI models](/agents/api-reference/using-ai-models/). Store a user's chat history to use as context for future conversations, summarize documents to bootstrap an Agent's knowledge base, and/or use data from your Agent's [web browsing](/agents/api-reference/browse-the-web/) tasks to enhance your Agent's capabilities. You can use the Agent's own [SQL database](/agents/api-reference/store-and-sync-state) as the source of truth for your data and store embeddings in [Vectorize](/vectorize/) (or any other vector-enabled database) to allow your Agent to retrieve relevant information. ### Vector search :::note If you're brand-new to vector databases and Vectorize, visit the [Vectorize tutorial](/vectorize/get-started/intro/) to learn the basics, including how to create an index, insert data, and generate embeddings. ::: You can query a vector index (or indexes) from any method on your Agent: any Vectorize index you attach is available on `this.env` within your Agent. If you've [associated metadata](/vectorize/best-practices/insert-vectors/#metadata) with your vectors that maps back to data stored in your Agent, you can then look up the data directly within your Agent using `this.sql`. Here's an example of how to give an Agent retrieval capabilties: <TypeScriptExample> ```ts import { Agent } from "agents"; interface Env { AI: Ai; VECTOR_DB: Vectorize; } export class RAGAgent extends Agent<Env> { // Other methods on our Agent // ... // async queryKnowledge(userQuery: string) { // Turn a query into an embedding const queryVector = await this.env.AI.run('@cf/baai/bge-base-en-v1.5', { text: [userQuery], }); // Retrieve results from our vector index let searchResults = await this.env.VECTOR_DB.query(queryVector.data[0], { topK: 10, returnMetadata: 'all', }); let knowledge = []; for (const match of searchResults.matches) { console.log(match.metadata); knowledge.push(match.metadata); } // Use the metadata to re-associate the vector search results // with data in our Agent's SQL database let results = this.sql`SELECT * FROM knowledge WHERE id IN (${knowledge.map((k) => k.id)})`; // Return them return results; } } ``` </TypeScriptExample> You'll also need to connect your Agent to your vector indexes: <WranglerConfig> ```jsonc { // ... "vectorize": [ { "binding": "VECTOR_DB", "index_name": "your-vectorize-index-name" } ] // ... } ``` </WranglerConfig> If you have multiple indexes you want to make available, you can provide an array of `vectorize` bindings. #### Next steps * Learn more on how to [combine Vectorize and Workers AI](/vectorize/get-started/embeddings/) * Review the [Vectorize query API](/vectorize/reference/client-api/) * Use [metadata filtering](/vectorize/reference/metadata-filtering/) to add context to your results --- # Run Workflows URL: https://developers.cloudflare.com/agents/api-reference/run-workflows/ import { MetaInfo, Render, Type, TypeScriptExample, WranglerConfig } from "~/components"; Agents can trigger asynchronous [Workflows](/workflows/), allowing your Agent to run complex, multi-step tasks in the background. This can include post-processing files that a user has uploaded, updating the embeddings in a [vector database](/vectorize/), and/or managing long-running user-lifecycle email or SMS notification workflows. Because an Agent is just like a Worker script, it can create Workflows defined in the same project (script) as the Agent _or_ in a different project. :::note[Agents vs. Workflows] Agents and Workflows have some similarities: they can both run tasks asynchronously. For straightforward tasks that are linear or need to run to completion, a Workflow can be ideal: steps can be retried, they can be cancelled, and can act on events. Agents do not have to run to completion: they can loop, branch and run forever, and they can also interact directly with users (over HTTP or WebSockets). An Agent can be used to trigger multiple Workflows as it runs, and can thus be used to co-ordinate and manage Workflows to achieve its goals. ::: ## Trigger a Workflow An Agent can trigger one or more Workflows from within any method, whether from an incoming HTTP request, a WebSocket connection, on a delay or schedule, and/or from any other action the Agent takes. Triggering a Workflow from an Agent is no different from [triggering a Workflow from a Worker script](/workflows/build/trigger-workflows/): <TypeScriptExample> ```ts interface Env { MY_WORKFLOW: Workflow; MyAgent: AgentNamespace<MyAgent>; } export class MyAgent extends Agent<Env> { async onRequest(request: Request) { let userId = request.headers.get("user-id"); // Trigger a schedule that runs a Workflow // Pass it a payload let { taskId } = await this.schedule(300, "runWorkflow", { id: userId, flight: "DL264", date: "2025-02-23" }); } async runWorkflow(data) { let instance = await env.MY_WORKFLOW.create({ id: data.id, params: data, }) // Schedule another task that checks the Workflow status every 5 minutes... await this.schedule("*/5 * * * *", "checkWorkflowStatus", { id: instance.id }); } } export class MyWorkflow extends WorkflowEntrypoint<Env> { async run(event: WorkflowEvent<Params>, step: WorkflowStep) { // Your Workflow code here } } ``` </TypeScriptExample> You'll also need to make sure your Agent [has a binding to your Workflow](/workflows/build/trigger-workflows/#workers-api-bindings) so that it can call it: <WranglerConfig> ```jsonc { // ... // Create a binding between your Agent and your Workflow "workflows": [ { // Required: "name": "EMAIL_WORKFLOW", "class_name": "MyWorkflow", // Optional: set the script_name field if your Workflow is defined in a // different project from your Agent "script_name": "email-workflows" } ], // ... } ``` </WranglerConfig> ## Trigger a Workflow from another project You can also call a Workflow that is defined in a different Workers script from your Agent by setting the `script_name` property in the `workflows` binding of your Agent: <WranglerConfig> ```jsonc { // Required: "name": "EMAIL_WORKFLOW", "class_name": "MyWorkflow", // Optional: set tthe script_name field if your Workflow is defined in a // different project from your Agent "script_name": "email-workflows" } ``` </WranglerConfig> Refer to the [cross-script calls](/workflows/build/workers-api/#cross-script-calls) section of the Workflows documentation for more examples. --- # Schedule tasks URL: https://developers.cloudflare.com/agents/api-reference/schedule-tasks/ import { MetaInfo, Render, Type, TypeScriptExample, WranglerConfig } from "~/components"; An Agent can schedule tasks to be run in the future by calling `this.schedule(when, callback, data)`, where `when` can be a delay, a `Date`, or a cron string; `callback` the function name to call, and `data` is an object of data to pass to the function. Scheduled tasks can do anything a request or message from a user can: make requests, query databases, send emails, read+write state: scheduled tasks can invoke any regular method on your Agent. ### Scheduling tasks You can call `this.schedule` within any method on an Agent, and schedule tens-of-thousands of tasks per individual Agent: <TypeScriptExample> ```ts import { Agent } from "agents" export class SchedulingAgent extends Agent { async onRequest(request) { // Handle an incoming request // Schedule a task 5 minutes from now // Calls the "checkFlights" method let { taskId } = await this.schedule(600, "checkFlights", { flight: "DL264", date: "2025-02-23" }); return Response.json({ taskId }); } async checkFlights(data) { // Invoked when our scheduled task runs // We can also call this.schedule here to schedule another task } } ``` </TypeScriptExample> :::caution Tasks that set a callback for a method that does not exist will throw an exception: ensure that the method named in the `callback` argument of `this.schedule` exists on your `Agent` class. ::: You can schedule tasks in multiple ways: <TypeScriptExample> ```ts // schedule a task to run in 10 seconds let task = await this.schedule(10, "someTask", { message: "hello" }); // schedule a task to run at a specific date let task = await this.schedule(new Date("2025-01-01"), "someTask", {}); // schedule a task to run every 10 seconds let { id } = await this.schedule("*/10 * * * *", "someTask", { message: "hello" }); // schedule a task to run every 10 seconds, but only on Mondays let task = await this.schedule("0 0 * * 1", "someTask", { message: "hello" }); // cancel a scheduled task this.cancelSchedule(task.id); ``` </TypeScriptExample> Calling `await this.schedule` returns a `Schedule`, which includes the task's randomly generated `id`. You can use this `id` to retrieve or cancel the task in the future. It also provides a `type` property that indicates the type of schedule, for example, one of `"scheduled" | "delayed" | "cron"`. :::note[Maximum scheduled tasks] Each task is mapped to a row in the Agent's underlying [SQLite database](/durable-objects/api/sql-storage/), which means that each task can be up to 2 MB in size. The maximum number of tasks must be `(task_size * tasks) + all_other_state < maximum_database_size` (currently 1GB per Agent). ::: ### Managing scheduled tasks You can get, cancel and filter across scheduled tasks within an Agent using the scheduling API: <TypeScriptExample> ```ts // Get a specific schedule by ID // Returns undefined if the task does not exist let task = await this.getSchedule(task.id) // Get all scheduled tasks // Returns an array of Schedule objects let tasks = this.getSchedules(); // Cancel a task by its ID // Returns true if the task was cancelled, false if it did not exist await this.cancelSchedule(task.id); // Filter for specific tasks // e.g. all tasks starting in the next hour let tasks = this.getSchedules({ timeRange: { start: new Date(Date.now()), end: new Date(Date.now() + 60 * 60 * 1000), } }); ``` </TypeScriptExample> --- # Store and sync state URL: https://developers.cloudflare.com/agents/api-reference/store-and-sync-state/ import { MetaInfo, Render, Type, TypeScriptExample, WranglerConfig } from "~/components"; Every Agent has built-in state management capabilities, including built-in storage and synchronization between the Agent and frontend applications. State within an Agent is: * Persisted across Agent restarts: data is permanently stored within an Agent. * Automatically serialized/deserialized: you can store any JSON-serializable data. * Immediately consistent within the Agent: read your own writes. * Thread-safe for concurrent updates * Fast: state is colocated wherever the Agent is running. Reads and writes do not need to traverse the network. Agent state is stored in a SQL database that is embedded within each individual Agent instance: you can interact with it using the higher-level `this.setState` API (recommended), which allows you to sync state and trigger events on state changes, or by directly querying the database with `this.sql`. #### State API Every Agent has built-in state management capabilities. You can set and update the Agent's state directly using `this.setState`: <TypeScriptExample> ```ts import { Agent } from "agents"; export class MyAgent extends Agent { // Update state in response to events async incrementCounter() { this.setState({ ...this.state, counter: this.state.counter + 1, }); } // Handle incoming messages async onMessage(message) { if (message.type === "update") { this.setState({ ...this.state, ...message.data, }); } } // Handle state updates onStateUpdate(state, source: "server" | Connection) { console.log("state updated", state); } } ``` </TypeScriptExample> If you're using TypeScript, you can also provide a type for your Agent's state by passing in a type as a [type parameter](https://www.typescriptlang.org/docs/handbook/2/generics.html#using-type-parameters-in-generic-constraints) as the _second_ type parameter to the `Agent` class definition. <TypeScriptExample> ```ts import { Agent } from "agents"; interface Env {} // Define a type for your Agent's state interface FlightRecord { id: string; departureIata: string; arrival: Date; arrivalIata: string; price: number; } // Pass in the type of your Agent's state export class MyAgent extends Agent<Env, FlightRecord> { // This allows this.setState and the onStateUpdate method to // be typed: async onStateUpdate(state: FlightRecord) { console.log("state updated", state); } async someOtherMethod() { this.setState({ ...this.state, price: this.state.price + 10, }); } } ``` </TypeScriptExample> ### Set the initial state for an Agent You can also set the initial state for an Agent via the `initialState` property on the `Agent` class: <TypeScriptExample> ```ts type State = { counter: number; text: string; color: string; }; class MyAgent extends Agent<Env, State> { // Set a default, initial state initialState = { counter: 0, text: "", color: "#3B82F6", }; doSomething() { console.log(this.state); // {counter: 0, text: "", color: "#3B82F6"}, if you haven't set the state yet } } ``` </TypeScriptExample> Any initial state is synced to clients connecting via [the `useAgent` hook](#synchronizing-state). ### Synchronizing state Clients can connect to an Agent and stay synchronized with its state using the React hooks provided as part of `agents/react`. A React application can call `useAgent` to connect to a named Agent over WebSockets at <TypeScriptExample> ```ts import { useState } from "react"; import { useAgent } from "agents/react"; function StateInterface() { const [state, setState] = useState({ counter: 0 }); const agent = useAgent({ agent: "thinking-agent", name: "my-agent", onStateUpdate: (newState) => setState(newState), }); const increment = () => { agent.setState({ counter: state.counter + 1 }); }; return ( <div> <div>Count: {state.counter}</div> <button onClick={increment}>Increment</button> </div> ); } ``` </TypeScriptExample> The state synchronization system: * Automatically syncs the Agent's state to all connected clients * Handles client disconnections and reconnections gracefully * Provides immediate local updates * Supports multiple simultaneous client connections Common use cases: * Real-time collaborative features * Multi-window/tab synchronization * Live updates across multiple devices * Maintaining consistent UI state across clients * When new clients connect, they automatically receive the current state from the Agent, ensuring all clients start with the latest data. ### SQL API Every individual Agent instance has its own SQL (SQLite) database that runs _within the same context_ as the Agent itself. This means that inserting or querying data within your Agent is effectively zero-latency: the Agent doesn't have to round-trip across a continent or the world to access its own data. You can access the SQL API within any method on an Agent via `this.sql`. The SQL API accepts template literals, and <TypeScriptExample> ```ts export class MyAgent extends Agent<Env> { async onRequest(request: Request) { let userId = new URL(request.url).searchParams.get('userId'); // 'users' is just an example here: you can create arbitrary tables and define your own schemas // within each Agent's database using SQL (SQLite syntax). let user = await this.sql`SELECT * FROM users WHERE id = ${userId}` return Response.json(user) } } ``` </TypeScriptExample> You can also supply a [TypeScript type argument](https://www.typescriptlang.org/docs/handbook/2/generics.html#using-type-parameters-in-generic-constraints) the query, which will be used to infer the type of the result: ```ts type User = { id: string; name: string; email: string; }; export class MyAgent extends Agent<Env> { async onRequest(request: Request) { let userId = new URL(request.url).searchParams.get('userId'); // Supply the type paramter to the query when calling this.sql // This assumes the results returns one or more User rows with "id", "name", and "email" columns const user = await this.sql<User>`SELECT * FROM users WHERE id = ${userId}`; return Response.json(user) } } ``` You do not need to specify an array type (`User[]` or `Array<User>`) as `this.sql` will always return an array of the specified type. Providing a type parameter does not validate that the result matches your type definition. In TypeScript, properties (fields) that do not exist or conform to the type you provided will be dropped. If you need to validate incoming events, we recommend a library such as [zod](https://zod.dev/) or your own validator logic. :::note Learn more about the zero-latency SQL storage that powers both Agents and Durable Objects [on our blog](https://blog.cloudflare.com/sqlite-in-durable-objects/). ::: The SQL API exposed to an Agent is similar to the one [within Durable Objects](/durable-objects/api/sql-storage/): Durable Object SQL methods available on `this.ctx.storage.sql`. You can use the same SQL queries with the Agent's database, create tables, and query data, just as you would with Durable Objects or [D1](/d1/). ### Use Agent state as model context You can combine the state and SQL APIs in your Agent with its ability to [call AI models](/agents/api-reference/using-ai-models/) to include historical context within your prompts to a model. Modern Large Language Models (LLMs) often have very large context windows (up to millions of tokens), which allows you to pull relevant context into your prompt directly. For example, you can use an Agent's built-in SQL database to pull history, query a model with it, and append to that history ahead of the next call to the model: <TypeScriptExample> ```ts export class ReasoningAgent extends Agent<Env> { async callReasoningModel(prompt: Prompt) { let result = this.sql<History>`SELECT * FROM history WHERE user = ${prompt.userId} ORDER BY timestamp DESC LIMIT 1000`; let context = []; for await (const row of result) { context.push(row.entry); } const client = new OpenAI({ apiKey: this.env.OPENAI_API_KEY, }); // Combine user history with the current prompt const systemPrompt = prompt.system || 'You are a helpful assistant.'; const userPrompt = `${prompt.user}\n\nUser history:\n${context.join('\n')}`; try { const completion = await client.chat.completions.create({ model: this.env.MODEL || 'o3-mini', messages: [ { role: 'system', content: systemPrompt }, { role: 'user', content: userPrompt }, ], temperature: 0.7, max_tokens: 1000, }); // Store the response in history this .sql`INSERT INTO history (timestamp, user, entry) VALUES (${new Date()}, ${prompt.userId}, ${completion.choices[0].message.content})`; return completion.choices[0].message.content; } catch (error) { console.error('Error calling reasoning model:', error); throw error; } } } ``` </TypeScriptExample> This works because each instance of an Agent has its _own_ database, the state stored in that database is private to that Agent: whether it's acting on behalf of a single user, a room or channel, or a deep research tool. By default, you don't have to manage contention or reach out over the network to a centralized database to retrieve and store state. ### Next steps * Review the [API documentation](/agents/api-reference/agents-api/) for the Agents class to learn how to define them. * [Build a chat Agent](/agents/getting-started/build-a-chat-agent/) using the Agents SDK and deploy it to Workers. * Learn more [using WebSockets](/agents/api-reference/websockets/) to build interactive Agents and stream data back from your Agent. * [Orchestrate asynchronous workflows](/agents/api-reference/run-workflows) from your Agent by combining the Agents SDK and [Workflows](/workflows). --- # Using AI Models URL: https://developers.cloudflare.com/agents/api-reference/using-ai-models/ import { AnchorHeading, MetaInfo, Render, Type, TypeScriptExample, WranglerConfig } from "~/components"; Agents can communicate with AI models hosted on any provider, including: * [Workers AI](/workers-ai/) * The [AI SDK](https://sdk.vercel.ai/docs/ai-sdk-core/overview) * [OpenAI](https://platform.openai.com/docs/quickstart?language=javascript) * [Anthropic](https://docs.anthropic.com/en/api/client-sdks#typescript) * [Google's Gemini](https://ai.google.dev/gemini-api/docs/openai) You can also use the model routing features in [AI Gateway](/ai-gateway/) to route across providers, eval responses, and manage AI provider rate limits. Because Agents are built on top of [Durable Objects](/durable-objects/), each Agent or chat session is associated with a stateful compute instance. Traditional serverless architectures often present challenges for persistent connections needed in real-time applications like chat. A user can disconnect during a long-running response from a modern reasoning model (such as `o3-mini` or DeepSeek R1), or lose conversational context when refreshing the browser. Instead of relying on request-response patterns and managing an external database to track & store conversation state, state can be stored directly within the Agent. If a client disconnects, the Agent can write to its own distributed storage, and catch the client up as soon as it reconnects: even if it's hours or days later. ## Calling AI Models You can call models from any method within an Agent, including from HTTP requests using the [`onRequest`](/agents/api-reference/agents-api/) handler, when a [scheduled task](/agents/api-reference/schedule-tasks/) runs, when handling a WebSocket message in the [`onMessage`](/agents/api-reference/websockets/) handler, or from any of your own methods. Importantly, Agents can call AI models on their own — autonomously — and can handle long-running responses that can take minutes (or longer) to respond in full. ### Long-running model requests {/*long-running-model-requests*/} Modern [reasoning models](https://platform.openai.com/docs/guides/reasoning) or "thinking" model can take some time to both generate a response _and_ stream the response back to the client. Instead of buffering the entire response, or risking the client disconecting, you can stream the response back to the client by using the [WebSocket API](/agents/api-reference/websockets/). <TypeScriptExample filename="src/index.ts"> ```ts import { Agent } from "agents" import { OpenAI } from "openai" export class MyAgent extends Agent<Env> { async onConnect(connection: Connection, ctx: ConnectionContext) { // } async onMessage(connection: Connection, message: WSMessage) { let msg = JSON.parse(message) // This can run as long as it needs to, and return as many messages as it needs to! await queryReasoningModel(connection, msg.prompt) } async queryReasoningModel(connection: Connection, userPrompt: string) { const client = new OpenAI({ apiKey: this.env.OPENAI_API_KEY, }); try { const stream = await client.chat.completions.create({ model: this.env.MODEL || 'o3-mini', messages: [{ role: 'user', content: userPrompt }], stream: true, }); // Stream responses back as WebSocket messages for await (const chunk of stream) { const content = chunk.choices[0]?.delta?.content || ''; if (content) { connection.send(JSON.stringify({ type: 'chunk', content })); } } // Send completion message connection.send(JSON.stringify({ type: 'done' })); } catch (error) { connection.send(JSON.stringify({ type: 'error', error: error })); } } } ``` </TypeScriptExample> You can also persist AI model responses back to [Agent's internal state](/agents/api-reference/store-and-sync-state/) by using the `this.setState` method. For example, if you run a [scheduled task](/agents/api-reference/schedule-tasks/), you can store the output of the task and read it later. Or, if a user disconnects, read the message history back and send it to the user when they reconnect. ### Workers AI ### Hosted models You can use [any of the models available in Workers AI](/workers-ai/models/) within your Agent by [configuring a binding](/workers-ai/configuration/bindings/). Workers AI supports streaming responses out-of-the-box by setting `stream: true`, and we strongly recommend using them to avoid buffering and delaying responses, especially for larger models or reasoning models that require more time to generate a response. <TypeScriptExample filename="src/index.ts"> ```ts import { Agent } from "agents" interface Env { AI: Ai; } export class MyAgent extends Agent<Env> { async onRequest(request: Request) { const response = await env.AI.run( "@cf/deepseek-ai/deepseek-r1-distill-qwen-32b", { prompt: "Build me a Cloudflare Worker that returns JSON.", stream: true, // Stream a response and don't block the client! } ); // Return the stream return new Response(answer, { headers: { "content-type": "text/event-stream" } }) } } ``` </TypeScriptExample> Your wrangler configuration will need an `ai` binding added: <WranglerConfig> ```toml [ai] binding = "AI" ``` </WranglerConfig> ### Model routing You can also use the model routing features in [AI Gateway](/ai-gateway/) directly from an Agent by specifying a [`gateway` configuration](/ai-gateway/providers/workersai/) when calling the AI binding. :::note Model routing allows you to route requests to different AI models based on whether they are reachable, rate-limiting your client, and/or if you've exceeded your cost budget for a specific provider. ::: <TypeScriptExample filename="src/index.ts"> ```ts import { Agent } from "agents" interface Env { AI: Ai; } export class MyAgent extends Agent<Env> { async onRequest(request: Request) { const response = await env.AI.run( "@cf/deepseek-ai/deepseek-r1-distill-qwen-32b", { prompt: "Build me a Cloudflare Worker that returns JSON." }, { gateway: { id: "{gateway_id}", // Specify your AI Gateway ID here skipCache: false, cacheTtl: 3360, }, }, ); return Response.json(response) } } ``` </TypeScriptExample> Your wrangler configuration will need an `ai` binding added. This is shared across both Workers AI and AI Gateway. <WranglerConfig> ```toml [ai] binding = "AI" ``` </WranglerConfig> Visit the [AI Gateway documentation](/ai-gateway/) to learn how to configure a gateway and retrieve a gateway ID. ### AI SDK The [AI SDK](https://sdk.vercel.ai/docs/introduction) provides a unified API for using AI models, including for text generation, tool calling, structured responses, image generation, and more. To use the AI SDK, install the `ai` package and use it within your Agent. The example below shows how it use it to generate text on request, but you can use it from any method within your Agent, including WebSocket handlers, as part of a scheduled task, or even when the Agent is initialized. ```sh npm install ai @ai-sdk/openai ``` <TypeScriptExample filename="src/index.ts"> ```ts import { Agent } from "agents" import { generateText } from 'ai'; import { openai } from '@ai-sdk/openai'; export class MyAgent extends Agent<Env> { async onRequest(request: Request): Promise<Response> { const { text } = await generateText({ model: openai("o3-mini"), prompt: "Build me an AI agent on Cloudflare Workers", }); return Response.json({modelResponse: text}) } } ``` </TypeScriptExample> ### OpenAI compatible endpoints Agents can call models across any service, including those that support the OpenAI API. For example, you can use the OpenAI SDK to use one of [Google's Gemini models](https://ai.google.dev/gemini-api/docs/openai#node.js) directly from your Agent. Agents can stream responses back over HTTP using Server Sent Events (SSE) from within an `onRequest` handler, or by using the native [WebSockets](/agents/api-reference/websockets/) API in your Agent to responses back to a client, which is especially useful for larger models that can take over 30+ seconds to reply. <TypeScriptExample filename="src/index.ts"> ```ts import { Agent } from "agents" import { OpenAI } from "openai" export class MyAgent extends Agent<Env> { async onRequest(request: Request): Promise<Response> { const openai = new OpenAI({ apiKey: this.env.GEMINI_API_KEY, baseURL: "https://generativelanguage.googleapis.com/v1beta/openai/" }); // Create a TransformStream to handle streaming data let { readable, writable } = new TransformStream(); let writer = writable.getWriter(); const textEncoder = new TextEncoder(); // Use ctx.waitUntil to run the async function in the background // so that it doesn't block the streaming response ctx.waitUntil( (async () => { const stream = await openai.chat.completions.create({ model: "4o", messages: [{ role: "user", content: "Write me a Cloudflare Worker." }], stream: true, }); // loop over the data as it is streamed and write to the writeable for await (const part of stream) { writer.write( textEncoder.encode(part.choices[0]?.delta?.content || ""), ); } writer.close(); })(), ); // Return the readable stream back to the client return new Response(readable) } } ``` </TypeScriptExample> --- # Using WebSockets URL: https://developers.cloudflare.com/agents/api-reference/websockets/ import { MetaInfo, Render, Type, TypeScriptExample, WranglerConfig } from "~/components"; Users and clients can connect to an Agent directly over WebSockets, allowing long-running, bi-directional communication with your Agent as it operates. To enable an Agent to accept WebSockets, define `onConnect` and `onMessage` methods on your Agent. * `onConnect(connection: Connection, ctx: ConnectionContext)` is called when a client establishes a new WebSocket connection. The original HTTP request, including request headers, cookies, and the URL itself, are available on `ctx.request`. * `onMessage(connection: Connection, message: WSMessage)` is called for each incoming WebSocket message. Messages are one of `ArrayBuffer | ArrayBufferView | string`, and you can send messages back to a client using `connection.send()`. You can distinguish between client connections by checking `connection.id`, which is unique for each connected client. Here's an example of an Agent that echoes back any message it receives: <TypeScriptExample> ```ts import { Agent, Connection } from "agents"; export class ChatAgent extends Agent { async onConnect(connection: Connection, ctx: ConnectionContext) { // Connections are automatically accepted by the SDK. // You can also explicitly close a connection here with connection.close() // Access the Request on ctx.request to inspect headers, cookies and the URL } async onMessage(connection: Connection, message: WSMessage) { // const response = await longRunningAITask(message) await connection.send(message) } } ``` </TypeScriptExample> ### Connecting clients The Agent framework includes a useful helper package for connecting directly to your Agent (or other Agents) from a client application. Import `agents/client`, create an instance of `AgentClient` and use it to connect to an instance of your Agent: <TypeScriptExample> ```ts import { AgentClient } from "agents/client"; const connection = new AgentClient({ agent: "dialogue-agent", name: "insight-seeker", }); connection.addEventListener("message", (event) => { console.log("Received:", event.data); }); connection.send( JSON.stringify({ type: "inquiry", content: "What patterns do you see?", }) ); ``` </TypeScriptExample> ### React clients React-based applications can import `agents/react` and use the `useAgent` hook to connect to an instance of an Agent directly: <TypeScriptExample> ```ts import { useAgent } from "agents/react"; function AgentInterface() { const connection = useAgent({ agent: "dialogue-agent", name: "insight-seeker", onMessage: (message) => { console.log("Understanding received:", message.data); }, onOpen: () => console.log("Connection established"), onClose: () => console.log("Connection closed"), }); const inquire = () => { connection.send( JSON.stringify({ type: "inquiry", content: "What insights have you gathered?", }) ); }; return ( <div className="agent-interface"> <button onClick={inquire}>Seek Understanding</button> </div> ); } ``` </TypeScriptExample> The `useAgent` hook automatically handles the lifecycle of the connection, ensuring that it is properly initialized and cleaned up when the component mounts and unmounts. You can also [combine `useAgent` with `useState`](/agents/api-reference/store-and-sync-state/) to automatically synchronize state across all clients connected to your Agent. ### Handling WebSocket events Define `onError` and `onClose` methods on your Agent to explicitly handle WebSocket client errors and close events. Log errors, clean up state, and/or emit metrics: <TypeScriptExample> ```ts import { Agent, Connection } from "agents"; export class ChatAgent extends Agent { // onConnect and onMessage methods // ... // WebSocket error and disconnection (close) handling. async onError(connection: Connection, error: unknown): Promise<void> { console.error(`WS error: ${error}`); } async onClose(connection: Connection, code: number, reason: string, wasClean: boolean): Promise<void> { console.log(`WS closed: ${code} - ${reason} - wasClean: ${wasClean}`); connection.close(); } } ``` </TypeScriptExample> --- # Getting started URL: https://developers.cloudflare.com/agents/getting-started/ import { DirectoryListing } from "~/components" <DirectoryListing /> --- # Testing your Agents URL: https://developers.cloudflare.com/agents/getting-started/testing-your-agent/ import { Render, PackageManagers, WranglerConfig } from "~/components" Because Agents run on Cloudflare Workers and Durable Objects, they can be tested using the same tools and techniques as Workers and Durable Objects. ## Writing and running tests ### Setup :::note The `agents-starter` template and new Cloudflare Workers projects already include the relevant `vitest` and `@cloudflare/vitest-pool-workers` packages, as well as a valid `vitest.config.js` file. ::: Before you write your first test, install the necessary packages: ```sh npm install vitest@~3.0.0 --save-dev --save-exact npm install @cloudflare/vitest-pool-workers --save-dev ``` Ensure that your `vitest.config.js` file is identical to the following: ```js import { defineWorkersConfig } from "@cloudflare/vitest-pool-workers/config"; export default defineWorkersConfig({ test: { poolOptions: { workers: { wrangler: { configPath: "./wrangler.toml" }, }, }, }, }); ``` ### Add the Agent configuration Add a `durableObjects` configuration to `vitest.config.js` with the name of your Agent class: ```js import { defineWorkersConfig } from '@cloudflare/vitest-pool-workers/config'; export default defineWorkersConfig({ test: { poolOptions: { workers: { main: './src/index.ts', miniflare: { durableObjects: { NAME: 'MyAgent', }, }, }, }, }, }); ``` ### Write a test :::note Review the [Vitest documentation](https://vitest.dev/) for more information on testing, including the test API reference and advanced testing techniques. ::: Tests use the `vitest` framework. A basic test suite for your Agent can validate how your Agent responds to requests, but can also unit test your Agent's methods and state. ```ts import { env, createExecutionContext, waitOnExecutionContext, SELF } from 'cloudflare:test'; import { describe, it, expect } from 'vitest'; import worker from '../src'; import { Env } from '../src'; interface ProvidedEnv extends Env {} describe('make a request to my Agent', () => { // Unit testing approach it('responds with state', async () => { // Provide a valid URL that your Worker can use to route to your Agent // If you are using routeAgentRequest, this will be /agent/:agent/:name const request = new Request<unknown, IncomingRequestCfProperties>('http://example.com/agent/my-agent/agent-123'); const ctx = createExecutionContext(); const response = await worker.fetch(request, env, ctx); await waitOnExecutionContext(ctx); expect(await response.text()).toMatchObject({ hello: 'from your agent' }); }); it('also responds with state', async () => { const request = new Request('http://example.com/agent/my-agent/agent-123'); const response = await SELF.fetch(request); expect(await response.text()).toMatchObject({ hello: 'from your agent' }); }); }); ``` ### Run tests Running tests is done using the `vitest` CLI: ```sh $ npm run test # or run vitest directly $ npx vitest ``` ```sh output MyAgent ✓ should return a greeting (1 ms) Test Files 1 passed (1) ``` Review the [documentation on testing](/workers/testing/vitest-integration/get-started/write-your-first-test/) for additional examples and test configuration. ## Running Agents locally You can also run an Agent locally using the `wrangler` CLI: ```sh $ npx wrangler dev ``` ```sh output Your Worker and resources are simulated locally via Miniflare. For more information, see: https://developers.cloudflare.com/workers/testing/local-development. Your worker has access to the following bindings: - Durable Objects: - MyAgent: MyAgent Starting local server... [wrangler:inf] Ready on http://localhost:53645 ``` This spins up a local development server that runs the same runtime as Cloudflare Workers, and allows you to iterate on your Agent's code and test it locally without deploying it. Visit the [`wrangler dev`](https://developers.cloudflare.com/workers/wrangler/commands/#dev) docs to review the CLI flags and configuration options. --- # Calling LLMs URL: https://developers.cloudflare.com/agents/concepts/calling-llms/ import { Render } from "~/components"; ### Understanding LLM providers and model types Different LLM providers offer models optimized for specific types of tasks. When building AI systems, choosing the right model is crucial for both performance and cost efficiency. #### Reasoning Models Models like OpenAI's o1, Anthropic's Claude, and DeepSeek's R1 are particularly well-suited for complex reasoning tasks. These models excel at: - Breaking down problems into steps - Following complex instructions - Maintaining context across long conversations - Generating code and technical content For example, when implementing a travel booking system, you might use a reasoning model to analyze travel requirements and generate appropriate booking strategies. #### Instruction Models Models like GPT-4 and Claude Instant are optimized for following straightforward instructions efficiently. They work well for: - Content generation - Simple classification tasks - Basic question answering - Text transformation These models are often more cost-effective for straightforward tasks that do not require complex reasoning. --- # Human in the Loop URL: https://developers.cloudflare.com/agents/concepts/human-in-the-loop/ import { Render, Note, Aside } from "~/components"; ### What is Human-in-the-Loop? Human-in-the-Loop (HITL) workflows integrate human judgment and oversight into automated processes. These workflows pause at critical points for human review, validation, or decision-making before proceeding. This approach combines the efficiency of automation with human expertise and oversight where it matters most.  #### Understanding Human-in-the-Loop workflows In a Human-in-the-Loop workflow, processes are not fully automated. Instead, they include designated checkpoints where human intervention is required. For example, in a travel booking system, a human may want to confirm the travel before an agent follows through with a transaction. The workflow manages this interaction, ensuring that: 1. The process pauses at appropriate review points 2. Human reviewers receive necessary context 3. The system maintains state during the review period 4. Review decisions are properly incorporated 5. The process continues once approval is received ### Best practices for Human-in-the-Loop workflows #### Long-Term State Persistence Human review processes do not operate on predictable timelines. A reviewer might need days or weeks to make a decision, especially for complex cases requiring additional investigation or multiple approvals. Your system needs to maintain perfect state consistency throughout this period, including: - The original request and context - All intermediate decisions and actions - Any partial progress or temporary states - Review history and feedback :::note[Tip] [Durable Objects](/durable-objects/) provide an ideal solution for managing state in Human-in-the-Loop workflows, offering persistent compute instances that maintain state for hours, weeks, or months. ::: #### Continuous Improvement Through Evals Human reviewers play a crucial role in evaluating and improving LLM performance. Implement a systematic evaluation process where human feedback is collected not just on the final output, but on the LLM's decision-making process. This can include: - Decision Quality Assessment: Have reviewers evaluate the LLM's reasoning process and decision points, not just the final output. - Edge Case Identification: Use human expertise to identify scenarios where the LLM's performance could be improved. - Feedback Collection: Gather structured feedback that can be used to fine-tune the LLM or adjust the workflow. [AI Gateway](/ai-gateway/evaluations/add-human-feedback/) can be a useful tool for setting up an LLM feedback loop. #### Error handling and recovery Robust error handling is essential for maintaining workflow integrity. Your system should gracefully handle various failure scenarios, including reviewer unavailability, system outages, or conflicting reviews. Implement clear escalation paths for handling exceptional cases that fall outside normal parameters. The system should maintain stability during paused states, ensuring that no work is lost even during extended review periods. Consider implementing automatic checkpointing that allows workflows to be resumed from the last stable state after any interruption. --- # Tools URL: https://developers.cloudflare.com/agents/concepts/tools/ ### What are tools? Tools enable AI systems to interact with external services and perform actions. They provide a structured way for agents and workflows to invoke APIs, manipulate data, and integrate with external systems. Tools form the bridge between AI decision-making capabilities and real-world actions. ### Understanding tools In an AI system, tools are typically implemented as function calls that the AI can use to accomplish specific tasks. For example, a travel booking agent might have tools for: - Searching flight availability - Checking hotel rates - Processing payments - Sending confirmation emails Each tool has a defined interface specifying its inputs, outputs, and expected behavior. This allows the AI system to understand when and how to use each tool appropriately. ### Common tool patterns #### API integration tools The most common type of tools are those that wrap external APIs. These tools handle the complexity of API authentication, request formatting, and response parsing, presenting a clean interface to the AI system. #### Model Context Protocol (MCP) The [Model Context Protocol](https://modelcontextprotocol.io/introduction) provides a standardized way to define and interact with tools. Think of it as an abstraction on top of APIs designed for LLMs to interact with external resources. MCP defines a consistent interface for: - **Tool Discovery**: Systems can dynamically discover available tools - **Parameter Validation**: Tools specify their input requirements using JSON Schema - **Error Handling**: Standardized error reporting and recovery - **State Management**: Tools can maintain state across invocations #### Data processing tools Tools that handle data transformation and analysis are essential for many AI workflows. These might include: - CSV parsing and analysis - Image processing - Text extraction - Data validation --- # Concepts URL: https://developers.cloudflare.com/agents/concepts/ import { DirectoryListing } from "~/components"; <DirectoryListing /> --- # Agents URL: https://developers.cloudflare.com/agents/concepts/what-are-agents/ import { Render } from "~/components"; ### What are agents? An agent is an AI system that can autonomously execute tasks by making decisions about tool usage and process flow. Unlike traditional automation that follows predefined paths, agents can dynamically adapt their approach based on context and intermediate results. Agents are also distinct from co-pilots (e.g. traditional chat applications) in that they can fully automate a task, as opposed to simply augmenting and extending human input. - **Agents** → non-linear, non-deterministic (can change from run to run) - **Workflows** → linear, deterministic execution paths - **Co-pilots** → augmentative AI assistance requiring human intervention ### Example: Booking vacations If this is your first time working with, or interacting with agents, this example will illustrate how an agent works within a context like booking a vacation. If you are already familiar with the topic, read on. Imagine you're trying to book a vacation. You need to research flights, find hotels, check restaurant reviews, and keep track of your budget. #### Traditional workflow automation A traditional automation system follows a predetermined sequence: - Takes specific inputs (dates, location, budget) - Calls predefined API endpoints in a fixed order - Returns results based on hardcoded criteria - Cannot adapt if unexpected situations arise  #### AI Co-pilot A co-pilot acts as an intelligent assistant that: - Provides hotel and itinerary recommendations based on your preferences - Can understand and respond to natural language queries - Offers guidance and suggestions - Requires human decision-making and action for execution  #### Agent An agent combines AI's ability to make judgements and call the relevant tools to execute the task. An agent's output will be nondeterministic given: - Real-time availability and pricing changes - Dynamic prioritization of constraints - Ability to recover from failures - Adaptive decision-making based on intermediate results  An agents can dynamically generate an itinerary and execute on booking reservations, similarly to what you would expect from a travel agent. ### Three primary components of agent systems: - **Decision Engine**: Usually an LLM (Large Language Model) that determines action steps - **Tool Integration**: APIs, functions, and services the agent can utilize - **Memory System**: Maintains context and tracks task progress #### How agents work Agents operate in a continuous loop of: 1. **Observing** the current state or task 2. **Planning** what actions to take, using AI for reasoning 3. **Executing** those actions using available tools (often APIs or [MCPs](https://modelcontextprotocol.io/introduction)) 4. **Learning** from the results (storing results in memory, updating task progress, and preparing for next iteration) --- # Workflows URL: https://developers.cloudflare.com/agents/concepts/workflows/ import { Render } from "~/components"; ## What are workflows? A workflow is the orchestration layer that coordinates how an agent's components work together. It defines the structured paths through which tasks are processed, tools are called, and results are managed. While agents make dynamic decisions about what to do, workflows provide the underlying framework that governs how those decisions are executed. ### Understanding workflows in agent systems Think of a workflow like the operating procedures of a company. The company (agent) can make various decisions, but how those decisions get implemented follows established processes (workflows). For example, when you book a flight through a travel agent, they might make different decisions about which flights to recommend, but the process of actually booking the flight follows a fixed sequence of steps. Let's examine a basic agent workflow: ### Core components of a workflow A workflow typically consists of several key elements: 1. **Input Processing** The workflow defines how inputs are received and validated before being processed by the agent. This includes standardizing formats, checking permissions, and ensuring all required information is present. 2. **Tool Integration** Workflows manage how external tools and services are accessed. They handle authentication, rate limiting, error recovery, and ensuring tools are used in the correct sequence. 3. **State Management** The workflow maintains the state of ongoing processes, tracking progress through multiple steps and ensuring consistency across operations. 4. **Output Handling** Results from the agent's actions are processed according to defined rules, whether that means storing data, triggering notifications, or formatting responses. --- # Platform URL: https://developers.cloudflare.com/agents/platform/ import { DirectoryListing } from "~/components"; <DirectoryListing /> --- # Limits URL: https://developers.cloudflare.com/agents/platform/limits/ import { Render } from "~/components" Limits that apply to authoring, deploying, and running Agents are detailed below. Many limits are inherited from those applied to Workers scripts and/or Durable Objects, and are detailed in the [Workers limits](/workers/platform/limits/) documentation. | Feature | Limit | | ----------------------------------------- | ----------------------- | | Max concurrent (running) Agents per account | Tens of millions+ [^1] | Max definitions per account | ~250,000+ [^2] | Max state stored per unique Agent | 1 GB | | Max compute time per Agent | 30 seconds (refreshed per HTTP request / incoming WebSocket message) [^3] | | Duration (wall clock) per step [^3] | Unlimited (e.g. waiting on a database call or an LLM response) | --- [^1]: Yes, really. You can have tens of millions of Agents running concurrently, as each Agent is mapped to a [unique Durable Object](/durable-objects/what-are-durable-objects/) (actor). [^2]: You can deploy up to [500 scripts per account](/workers/platform/limits/), but each script (project) can define multiple Agents. Each deployed script can be up to 10 MB on the [Workers Paid Plan](/workers/platform/pricing/#workers) [^3]: Compute (CPU) time per Agent is limited to 30 seconds, but this is refreshed when an Agent receives a new HTTP request, runs a [scheduled task](/agents/api-reference/schedule-tasks/), or an incoming WebSocket message. <Render file="limits_increase" product="workers" /> --- # Guides URL: https://developers.cloudflare.com/agents/guides/ import { DirectoryListing } from "~/components" <DirectoryListing /> --- # Build a remote MCP server URL: https://developers.cloudflare.com/agents/guides/remote-mcp-server/ import { Details, Render, PackageManagers } from "~/components"; ## Deploy your first MCP server This guide will walk you through how to deploy an [example MCP server](https://github.com/cloudflare/ai/tree/main/demos/remote-mcp-server) to your Cloudflare account. You will then customize this example to suit your needs. To get started, run the following command to create a new MCP server: <PackageManagers type="create" pkg="cloudflare@latest" args={"my-mcp-server --template=cloudflare/ai/demos/remote-mcp-server"} /> Now, you have the MCP server setup, with dependencies installed. Move into that project folder: ```sh cd my-mcp-server ``` ### Local development In the directory of your new project, run the following command to start the development server: ```sh npm start ``` Your MCP server is now running on `http://localhost:8787/sse`. In a new terminal, run the [MCP inspector](https://github.com/modelcontextprotocol/mcp-inspector). The MCP inspector is an interactive MCP client that allows you to connect to your MCP server and invoke tools from a web browser. ```sh npx @modelcontextprotocol/inspector@latest ``` Open the MCP inspector in your web browser: ```sh open http://localhost:5173 ``` In the inspector, enter the URL of your MCP server, `http://localhost:8787/sse`, and click **Connect**:  You will be redirected to an example OAuth login page. Enter any username and password and click "Log in and approve" to continue. (you can add your own authentication and/or authorization provider to replace this. Refer to the [authorization](/agents/model-context-protocol/authorization/) section for details on how to do this.)  Once you have logged in, you will be redirected back to the inspector. You should see the "List Tools" button, which will list the tools that your MCP server exposes.  ### Deploy your MCP server You can deploy your MCP server to Cloudflare using the following [Wrangler CLI command](/workers/wrangler) within the example project: ```sh npx wrangler@latest deploy ``` If you have already [connected a git repository](/workers/ci-cd/builds/) to the Worker with your MCP server, you can deploy your MCP server by pushing a change or merging a pull request to the main branch of the repository. After deploying, take the URL of your deployed MCP server, and enter it in the MCP inspector running on `http://localhost:5173`. You now have a remote MCP server, deployed to Cloudflare, that MCP clients can connect to. ### Connect your remote MCP server to Claude and other MCP Clients via a local proxy Now that your MCP server is running with OAuth authentication, you can use the [`mcp-remote` local proxy](https://www.npmjs.com/package/mcp-remote) to connect Claude Desktop or other MCP clients to it — even though these tools aren't yet _remote_ MCP clients, and don't support remote transport or authorization on the client side. This lets you to test what an interaction with your OAuth-enabled MCP server will be like with a real MCP client. Update your Claude Desktop configuration to point to the URL of your MCP server. You can use either the `localhost:8787/sse` URL, or the URL of your deployed MCP server: ```json { "mcpServers": { "math": { "command": "npx", "args": [ "mcp-remote", "https://your-worker-name.your-account.workers.dev/sse" ] } } } ``` Restart Claude Desktop and complete the authentication flow again. Once this is done, Claude will be able to make calls to your remote MCP server. you can test this by asking Claude to use one of your tools. For example: "Could you use the math tool to add 23 and 19?". Claude should invoke the tool and show the result generated by the MCP server. Learn more about other ways of using remote MCP servers with MCP clients here in [this section](/agents/guides/test-remote-mcp-server). ## Add Authentication The example MCP server you just deployed above acts as an OAuth provider to MCP clients, handling authorization, but has a placeholder authentication flow. It lets you enter any username and password to log in, and doesn't actually authenticate you against any user database. In the next section, you will add a real authentication provider to your MCP server. Following these steps will show you more clearly how to integrate it with your MCP server. We'll use GitHub in this example, but you can use any OAuth provider that supports the OAuth 2.0 specification, including Google, Slack, Stytch, Auth0, and more. ### Step 1 — Create and deploy a new MCP server Run the following command to create a new MCP server: <PackageManagers type="create" pkg="cloudflare@latest" args={ "my-mcp-server-github-auth --template=cloudflare/ai/demos/remote-mcp-github-oauth" } /> Now, you have the MCP server setup, with dependencies installed. Move into that project folder: ```sh cd my-mcp-server-github-auth ``` Then, run the following command to deploy the MCP server: ```sh npx wrangler@latest deploy ``` You'll notice that in the example MCP server, if you open `src/index.ts`, the primary difference is that the `defaultHandler` is set to the `GitHubHandler`: ```ts ins="OAuthProvider.GitHubHandler" import GitHubHandler from "./github-handler"; export default new OAuthProvider({ apiRoute: "/sse", apiHandler: MyMCP.Router, defaultHandler: GitHubHandler, authorizeEndpoint: "/authorize", tokenEndpoint: "/token", clientRegistrationEndpoint: "/register", }); ``` This will ensure that your users are redirected to GitHub to authenticate. To get this working though, you need to create OAuth client apps in the steps below. ### Step 2 — Create an OAuth App You'll need to create two [GitHub OAuth Apps](https://docs.github.com/en/apps/oauth-apps/building-oauth-apps/creating-an-oauth-app) to use GitHub as an authentication provider for your MCP server — one for local development, and one for production. #### First create a new OAuth App for local development Navigate to [github.com/settings/developers](https://github.com/settings/developers) to create a new OAuth App with the following settings: - **Application name**: `My MCP Server (local)` - **Homepage URL**: `http://localhost:8787` - **Authorization callback URL**: `http://localhost:8787/callback` For the OAuth app you just created, add the client ID of the OAuth app as `GITHUB_CLIENT_ID` and generate a client secret, adding it as `GITHUB_CLIENT_SECRET` to a `.dev.vars` file in the root of your project, which [will be used to set secrets in local development](/workers/configuration/secrets/). ```sh touch .dev.vars echo 'GITHUB_CLIENT_ID="your-client-id"' >> .dev.vars echo 'GITHUB_CLIENT_SECRET="your-client-secret"' >> .dev.vars cat .dev.vars ``` #### Next, run your MCP server locally Run the following command to start the development server: ```sh npm start ``` Your MCP server is now running on `http://localhost:8787/sse`. In a new terminal, run the [MCP inspector](https://github.com/modelcontextprotocol/mcp-inspector). The MCP inspector is an interactive MCP client that allows you to connect to your MCP server and invoke tools from a web browser. ```sh npx @modelcontextprotocol/inspector@latest ``` Open the MCP inspector in your web browser: ```sh open http://localhost:5173 ``` In the inspector, enter the URL of your MCP server, `http://localhost:8787/sse`, and click **Connect**: You should be redirected to a GitHub login or authorization page. After authorizing the MCP Client (the inspector) access to your GitHub account, you will be redirected back to the inspector. You should see the "List Tools" button, which will list the tools that your MCP server exposes. #### Second — create a new OAuth App for production You'll need to repeat these steps to create a new OAuth App for production. Navigate to [github.com/settings/developers](https://github.com/settings/developers) to create a new OAuth App with the following settings: - **Application name**: `My MCP Server (production)` - **Homepage URL**: Enter the workers.dev URL of your deployed MCP server (ex: `worker-name.account-name.workers.dev`) - **Authorization callback URL**: Enter the `/callback` path of the workers.dev URL of your deployed MCP server (ex: `worker-name.account-name.workers.dev/callback`) For the OAuth app you just created, add the client ID and client secret, using Wrangler CLI: ```sh wrangler secret put GITHUB_CLIENT_ID ``` ```sh wrangler secret put GITHUB_CLIENT_SECRET ``` #### Finally, connect to your MCP server Now that you've added the ID and secret of your production OAuth app, you should now be able to connect to your MCP server running at `worker-name.account-name.workers.dev/sse` using the MCP inspector or ([other MCP clients](/agents/guides/remote-mcp-server/#connect-your-mcp-server-to-claude-and-other-mcp-clients)), and authenticate with GitHub. ## Next steps - Add [tools](/agents/model-context-protocol/tools/) to your MCP server. - Customize your MCP Server's [authentication and authorization](/agents/model-context-protocol/authorization/). --- # Test a Remote MCP Server URL: https://developers.cloudflare.com/agents/guides/test-remote-mcp-server/ import { Render } from "~/components"; Remote, authorized connections are an evolving part of the [Model Context Protocol (MCP) specification](https://spec.modelcontextprotocol.io/specification/draft/basic/authorization/). Not all MCP clients support remote connections yet. This guide will show you options for how to start using your remote MCP server with MCP clients that support remote connections. If you haven't yet created and deployed a remote MCP server, you should follow the [Build a Remote MCP Server](/agents/guides/remote-mcp-server/) guide first. ## The Model Context Protocol (MCP) inspector The [`@modelcontextprotocol/inspector` package](https://github.com/modelcontextprotocol/inspector) is a visual testing tool for MCP servers. You can run it locally by running the following command: ```bash npx @modelcontextprotocol/inspector ``` Then, enter the URL of your remote MCP server. You can use an MCP server running on your local machine on localhost, or you can use a remote MCP server running on Cloudflare.  Once you have authenticated, you will be redirected back to the inspector. You should see the "List Tools" button, which will list the tools that your MCP server exposes.  ## Connect your remote MCP server to Claude Desktop via a local proxy Even though [Claude Desktop](https://claude.ai/download) doesn't yet support remote MCP clients, you can use the [`mcp-remote` local proxy](https://www.npmjs.com/package/mcp-remote) to connect it to your remote MCP server. This lets you to test what an interaction with your remote MCP server will be like with a real-world MCP client. 1. Open Claude Desktop and navigate to Settings -> Developer -> Edit Config. This opens the configuration file that controls which MCP servers Claude can access. 2. Replace the content with a configuration like this: ```json { "mcpServers": { "math": { "command": "npx", "args": ["mcp-remote", "http://my-mcp-server.my-account.workers.dev/sse"] } } } ``` This tells Claude to communicate with your MCP server running at `http://localhost:8787/sse`. 3. Save the file and restart Claude Desktop (command/ctrl + R). When Claude restarts, a browser window will open showing your OAuth login page. Complete the authorization flow to grant Claude access to your MCP server. Once authenticated, you'll be able to see your tools by clicking the tools icon in the bottom right corner of Claude's interface. ## Connect your remote MCP server to Cursor To connect [Cursor](https://www.cursor.com/) with your remote MCP server, choose `Type`: "Command" and in the `Command` field, combine the command and args fields into one (e.g.`npx mcp-remote https://your-worker-name.your-account.workers.dev/sse`). ## Connect your remote MCP server to Windsurf You can connect your remote MCP server to [Windsurf](https://codeium.com/windsurf) by editing the [`mcp_config.json` file](https://docs.codeium.com/windsurf/mcp), and adding the following configuration: ```json { "mcpServers": { "math": { "command": "npx", "args": ["mcp-remote", "http://my-mcp-server.my-account.workers.dev/sse"] } } } ``` --- # Authorization URL: https://developers.cloudflare.com/agents/model-context-protocol/authorization/ import { DirectoryListing } from "~/components"; When building a [Model Context Protocol (MCP)](https://modelcontextprotocol.io) server, you need both a way to allow users to login (authentication) and allow them to grant the MCP client access to resources on their account (authorization). <diagram> </diagram> The Model Context Protocol uses [a subset of OAuth 2.1 for authorization](https://spec.modelcontextprotocol.io/specification/draft/basic/authorization/). OAuth allows your users to grant limited access to resources, without them having to share API keys or other credentials. Cloudflare provides an [OAuth Provider Library](https://github.com/cloudflare/workers-oauth-provider) that implements the provider side of the OAuth 2.1 protocol, allowing you to easily add authorization to your MCP server. You can use the OAuth Provider Library in three ways: 1. **Your Worker handles authorization itself.** Your MCP server, running on Cloudflare, handles the complete OAuth flow. ([Example](/agents/guides/remote-mcp-server/)) 2. **Integrate directly with a third-party OAuth provider**, such as GitHub or Google. 3. **Integrate with your own OAuth provider**, including authorization-as-a-service providers you might already rely on, such as Stytch and Auth0. The following sections describe each of these options and link to runnable code examples for each. ## Authorization options ### (1) Your MCP Server handles authorization and authentication itself Your MCP Server, using the [OAuth Provider Library](https://github.com/cloudflare/workers-oauth-provider), can handle the complete OAuth authorization flow, without any third-party involvement. The [Workers OAuth Provider Library](https://github.com/cloudflare/workers-oauth-provider) is a Cloudflare Worker that implements a [`fetch()` handler](/workers/runtime-apis/handlers/fetch/), and handles incoming requests to your MCP server. You provide your own handlers for your MCP Server's API, and autentication and authorization logic, and URI paths for the OAuth endpoints, as shown below: ```ts export default new OAuthProvider({ apiRoute: "/mcp", // Your MCP server: apiHandler: MyMCPServer.Router, // Your handler for authentication and authorization: defaultHandler: MyAuthHandler, authorizeEndpoint: "/authorize", tokenEndpoint: "/token", clientRegistrationEndpoint: "/register", }); ``` Refer to the [getting started example](/agents/guides/remote-mcp-server/) for a complete example of the `OAuthProvider` in use, with a mock authentication flow. The authorization flow in this case works like this: ```mermaid sequenceDiagram participant B as User-Agent (Browser) participant C as MCP Client participant M as MCP Server (your Worker) C->>M: MCP Request M->>C: HTTP 401 Unauthorized Note over C: Generate code_verifier and code_challenge C->>B: Open browser with authorization URL + code_challenge B->>M: GET /authorize Note over M: User logs in and authorizes M->>B: Redirect to callback URL with auth code B->>C: Callback with authorization code C->>M: Token Request with code + code_verifier M->>C: Access Token (+ Refresh Token) C->>M: MCP Request with Access Token Note over C,M: Begin standard MCP message exchange ``` Remember — [authentication is different from authorization](https://www.cloudflare.com/learning/access-management/authn-vs-authz/). Your MCP Server can handle authorization itself, while still relying on an external authentication service to first authenticate users. The [example](/agents/guides/remote-mcp-server) in getting started provides a mock authentdcation flow. You will need to implement your own authentication handler — either handling authentication yourself, or using an external authentication services. ### (2) Third-party OAuth Provider The [OAuth Provider Library](https://github.com/cloudflare/workers-oauth-provider) can be configured to use a third-party OAuth provider, such as GitHub or Google. You can see a complete example of this in the [GitHub example](/agents/guides/remote-mcp-server/#add-authentication). When you use a third-party OAuth provider, you must provide a handler to the `OAuthProvider` that implements the OAuth flow for the third-party provider. ```ts ins="defaultHandler: MyAuthHandler," import MyAuthHandler from "./auth-handler"; export default new OAuthProvider({ apiRoute: "/mcp", // Your MCP server: apiHandler: MyMCPServer.Router, // Replace this handler with your own handler for authentication and authorization with the third-party provider: defaultHandler: MyAuthHandler, authorizeEndpoint: "/authorize", tokenEndpoint: "/token", clientRegistrationEndpoint: "/register", }); ``` Note that as [defined in the Model Context Protocol specification](https://spec.modelcontextprotocol.io/specification/draft/basic/authorization/#292-flow-description) when you use a third-party OAuth provider, the MCP Server (your Worker) generates and issues its own token to the MCP client: ```mermaid sequenceDiagram participant B as User-Agent (Browser) participant C as MCP Client participant M as MCP Server (your Worker) participant T as Third-Party Auth Server C->>M: Initial OAuth Request M->>B: Redirect to Third-Party /authorize B->>T: Authorization Request Note over T: User authorizes T->>B: Redirect to MCP Server callback B->>M: Authorization code M->>T: Exchange code for token T->>M: Third-party access token Note over M: Generate bound MCP token M->>B: Redirect to MCP Client callback B->>C: MCP authorization code C->>M: Exchange code for token M->>C: MCP access token ``` Read the docs for the [Workers oAuth Provider Library](https://github.com/cloudflare/workers-oauth-provider) for more details. ### (3) Bring your own OAuth Provider If your application already implements an Oauth Provider itself, or you use Stytch, Auth0, or authorization-as-a-service provider, you can use this in the same way that you would use a third-party OAuth provider, described above in (2). ## Next steps - [Learn how to use the Workers OAuth Provider Library](https://github.com/cloudflare/workers-oauth-provider) - Learn how to use a third-party OAuth provider, using the [GitHub](/agents/guides/remote-mcp-server/#add-authentication) example MCP server. --- # Model Context Protocol (MCP) URL: https://developers.cloudflare.com/agents/model-context-protocol/ You can build and deploy [Model Context Protocol (MCP)](https://modelcontextprotocol.io/) servers on Cloudflare. ## What is the Model Context Protocol (MCP)? [Model Context Protocol (MCP)](https://modelcontextprotocol.io) is an open standard that connects AI systems with external applications. Think of MCP like a USB-C port for AI applications. Just as USB-C provides a standardized way to connect your devices to various accessories, MCP provides a standardized way to connect AI agents to different services. ### MCP Terminology - **MCP Hosts**: AI assistants (like [Claude](http://claude.ai) or [Cursor](http://cursor.com)), AI agents, or applications that need to access external capabilities. - **MCP Clients**: Clients embedded within the MCP hosts that connect to MCP servers and invoke tools. Each MCP client instance has a single connection to an MCP server. - **MCP Servers**: Applications that expose [tools](/agents/model-context-protocol/tools/), [prompts](https://modelcontextprotocol.io/docs/concepts/prompts), and [resources](https://modelcontextprotocol.io/docs/concepts/resources) that MCP clients can use. ### Remote vs. local MCP connections The MCP standard supports two modes of operation: - **Remote MCP connections**: MCP clients connect to MCP servers over the Internet, establishing a [long-lived connection using HTTP and Server-Sent Events (SSE)](/agents/model-context-protocol/transport/), and authorizing the MCP client access to resources on the user's account using [OAuth](/agents/model-context-protocol/authorization/). - **Local MCP connections**: MCP clients connect to MCP servers on the same machine, using [stdio](https://spec.modelcontextprotocol.io/specification/draft/basic/transports/#stdio) as a local transport method. ### Get Started Go to the [Getting Started](/agents/guides/remote-mcp-server/) guide to learn how to build and deploy your first remote MCP server to Cloudflare. --- # Tools URL: https://developers.cloudflare.com/agents/model-context-protocol/tools/ import { Render, TypeScriptExample } from "~/components"; Model Context Protocol (MCP) tools are functions that a [MCP Server](/agents/model-context-protocol) provides and MCP clients can call. When you build MCP Servers with the `@cloudflare/model-context-protocol` package, you can define tools the [same way as shown in the `@modelcontextprotocol/typescript-sdk` package's examples](https://github.com/modelcontextprotocol/typescript-sdk?tab=readme-ov-file#tools). For example, the following code from [this example MCP server](https://github.com/cloudflare/ai/tree/main/demos/remote-mcp-server) defines a simple MCP server that adds two numbers together: <TypeScriptExample> ```ts title="src/index.ts" import { McpServer } from "@modelcontextprotocol/sdk/server/mcp"; import { DurableMCP } from "@cloudflare/model-context-protocol"; export class MyMCP extends DurableMCP { server = new McpServer({ name: "Demo", version: "1.0.0" }); async init() { this.server.tool( "add", { a: z.number(), b: z.number() }, async ({ a, b }) => ({ content: [{ type: "text", text: String(a + b) }], }), ); } } ``` </TypeScriptExample> --- # Transport URL: https://developers.cloudflare.com/agents/model-context-protocol/transport/ import { Render } from "~/components"; The Model Context Protocol (MCP) specification defines [two standard transport mechanisms](https://spec.modelcontextprotocol.io/specification/draft/basic/transports/): 1. **stdio, communication over standard in and standard out** — designed for local MCP connections 2. **HTTP with Server-Sent Events (SSE)** — designed for remote MCP connections MCP Servers deployed to Cloudflare support remote MCP connections, using HTTP with Server-Sent Events (SSE) as transport. SSE requires a persistent HTTP connection, which is supported by Cloudflare [Durable Objects](/durable-objects/). Transport is configured and handled automatically. You don't need to configure anything — it just works. :::note Even if the MCP client you are using only supports local MCP connections, you can still connect it to a remote MCP server. Follow [this guide](/agents/guides/test-remote-mcp-server/) for instructions on how to connect to your remote MCP server from Claude Desktop, Cursor, Windsurf, and other local MCP clients, using the [`mcp-remote` local proxy](https://www.npmjs.com/package/mcp-remote). ::: ---