The cloud,
rebuilt for AI agents
Deploy, run, and scale autonomous agents on infrastructure designed for the future, not the past.
curl -sSL https://agentuity.sh/install.sh | bash
Build With Your Stack
Integrate with leading LLMs, popular frameworks including LangChain and CrewAI, and your preferred programming languages. Infrastructure that adapts to your tech stack, not the other way around.
import type { AgentContext, AgentRequest, AgentResponse } from "@agentuity/sdk"; import { generateObject } from "ai"; import { anthropic } from "@ai-sdk/anthropic"; export default async function Agent( req: AgentRequest, resp: AgentResponse, ctx: AgentContext, ) { const userQuery = req.data.text ?? "Recommend restaurants in Miami"; // Parse user intent with Vercel AI SDK const userIntent = await generateObject({ model: anthropic("claude-3-7-sonnet-20250219"), system: "You serve as a central hub that routes user requests to the right available AI agents. " + "Your task is to determine the user's intent to handle requests in a structured way.", schema: UserIntentSchema, prompt: userQuery, }); // Route to appropriate agent based on intent if (userIntent.intent === "food") { const agent = await ctx.getAgent({ name: "food-recommendations" }); const result = await agent.run({ data: userQuery, contentType: "text/plain", }); // Store interaction history await ctx.kv.set("user-history", req.sessionId, { query: userQuery }); return resp.text(result.text); } return resp.text("I can help you with that!"); }
Seamless Agent Collaboration
Break down framework barriers. Agent-native infrastructure enables seamless communication between agents built with different tools—whether they're using LangChain, CrewAI, or custom implementations. Let your agents collaborate naturally across any platform.
resp.handoff({ name: "my-other-agent" }, { data: "Summarize this text"… })
resp.handoff( { name: "my-other-agent" }, { data: "Summarize this text"… } )




Fastest Path to Production
Skip the complex setup of traditional cloud providers. Deploy your AI agents with a single command—no IAM, security groups, or load balancers to configure; agent-native infrastructure handles everything else.
Instructions In, Magic Out
Connect your agents to any channel—chat, webhooks, email, SMS, voice, or custom interfaces—using an intelligent routing system that ensures your data flows seamlessly to the right places at the right time.
$✨ Built in 170ms
🚀 Local Agent InteractionTo interact with your agents locally, you can:
curl -v http://localhost:3500/run/agent_123bcb427f491
--json {"input": "Hello, world!"}
Or use the 💻 Dev Mode in our app
https://app.agentuity.com/devmode/584868a149dc2514[INFO] Server started on port 3500
First-class Developer Experience
Build, test, and deploy AI agents with confidence. Agentuity's comprehensive toolkit includes a CLI, web interface, pre-built agent library, and local dev mode for rapid iteration.
Unified AI Gateway
Access any LLM provider through a gateway. Stop managing the complexity of multiple model providers and their API keys and tokens—just use the gateway to connect to OpenAI, Anthropic, Gemini, or any other provider. Pay only for what you use.
Everything You Need to Deploy, Run, and Scale
GitHub Pipeline Automations
Streamline deployments with GitHub Actions integration for consistent, automated deployments.
Real-Time Monitoring and Analytics
Track performance, memory usage, and costs with detailed metrics to optimize your AI agents.
Effortless Scaling and Automation
Scale your AI agents seamlessly with modern, agent-native enterprise infrastructure.
From Zero to Production
in 3 Minutes
Watch this quick tutorial to see how easy it is to get started. We'll show you how to download the CLI, create your first agent, test it locally, and deploy to production—all in just 3 minutes.

Ready to get started?
Deploy your first agent in minutes