AI Agent SDK

Build agents. Skip the infrastructure.

Hosted runtime, 150+ integrations, scheduling, memory, and email inbox. Ship autonomous agents to production in minutes.

Stop building agent infrastructure

Everything an agent needs to run in production. None of the setup.

Hosted agent runtime

Agents run in isolated sandboxes. You ship the workflow, not the infra.

Human-in-the-loop

Require approval before sensitive actions. Keep the speed without giving up control.

150+ managed integrations

Gmail, Slack, Notion, HubSpot, Google Ads, Linear. OAuth and token refresh handled.

Scheduled runs

Set the cadence once. Daily, weekly, or hourly runs happen automatically.

Email inbox

Forward an email, get work done automatically. Every agent gets their own inbox.

Agent memory

Agents remember past conversations and build on them. Per-user scoping for multi-tenant apps.

File handling

Agents generate reports and spreadsheets, downloadable through the API.

From zero to deployed agent in minutes

Create a teammate, schedule it, give it an inbox. Everything runs in a hosted sandbox.

pip install m8tes

From weeks of engineering to minutes

Other SDKs give you building blocks. We give you deployed agents.

With OpenAI Agents SDK, Claude Code SDK, etc.

  • Build and maintain sandboxed execution
  • Implement OAuth for every app you connect
  • Write your own scheduling and trigger system
  • Build approval flows so agents don't go rogue
  • Design memory that persists across executions
  • Set up real-time streaming for your UI
  • Handle file output and delivery
  • Build webhook infrastructure for agent events
  • Isolate user data for multi-tenant apps
  • Give agents their own email inbox
  • Keep up with fast-moving AI models and APIs

Weeks of engineering before your first agent runs.

With the m8tes SDK

  • Hosted sandbox, no servers to manage
  • 150+ integrations with managed OAuth
  • Scheduling, webhook triggers, and email inbox built in
  • Human-in-the-loop so you stay in control
  • Persistent memory across executions
  • Real-time streaming out of the box
  • Generated files, downloadable via the API
  • Outbound webhooks for agent events
  • Per-user isolation for multi-tenant apps
  • Every agent gets its own @m8tes.ai inbox
  • Always up to date as AI evolves. Zero maintenance.

Live in minutes. Not weeks.

Integrate agents into any workflow

Run agents on demand, on a schedule, from an email, or from any external event.

Email inbox

Email your agent. It reads it and gets to work.

Schedule

Daily, weekly, or hourly. Set a cadence and your agent runs automatically.

Webhook

POST when a Stripe charge fails, a GitHub issue opens, or a form submits. Any event, any system.

On demand

Trigger runs from your code or the API.

Works with

GmailGmail
SlackSlack
NotionNotion
LinearLinear
HubSpotHubSpot
StripeStripe
Google SheetsGoogle Sheets
Google AdsGoogle Ads

and 150+ other apps

Not a framework. Not a library.

m8tes vs. LangChain, CrewAI, and other SDKs

LangChain, CrewAI, OpenAI Agents SDK, etc.

  • Orchestrate LLM calls locally. No runtime included.
  • You build and host the execution environment
  • You implement OAuth for every integration
  • You write scheduling, triggers, and memory from scratch
  • You add human-in-the-loop and approval flows yourself

m8tes

  • Hosted runtime. No infra to manage.
  • 150+ integrations with managed OAuth
  • Scheduling, webhooks, and email triggers built in
  • Persistent memory out of the box
  • Human-in-the-loop built in. Agents ask before acting.

Build for your users. Automate for your team.

Build AI features for your users

Customer-facing assistants

Give each user their own agent with isolated memory, tools, and permissions. Multi-tenant without custom plumbing.

Automated onboarding

New signup triggers a run that provisions accounts, sends welcome emails, and schedules a check-in.

In-app research agent

Users describe what they need. The agent searches, synthesizes, and returns structured results in your UI.

Automate internal workflows

Revenue reporting

Pull MRR from Stripe, update the sheet, and post the summary to Slack. No more manual Monday reporting.

Support triage

Classify inbound tickets, draft replies, escalate blockers to the right Slack channel. Runs 24/7 on a schedule.

Ad spend monitoring

Check Google Ads weekly. Pause ads burning budget on low-converting keywords. Alert the team.

Go deeper on AI agent development

Architecture, patterns, and production setup.

How the AI agent SDK works

Teammate, task, run

The SDK is built around three concepts. A teammate is a reusable agent with a name, instructions, and connected apps. A task defines what the agent should do. A run is a single execution of that task.

You configure a teammate once. Then trigger runs on demand, on a schedule, or from a webhook. Each run streams events in real time. Show progress in your UI or pipe results downstream.

Hosted sandboxed execution

Every run executes in an isolated sandbox with its own filesystem and network. The runtime handles tool calls, OAuth token refresh, rate limiting, and retries. You never touch infrastructure.

Streaming and real-time events

The SDK uses server-sent events (SSE) to stream run progress. You get events for tool calls, text output, file creation, and completion. Build real-time UIs or process events in a backend pipeline.

Building multi-tenant AI agents

Set user_id on any run to scope it to a specific end user. Memory, task history, and tool access are strictly isolated. One user never sees another user's data.

User isolation with user_id

Set user_id on any run to scope it to a specific end user. Memory, task history, and tool access are strictly isolated. One user never sees another user's data.

Per-user memory and tool scoping

Each end user gets their own memory context. The agent remembers past interactions with that user and builds on them. Scope tool permissions per user for fine-grained access control.

Embedding AI agents in your product

Every user gets their own agent experience inside your app. Isolated runs, memory, and tool access. You control the instructions and capabilities. Your users just see results.

Production patterns for AI agent development

Three modes. Autonomous runs without asking. Approval asks before each sensitive action. Plan proposes first; you approve, then it executes. Start with approval. Move to autonomous when you're ready.

Permission modes for safe execution

Three modes. Autonomous runs without asking. Approval asks before each sensitive action. Plan proposes first; you approve, then it executes. Start with approval. Move to autonomous when you're ready.

You can change the mode after a run starts. Use client.runs.update_permission_mode(run_id, permission_mode="approval") when you need to tighten control mid-run or switch back to autonomous once the risky step is done.

Scheduling, webhooks, and email triggers

Cron schedules handle recurring work. Webhooks let you trigger runs from any external event (Stripe charge failed, GitHub issue opened, form submitted). Email triggers let users forward messages directly to an agent.

Monitoring and output validation

Every run produces structured logs showing each tool call, its inputs, and its outputs. Use them to debug, audit, and improve. Output files (CSV, PDF, reports) are downloadable through the API.

Track account-level consumption too. client.auth.get_usage() gives you runs, spend, limits, and billing period boundaries so you can surface usage in your own dashboards or alerts.

FAQ

Common questions about building AI agents with the SDK.

Explore more

Build your first AI agent today

Free to start. No infrastructure to manage. Deploy in minutes.