Select the LLM that powers Agent. You can use a local Ollama model or a cloud provider.
Loading available models…
Optional — you can always configure these later in Settings.
Your agent is configured and ready to go.
LLM Models
When enabled, the model's <think> reasoning is shown in a collapsible block during streaming. Default: hidden.
Recommended Model Profile
Main glm-5.1:cloudFast gpt-oss:20b-cloudFallback qwen3.5:9b
Store your Ollama Cloud API key under API Keys if you use cloud variants. You can still sign in locally to pull cloud models through the Ollama runtime.
Download Ollama Models
Pull models from the Ollama library. Requires Ollama running locally. Cloud variants are available after signing in at ollama.com.
Installed Models
API Keys
Provider keys unlock additional model families. Keys are stored locally and never sent to third parties.
llama-server starts automatically when you use a llama.cpp model. Stop to free memory.
Co-Pilot
You can configure each Co-Pilot before turning Co-Pilot on. Initiative stays shared; schedule and instructions are now set per Co-Pilot.
Checks today's agenda, recent email, and drafts, then proposes priorities and next actions.
Creates a recurring Co-Pilot Morning Brief schedule.
Co-Pilot is off. Turn it on to schedule a daily Morning Brief.
Checks open loops across recent email and upcoming meetings, then shows what still likely needs a reply, decision, or preparation.
Creates a recurring midday or end-of-day follow-up sweep.
Open loops, pending replies, upcoming prep, and stalled decisions.
Co-Pilot is off. Turn it on to schedule a daily Follow-up sweep.
Context Sources
Checking Outlook availability…
Checking Google authorization…
Durable wiki memory for important threads and ongoing context.
No relevant memory stored yet.
Outlook
Persistent folder context is managed centrally in Settings → Context.
Google Workspace
Provide OAuth 2.0 credentials from Google Cloud Console to enable Gmail, Calendar and Drive skills.
Persistent label context is managed centrally in Settings → Context.
Brave Search
Microsoft OneDrive
No OAuth or Azure setup required — reads directly from your synced OneDrive folder.
Register an app in Azure Portal → App Registrations. Add Files.ReadWrite.All and User.Read delegated permissions. Set the redirect URI below as a Web platform redirect.
Browser Control
Browser navigation stays inside the app preview. External Chromium windows are disabled.
MCP Servers
Connect external tool servers via Model Context Protocol. Add from presets or manually.
Webhooks
Define data endpoints the agent can call to fetch information. Each webhook becomes available as a named data source the agent can query during runs.
Open WhatsApp → Linked Devices → Link a Device → Scan this QR
📱 Linked: —
Send Test Message
Available only after WhatsApp is connected. The app sends a test message to your linked WhatsApp.
Uses WhatsApp Web (linked device). No Meta Business API needed. Setup and connection are handled automatically — no external dependencies required.
Slack
Create a Slack app at api.slack.com. Add scopes: chat:write, channels:read, app_mentions:read, im:history. Enable Socket Mode for the App Token (xapp-…).
Telegram
Create a bot via @BotFather on Telegram. Send /newbot and paste the token here. Messages sent to your bot will be executed as agent goals.
Run Options
Server Control
Use Start Server if the backend has dropped completely. Reset Server is for restarting a backend that is still partly reachable.
When enabled, closing the last browser tab will stop the server and free system resources. A confirmation dialog is shown first. Default: on.
System Check
Test all configured integrations at once.
Run Policy & Budgets
Updates
Automatic version check and one-click update install. Existing user settings are preserved.
Built-in Skills
Custom Skills (created by agent · persisted in tools/ folder)
No custom skills yet. The agent can create skills with create_skill.
MCP Skills (discovered from configured MCP servers)
No MCP skills discovered yet. Add or refresh MCP servers in Settings.
System Log
Self-improvement events from the last 30 days — corrections, reflections, memory maintenance, and server starts.
Automatic Learnings
A stronger model reviews proposed learnings automatically and either promotes them into durable rules or discards them.
No reviewed learnings yet.
No system events recorded yet.
Overview
AgentOrchestrator is a local-first work assistant that combines chat, email, calendar, browser automation, file handling, cloud context, scheduled jobs, and multi-step execution in one interface.
You use the composer like a normal chat app, but the runtime behaves more like an operations layer than a plain chatbot. For every message it decides whether it should answer immediately, use a deterministic functional path, or launch a full multi-stage workflow.
In practice this means the same input box can handle very different kinds of work: answering a question, opening the right email, updating a calendar event, summarizing an attachment, searching connected context sources, or running a multi-step agent plan with visible progress.
What the product is optimized for
open the first one or move it to Friday can reuse the right records from the prior step.Current Setup Flow
The recommended model profile is glm-5.1:cloud for main work, gpt-oss:20b-cloud for fast work, and qwen3.5:9b as fallback. If you use Ollama Cloud models, the local Ollama runtime must already be signed in with ollama signin.
Interface Layout
Conversations
Every fresh chat starts from the How can I help? welcome state. A newly opened browser tab or window also starts clean, so the product does not silently attach you to an older conversation unless you explicitly continue one from History.
Composer
Live Workflows
Agent Pipeline
Complex work uses a staged pipeline so planning, information gathering, execution, and quality checks do not collapse into a single opaque model response.
The product does not always use every stage. A simple request can bypass most of this. A harder task can loop through planning, execution, and validation more than once before it returns a final answer.
Fast Paths
Safety
Co-Pilot
Co-Pilot is the proactive briefing layer. It reads your connected inboxes and calendars and turns them into structured summaries.
copilot: morning brief or another configured brief name.Skills
Skills are the tools the agent can call. The app separates them into Built-in Skills and Custom Skills.
tools/ folder and can be turned on or off in Settings.Custom Skill Basics
run(args) entry point.tools/_std.py for consistent success and error output.Settings
Use the gear icon to open Settings. The most important sections are Models, Integrations, Channels, Server, and Updates.
Models
Integrations
Channels
Channel commands include status, pause, resume, stop, schedule: <interval> <goal>, schedules, unschedule: <id>, copilot: <name>, and help. In practice, normal natural-language requests still work best, and commands are mostly useful for messaging channels where you want quick operational control.
Server and Updates
Shortcuts and UI Tips
| Enter | Send message |
| Shift + Enter | Insert a new line |
| Escape | Close the current overlay |
Technical: Memory and Context
The system uses several context, memory, routing, execution, and persistence layers at once. They are not interchangeable. Some exist to answer the current request, some exist to preserve continuity, and some exist only so the runtime can inspect itself and improve future runs.
Reading guide
Architecture at a glance
Composer/UI -> request packaging + attachments + session binding -> backend routing + task understanding + context assembly -> direct reply OR deterministic fast path OR full orchestrator -> validation + fallback/replan/clarification when needed -> final report + session history + memory/log side effects Cross-cutting layers: settings/keyring | managed integrations | tool registry | memory/wiki/user profile | websocket events
End-to-end request lifecycle
What happens before the model is even called
settings.json, resolves secret placeholders from keyring where needed, and determines which providers, channels, policies, and models are currently active.Context Layers
| Current request | Your latest message, live toggles, selected models, browser state, and anything attached to the current send action. |
| Current session | The active conversation history, recent tool results, run state, and short-range continuity used for follow-up questions. |
| Attached context | Files, folders, Google Drive items, and OneDrive items attached to the active request. |
| Managed context sources | Mailbox folders, labels, local folders, and cloud folders configured in Settings > Context so the agent can search them when relevant. |
| Relevant Memory | Durable user-facing wiki memory such as ongoing topics, important reference threads, and stable preferences that matter in future dialogue. |
| User Identity | The structured profile in user.md. This stores durable personal facts and communication preferences. |
| Semantic history recall | Search across earlier sessions when older work is relevant but not already present in the active session. |
| System Log | Operational traces for learning, reflection, corrections, memory maintenance, and server events. This is diagnostic, not conversational memory. |
How routing chooses the execution path
open the second one, summarize the PDF, or move that meeting.The practical goal of this routing layer is to keep simple work fast and stable while still allowing complex tasks to expand into a real agent run when necessary.
What the memory layers are for
Functional section map
How the orchestrator actually runs
The orchestrator is policy-bound. It tracks maximum steps, maximum tool calls, validation retries, token budgets, runtime budgets, approval rules, disabled tools, and dynamic policy adjustments based on earlier outcomes. This means the runtime is not only trying to solve the task, but also trying to do so within operational constraints.
Tooling model under the hood
tools/ folder and are loaded dynamically so you can add project-specific functionality without rewriting the core app.How email and calendar stay reliable
the first one, that draft, or the meeting tomorrow without re-inventing selection logic each time.How attachments and external context are processed
How memory, learning, and reflection differ
Learning and System Log
What the frontend is really doing during a run
A concrete mental model
Think of the app as five cooperating systems running behind one chat box: a UI session manager, a routing layer, a set of deterministic business runtimes, an agent orchestrator, and a memory and learning layer. The product feels simple from the front, but under the hood it keeps deciding which of those systems should take the lead for the current message.
That is the core design principle of AgentOrchestrator: do not send everything to one giant model prompt. Instead, preserve structured state, choose the narrowest correct path, expose progress, and only expand into a full autonomous workflow when the task truly requires it.
AgentOrchestrator can perform autonomous actions on your system, including reading and writing files, executing code, sending emails, and fetching external data via the internet.
By using this application you acknowledge the following:
Ensure you have the appropriate permissions for all actions performed by the agent.
Your AI-powered autonomous assistant. Let's set up the essentials — it only takes a minute.
The recommended setup uses Ollama plus optional cloud providers. You can change everything later in Settings.
Choose where AgentOrchestrator stores its data — output files, memory, skills, and settings.
All agent data will be stored in this folder.
These folders will be created automatically:
Ollama runs locally on your machine. The recommended profile uses GLM as the main cloud model, GPT-OSS as the fast cloud model, and local Qwen as fallback.
Local runtime with optional Ollama Cloud access for GPT-OSS cloud models.
Recommended profile: glm-5.1:cloud main · gpt-oss:20b-cloud fast · qwen3.5:9b fallback.
For Ollama Cloud models: sign in locally first with ollama signin. The Qwen fallback stays local.
ollama signin
GPT-4.1, o3, o4-mini and more cloud models.
Gemini 2.5 Pro, Flash and more from Google.
NVIDIA-hosted models via build.nvidia.com.
Agents can search the web. A built-in fallback works without a key, but Brave gives better results.
DuckDuckGo-based search. Works out of the box, no key needed.
Premium search results with Brave. Free tier available.
Connect messaging channels so you can control your agent from anywhere.
Control the agent from Slack via Socket Mode.
Create a Slack app at api.slack.com. Enable Socket Mode.
Send commands via WhatsApp. Turn it on here, then scan the QR code later in Settings.
No credentials required. After setup, go to Settings → Channels → WhatsApp to scan the QR code with your phone.
Control the agent from a Telegram bot chat.
Create a bot via @BotFather on Telegram and paste the token here.
Connect external services so agents can read email, manage calendar, and access files.
Gmail, Google Calendar & Drive access for agents.
Get credentials from Google Cloud Console. You can authorize after setup in Settings → Integrations.
Read & send email via Outlook Desktop. No key needed.
Here's a summary of your configuration:
You can change any of these later in Settings (gear icon).
The backend server is not running.
Checking launcher...