Skip to content

How AI Works with Crow

Crow is an AI-powered platform, but it is not tied to any single AI provider. Multiple AI systems can connect to the same Crow instance simultaneously, sharing a single database of memories, projects, blog posts, files, and contacts. This page explains the three connection patterns and how they fit together.

Three Ways to Connect

1. External MCP (Direct Connection)

AI platforms that support the Model Context Protocol connect directly to Crow's MCP servers. This is the primary pattern for Claude.ai, ChatGPT, Gemini, Cursor, Windsurf, and other desktop/web AI tools.

AI Platform ──► MCP Transport ──► Crow MCP Servers ──► SQLite Database
  (Claude,        (stdio or           (memory,
   ChatGPT,        Streamable HTTP)    projects,
   Cursor)                             sharing,
                                       storage,
                                       blog)

How it works: The AI platform spawns Crow's MCP servers as child processes (stdio) or connects to the gateway over HTTP. The platform's built-in AI calls Crow tools directly — store memories, search projects, publish blog posts, etc.

Best for: Deep work sessions, platform-specific features (Claude's artifacts, Cursor's code editing), using the AI you already pay for.

Setup: Add MCP server entries to the platform's configuration file. See Platforms for per-platform guides.

2. BYOAI Chat (Crow's Nest)

Crow's built-in AI Chat lets you use any AI provider through the Crow's Nest web dashboard. Crow handles the AI provider connection and tool dispatch internally.

User ──► Crow's Nest ──► /api/chat ──► AI Provider API ──► Tool Executor ──► MCP Servers ──► Database
           (browser)      (gateway)    (OpenAI, Anthropic,   (in-process
                                        Google, Ollama,       MCP Client)
                                        OpenRouter)

How it works: You configure an AI provider (API key + model) in Settings. When you send a message, the gateway forwards it to the provider. When the AI responds with tool calls, the gateway's tool executor dispatches them to Crow's MCP servers in-process and feeds the results back to the AI. The response streams to your browser via Server-Sent Events.

Best for: Quick interactions from the dashboard, using free/cheap AI providers (Ollama for fully local, OpenRouter for budget models), accessing Crow from devices without a native AI client.

Setup: See the AI Providers (BYOAI) guide.

3. Native Bots (CrowClaw)

AI bots that live on Discord, WhatsApp, Telegram, and other chat platforms — managed directly from the Crow dashboard. The CrowClaw extension handles bot lifecycle, AI configuration, skill deployment, and monitoring.

Chat Platform ──► OpenClaw Engine ──► Crow MCP Servers ──► SQLite Database
  (Discord,         (managed by            ▲
   WhatsApp,         CrowClaw)             │
   Telegram)                          CrowClaw Panel
                                     (dashboard UI)

How it works: Install the CrowClaw extension, create a bot from the dashboard, and deploy it. CrowClaw auto-configures the bot's AI provider from Crow's existing AI profiles (BYOAI bridge), deploys skills, and manages the bot's systemd service. The bot appears in the Messages panel alongside peers and AI chat — one inbox for everything.

Bots aren't just chat interfaces — they can control Crow apps. Since the bot connects via MCP, it has access to the same tool suite as Claude or ChatGPT: memory, projects, blog, sharing, storage, and extensions. A household bot can track expenses and manage a pantry. A research bot can monitor RSS feeds and store findings in projects. A blog bot can publish scheduled posts.

Best for: Accessing Crow from mobile chat apps, multi-platform presence, automated workflows, collaborative use through shared channels.

Setup: See the Bot Management guide or install CrowClaw from Extensions.

What They All Share

All three connection patterns access the same database, same tools, and same data. A memory stored from Claude.ai is instantly searchable from the Crow's Nest AI Chat and from a bot on Discord. A blog post drafted in Cursor can be published from ChatGPT.

ResourceShared across all connections
MemoriesFull-text searchable, tagged, scored by importance
Projects & sourcesResearch, data connectors, notes, citations
Blog postsDrafts, published posts, themes, RSS feeds
FilesS3-compatible storage (when MinIO is configured)
Contacts & messagesPeer identities, Nostr DMs, shared items
Behavioral contextcrow.md identity, protocols, and custom sections

This is the core value proposition: use whichever AI interface fits the moment, and your data follows you.

Tool Routing

Crow exposes 49+ individual MCP tools across its five servers. Loading all of them into an AI's context window is wasteful — most interactions only need a few tools.

The Router Pattern

The gateway's /router/mcp endpoint consolidates all tools into 7 category tools, reducing context usage by approximately 75%:

Category ToolRoutes ToActions
crow_memoryMemory serverstore, search, recall, list, update, delete, stats, context
crow_projectsProject servercreate, list, update, add sources/notes, search, bibliography
crow_blogBlog servercreate, edit, publish, unpublish, themes, export, share
crow_sharingSharing serverinvite, accept, share, inbox, send message, contacts, revoke
crow_storageStorage serverupload, list, get URL, delete, stats
crow_toolsExternal integrationsGitHub, Brave Search, Slack, and other connected services
crow_discoverSchema lookupReturns full parameter schemas on demand

Each category tool accepts an action parameter (e.g., crow_memory with action: "store_memory") and a params object. The crow_discover tool lets the AI inspect available actions and their full schemas without loading everything upfront.

Where Routing Applies

  • Gateway HTTP (/router/mcp): Used by remote MCP clients and the BYOAI tool executor
  • crow-core (servers/core/): Stdio equivalent for local deployments — starts with memory tools, activates other servers on demand
  • Direct endpoints (/memory/mcp, /projects/mcp, etc.): Still available for clients that prefer full tool definitions

Messaging and Sharing

Crow includes a peer-to-peer layer for communication and data sharing between users, independent of any AI provider.

Nostr (Messaging)

Direct messages between Crow users are encrypted with NIP-44 and relayed through Nostr relays. Messages appear in the Crow's Nest Messages panel and are accessible from any connected AI via the crow_send_message and crow_inbox tools.

Hypercore (Data Sync)

Shared items (memories, projects, sources) replicate between peers over Hypercore append-only feeds. Peers discover each other via Hyperswarm DHT with NAT holepunching — no central server required.

Peer Relay (Offline Delivery)

When a peer is offline, messages and shares can be held by an opt-in peer relay for later delivery. The relay stores encrypted payloads and forwards them when the recipient comes online.

Choosing a Pattern

ScenarioRecommended Pattern
Deep research session with ClaudeExternal MCP (stdio)
Quick memory lookup from your phoneBYOAI Chat (Crow's Nest)
Team collaboration on DiscordNative Bots (CrowClaw)
Code project with AI assistanceExternal MCP via Cursor or Claude Code
Fully local, no cloud AIBYOAI Chat with Ollama
Access from multiple chat appsNative Bots (Discord + WhatsApp + Telegram)
Automated household managementNative Bots with skills
All of the above, simultaneouslyAll three — they share one database

Next Steps

Released under the MIT License.