
Stop re-explaining yourself to your AI.
MythOS is the shared memory between you and every AI you use. Capture what you read. Connect what you think. Publish what matters — across every model, every agent, every channel.
Your library, one click from every AI.

You already have a memory. It's just scattered across Notion pages, voice memos, half-written drafts, and every AI chat you've ever had.
Write once. Remember forever.
Markdown editor with @mentions, #hashtags, collapsible headings, and images that paste in like email. Your memos link to each other automatically. Your library builds itself.

Build your library with your AI, not around it.
MythOS is designed to be co-authored. Ask Claude to research a topic — it writes the memo and files it for you. Use Cursor, Windsurf, or any MCP-native app, and your library follows. Every conversation ends with something permanent.
Connect via MCP →One library. Every workflow.
Drop a memo into any automation. Zapier reads from MythOS. Your n8n flow pulls your notes before firing. Your agent loads your library as context before replying. Your memory becomes the single source of truth the rest of your stack can actually reference.
See integrations →Talk to what you've written.
Every memo becomes searchable by meaning, not just keywords. Ask your library questions. Get answers grounded in your own words, with citations to the exact memo where you wrote them.
Try the chat →Your library, out loud.
Publish any memo with one toggle. Readers get your prose, your mentions, your tags, and a chat panel that speaks in your voice — grounded in your own memos. Every memo is a live page. Every library is a live conversation.
See a public memo →MythOS MCP
Objective
MythOS MCP — Model Context Protocol — connects your MythOS knowledge library to AI assistants like Claude, Cursor, and Windsurf.
It lets you search memos, read content, and sync changes — all from within your AI tools. This is the Memory layer of The Augmentation Stack.
One source. Three channels.
Turn any public memo into a discussion space — upvote, comment, fork. Send it as a newsletter issue to your subscribers. Get surfaced in search — Google and every AI indexer, thanks to JSON-LD and /llms.txt. Your audience doesn't come to the platform; the platform goes to them.
MythOS MCP — how I connect every AI to my library
A rundown of connecting Claude, Cursor, and Windsurf via MCP. Every conversation ends with a memo filed in my library. Feedback welcome.
Your memory shouldn't die when the AI does.
This week: connecting every AI you use to a single portable memory layer. One memo, many models. Plus: three experiments in agent-readable notes…
MythOS MCP — Connect every AI to your library
Model Context Protocol connects your MythOS knowledge library to AI assistants like Claude, Cursor, and Windsurf. Search memos, read content, sync changes…
Addressable by every AI you'll ever use.
MythOS publishes an llms.txt, serves an agent-scoped API, and speaks OpenAPI. Your Claude reads your library. Your ChatGPT reads your library. Your next model, trained next year, reads your library. Your memory doesn't die when the AI does.
Simple tiers. Yours forever.
Free to start. Upgrade when your library outgrows you. Export everything anytime.
- Everything in Scribe
- MCP server access
- Hash ingestion
- Newsletter publishing
- Unlimited collaborators
- API access (60 req/min)
- Everything in Scholar
- Custom landing page builder
- Priority support
- Full API access
- Custom community
Before you ask.
Is this related to Anthropic's Claude "MythOS"?—
No. We've been MythOS since 2017 — years before Anthropic existed. If anything, it's the other way around.
What if I want to leave?—
Export everything as markdown anytime, including media. Your library is yours; we're just hosting it.
Do I need to know markdown?—
No. The editor works like any rich-text tool — bold, lists, checkboxes, images. Markdown is what we save underneath so it stays portable forever.
What AIs does this work with?—
Claude, ChatGPT, Gemini, Grok, Perplexity — anything that can speak HTTP or MCP. We publish an llms.txt and an agent-scoped API. Your next model will work too.
Can I bring my own AI key?—
Yes. Free and Pro both support BYO keys for OpenAI or Anthropic. Pro also includes hosted AI if you'd rather not manage keys.
Is my library private by default?—
Yes. Every memo starts private. You choose what to make public, unlisted, or shared with specific contacts via permission tags.