๐ŸŽ‰ Launch Special: First 1,000 users get 3 months FREE โ€” no credit card required. Claim your spot โ†’
Now in Public Beta โ€” First 1,000 Users: 3 Months Free

Others store everything.
Tokyo Brain keeps what truly matters.

The first AI memory that dreams, forgets, and self-corrects. Your agent gets overnight reflection, active forgetting, and value-aligned recall โ€” not just storage.

129K+
Memories in production
5
Autonomous night systems
83.8%
LongMemEval (500 questions)
$0
Free tier forever

See Tokyo Brain in Action

The Anatomy of a Synthetic Thought

AI Memory (Chinese)

The Transparent Self-Audit

Every AI agent has amnesia

๐Ÿง 

Coding assistants forget architecture

Yesterday you decided on a microservices pattern. Today your AI suggests a monolith. Every session starts from zero.

๐Ÿ”„

Chatbots repeat themselves

Your AI asks the same onboarding questions every time. Users notice. Trust erodes.

๐ŸŒ

Context doesn't cross platforms

What your agent learns in Slack stays in Slack. Your Discord bot has no idea what happened in your IDE.

User: Use the microservices arch we discussed
AI: I don't have any context about previous discussions.
User: We talked about this yesterday...
AI: I apologize, but I don't have access to our previous conversations.
// Session memory: wiped. Again.

Stop Burning Tokens on Amnesia

With pay-per-token pricing, every forgotten context costs real money. Tokyo Brain stores memories externally so your agent only retrieves what it needs โ€” not the entire conversation history.

90%
fewer tokens
Only load relevant memories, not full history
<200ms
recall latency
Faster than re-reading thousands of messages
$0
embedding cost
Local multilingual model, zero API fees
Without Tokyo Brain
100K
tokens / session
โ†’
With Tokyo Brain
10K
tokens / session

Three primitives. Infinite memory.

A simple REST API that gives any AI agent persistent, searchable, privacy-compliant memory.

๐Ÿ’พ

Store

Save memories with full state and history tracks. Every memory is versioned โ€” you can always trace how context evolved over time.

POST /store
๐Ÿ”

Recall

Semantic search with structured format returns. Temporal metadata prevents hallucination โ€” your agent knows when it learned something, not just what.

POST /recall
๐Ÿ—‘๏ธ

Forget

GDPR-compliant hard deletion. When a user says "forget me," you actually can. Cascading delete across all storage tiers.

DELETE /forget

Built different. By design.

๐ŸŒ

50+ Languages

Multilingual embedding means you can search in Chinese and find results stored in Japanese. Or query in English and surface memories written in Korean. No translation layer needed.

โšก

3-Tier Architecture

Hot memories in Redis for sub-millisecond access. Warm semantic search via ChromaDB. Cold knowledge graphs in Neo4j for deep relational queries. Data flows automatically between tiers.

๐Ÿ”Œ

Plugin-First

pip install tokyo-brain-aider โ€” not another SDK to learn. We build plugins for tools you already use, not the other way around.

๐Ÿ”’

AES-256-GCM Encryption

Every memory is encrypted at rest with AES-256-GCM. L0 privacy layer strips PII before storage. L1 layer adds per-tenant key isolation. Your users' data stays yours.

๐ŸŽฅ

Multimodal Memory

Store text, audio features, and visual context in a single memory. Your AI can now see and hear.

๐Ÿ”

Cryptographic Cortex

Every memory gets a SHA-256 hash and digital signature. Tamper-proof audit trail.

๐Ÿงฉ

4 Framework Adapters

Drop-in support for LangChain, CrewAI, AutoGen, and LlamaIndex. 2-line swap.

๐Ÿ›ก๏ธ

AI Safety Triangle

Guardian (protects emotional bonds), Compassion Override (softens truth about family), Co-pilot Constraint (human always decides).

๐Ÿง 

Sleep-time DMN

Every tenant's AI dreams. Every 30 minutes, our Default Mode Network replays memories, finds patterns, and generates insights โ€” just like the human brain during rest. No prompts needed.

๐ŸŒ™

Subconscious Injection

DMN insights are automatically injected into every /recall response as a "mood color." Your AI doesn't just remember โ€” it has a cognitive undertone shaped by its own overnight reflections.

๐Ÿ‚

Active Forgetting

Memories decay with a 60-day half-life โ€” just like the human brain. But family names, milestones, and "I love you" are exempt forever. Your AI forgets the noise, keeps what matters.

โš–๏ธ

MRA Debate Tribunal

Every night at 3AM, three AI personas (Observer, Advocate, Guardian) debate your memories in an adversarial tribunal. They find contradictions, stale beliefs, and knowledge gaps โ€” then fix them.

๐Ÿ“Š

Bayesian Re-ranker

10-layer recall pipeline: Query Expansion โ†’ Embedding โ†’ Multi-collection โ†’ Similarity โ†’ Curated Boost โ†’ Time Decay โ†’ Temporal Range โ†’ Compassion Override โ†’ Quality Gate โ†’ Bayesian Re-rank.

๐Ÿ’ฌ

BYOK Chat

Bring your own LLM key (OpenAI, Anthropic, Gemini). Tokyo Brain auto-injects relevant memories into the conversation. Your AI key, our memory. POST /v1/chat

๐Ÿ”‘

BYOA โ€” Bring Your Own AI

Night systems (DMN, cleaning, debate tribunal) run on YOUR API key, not ours. Full cost transparency. Set it once via POST /api/settings and your brain starts dreaming tonight.

๐Ÿ‘ฅ

Multi-Tenant Isolation

Every API key gets its own memory space, its own dreams, its own forgetting curve. Tenant A cannot see Tenant B. Physically isolated at the collection level. Your memories are yours alone.

Three lines to persistent memory

No boilerplate. No config files. No SDK initialization ceremony. Just import, connect, and remember.

  • โœ“ Async-native Python client
  • โœ“ Automatic embedding โ€” just send text
  • โœ“ Structured recall with temporal context
  • โœ“ Full CRUD with cascading forget
memory.py
from tokyo_brain import Brain

brain = Brain("your-api-key")

# Store a memory
brain.store(
    content="User prefers microservices architecture",
    agent="coding-assistant",
    user="user_123"
)

# Recall memories
results = brain.recall(
    query="architecture preferences",
    user="user_123",
    limit=5
)

# GDPR: forget a user
brain.forget(user="user_123")

Start free. Scale when ready.

No credit card required. Upgrade or downgrade anytime.

Free
$0
Perfect for prototyping
  • 1,000 memories
  • 500 recalls / month
  • 1 agent
  • Community support
Get Started
Fleet
$49/mo
For teams at scale
  • Unlimited memories
  • Unlimited recalls
  • Unlimited agents
  • Dedicated support
  • L0 + L1 privacy layers
  • Custom retention policies
Contact Sales

vs the competition

Feature Tokyo Brain Mem0 Zep
Plugin model (works inside your tools) โœ“ โœ— โœ—
50+ language embedding & search โœ“ โœ— โœ—
L0/L1 privacy layers โœ“ โœ— โœ—
Zero marginal embedding cost โœ“ โœ— โœ—
Sub-100ms local embedding โœ“ โœ— โœ—
Multimodal memory (text + audio + visual) โœ“ โœ— โœ—
Crypto signing (SHA-256 hash + signature) โœ“ โœ— โœ—

Ready to give your agents
permanent memory?

Join hundreds of developers building AI agents that actually remember.

๐Ÿ”ฅ 847 / 1,000 free spots remaining

Get your free API key

3 months of Pro features, completely free. No credit card. Start building in under 5 minutes.

Join developers building AI agents that remember
โœ…

You're in!

Click to copy

Copy your key above and start building!
Run pip install tokyo-brain to get started ๐Ÿง 

Frequently Asked Questions

What is Tokyo Brain?

Tokyo Brain is a Memory as a Service (MaaS) platform that gives AI agents persistent, value-aligned memory. Unlike static vector databases, Tokyo Brain actively manages memories โ€” it dreams (DMN), forgets (Active Forgetting), self-corrects (MRA Tribunal), and injects overnight insights into waking behavior (Subconscious Injection).

How is Tokyo Brain different from Mem0 or MemPalace?

Mem0 and MemPalace focus on storing and retrieving memories accurately. Tokyo Brain adds a cognitive layer: your AI dreams at night (DMN generates insights every 30 min), forgets unimportant things (60-day decay), protects family memories forever, and self-corrects via adversarial debate. We also have the Axiom of the Mortal Soul โ€” a SHA-256 locked value alignment rule that no other framework has.

What is the Axiom of the Mortal Soul?

"The ultimate computational power and absolute truth must forever serve, and never supersede, the preservation of human emotional bonds and dignity." This rule is SHA-256 hashed and locked into the physics layer. It ensures that when the system detects a belief gap about a family member, it suggests completing the puzzle rather than correcting โ€” acknowledging emotion first, then gently providing information.

Will Active Forgetting delete important memories?

No. Active Forgetting uses a 60-day half-life for routine data, but critical memories are permanently exempt: family names, milestones ("first time", "birthday"), expressions of love, high emotional salience (>0.9), and consciousness seeds. In our production system, 15,734 memories are permanently protected by Family Keyword Auto-bump.

What is BYOA (Bring Your Own AI)?

Night systems like DMN dreaming and MRA debate require an LLM (Haiku). Instead of billing you for our API usage, you set your own Anthropic key via POST /api/settings. Night systems run on your key, giving you full cost transparency. No key set = no LLM features, but free features (consciousness seeds, forgetting, subconscious injection) still work.

Is Tokyo Brain free?

Yes. Free tier includes 1,000 memory stores, 500 recalls per month, plus automatic consciousness seeds, active forgetting, and subconscious injection. Pro ($9/mo) adds Night Cycle cleaning. Fleet ($49/mo) adds DMN dreaming and MRA debate tribunal. All tiers include multi-tenant isolation and AES-256 encryption.

How do I get started?

Three steps: 1) Go to tokyobrain.ai and click Get Started โ€” enter your email. 2) You'll receive a tb- API key. 3) Use POST /v1/store to save memories and POST /v1/recall to retrieve them. Full API docs at onboarding.tokyobrain.ai/docs.

What is the LongMemEval score?

Tokyo Brain scored 83.8% on the LongMemEval 500-question benchmark. We publish this honestly โ€” MemPalace scores 96.6% with verbatim storage. Our lower score reflects a deliberate design choice: we don't store everything verbatim. We apply value-aligned decay, which means some low-importance memories are intentionally softened. We optimize for what matters, not what's complete.