Skip to main content
BenchmarkedBeta: Free Access

Magic post-it notes for your AI tools.

Remembering stuff saves you 43% in token costs.

Stompy gives your AI persistent memory. Architecture decisions, coding conventions, project context — stored once, recalled everywhere. Across Claude, Gemini, Cursor, and Codex.

Stompy the elephant waving hello

Works with every major AI coding tool

C
Claude DesktopMCP
CC
Claude CodeMCP
G
Gemini CLICLI
Cx
Codex CLICLI
Cu
CursorIDE
W
WindsurfIDE
Cl
ClineIDE
N8
N8NAutomation

MCP, CLI, or REST API — if your AI can talk, Stompy can remember

And any future model that supports MCP — Stompy grows with the ecosystem.

Real Usage

8 months of daily dogfooding. Here's what changed.

Stompy searching through code

Six agents, zero re-exploration

Swarm benchmark: the lead agent recalled architecture from memory, skipping 15 minutes of codebase scanning. Same quality, 43% less cost.

Stompy asking a question

Caught a contradiction we missed

Conflict detection found a bcrypt vs argon2 policy clash across sessions. Flagged before it shipped.

Stompy taking notes

Monday morning, zero re-explaining

New session recalled auth architecture, rate limits, and API conventions from two weeks ago. Started coding immediately.

Get Started

Three steps. Two minutes.

1

Add one line

Open Claude Desktop → Settings → Connectors → Add custom connector:

MCP Connector URL
https://mcp.stompy.ai

Click "Connect" and sign in. Works with Cursor, Codex, and any MCP client too.

2

Your AI builds your knowledge base

Tell your AI to remember architecture decisions, coding conventions, project context. It stores what matters via Stompy — no manual uploads, no copy-pasting into config files.

3

Every session builds on the last

Next session, your AI recalls everything. Architecture, rules, conventions. Across Claude, Gemini, Cursor, and Codex. No copy-pasting context. No re-explaining.

The Numbers

Proof, not promises.

43%

cost reduction across Opus & Sonnet swarm workflows

40/40

identical quality scores, Stompy vs baseline

$0.101/pt

lowest cost per quality point (Phase 2)

3 phases, 33+ runs, 3 models. N=1 per cell — promising signal, not statistical proof. Read the full methodology →

Beta: Free for Early Users

Simple, Transparent Pricing

Choose the plan that fits your usage. All plans include core Stompy features.

MonthlyAnnualSave 33%

Free

For casual users & evaluation

$0.00/mo

Usage Limits

  • 10 hrs/month active time
  • 512MB storage
  • Unlimited projects
  • Semantic search
  • Document ingestion
  • Community support
Most Popular

Pro

For regular users & hobbyists

$4.99/mo

Billed annually ($60/year)

Usage Limits

  • 75 hrs/month (~2.5 hrs/day)
  • 10GB storage
  • Everything in Free
  • Priority support
  • Advanced analytics
  • API access

Power

For heavy users & developers

$19.99/mo

Billed annually ($240/year)

Usage Limits

  • 300 hrs/month (~10 hrs/day)
  • 50GB storage
  • Everything in Pro
  • Extended usage hours
  • Bulk operations
  • Priority queue

Enterprise

For teams & organizations

Custom

Usage Limits

  • Unlimited active time
  • Dedicated compute
  • Everything in Power
  • Team collaboration
  • SSO & SAML
  • Dedicated support
Contact Us

All plans include

PostgreSQL databaseS3 document storageVoyageAI embeddingsOpenRouter Vision

We'll email you when it's ready. We won't email you about anything else.

FAQ

Frequently Asked Questions

Everything you need to know about Stompy

Getting Started

In Claude Desktop, go to Settings → Integrations → Add Custom MCP Server → enter https://mcp.stompy.ai → sign in with OAuth. That's it — 27 persistent memory tools are instantly available. For Gemini CLI, Cursor, or Codex, see /downloads for per-client setup guides.

MCP (Model Context Protocol) is an open standard for AI tool integration, used by Claude, Gemini, Cursor, and others. Stompy exposes 27 MCP tools that your AI can call directly—lock_context to store knowledge, context_search to find it semantically, recall_batch to fetch multiple contexts at once, project_brief for an AI-synthesized project overview, ticket to track tasks, detect_conflicts to find contradictions. You add one config line, your AI gets persistent memory. Not using MCP? Stompy also works via CLI commands and REST API.

No. That's the point. Stompy is fully hosted: PostgreSQL databases, S3 document storage, VoyageAI embeddings, OpenRouter Vision, NLI models—all managed. Add one block to your MCP client config (Claude Desktop, Gemini CLI, Cursor, or Codex) with your API key, and you're done. See /downloads for per-client setup guides.

Yes. Stompy connects via MCP, CLI, or REST API. Context stored by Claude is immediately available to Gemini, Codex, or any other tool. Same API key, same projects.

No. One Stompy account, one API key. Each tool connects through whichever path it supports — MCP for native integration, CLI for terminal access, REST API for everything else.

Local means you manage PostgreSQL, configure VoyageAI API keys, set up S3, handle NLI models, run Ollama. Stompy is instant: one npm command, zero infrastructure. Automatic backups, managed scaling, no DevOps. Focus on building, not maintaining memory servers.

Features & Capabilities

When you call context_search, Stompy uses VoyageAI's voyage-3.5-lite embeddings to convert your query into a vector, then finds contexts with similar meaning—not just keyword matches. Ask 'how do we handle auth?' and it finds your JWT documentation even if it never mentions 'auth'. Similarity scores show match quality (92% match, 78% match).

Stompy stores content as contexts via lock_context. You can store text, code snippets, specifications, and rules as versioned, searchable snapshots. File uploads are available through the web dashboard with async processing. All content gets VoyageAI embeddings for semantic search.

Each project gets its own PostgreSQL schema—completely isolated database tables. Every tool accepts a project parameter, so you just say context_search('auth', project='client-a') and it searches only that project. Zero bleed between projects. Automatic handover loading resumes where your last session left off.

detect_conflicts uses Natural Language Inference (NLI) to find contradictions. If you stored 'always use bcrypt' last month and 'use argon2 for passwords' today, NLI catches the contradiction. resolve_conflict lets you pick the winner: keep_a, keep_b, merge, or dismiss.

Your memory stays. Stompy is the constant. Switch from Claude to Gemini tomorrow — your architecture decisions, rules, and project context are all still there.

Technical Details

Before storing content with lock_context, Stompy analyzes novelty against existing contexts. Content with <10% new information is rejected (mostly redundant). 10-30% novelty triggers a warning. This keeps your knowledge base clean without duplicates. Use force_store=True to bypass if needed.

Yes! db_query runs read-only SELECT statements against your project's PostgreSQL schema. Auto-adds LIMIT 100 for safety. db_schema lets you inspect table structures. Full SQL power with guardrails.

You have full control. Use context_explore to see all stored contexts, db_query to inspect raw data, project_delete to remove a project entirely. The web dashboard provides a visual interface for browsing projects, contexts, and sessions.

No. Stompy uses MCP (Model Context Protocol), an open standard implemented by Claude, Gemini, Codex, Cursor, and others. We're actively adding support for new models as they launch. Your memory is yours — not locked to any AI company.

Stompy celebrating

Give your AI a memory.

One config line. Every session builds on the last.

"No memories yet. Which is fine. Totally fine. Stompy's just here. Waiting."