AI Memory Compared: How Ditto, ChatGPT, Claude, and Gemini Handle Long-Term Memory
You spend an hour walking an AI assistant through your project — the tech stack, the trade-offs, why you chose Postgres over Mongo. Next week you start a new chat and it asks what programming language you use.
That gap — between what an AI could know about you and what it actually retains — is the core problem every major assistant is now racing to solve. ChatGPT builds a two-layer profile from your conversations. Claude stores transparent Markdown files. Gemini taps into your entire Google ecosystem. Ditto builds a persistent knowledge graph you can actually see, search, and take with you.
But “memory” means very different things in each system. If you’re choosing an AI assistant for ongoing work — coding, research, writing, project management — the way it handles memory will shape your entire experience.
This post breaks down how each system works in 2026, what you can see and control, and where each one falls short.
ChatGPT Memory: Two Layers, One Black Box
ChatGPT’s memory has evolved significantly since its initial launch in early 2024. It now operates as a two-layer system:
- Saved Memories: Explicit facts ChatGPT stores — either automatically extracted from conversations or when you say “remember this.” Viewable and deletable in Settings > Personalization > Manage Memories.
- Chat History Reference: Introduced April 2025, ChatGPT now references all past conversations to personalize responses — not just saved memories. It builds an evolving user profile that includes device type, model usage patterns, activity habits, and more.
In October 2025, OpenAI overhauled memory management. You can now search and sort saved memories via a search bar in Settings. Automatic memory management prioritizes memories by recency and frequency, preventing the “memory full” error that frustrated early adopters. Less important memories are shown in gray but not deleted.
Projects add scoped memory — organized workspaces where ChatGPT can reference other conversations within the same project. Project sharing is available on all tiers, with support for Slack channels, Google Drive files, and notes as project sources.
Proactive memory arrived with Pulse (September 2025) — a daily briefing feature that uses your memories and connected apps to deliver personalized morning cards. Currently Pro-only ($200/month).
OpenAI also fully adopted MCP as of October 2025, supporting read and write actions via custom apps and connectors in ChatGPT.
What’s good: It works automatically with zero effort. The two-layer system means ChatGPT picks up on patterns you’d never think to explicitly tell it. Memory search and management have real UI now.
What’s missing: You can’t see how the chat history profile influences responses — security researchers have raised concerns about the opaque “dossier” it builds. Custom GPTs still lack persistent memory. And if you switch to another AI tool, your memories stay locked in OpenAI’s ecosystem — their data export produces raw JSON that’s essentially unusable for import elsewhere.
Claude Memory: From Manual to Automatic
Claude’s memory story has changed dramatically. What started as manual-only Projects has become a full persistent memory system — and the transparency of its approach sets it apart.
Memory launched for Team and Enterprise plans in September 2025, expanded to all paid tiers with automatic memory in October 2025, and was made free for all users in March 2026. Anthropic also shipped a memory import tool for migrating from ChatGPT and Gemini.
How it works:
- Uses a transparent, file-based approach with Markdown files — not opaque vector databases
- Memory is opt-in and user-controlled: you can download memories per-project and export them
- Projects remain a core feature — self-contained workspaces with uploaded files, custom instructions, and shared context
- Skills add organized folders of instructions and scripts that Claude loads dynamically — including pre-built skills for PowerPoint, Excel, Word, and PDF
- Compaction enables infinite-length conversations — when chats approach the context limit, Claude automatically summarizes earlier messages and continues
Context windows: Current Claude models support 200K tokens standard, with a 1M token beta available for select models.
MCP: Claude Desktop supports MCP integrations, and Claude Code acts as a full MCP client connecting to external tools. Anthropic also offers a dedicated Memory Tool in the API for developers.
What’s good: Transparent memory you can see, edit, and export. The Markdown-based approach means your memories are inherently portable. The import tool makes switching from competitors easy. Skills and Projects provide structured organization.
What’s missing: Memory is newer than ChatGPT’s, so it’s still maturing. Cross-project memory and shared knowledge bases are limited — Anthropic is building Knowledge Bases for Claude Cowork but they’re not released yet. No visual memory graph or relationship mapping.
Gemini Memory: Your Google Life, Contextualized
Gemini takes a fundamentally different approach — it leverages your existing Google ecosystem as memory:
- Persistent memory: Gemini stores user preferences and facts across conversations using vector embeddings, with time-tagged entries that distinguish between current and historical context.
- Learning from past chats: Launched August 2025, Gemini can learn from your conversation history over time, complemented by a “Temporary Chat” mode for private conversations.
- Personal Intelligence: Launched January 2026, this connects Gemini to Gmail, Google Photos, Search history, and YouTube history. Gemini can reason across your data proactively — like suggesting winter tires after seeing family road trip photos, or pulling your license plate number from a photo.
- Google Workspace integration: Generally available since May 2025 — type
@Gmail,@Drive, or@Calendarto pull data from any Google service directly into conversations. - NotebookLM integration: As of December 2025, you can attach NotebookLM notebooks as live data sources in Gemini. Notebooks support up to 300 sources and have significantly longer conversation memory than standard chats.
- Deep Research: An agentic research system that leverages massive context windows to analyze 100+ sources and produce comprehensive multi-page reports with citations. It can pull from your Workspace content — emails, chats, and Drive docs — automatically.
Context windows: Gemini leads with massive windows — up to 1M tokens standard (2M via API) on current flagship models. Gemini 2.5 Pro achieves 99.7% recall at 1M tokens.
Gems: Custom AI personas that can now be backed by NotebookLM notebooks — upgrading from 10-file limits to 300-source knowledge bases that update dynamically.
Google has also fully embraced MCP with managed servers for Google Maps, BigQuery, and more — and their MCP servers work with Claude and ChatGPT as clients too.
What’s good: If you live in the Google ecosystem, no other AI has this kind of contextual access. Personal Intelligence turns years of emails, photos, and searches into instant context. The massive context windows and Deep Research are genuinely best-in-class for research tasks.
What’s missing: Memory is tied to the Google ecosystem — useful only if you’re a heavy Google user. The “memory” is more about accessing existing data than building new understanding. There’s no visible knowledge graph or relationship mapping between topics. Personal Intelligence is only available on AI Pro (249.99/mo) in the U.S. And the privacy implications of connecting your entire digital life to an AI are significant — even with Google’s opt-in controls.
Ditto Memory: Visible, Searchable, and Portable
Ditto wasn’t built to compete with ChatGPT and then bolt on memory later. Memory is the foundation — every feature flows from the idea that your AI should build a persistent understanding of you over time. Where other assistants treat memory as a feature to check off, Ditto treats it as the entire point.
How It Works in Practice
Every conversation with Ditto builds your personal knowledge graph automatically. Here’s what that actually looks like:
Say you’ve been using Ditto for a month — chatting about a React migration at work, planning a trip to Portugal, and brainstorming a side project. Open your knowledge graph and you’ll see an interactive network: nodes for “React,” “SolidJS,” “Portugal,” “Lisbon restaurants,” and “side project” — with weighted connections showing how these topics relate based on your actual conversations. The graph grows and evolves as you talk. Patterns in your thinking you didn’t know existed start to emerge.
Now ask Ditto “what was that architecture decision we discussed last week?” Here’s what happens behind the scenes: a trained neural network predicts optimal retrieval weights for your specific query in under a millisecond. It dynamically balances three signals — semantic similarity, recency, and discussion frequency — based on what your query actually needs. A question about “yesterday” prioritizes recent memories. A question about “that recurring bug” prioritizes frequently discussed topics. A question about “Python projects” leans on semantic similarity.
This is the learned retrieval weights system, described in detail in Omar’s engineering post. One database round-trip. All scoring is atomic in SQL. No black box — and no knobs for you to fiddle with. It just works.
Transparent Retrieval
Most AI assistants use your memories silently — you never know which memories influenced a response or why they were chosen. Ditto shows you everything:
- Retrieved memory cards: Expandable panels on every response showing exactly which memories were used
- Predicted weights: The similarity, recency, and frequency percentages the system chose for your query
- Per-memory scores: Color-coded bars breaking down each memory’s contribution (blue for similarity, amber for recency, emerald for frequency)
- Predicted intent: Whether your query was classified as semantic, temporal, or frequency-oriented
This transparency isn’t just a nice-to-have — it builds trust. When you can see why Ditto surfaced a specific memory, you understand how the system thinks and can work with it more effectively.
Context Steering with Ditto Threads
Sometimes passive memory isn’t enough — you need to tell your AI exactly what to focus on. Ditto Threads let you attach subjects, memories, and notes directly to a conversation — creating persistent, context-rich threads you can return to anytime.
Picture this: you’re switching between three projects during the day — a client deliverable, an open-source contribution, and a personal coding project. Instead of hoping the AI retrieves the right context each time, you attach relevant subjects from your knowledge graph, pin key decision memories, and add freeform notes to each thread. Switch threads, switch contexts — no confusion, no bleed-through between topics. And unlike every other AI’s threading, Ditto Threads never go stale. Come back a month later and the AI picks up right where you left off, with full context.
The Full Feature Set
- Persistent storage: Every conversation is saved — short-term memory for recent context, long-term memory via semantic search across your entire history
- Knowledge graph: Automatic extraction of subjects, people, topics, and concepts with an interactive network visualization
- Semantic search: Find “that API architecture discussion from last month” by meaning, not keywords
- Multi-provider models: Use OpenAI, Anthropic, Google, Meta, xAI, and more — switch models mid-conversation without losing context
- MCP portability: Ditto is both an MCP server and client. Your memories work in Claude, Cursor, Cline, and any MCP-compatible tool. Connect external servers like Zapier and GitHub to extend Ditto’s capabilities
- Personality adaptation: Big Five, MBTI, and DISC assessments that shape how Ditto communicates with you
What’s good: Memory is the product, not an add-on. You can see your knowledge graph grow, search across months of conversations by meaning, inspect exactly why any memory was retrieved, and use your memories in other AI tools via MCP. Because Ditto works with models from OpenAI, Anthropic, Google, and Meta, your memories are never tied to a single provider’s capabilities or pricing changes — pick the best model for each task, and your memory follows you.
What’s missing: The ecosystem of plugins and integrations is still growing. Proactive features like daily briefings and weekly reflections are on the roadmap but not shipped yet. And because Ditto is model-agnostic, context window limits depend on whichever provider you choose.
Feature-by-Feature Comparison
| Capability | Ditto | ChatGPT | Claude | Gemini |
|---|---|---|---|---|
| Automatic memory | Yes — every conversation saved | Yes — two-layer system (facts + chat history profile) | Yes — Markdown-based, free for all users | Yes — vector embeddings + Google data |
| Memory visibility | Full knowledge graph visualization | Hidden list in Settings; opaque profile | Transparent Markdown files | List in Settings; ecosystem data implicit |
| Cross-conversation search | Semantic search across all history | Search bar in Settings (Oct 2025) | Per-project; Knowledge Bases coming | Via Google Search history integration |
| Context steering | Ditto Threads (attach subjects, memories, notes) | Projects + Custom Instructions | Projects + Skills + system prompt | Gems + NotebookLM notebooks |
| Memory transparency | Shows retrieved memories with scored weights | No indication of memory use | Shows memory source files | No indication of memory use |
| Multi-provider models | Yes — OpenAI, Anthropic, Google, Meta, more | GPT/o-series models only | Claude models only | Gemini models only |
| Memory portability (MCP) | Full MCP server + client | MCP apps (no memory export) | MCP client + memory export | Managed MCP servers (no memory export) |
| Personality adaptation | Big Five, MBTI, DISC assessments | Custom Instructions | System prompt + Skills | Personal Intelligence (Google data) |
| Organization tools | Bookmarks, collections, goals | Projects, chat folders | Projects, Skills, Artifacts | Gems, NotebookLM |
| Knowledge graph | Interactive network visualization | No | No | No |
| Context window | Uses provider’s window | 128K–200K tokens | 200K–1M tokens | 1M–2M tokens |
| Proactive features | Coming soon | Pulse (Pro only, $200/mo) | Cowork (research preview) | Personal Intelligence ($19.99+/mo) |
| Memory import | N/A — provider-agnostic | No import tool | Import from ChatGPT/Gemini | Import (beta) |
When Each Tool Works Best
Choose ChatGPT if:
- You want zero-effort memory that works in the background
- You primarily use GPT/o-series models and don’t need to switch providers
- You want proactive daily briefings (Pulse) and are on the Pro plan
- You’re comfortable with opaque memory — not seeing exactly what’s stored or how it’s used
Choose Claude if:
- You want transparent, file-based memory you can inspect and export
- You need powerful context windows (up to 1M tokens in beta)
- You’re switching from ChatGPT — Claude’s memory import tool makes migration easy
- You value Skills and Projects for structured, repeatable workflows
Choose Gemini if:
- You live in the Google ecosystem (Gmail, Drive, Calendar, Photos)
- You want your AI to reason across your existing digital life via Personal Intelligence
- You need massive context windows (up to 2M tokens) or Deep Research for multi-source analysis
- You use NotebookLM for knowledge management
Choose Ditto if:
- You want to see and understand how your AI remembers you — with an actual knowledge graph
- You work across multiple topics and need connections between them mapped visually
- You want to use your memories across different AI tools (via MCP)
- You value transparent retrieval — knowing exactly which memories shaped each response
- You want model flexibility — pick the best model for each task from any provider
The Portability Problem
Here’s something most people don’t think about when choosing an AI assistant: lock-in.
If you build up months of conversations in ChatGPT, those memories are locked in OpenAI’s system. Their data export produces raw JSON that’s essentially unusable for import elsewhere. Claude is better — Anthropic stores memories as transparent Markdown files you can download, and they’ve built an import tool to pull memories from other providers. Gemini’s memories are deeply tied to Google services — useful within that ecosystem, but not portable outside it.
Ditto solves this with MCP integration. Your Ditto memories are accessible from any MCP-compatible AI tool. Connect Ditto’s MCP server to Claude Desktop, Cursor, or Cline, and those tools can search your memory, retrieve context, and even save new memories back to Ditto.
One memory system that works everywhere. No lock-in.
What’s Coming Next
Ditto’s roadmap is focused on making memory even more proactive:
- Conversation Sessions — named, switchable threads for different areas of your life
- Weekly Reflections — auto-generated summaries of your topics and goal progress
- Proactive Nudges — “You mentioned your interview is tomorrow — good luck!”
- Voice Mode — real-time voice conversations with full memory continuity
Try It Yourself
Sign up for Ditto — free, no credit card — and have five conversations across different topics. Then open your knowledge graph. Watch the nodes form, the connections emerge, the patterns in your thinking you didn’t know existed.
That’s the moment most people get it: this is what memory should feel like.
If you’re already in Claude or Cursor, connect Ditto via MCP and start building your memory graph without changing your workflow.
Questions? We’re a small, stubborn team that reads every message — reach out at support@heyditto.ai.
Sources
- ChatGPT Memory Launch (OpenAI Help Center)
- ChatGPT Memory Improvements (MacRumors, April 2025)
- How ChatGPT Remembers You: A Deep Dive (Embrace The Red)
- ChatGPT Memory Dossier Criticism (Simon Willison)
- ChatGPT Memory Management Update (Blockchain News, October 2025)
- OpenAI Automatic Memory Management (October 2025)
- ChatGPT Projects (OpenAI Help Center)
- ChatGPT Project Sharing (AI Fire)
- ChatGPT Pulse (TechCrunch, September 2025)
- ChatGPT MCP Support (InfoQ, October 2025)
- ChatGPT Connectors (OpenAI Help Center)
- Custom GPTs Memory Limitations (Lead With AI)
- ChatGPT Token Limits and Context Windows (Data Studios)
- ChatGPT Data Export (OpenAI Help Center)
- Claude Persistent Memory (Reworked, September 2025)
- Claude Automatic Memory for All Paid Tiers (MacRumors, October 2025)
- Claude Memory Free for All Users (Engadget, March 2026)
- Claude Memory Import Tool (MacRumors, March 2026)
- Claude Memory Architecture Deep Dive (Skywork AI)
- Claude Memory Export (VentureBeat)
- Claude Skills and 2026 Launches (Medium)
- Claude Context Windows (Anthropic Docs)
- Claude Memory Tool API (Anthropic Docs)
- Claude Compaction / Infinite Chats (DEV Community)
- Claude Code MCP Docs
- Claude Knowledge Bases (TestingCatalog)
- Anthropic Import Tool (SiliconANGLE, March 2026)
- Gemini Long-Term Memory Features (VCSolutions)
- Gemini Temporary Chats and Personalization (Google Blog)
- Gemini Personal Intelligence (Google Blog, January 2026)
- Gemini Personal Intelligence (TechCrunch, January 2026)
- Google Workspace Apps for Gemini (Google Workspace Updates, May 2025)
- NotebookLM Integration with Gemini (Phandroid, December 2025)
- NotebookLM 300 Sources (Gadget Hacks)
- NotebookLM Source Memory Upgrade (Geeky Gadgets)
- Gemini Deep Research (Google)
- Deep Research Workspace Integration (Google Workspace Updates, November 2025)
- Gemini Long Context (Google AI Docs)
- Gemini 2.5 Pro Recall at 1M Tokens (Latenode)
- Gemini Gems (Google)
- NotebookLM as Source in Gemini (Google Workspace Updates, January 2026)
- Google AI Pro and Ultra Features (9to5Google, February 2026)
- Google MCP Support for Google Services (Google Cloud Blog)
- Google Embraces MCP (The New Stack)
- ChatGPT Memory Portability Discussion (OpenAI Community)
- Learned Retrieval Weights: How Ditto Picks the Right Memories (Ditto Engineering Blog)
- Connect Ditto to Any AI Assistant: MCP Integration Guide (Ditto Blog)
- Ditto Threads: AI Chat Threads That Never Go Stale (Ditto Blog)