Memoria - Because AI Has Amnesia

A personal memory layer that makes AI think like me, talk like me, and actually remember what I told it yesterday.

Memoria - Shubham Shrivastava's side project
Memoria - Shubham Shrivastava's side project
Memoria - Shubham Shrivastava's side project

Type

Type

Side Project

Started

Started

May 2025

Stage

Stage

v1 → v2 Rebuild

Stack

Stack

Python, Postgres, pgvector, MCP

Status

Status

Building in Public

THE IDEA

What if there was one memory layer that sat above all models?

Not ChatGPT's memory. Not Claude's memory. My memory.

A unified context layer that knows who I am, what I'm working on, how I think, and how I communicate — then passes the right context to whatever model I'm using.

One brain. Multiple models. No fragmentation.

6

AI tools I use daily

1hr

Time wasted daily re-explaining context

1

Memory layer to rule them all

THE REAL PROBLEM

I use ChatGPT, Claude, Cursor, Perplexity, Figma AI, Lovable — sometimes all in the same day.

Each one has memory. ChatGPT remembers things. Claude remembers things. Cursor has context.

But none of them talk to each other.

My ChatGPT knows I'm Head of Product at Neoflo. My Claude doesn't. My Cursor has no idea what I discussed in Perplexity. Every tool is a silo. Every memory is fragmented.

It's not that AI forgets. It's that AI remembers in isolated pockets that never connect.

I did the math. I waste roughly one hour a day re-explaining context that already exists somewhere. That's 365 hours a year — repeating myself to tools that technically "remember", just not together.

WHY EXISTING AI MEMORY FALLS SHORT

Most AI tools treat memory as:

  • App-specific

  • Model-specific

  • Implicit

  • Opaque

There are no clear rules for:

  • What's worth remembering vs noise

  • How memory updates when things change

  • How contradictions are handled

  • How history is preserved

  • How much context should be injected back

Memory behaves like a chat log or vector dump, not a dependable system.

And because memory lives inside each product, switching tools means starting over.

THE JOURNEY

THE JOURNEY

THE JOURNEY

May 2025 — MVP v1

Built the first working version. It had a chat UI, memory stored in a database, vector embeddings, semantic search, and context retrieval injected into prompts.

It worked. But the extraction was dumb.

Everything went into the database — no filtering, no lifecycle, no relevance scoring, no update logic. Memory got noisy fast: too much context, not enough signal, duplicates, contradictions, outdated facts showing up in responses.

The failure wasn't semantic search. The failure was lack of memory governance.

The "Build Everything from Scratch" Phase

My next instinct was to formalize it as a full product: custom UI, direct model API usage, bring-your-own API keys, database + embeddings, retrieval-augmented prompting.

In short: Memoria becomes the app.

This approach is valid — but flawed for this problem.

Why it wasn't ideal:

  1. It becomes a platform build, not a memory problem. I'd spend time rebuilding chat UX, auth, model switching, cost controls — things ChatGPT, Claude, and Cursor already do well.

  2. Embeddings don't solve memory correctness. Vector search helps retrieval, but it doesn't answer: what's worth saving? How do preferences update? How do time-based facts change? How do we prevent contradictions?

The real problem wasn't better retrieval. It was knowing what to store, how to update it, and when to forget it.

The Shift in Thinking

That's when the core realization clicked:

Memory is not an LLM problem. It's a systems and policy problem.

LLMs are great at understanding language. They are terrible at being consistent, auditable decision-makers.

So I stopped trying to make the model "smart about memory" and instead designed memory as infrastructure.

THE SOLUTION — MEMORIA v2

Memoria is now a personal AI memory layer, not an app.

It sits above models and tools. It does two things deliberately and separately:

1. Saving Memory Correctly

Memoria uses a governed memory pipeline.

LLMs propose meaning. Deterministic code decides policy.

The pipeline:

  • Semantic chunking — extract atomic, reusable memory candidates

  • Worthiness decisions — decide what deserves long-term memory

  • Normalization — canonical names, neutral tense, explicit facts

  • Deduplication — prevent repetition and contradiction

  • Update vs append rules:

    • Preferences → single source of truth

    • Decisions → immutable history

    • Time-varying facts (e.g. jobs) → timelines with validity windows

  • Metadata & lifecycle — importance, timestamps, expiry

  • Forgetting — outdated memory is hidden, not deleted

Memory stops being a log. It becomes small, explainable, and trustworthy.

2. Passing Context Intelligently

Saving memory is only half the problem. Using it correctly matters just as much.

Skills: Skills define how the AI should think and respond — tone, structure, reasoning style, communication preferences. They are persistent and behavioral, not data.

MCP (Model Context Protocol): MCP exposes memory as a universal interface — model-agnostic, pull-based, structured, portable across tools. Instead of injecting giant prompts, models query exactly what they need.

Retrieval discipline:

  • Strict limits (top 5–7 items)

  • Profile memory separated from task memory

  • Relevance > importance > recency ordering

The result: new chats already know me, switching models doesn't reset context, relevant background is pulled automatically, and the AI sounds like me.

It never asks "Can you give me more context?"

WHAT MEMORIA IS (AND ISN'T)

Memoria is:

  • A personal memory system

  • Model-agnostic

  • Explainable

  • Long-term stable

  • Boring by design (that's good)

Memoria is not:

  • A chat app

  • An LLM deciding what to remember

  • A vector dump

  • A journal

WHERE THIS IS GOING

Phase 1: Personal Memory (Now) Use my website and work as seed data. Build a memory that understands who I am, how I work, and how I think.

Phase 2: Agent Access Expose memory via API. Automations and agents don't start from zero — they query my memory layer and get instant context.

Phase 3: Public Chatbot A chatbot on my website that sounds like me, answers in my voice, understands my work and thinking, and can schedule calls or route conversations. Not a generic bot — a system backed by my memory.

THE END STATE

I become the API.

My knowledge, my preferences, my way of thinking — accessible to any system that needs it.

THE CORE INSIGHT

The hard part isn't storing memory.

It's knowing what to store, how to update it, and when to forget it.

Memoria exists to solve exactly that.


Building in public. Breaking things weekly.