Lesson format
Reading lesson with structured written material and clear navigation inside the course flow.
Reading lesson with structured written material and clear navigation inside the course flow.
This lesson is free and open to all visitors.
CLAUDE.md and Obsidian are both human-curated: you decide what to write, where to put it, what to link. That is their strength and their ceiling. At some point your archive — transcripts, meeting recordings, research papers, customer emails — becomes too large to curate. Every week adds content you'll never link by hand.
That is where vector memory earns its place. Not as a replacement for the first two layers, but as a third layer that handles what they can't: retrieving meaning from huge unstructured archives.
A vector database stores text as embeddings — long lists of numbers that capture what the text means. Similar meanings produce similar number lists. When you ask a question, the question is also turned into a vector, and the database returns the chunks of text whose vectors are closest to it. It is search by meaning, not keyword.
Obsidian works on explicit relationships you wrote down. A wikilink is there because you put it there. A tag is there because you tagged it. If you forgot to tag a note with #ai, Obsidian cannot know it is about AI — unless you grep the word "AI" in the content, which is literal, not semantic.
Vector search finds notes about AI even when the word "AI" doesn't appear, because the meaning of the text is similar to the query. That is a fundamentally different capability. It is cheaper than curating, but it only returns what it saw at ingest — no reasoning over structure, no links.
Pinecone is a managed vector database. You pay a small amount per month, you POST your text chunks with their embeddings, and you query by sending a new text (or its embedding) and getting back the top-k most similar chunks. Alternatives: Weaviate, Chroma, Milvus, Supabase Vector. Pinecone is the most common default for small teams because it is fully managed and has a simple REST API.
Notice what just happened: Claude looked at five chunks out of millions. It didn't scan the archive. It asked the archive "what here is about this?" and got a short list back. That is the fundamental move vector search enables.
Embeddings are produced by a specialized embedding model (e.g., OpenAI's text-embedding-3-large, Voyage's voyage-3-large, Cohere's embed). They capture semantic meaning — so "our pricing is flexible" and "we offer custom quotes" end up close in vector space even though the words differ.
They do not capture structure. The embedding of a paragraph doesn't know whether that paragraph is a header, footnote, or a question. For that, you need metadata stored alongside the vector (which Pinecone supports).
You need vector memory when:
You do not need it when:
Vector memory is the right tool for a huge, messy pile you'll never clean. If your business isn't yet producing a huge messy pile, don't build the tool for it — build the smaller, cheaper tools first.