Memvid is not another vector database. It’s a portable, single-file memory system with flexible embedding options. Use local models for privacy or connect to external providers like OpenAI, NVIDIA, and more.
The Problem with Traditional RAG
Most memory systems today follow the same pattern: This approach has serious limitations:| Problem | Impact |
|---|---|
| API dependency | Can’t work everyhwere, costs money per query |
| Embedding drift | Model updates break your index |
| No exact matching | ”Error 404” doesn’t find “error 404” |
| Black box relevance | Hard to debug why results are wrong |
| Cold start | Need to embed everything before first search |
The Memvid Innovation
Memvid takes a completely different approach: Flexible embedding options. Memvid combines multiple search strategies:- BM25 lexical search - Battle-tested, fast, explainable
- Vector embeddings - Local models (Nomic, BGE, GTE) or external APIs (OpenAI, NVIDIA)
- SimHash deduplication - Find near-duplicates instantly
- Time-aware retrieval - When something was added matters
- Hybrid search - Combines lexical + semantic for best results
How Frames Change Everything
Traditional systems store “chunks” - arbitrary text splits. Memvid stores Frames - structured units of memory:| Field | Example |
|---|---|
content | ”The quarterly revenue exceeded $10M…” |
uri | reports/q4-2024.pdf |
timestamp | 2024-12-15T10:30:00Z |
content_hash | blake3(…) |
simhash | 0x8f3a2b1c… |
- What it contains (content + hash)
- Where it came from (URI + metadata)
- When it was created (timestamp)
- What it’s similar to (SimHash)
- What entities it mentions (Logic Mesh)
- How it connects to other frames (relationships)
Multiple Indices, Flexible Providers
Every.mv2 file contains multiple search indices:
| Index | Purpose | How It Works |
|---|---|---|
| Lexical (BM25) | Keyword search | TF-IDF scoring, exact matches |
| Vector | Semantic similarity | Local or external embedding models |
| SimHash | Near-duplicate detection | 64-bit locality-sensitive hash |
| Time | Temporal queries | B-tree on timestamps |
| Logic Mesh | Entity-relationship | Triple store (subject-predicate-object) |
- Local models (Nomic, BGE-small, BGE-base, GTE-large) - Fast, private, works offline
- External APIs (OpenAI, NVIDIA) - Higher quality, no local compute needed
Choosing Your Embedding Provider
Memvid supports multiple embedding providers. Choose based on your needs:| Provider | Models | Best For |
|---|---|---|
| Local (Default) | bge-small, bge-base, nomic, gte-large | Privacy, offline use, no API costs |
| OpenAI | openai, openai-small, openai-ada | Highest quality, multilingual |
| NVIDIA | nvidia | Enterprise, high throughput |
Logic Mesh: Relationships Without ML
Traditional systems need expensive NER models for entity extraction. Memvid’s Logic Mesh uses:- Rule-based extraction - Fast, free, no API
- Pattern matching - Dates, emails, numbers
- Co-occurrence - Entities mentioned together
- Temporal reasoning - When facts changed
SimHash: Smart Deduplication
Instead of comparing embeddings, Memvid uses SimHash - a locality-sensitive hash that detects near-duplicates:- Instant - O(1) comparison
- No API calls - Computed locally
- **Works anywhere **- No internet needed
- Deterministic - Same input = same hash
Real-World Performance
Memvid’s local-first approach delivers fast performance:| Operation | Traditional Vector DB | Memvid (Local Embeddings) |
|---|---|---|
| Ingest 1,000 docs | 5-10 minutes | 30 seconds |
| First search | After embedding | Instant |
| Offline search | ❌ | ✅ |
| Cost per query | $0.0001+ | $0 |
| Single file portability | ❌ | ✅ |
Getting Started
Get started with Memvid in minutes:Switching Embedding Providers
Need higher quality embeddings? Switch to an external provider:Next Steps
5-Minute Quickstart
Get up and running fast
Frame Architecture
Deep dive into frames
Logic Mesh
Entity-relationship graphs
Deduplication
SimHash and content hashing