Chroma
Embedded vector database for AI apps. Runs in-process like SQLite, prototype-first.
VISIT CHROMAQuick facts
- CategoryEmbedded
- EnginePython / Rust core
- PricingOpen source
- LicenseApache-2.0
- Created2022
- GitHub stars22.4k
- Hybrid searchNo
- Edge-readyNo
- Multi-tenantSingle-tenant
- Max dimensionsunlimited
What it is
Chroma runs in-process — like SQLite for vectors. Python-first SDK, simple API, optional client-server mode for production. The default vector DB for prototype RAG apps and LangChain / LlamaIndex tutorials. Distinct from purpose-built engines: Chroma is for the hobbyist-to-mid-scale tier, not for big production workloads.
Best for
- Prototype RAG apps where adding any infrastructure is overhead
- Single-tenant Python tools embedded in another app
- Small to mid-scale workloads under ~5M vectors
When not to pick it
Skip Chroma for high-traffic production workloads — it does not yet match Pinecone or Qdrant on latency and throughput at scale. Skip for non-Python stacks; the Python SDK is the path of least resistance.
My take
Chroma is the right pick for prototypes and Python tools. For serious RAG production, graduate to pgvector (if Postgres-aligned) or Qdrant (if not).
Links
Compare Chroma side-by-side
Similar tools you should also consider
If Chroma is your pick — the next conversation is short
The 30-min call is where your vector-DB choice becomes a real RAG architecture, a chunking + reranking strategy that actually works for your corpus, and a price range you can take to your stakeholders. Describe your data shape, your query patterns, your latency budget. I tell you whether Chroma is genuinely your fit.