In 2023, every AI investor was looking for the vector database play. Pinecone, Weaviate, Qdrant, Chroma, Milvus, LanceDB, Marqo, Vald, Vespa, and a dozen more all raised money on the premise that vector search was a standalone infrastructure category. Two years later, the consolidation is obvious and the survivors are defined.

Why Vector Databases Had a Moment

Retrieval-Augmented Generation (RAG) became the dominant pattern for building LLM applications in 2023. You embed your documents, store the embeddings in a vector database, and at query time retrieve the most semantically similar documents to give the LLM context.

This pattern requires vector similarity search - finding the k nearest neighbors in high-dimensional embedding space efficiently. Traditional databases did not support this natively. Specialized vector databases built exactly this and nothing else.

The bet was: RAG adoption will be so large that vector databases become essential infrastructure, like Redis for caching or Kafka for event streaming.

What Actually Happened

The bet on RAG volume was correct. The bet on standalone vector databases was mostly wrong, for two reasons.

Reason 1: Existing databases added vector support.

Postgres added pgvector in 2021. By 2024, pgvector with HNSW indexes was competitive with standalone vector databases for most use cases. A table with 1 million embeddings, properly indexed, searches in 30-50ms. Teams already running Postgres needed no additional infrastructure.

MySQL, Redis (via Redis Search), Elasticsearch, and MongoDB all added vector search. Azure Cognitive Search, Supabase, and Neon all support pgvector natively. If you already use any of these, you have vector search.

Reason 2: The differentiation was not enough.

A standalone vector database had to be significantly better than pgvector to justify adding another service to your stack. For most applications at moderate scale, it was not. Billion-scale vector search with sub-10ms latency is a real use case, but it is not most companies’ use case.

The Survivors

Pinecone is the market leader by revenue and brand recognition. They serve enterprise customers who want a managed, scalable vector search with no operational overhead. The product is polished, the API is clean, and they have succeeded in making vector search a hosted service with clear SLAs. Raised at $750M valuation. Customers who care less about per-query cost and more about reliability and support continue to use it.

Weaviate built beyond pure vector search, adding filtering, generative search, and multi-modal support. They are open-source with a managed cloud offering. The open-source strategy creates adoption that converts to revenue, similar to Elastic’s model.

Qdrant is growing among developers who self-host. Rust-based, performant, and straightforward to operate. The self-hosted story is strong and they have a managed offering for teams that want it. Popular in the European market partly due to data residency.

Chroma became the default for development and prototyping - easy to start, in-memory or persisted, no server required. But for production, most Chroma users graduate to pgvector or a managed service. Chroma did not build the enterprise features to retain production workloads.

The Consolidation Pattern

Company Status
Pinecone Profitable, enterprise focus
Weaviate Growing, open-source moat
Qdrant Growing, self-hosted focus
Chroma Prototyping tool, production conversion low
Milvus / Zilliz Large scale, mostly Asia-Pacific focus
Others Most pivoted or shutting down

Several early-stage vector database startups did not raise their Series B. The market did not develop the way they expected - not because vector search is not useful, but because it got absorbed into existing databases.

Where Specialization Still Wins

There are still cases where a standalone vector database outperforms pgvector:

Scale: Above 100 million vectors, pgvector’s HNSW index consumes significant memory and index build times are long. Pinecone, Qdrant, and Weaviate handle this more efficiently.

Multi-modal search: Searching across text, images, and structured data simultaneously is not a use case pgvector handles well. Weaviate’s multi-modal capabilities are genuinely differentiated.

Hybrid search at scale: Combining vector search with keyword search, filtering, and faceting at massive scale requires purpose-built engines.

Real-time updates: pgvector’s HNSW index reindexes on writes. At high write rates, index maintenance becomes a bottleneck. Purpose-built systems handle this better.

The Outcome for Developers

The practical takeaway: default to pgvector. If you are already on Postgres, it is already in your database. Add the extension, create an index, and vector search works. No new service, no new API key, no new pricing model.

CREATE EXTENSION vector;

ALTER TABLE documents ADD COLUMN embedding VECTOR(1536);
CREATE INDEX ON documents USING hnsw(embedding vector_cosine_ops);

SELECT id, content
FROM documents
ORDER BY embedding <=> query_embedding
LIMIT 10;

Switch to a standalone vector database when you hit real scale limits with pgvector, when you need multi-modal search, or when your write rate makes HNSW maintenance too expensive.

Bottom Line

The vector database market consolidated because existing databases added vector search. pgvector on Postgres handles most production use cases adequately, and the teams that choose it avoid adding a new service to their stack. The survivors - Pinecone for enterprise managed, Weaviate for multi-modal and open-source, Qdrant for self-hosted performance - serve real niches that pgvector does not cover. The category is not disappearing, but the “every app needs a vector database” narrative was wrong. Most apps need vector search, and many already have it in their existing database.