A database optimized for storing and searching high-dimensional vector embeddings.
A vector database is a specialized data storage system designed to efficiently index, store, and query high-dimensional vectors — numerical arrays that encode the semantic or perceptual content of data such as text, images, audio, or user behavior. Unlike traditional relational databases optimized for exact lookups on structured fields, vector databases are built around approximate nearest neighbor (ANN) search algorithms, which identify the most semantically similar items to a query vector without exhaustively scanning every record. Common indexing strategies include HNSW (Hierarchical Navigable Small World graphs), IVF (Inverted File Index), and product quantization, each offering different tradeoffs between speed, memory, and recall accuracy.
The practical need for vector databases grew directly from the widespread adoption of embedding models — neural networks that map raw data into dense vector representations where geometric proximity reflects semantic similarity. When a language model encodes a sentence as a 768-dimensional vector, or a vision model encodes an image as a 2048-dimensional feature vector, the downstream task often requires finding other vectors nearby in that space. Doing this at scale — across millions or billions of embeddings — demands infrastructure that relational and document databases simply cannot provide efficiently.
Vector databases have become foundational to retrieval-augmented generation (RAG) pipelines, where a language model's responses are grounded by retrieving relevant documents from a vector index at inference time. They also power semantic search engines, recommendation systems, duplicate detection, and multimodal retrieval applications. Purpose-built systems like Pinecone, Weaviate, Milvus, and Qdrant emerged alongside this demand, while established databases such as PostgreSQL (via pgvector) and Redis added vector search extensions to remain competitive.
The rise of vector databases reflects a broader shift in how AI systems manage knowledge: rather than hard-coding information into model weights, practitioners increasingly store knowledge externally as searchable embeddings. This makes systems more updatable, auditable, and scalable. As embedding models grow more capable and deployment contexts more diverse, vector databases have become a core piece of production AI infrastructure, sitting at the intersection of machine learning and systems engineering.