A vector database stores text, images, and other data as numerical vectors (embeddings) and provides fast search based on semantic similarity.
Traditional relational databases search by keyword matching. Searching for "dog" returns rows containing the string "dog" but misses "puppy" or "canine." Vector databases break through this limitation by converting text into vectors of hundreds to thousands of dimensions using embedding models and searching by "semantic distance" from the query.
Demand surged alongside the proliferation of RAG (Retrieval-Augmented Generation). To feed external knowledge to an LLM, relevant documents must first be retrieved at speed. Vector databases serve as the infrastructure for this search layer.
Major implementations include purpose-built databases like Pinecone and Weaviate, and extensions to existing databases like PostgreSQL's pgvector. pgvector is also integrated into Supabase, offering the advantage of embedding vector search within existing RDB operational workflows.
HNSW (Hierarchical Navigable Small World) is widely used as an indexing algorithm. The trade-off between search accuracy and speed, along with chunk size design and metadata filtering, are where practical expertise makes a difference.


Embedding is a technique that transforms unstructured data such as text, images, and audio into fixed-length numerical vectors while preserving semantic relationships.

Hybrid search is a technique that combines keyword-based full-text search (such as BM25) with vector search (semantic search), leveraging the strengths of both to improve retrieval accuracy.

A next-generation RAG architecture that combines knowledge graphs and vector search, leveraging relationships between entities to improve retrieval accuracy.


Gemini Embedding 2 is a multimodal embedding model developed by Google, capable of converting text, images, video, audio, and documents into a single vector space.