How HNSW works?
HNSW (Hierarchical Navigable Small World) is a graph-based approximate nearest neighbor (ANN) search algorithm used in LanceDB. It organizes high-dimensional vectors into a multi-layered graph, enabling fast search with logarithmic complexity. Let’s break down how HNSW works step by step.
Step 1: What is HNSW?
HNSW is a graph-based algorithm designed for efficient nearest neighbor search in high-dimensional spaces. Instead of scanning all vectors, it builds a multi-layered graph where:
- Nodes represent stored vectors.
- Edges connect similar vectors.
- Higher layers contain fewer, more representative nodes (coarse search).
- Lower layers contain dense, fine-grained connections (fine search).
Key Idea
The algorithm works by:
- Starting at a high-level node (coarse search).
- Moving toward more similar nodes (greedy search).
- Descending to lower layers (fine search).
- Stopping when the best neighbor is found.
Step 2: Constructing the HNSW Graph
2.1 Nodes and Edge Connections
Basic Definition
Suppose we store
Mathematical Construction
For each new vector
- Select nearest neighbors: Find
nearest neighbors among existing nodes using cosine similarity: - Form connections: Connect
to the top- most similar vectors. - Use layers for scalability: Assign each node to multiple layers using a probability function:
where is a constant controlling how many vectors appear at high levels.
Layered Structure
- Layer 0 (Base layer): Contains all vectors with fine-grained connections.
- Layer 1, 2, …: Contains a subset of vectors, each linking to important “hub” nodes.
Step 3: Searching in the HNSW Graph
Once the multi-layered graph is built, HNSW finds the closest vector to a query using a graph traversal strategy.
3.1 Search Algorithm
Given a query vector
- Start at the top layer:
- Pick a random entry node at the highest layer.
- Compute cosine similarity between
and that node.
- Greedy Search at Each Layer:
- Move to the most similar neighbor.
- Repeat until no closer neighbor is found.
- Descend to the Next Layer:
- Once the best node at the current layer is found, descend to the next lower layer and continue searching.
- Refine the Search at the Base Layer:
- When reaching Layer 0, HNSW expands the search locally using exact nearest-neighbor search.
Step 4: Mathematical Example of HNSW Search
Let’s consider a dataset with 4 stored vectors in 2D:
A query vector is:
Graph Structure
Assume the following edges exist (based on cosine similarity):
is connected to and . is connected to and . is connected to and . is connected to and .
Search Process
- Start at the top layer:
- Suppose we start at
.
- Suppose we start at
- Greedy Search:
- Compute cosine similarity between
and : . - Compare with neighbors
and : . .
- Since
is closer than its neighbors, we stay at .
- Compute cosine similarity between
- Descend and Expand Search:
- At Layer 0, check more neighbors:
- Compare
to and . - If no closer node is found, return
as the best match.
- Compare
- At Layer 0, check more neighbors:
Step 5: Why HNSW is Fast
HNSW prunes the search space by:
- Using hierarchical layers: Higher layers reduce the number of comparisons.
- Greedy traversal: Instead of scanning all
vectors, it only checks a small subset. - Efficient refinement: The base layer (Layer 0) ensures that the final result is highly accurate.
Computational Complexity
- Brute-force search:
. - HNSW search:
.
HNSW is exponentially faster because it follows precomputed graph links instead of scanning everything.
Final Takeaways
- HNSW builds a graph where similar vectors are connected.
- At search time, it follows the best similarity path instead of scanning all vectors.
- The result is found in
time, making it much faster than brute-force search.