Euclidean β’ Angular β’ Bitwise β’ Haversine β’ User-Defined Metrics
C++ 11 β’
Python 3 β’
JavaScript β’
Java β’
Rust β’
C 99 β’
Objective-C β’
Swift β’
C# β’
GoLang β’
Wolfram
Linux β’ MacOS β’ Windows β’ iOS β’ Docker β’ WebAssembly
- β Benchmark-topping performance.
- β Simple and extensible single C++11 header implementation.
- β SIMD-optimized and user-defined metrics with JIT compilation.
- β Variable dimensionality vectors for unique applications, including search over compressed data.
- β Bitwise Tanimoto and Sorensen coefficients for Genomics and Chemistry applications.
- β
Hardware-agnostic
f16
&i8
- half-precision & quarter-precision support. - β View large indexes from disk without loading into RAM.
- β
Space-efficient point-clouds with
uint40_t
, accommodating 4B+ size. - β Compatible with OpenMP and custom "executors", for fine-grained control over CPU utilization.
- β Heterogeneous lookups, renaming/relabeling, and on-the-fly deletions.
- β Near-real-time clustering and sub-clusterings for Tens or Millions of clusters.
- β Semantic Search and Joins.
FAISS is a widely recognized standard for high-performance vector search engines. USearch and FAISS both employ the same HNSW algorithm, but they differ significantly in their design principles. USearch is compact and broadly compatible without sacrificing performance, with a primary focus on user-defined metrics and fewer dependencies.
FAISS | USearch | |
---|---|---|
Implementation | 84 K SLOC in faiss/ |
3 K SLOC in usearch/ |
Supported metrics | 9 fixed metrics | Any User-Defined metrics |
Supported languages | C++, Python | 10 languages |
Supported ID types | uint32_t , uint64_t |
uint32_t , uint40_t , uint64_t |
Dependencies | BLAS, OpenMP | None |
Bindings | SWIG | Native |
Acceleration | Learned Quantization | Downcasting |
Base functionality is identical to FAISS, and the interface must be familiar if you have ever investigated Approximate Nearest Neighbors search:
$ pip install usearch
import numpy as np
from usearch.index import Index
index = Index(
ndim=3, # Define the number of dimensions in input vectors
metric='cos', # Choose 'l2sq', 'haversine' or other metric, default = 'ip'
dtype='f32', # Quantize to 'f16' or 'i8' if needed, default = 'f32'
connectivity=16, # Optional: How frequent should the connections in the graph be
expansion_add=128, # Optional: Control the recall of indexing
expansion_search=64, # Optional: Control the quality of search
)
vector = np.array([0.2, 0.6, 0.4])
index.add(42, vector)
matches: Matches = index.search(vector, 10)
assert len(index) == 1
assert len(matches) == 1
assert matches[0].key == 42
assert matches[0].distance <= 0.001
assert np.allclose(index[42], vector)
Comparing the performance of FAISS against USearch on 1 Million 96-dimensional vectors from the famous Deep1B dataset, once can expect the following numbers on modern AWS c7g.metal
instances.
FAISS, f32 |
USearch, f32 |
USearch, f16 |
USearch, i8 |
|
---|---|---|---|---|
Batch Insert | 16 K/s | 73 K/s | 100 K/s | 104 K/s +550% |
Batch Search | 82 K/s | 103 K/s | 113 K/s | 134 K/s +63% |
Bulk Insert | 76 K/s | 105 K/s | 115 K/s | 202 K/s +165% |
Bulk Search | 118 K/s | 174 K/s | 173 K/s | 304 K/s +157% |
Recall @ 10 | 99% | 99.2% | 99.1% | 99.2% |
HNSW was configured with identical hyper-parameters: connectivity
M=16
, expansion @ constructionefConstruction=128
, and expansion @ searchef=64
. Batch size is 256. Jump to the Performance Tuning section to read about the effects of those hyper-parameters.
While most vector search packages concentrate on just a couple of metrics - "Inner Product distance" and "Euclidean distance," USearch extends this list to include any user-defined metrics. This flexibility allows you to customize your search for a myriad of applications, from computing geo-spatial coordinates with the rare Haversine distance to creating custom metrics for composite embeddings from multiple AI models.
Unlike older approaches indexing high-dimensional spaces, like KD-Trees and Locality Sensitive Hashing, HNSW doesn't require vectors to be identical in length. They only have to be comparable. So you can apply it in obscure applications, like searching for similar sets or fuzzy text matching, using GZip as a distance function.
Read more about JIT and UDF in USearch Python SDK.
Training a quantization model and dimension-reduction is a common approach to accelerate vector search. Those, however, are only sometimes reliable, can significantly affect the statistical properties of your data, and require regular adjustments if your distribution shifts.
Instead, we have focused on high-precision arithmetic over low-precision downcasted vectors.
The same index, and add
and search
operations will automatically down-cast or up-cast between f32_t
, f16_t
, f64_t
, and i8_t
representations, even if the hardware doesn't natively support it.
Continuing the topic of memory efficiency, we provide a uint40_t
to allow collection with over 4B+ vectors without allocating 8 bytes for every neighbor reference in the proximity graph.
USearch supports multiple forms of serialization:
- Into a file defined with a path.
- Into a stream defined with a callback, serializing or reconstructing incrementally.
- Into a buffer of fixed length, or a memory-mapped file, that supports random access.
The latter allows you to serve indexes from external memory, enabling you to optimize your server choices for indexing speed and serving costs. This can result in 20x cost reduction on AWS and other public clouds.
index.save("index.usearch")
loaded_copy = index.load("index.usearch")
view = Index.restore("index.usearch", view=True)
other_view = Index(ndim=..., metric=CompiledMetric(...))
other_view.view("index.usearch")
Approximate search methods, such as HNSW, are predominantly used when an exact brute-force search becomes too resource-intensive.
This typically occurs when you have millions of entries in a collection.
For smaller collections, we offer a more direct approach with the search
method.
from usearch.index import search, MetricKind, Matches, BatchMatches
import numpy as np
# Generate 10'000 random vectors with 1024 dimensions
vectors = np.random.rand(10_000, 1024).astype(np.float32)
vector = np.random.rand(1024).astype(np.float32)
one_in_many: Matches = search(vectors, vector, 50, MetricKind.L2sq, exact=True)
many_in_many: BatchMatches = search(vectors, vectors, 50, MetricKind.L2sq, exact=True)
By passing the exact=True
argument, the system bypasses indexing altogether and performs a brute-force search through the entire dataset using SIMD-optimized similarity metrics from SimSIMD.
When compared to FAISS's IndexFlatL2
in Google Colab, USearch may offer up to a 20x performance improvement:
faiss.IndexFlatL2
: 55.3 ms.usearch.index.search
: 2.54 ms.
For larger workloads targeting billions or even trillions of vectors, parallel multi-index lookups become invaluable. These lookups prevent the need to construct a single, massive index, allowing users to query multiple smaller ones instead.
from usearch.index import Indexes
multi_index = Indexes(
indexes: Iterable[usearch.index.Index] = [...],
paths: Iterable[os.PathLike] = [...],
view: bool = False,
threads: int = 0,
)
multi_index.search(...)
Once the index is constructed, it can be used to cluster entries much faster.
In essense, the Index
itself can be seen as a clustering, and it allows iterative deepening.
clustering = index.cluster(
min_count=10, # Optional
max_count=15, # Optional
threads=..., # Optional
)
# Get the clusters and their sizes
centroid_keys, sizes = clustering.centroids_popularity
# Use Matplotlib draw a histogram
clustering.plot_centroids_popularity()
# Export a NetworkX graph of the clusters
g = clustering.network
# Get members of a specific cluster
first_members = clustering.members_of(centroid_keys[0])
# Deepen into that cluster spliting it into more parts, all same arguments supported
sub_clustering = clustering.subcluster(min_count=..., max_count=...)
Using Scikit-Learn, on a 1 Million point dataset, one may expect queries to take anywhere from minutes to hours, depending on the number of clusters you want to highlight. For 50'000 clusters the performance difference between USearch and conventional clustering methods may easily reach 100x.
One of the big questions these days is how will AI change the world of databases and data management.
Most databases are still struggling to implement high-quality fuzzy search, and the only kind of joins they know are deterministic.
A join
is different from searching for every entry, as it requires a one-to-one mapping, banning collisions among separate search results.
Exact Search | Fuzzy Search | Semantic Search ? |
---|---|---|
Exact Join | Fuzzy Join ? | Semantic Join ?? |
Using USearch one can implement sub-quadratic complexity approximate, fuzzy, and semantic joins. This can come in handy in any fuzzy-matching tasks, common to Database Management Software.
men = Index(...)
women = Index(...)
pairs: dict = men.join(women, max_proposals=0, exact=False)
Read more in post: From Dating to Vector Search - "Stable Marriages" on a Planetary Scale π©ββ€οΈβπ¨
By now, the core functionality is supported across all bindings. Broader functionality is ported per request.
C++ 11 | Python 3 | C 99 | Java | JavaScript | Rust | GoLang | Swift | |
---|---|---|---|---|---|---|---|---|
Add, search | β | β | β | β | β | β | β | β |
Save, load, view | β | β | β | β | β | β | β | β |
User-defined metrics | β | β | β | β | β | β | β | β |
Joins | β | β | β | β | β | β | β | β |
Variable-length vectors | β | β | β | β | β | β | β | β |
4B+ capacities | β | β | β | β | β | β | β | β |
AI has a growing number of applications, but one of the coolest classic ideas is to use it for Semantic Search. One can take an encoder model, like the multi-modal UForm, and a web-programming framework, like UCall, and build a text-to-image search platform in just 20 lines of Python.
import ucall
import uform
import usearch
import numpy as np
import PIL as pil
server = ucall.Server()
model = uform.get_model('unum-cloud/uform-vl-multilingual')
index = usearch.index.Index(ndim=256)
@server
def add(key: int, photo: pil.Image.Image):
image = model.preprocess_image(photo)
vector = model.encode_image(image).detach().numpy()
index.add(key, vector.flatten(), copy=True)
@server
def search(query: str) -> np.ndarray:
tokens = model.preprocess_text(query)
vector = model.encode_text(tokens).detach().numpy()
matches = index.search(vector.flatten(), 3)
return matches.keys
server.run()
A more complete demo with Streamlit is available on GitHub. We have pre-processed some commonly used datasets, cleaned the images, produced the vectors, and pre-built the index.
Dataset | Modalities | Images | Download |
---|---|---|---|
Unsplash | Images & Descriptions | 25 K | HuggingFace / Unum |
Conceptual Captions | Images & Descriptions | 3 M | HuggingFace / Unum |
Arxiv | Titles & Abstracts | 2 M | HuggingFace / Unum |
Comparing molecule graphs and searching for similar structures is expensive and slow. It can be seen as a special case of the NP-Complete Subgraph Isomorphism problem. Luckily, domain-specific approximate methods exist. The one commonly used in Chemistry, is to generate structures from SMILES, and later hash them into binary fingerprints. The latter are searchable with bitwise similarity metrics, like the Tanimoto coefficient. Below is an example using the RDKit package.
from usearch.index import Index, MetricKind
from rdkit import Chem
from rdkit.Chem import AllChem
import numpy as np
molecules = [Chem.MolFromSmiles('CCOC'), Chem.MolFromSmiles('CCO')]
encoder = AllChem.GetRDKitFPGenerator()
fingerprints = np.vstack([encoder.GetFingerprint(x) for x in molecules])
fingerprints = np.packbits(fingerprints, axis=1)
index = Index(ndim=2048, metric=MetricKind.Tanimoto)
keys = np.arange(len(molecules))
index.add(keys, fingerprints)
matches = index.search(fingerprints, 10)
With Objective-C and Swift iOS bindings, USearch can be easily used in mobile applications. The SwiftVectorSearch project illustrates how to build a dynamic, real-time search system on iOS. In this example, we use 2-dimensional vectorsβencoded as latitude and longitudeβto find the closest Points of Interest (POIs) on a map. The search is based on the Haversine distance metric, but can easily be extended to support high-dimensional vectors.
- GPTCache: Python.
- LangChain: Python and JavaScipt.
- ClickHouse: C++.
- Microsoft Semantic Kernel: Python and C#.
@software{Vardanian_USearch_2022,
doi = {10.5281/zenodo.7949416},
author = {Vardanian, Ash},
title = {{USearch by Unum Cloud}},
url = {https://github.com/unum-cloud/usearch},
version = {2.3.1},
year = {2022},
month = jun,
}