Part VI Storing and Searching Vectors in Coherence

With the increased popularity of Gen AI, and Retrieval Augmented Generation (RAG) use cases in particular, the need for a way to store and search a large number of dense vector embeddings efficiently is larger than ever.

Vector DB

Coherence already provides a number of features that make this possible, such as efficient serialization format, filters, and aggregators, which allow you to search across large data sets in parallel by using all the CPU cores in a cluster, and gRPC proxy, which allows remote clients written in any supported language, including Python, to access data in a Coherence cluster efficiently.

This release adds the missing bits that turn Coherence into a full-fledged Vector Database:
  1. Built-in support for Vector Types: float32, int8, and bit dense vectors of arbitrary dimension.
  2. Built-in support for Semantic Search, including:
    • HNSW indexing
    • Binary Quantization
    • Index-Optimized Exact Searches
    • Metadata Filtering
  3. Built-in support for Document Chunks, addressing a common RAG use case.
  4. Integration with LangChain and LangChain4j.
  5. Integration with Spring AI.

Part VI contains the following chapters and provides more details about each of these features: