chromem

package module
v0.6.0 Latest Latest
Warning

This package is not in the latest version of its module.

Go to latest
Published: Apr 25, 2024 License: AGPL-3.0 Imports: 24 Imported by: 6

README

chromem-go

Go Reference Build status Go Report Card GitHub Releases

Embeddable vector database for Go with Chroma-like interface and zero third-party dependencies. In-memory with optional persistence.

Because chromem-go is embeddable it enables you to add retrieval augmented generation (RAG) and similar embeddings-based features into your Go app without having to run a separate database. Like when using SQLite instead of PostgreSQL/MySQL/etc.

It's not a library to connect to Chroma and also not a reimplementation of it in Go. It's a database on its own.

The focus is not scale (millions of documents) or number of features, but simplicity and performance for the most common use cases. On a mid-range 2020 Intel laptop CPU you can query 1,000 documents in 0.3 ms and 100,000 documents in 40 ms, with very few and small memory allocations. See Benchmarks for details.

⚠️ The project is in beta, under heavy construction, and may introduce breaking changes in releases before v1.0.0. All changes are documented in the CHANGELOG.

Contents

  1. Use cases
  2. Interface
  3. Features + Roadmap
  4. Intallation
  5. Usage
  6. Benchmarks
  7. Development
  8. Motivation
  9. Related projects

Use cases

With a vector database you can do various things:

  • Retrieval augmented generation (RAG), question answering (Q&A)
  • Text and code search
  • Recommendation systems
  • Classification
  • Clustering

Let's look at the RAG use case in more detail:

RAG

The knowledge of large language models (LLMs) - even the ones with with 30 billion, 70 billion paramters and more - is limited. They don't know anything about what happened after their training ended, they don't know anything about data they were not trained with (like your company's intranet, Jira / bug tracker, wiki or other kinds of knowledge bases), and even the data they do know they often can't reproduce it exactly, but start to hallucinate instead.

Fine-tuning an LLM can help a bit, but it's more meant to improve the LLMs reasoning about specific topics, or reproduce the style of written text or code. Fine-tuning does not add knowledge 1:1 into the model. Details are lost or mixed up. And knowledge cutoff (about anything that happened after the fine-tuning) isn't solved either.

=> A vector database can act as the the up-to-date, precise knowledge for LLMs:

  1. You store relevant documents that you want the LLM to know in the database.
  2. The database stores the embeddings alongside the documents, which you can either provide or can be created by specific "embedding models" like OpenAI's text-embedding-3-small.
    • chromem-go can do this for you and supports multiple embedding providers and models out-of-the-box.
  3. Later, when you want to talk to the LLM, you first send the question to the vector DB to find similar/related content. This is called "nearest neighbor search".
  4. In the question to the LLM, you provide this content alongside your question.
  5. The LLM can take this up-to-date precise content into account when answering.

Check out the example code to see it in action!

Interface

Our original inspiration was the Chroma interface, whose core API is the following (taken from their README):

Chroma core interface
import chromadb
# setup Chroma in-memory, for easy prototyping. Can add persistence easily!
client = chromadb.Client()

# Create collection. get_collection, get_or_create_collection, delete_collection also available!
collection = client.create_collection("all-my-documents")

# Add docs to the collection. Can also update and delete. Row-based API coming soon!
collection.add(
    documents=["This is document1", "This is document2"], # we handle tokenization, embedding, and indexing automatically. You can skip that and add your own embeddings as well
    metadatas=[{"source": "notion"}, {"source": "google-docs"}], # filter on these!
    ids=["doc1", "doc2"], # unique for each doc
)

# Query/search 2 most similar results. You can also .get by id
results = collection.query(
    query_texts=["This is a query document"],
    n_results=2,
    # where={"metadata_field": "is_equal_to_this"}, # optional filter
    # where_document={"$contains":"search_string"}  # optional filter
)

Our Go library exposes the same interface:

chromem-go equivalent
package main

import "github.com/philippgille/chromem-go"

func main() {
    // Set up chromem-go in-memory, for easy prototyping. Can add persistence easily!
    // We call it DB instead of client because there's no client-server separation. The DB is embedded.
    db := chromem.NewDB()

    // Create collection. GetCollection, GetOrCreateCollection, DeleteCollection also available!
    collection, _ := db.CreateCollection("all-my-documents", nil, nil)

    // Add docs to the collection. Update and delete will be added in the future.
    // Can be multi-threaded with AddConcurrently()!
    // We're showing the Chroma-like method here, but more Go-idiomatic methods are also available!
    _ = collection.Add(ctx,
        []string{"doc1", "doc2"}, // unique ID for each doc
        nil, // We handle embedding automatically. You can skip that and add your own embeddings as well.
        []map[string]string{{"source": "notion"}, {"source": "google-docs"}}, // Filter on these!
        []string{"This is document1", "This is document2"},
    )

    // Query/search 2 most similar results. Getting by ID will be added in the future.
    results, _ := collection.Query(ctx,
        "This is a query document",
        2,
        map[string]string{"metadata_field": "is_equal_to_this"}, // optional filter
        map[string]string{"$contains": "search_string"},         // optional filter
    )
}

Initially chromem-go started with just the four core methods, but we added more over time. We intentionally don't want to cover 100% of Chroma's API surface though.
We're providing some alternative methods that are more Go-idiomatic instead.

For the full interface see the Godoc: https://pkg.go.dev/github.com/philippgille/chromem-go

Features

  • Zero dependencies on third party libraries
  • Embeddable (like SQLite, i.e. no client-server model, no separate DB to maintain)
  • Multi-threaded processing (when adding and querying documents), making use of Go's native concurrency features
  • Experimental WebAssembly binding
  • Embedding creators:
  • Similarity search:
    • Exhaustive nearest neighbor search using cosine similarity (sometimes also called exact search or brute-force search or FLAT index)
  • Filters:
    • Document filters: $contains, $not_contains
    • Metadata filters: Exact matches
  • Storage:
    • In-memory
    • Optional immediate persistence (writes one file for each added collection and document, encoded as gob, optionally gzip-compressed)
    • Backups: Export and import of the entire DB to/from a single file (encoded as gob, optionally gzip-compressed and AES-GCM encrypted)
  • Data types:
    • Documents (text)
Roadmap
  • Performance:
    • Use SIMD for dot product calculation on supported CPUs (draft PR: #48)
    • Add roaring bitmaps to speed up full text filtering
  • Embedding creators:
    • Add an EmbeddingFunc that downloads and shells out to llamafile
  • Similarity search:
    • Approximate nearest neighbor search with index (ANN)
      • Hierarchical Navigable Small World (HNSW)
      • Inverted file flat (IVFFlat)
  • Filters:
    • Operators ($and, $or etc.)
  • Storage:
    • JSON as second encoding format
    • Write-ahead log (WAL) as second file format)
    • Optional remote storage (S3, PostgreSQL, ...)
  • Data types:
    • Images
    • Videos

Installation

go get github.com/philippgille/chromem-go@latest

Usage

See the Godoc for a reference: https://pkg.go.dev/github.com/philippgille/chromem-go

For full, working examples, using the vector database for retrieval augmented generation (RAG) and semantic search and using either OpenAI or locally running the embeddings model and LLM (in Ollama), see the example code.

Quickstart

This is taken from the "minimal" example:

package main

import (
 "context"
 "fmt"
 "runtime"

 "github.com/philippgille/chromem-go"
)

func main() {
  ctx := context.Background()

  db := chromem.NewDB()

  c, err := db.CreateCollection("knowledge-base", nil, nil)
  if err != nil {
    panic(err)
  }

  err = c.AddDocuments(ctx, []chromem.Document{
    {
      ID:      "1",
      Content: "The sky is blue because of Rayleigh scattering.",
    },
    {
      ID:      "2",
      Content: "Leaves are green because chlorophyll absorbs red and blue light.",
    },
  }, runtime.NumCPU())
  if err != nil {
    panic(err)
  }

  res, err := c.Query(ctx, "Why is the sky blue?", 1, nil, nil)
  if err != nil {
    panic(err)
  }

  fmt.Printf("ID: %v\nSimilarity: %v\nContent: %v\n", res[0].ID, res[0].Similarity, res[0].Content)
}

Output:

ID: 1
Similarity: 0.6833369
Content: The sky is blue because of Rayleigh scattering.

Benchmarks

Benchmarked on 2024-03-17 with:

  • Computer: Framework Laptop 13 (first generation, 2021)
  • CPU: 11th Gen Intel Core i5-1135G7 (2020)
  • Memory: 32 GB
  • OS: Fedora Linux 39
    • Kernel: 6.7
$ go test -benchmem -run=^$ -bench .
goos: linux
goarch: amd64
pkg: github.com/philippgille/chromem-go
cpu: 11th Gen Intel(R) Core(TM) i5-1135G7 @ 2.40GHz
BenchmarkCollection_Query_NoContent_100-8          13164      90276 ns/op     5176 B/op       95 allocs/op
BenchmarkCollection_Query_NoContent_1000-8          2142     520261 ns/op    13558 B/op      141 allocs/op
BenchmarkCollection_Query_NoContent_5000-8           561    2150354 ns/op    47096 B/op      173 allocs/op
BenchmarkCollection_Query_NoContent_25000-8          120    9890177 ns/op   211783 B/op      208 allocs/op
BenchmarkCollection_Query_NoContent_100000-8          30   39574238 ns/op   810370 B/op      232 allocs/op
BenchmarkCollection_Query_100-8                    13225      91058 ns/op     5177 B/op       95 allocs/op
BenchmarkCollection_Query_1000-8                    2226     519693 ns/op    13552 B/op      140 allocs/op
BenchmarkCollection_Query_5000-8                     550    2128121 ns/op    47108 B/op      173 allocs/op
BenchmarkCollection_Query_25000-8                    100   10063260 ns/op   211705 B/op      205 allocs/op
BenchmarkCollection_Query_100000-8                    30   39404005 ns/op   810295 B/op      229 allocs/op
PASS
ok   github.com/philippgille/chromem-go 28.402s

Development

  • Build: go build ./...
  • Test: go test -v -race -count 1 ./...
  • Benchmark:
    • go test -benchmem -run=^$ -bench . (add > bench.out or similar to write to a file)
    • With profiling: go test -benchmem -run ^$ -cpuprofile cpu.out -bench .
      • (profiles: -cpuprofile, -memprofile, -blockprofile, -mutexprofile)
  • Compare benchmarks:
    1. Install benchstat: go install golang.org/x/perf/cmd/benchstat@latest
    2. Compare two benchmark results: benchstat before.out after.out

Motivation

In December 2023, when I wanted to play around with retrieval augmented generation (RAG) in a Go program, I looked for a vector database that could be embedded in the Go program, just like you would embed SQLite in order to not require any separate DB setup and maintenance. I was surprised when I didn't find any, given the abundance of embedded key-value stores in the Go ecosystem.

At the time most of the popular vector databases like Pinecone, Qdrant, Milvus, Chroma, Weaviate and others were not embeddable at all or only in Python or JavaScript/TypeScript.

Then I found @eliben's blog post and example code which showed that with very little Go code you could create a very basic PoC of a vector database.

That's when I decided to build my own vector database, embeddable in Go, inspired by the ChromaDB interface. ChromaDB stood out for being embeddable (in Python), and by showing its core API in 4 commands on their README and on the landing page of their website.

  • Shoutout to @eliben whose blog post and example code inspired me to start this project!
  • Chroma: Looking at Pinecone, Qdrant, Milvus, Weaviate and others, Chroma stood out by showing its core API in 4 commands on their README and on the landing page of their website. It was also putting the most emphasis on its embeddability (in Python).
  • The big, full-fledged client-server-based vector databases for maximum scale and performance:
    • Pinecone: Closed source
    • Qdrant: Written in Rust, not embeddable in Go
    • Milvus: Written in Go and C++, but not embeddable as of December 2023
    • Weaviate: Written in Go, but not embeddable in Go as of March 2024 (only in Python and JavaScript/TypeScript and that's experimental)
  • Some non-specialized SQL, NoSQL and Key-Value databases added support for storing vectors and (some of them) querying based on similarity:
    • pgvector extension for PostgreSQL: Client-server model
    • Redis (1, 2): Client-server model
    • sqlite-vss extension for SQLite: Embedded, but the Go bindings require CGO. There's a CGO-free Go library for SQLite, but then it's without the vector search extension.
    • DuckDB has a function to calculate cosine similarity (1): Embedded, but the Go bindings use CGO
    • MongoDB's cloud platform offers a vector search product (1): Client-server model
  • Some libraries for vector similarity search:
    • Faiss: Written in C++; 3rd party Go bindings use CGO
    • Annoy: Written in C++; Go bindings use CGO (1)
    • USearch: Written in C++; Go bindings use CGO
  • Some orchestration libraries, inspired by the Python library LangChain, but with no or only rudimentary embedded vector DB:

Documentation

Index

Constants

View Source
const (
	InputTypeCohereSearchDocumentPrefix string = "search_document: "
	InputTypeCohereSearchQueryPrefix    string = "search_query: "
	InputTypeCohereClassificationPrefix string = "classification: "
	InputTypeCohereClusteringPrefix     string = "clustering: "
)

Prefixes for external use.

View Source
const BaseURLOpenAI = "https://api.openai.com/v1"

Variables

This section is empty.

Functions

This section is empty.

Types

type Collection

type Collection struct {
	Name string
	// contains filtered or unexported fields
}

Collection represents a collection of documents. It also has a configured embedding function, which is used when adding documents that don't have embeddings yet.

func (*Collection) Add

func (c *Collection) Add(ctx context.Context, ids []string, embeddings [][]float32, metadatas []map[string]string, contents []string) error

Add embeddings to the datastore.

  • ids: The ids of the embeddings you wish to add
  • embeddings: The embeddings to add. If nil, embeddings will be computed based on the contents using the embeddingFunc set for the Collection. Optional.
  • metadatas: The metadata to associate with the embeddings. When querying, you can filter on this metadata. Optional.
  • contents: The contents to associate with the embeddings.

This is a Chroma-like method. For a more Go-idiomatic one, see [AddDocuments].

func (*Collection) AddConcurrently added in v0.2.0

func (c *Collection) AddConcurrently(ctx context.Context, ids []string, embeddings [][]float32, metadatas []map[string]string, contents []string, concurrency int) error

AddConcurrently is like Add, but adds embeddings concurrently. This is mostly useful when you don't pass any embeddings so they have to be created. Upon error, concurrently running operations are canceled and the error is returned.

This is a Chroma-like method. For a more Go-idiomatic one, see [AddDocuments].

func (*Collection) AddDocument added in v0.4.0

func (c *Collection) AddDocument(ctx context.Context, doc Document) error

AddDocument adds a document to the collection. If the document doesn't have an embedding, it will be created using the collection's embedding function.

func (*Collection) AddDocuments added in v0.4.0

func (c *Collection) AddDocuments(ctx context.Context, documents []Document, concurrency int) error

AddDocuments adds documents to the collection with the specified concurrency. If the documents don't have embeddings, they will be created using the collection's embedding function. Upon error, concurrently running operations are canceled and the error is returned.

func (*Collection) Count added in v0.4.0

func (c *Collection) Count() int

Count returns the number of documents in the collection.

func (*Collection) Delete added in v0.6.0

func (c *Collection) Delete(_ context.Context, where, whereDocument map[string]string, ids ...string) error

Delete removes document(s) from the collection.

  • where: Conditional filtering on metadata. Optional.
  • whereDocument: Conditional filtering on documents. Optional.
  • ids: The ids of the documents to delete. If empty, all documents are deleted.

func (*Collection) Query

func (c *Collection) Query(ctx context.Context, queryText string, nResults int, where, whereDocument map[string]string) ([]Result, error)

Performs an exhaustive nearest neighbor search on the collection.

  • queryText: The text to search for. Its embedding will be created using the collection's embedding function.
  • nResults: The number of results to return. Must be > 0.
  • where: Conditional filtering on metadata. Optional.
  • whereDocument: Conditional filtering on documents. Optional.

func (*Collection) QueryEmbedding added in v0.5.0

func (c *Collection) QueryEmbedding(ctx context.Context, queryEmbedding []float32, nResults int, where, whereDocument map[string]string) ([]Result, error)

Performs an exhaustive nearest neighbor search on the collection.

  • queryEmbedding: The embedding of the query to search for. It must be created with the same embedding model as the document embeddings in the collection.
  • nResults: The number of results to return. Must be > 0.
  • where: Conditional filtering on metadata. Optional.
  • whereDocument: Conditional filtering on documents. Optional.

type DB added in v0.3.0

type DB struct {
	// contains filtered or unexported fields
}

DB is the chromem-go database. It holds collections, which hold documents.

+----+    1-n    +------------+    n-n    +----------+
| DB |-----------| Collection |-----------| Document |
+----+           +------------+           +----------+

func NewDB added in v0.3.0

func NewDB() *DB

NewDB creates a new in-memory chromem-go DB. While it doesn't write files when you add collections and documents, you can still use DB.Export and DB.Import to export and import the the entire DB from a file.

func NewPersistentDB added in v0.4.0

func NewPersistentDB(path string, compress bool) (*DB, error)

NewPersistentDB creates a new persistent chromem-go DB. If the path is empty, it defaults to "./chromem-go". If compress is true, the files are compressed with gzip.

The persistence covers the collections (including their documents) and the metadata. However it doesn't cover the EmbeddingFunc, as functions can't be serialized. When some data is persisted and you create a new persistent DB with the same path, you'll have to provide the same EmbeddingFunc as before when getting an existing collection and adding more documents to it.

Currently the persistence is done synchronously on each write operation, and each document addition leads to a new file, encoded as gob. In the future we will make this configurable (encoding, async writes, WAL-based writes, etc.).

In addition to persistence for each added collection and document you can use DB.Export and DB.Import to export and import the entire DB to/from a file, which also works for the pure in-memory DB.

func (*DB) CreateCollection added in v0.3.0

func (db *DB) CreateCollection(name string, metadata map[string]string, embeddingFunc EmbeddingFunc) (*Collection, error)

CreateCollection creates a new collection with the given name and metadata.

  • name: The name of the collection to create.
  • metadata: Optional metadata to associate with the collection.
  • embeddingFunc: Optional function to use to embed documents. Uses the default embedding function if not provided.

func (*DB) DeleteCollection added in v0.4.0

func (db *DB) DeleteCollection(name string) error

DeleteCollection deletes the collection with the given name. If the collection doesn't exist, this is a no-op. If the DB is persistent, it also removes the collection's directory. You shouldn't hold any references to the collection after calling this method.

func (*DB) Export added in v0.5.0

func (db *DB) Export(filePath string, compress bool, encryptionKey string) error

Export exports the DB to a file at the given path. The file is encoded as gob, optionally compressed with flate (as gzip) and optionally encrypted with AES-GCM. This works for both the in-memory and persistent DBs. If the file exists, it's overwritten, otherwise created.

  • filePath: If empty, it defaults to "./chromem-go.gob" (+ ".gz" + ".enc")
  • compress: Optional. Compresses as gzip if true.
  • encryptionKey: Optional. Encrypts with AES-GCM if provided. Must be 32 bytes long if provided.

func (*DB) GetCollection added in v0.4.0

func (db *DB) GetCollection(name string, embeddingFunc EmbeddingFunc) *Collection

GetCollection returns the collection with the given name. The embeddingFunc param is only used if the DB is persistent and was just loaded from storage, in which case no embedding func is set yet (funcs are not (de-)serializable). It can be nil, in which case the default one will be used. The returned collection is a reference to the original collection, so any methods on the collection like Add() will be reflected on the DB's collection. Those operations are concurrency-safe. If the collection doesn't exist, this returns nil.

func (*DB) GetOrCreateCollection added in v0.4.0

func (db *DB) GetOrCreateCollection(name string, metadata map[string]string, embeddingFunc EmbeddingFunc) (*Collection, error)

GetOrCreateCollection returns the collection with the given name if it exists in the DB, or otherwise creates it. When creating:

  • name: The name of the collection to create.
  • metadata: Optional metadata to associate with the collection.
  • embeddingFunc: Optional function to use to embed documents. Uses the default embedding function if not provided.

func (*DB) Import added in v0.5.0

func (db *DB) Import(filePath string, encryptionKey string) error

Import imports the DB from a file at the given path. The file must be encoded as gob and can optionally be compressed with flate (as gzip) and encrypted with AES-GCM. This works for both the in-memory and persistent DBs. Existing collections are overwritten.

- filePath: Mandatory, must not be empty - encryptionKey: Optional, must be 32 bytes long if provided

func (*DB) ListCollections added in v0.4.0

func (db *DB) ListCollections() map[string]*Collection

ListCollections returns all collections in the DB, mapping name->Collection. The returned map is a copy of the internal map, so it's safe to directly modify the map itself. Direct modifications of the map won't reflect on the DB's map. To do that use the DB's methods like CreateCollection() and DeleteCollection(). The map is not an entirely deep clone, so the collections themselves are still the original ones. Any methods on the collections like Add() for adding documents will be reflected on the DB's collections and are concurrency-safe.

func (*DB) Reset added in v0.4.0

func (db *DB) Reset() error

Reset removes all collections from the DB. If the DB is persistent, it also removes all contents of the DB directory. You shouldn't hold any references to old collections after calling this method.

type Document added in v0.4.0

type Document struct {
	ID        string
	Metadata  map[string]string
	Embedding []float32
	Content   string
}

Document represents a single document.

func NewDocument added in v0.4.0

func NewDocument(ctx context.Context, id string, metadata map[string]string, embedding []float32, content string, embeddingFunc EmbeddingFunc) (Document, error)

NewDocument creates a new document, including its embeddings. Metadata is optional. If the embeddings are not provided, they are created using the embedding function. You can leave the content empty if you only want to store embeddings. If embeddingFunc is nil, the default embedding function is used.

If you want to create a document without embeddings, for example to let Collection.AddDocuments create them concurrently, you can create a document with `chromem.Document{...}` instead of using this constructor.

type EmbeddingFunc

type EmbeddingFunc func(ctx context.Context, text string) ([]float32, error)

EmbeddingFunc is a function that creates embeddings for a given text. chromem-go will use OpenAI`s "text-embedding-3-small" model by default, but you can provide your own function, using any model you like. The function must return a *normalized* vector, i.e. the length of the vector must be 1. OpenAI's and Mistral's embedding models do this by default. Some others like Nomic's "nomic-embed-text-v1.5" don't.

func NewEmbeddingFuncCohere added in v0.5.0

func NewEmbeddingFuncCohere(apiKey string, model EmbeddingModelCohere) EmbeddingFunc

NewEmbeddingFuncCohere returns a function that creates embeddings for a text using Cohere's API. One important difference to OpenAI's and other's APIs is that Cohere differentiates between document embeddings and search/query embeddings. In order for this embedding func to do the differentiation, you have to prepend the text with either "search_document" or "search_query". We'll cut off that prefix before sending the document/query body to the API, we'll just use the prefix to choose the right "input type" as they call it.

When you set up a chromem-go collection with this embedding function, you might want to create the document separately with NewDocument and then cut off the prefix before adding the document to the collection. Otherwise when you query the collection, the returned documents will still have the prefix in their content.

cohereFunc := chromem.NewEmbeddingFuncCohere(cohereApiKey, chromem.EmbeddingModelCohereEnglishV3)
content := "The sky is blue because of Rayleigh scattering."
// Create the document with the prefix.
contentWithPrefix := chromem.InputTypeCohereSearchDocumentPrefix + content
doc, _ := NewDocument(ctx, id, metadata, nil, contentWithPrefix, cohereFunc)
// Remove the prefix so that later query results don't have it.
doc.Content = content
_ = collection.AddDocument(ctx, doc)

This is not necessary if you don't keep the content in the documents, as chromem-go also works when documents only have embeddings. You can also keep the prefix in the document, and only remove it after querying.

We plan to improve this in the future.

func NewEmbeddingFuncDefault added in v0.3.0

func NewEmbeddingFuncDefault() EmbeddingFunc

NewEmbeddingFuncDefault returns a function that creates embeddings for a text using OpenAI`s "text-embedding-3-small" model via their API. The model supports a maximum text length of 8191 tokens. The API key is read from the environment variable "OPENAI_API_KEY".

func NewEmbeddingFuncJina added in v0.3.0

func NewEmbeddingFuncJina(apiKey string, model EmbeddingModelJina) EmbeddingFunc

NewEmbeddingFuncJina returns a function that creates embeddings for a text using the Jina API.

func NewEmbeddingFuncLocalAI added in v0.3.0

func NewEmbeddingFuncLocalAI(model string) EmbeddingFunc

NewEmbeddingFuncLocalAI returns a function that creates embeddings for a text using the LocalAI API. You can start a LocalAI instance like this:

docker run -it -p 127.0.0.1:8080:8080 localai/localai:v2.7.0-ffmpeg-core bert-cpp

And then call this constructor with model "bert-cpp-minilm-v6". But other embedding models are supported as well. See the LocalAI documentation for details.

func NewEmbeddingFuncMistral added in v0.3.0

func NewEmbeddingFuncMistral(apiKey string) EmbeddingFunc

NewEmbeddingFuncMistral returns a function that creates embeddings for a text using the Mistral API.

func NewEmbeddingFuncMixedbread added in v0.3.0

func NewEmbeddingFuncMixedbread(apiKey string, model EmbeddingModelMixedbread) EmbeddingFunc

NewEmbeddingFuncMixedbread returns a function that creates embeddings for a text using the mixedbread.ai API.

func NewEmbeddingFuncOllama added in v0.4.0

func NewEmbeddingFuncOllama(model string, baseURLOllama string) EmbeddingFunc

NewEmbeddingFuncOllama returns a function that creates embeddings for a text using Ollama's embedding API. You can pass any model that Ollama supports and that supports embeddings. A good one as of 2024-03-02 is "nomic-embed-text". See https://ollama.com/library/nomic-embed-text baseURLOllama is the base URL of the Ollama API. If it's empty, "http://localhost:11434/api" is used.

func NewEmbeddingFuncOpenAI added in v0.3.0

func NewEmbeddingFuncOpenAI(apiKey string, model EmbeddingModelOpenAI) EmbeddingFunc

NewEmbeddingFuncOpenAI returns a function that creates embeddings for a text using the OpenAI API.

func NewEmbeddingFuncOpenAICompat added in v0.3.0

func NewEmbeddingFuncOpenAICompat(baseURL, apiKey, model string, normalized *bool) EmbeddingFunc

NewEmbeddingFuncOpenAICompat returns a function that creates embeddings for a text using an OpenAI compatible API. For example:

The `normalized` parameter indicates whether the vectors returned by the embedding model are already normalized, as is the case for OpenAI's and Mistral's models. The flag is optional. If it's nil, it will be autodetected on the first request (which bears a small risk that the vector just happens to have a length of 1).

type EmbeddingModelCohere added in v0.5.0

type EmbeddingModelCohere string
const (
	EmbeddingModelCohereMultilingualV2      EmbeddingModelCohere = "embed-multilingual-v2.0"
	EmbeddingModelCohereEnglishLightV2      EmbeddingModelCohere = "embed-english-light-v2.0"
	EmbeddingModelCohereEnglishV2           EmbeddingModelCohere = "embed-english-v2.0"
	EmbeddingModelCohereMultilingualLightV3 EmbeddingModelCohere = "embed-multilingual-light-v3.0"
	EmbeddingModelCohereEnglishLightV3      EmbeddingModelCohere = "embed-english-light-v3.0"
	EmbeddingModelCohereMultilingualV3      EmbeddingModelCohere = "embed-multilingual-v3.0"
	EmbeddingModelCohereEnglishV3           EmbeddingModelCohere = "embed-english-v3.0"
)

type EmbeddingModelJina added in v0.3.0

type EmbeddingModelJina string
const (
	EmbeddingModelJina2BaseEN   EmbeddingModelJina = "jina-embeddings-v2-base-en"
	EmbeddingModelJina2BaseDE   EmbeddingModelJina = "jina-embeddings-v2-base-de"
	EmbeddingModelJina2BaseCode EmbeddingModelJina = "jina-embeddings-v2-base-code"
	EmbeddingModelJina2BaseZH   EmbeddingModelJina = "jina-embeddings-v2-base-zh"
)

type EmbeddingModelMixedbread added in v0.3.0

type EmbeddingModelMixedbread string
const (
	EmbeddingModelMixedbreadUAELargeV1          EmbeddingModelMixedbread = "UAE-Large-V1"
	EmbeddingModelMixedbreadBGELargeENV15       EmbeddingModelMixedbread = "bge-large-en-v1.5"
	EmbeddingModelMixedbreadGTELarge            EmbeddingModelMixedbread = "gte-large"
	EmbeddingModelMixedbreadE5LargeV2           EmbeddingModelMixedbread = "e5-large-v2"
	EmbeddingModelMixedbreadMultilingualE5Large EmbeddingModelMixedbread = "multilingual-e5-large"
	EmbeddingModelMixedbreadMultilingualE5Base  EmbeddingModelMixedbread = "multilingual-e5-base"
	EmbeddingModelMixedbreadAllMiniLML6V2       EmbeddingModelMixedbread = "all-MiniLM-L6-v2"
	EmbeddingModelMixedbreadGTELargeZh          EmbeddingModelMixedbread = "gte-large-zh"
)

type EmbeddingModelOpenAI added in v0.3.0

type EmbeddingModelOpenAI string
const (
	EmbeddingModelOpenAI2Ada   EmbeddingModelOpenAI = "text-embedding-ada-002"
	EmbeddingModelOpenAI3Small EmbeddingModelOpenAI = "text-embedding-3-small"
	EmbeddingModelOpenAI3Large EmbeddingModelOpenAI = "text-embedding-3-large"
)

type Result

type Result struct {
	ID        string
	Metadata  map[string]string
	Embedding []float32
	Content   string

	// The cosine similarity between the query and the document.
	// The higher the value, the more similar the document is to the query.
	// The value is in the range [-1, 1].
	Similarity float32
}

Result represents a single result from a query.

Directories

Path Synopsis
example module
examples
minimal Module

Jump to

Keyboard shortcuts

? : This menu
/ : Search site
f or F : Jump to
y or Y : Canonical URL