Vector Search Using the Deep Lake Tensor Database

Running Vector Search in the Deep Lake Tensor Database

How to Run Vector Search in the Deep Lake Tensor Database

The REST API is currently in Alpha, and the syntax may change without announcement.

Deep Lake offers a Managed Tensor Database for querying datasets, including vector search. The database is serverless, which simplifies self-hosting and substantially lowers costs. This tutorial demonstrates how to execute embedding search on a Deep Lake dataset using the REST API for the Tensor Database.

Creating the Dataset

A Deep Lake dataset containing embeddings can be created using a variety of APIs, including Deep Lake's LangChain integration.

Most importantly, when a dataset is stored in the Deep Lake Managed Tensor Database, the dataset dataset_path and runtime must be specified accordingly:

dataset_path = "hub://<org_id>/<dataset_name>"
runtime = {"db_engine": True}

# from langchain.vectorstores import DeepLake
# db = DeepLake.from_documents(texts, embeddings, dataset_path=dataset_path, runtime = runtime)

In this tutorial, the dataset has already been created can be found here.

Let's query our dataset stored in the Managed Tensor Database using the REST API. The steps are:

  1. Define the authentication tokens and search terms

  2. Embed the search search term using OpenAI

  3. Reformat the embedding to an embedding_search string that can be passed to the REST API request.

  4. Create the query string using Deep Lake TQL. The dataset_path and embedding_search are a part of the query string.

  5. Submit the request and print the response data data

import requests
import openai
import os

# Tokens should be set in environmental variables.
ACTIVELOOP_TOKEN = os.environ['ACTIVELOOP_TOKEN']
DATASET_PATH = 'hub://activeloop/twitter-algorithm'
ENDPOINT_URL = 'https://app.activeloop.ai/api/query/v1'
SEARCH_TERM = 'What do the trust and safety models do?'
# os.environ['OPENAI_API_KEY'] OPEN AI TOKEN should also exist in env variables

# The headers contains the user token
headers = {
    "Authorization": f"Bearer {ACTIVELOOP_TOKEN}",
}

# Embed the search term
embedding = openai.Embedding.create(input=SEARCH_TERM, model="text-embedding-ada-002")["data"][0]["embedding"]

# Format the embedding array or list as a string, so it can be passed in the REST API request.
embedding_string = ",".join([str(item) for item in embedding])

# Create the query using TQL
query = f"select * from (select text, cosine_similarity(embedding, ARRAY[{embedding_string}]) as score from \"{dataset_path}\") order by score desc limit 5"
          
# Submit the request                              
response = requests.post(ENDPOINT_URL, json={"query": query}, headers=headers)

data = response.json()

print(data)

Congrats! You performed a vector search using the Deep Lake Managed Database! 🎉

Last updated