Vector Search Options
Overview of Vector Search Options in Deep Lake
Overview of Vector Search Options in Deep Lake
Deep Lake offers a variety of vector search options depending on the Storage Location of the Vector Store and infrastructure that should run the computations.
Storage Location | Compute Location | Execution Algorithm |
---|---|---|
Vector search can occur via a variety of APIs in Deep Lake. They are explained in the links below:
The optimal option for search execution is automatically selected based on the Vector Stores storage location. The different computation options are explained below.
Deep Lake OSS offers query execution logic that run on the client (your machine) using OSS code in Python. This compute logic is accessible in all Deep Lake Python APIs and is available for Vector Stores stored in any location. See individual APIs below for details.
Deep Lake Compute Engine offers query execution logic that run on the client (your machine) using C++ Code that is called via Python API. This compute logic is accessible in all Deep Lake Python APIs and is only available for Vector Stores stored Deep Lake storage or in user clouds connected to Deep Lake. See individual APIs below for details.
To run queries using Compute Engine, make sure to !pip install "deeplake[enterprise]"
.
Deep Lake offers a Managed Tensor Database that executes queries on Deep Lake infrastructure while running Compute Engine under-the-hood. This compute logic is accessible in all Deep Lake Python APIs and is only available for Vector Stores stored in the Deep Lake Managed Tensor Database. See individual APIs below for details.
In memory or local
Client-side
Deep Lake OSS Python Code
User cloud (must be connected to Deep Lake)
Client-side
Deep Lake C++ Compute Engine
Deep Lake Storage
Client-side
Deep Lake C++ Compute Engine
Deep Lake Managed Tensor Database
Managed Database
Deep Lake C++ Compute Engine