Training Models at Scale
Train models at scale using Deep Lake
Train models at scale using Deep Lake
There are several Deep Lake related tuning parameters that affect the speed of Deep Lake OSS and Enterprise dataloaders. The plot below shows performance of the Deep Lake dataloaders under different scenarios, and it is discussed in detail below.
The Deep Lake Enterprise dataloader streams data faster compared to the OSS dataloaer, due to its C++ implementation that optimizes asynchronous data fetching and decompression.
The Enterprise dataloader is ~1.5-3X faster compared to the OSS version, depending on the complexity of the transform and the number of workers available for parallelization.
Distributed training is only available in the Enterprise dataloader.
num_workers
Both the OSS and Enterprise dataloaders in Deep Lake have a num_workers
parameter that parallelizes the data fetching, decompression, and transformation.
Increasing num_workers
will not improve performance in GPU-bottlenecked scenarios. Therefore, we recommend starting with 2-4 workers, and increasing num_workers
it if the GPU utilization is low.
Increasing num_workers
beyond the number of CPUs on a machine does not improve performance.
It is common for GPU machines to have 8x CPUs per GPU
Increasing num_workers
linearly improves streaming speed, with diminishing returns beyond 8+ workers.
Increasing num_workers
beyond 16 is generally unnecessary, unless you are running complex transformations.
decode_method
for imagesFaster dataloading is achieved by minimizing the amount of operations that take place before data is delivered to the GPU. It is important to the decode_method
parameter in the OSS and Enterprise dataloaders based on the following guidelines:
When transforming images using tools the require numpy arrays as inputs, such as Albumentations, decode_method
should be to numpy, which is the default (No parameters changes are needed)
When transforming images using tools the require PIL images as inputs, such as torchvision transforms, decode_method
should be to {'image_tensor_name': 'pil'}
. torchvision.transforms.ToPIL()
should be removed from the top of the transforms stack.
Leaving the decode_method
as numpy may decrease dataloading speed by up to 2X, because the image is decoded to a numpy array and then re-encoded as a PIL image, instead of being directly decoded to a PIL image.
The torchvision transforms used to create the comparison in the plot above are: