Numpy Reader

Overview

This example shows how to read Numpy array files (*.npy), with DALI’s readers.numpy reader. This notebook also shows how to use DALI to load numpy files directly to GPU memory, thanks to NVIDIA GPUDirect Storage, and how to use the region-of-interest (ROI) API to load regions of the array.

The reader extracts the shape and the data type information directly from the files. Please note that only Numpy v1 (and not v2) files are currently supported. Numpy v1 are the most commonly used. See the numpy file format specification for more details.

The operator returns arrays with shapes taken from the files. DALI tensors are always stored in C (row-major) order. If the files contain the data in FORTRAN (column-major) order, the operator will automatically transpose the data to C order. This transposition adds significant time to the loading process. Therefore, we recommend storing files in C order when possible.

Test data

In this example we are going to work with 3D MRI data, available in the DALI_extra repository. The environment variable DALI_EXTRA_PATH should point to the location of DALI_extra. Ensure that you check out the correct release tag that corresponds to the installed version of DALI.

We will be working with 3D MRI scans in two formats:

  1. 2D slices, with one file per slice:

npy_2d_slices
└── STU00001
    ├── SER00001
    │   ├── 0.npy
    │   ├── 1.npy
    │   ├── 2.npy
    │   ├── 3.npy
    │   └── 4.npy
  1. 3D arrays with the whole volume

npy_3d
└── STU00001
    ├── SER00001.npy
[1]:
import os

batch_size = 4  # to be used in pipelines
dali_extra_dir = os.environ['DALI_EXTRA_PATH']
data_dir_2d = os.path.join(dali_extra_dir, 'db', '3D', 'MRI', 'Knee', 'npy_2d_slices', 'STU00001')
data_dir_3d = os.path.join(dali_extra_dir, 'db', '3D', 'MRI', 'Knee', 'npy_3d', 'STU00001')

Let us introduce a utility function to plot batches of samples:

[2]:
# misc python stuff
import numpy as np
from glob import glob
import shutil
import tempfile

# visualization
from PIL import Image
import matplotlib.pyplot as plt

def plot_batch(np_arrays, nsamples=None):
    if nsamples is None:
        nsamples = len(np_arrays)
    fig, axvec = plt.subplots(nrows=1, ncols=nsamples, figsize=(10, 10 * nsamples))
    for i in range(nsamples):
        ax = axvec[i]
        ax.tick_params(left=False, bottom=False, labelleft=False, labelbottom=False)
        ax.imshow(Image.fromarray(np_arrays[i]))
    plt.tight_layout()

We can check that the data is equivalent in both representations:

[3]:
ser00001_2d = [np.load(os.path.join(data_dir_2d, 'SER00001', f'{i}.npy')) for i in range(4)]
plot_batch(ser00001_2d)
../../../_images/examples_general_data_loading_numpy_reader_6_0.png
[4]:
ser00001_3d = np.load(os.path.join(data_dir_3d, 'SER00001.npy'))
plot_batch([ser00001_3d[0], ser00001_3d[1], ser00001_3d[2], ser00001_3d[3]])
../../../_images/examples_general_data_loading_numpy_reader_7_0.png

Usage

The readers.numpy reader can be configured in different ways. Let us see some examples.

Glob filter

Read all files under a specied directory specified by the argument file_root, that match a particular glob string, specified by the file_filter argument.

[5]:
from nvidia.dali import pipeline_def, fn

data_dir = os.path.join(data_dir_2d, 'SER00001')

@pipeline_def(batch_size=batch_size, num_threads=3, device_id=0)
def pipe1():
    data = fn.readers.numpy(device='cpu', file_root=data_dir, file_filter='*.npy')
    return data

def run(p):
    p.build()  # build the pipeline
    outputs = p.run()  # Run once
    # Getting the batch as a list of numpy arrays, for displaying
    batch = [np.array(outputs[0][s]) for s in range(batch_size)]
    return batch

data1 = run(pipe1())
plot_batch(data1)
../../../_images/examples_general_data_loading_numpy_reader_10_0.png

Text file with a list of file paths

Read all files specified by a text file where each line contains a path to a file, relative to the file_root directory.

[6]:
# Used to verify that two batches are identical
def assert_all_equal(batch1, batch2):
    assert len(batch1) == len(batch2)
    batch_size = len(batch1)
    for i in range(batch_size):
        np.testing.assert_array_equal(batch1[i], batch2[i])

# Listing all *.npy files in data_dir
files = sorted([f for f in os.listdir(data_dir) if '.npy' in f])

# Defining a pipeline taking the file list file path as an argument
@pipeline_def(batch_size=batch_size, num_threads=3, device_id=0)
def pipe2(file_list_path):
    data = fn.readers.numpy(device='cpu', file_root=data_dir, file_list=filelist_path)
    return data

# Creating a temporary directory to contain the "file list" file.
with tempfile.TemporaryDirectory() as tmp_dir:
    # Creating a file_list text file, which contains the names of the data files to be loaded
    filelist_path = os.path.join(tmp_dir, "test_file_list.txt")
    with open(filelist_path, "w") as f:
        f.writelines("\n".join(files))

    # Displaying the contents of the file we just created
    print("File list contents:")
    with open(filelist_path, 'r') as f:
        print(f.read())
    print('\n')

    data2 = run(pipe2(file_list_path=filelist_path))
    assert_all_equal(data1, data2)
File list contents:
0.npy
1.npy
2.npy
3.npy
4.npy


List of file paths

Read all files specified by a list of strings containing file paths, relative to the file_root directory.

[7]:
# Defining a pipeline taking a list of files directly as a keyword argument
@pipeline_def(batch_size=batch_size, num_threads=3, device_id=0)
def pipe3():
    data = fn.readers.numpy(device='cpu', file_root=data_dir, files=files)
    return data

data3 = run(pipe3())
assert_all_equal(data1, data3)

Higher dimensionality

Numpy reader can also be used to load data of higher dimensionality (up to 6 dimensions). Let’s demonstrate with our 3D MRI arrays

[8]:
@pipeline_def(batch_size=batch_size, num_threads=3, device_id=0)
def pipe_3d():
    data = fn.readers.numpy(device='cpu', file_root=data_dir_3d, file_filter='*.npy')
    return data

data_3d = run(pipe_3d())
for s in range(len(data_3d)):
    print(f"Sample {s} shape: {data_3d[s].shape}")

# Displaying first 4 slices of the first sample in the batch
plot_batch([data_3d[0][0, :, :], data_3d[0][1, :, :], data_3d[0][2, :, :], data_3d[0][3, :, :]])
Sample 0 shape: (5, 512, 512)
Sample 1 shape: (9, 512, 512)
Sample 2 shape: (2, 512, 512)
Sample 3 shape: (38, 320, 320)
../../../_images/examples_general_data_loading_numpy_reader_16_1.png

Region-of-interest (ROI) API

In the example shown above, we see that the relevant data is concentrated in the upper left quadrant of the image and the rest does not contain useful information. This is not true for all the images in the dataset, but it will serve us as a good example to demonstrate ROI reading.

Numpy reader allows the user to specify a region of interest, equivalent to the arguments specified to slice operation. The benefit is that the reader will only read the relevant part of the file, saving I/O bandwidth and memory utilization. Note that setting the dont_use_mmap argument to False will negate that performance benefit, with the entire file being read first and then sliced.

The ROI can be specified in absolute or relative terms, and can be specified on a subset of the array’s axes. For dimensions not specified in the ROI, the whole extent of the array shall be used. For full documentation on ROI decoding, refer to `readers.numpy documentation <https://docs.nvidia.com/deeplearning/dali/user-guide/docs/supported_ops.html#nvidia.dali.fn.readers.numpy>`__ (see arguments roi_start, rel_roi_start, roi_end, rel_roi_end, roi_shape, rel_roi_shape, axes).

Let us see a few practical examples:

ROI start and end, in absolute coordinates

[9]:
# One file is enough for this example
files = ['0.npy']

@pipeline_def(batch_size=batch_size, num_threads=3, device_id=0)
def pipe_roi1():
    data = fn.readers.numpy(device='cpu', file_root=data_dir, files=files,
                            roi_start=[30, 30], roi_end=[230, 230])
    return data

data_roi1 = run(pipe_roi1())
plot_batch(data_roi1)
../../../_images/examples_general_data_loading_numpy_reader_19_0.png

ROI start and end, in relative coordinates

[10]:
@pipeline_def(batch_size=batch_size, num_threads=3, device_id=0)
def pipe_roi2():
    data = fn.readers.numpy(device='cpu', file_root=data_dir, files=files,
                            rel_roi_start=[0.1, 0.01], rel_roi_end=[0.4, 0.5])
    return data

data_roi2 = run(pipe_roi2())
plot_batch(data_roi2)
../../../_images/examples_general_data_loading_numpy_reader_21_0.png

Specifying a subset of the array’s axes

[11]:
@pipeline_def(batch_size=batch_size, num_threads=3, device_id=0)
def pipe_roi3(rel_roi_start, rel_roi_end, roi_axes):
    data = fn.readers.numpy(device='cpu', file_root=data_dir, files=files,
                            rel_roi_start=rel_roi_start, rel_roi_end=rel_roi_end, roi_axes=roi_axes)
    return data

p1 = pipe_roi3(rel_roi_start=[0.1], rel_roi_end=[0.4], roi_axes=(0,))
data_roi3_1 = run(p1)
plot_batch(data_roi3_1)

p2 = pipe_roi3(rel_roi_start=[0.1], rel_roi_end=[0.4], roi_axes=(1,))
data_roi3_2 = run(p2)
plot_batch(data_roi3_2)
../../../_images/examples_general_data_loading_numpy_reader_23_0.png
../../../_images/examples_general_data_loading_numpy_reader_23_1.png

Out of bounds access

Additionally, the user can choose what happens when a ROI falls outside of the bounds of the array. By default, out-of-bounds access results in an error, but it can be controlled by the out_of_bounds_policy argument.

Let us see an example of a pipeline configured to pad with zeros when the ROI falls out of bounds:

[12]:
@pipeline_def(batch_size=batch_size, num_threads=3, device_id=0)
def pipe_roi_oob():
    data = fn.readers.numpy(device='cpu', file_root=data_dir, files=files,
                            roi_start=(-150, -150), roi_end=(400, 400),
                            out_of_bounds_policy='pad', fill_value=0,
                            shard_id=0, num_shards=1)
    return data

data_oob = run(pipe_roi_oob())
plot_batch(data_oob)
../../../_images/examples_general_data_loading_numpy_reader_25_0.png

GPUDirect Storage Support

DALI Numpy Reader supports GPUDirect Storage (GDS) via libcufile. GDS enables a direct data path between storage and GPU memory and avoids extra copies through a bounce buffer in the CPU’s memory.

In order to enable GDS support in DALI, make sure GDS is installed. On Systems with CUDA 11.4 or newer, GDS is already installed as part of the CUDA 11.4 SDK. For older CUDA releases, please install GDS separately (follow the link above for instructions).

Once GDS is installed, it can be used by simply switching the device of the reader to "gpu". Note that if GDS is not available, you will likely see a CUDA Driver API error when trying to execute the pipeline.

[13]:
@pipeline_def(batch_size=batch_size, num_threads=3, device_id=0)
def pipe_gds():
    data = fn.readers.numpy(device='gpu', file_root=data_dir, files=files)
    return data

p = pipe_gds()
p.build()
pipe_out = p.run()

data_gds = pipe_out[0].as_cpu().as_array()  # as_cpu() to copy the data back to CPU memory
print(data_gds.shape)
plot_batch(data_gds)
(4, 512, 512)
../../../_images/examples_general_data_loading_numpy_reader_28_1.png