Inference Server Release 18.07 Beta
The NVIDIA container image of the Inference Server, release 18.07, is available as a beta release.
Contents of the Inference Server
This container image contains the Inference Server executable in /opt/inference_server.
The container also includes the following:
- Ubuntu 16.04
- NVIDIA CUDA 9.0.176 (see Errata section and 2.1) including CUDA® Basic Linear Algebra Subroutines library™ (cuBLAS) 9.0.425
- NVIDIA CUDA® Deep Neural Network library™ (cuDNN) 7.1.4
- NCCL 2.2.13 (optimized for NVLink™ )
- TensorRT 4.0.1
Key Features and Enhancements
This Inference Server release includes the following key features and
enhancements.
- The Inference Server container image version 18.07 is based on NVIDIA Inference Server 0.4.0 beta, TensorFlow 1.8.0, and Caffe2 0.8.1.
- Latest version of CUDA® Basic Linear Algebra Subroutines library™ (cuBLAS) 9.0.425.
- Support added for TensorFlow SavedModel format.
- Support added for gRPC in addition to existing HTTP REST.
- Ubuntu 16.04 with June 2018 updates
Known Issues
This is a beta release of the Inference Server. All features are expected to be available, however, some aspects of functionality and performance will likely be limited compared to a non-beta release.