Release Notes
Nsight Deep Learning Designer Release Notes.
Release notes and system requirements.
Updates in 2024.1
Nsight Deep Learning Designer has been completely revamped. In this release, we’ve embraced ONNX to provide broader and more flexible supports for DL models. More importantly, we have adopted TensorRT, NVIDIA’s signature inference solution, to replace NvNeural as the underlying inference engine. Some notable changes in this release:
Nsight Deep Learning Designer is now a full-fledged ONNX model editor. ONNX models can be visualized, edited, and profiled.
We revamped how nodes are represented on the canvas.
A new layout algorithm to arrange nodes on the canvas in a more compact layout was added.
Added support for ONNX opset version 1 to 20, and ONNX Runtime Contrib opset version 1.
Added support for subgraphs and local functions in ONNX models.
Added support to extract subgraphs or a selection of nodes to a standalone ONNX model.
Added support to export a TensorRT engine from an ONNX model.
Added support for visualizing TensorRT engine graphs.
Added support for TensorRT and ONNX Runtime profiling for ONNX models. Profiling generates a report document with detailed inference performances of the model.
Added a new model validator based on Polygraphy to report any errors, warnings, or issues caused by the current model structure.
Added support to run custom tools on ONNX models from Nsight Deep Learning Designer.
Added batch modification actions for ONNX models: convert model to FP16, sanitize model, and batch tensor conversion.
Nsight Deep Learning Designer depends on the following components and will download some of them automatically when needed:
Component |
Version |
---|---|
CUDA |
12.7 |
cuDNN |
8.9.7 |
DirectML |
1.15.2 |
Nsight Systems |
2024.6.1 |
ONNX IR |
9 |
ONNX GraphSurgeon |
0.5.1 |
ONNX Opsets |
1 to 20 |
ONNX Runtime |
1.17.3 |
Polygraphy |
0.49.9 |
TensorRT |
10.7 |
WinPixEventRuntime |
1.0.240308001 |
As part of the transition to ONNX and TensorRT, we deprecated the following:
Editing and visualizing NvNeural models
Importing/exporting NvNeural models from/to PyTorch
NvNeural SDK
Analysis mode and channel inspector
Updates in 2022.2
NVIDIA Nsight Deep Learning Designer changes in version 2022.2:
We added support to launch the PyTorch exporter from a virtual environment (Conda or virtualenv).
We improved the overall performance of Channel Inspector by separating the visualization of per-layer weights from per-layer features and changing the way how we visualize NxCx1x1 weights.
We added an experimental feature that allows users to directly import existing PyTorch models into NVIDIA Nsight Deep Learning Designer without starting from scratch.
We switched to using cubic curves to represent layer links to reduce path-finding overhead.
We added support to visualize inference results of classification networks in the Analysis Mode.
We added support for custom padding (in addition to the current
same
andvalid
options) to the relevant layers.We fixed numerous bugs.
We have also decided to remove some uncommonly used operators as we modernize the inference library and unify our model exporters. The following layers are now deprecated and will be removed in a future product release:
Local response normalization (LRN) layer: Only
region
values ofwithin
are being removed. Normalizationacross
channels is still fully supported.Mono-four-stack layer: Replace with a custom layer.
Mono-to-RGB layer: Replace with a custom layer.
Network layer: Import the subnetwork as a template instead.
Output layer: Use of output layers for tensor slicing (the
width
,height
,channels
, andoffset
parameters) is deprecated. Use an explicit slice layer if these operations are required.
The following activations are now deprecated:
Leaky sigmoid activation: Replace with a custom layer.
Leaky tanh activation: Replace with a custom layer.
ReLU activation with
alpha
values other than zero is being removed. Normal ReLU is still fully supported. The element-wisemax
layer can replace these activations.
NvNeural changes in version 2022.2:
Changed the signature of
nvneural::XmlNetworkBuilder::createLayerObject
to receive the original serialized type used to select the layer object being instantiated. Custom classes deriving fromXmlNetworkBuilder
must be updated if they override this function.
Updates in 2022.1
NVIDIA Nsight Deep Learning Designer changes in version 2022.1:
Added support to save all tensors in the analysis mode.
Added support for using nested templates to construct hierarchical network graphs.
We significantly improved the performance of the type-checking process in the Editor.
Fixed a bug that prevented PyTorch exports on Linux from succeeding.
Removed clamping behavior from the Affine layer. It no longer restricts the values of its
scale
andoffset
parameters. Theoptions_on
parameter has been deprecated; users wishing to hide interactive controls for this layer during analysis should set the newinclude_ui
parameter to false.Fixed a bug that blocked FP16 inference when fusing 7x7 convolutions with batch normalizations.
NvNeural changes in version 2022.1:
Added a new analysis layer: Signal Injector.
Added a new Input (Constant) layer which supports direct embedding of scalar constants.
Optimized the performance of the BatchNorm layer.
Optimized the performance of the Upscale layer.
Added support for downscaling and fixed-size scaling to the Upscale layer.
The NvRTC wrapper in
nvneural::ICudaRuntimeCompiler
has been replaced with a stub when type-checking networks from the GUI. Plugins that rely on the ability to execute generated kernel code during initialization ornvneural::ILayer::reshape
should call NvRTC directly, but for performance reasons we do not recommend this approach.The
forward()
function in the exported PyTorch class now takes in keyword-only arguments. User should explicitly name the input paramters while calling the model/function.The
INetwork::inferenceSubgraph
method now applies queued reshape operations. Queued reshapes are not cleared upon failure and will continue to blockinference
andinferenceSubgraph
calls until they succeed.
Updates in 2021.2
NVIDIA Nsight Deep Learning Designer changes in version 2021.2:
The Channel Inspector can display summary and per-channel statistics about a layer’s output tensor: mean, minimum/maximum, standard deviation, and sparsity (percentage of tensor elements close to zero).
Output tensor shapes are now visible during editing.
ConverenceNG can now save network outputs as .npy files.
Users can now expand or collapse the parameters list of a layer glyph in the editor view.
In Channel Inspector, users can now toggle a checkbox to perform auto scale and shift of the channels.
Users can now save a template as a file that can be imported into another model.
We have added more data to network profiling reports:
Per-layer device memory footprints
Whole-network device memory footprint
Percentage view for inference timings
Layers’ distance from the nearest input, for sorting by network depth
Template-level inference timings
NvNeural changes in version 2021.2:
Plugin initialization has been refactored to reduce its reliance on translation-unit-scoped static initialization. The ExportPlugin framework now expects user plugin code to provide an implementation of the function
void nvneural::plugin::InitializePluginTypes()
. This function should call static ClassRegistry methods to make its export types visible to the client application.The SkipConcatenation optimization has been rewritten. Custom concatenation layers should implement
nvneural::IConcatenationLayer2
to participate in this optimization.We added two new analysis layers: Saliency Generator and Saliency Mix. The Saliency Generator layer converts their input tensors into a single-channel tensor with the same H and W as their inputs. The saliency mix layer simply overlays saliency information (output of a Saliency Generator) onto another input tensor.
Known Issues
Nsight Deep Learning Designer’s profiling features require access to GPU performance counters, which by default requires local administrator privileges. See the following page for details on how to configure your system to allow profiling without elevation. Profiling without the appropriate permissions will generate error messages of the form
ERR: xxx: Error 19 returned from Perfworks
.The values of operations/second calculated from this NVIDIA developer tools site and generated by using the data center monitoring tools are not calculated in the same way as the operations/second used for export control purposes and should not be relied upon to assess performance against the export control limits.
The screenshot feature in the Send Feedback dialog does not work on Linux systems using Wayland as a display manager.
The ONNX Runtime DirectML profiling provider does not support machines that use a software display adapter.
There is no support for COMPLEX64 and COMPLEX128 ONNX data types.
ONNX models’ inputs and outputs only support Tensor types.
Link annotations can overlap each other if the links are too close together.
Model validation fails when processing an ONNX model with an invalid SparseTensor.
TensorRT profiling does not support control-flow operators such as Loop and If.
Closing an unsaved local function document with the [x] button in the document tab will not allow the document to be closed if the user elects to save changes. Save the local function edits with the ‘Confirm Local Function Edits’ button or the ‘Ctrl-S’ keyboard shortcut.
Platform Support
Linux
Nsight Deep Learning Designer supports Linux x86_64 systems running Ubuntu 20.04 LTS or newer, with GLIBC version 2.29 or higher.
The Nsight Deep Learning Designer host application requires several packages to be installed to enable Qt. Please refer to the Qt for X11 Requirements. When executing Nsight Deep Learning Designer with missing dependencies, an error message with information on the missing packages is shown. Note that only one package will be shown at a time, even though multiple may be missing from your system. The following command installs needed packages for Nsight Deep Learning Designer on X11:
Ubuntu 20.04:
apt install libopengl0 libxcb-icccm4 libxcb-image0 libxcb-keysyms1 libxcb-render-util0 libxcb-xinerama0 libxcb-xkb1 libxkbcommon-x11-0 libxcb-cursor0
NVIDIA L4T
Nsight Deep Learning Designer supports NVIDIA L4T arm64 systems running Ubuntu 20.04 LTS or newer, with GLIBC version 2.29 or higher.
The Nsight Deep Learning Designer host application requires several packages to be installed to enable Qt. Please refer to the Qt for X11 Requirements. When executing Nsight Deep Learning Designer with missing dependencies, an error message with information on the missing packages is shown. Note that only one package will be shown at a time, even though multiple may be missing from your system. The following command installs needed packages for Nsight Deep Learning Designer on X11:
Ubuntu 20.04:
apt install libopengl0 libxcb-icccm4 libxcb-image0 libxcb-keysyms1 libxcb-render-util0 libxcb-xinerama0 libxcb-xkb1 libxkbcommon-x11-0 libxcb-cursor0
Windows
Nsight Deep Learning Designer supports Windows x86_64 systems running:
Windows 10: 20H1 or newer.
Windows 11: 21H2 or newer.
GPU Support
NVIDIA Nsight Deep Learning Designer requires an NVIDIA GPU to run:
Linux x86_64
GeForce GPUs: GeForce RTX 2000 series, RTX 3000 series, or RTX 4000 series.
Data Center GPUs: A100 or H100.
NVIDIA L4T arm64
Embedded Systems: Jetson Orin.
Windows x86_64
GeForce GPUs: GeForce RTX 2000 series, RTX 3000 series, or RTX 4000 series.
Recommended Display Driver
You must have a recent NVIDIA display driver installed on your system to run Nsight Deep Learning Designer. The following display drivers are recommended:
Windows: Release 560.00 or newer.
Linux: Release 560.00 or newer.
NVIDIA L4T: NVIDIA JetPack SDK 6.1. Note: When targeting the NVIDIA L4T platform, the user (local or remote) needs to be a member of the debug group in order to profile.
Notices
Notice
This document is provided for information purposes only and shall not be regarded as a warranty of a certain functionality, condition, or quality of a product. NVIDIA Corporation (“NVIDIA”) makes no representations or warranties, expressed or implied, as to the accuracy or completeness of the information contained in this document and assumes no responsibility for any errors contained herein. NVIDIA shall have no liability for the consequences or use of such information or for any infringement of patents or other rights of third parties that may result from its use. This document is not a commitment to develop, release, or deliver any Material (defined below), code, or functionality.
NVIDIA reserves the right to make corrections, modifications, enhancements, improvements, and any other changes to this document, at any time without notice.
Customer should obtain the latest relevant information before placing orders and should verify that such information is current and complete.
NVIDIA products are sold subject to the NVIDIA standard terms and conditions of sale supplied at the time of order acknowledgement, unless otherwise agreed in an individual sales agreement signed by authorized representatives of NVIDIA and customer (“Terms of Sale”). NVIDIA hereby expressly objects to applying any customer general terms and conditions with regards to the purchase of the NVIDIA product referenced in this document. No contractual obligations are formed either directly or indirectly by this document.
NVIDIA products are not designed, authorized, or warranted to be suitable for use in medical, military, aircraft, space, or life support equipment, nor in applications where failure or malfunction of the NVIDIA product can reasonably be expected to result in personal injury, death, or property or environmental damage. NVIDIA accepts no liability for inclusion and/or use of NVIDIA products in such equipment or applications and therefore such inclusion and/or use is at customer’s own risk.
NVIDIA makes no representation or warranty that products based on this document will be suitable for any specified use. Testing of all parameters of each product is not necessarily performed by NVIDIA. It is customer’s sole responsibility to evaluate and determine the applicability of any information contained in this document, ensure the product is suitable and fit for the application planned by customer, and perform the necessary testing for the application in order to avoid a default of the application or the product. Weaknesses in customer’s product designs may affect the quality and reliability of the NVIDIA product and may result in additional or different conditions and/or requirements beyond those contained in this document. NVIDIA accepts no liability related to any default, damage, costs, or problem which may be based on or attributable to: (i) the use of the NVIDIA product in any manner that is contrary to this document or (ii) customer product designs.
No license, either expressed or implied, is granted under any NVIDIA patent right, copyright, or other NVIDIA intellectual property right under this document. Information published by NVIDIA regarding third-party products or services does not constitute a license from NVIDIA to use such products or services or a warranty or endorsement thereof. Use of such information may require a license from a third party under the patents or other intellectual property rights of the third party, or a license from NVIDIA under the patents or other intellectual property rights of NVIDIA.
Reproduction of information in this document is permissible only if approved in advance by NVIDIA in writing, reproduced without alteration and in full compliance with all applicable export laws and regulations, and accompanied by all associated conditions, limitations, and notices.
THIS DOCUMENT AND ALL NVIDIA DESIGN SPECIFICATIONS, REFERENCE BOARDS, FILES, DRAWINGS, DIAGNOSTICS, LISTS, AND OTHER DOCUMENTS (TOGETHER AND SEPARATELY, “MATERIALS”) ARE BEING PROVIDED “AS IS.” NVIDIA MAKES NO WARRANTIES, EXPRESSED, IMPLIED, STATUTORY, OR OTHERWISE WITH RESPECT TO THE MATERIALS, AND EXPRESSLY DISCLAIMS ALL IMPLIED WARRANTIES OF NONINFRINGEMENT, MERCHANTABILITY, AND FITNESS FOR A PARTICULAR PURPOSE. TO THE EXTENT NOT PROHIBITED BY LAW, IN NO EVENT WILL NVIDIA BE LIABLE FOR ANY DAMAGES, INCLUDING WITHOUT LIMITATION ANY DIRECT, INDIRECT, SPECIAL, INCIDENTAL, PUNITIVE, OR CONSEQUENTIAL DAMAGES, HOWEVER CAUSED AND REGARDLESS OF THE THEORY OF LIABILITY, ARISING OUT OF ANY USE OF THIS DOCUMENT, EVEN IF NVIDIA HAS BEEN ADVISED OF THE POSSIBILITY OF SUCH DAMAGES. Notwithstanding any damages that customer might incur for any reason whatsoever, NVIDIA’s aggregate and cumulative liability towards customer for the products described herein shall be limited in accordance with the Terms of Sale for the product.
VESA DisplayPort
DisplayPort and DisplayPort Compliance Logo, DisplayPort Compliance Logo for Dual-mode Sources, and DisplayPort Compliance Logo for Active Cables are trademarks owned by the Video Electronics Standards Association in the United States and other countries.
HDMI
HDMI, the HDMI logo, and High-Definition Multimedia Interface are trademarks or registered trademarks of HDMI Licensing LLC.
OpenCL
OpenCL is a trademark of Apple Inc. used under license to the Khronos Group Inc.
Trademarks
NVIDIA and the NVIDIA logo are trademarks and/or registered trademarks of NVIDIA Corporation in the Unites States and other countries. Other company and product names may be trademarks of the respective companies with which they are associated.