High Performance Inferencing with TensorRT
by Rich Brueckner from High-Performance Computing News Analysis | insideHPC on (#3E26D)
Chris Gottbrath from NVIDIA gave this talk at SC17 in Denver. "This talk will introduce the TensorRT Programmable Inference Accelerator which enables high throughput and low latency inference on clusters with NVIDIA V100, P100, P4 or P40 GPUs. TensorRT is both an optimizer and runtime - users provide a trained neural network and can easily creating highly efficient inference engines that can be incorporated into larger applications and services."
The post High Performance Inferencing with TensorRT appeared first on insideHPC.