Article 3E26D High Performance Inferencing with TensorRT

High Performance Inferencing with TensorRT

by
Rich Brueckner
from High-Performance Computing News Analysis | insideHPC on (#3E26D)
chris-150x150.jpeg

Chris Gottbrath from NVIDIA gave this talk at SC17 in Denver. "This talk will introduce the TensorRT Programmable Inference Accelerator which enables high throughput and low latency inference on clusters with NVIDIA V100, P100, P4 or P40 GPUs. TensorRT is both an optimizer and runtime - users provide a trained neural network and can easily creating highly efficient inference engines that can be incorporated into larger applications and services."

The post High Performance Inferencing with TensorRT appeared first on insideHPC.

External Content
Source RSS or Atom Feed
Feed Location http://insidehpc.com/feed/
Feed Title High-Performance Computing News Analysis | insideHPC
Feed Link https://insidehpc.com/
Reply 0 comments