NVIDIA Webinar
In this webinar series, we will do an in-depth presentation of Triton and its main components.
We also will demonstrate how to quick-start and use Triton on real word applications on-prem, on the cloud, or in a mixed environment.
In addition, we provide you with the scripts and code to jumpstart your Triton expertise.
TensorRT is an optimizing compiler of neural networks for inference.
NVIDIA TensorRT-based applications perform up to 36X faster than CPU-only platforms during inference, enabling you to optimize neural network models trained on all major frameworks, calibrate for lower precision with high accuracy, and deploy to hyperscale data centers, embedded platforms, or
automotive product platforms.
The NVIDIA Triton™ Inference Server is an open-source, high-performance inference-serving software that helps standardize model deployment and execution and delivers fast and scalable AI in production.
Triton supports any ML/DL models, frameworks, and hardware (CPU, GPU, servers, cloud, edge, embedded) and can work with any middleware (bare metal, virtualization, containers, microservices).
maincontent goes here
Content goes here
content goes here
Content goes here
Content goes here
Content here
Webinar: Description here
Date & Time: Wednesday, April 22, 2018