AI Inference
Offer
Back to top

Introduction

Date: Wednesday, July 26, 2023
Time: 16:00 - 17:00 CEST
Duration: 1 hour


In this webinar series, we will do an in-depth presentation of Triton and its main components. We also will demonstrate how to quick-start and use Triton on real word applications on-prem, on the cloud, or in a mixed environment. In addition, we provide you with the scripts and code to jumpstart your Triton expertise.

TensorRT is an optimizing compiler of neural networks for inference. NVIDIA TensorRT-based applications perform up to 36X faster than CPU-only platforms during inference, enabling you to optimize neural network models trained on all major frameworks, calibrate for lower precision with high accuracy, and deploy to hyperscale data centers, embedded platforms, or automotive product platforms.

The NVIDIA Triton™ Inference Server is an open-source, high-performance inference-serving software that helps standardize model deployment and execution and delivers fast and scalable AI in production. Triton supports any ML/DL models, frameworks, and hardware (CPU, GPU, servers, cloud, edge, embedded) and can work with any middleware (bare metal, virtualization, containers, microservices).

In this webinar you will learn:
  • Overview of TensorRT and its integration with Pytorch and TensorFlow
  • Rapid Prototyping and Testing of ML Models With PyTriton
  • A complete, reproducible demo of how one can take a Pytorch model, compile with TensorRT and inference in production in Triton.
  • Review and Real Enterprise Customer Cases
Learn more about Triton Inference Server, Triton docs,

Webinar Registration

THANK YOU FOR REGISTERING FOR THE WEBINAR



You will receive an email with instructions on how to join the webinar shortly.

Main Content

maincontent goes here

Content

Content goes here

Content

content goes here

main image description

Content

Content goes here

Content

DGX Station Datasheet

Get a quick low-down and technical specs for the DGX Station.
DGX Station Whitepaper

Dive deeper into the DGX Station and learn more about the architecture, NVLink, frameworks, tools and more.
DGX Station Whitepaper

Dive deeper into the DGX Station and learn more about the architecture, NVLink, frameworks, tools and more.
DGX Station Whitepaper

Dive deeper into the DGX Station and learn more about the architecture, NVLink, frameworks, tools and more.

Content

Content goes here

Speakers

Dmitry Mironov
Solutions Architect, NVIDIA
Dmitry Mironov is an AI Solutions Architect at NVIDIA. He helps customers use the GPUs efficiently, and helps speed up various pipelines in CV, NLP, Conversational AI and Data Science. Before NVIDIA, Dmitry served as a CTO and Co-founder of a startup. He had been integrating Computer Vision into gold mining, transportation, energy and other industries.
Dora Csillag
Senior Solution Architect – Deep Learning and Inference, NVIDIA
Dora Csillag is a Senior Solutions Architect at NVIDIA. She helps customers building and deploying AI solutions, while focusing on Inference in the field of Computer Vision and Generative AI. She was previously applied researcher in Video Analytics, for Road Safety, Border Controls, and other industries.
Presenter 3 Name
Presenter 3 Title
Presenter 3 Bio
Presenter 4 Name
Job Title 4
Presenter 4 Bio


Main CTA for lightbox form use class="lightbox" CTA

Content Title

Content here

Register

Webinar: Description here

Date & Time: Wednesday, April 22, 2018