WEBINAR
AI inference can deliver faster, more accurate predictions to organizations of all sizes—but building a platform for production AI inference is hard.
Real-world use cases require different types of AI model architectures, and the models can contain hundreds of millions of parameters.
Models are trained in different frameworks (TensorFlow, PyTorch, XGBoost, Python, and others) and have different formats.
Applications have different requirements (real-time low latency, high-throughput batch, or streaming inputs), and then there are different execution environments (CPUs, GPUs, in the cloud, on premises, at the edge).
High-performance inference on specific hardware or in-framework is challenging because of competing constraints like latency, accuracy, throughput, and memory size that modern AI applications demand.
Join our webinar to explore how NVIDIA’s inference solution, including
open-source NVIDIA Triton™ Inference Server and NVIDIA® TensorRT™, delivers fast and scalable AI inference in production.
maincontent goes here
Content goes here
content goes here
Content goes here
Content goes here
Shankar is Senior Product Marketing manager in the data center GPU team at NVIDIA. He is responsible for GPU software infrastructure marketing to help IT and DevOps easily adopt and seamlessly integrate GPUs in their infrastructure. Before NVIDIA, he held engineering, operations and marketing positions in both small and large technology companies. He holds business and engineering degrees.
Jay Rodge is a product marketing manager for deep learning and inference products at NVIDIA driving launches and product marketing initiatives. Jay received his master’s degree in computer science from Illinois Tech, Chicago with a focus on computer vision and NLP. Before NVIDIA, Jay was an AI research intern at BMW Group solving problems using computer vision for BMW’s largest manufacturing plant.
Presenter 3 Bio
Presenter 4 Bio
Content here
Webinar: Description here
Date & Time: Wednesday, April 22, 2018