LIVE WEBINAR
Maximizing GPU Utilization for Data Center Inference with NVIDIA TensorRT Inference Server on GKE with Kubeflow
Whether it’s performing object detection in images or video, recommending restaurants, or translating the spoken word, inference is the mechanism that allows applications to derive valuable information from trained AI models. But many inference solutions are one-off designs that lack the performance and flexibility to be seamlessly deployed in modern data center environments.
NVIDIA TensorRT Inference Server lets you leverage inference in your application without needing to reinvent the wheel. Delivered as a ready-to-deploy container from NGC, NVIDIA’s registry for GPU-accelerated software containers, and as an open source project, NVIDIA TensorRT Inference Server is a microservice that enables applications to use AI models in data center production. It maximizes GPU utilization, supports all popular AI frameworks and model types, and provides packaging and documentation for deployment using Kubeflow.
maincontent goes here
Content goes here
content goes here
Content goes here
Content goes here
Product Manager, NVIDIA
Tripti Singhal is a product manager on the NVIDIA Deep Learning Software team working on TensorRT Inference Server. She was a Deep Learning Solutions Architect at NVIDIA prior to moving to the product team and received her bachelor's degree in Computer Science at University of California, Santa Barbara.
Add Presenter 2's Name (John Smith)
Add Presenter 2's Title (ex: CMO, ABC Company)
Add Presenter 3's Name (John Smith)
Add Presenter 3's Title (ex: CMO, ABC Company)
Add Presenter 4's Name (John Smith)
Add Presenter 4's Title (ex: CMO, ABC Company)
Content here
Webinar: Description here
Date & Time: Wednesday, April 22, 2018