NVIDIA Triton Inference Server
Model serving with Triton Inference Server
Out of date
This guide contains outdated information pertaining to Kubeflow 1.0. This guide needs to be updated for Kubeflow 1.1.
Kubeflow currently doesn’t have a specific guide for NVIDIA Triton Inference Server. Note that Triton was previously known as the TensorRT Inference Server. See the NVIDIA documentation for instructions on running NVIDIA inference server on Kubernetes.
Last modified 03.08.2020: Added outdated banner to non-index docs unchanged in last 30d (#2072) (e56f3650)