Deploying the optimal model

In this section, we cover the steps to download your chosen model as a Docker image or ONNX file, deploy your TensorRT model to the Triton Inference Server and run inference through the server .

Last updated