Category: Kubernetes, artificial-intelligence

This tutorial is the latest part of a series where we build an end-to-end stack to perform machine learning inference at the edge, built on thehttps://docs.rke2.io/ Kubernetes distribution. Refer to the that tutorial if you plan to run Nvidia Triton Inference Server on a GPU-powered host.

Since we plan to run the inference at the edge, we will utilize https://min.io/?utm_content=inline-mention as an S3-compatible object storage service. We will have a fully configured MinIO object storage service running on an RKE2 cluster by the end of this tutorial.

We are now ready to use MinIO running on RKE2 as the model registry for Nvidia Triton Inference Server.

Related Articles