A repository demonstrating deploying ML models using Triton + FastAPI + Docker
-
Updated
Mar 18, 2024 - Jupyter Notebook
A repository demonstrating deploying ML models using Triton + FastAPI + Docker
Model Deployment Operator is a prototype system designed to automate the deployment and management of Nvidia Triton models on Kubernetes
Nvidia Triton server gRPC client for Node
Triton inference server with Python backend and transformers
Hub with hands-on resources on building Computer Vision and Deep Learning solutions that process Video/Image data.
Add a description, image, and links to the nvidia-triton topic page so that developers can more easily learn about it.
To associate your repository with the nvidia-triton topic, visit your repo's landing page and select "manage topics."