🦄 Distributed Inference on Kubernetes with DRA and MIG
-
Updated
Sep 21, 2024 - Shell
🦄 Distributed Inference on Kubernetes with DRA and MIG
Unofficial example on creating Multi-Instance GPU (MIG) instances with NVIDIA Management Library (NVML) Go bindings.
Unofficial minimal instructions for managing NVIDIA Multi-Instance GPU (MIG) in a docker container
Tools for training, analysis and execution of an optimized task scheduling RL agent on GPUs with Multi-Instance GPU (MIG).
A Kubernetes serving manager for machine learning inference system enabled with NVIDIA MIG/MPS GPU-Sharing support
C++ implementation of FAR scheduler, an approach for optimized GPU co-execution of tasks using the hardware-level partitioning MIG (Multi-Instance GPU).
A Kubernetes serving manager for machine learning inference system enabled with NVIDIA MIG/MPS GPU-Sharing support
Add a description, image, and links to the multi-instance-gpu topic page so that developers can more easily learn about it.
To associate your repository with the multi-instance-gpu topic, visit your repo's landing page and select "manage topics."