MuxFlow: Efficient and Safe GPU Sharing in Large-Scale Production Deep Learning Clusters
Large-scale GPU clusters are widely-used to speed up both latency-critical (online) and best-effort (offline) deep learning (DL) workloads. However, most DL clusters either dedicate each GPU to one workload or share workloads in time, leading to very low GPU resource utilization. We present MuxFlow,...
Gespeichert in:
Hauptverfasser: | , , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Large-scale GPU clusters are widely-used to speed up both latency-critical
(online) and best-effort (offline) deep learning (DL) workloads. However, most
DL clusters either dedicate each GPU to one workload or share workloads in
time, leading to very low GPU resource utilization. We present MuxFlow, the
first production cluster system that supports efficient and safe space-sharing
for DL workloads. NVIDIA MPS provides an opportunity to share multiple
workloads in space on widely-deployed NVIDIA GPUs, but it cannot guarantee the
performance and safety of online workloads. MuxFlow introduces a two-level
protection mechanism for memory and computation to guarantee the performance of
online workloads. Based on our practical error analysis, we design a mixed
error-handling mechanism to guarantee the safety of online workloads. MuxFlow
further proposes dynamic streaming multiprocessor (SM) allocation and
matching-based scheduling to improve the efficiency of offline workloads.
MuxFlow has been deployed at CompanyX's clusters with more than 20,000 GPUs.
The deployment results indicate that MuxFlow substantially improves the GPU
utilization from 26$\%$ to 76$\%$, SM activity from 16$\%$ to 33$\%$, and GPU
memory from 42$\%$ to 48$\%$. |
---|---|
DOI: | 10.48550/arxiv.2303.13803 |