Optimal Workload Placement on Multi-Instance GPUs
There is an urgent and pressing need to optimize usage of Graphical Processing Units (GPUs), which have arguably become one of the most expensive and sought after IT resources. To help with this goal, several of the current generation of GPUs support a partitioning feature, called Multi-Instance GPU...
Gespeichert in:
Hauptverfasser: | , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | There is an urgent and pressing need to optimize usage of Graphical
Processing Units (GPUs), which have arguably become one of the most expensive
and sought after IT resources. To help with this goal, several of the current
generation of GPUs support a partitioning feature, called Multi-Instance GPU
(MIG) to allow multiple workloads to share a GPU, albeit with some constraints.
In this paper we investigate how to optimize the placement of Large Language
Model (LLM)-based AI Inferencing workloads on GPUs. We first identify and
present several use cases that are encountered in practice that require
workloads to be efficiently placed or migrated to other GPUs to make room for
incoming workloads. The overarching goal is to use as few GPUs as possible and
to further minimize memory and compute wastage on GPUs that are utilized. We
have developed two approaches to address this problem: an optimization method
and a heuristic method. We benchmark these with two workload scheduling
heuristics for multiple use cases. Our results show up to 2.85x improvement in
the number of GPUs used and up to 70% reduction in GPU wastage over baseline
heuristics. We plan to enable the SRE community to leverage our proposed method
in production environments. |
---|---|
DOI: | 10.48550/arxiv.2409.06646 |