Taming Throughput-Latency Tradeoff in LLM Inference with Sarathi-Serve
Each LLM serving request goes through two phases. The first is prefill which processes the entire input prompt and produces the first output token and the second is decode which generates the rest of output tokens, one-at-a-time. Prefill iterations have high latency but saturate GPU compute due to p...
Gespeichert in:
Hauptverfasser: | , , , , , , , |
---|---|
Format: | Artikel |
Sprache: | eng |
Schlagworte: | |
Online-Zugang: | Volltext bestellen |
Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
Zusammenfassung: | Each LLM serving request goes through two phases. The first is prefill which
processes the entire input prompt and produces the first output token and the
second is decode which generates the rest of output tokens, one-at-a-time.
Prefill iterations have high latency but saturate GPU compute due to parallel
processing of the input prompt. In contrast, decode iterations have low latency
but also low compute utilization because a decode iteration processes only a
single token per request. This makes batching highly effective for decodes and
consequently for overall throughput. However, batching multiple requests leads
to an interleaving of prefill and decode iterations which makes it challenging
to achieve both high throughput and low latency.
We introduce an efficient LLM inference scheduler, Sarathi-Serve, to address
this throughput-latency tradeoff. Sarathi-Serve introduces chunked-prefills
which splits a prefill request into near equal sized chunks and creates
stall-free schedules that adds new requests in a batch without pausing ongoing
decodes. Stall-free scheduling unlocks the opportunity to improve throughput
with large batch sizes while minimizing the effect of batching on latency.
Furthermore, uniform batches in Sarathi-Serve ameliorate the imbalance between
iterations resulting in minimal pipeline bubbles.
Our techniques yield significant improvements in inference performance across
models and hardware under tail latency constraints. For Mistral-7B on single
A100 GPUs, we achieve 2.6x higher serving capacity and up to 3.7x higher
serving capacity for the Yi-34B model on two A100 GPUs as compared to vLLM.
When used with pipeline parallelism on Falcon-180B, Sarathi-Serve provides up
to 5.6x gain in the end-to-end serving capacity. The source code for
Sarathi-Serve is available at https://github.com/microsoft/sarathi-serve. |
---|---|
DOI: | 10.48550/arxiv.2403.02310 |