[CS] Yihua Cheng Dissertation Defense/Feb 14, 2025
via cs
cs at mailman.cs.uchicago.edu
Wed Feb 12 12:01:23 CST 2025
This is an announcement of Yihua Cheng's Dissertation Defense.
===============================================
Candidate: Yihua Cheng
Date: Friday, February 14, 2025
Time: 3 pm CST
Remote Location: https://uchicago.zoom.us/j/3150893650?pwd=RFgxNEE2MUQ0QzlsVXF3Ym94bDQ2Zz09
Location: JCL 298
Title: A Scalable Approach to Distributed Large Language Model Inference
Abstract: As the use of large language models (LLMs) expands rapidly, so does the intensity and scale of the workloads required to query LLMs. Thus, the requirements for serving LLMs evolve beyond single-instance deployment to large-scale distributed deployment. As most of today's LLM serving system optimizations focus only on speeding up a single serving instance, key techniques for distributed LLM deployments are still missing.
The key contribution of this dissertation is the design and implementation of an efficient system for distributed LLM serving engine deployment. Our thesis is that by decoupling the inference states (KV caches) from the LLM serving engine, the performance of distributed LLM inference can be substantially improved. Unlike prior work, which treats the LLM serving engine as a black box and builds global orchestrators serving engines, our approach uses a separate module to transfer, store, and share the KV caches across different serving engines.
To prove this thesis, this dissertation provides a suite of techniques to address the following fundamental challenges. First, we need an efficient way to offload the KV caches from serving engines and insert them back. Second, we need a scalable solution that is compatible with heterogeneous hardware and infrastructure stacks.
Our key insight is that many KV caches can be reused across serving engines in the large-scale distributed setup. With the reused KV cache, the LLM can skip computationally intensive prefills and start generating outputs immediately. We have developed algorithms to properly transmit, store, and reuse the KV caches in the distributed serving system. We have shown that our solution can substantially reduce the computational cost on real-world workloads, as well as provide better production-ready guarantees, including faster auto-scaling and higher service availability. We have also open-sourced our solution at {https://github.com/vllm-project/production-stack}.
Advisor: Junchen Jiang
Committee Members: Junchen Jiang, Kexin Pei, Hui Zhang
More information about the cs
mailing list