-
Notifications
You must be signed in to change notification settings - Fork 488
Description
What you would like to be added?
Inspired by this research paper Vidur: A Large-Scale Simulation Framework For LLM Inference
Optimizing the deployment of Large language models (LLMs) is expensive today since it requires experimentally running an application workload against an LLM implementation while exploring large configuration space formed by system knobs such as parallelization strategies, batching techniques, and scheduling policies.
we present Vidur-Search, a configuration search tool that helps optimize LLM deployment. Vidur-Search uses Vidur
to automatically identify the most cost-effective deployment configuration that meets application performance
constraints. For example, Vidur-Search finds the best deployment configuration for LLaMA2-70B in one hour on
a CPU machine, in contrast to a deployment-based exploration which would require 42K GPU hours – costing
218K dollars.
Why is this needed?
Not sure if it is in the scope of katib, but glad to raise an issue here.
Love this feature?
Give it a 👍 We prioritize the features with most 👍