To help with development and testing we have developed a light weight vLLM simulator. It does not truly run inference, but it does emulate responses to the HTTP REST endpoints of vLLM. Currently it supports partial OpenAI-compatible API:
- /v1/chat/completions
- /v1/completions
- /v1/models
In addition, a set of the vLLM HTTP endpoints are suppored as well. These include:
Endpoint | Description |
---|---|
/v1/load_lora_adapter | simulates the dynamic registration of a LoRA adapter |
/v1/unload_lora_adapter | simulates the dynamic unloading and unregistration of a LoRA adapter |
/metrics | exposes Prometheus metrics. See the table below for details |
/health | standard health check endpoint |
/ready | standard readiness endpoint |
In addition, it supports a subset of vLLM's Prometheus metrics. These metrics are exposed via the /metrics HTTP REST endpoint. Currently supported are the following metrics:
Metric | Description |
---|---|
vllm:gpu_cache_usage_perc | The fraction of KV-cache blocks currently in use (from 0 to 1). Currently this value will always be zero. |
vllm:lora_requests_info | Running stats on LoRA requests |
vllm:num_requests_running | Number of requests currently running on GPU |
vllm:num_requests_waiting | Prometheus metric for the number of queued requests |
The simulated inference has no connection with the model and LoRA adapters specified in the command line parameters or via the /v1/load_lora_adapter HTTP REST endpoint. The /v1/models endpoint returns simulated results based on those same command line parameters and those loaded via the /v1/load_lora_adapter HTTP REST endpoint.
The simulator supports two modes of operation:
echo
mode: the response contains the same text that was received in the request. For/v1/chat/completions
the last message for the role=user
is used.random
mode: the response is randomly chosen from a set of pre-defined sentences.
Timing of the response is defined by the time-to-first-token
and inter-token-latency
parameters. In case P/D is enabled for a request, kv-cache-transfer-latency
will be used instead of time-to-first-token
.
For a request with stream=true
: time-to-first-token
or kv-cache-transfer-latency
defines the delay before the first token is returned, inter-token-latency
defines the delay between subsequent tokens in the stream.
For a requst with stream=false
: the response is returned after delay of <time-to-first-token> + (<inter-token-latency> * (<number_of_output_tokens> - 1))
or <kv-cache-transfer-latency> + (<inter-token-latency> * (<number_of_output_tokens> - 1))
in P/D case
It can be run standalone or in a Pod for testing under packages such as Kind.
API responses contains a subset of the fields provided by the OpenAI API.
Click to show the structure of requests/responses
/v1/chat/completions
- request
- stream
- model
- messages
- role
- content
- response
- id
- created
- model
- choices
- index
- finish_reason
- message
- request
/v1/completions
- request
- stream
- model
- prompt
- max_tokens (for future usage)
- response
- id
- created
- model
- choices
- text
- request
/v1/models
- response
- object (list)
- data
- id
- object (model)
- created
- owned_by
- root
- parent
- response
For more details see the vLLM documentation
config
: the path to a yaml configuration file that can contain the simulator's command line parameters. If a parameter is defined in both the config file and the command line, the command line value overwrites the configuration file value. An example configuration file can be found atmanifests/config.yaml
port
: the port the simulator listents on, default is 8000model
: the currently 'loaded' model, mandatoryserved-model-name
: model names exposed by the API (a list of space-separated strings)lora-modules
: a list of LoRA adapters (a list of space-separated JSON strings): '{"name": "name", "path": "lora_path", "base_model_name": "id"}', optional, empty by defaultmax-loras
: maximum number of LoRAs in a single batch, optional, default is onemax-cpu-loras
: maximum number of LoRAs to store in CPU memory, optional, must be >= than max-loras, default is max-lorasmax-model-len
: model's context window, maximum number of tokens in a single request including input and output, optional, default is 1024max-num-seqs
: maximum number of sequences per iteration (maximum number of inference requests that could be processed at the same time), default is 5mode
: the simulator mode, optional, by defaultrandom
echo
: returns the same text that was sent in the requestrandom
: returns a sentence chosen at random from a set of pre-defined sentences
time-to-first-token
: the time to the first token (in milliseconds), optional, by default zerotime-to-first-token-std-dev
: standard deviation for time before the first token will be returned, in milliseconds, optional, default is 0, can't be more than 30% oftime-to-first-token
, will not cause the actual time to first token to differ by more than 70% fromtime-to-first-token
inter-token-latency
: the time to 'generate' each additional token (in milliseconds), optional, by default zerointer-token-latency-std-dev
: standard deviation for time between generated tokens, in milliseconds, optional, default is 0, can't be more than 30% ofinter-token-latency
, will not cause the actual inter token latency to differ by more than 70% frominter-token-latency
kv-cache-transfer-latency
: time for KV-cache transfer from a remote vLLM (in milliseconds), by default zero. Usually much shorter thantime-to-first-token
kv-cache-transfer-latency-std-dev
: standard deviation for time to "transfer" kv-cache from another vLLM instance in case P/D is activated, in milliseconds, optional, default is 0, can't be more than 30% ofkv-cache-transfer-latency
, will not cause the actual latency to differ by more than 70% fromkv-cache-transfer-latency
prefill-overhead
: constant overhead time for prefill (in milliseconds), optional, by default zero, used in calculating time to first token, this will be ignored iftime-to-first-token
is not0
prefill-time-per-token
: time taken to generate each token during prefill (in milliseconds), optional, by default zero, this will be ignored iftime-to-first-token
is not0
prefill-time-std-dev
: similar totime-to-first-token-std-dev
, but is applied on the final prefill time, which is calculated byprefill-overhead
,prefill-time-per-token
, and number of prompt tokens, this will be ignored iftime-to-first-token
is not0
kv-cache-transfer-time-per-token
: time taken to transfer cache for each token in case P/D is enabled (in milliseconds), optional, by default zero, this will be ignored ifkv-cache-transfer-latency
is not0
kv-cache-transfer-time-std-dev
: similar totime-to-first-token-std-dev
, but is applied on the final kv cache transfer time in case P/D is enabled (in milliseconds), which is calculated bykv-cache-transfer-time-per-token
and number of prompt tokens, this will be ignored ifkv-cache-transfer-latency
is not0
time-factor-under-load
: a multiplicative factor that affects the overall time taken for requests when parallelrequests are being processed. The value of this factor must be >= 1.0, with a default of 1.0. If this factor is 1.0, no extra time is added. When the factor is x (where x > 1.0) and there aremax-num-seqs
requests, the total time will be multiplied by x. The extra time then decreases multiplicatively to 1.0 when the number of requests is less than MaxNumSeqs.seed
: random seed for operations (if not set, current Unix time in nanoseconds is used)
max-tool-call-integer-param
: the maximum possible value of integer parameters in a tool call, optional, defaults to 100min-tool-call-integer-param
: the minimum possible value of integer parameters in a tool call, optional, defaults to 0max-tool-call-number-param
: the maximum possible value of number (float) parameters in a tool call, optional, defaults to 100min-tool-call-number-param
: the minimum possible value of number (float) parameters in a tool call, optional, defaults to 0max-tool-call-array-param-length
: the maximum possible length of array parameters in a tool call, optional, defaults to 5min-tool-call-array-param-length
: the minimum possible length of array parameters in a tool call, optional, defaults to 1tool-call-not-required-param-probability
: the probability to add a parameter, that is not required, in a tool call, optional, defaults to 50object-tool-call-not-required-field-probability
: the probability to add a field, that is not required, in an object in a tool call, optional, defaults to 50
enable-kvcache
: if true, the KV cache support will be enabled in the simulator. In this case, the KV cache will be simulated, and ZQM events will be published when a KV cache block is added or evicted.kv-cache-size
: the maximum number of token blocks in kv cacheblock-size
: token block size for contiguous chunks of tokens, possible values: 8,16,32,64,128tokenizers-cache-dir
: the directory for caching tokenizershash-seed
: seed for hash generation (if not set, is read from PYTHONHASHSEED environment variable)zmq-endpoint
: ZMQ address to publish eventszmq-max-connect-attempts
: the maximum number of ZMQ connection attempts, defaults to 0, maximum: 10event-batch-size
: the maximum number of kv-cache events to be sent together, defaults to 16
failure-injection-rate
: probability (0-100) of injecting failures, optional, default is 0failure-types
: list of specific failure types to inject (rate_limit, invalid_api_key, context_length, server_error, invalid_request, model_not_found), optional, if empty all types are used
-
fake-metrics
: represents a predefined set of metrics to be sent to Prometheus as a substitute for the real metrics. When specified, only these fake metrics will be reported — real metrics and fake metrics will never be reported together. The set should include values forrunning-requests
waiting-requests
kv-cache-usage
loras
- an array containing LoRA information objects, each with the fields:running
(a comma-separated list of LoRAs in use by running requests),waiting
(a comma-separated list of LoRAs to be used by waiting requests), andtimestamp
(seconds since Jan 1 1970, the timestamp of this metric).
Example: {"running-requests":10,"waiting-requests":30,"kv-cache-usage":0.4,"loras":[{"running":"lora4,lora2","waiting":"lora3","timestamp":1257894567},{"running":"lora4,lora3","waiting":"","timestamp":1257894569}]}
data-parallel-size
: number of ranks to run in Data Parallel deployment, from 1 to 8, default is 1. The ports will be assigned as follows: rank 0 will run on the configuredport
, rank 1 onport
+1, etc.
dataset-path
: Optional local file path to the SQLite database file used for generating responses from a dataset.- If not set, hardcoded preset responses will be used.
- If set but the file does not exist the
dataset-url
will be used to download the database to the path specified bydataset-path
. - If the file exists but is currently occupied by another process, responses will be randomly generated from preset text (the same behavior as if the path were not set).
- Responses are retrieved from the dataset by the hash of the conversation history, with a fallback to a random dataset response, constrained by the maximum output tokens and EoS token handling, if no matching history is found.
- Refer to llm-d converted ShareGPT for detailed information on the expected format of the SQLite database file.
dataset-url
: Optional URL for downloading the SQLite database file used for response generation.- This parameter is only used if the
dataset-path
is also set and the file does not exist at that path. - If the file needs to be downloaded, it will be saved to the location specified by
dataset-path
. - If the file already exists at the
dataset-path
, it will not be downloaded again - Example URL
https://huggingface.co/datasets/hf07397/inference-sim-datasets/resolve/91ffa7aafdfd6b3b1af228a517edc1e8f22cd274/huggingface/ShareGPT_Vicuna_unfiltered/conversations.sqlite3
- This parameter is only used if the
dataset-in-memory
: If true, the entire dataset will be loaded into memory for faster access. This may require significant memory depending on the size of the dataset. Default is false.
In addition, as we are using klog, the following parameters are available:
add_dir_header
: if true, adds the file directory to the header of the log messagesalsologtostderr
: log to standard error as well as files (no effect when -logtostderr=true)log_backtrace_at
: when logging hits line file:N, emit a stack trace (default :0)log_dir
: if non-empty, write log files in this directory (no effect when -logtostderr=true)log_file
: if non-empty, use this log file (no effect when -logtostderr=true)log_file_max_size
: defines the maximum size a log file can grow to (no effect when -logtostderr=true). Unit is megabytes. If the value is 0, the maximum file size is unlimited. (default 1800)logtostderr
: log to standard error instead of files (default true)one_output
: if true, only write logs to their native severity level (vs also writing to each lower severity level; no effect when -logtostderr=true)skip_headers
: if true, avoid header prefixes in the log messagesskip_log_headers
: if true, avoid headers when opening log files (no effect when -logtostderr=true)stderrthreshold
: logs at or above this threshold go to stderr when writing to files and stderr (no effect when -logtostderr=true or -alsologtostderr=true) (default 2)v
: number for the log level verbosityvmodule
: comma-separated list of pattern=N settings for file-filtered logging
POD_NAME
: the simulator pod name. If defined, the response will contain the HTTP headerx-inference-pod
with this valuePOD_NAMESPACE
: the simulator pod namespace. If defined, the response will contain the HTTP headerx-inference-namespace
with this value
max-running-requests
was replaced bymax-num-seqs
lora
was replaced bylora-modules
, which is now a list of JSON strings, e.g, '{"name": "name", "path": "lora_path", "base_model_name": "id"}'
To build a Docker image of the vLLM Simulator, run:
make image-build
Please note that the default image tag is ghcr.io/llm-d/llm-d-inference-sim:dev
.
The following environment variables can be used to change the image tag: REGISTRY
, SIM_TAG
, IMAGE_TAG_BASE
or IMG
.
Note: On macOS, use make image-build TARGETOS=linux
to pull the correct base image.
To run the vLLM Simulator image under Docker, run:
docker run --rm --publish 8000:8000 ghcr.io/llm-d/llm-d-inference-sim:dev --port 8000 --model "Qwen/Qwen2.5-1.5B-Instruct" --lora-modules '{"name":"tweet-summary-0"}' '{"name":"tweet-summary-1"}'
Note: To run the vLLM Simulator with the latest release version, in the above docker command replace dev
with the current release which can be found on GitHub.
Note: The above command exposes the simulator on port 8000, and serves the Qwen/Qwen2.5-1.5B-Instruct model.
To build the vLLM simulator to run locally as an executable, run:
make build
To run the vLLM simulator in a standalone test environment, run:
./bin/llm-d-inference-sim --model my_model --port 8000
To run the vLLM simulator in a Kubernetes cluster, run:
kubectl apply -f manifests/deployment.yaml
When testing locally with kind, build the docker image with make build-image
then load into the cluster:
kind load --name kind docker-image ghcr.io/llm-d/llm-d-inference-sim:dev
Update the deployment.yaml
file to use the dev tag.
To verify the deployment is available, run:
kubectl get deployment vllm-llama3-8b-instruct
kubectl get service vllm-llama3-8b-instruct-svc
Use kubectl port-forward
to expose the service on your local machine:
kubectl port-forward svc/vllm-llama3-8b-instruct-svc 8000:8000
Test the API with curl
curl -X POST http://localhost:8000/v1/chat/completions \
-H "Content-Type: application/json" \
-d '{
"model": "meta-llama/Llama-3.1-8B-Instruct",
"messages": [
{"role": "user", "content": "Hello!"}
]
}'