Skip to content

llm-d/llm-d-inference-sim

Repository files navigation

Go Report Card License Join Slack

vLLM Simulator

To help with development and testing we have developed a light weight vLLM simulator. It does not truly run inference, but it does emulate responses to the HTTP REST endpoints of vLLM. Currently it supports partial OpenAI-compatible API:

  • /v1/chat/completions
  • /v1/completions
  • /v1/models

In addition, a set of the vLLM HTTP endpoints are suppored as well. These include:

Endpoint Description
/v1/load_lora_adapter simulates the dynamic registration of a LoRA adapter
/v1/unload_lora_adapter simulates the dynamic unloading and unregistration of a LoRA adapter
/metrics exposes Prometheus metrics. See the table below for details
/health standard health check endpoint
/ready standard readiness endpoint

In addition, it supports a subset of vLLM's Prometheus metrics. These metrics are exposed via the /metrics HTTP REST endpoint. Currently supported are the following metrics:

Metric Description
vllm:gpu_cache_usage_perc The fraction of KV-cache blocks currently in use (from 0 to 1). Currently this value will always be zero.
vllm:lora_requests_info Running stats on LoRA requests
vllm:num_requests_running Number of requests currently running on GPU
vllm:num_requests_waiting Prometheus metric for the number of queued requests

The simulated inference has no connection with the model and LoRA adapters specified in the command line parameters or via the /v1/load_lora_adapter HTTP REST endpoint. The /v1/models endpoint returns simulated results based on those same command line parameters and those loaded via the /v1/load_lora_adapter HTTP REST endpoint.

The simulator supports two modes of operation:

  • echo mode: the response contains the same text that was received in the request. For /v1/chat/completions the last message for the role=user is used.
  • random mode: the response is randomly chosen from a set of pre-defined sentences.

Timing of the response is defined by the time-to-first-token and inter-token-latency parameters. In case P/D is enabled for a request, kv-cache-transfer-latency will be used instead of time-to-first-token.

For a request with stream=true: time-to-first-token or kv-cache-transfer-latency defines the delay before the first token is returned, inter-token-latency defines the delay between subsequent tokens in the stream.

For a requst with stream=false: the response is returned after delay of <time-to-first-token> + (<inter-token-latency> * (<number_of_output_tokens> - 1)) or <kv-cache-transfer-latency> + (<inter-token-latency> * (<number_of_output_tokens> - 1)) in P/D case

It can be run standalone or in a Pod for testing under packages such as Kind.

Limitations

API responses contains a subset of the fields provided by the OpenAI API.

Click to show the structure of requests/responses
  • /v1/chat/completions
    • request
      • stream
      • model
      • messages
        • role
        • content
    • response
      • id
      • created
      • model
      • choices
        • index
        • finish_reason
        • message
  • /v1/completions
    • request
      • stream
      • model
      • prompt
      • max_tokens (for future usage)
    • response
      • id
      • created
      • model
      • choices
        • text
  • /v1/models
    • response
      • object (list)
      • data
        • id
        • object (model)
        • created
        • owned_by
        • root
        • parent

For more details see the vLLM documentation

Command line parameters

  • config: the path to a yaml configuration file that can contain the simulator's command line parameters. If a parameter is defined in both the config file and the command line, the command line value overwrites the configuration file value. An example configuration file can be found at manifests/config.yaml
  • port: the port the simulator listents on, default is 8000
  • model: the currently 'loaded' model, mandatory
  • served-model-name: model names exposed by the API (a list of space-separated strings)
  • lora-modules: a list of LoRA adapters (a list of space-separated JSON strings): '{"name": "name", "path": "lora_path", "base_model_name": "id"}', optional, empty by default
  • max-loras: maximum number of LoRAs in a single batch, optional, default is one
  • max-cpu-loras: maximum number of LoRAs to store in CPU memory, optional, must be >= than max-loras, default is max-loras
  • max-model-len: model's context window, maximum number of tokens in a single request including input and output, optional, default is 1024
  • max-num-seqs: maximum number of sequences per iteration (maximum number of inference requests that could be processed at the same time), default is 5
  • mode: the simulator mode, optional, by default random
    • echo: returns the same text that was sent in the request
    • random: returns a sentence chosen at random from a set of pre-defined sentences

  • time-to-first-token: the time to the first token (in milliseconds), optional, by default zero
  • time-to-first-token-std-dev: standard deviation for time before the first token will be returned, in milliseconds, optional, default is 0, can't be more than 30% of time-to-first-token, will not cause the actual time to first token to differ by more than 70% from time-to-first-token
  • inter-token-latency: the time to 'generate' each additional token (in milliseconds), optional, by default zero
  • inter-token-latency-std-dev: standard deviation for time between generated tokens, in milliseconds, optional, default is 0, can't be more than 30% of inter-token-latency, will not cause the actual inter token latency to differ by more than 70% from inter-token-latency
  • kv-cache-transfer-latency: time for KV-cache transfer from a remote vLLM (in milliseconds), by default zero. Usually much shorter than time-to-first-token
  • kv-cache-transfer-latency-std-dev: standard deviation for time to "transfer" kv-cache from another vLLM instance in case P/D is activated, in milliseconds, optional, default is 0, can't be more than 30% of kv-cache-transfer-latency, will not cause the actual latency to differ by more than 70% from kv-cache-transfer-latency

  • prefill-overhead: constant overhead time for prefill (in milliseconds), optional, by default zero, used in calculating time to first token, this will be ignored if time-to-first-token is not 0
  • prefill-time-per-token: time taken to generate each token during prefill (in milliseconds), optional, by default zero, this will be ignored if time-to-first-token is not 0
  • prefill-time-std-dev: similar to time-to-first-token-std-dev, but is applied on the final prefill time, which is calculated by prefill-overhead, prefill-time-per-token, and number of prompt tokens, this will be ignored if time-to-first-token is not 0
  • kv-cache-transfer-time-per-token: time taken to transfer cache for each token in case P/D is enabled (in milliseconds), optional, by default zero, this will be ignored if kv-cache-transfer-latency is not 0
  • kv-cache-transfer-time-std-dev: similar to time-to-first-token-std-dev, but is applied on the final kv cache transfer time in case P/D is enabled (in milliseconds), which is calculated by kv-cache-transfer-time-per-token and number of prompt tokens, this will be ignored if kv-cache-transfer-latency is not 0

  • time-factor-under-load: a multiplicative factor that affects the overall time taken for requests when parallelrequests are being processed. The value of this factor must be >= 1.0, with a default of 1.0. If this factor is 1.0, no extra time is added. When the factor is x (where x > 1.0) and there are max-num-seqs requests, the total time will be multiplied by x. The extra time then decreases multiplicatively to 1.0 when the number of requests is less than MaxNumSeqs.
  • seed: random seed for operations (if not set, current Unix time in nanoseconds is used)

  • max-tool-call-integer-param: the maximum possible value of integer parameters in a tool call, optional, defaults to 100
  • min-tool-call-integer-param: the minimum possible value of integer parameters in a tool call, optional, defaults to 0
  • max-tool-call-number-param: the maximum possible value of number (float) parameters in a tool call, optional, defaults to 100
  • min-tool-call-number-param: the minimum possible value of number (float) parameters in a tool call, optional, defaults to 0
  • max-tool-call-array-param-length: the maximum possible length of array parameters in a tool call, optional, defaults to 5
  • min-tool-call-array-param-length: the minimum possible length of array parameters in a tool call, optional, defaults to 1
  • tool-call-not-required-param-probability: the probability to add a parameter, that is not required, in a tool call, optional, defaults to 50
  • object-tool-call-not-required-field-probability: the probability to add a field, that is not required, in an object in a tool call, optional, defaults to 50

  • enable-kvcache: if true, the KV cache support will be enabled in the simulator. In this case, the KV cache will be simulated, and ZQM events will be published when a KV cache block is added or evicted.
  • kv-cache-size: the maximum number of token blocks in kv cache
  • block-size: token block size for contiguous chunks of tokens, possible values: 8,16,32,64,128
  • tokenizers-cache-dir: the directory for caching tokenizers
  • hash-seed: seed for hash generation (if not set, is read from PYTHONHASHSEED environment variable)
  • zmq-endpoint: ZMQ address to publish events
  • zmq-max-connect-attempts: the maximum number of ZMQ connection attempts, defaults to 0, maximum: 10
  • event-batch-size: the maximum number of kv-cache events to be sent together, defaults to 16

  • failure-injection-rate: probability (0-100) of injecting failures, optional, default is 0
  • failure-types: list of specific failure types to inject (rate_limit, invalid_api_key, context_length, server_error, invalid_request, model_not_found), optional, if empty all types are used

  • fake-metrics: represents a predefined set of metrics to be sent to Prometheus as a substitute for the real metrics. When specified, only these fake metrics will be reported — real metrics and fake metrics will never be reported together. The set should include values for

    • running-requests
    • waiting-requests
    • kv-cache-usage
    • loras - an array containing LoRA information objects, each with the fields: running (a comma-separated list of LoRAs in use by running requests), waiting (a comma-separated list of LoRAs to be used by waiting requests), and timestamp (seconds since Jan 1 1970, the timestamp of this metric).

    Example: {"running-requests":10,"waiting-requests":30,"kv-cache-usage":0.4,"loras":[{"running":"lora4,lora2","waiting":"lora3","timestamp":1257894567},{"running":"lora4,lora3","waiting":"","timestamp":1257894569}]}


  • data-parallel-size: number of ranks to run in Data Parallel deployment, from 1 to 8, default is 1. The ports will be assigned as follows: rank 0 will run on the configured port, rank 1 on port+1, etc.

  • dataset-path: Optional local file path to the SQLite database file used for generating responses from a dataset.
    • If not set, hardcoded preset responses will be used.
    • If set but the file does not exist the dataset-url will be used to download the database to the path specified by dataset-path.
    • If the file exists but is currently occupied by another process, responses will be randomly generated from preset text (the same behavior as if the path were not set).
    • Responses are retrieved from the dataset by the hash of the conversation history, with a fallback to a random dataset response, constrained by the maximum output tokens and EoS token handling, if no matching history is found.
    • Refer to llm-d converted ShareGPT for detailed information on the expected format of the SQLite database file.
  • dataset-url: Optional URL for downloading the SQLite database file used for response generation.
    • This parameter is only used if the dataset-path is also set and the file does not exist at that path.
    • If the file needs to be downloaded, it will be saved to the location specified by dataset-path.
    • If the file already exists at the dataset-path, it will not be downloaded again
    • Example URL https://huggingface.co/datasets/hf07397/inference-sim-datasets/resolve/91ffa7aafdfd6b3b1af228a517edc1e8f22cd274/huggingface/ShareGPT_Vicuna_unfiltered/conversations.sqlite3
  • dataset-in-memory: If true, the entire dataset will be loaded into memory for faster access. This may require significant memory depending on the size of the dataset. Default is false.

In addition, as we are using klog, the following parameters are available:

  • add_dir_header: if true, adds the file directory to the header of the log messages
  • alsologtostderr: log to standard error as well as files (no effect when -logtostderr=true)
  • log_backtrace_at: when logging hits line file:N, emit a stack trace (default :0)
  • log_dir: if non-empty, write log files in this directory (no effect when -logtostderr=true)
  • log_file: if non-empty, use this log file (no effect when -logtostderr=true)
  • log_file_max_size: defines the maximum size a log file can grow to (no effect when -logtostderr=true). Unit is megabytes. If the value is 0, the maximum file size is unlimited. (default 1800)
  • logtostderr: log to standard error instead of files (default true)
  • one_output: if true, only write logs to their native severity level (vs also writing to each lower severity level; no effect when -logtostderr=true)
  • skip_headers: if true, avoid header prefixes in the log messages
  • skip_log_headers: if true, avoid headers when opening log files (no effect when -logtostderr=true)
  • stderrthreshold: logs at or above this threshold go to stderr when writing to files and stderr (no effect when -logtostderr=true or -alsologtostderr=true) (default 2)
  • v: number for the log level verbosity
  • vmodule: comma-separated list of pattern=N settings for file-filtered logging

Environment variables

  • POD_NAME: the simulator pod name. If defined, the response will contain the HTTP header x-inference-pod with this value
  • POD_NAMESPACE: the simulator pod namespace. If defined, the response will contain the HTTP header x-inference-namespace with this value

Migrating from releases prior to v0.2.0

  • max-running-requests was replaced by max-num-seqs
  • lora was replaced by lora-modules, which is now a list of JSON strings, e.g, '{"name": "name", "path": "lora_path", "base_model_name": "id"}'

Working with docker image

Building

To build a Docker image of the vLLM Simulator, run:

make image-build

Please note that the default image tag is ghcr.io/llm-d/llm-d-inference-sim:dev.
The following environment variables can be used to change the image tag: REGISTRY, SIM_TAG, IMAGE_TAG_BASE or IMG.

Note: On macOS, use make image-build TARGETOS=linux to pull the correct base image.

Running

To run the vLLM Simulator image under Docker, run:

docker run --rm --publish 8000:8000 ghcr.io/llm-d/llm-d-inference-sim:dev  --port 8000 --model "Qwen/Qwen2.5-1.5B-Instruct"  --lora-modules '{"name":"tweet-summary-0"}' '{"name":"tweet-summary-1"}'

Note: To run the vLLM Simulator with the latest release version, in the above docker command replace dev with the current release which can be found on GitHub.

Note: The above command exposes the simulator on port 8000, and serves the Qwen/Qwen2.5-1.5B-Instruct model.

Standalone testing

Building

To build the vLLM simulator to run locally as an executable, run:

make build

Running

To run the vLLM simulator in a standalone test environment, run:

./bin/llm-d-inference-sim --model my_model --port 8000

Kubernetes testing

To run the vLLM simulator in a Kubernetes cluster, run:

kubectl apply -f manifests/deployment.yaml

When testing locally with kind, build the docker image with make build-image then load into the cluster:

kind load --name kind docker-image ghcr.io/llm-d/llm-d-inference-sim:dev

Update the deployment.yaml file to use the dev tag.

To verify the deployment is available, run:

kubectl get deployment vllm-llama3-8b-instruct
kubectl get service vllm-llama3-8b-instruct-svc

Use kubectl port-forward to expose the service on your local machine:

kubectl port-forward svc/vllm-llama3-8b-instruct-svc 8000:8000

Test the API with curl

curl -X POST http://localhost:8000/v1/chat/completions \
  -H "Content-Type: application/json" \
  -d '{
    "model": "meta-llama/Llama-3.1-8B-Instruct",
    "messages": [
      {"role": "user", "content": "Hello!"}
    ]
  }'

About

A light weight vLLM simulator, for mocking out replicas.

Topics

Resources

License

Stars

Watchers

Forks

Packages

 
 
 

Languages