@@ -268,7 +268,7 @@ class ModelConfig:
268
268
It can be a branch name, a tag name, or a commit id. If unspecified, will
269
269
use the default version."""
270
270
rope_scaling : dict [str , Any ] = field (default_factory = dict )
271
- """RoPE scaling configuration in JSON format . For example,
271
+ """RoPE scaling configuration. For example,
272
272
`{"rope_type":"dynamic","factor":2.0}`."""
273
273
rope_theta : Optional [float ] = None
274
274
"""RoPE theta. Use with `rope_scaling`. In some cases, changing the RoPE
@@ -346,30 +346,28 @@ class ModelConfig:
346
346
(stored in `~/.huggingface`)."""
347
347
hf_overrides : HfOverrides = field (default_factory = dict )
348
348
"""If a dictionary, contains arguments to be forwarded to the Hugging Face
349
- config. If a callable, it is called to update the HuggingFace config. When
350
- specified via CLI, the argument must be a valid JSON string."""
349
+ config. If a callable, it is called to update the HuggingFace config."""
351
350
mm_processor_kwargs : Optional [dict [str , Any ]] = None
352
351
"""Arguments to be forwarded to the model's processor for multi-modal data,
353
352
e.g., image processor. Overrides for the multi-modal processor obtained
354
353
from `AutoProcessor.from_pretrained`. The available overrides depend on the
355
354
model that is being run. For example, for Phi-3-Vision: `{"num_crops": 4}`.
356
- When specified via CLI, the argument must be a valid JSON string. """
355
+ """
357
356
disable_mm_preprocessor_cache : bool = False
358
357
"""If `True`, disable caching of the multi-modal preprocessor/mapper (not
359
358
recommended)."""
360
359
override_neuron_config : dict [str , Any ] = field (default_factory = dict )
361
360
"""Initialize non-default neuron config or override default neuron config
362
361
that are specific to Neuron devices, this argument will be used to
363
362
configure the neuron config that can not be gathered from the vllm
364
- arguments. e.g. `{"cast_logits_dtype": "bloat16"}`. When specified via CLI,
365
- the argument must be a valid JSON string."""
363
+ arguments. e.g. `{"cast_logits_dtype": "bloat16"}`."""
366
364
pooler_config : Optional ["PoolerConfig" ] = field (init = False )
367
365
"""Pooler config which controls the behaviour of output pooling in pooling
368
366
models."""
369
367
override_pooler_config : Optional [Union [dict , "PoolerConfig" ]] = None
370
368
"""Initialize non-default pooling config or override default pooling config
371
369
for the pooling model. e.g. `{"pooling_type": "mean", "normalize": false}`.
372
- When specified via CLI, the argument must be a valid JSON string. """
370
+ """
373
371
logits_processor_pattern : Optional [str ] = None
374
372
"""Optional regex pattern specifying valid logits processor qualified names
375
373
that can be passed with the `logits_processors` extra completion argument.
@@ -385,8 +383,7 @@ class ModelConfig:
385
383
"""Overrides or sets generation config. e.g. `{"temperature": 0.5}`. If
386
384
used with `--generation-config auto`, the override parameters will be
387
385
merged with the default config from the model. If used with
388
- `--generation-config vllm`, only the override parameters are used.
389
- When specified via CLI, the argument must be a valid JSON string."""
386
+ `--generation-config vllm`, only the override parameters are used."""
390
387
enable_sleep_mode : bool = False
391
388
"""Enable sleep mode for the engine (only cuda platform is supported)."""
392
389
model_impl : Union [str , ModelImpl ] = ModelImpl .AUTO .value
@@ -1556,8 +1553,7 @@ class LoadConfig:
1556
1553
cache directory of Hugging Face."""
1557
1554
model_loader_extra_config : dict = field (default_factory = dict )
1558
1555
"""Extra config for model loader. This will be passed to the model loader
1559
- corresponding to the chosen load_format. This should be a JSON string that
1560
- will be parsed into a dictionary."""
1556
+ corresponding to the chosen load_format."""
1561
1557
ignore_patterns : Optional [Union [list [str ], str ]] = None
1562
1558
"""The list of patterns to ignore when loading the model. Default to
1563
1559
"original/**/*" to avoid repeated loading of llama's checkpoints."""
@@ -2826,7 +2822,6 @@ class MultiModalConfig:
2826
2822
"limit_mm_per_prompt" )
2827
2823
"""
2828
2824
The maximum number of input items allowed per prompt for each modality.
2829
- This should be a JSON string that will be parsed into a dictionary.
2830
2825
Defaults to 1 (V0) or 999 (V1) for each modality.
2831
2826
2832
2827
For example, to allow up to 16 images and 2 videos per prompt:
0 commit comments