|
7 | 7 | import ray
|
8 | 8 | from prometheus_client import REGISTRY
|
9 | 9 |
|
| 10 | +import vllm.envs as envs |
10 | 11 | from vllm import EngineArgs, LLMEngine
|
11 | 12 | from vllm.distributed import cleanup_dist_env_and_memory
|
12 | 13 | from vllm.engine.arg_utils import AsyncEngineArgs
|
@@ -141,8 +142,10 @@ def test_metric_set_tag_model_name(vllm_runner, model: str, dtype: str,
|
141 | 142 | stat_logger = vllm_model.model.llm_engine.stat_loggers['prometheus']
|
142 | 143 | metrics_tag_content = stat_logger.labels["model_name"]
|
143 | 144 |
|
| 145 | + if envs.VLLM_CI_USE_S3: |
| 146 | + model = f"{MODEL_WEIGHTS_S3_BUCKET}/{model}" |
144 | 147 | if served_model_name is None or served_model_name == []:
|
145 |
| - assert metrics_tag_content == f"{MODEL_WEIGHTS_S3_BUCKET}/{model}", ( |
| 148 | + assert metrics_tag_content == model, ( |
146 | 149 | f"Metrics tag model_name is wrong! expect: {model!r}\n"
|
147 | 150 | f"actual: {metrics_tag_content!r}")
|
148 | 151 | else:
|
@@ -215,8 +218,9 @@ def test_engine_log_metrics_regression(
|
215 | 218 | while engine.has_unfinished_requests():
|
216 | 219 | engine.step()
|
217 | 220 |
|
218 |
| - assert_metrics(f"{MODEL_WEIGHTS_S3_BUCKET}/{model}", engine, |
219 |
| - disable_log_stats, len(example_prompts)) |
| 221 | + if envs.VLLM_CI_USE_S3: |
| 222 | + model = f"{MODEL_WEIGHTS_S3_BUCKET}/{model}" |
| 223 | + assert_metrics(model, engine, disable_log_stats, len(example_prompts)) |
220 | 224 |
|
221 | 225 |
|
222 | 226 | @pytest.mark.parametrize("model", MODELS)
|
|
0 commit comments