Skip to content

Commit 8f58439

Browse files
ekagra-ranjanMu Huai
authored andcommitted
[Model] Ignore rotary embed load for Cohere model (vllm-project#17319)
Signed-off-by: Mu Huai <[email protected]>
1 parent 38b7166 commit 8f58439

File tree

1 file changed

+4
-0
lines changed

1 file changed

+4
-0
lines changed

vllm/model_executor/models/commandr.py

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -418,6 +418,10 @@ def load_weights(self, weights: Iterable[Tuple[str,
418418
loaded_params: Set[str] = set()
419419
for name, loaded_weight in weights:
420420

421+
# Skip loading rotary embeddings since vLLM has its own
422+
if "rotary_emb.inv_freq" in name:
423+
continue
424+
421425
if (self.quant_config is not None and
422426
(scale_name := self.quant_config.get_cache_scale(name))):
423427
# Loading kv cache quantization scales

0 commit comments

Comments
 (0)