We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent baeec70 commit 1297cc8Copy full SHA for 1297cc8
vllm/model_executor/models/llama.py
@@ -512,7 +512,10 @@ def __init__(
512
quant_config=quant_config,
513
)
514
if config.tie_word_embeddings:
515
- self.lm_head = self.model.embed_tokens
+ # Reverted logic to fix build issues;
516
+ # this may introduce GGUF-related bugs.
517
+ # self.lm_head = self.model.embed_tokens
518
+ self.lm_head.weight = self.model.embed_tokens.weight
519
520
logit_scale = getattr(config, "logit_scale", 1.0)
521
self.logits_processor = LogitsProcessor(self.unpadded_vocab_size,
0 commit comments