Skip to content

Commit 10b5b0e

Browse files
committed
Rename
1 parent 9f12890 commit 10b5b0e

File tree

15 files changed

+40
-36
lines changed

15 files changed

+40
-36
lines changed

vllm/model_executor/models/blip2.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@
2020
from .blip import (BlipVisionModel, dummy_image_for_blip,
2121
get_max_blip_image_tokens)
2222
from .interfaces import SupportsMultiModal, SupportsPP
23-
from .utils import (WeightLoader, init_vllm_registered_model,
23+
from .utils import (AutoWeightsLoader, init_vllm_registered_model,
2424
merge_multimodal_embeddings)
2525

2626
# We use this internally as placeholders since there is no image token
@@ -686,5 +686,5 @@ def sample(
686686
return self.language_model.sample(logits, sampling_metadata)
687687

688688
def load_weights(self, weights: Iterable[Tuple[str, torch.Tensor]]):
689-
loader = WeightLoader(self)
689+
loader = AutoWeightsLoader(self)
690690
loader.load_weights(weights)

vllm/model_executor/models/fuyu.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -41,7 +41,7 @@
4141
SequenceData)
4242

4343
from .interfaces import SupportsMultiModal, SupportsPP
44-
from .utils import WeightLoader, flatten_bn, merge_multimodal_embeddings
44+
from .utils import AutoWeightsLoader, flatten_bn, merge_multimodal_embeddings
4545

4646
# Cannot find the following 2 numbers from hf config.
4747
_IMAGE_TOKEN_ID = 71011
@@ -347,5 +347,5 @@ def sample(
347347
return next_tokens
348348

349349
def load_weights(self, weights: Iterable[Tuple[str, torch.Tensor]]):
350-
loader = WeightLoader(self)
350+
loader = AutoWeightsLoader(self)
351351
loader.load_weights(weights)

vllm/model_executor/models/gemma2.py

Lines changed: 5 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -40,7 +40,7 @@
4040
from vllm.sequence import IntermediateTensors
4141

4242
from .interfaces import SupportsLoRA, SupportsPP
43-
from .utils import (WeightLoader, is_pp_missing_parameter,
43+
from .utils import (AutoWeightsLoader, is_pp_missing_parameter,
4444
make_empty_intermediate_tensors_factory, make_layers)
4545

4646
logger = init_logger(__name__)
@@ -434,7 +434,8 @@ def sample(
434434
return next_tokens
435435

436436
def load_weights(self, weights: Iterable[Tuple[str, torch.Tensor]]):
437-
loader = WeightLoader(self,
438-
allow_missing_prefixes=None if
439-
self.config.tie_word_embeddings else ["lm_head"])
437+
loader = AutoWeightsLoader(
438+
self,
439+
allow_missing_prefixes=None
440+
if self.config.tie_word_embeddings else ["lm_head"])
440441
loader.load_weights(weights)

vllm/model_executor/models/internvl.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -31,7 +31,7 @@
3131
from .clip import (dummy_image_for_clip, dummy_seq_data_for_clip,
3232
get_clip_num_patches)
3333
from .interfaces import SupportsMultiModal, SupportsPP
34-
from .utils import (WeightLoader, flatten_bn, init_vllm_registered_model,
34+
from .utils import (AutoWeightsLoader, flatten_bn, init_vllm_registered_model,
3535
merge_multimodal_embeddings)
3636

3737
IMG_START = '<img>'
@@ -608,5 +608,5 @@ def sample(
608608
return self.language_model.sample(logits, sampling_metadata)
609609

610610
def load_weights(self, weights: Iterable[Tuple[str, torch.Tensor]]):
611-
loader = WeightLoader(self)
611+
loader = AutoWeightsLoader(self)
612612
loader.load_weights(weights)

vllm/model_executor/models/llama.py

Lines changed: 5 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -51,7 +51,7 @@
5151
from vllm.utils import is_hip
5252

5353
from .interfaces import SupportsLoRA, SupportsPP
54-
from .utils import (PPMissingLayer, WeightLoader, is_pp_missing_parameter,
54+
from .utils import (AutoWeightsLoader, PPMissingLayer, is_pp_missing_parameter,
5555
make_empty_intermediate_tensors_factory, make_layers)
5656

5757

@@ -550,9 +550,10 @@ def sample(self, logits: torch.Tensor,
550550
return next_tokens
551551

552552
def load_weights(self, weights: Iterable[Tuple[str, torch.Tensor]]):
553-
loader = WeightLoader(self,
554-
allow_missing_prefixes=None if
555-
self.config.tie_word_embeddings else ["lm_head"])
553+
loader = AutoWeightsLoader(
554+
self,
555+
allow_missing_prefixes=None
556+
if self.config.tie_word_embeddings else ["lm_head"])
556557
loader.load_weights(
557558
self.maybe_remap_mistral(name, loaded_weight)
558559
for name, loaded_weight in weights)

vllm/model_executor/models/llava.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -25,7 +25,7 @@
2525
from .siglip import (SiglipVisionModel, dummy_image_for_siglip,
2626
dummy_seq_data_for_siglip, get_max_siglip_image_tokens,
2727
input_processor_for_siglip)
28-
from .utils import (WeightLoader, flatten_bn, init_vllm_registered_model,
28+
from .utils import (AutoWeightsLoader, flatten_bn, init_vllm_registered_model,
2929
merge_multimodal_embeddings)
3030

3131

@@ -405,5 +405,5 @@ def sample(
405405
return self.language_model.sample(logits, sampling_metadata)
406406

407407
def load_weights(self, weights: Iterable[Tuple[str, torch.Tensor]]):
408-
loader = WeightLoader(self)
408+
loader = AutoWeightsLoader(self)
409409
loader.load_weights(weights)

vllm/model_executor/models/llava_next.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,7 @@
2828
from .siglip import (SiglipVisionModel, dummy_image_for_siglip,
2929
dummy_seq_data_for_siglip, get_siglip_image_feature_size,
3030
get_siglip_patch_grid_length, input_processor_for_siglip)
31-
from .utils import (WeightLoader, flatten_bn, init_vllm_registered_model,
31+
from .utils import (AutoWeightsLoader, flatten_bn, init_vllm_registered_model,
3232
merge_multimodal_embeddings)
3333

3434
# Result in the max possible feature size (2x2 grid of 336x336px tiles)
@@ -641,5 +641,5 @@ def sample(
641641
return self.language_model.sample(logits, sampling_metadata)
642642

643643
def load_weights(self, weights: Iterable[Tuple[str, torch.Tensor]]):
644-
loader = WeightLoader(self)
644+
loader = AutoWeightsLoader(self)
645645
loader.load_weights(weights)

vllm/model_executor/models/llava_next_video.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@
2727
from .interfaces import SupportsMultiModal, SupportsPP
2828
from .siglip import (SiglipVisionModel, dummy_image_for_siglip,
2929
dummy_seq_data_for_siglip)
30-
from .utils import (WeightLoader, init_vllm_registered_model,
30+
from .utils import (AutoWeightsLoader, init_vllm_registered_model,
3131
merge_multimodal_embeddings)
3232

3333
# For profile run
@@ -457,5 +457,5 @@ def sample(
457457
return self.language_model.sample(logits, sampling_metadata)
458458

459459
def load_weights(self, weights: Iterable[Tuple[str, torch.Tensor]]):
460-
loader = WeightLoader(self)
460+
loader = AutoWeightsLoader(self)
461461
loader.load_weights(weights)

vllm/model_executor/models/llava_onevision.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -34,7 +34,7 @@
3434
from .siglip import (SiglipVisionModel, dummy_seq_data_for_siglip,
3535
dummy_video_for_siglip, get_siglip_image_feature_size,
3636
get_siglip_patch_grid_length, input_processor_for_siglip)
37-
from .utils import (WeightLoader, flatten_bn, init_vllm_registered_model,
37+
from .utils import (AutoWeightsLoader, flatten_bn, init_vllm_registered_model,
3838
merge_multimodal_embeddings)
3939

4040
logger = init_logger(__name__)
@@ -871,5 +871,5 @@ def sample(
871871
return self.language_model.sample(logits, sampling_metadata)
872872

873873
def load_weights(self, weights: Iterable[Tuple[str, torch.Tensor]]):
874-
loader = WeightLoader(self)
874+
loader = AutoWeightsLoader(self)
875875
loader.load_weights(weights)

vllm/model_executor/models/paligemma.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@
2020
from .interfaces import SupportsMultiModal, SupportsPP
2121
from .siglip import (SiglipVisionModel, dummy_image_for_siglip,
2222
dummy_seq_data_for_siglip, get_max_siglip_image_tokens)
23-
from .utils import WeightLoader, merge_multimodal_embeddings
23+
from .utils import AutoWeightsLoader, merge_multimodal_embeddings
2424

2525
logger = init_logger(__name__)
2626

@@ -291,5 +291,5 @@ def sample(
291291
return self.language_model.sample(logits, sampling_metadata)
292292

293293
def load_weights(self, weights: Iterable[Tuple[str, torch.Tensor]]):
294-
loader = WeightLoader(self)
294+
loader = AutoWeightsLoader(self)
295295
loader.load_weights(weights)

0 commit comments

Comments
 (0)