Skip to content
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
7 changes: 4 additions & 3 deletions vllm/model_executor/layers/quantization/mxfp4.py
Original file line number Diff line number Diff line change
Expand Up @@ -194,6 +194,7 @@ def create_weights(self, layer: torch.nn.Module, num_experts: int,

self.intermediate_size = intermediate_size_per_partition_after_pad
self.hidden_size = hidden_size
self.intermediate_pad = (intermediate_size_per_partition_after_pad - intermediate_size_per_partition)
# Fused gate_up_proj (column parallel)
w13_weight = torch.nn.Parameter(
torch.zeros(
Expand Down Expand Up @@ -708,7 +709,7 @@ def apply(
sorted_expert_ids,
num_valid_ids,
top_k,
192, # n_pad_zeros
self.intermediate_pad // 64 * 64 * 2,
128, # k_pad_zeros
None, # sorted_weights
None,
Expand All @@ -725,7 +726,7 @@ def apply(
num_valid_ids,
top_k,
192, # n_pad_zeros
128, # k_pad_zeros
self.intermediate_pad // 128 * 128,
sorted_weights, # sorted_weights
None,
self.w2_scale_aiter_tensor,
Expand All @@ -750,4 +751,4 @@ def apply(
w1_precision=self.w13_precision_config,
w2_precision=self.w2_precision_config,
apply_router_weight_on_input=apply_router_weight_on_input,
)
)
Loading