We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 3e692fe commit 81ce2ecCopy full SHA for 81ce2ec
src/llmcompressor/transformers/compression/sparsity_config.py
@@ -183,7 +183,8 @@ def is_sparse24_bitmask_supported(
183
if not is_model_quantized(model):
184
logger.warning(
185
"Compressed Sparse-only 2:4 models are not supported in vLLM<=0.7.0, "
186
- "consider saving with disable_sparse_compression=True"
+ "consider saving with `disable_sparse_compression` set, "
187
+ "`model.save_pretrained(..., disable_sparse_compression=True)`"
188
)
189
return True
190
0 commit comments