We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent b06c4cf commit f326cd4Copy full SHA for f326cd4
examples/quantization_2of4_sparse_w4a16/llama7b_sparse_w4a16.py
@@ -54,8 +54,8 @@
54
warmup_ratio=warmup_ratio,
55
)
56
logger.info(
57
- "Note: llcompressor does not currently support running ",
58
- "compressed models in the marlin-24 format. The model ",
59
- "produced from this example can be run on vLLM with ",
60
- "dtype=torch.float16",
+ "llmcompressor does not currently support running compressed models in the marlin24 format." # noqa
+)
+logger.info(
+ "The model produced from this example can be run on vLLM with dtype=torch.float16"
61
0 commit comments