We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent 889230c commit c930a0fCopy full SHA for c930a0f
src/llmcompressor/modifiers/awq/base.py
@@ -113,7 +113,7 @@ class AWQModifier(Modifier, QuantizationMixin):
113
:param offload_device: offload cached args to this device, which reduces memory
114
requirements but requires more time to move data between cpu and execution
115
device. Defaults to None, so cached args are not offloaded. Consider setting
116
- to "cpu" if you are encountering OOM errors
+ to torch.device("cpu") if you are encountering OOM errors
117
:param max_chunk_memory: maximum memory to use for each chunk of input activations
118
:param duo_scaling: whether to use duo scaling, which uses both input activations
119
and weights to determine the scaling factor
0 commit comments