File tree Expand file tree Collapse file tree 1 file changed +3
-3
lines changed
src/llmcompressor/modifiers/awq Expand file tree Collapse file tree 1 file changed +3
-3
lines changed Original file line number Diff line number Diff line change @@ -111,9 +111,9 @@ class AWQModifier(Modifier, QuantizationMixin):
111111 device. Defaults to None, so cached args are not offloaded. Consider setting
112112 to torch.device("cpu") if you are encountering OOM errors
113113 :param duo_scaling: whether to use duo scaling, which uses both input activations
114- and weights to determine the scaling factor. Defaults to None
115- If False, only activations are used.
114+ and weights to determine the scaling factor. Defaults to True
116115 If True, both activations and weights are used.
116+ If False, only activations are used.
117117 If None, half the grid search is performed with duo_scaling=False and the
118118 other half is performed with duo_scaling=True.
119119 :param n_grid: when performing the best scales grid search for each mapping,
@@ -129,7 +129,7 @@ class AWQModifier(Modifier, QuantizationMixin):
129129 sequential_targets : str | list [str ] | None = None
130130 mappings : list [AWQMapping ] | None = None
131131 offload_device : torch .device | None = None
132- duo_scaling : bool | None = None
132+ duo_scaling : bool | None = True
133133 n_grid : int = 20
134134
135135 # Private vars set during validation
You can’t perform that action at this time.
0 commit comments