Skip to content

Commit f8e71da

Browse files
change default back to previous
Signed-off-by: Brian Dellabetta <bdellabe@redhat.com>
1 parent ac6f6d2 commit f8e71da

File tree

1 file changed

+3
-3
lines changed
  • src/llmcompressor/modifiers/awq

1 file changed

+3
-3
lines changed

src/llmcompressor/modifiers/awq/base.py

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -111,9 +111,9 @@ class AWQModifier(Modifier, QuantizationMixin):
111111
device. Defaults to None, so cached args are not offloaded. Consider setting
112112
to torch.device("cpu") if you are encountering OOM errors
113113
:param duo_scaling: whether to use duo scaling, which uses both input activations
114-
and weights to determine the scaling factor. Defaults to None
115-
If False, only activations are used.
114+
and weights to determine the scaling factor. Defaults to True
116115
If True, both activations and weights are used.
116+
If False, only activations are used.
117117
If None, half the grid search is performed with duo_scaling=False and the
118118
other half is performed with duo_scaling=True.
119119
:param n_grid: when performing the best scales grid search for each mapping,
@@ -129,7 +129,7 @@ class AWQModifier(Modifier, QuantizationMixin):
129129
sequential_targets: str | list[str] | None = None
130130
mappings: list[AWQMapping] | None = None
131131
offload_device: torch.device | None = None
132-
duo_scaling: bool | None = None
132+
duo_scaling: bool | None = True
133133
n_grid: int = 20
134134

135135
# Private vars set during validation

0 commit comments

Comments
 (0)