Skip to content
This repository was archived by the owner on Oct 25, 2024. It is now read-only.

Commit 5e5e17c

Browse files
authored
Update WOQ AutoRoundConfig parameter (#1568)
1 parent 0e13607 commit 5e5e17c

File tree

3 files changed

+5
-5
lines changed

3 files changed

+5
-5
lines changed

examples/huggingface/pytorch/text-generation/quantization/run_generation_cpu_woq.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -154,7 +154,7 @@
154154
help="minmax learning rate, if None,it will beset to be the same with lr",
155155
)
156156
parser.add_argument(
157-
"--enable_quanted_input",
157+
"--disable_quanted_input",
158158
action="store_true",
159159
help="whether to use the output of quantized block to tune the next block",
160160
)
@@ -286,7 +286,7 @@
286286
calib_len=args.calib_len,
287287
lr=args.lr,
288288
minmax_lr=args.minmax_lr,
289-
enable_quanted_input=args.enable_quanted_input,
289+
disable_quanted_input=args.disable_quanted_input,
290290
use_ipex=args.use_ipex,
291291
)
292292
else:

intel_extension_for_transformers/transformers/llm/quantization/utils.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -527,7 +527,7 @@ def default_calib_func(model):
527527
"seqlen": config.calib_len,
528528
"iters": config.iters,
529529
"scale_dtype": config.scale_dtype,
530-
"enable_quanted_input": config.enable_quanted_input,
530+
"enable_quanted_input": not config.disable_quanted_input,
531531
"lr": config.lr,
532532
"minmax_lr": config.minmax_lr,
533533
}

intel_extension_for_transformers/transformers/utils/config.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1056,7 +1056,7 @@ def __init__(
10561056
sym: bool = False,
10571057
lr: float = None,
10581058
minmax_lr: float = None,
1059-
enable_quanted_input: bool = True,
1059+
disable_quanted_input: bool = False,
10601060
nsamples: int = 512,
10611061
iters: int = 200,
10621062
use_ggml: bool = False,
@@ -1083,7 +1083,7 @@ def __init__(
10831083
self.group_size = group_size
10841084
self.lr = lr
10851085
self.minmax_lr = minmax_lr
1086-
self.enable_quanted_input = enable_quanted_input
1086+
self.disable_quanted_input = disable_quanted_input
10871087
self.iters = iters
10881088
self.llm_int8_skip_modules = (
10891089
llm_int8_skip_modules if llm_int8_skip_modules else []

0 commit comments

Comments
 (0)