Skip to content

Commit 6405f2b

Browse files
committed
minor
Signed-off-by: realAsma <akuriparambi@nvidia.com>
1 parent 8e56285 commit 6405f2b

File tree

1 file changed

+2
-0
lines changed

1 file changed

+2
-0
lines changed

modelopt/torch/quantization/algorithms.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -961,6 +961,8 @@ def run_search_with_stats(self, max_weight_size, verbose=False):
961961
AutoQuantize uses Linear Programming Solver to find the optimal quantization configuration which
962962
minimizes the sum of per-layer auto_quantize scores while meeting the specified constraint.
963963
"""
964+
# TODO: Do this only for rank 0 in the respective pipeline group
965+
964966
for lower_bound in [None, 0.99, 0.90]:
965967
# The LP solver for auto_quantize sometimes fails to find a solution if a lower bound is not
966968
# specified. I dont know why this happens.

0 commit comments

Comments
 (0)