|
41 | 41 | from nncf.quantization.advanced_parameters import AdvancedCompressionParameters |
42 | 42 | from nncf.quantization.advanced_parameters import AdvancedQuantizationParameters |
43 | 43 | from nncf.quantization.advanced_parameters import convert_to_dict_recursively |
| 44 | +from nncf.quantization.advanced_parameters import is_weight_compression_needed |
44 | 45 | from nncf.quantization.algorithms.accuracy_control.algorithm import QuantizationAccuracyRestorer |
45 | 46 | from nncf.quantization.algorithms.accuracy_control.algorithm import calculate_accuracy_drop |
46 | 47 | from nncf.quantization.algorithms.accuracy_control.evaluator import Evaluator |
@@ -121,7 +122,7 @@ def _extract_all_subgraphs(model: ov.Model, current_id: str) -> None: |
121 | 122 | quantization_algorithm, model, graphs, main_model_graph_id, calibration_dataset, subset_size, 1 |
122 | 123 | ) |
123 | 124 |
|
124 | | - if advanced_parameters is None or advanced_parameters.is_weight_compression_needed(): |
| 125 | + if is_weight_compression_needed(advanced_parameters): |
125 | 126 | compress_quantize_weights_transformation(quantized_model) |
126 | 127 |
|
127 | 128 | dump_parameters( |
@@ -168,7 +169,7 @@ def native_quantize_impl( |
168 | 169 | warning_model_no_batchwise_support(graph, advanced_parameters, model_type, OPERATIONS_OUTPUT_HAS_NO_BATCH_AXIS) |
169 | 170 | quantized_model = quantization_algorithm.apply(model, graph, dataset=calibration_dataset) |
170 | 171 |
|
171 | | - if advanced_parameters is None or advanced_parameters.is_weight_compression_needed(): |
| 172 | + if is_weight_compression_needed(advanced_parameters): |
172 | 173 | compress_quantize_weights_transformation(quantized_model) |
173 | 174 |
|
174 | 175 | dump_parameters( |
@@ -214,7 +215,7 @@ def quantize_with_accuracy_control_impl( |
214 | 215 | else: |
215 | 216 | copied_parameters = deepcopy(advanced_quantization_parameters) |
216 | 217 |
|
217 | | - compress_weights = copied_parameters.is_weight_compression_needed() |
| 218 | + compress_weights = is_weight_compression_needed(copied_parameters) |
218 | 219 | copied_parameters.backend_params[AdvancedQuantizationParameters.COMPRESS_WEIGHTS] = False |
219 | 220 |
|
220 | 221 | quantized_model = quantize_impl( |
|
0 commit comments