You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
fix(quantization): Skip weight initialization for quantized models
This commit addresses the RuntimeError encountered when loading llmcompressor W8A8 quantized models, where `torch.nn.init.normal_()` is called on `int8` tensors during weight initialization.
Fixes#39366
Signed-off-by: Mauricio Harley <mauricioharley@gmail.com>
0 commit comments