From e0c3cee17085914bbe505c159beeb8ae39bc37dd Mon Sep 17 00:00:00 2001 From: mobicham <37179323+mobicham@users.noreply.github.com> Date: Fri, 10 May 2024 19:29:35 +0200 Subject: [PATCH] hqq - fix weight check in check_quantized_param (#30748) * hqq - fix weight check in check_quantized_param * ruff format --- src/transformers/quantizers/quantizer_hqq.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/transformers/quantizers/quantizer_hqq.py b/src/transformers/quantizers/quantizer_hqq.py index 06949d059a..14be75369d 100755 --- a/src/transformers/quantizers/quantizer_hqq.py +++ b/src/transformers/quantizers/quantizer_hqq.py @@ -101,7 +101,7 @@ class HqqHfQuantizer(HfQuantizer): ) -> bool: module, tensor_name = get_module_from_name(model, param_name) - return isinstance(module, torch.nn.Linear) + return isinstance(module, torch.nn.Linear) and (tensor_name == "weight") def create_quantized_param( self,