From 89aa77d50afd0d82f1eb3091ccf489d20ea61981 Mon Sep 17 00:00:00 2001 From: wenhuach21 Date: Tue, 22 Oct 2024 15:40:16 +0800 Subject: [PATCH] fix line too long issue --- auto_round/quantizer.py | 6 ++++-- 1 file changed, 4 insertions(+), 2 deletions(-) diff --git a/auto_round/quantizer.py b/auto_round/quantizer.py index 61833544..411e1ab1 100644 --- a/auto_round/quantizer.py +++ b/auto_round/quantizer.py @@ -243,7 +243,8 @@ def __init__(self, orig_layer, enable_minmax_tuning=True, enable_norm_bias_tunin self.params = {} if self.act_quant: - self.act_quant_func, self.act_data_type = get_quant_func(self.orig_layer.data_type, self.act_bits, self.act_sym) + self.act_quant_func, self.act_data_type = get_quant_func(self.orig_layer.data_type, self.act_bits, + self.act_sym) self.q_scale_thresh = 1e-5 @@ -421,7 +422,8 @@ def __init__(self, orig_layer, enable_minmax_tuning=True, enable_norm_bias_tunin self.act_quant = self.act_bits <= 8 self.weight_quant_func, self.data_type = get_quant_func(self.orig_layer.data_type, self.bits, self.sym) if self.act_quant: - self.act_quant_func, self.act_data_type = get_quant_func(self.orig_layer.data_type, self.act_bits, self.act_sym) + self.act_quant_func, self.act_data_type = get_quant_func(self.orig_layer.data_type, self.act_bits, + self.act_sym) self.q_scale_thresh = 1e-5 weight_dtype = torch.float32