Skip to content

Commit

Permalink
Revert "IPEXQuantLinear supports group_size=-1"
Browse files Browse the repository at this point in the history
This reverts commit fba1efb.
  • Loading branch information
ZX-ModelCloud committed Dec 13, 2024
1 parent fba1efb commit 1305c2e
Showing 1 changed file with 1 addition and 7 deletions.
8 changes: 1 addition & 7 deletions gptqmodel/nn_modules/qlinear/ipex.py
Original file line number Diff line number Diff line change
Expand Up @@ -51,7 +51,7 @@ def convert_dtype_torch2str(dtype):

class IPEXQuantLinear(BaseQuantLinear):
SUPPORTS_BITS = [4]
SUPPORTS_GROUP_SIZE = [-1, 16, 32, 64, 128]
SUPPORTS_GROUP_SIZE = [16, 32, 64, 128]
SUPPORTS_DESC_ACT = [True, False]
SUPPORTS_SYM = [True, False]
SUPPORTS_SHARDS = True
Expand Down Expand Up @@ -81,12 +81,6 @@ def __init__(
):
super().__init__(bits=bits, group_size=group_size, sym=sym, desc_act=desc_act, infeatures=infeatures, outfeatures=outfeatures, **kwargs)

# Normalize group_size
if group_size != -1:
group_size = group_size
else:
group_size = infeatures

if weight_dtype is None:
weight_dtype = torch.float16 if is_torch_support_xpu() else torch.bfloat16

Expand Down

0 comments on commit 1305c2e

Please sign in to comment.