Skip to content
Open
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
14 changes: 8 additions & 6 deletions auto_round/compressors/utils.py
Original file line number Diff line number Diff line change
Expand Up @@ -334,13 +334,15 @@ def normalize_item(item: Union[str, dict, "QuantizationScheme"], layer_name: str

# In AutoScheme with mixed gguf:q4_k_m, the super_group_size of gguf:q8_0 layer is None,
# which should not be filled by default q4km again
if fill_default_value:
tmp_scheme_keys = scheme_keys
else:
tmp_scheme_keys = extra_scheme_keys
for cfg in layer_config.values():
for key in tmp_scheme_keys:
cfg.setdefault(key, copy.deepcopy(default_dict.get(key)))
for key in scheme_keys:
if fill_default_value:
cfg.setdefault(key, copy.deepcopy(default_dict.get(key)))
else:
if key in extra_scheme_keys:
cfg.setdefault(key, copy.deepcopy(default_dict.get(key)))
else:
cfg.setdefault(key, None)

# 5. collect supported modules
embedding_types = (torch.nn.Embedding,)
Expand Down