Skip to content
This repository has been archived by the owner on Oct 25, 2024. It is now read-only.

Commit

Permalink
Update modeling_auto.py
Browse files Browse the repository at this point in the history
Signed-off-by: Wang, Chang <[email protected]>
  • Loading branch information
changwangss authored Aug 5, 2024
1 parent 7914ccc commit 1de8b17
Showing 1 changed file with 0 additions and 2 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -161,9 +161,7 @@ def build_woq_model(model, quantization_config):
from neural_compressor.adaptor.torch_utils.util import set_module
weight_dtype = quantization_config.weight_dtype
for n, m in model.named_modules():
print(n)
if n in quantization_config.llm_int8_skip_modules:
# import pdb;pdb.set_trace();
continue
if isinstance(m, torch.nn.Linear):
zp = getattr(
Expand Down

0 comments on commit 1de8b17

Please sign in to comment.