Skip to content

Commit

Permalink
Fix (example/llm): disable embedded lookup quantization
Browse files Browse the repository at this point in the history
  • Loading branch information
nickfraser committed Aug 19, 2024
1 parent b46a001 commit 8b18edc
Showing 1 changed file with 1 addition and 3 deletions.
4 changes: 1 addition & 3 deletions src/brevitas_examples/llm/main.py
Original file line number Diff line number Diff line change
Expand Up @@ -145,8 +145,6 @@
help='Group size for per_group input quantization. Default: 64.')
parser.add_argument(
'--quantize-input-zero-point', action='store_true', help='Quantize input zero-point.')
parser.add_argument(
'--quantize-embedding', action='store_true', help='Quantize first nn.Embedding layer.')
parser.add_argument(
'--quantize-last-layer', action='store_true', help='Quantize last nn.Linear layer.')
parser.add_argument('--gptq', action='store_true', help='Apply GPTQ.')
Expand Down Expand Up @@ -403,7 +401,7 @@ def main():
dtype=dtype,
device=device,
input_quant_format=args.input_quant_format,
quantize_embedding=args.quantize_embedding)
quantize_embedding=False)
if not args.quantize_last_layer:
name_blacklist += ["lm_head"]
model = layerwise_quantize(
Expand Down

0 comments on commit 8b18edc

Please sign in to comment.