You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I've noticed that the new convert_checkpoint.py scripts (such as the one for llama) have some quantization options built in, though some, like fp8, are missing. The readme for llama suggests to use quantize.py for fp8 post-training quantization instead.
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
-
I've noticed that the new convert_checkpoint.py scripts (such as the one for llama) have some quantization options built in, though some, like fp8, are missing. The readme for llama suggests to use quantize.py for fp8 post-training quantization instead.
Is there any reason these two are split?
Beta Was this translation helpful? Give feedback.
All reactions