-
-
Notifications
You must be signed in to change notification settings - Fork 898
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Various bugs with ORPO #2105
Comments
To improve the mapping speed (1.), .map() can be improved with the axolotl/src/axolotl/utils/data/rl.py Lines 60 to 75 in 724b660
to: data_set = data_set.map(
ds_transform_fn,
desc="Mapping RL Dataset",
num_proc=os.cpu_count() # or whatever value
) However this change requires to install |
Hey, thanks for pointing these out.
I'll create a PR to address these. Btw, this |
Please check that this issue hasn't been reported before.
Expected Behavior
Training a model using ORPO and found some strange behaviors.
Current behaviour
[2024-11-26 10:26:55,772] [INFO] [peft.tuners.tuners_utils.__init__:171] [PID:2760612] Already found a `peft_config` attribute in the model. This will lead to having multiple adapters in the model. Make sure to know what you are doing! max_steps is given, it will override any value given in num_train_epochs
(max_steps is empty in the config)
Could not estimate the number of tokens of the input, floating-point operations will not be computed
warmup_ratio
param doesn't work, I have to rely onwarmup_steps
Steps to reproduce
Config yaml
No response
Possible solution
No response
Which Operating Systems are you using?
Python Version
3.11
axolotl branch-commit
last
Acknowledgements
The text was updated successfully, but these errors were encountered: