v0.0.2
Pre-releaseTurbo-Alignment v0.0.2 Release Notes π
What's New π
-
π€ Add SimPO and ORPO Trainers
-
π§ Fix SLiC-HF Trainer
- We've resolved issues that prevented the use of the SLiC-HF trainer.
Documentation and Tutorials π
- π Add Multimodal Tutorial and Docs
- Enhance your skills with our new tutorials and documentation designed for multimodal training pipelines.
Improvements and Fixes π οΈ
-
π Fix Embeddings Initialization Strategy for GPT-NeoX
- Token embeddings initialization now supports both LLama and GPT-NeoX architectures.
-
π Fix Multiple Logits in Chat Generator
- The chat generator has been updated to handle multiple logits.
-
π Fix Type of ID for Answer Message in
rewards.py
- Answer message IDs are now correctly typed as strings to match pydantic model expectations.
-
π Add
stop_strings
into Chat Generator and Fix Multiple EOS Problem- Now, the chat generator manages multiple End-Of-Sequence tokens and incorporates using strings as EOS tokens.
-
π
.to('cpu')
Per Batch into Chat Generator- We've optimized memory usage for batched chat generation by moving batches individually to the CPU.
-
β Fix
model.config.use_cache
When Not Using Grad Checkpointing- The KV-Cache now behaves predictably, ensuring proper functionality even in absence of gradient checkpointing.
Full Changelog π
You can view the complete list of changes in this release by visiting the changelog on GitHub: Full Changelog.
New Contributors π
We hope you enjoy these updates! As always, we welcome your feedback and contributions to make Turbo-Alignment even better.
Don't forget to star βοΈ the repo if you find it useful, and watch it for future updates.
Thank you for supporting Turbo-Alignment! π
Need help or have questions? Reach out to us on GitHub Issues, and weβll be there to support you.
Installation
Upgrade to the latest Turbo-Alignment release with:
pip install turbo-alignment==0.0.2
β Turbo-Alignment Team π€«