Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Distributed fine tuning of LLMs #49

Open
Tracked by #38
Shreyanand opened this issue May 25, 2023 · 0 comments
Open
Tracked by #38

Distributed fine tuning of LLMs #49

Shreyanand opened this issue May 25, 2023 · 0 comments
Assignees
Labels
help wanted Extra attention is needed

Comments

@Shreyanand
Copy link
Member

Shreyanand commented May 25, 2023

The finetuning notebook uses 1 GPU and LoRA technique to fine tune a T5 model with 3B parameters. The task to be completed in this issue is to fine tune the same model (or 7B version of the model) on multiple GPU nodes. Use Instascale and Codeflare to schedule the training job and retrieve the finetuned model. Create a notebook that demos this.

@Shreyanand Shreyanand changed the title How can we fine tune small models with limited resources, and can we use ray? Distributed fine tuning of LLMs Jul 13, 2023
@Shreyanand Shreyanand self-assigned this Jul 18, 2023
@Shreyanand Shreyanand added the help wanted Extra attention is needed label Mar 22, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
help wanted Extra attention is needed
Projects
None yet
Development

No branches or pull requests

1 participant