-
Notifications
You must be signed in to change notification settings - Fork 163
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Clarification Needed on Utillization of Tokenization in the Fine-Tuning Module || InternLM-XComposer2d5 #431
Comments
|
Thanks for the response @yuhangzang I have a few more questions. I am failing to use multi-GPUs for training. Case 1:I tried the Lora fintuning on the sample dataset on the single A100. Lora-finetuning works on a single 80GB A100 machine.
This works properly. Case 2:I have 8 X L4 machines. (23 GBs X 8 = 184 GBs of GPU memory)
What changes do I need to make for this to work? |
Our code is tested in 8 A100 GPUs (80GB). You may set a small value of |
Hello Fellow Developers,
I am working on implementing the evaluation code in the current fine-tuning module and noticed something regarding the tokenizer.
While the tokenizer is passed into the make_supervised_data_module function, it doesn't seem to be utilized in the DataCollatorForSupervisedDataset.
Since DataCollatorForSupervisedDataset serves as the custom data collator, if the tokenizer isn’t used there, what is being employed for tokenization? This brings up the concern of whether the fine-tuning script is functioning as intended.
Could you please clarify this?
> Also, when are you planning to release the evaluation code?
Thanks in Advance.
The text was updated successfully, but these errors were encountered: