Replies: 1 comment 1 reply
-
For this section : lora_target_modules:
I see you have two v_projs above, is that just a typo? One should be k_proj? Tom |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I'm attempting to fine-tune Llama-3-70B on my personal data with qlora r: 512 and qlora_alpha: 256. It's currently using 8 H100 80GB SXM5, and I'm wondering if this is normal to consume this much compute for qlora or if I've made a mistake somewhere. Is it possible to optimize the pipeline with the same sequence length and lora r & alpha?
Here is my full config
Beta Was this translation helpful? Give feedback.
All reactions