Load Phi 3 small on Nvidia Tesla V100 - Flash Attention #1956
Replies: 1 comment 4 replies
-
Hey, this seems to be unrelated to PEFT, right? So please open a discussion for transformers instead, which looks like the right place.
ONNX is not for training, so it's not an option here. |
Beta Was this translation helpful? Give feedback.
4 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi,
I would like to inquire about the possibility of uploading and fine tuning a Phi 3 8k small. When I load the model, I get an error about missing Flash attention. If I want to install the given package, I get this error :
But I have the required version of pytorch and CUDA (torch 2.3.1 and cuda 12.1)
Is it because I am using a Tesla V100 graphics card? Is there any way to load the model also with this graphics card?
I found this in the documentation for the Phi 3 mini on Huggingface:
Does this also apply to the Phi3 Small 8k?? Beacause when I try to load it, the error occurs
Or should I try the ONNX version or it is just for inference?
Thank you.
Beta Was this translation helpful? Give feedback.
All reactions