Replies: 1 comment
-
Try this tutorial and see if it works for your model: ggerganov/llama.cpp#7712 |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I want to run a BERT-like model on my CPU using GGML to see if it's faster than the PyTorch version. I can't find any documentation on how to use GGML. What's the easiest way of doing something like this?
Beta Was this translation helpful? Give feedback.
All reactions