Help! Want a toy example to run matmul with q40 weight by cuda kernel #9435
-
Sorry, i am not familiar with the library, I want to run a matmul between a tensor created by pytorch and the q40 weight read from gguf. |
Beta Was this translation helpful? Give feedback.
Answered by
slaren
Sep 11, 2024
Replies: 1 comment 1 reply
-
Some resources: |
Beta Was this translation helpful? Give feedback.
1 reply
Answer selected by
Eutenacity
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Some resources:
https://huggingface.co/blog/introduction-to-ggml
https://github.com/ggerganov/ggml/blob/master/examples/simple/simple-backend.cpp