Replies: 1 comment
-
It is expected, the reason is that there are more kernels being compiled, which increases the size of the binary. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi there, awesome work! One question, I started compiling llama.cpp on windows with cuda a few weeks ago, back then the finished compile was around 36MB. Now it made a big jump and keeps growing, currently its at 80MB for the dll - does anybody know if that is correct behaviour? These are my compile steps:
set FORCE_CMAKE=1 && set "CMAKE_ARGS=-DLLAMA_CUBLAS=on -DLLAMA_AVX=off -DLLAMA_AVX2=off -DLLAMA_FMA=off"
python -m pip install .[all]
Beta Was this translation helpful? Give feedback.
All reactions