-
-
Notifications
You must be signed in to change notification settings - Fork 58
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Performence on terminal is very different within llamachat #24
Comments
thanks @fsrmlgd -- one thing might be to try tweaking the number of CPU cores used for prediction in LlamaChat (in Settings > General) if this is different to llama.cpp |
Thanks for your reply. The app is very helpful. llamachat.mp4warp.movAs you can see the cat on top bar, It seems ridiculous. I change nothing about default settings. |
If you need, some models can be directly used on PC are listed here. |
thanks @fsrmlgd, will take a look into this! |
I set up vicuna7b on an m1 Mac mini and am seeing the same thing. Text in llamachat generates very slowly just like in the video above. Mac was at 75% cpu usage. (and this is fine if llamas are still too heavy for low end systems, but that doesn’t seem like the case here.) |
@spacecowgoesmoo thanks, still getting round to this! |
Hello, here to say I experience the same issue. On terminal it works smoothly, but the GUI makes it very slow. btw amazing app design ! |
I find the performence when using llama.cpp in terminal is better in llamachat.
The answer speed in terminal (interactive mode) is clearly faster on the same cpu usage level.
It's a little strange. Maybe it can be solved.
The text was updated successfully, but these errors were encountered: