-
Notifications
You must be signed in to change notification settings - Fork 487
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fix #8, fix #41 Add llamacpp support #11
base: main
Are you sure you want to change the base?
Conversation
+1 for using abetlen/llama-cpp-python I think the devs are doing a very good job at supporting the latest |
I'm unable to get inference going with the branch - any tips? I've managed to get it load the alpaca models I have, and it clearly starts up llama.cpp in the background, but the only tokens I get back are single "\n" characters. Here's a bit of the log that shows inference occurring in case that helps!
|
I cannot found anything unusual from the log. Maybe you should add some log at server/lib/inference/init.py line line 623. And check if there is any problem with llama-cpp-python? |
Did your set the prompt template file for alpaca? Here is my as an example:
|
Aha, that prompt did the trick! Thanks! Do you have example prompts for other models? I didn't realize they used |
sure, I added some more in the Readme file. |
add basic llama.cpp support via abetlen/llama-cpp-python