Hi, I just build a llama model from llama.cpp, it can work on llama.cpp but llama-cpp-python #334
Replies: 1 comment
-
Moved to issue #336 |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hi, i just build my llama.cpp model . it works fine on llama.cpp, but when i move the model to llama-cpp-python by following the code like:
nllm = LlamaCpp(
model_path="./ggml-model-q4_0.bin",
callback_manager=callback_manager,
verbose=True,
n_ctx=2048
)
it is always encount an error like this:
llama.cpp: loading model from ./ggml-model-q4_0.bin
error loading model: unrecognized tensor type 14
llama_init_from_file: failed to load model
2023-06-07 15:15:52.963 Uncaught app exception
Traceback (most recent call last):
File "C:\Users\IamUser\AppData\Local\Programs\Python\Python310\lib\site-packages\streamlit\runtime\scriptrunner\script_runner.py", line 565, in _run_script
exec(code, module.dict)
File "D:\PythonPlayground\NN-Playgroud\chatbot_NPC_llama.py", line 39, in
nllm = LlamaCpp(
File "pydantic\main.py", line 341, in pydantic.main.BaseModel.init
pydantic.error_wrappers.ValidationError: 1 validation error for LlamaCpp
root
Could not load Llama model from path: ./ggml-model-q4_0.bin. Received error (type=value_error)
llama.cpp version : master-ffb06a3
llama-cpp-python: v0.157
it will work if i use older version of model and use v0.149 of llama-cpp-python
but i will like to use the lastest version. How can i solve this?
Thank you
Beta Was this translation helpful? Give feedback.
All reactions