We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
MacOS(M1, M2...)
=3.11
main
Installation From Source
Docker Installation
Docker Compose Installation
Cluster Installation
AutoDL Image
Other
Device: M2
LLM: ggml-model-q4_0.bin embedding: large-chinese
I download the model with
wget https://huggingface.co/TheBloke/vicuna-7B-v1.5-GGML/resolve/main/vicuna-7b-v1.5.ggmlv3.q4_K_M.bin -O models/ggml-model-q4_0.bin
and rewrite the .env file, but get this error:
.env
2023-10-02 19:45:30 xxxMacBook-Air.local pilot.model.llm.llama_cpp.llama_cpp[2889] INFO Cache capacity is 0 bytes 2023-10-02 19:45:30 xxxMacBook-Air.local pilot.model.llm.llama_cpp.llama_cpp[2889] INFO Load LLama model with params: {'model_path': '/xxx/DB-GPT/models/ggml-model-q4_0.bin', 'n_ctx': 4096, 'seed': -1, 'n_threads': None, 'n_batch': 512, 'use_mmap': True, 'use_mlock': False, 'low_vram': False, 'n_gpu_layers': 1000000000, 'n_gqa': None, 'logits_all': True, 'rms_norm_eps': 5e-06} gguf_init_from_file: invalid magic number 67676a74 error loading model: llama_model_loader: failed to load model from /xxx/DB-GPT/models/ggml-model-q4_0.bin llama_load_model_from_file: failed to load model 2023-10-02 19:45:30 xxxMacBook-Air.local pilot.model.cluster.worker.default_worker[2889] WARNING Model has been stopped!!
I follow the installation step by step but get this error
python pilot/server/dbgpt_server.py
No response
The text was updated successfully, but these errors were encountered:
llama.cpp stoped supporting for ggml model since ver 0.1.79, so use llama-cpp-python==0.1.78 or gguf model. :D
ggml
ver 0.1.79
llama-cpp-python==0.1.78
gguf
Sorry, something went wrong.
Aha, see here
feat(model): llama.cpp support new GGUF file format (#649)
f2427b1
Close #567 Close #644 Close #563 **Other** - Fix raise Exception when stop DB-GPT
Successfully merging a pull request may close this issue.
Search before asking
Operating system information
MacOS(M1, M2...)
Python version information
DB-GPT version
main
Related scenes
Installation Information
Installation From Source
Docker Installation
Docker Compose Installation
Cluster Installation
AutoDL Image
Other
Device information
Device: M2
Models information
LLM: ggml-model-q4_0.bin
embedding: large-chinese
What happened
I download the model with
and rewrite the
.env
file, but get this error:What you expected to happen
I follow the installation step by step but get this error
How to reproduce
python pilot/server/dbgpt_server.py
Additional context
No response
Are you willing to submit PR?
The text was updated successfully, but these errors were encountered: