diff --git a/assets/logo.svg b/assets/logo.svg
new file mode 100644
index 00000000..42f60cb7
--- /dev/null
+++ b/assets/logo.svg
@@ -0,0 +1,21 @@
+
diff --git a/models.md b/models.md
index a180490b..f50218b7 100644
--- a/models.md
+++ b/models.md
@@ -324,6 +324,55 @@ curl -X POST http://localhost:8080/v1/chat/completions \
+
+ OpenChat-3.5-0106
+
+Download the model
+
+```bash
+curl -LO https://huggingface.co/second-state/OpenChat-3.5-0106-GGUF/resolve/main/openchat-3.5-0106-Q5_K_M.gguf
+```
+
+Please check the sha256sum of the Downloaded model file to make sure it is correct.
+
+```bash
+shasum -a 256 openchat-3.5-0106-Q5_K_M.gguf
+output: c28f69693336ab63369451da7f1365e5003d79f3ac69566de72100a8299a967a
+```
+
+Chat with the model on the CLI
+
+```bash
+curl -LO https://github.com/second-state/LlamaEdge/releases/latest/download/llama-chat.wasm
+wasmedge --dir .:. --nn-preload default:GGML:AUTO:openchat-3.5-0106-Q5_K_M.gguf llama-chat.wasm -p openchat -r '<|end_of_turn|>'
+```
+
+Chat with the model via a web UI
+
+```bash
+curl -LO https://github.com/second-state/LlamaEdge/releases/latest/download/llama-api-server.wasm
+curl -LO https://github.com/second-state/chatbot-ui/releases/download/v0.1.0/chatbot-ui.tar.gz
+tar xzf chatbot-ui.tar.gz
+rm chatbot-ui.tar.gz
+
+wasmedge --dir .:. --nn-preload default:GGML:AUTO:openchat-3.5-0106-Q5_K_M.gguf llama-api-server.wasm -p openchat -r '<|end_of_turn|>'
+```
+
+Open your browser to http://localhost:8080 to start the chat!
+
+Send an API request to the server
+
+Test the API server from another terminal using the following command
+
+```bash
+curl -X POST http://localhost:8080/v1/chat/completions \
+ -H 'accept:application/json' \
+ -H 'Content-Type: application/json' \
+ -d '{"messages":[{"role":"system", "content": "You are a helpful assistant."}, {"role":"user", "content": "What is the capital of France?"}], "model":"OpenChat-3.5-0106"}'
+```
+
+
+
OpenChat-3.5-1210
diff --git a/run-llm.sh b/run-llm.sh
index 435ec5e7..c9ec8de6 100755
--- a/run-llm.sh
+++ b/run-llm.sh
@@ -45,18 +45,6 @@ prereq() {
# fi
fi
done
-
- # Check if libopenblas is installed
- if [[ "$OSTYPE" == "linux-gnu"* ]]; then
- if ! ldconfig -p | grep libopenblas &> /dev/null
- then
- printf "'libopenblas' is required for wasi-nn plugin to run.\n"
- exit 1
- fi
- elif [[ "$OSTYPE" == "darwin"* ]]; then
- # Todo check libopenblas in MacOS
- printf "" # Placeholder
- fi
}
install_wasmedge() {
@@ -105,8 +93,7 @@ BELLE-Llama2-13B-Chat::https://code.flows.network/webhook/iwYN1SdN3AmPgR5ao5Gt/B
MistralLite-7B::https://code.flows.network/webhook/iwYN1SdN3AmPgR5ao5Gt/mistrallite.Q5_K_M.gguf
Mistral-7B-Instruct-v0.1::https://code.flows.network/webhook/iwYN1SdN3AmPgR5ao5Gt/mistral-7b-instruct-v0.1.Q5_K_M.gguf
Mistral-7B-Instruct-v0.2::https://code.flows.network/webhook/iwYN1SdN3AmPgR5ao5Gt/mistral-7b-instruct-v0.2.Q4_0.gguf
-OpenChat-3.5-1210::https://code.flows.network/webhook/iwYN1SdN3AmPgR5ao5Gt/openchat-3.5-1210.Q5_K_M.gguf
-OpenChat-3.5::https://code.flows.network/webhook/iwYN1SdN3AmPgR5ao5Gt/openchat_3.5.Q5_K_M.gguf
+OpenChat-3.5-0106::https://code.flows.network/webhook/iwYN1SdN3AmPgR5ao5Gt/openchat-3.5-0106-Q5_K_M.gguf
Wizard-Vicuna::https://code.flows.network/webhook/iwYN1SdN3AmPgR5ao5Gt/wizard-vicuna-13b-ggml-model-q8_0.gguf
CausalLM-14B::https://code.flows.network/webhook/iwYN1SdN3AmPgR5ao5Gt/causallm_14b.Q5_1.gguf
TinyLlama-1.1B-Chat-v1.0::https://code.flows.network/webhook/iwYN1SdN3AmPgR5ao5Gt/tinyllama-1.1b-chat-v1.0.Q5_K_M.gguf
@@ -146,8 +133,7 @@ BELLE-Llama2-13B-Chat::56879e1fd6ee6a138286730e121f2dba1be51b8f7e261514a594dea89
MistralLite-7B::d06d149c24eea0446ea7aad596aca396fe7f3302441e9375d5bbd3fd9ba8ebea
Mistral-7B-Instruct-v0.1::c4b062ec7f0f160e848a0e34c4e291b9e39b3fc60df5b201c038e7064dbbdcdc
Mistral-7B-Instruct-v0.2::25d80b918e4432661726ef408b248005bebefe3f8e1ac722d55d0c5dcf2893e0
-OpenChat-3.5-1210::fbbb15ed13c630110b1790e0a0a621419fbf2f5564ead433c6ded9f8f196b95c
-OpenChat-3.5::3abf26b0f2ff11394351a23f8d538a1404a2afb69465a6bbaba8836fef51899d
+OpenChat-3.5-0106::c28f69693336ab63369451da7f1365e5003d79f3ac69566de72100a8299a967a
Wizard-Vicuna::681b6571e624fd211ae81308b573f24f0016f6352252ae98241b44983bb7e756
CausalLM-14B::8ddb4c04e6f0c06971e9b6723688206bf9a5b8ffc85611cc7843c0e8c8a66c4e
TinyLlama-1.1B-Chat-v1.0::aa54a5fb99ace5b964859cf072346631b2da6109715a805d07161d157c66ce7f
@@ -186,8 +172,7 @@ BELLE-Llama2-13B-Chat::belle-llama-2-chat
MistralLite-7B::mistrallite
Mistral-7B-Instruct-v0.1::mistral-instruct
Mistral-7B-Instruct-v0.2::mistral-instruct
-OpenChat-3.5-1210::openchat
-OpenChat-3.5::openchat
+OpenChat-3.5-0106::openchat
Wizard-Vicuna::vicuna-chat
CausalLM-14B::chatml
TinyLlama-1.1B-Chat-v1.0::chatml
@@ -214,7 +199,7 @@ Calm2-7B-Chat::vicuna-1.1-chat
Deepseek-Coder-6.7B::deepseek-coder
Deepseek-LLM-7B-Chat::deepseek-chat
SOLAR-10.7B-Instruct-v1.0::solar-instruct
-Mixtral-8x7B-Instruct-v0.1::mixtral-instruct
+Mixtral-8x7B-Instruct-v0.1::mistral-instruct
dolphin-2.6-phi-2::chatml
ELYZA-japanese-Llama-2-7b-instruct::llama-2-chat
ELYZA-japanese-Llama-2-7b-fast-instruct::llama-2-chat
@@ -234,8 +219,7 @@ Orca-2-13B::You are Orca, an AI language model created by Microsoft. You are a c
reverse_prompts='
MistralLite-7B::
-OpenChat-3.5-1210::<|end_of_turn|>
-OpenChat-3.5::<|end_of_turn|>
+OpenChat-3.5-0106::<|end_of_turn|>
Baichuan2-13B-Chat::用户:
OpenHermes-2.5-Mistral-7B::<|im_end|>
Dolphin-2.0-Mistral-7B::<|im_end|>
@@ -248,7 +232,7 @@ Zephyr-7B-Alpha::
Starling-LM-7B-alpha::<|end_of_turn|>
'
- model_names="Llama-2-7B-Chat Llama-2-13B-Chat BELLE-Llama2-13B-Chat MistralLite-7B Mistral-7B-Instruct-v0.1 Mistral-7B-Instruct-v0.2 OpenChat-3.5-1210 OpenChat-3.5 Wizard-Vicuna CausalLM-14B TinyLlama-1.1B-Chat-v1.0 TinyLlama-1.1B-Chat-v0.3 Baichuan2-13B-Chat OpenHermes-2.5-Mistral-7B Dolphin-2.0-Mistral-7B Dolphin-2.1-Mistral-7B Dolphin-2.2-Yi-34B Dolphin-2.2-Mistral-7B Dolphin-2.2.1-Mistral-7B Samantha-1.2-Mistral-7B Samantha-1.11-CodeLlama-34B Samantha-1.11-7B WizardLM-1.0-Uncensored-CodeLlama-34B WizardLM-7B-V1.0-Uncensored WizardLM-13B-V1.0-Uncensored WizardCoder-Python-7B-V1.0 Zephyr-7B-Alpha Orca-2-13B Neural-Chat-7B-v3-1 Starling-LM-7B-alpha Calm2-7B-Chat Deepseek-Coder-6.7B Deepseek-LLM-7B-Chat SOLAR-10.7B-Instruct-v1.0 Mixtral-8x7B-Instruct-v0.1 ELYZA-japanese-Llama-2-7b-fast-instruct ELYZA-japanese-Llama-2-7b-instruct dolphin-2.6-phi-2"
+ model_names="Llama-2-7B-Chat Llama-2-13B-Chat BELLE-Llama2-13B-Chat MistralLite-7B Mistral-7B-Instruct-v0.1 Mistral-7B-Instruct-v0.2 OpenChat-3.5-0106 Wizard-Vicuna CausalLM-14B TinyLlama-1.1B-Chat-v1.0 TinyLlama-1.1B-Chat-v0.3 Baichuan2-13B-Chat OpenHermes-2.5-Mistral-7B Dolphin-2.0-Mistral-7B Dolphin-2.1-Mistral-7B Dolphin-2.2-Yi-34B Dolphin-2.2-Mistral-7B Dolphin-2.2.1-Mistral-7B Samantha-1.2-Mistral-7B Samantha-1.11-CodeLlama-34B Samantha-1.11-7B WizardLM-1.0-Uncensored-CodeLlama-34B WizardLM-7B-V1.0-Uncensored WizardLM-13B-V1.0-Uncensored WizardCoder-Python-7B-V1.0 Zephyr-7B-Alpha Orca-2-13B Neural-Chat-7B-v3-1 Starling-LM-7B-alpha Calm2-7B-Chat Deepseek-Coder-6.7B Deepseek-LLM-7B-Chat SOLAR-10.7B-Instruct-v1.0 Mixtral-8x7B-Instruct-v0.1 ELYZA-japanese-Llama-2-7b-fast-instruct ELYZA-japanese-Llama-2-7b-instruct dolphin-2.6-phi-2"
# Convert model_names to an array
model_names_array=($model_names)