You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Dec 6, 2023. It is now read-only.
We can start from Mistral binaries, for example instead of returing a general Load failed we should return the ValueError below
| Traceback (most recent call last):
| File "starlette/responses.py", line 273, in wrap
| File "starlette/responses.py", line 262, in stream_response
| File "routes.py", line 64, in generate_chunk_based_response
| File "llama_cpp/llama.py", line 1540, in _convert_text_completion_chunks_to_chat
| File "llama_cpp/llama.py", line 947, in _create_completion
| ValueError: Requested tokens (3308) exceed context window of 512
+------------------------------------
The text was updated successfully, but these errors were encountered:
We can start from Mistral binaries, for example instead of returing a general
Load failed
we should return the ValueError belowThe text was updated successfully, but these errors were encountered: