Replies: 4 comments 2 replies
-
this might be related to issue #993 |
Beta Was this translation helpful? Give feedback.
-
You could try using the |
Beta Was this translation helpful? Give feedback.
-
@mgroeber9110 thanks, just tried it but unfortunatelly llama still stops randomly and requires Enter key to be pressed until it can finally return control to the user. |
Beta Was this translation helpful? Give feedback.
-
Also, make sure you're using |
Beta Was this translation helpful? Give feedback.
-
So running various models with chat like behavior works nicely, however I always have to trigger llama.cpp to give more tokens by pressing enter.
I couldn't find a parameter to improve that behavior and it seems nobody else has this problem? I'm highly confused about that.
My expectation would be that the output runs without interruption until the "reverse prompt" appears and I can enter something again.
So is there a way to give control to llama until the reverse prompt appears again?
Beta Was this translation helpful? Give feedback.
All reactions