-
Notifications
You must be signed in to change notification settings - Fork 34
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Bot is consistently dropping responses. Best way to work around this? #68
Comments
Hallucinations of the other user's messages is present in pretty much every model. There are some mitigations, but you should also be adding your specific model's stopping strings in the config.yml. |
The response is dropped because oobabot is detecting it is a garbage reply and decides not to send it. |
I've gotten that error too sometimes. |
No, it did not generate an empty message, I am sure of it. It is because the bot is hallucinating user input in it's replies. Here is an example I just generated. From the logs, their response to "How are you" (this continues on for at least 3x the length, I didn't bother copying it all though.)
Which leads to the error |
I was wrong earlier, I misread the error.
Log entry 1 is common and is due to the bot attempting to "continue" the dialogue (as instructed by the default prompt). The error code itself can be seen here: oobabot/src/oobabot/discord_bot.py Line 440 in c5546b1
|
From what I'm able to tell it usually filters out when the bot tries to predict multiple peoples lines. This happens from me from time to time and requires a bit of readjusting the settings until it's reigned the model back in. Unfortunately I have not been able to figure out how to solve this issue when the model is Llama 2 13B. |
I've found this to be challenging as well. One thing which I've found help is more context within the channel.... when it's a new channel, or just me and the bot, then I find the bot has less context to work with and hallucinates in this way. So one option is... just wait and see if it goes away. Now to give the complete opposite advice -- sometimes if the bot is really stuck on the idea of replying as someone else, a One thing which might solve this more robustly is for us to move to the chat-api provided by newer versions of text-generation-webui. This uses a prompt library built into textgen, which is better tuned to a variety of models. I think this is something we should definitely something we should do soon, though it may be a few weeks until I can get to it personally. In terms of models, lately I've been using Anyway, sorry for the frustration and I hope some of this helps! |
|
Hello, I have been running oobabot on a private server for a bit now. My typical setup is a 13b parameter ggml file split across cpu and gpu. when it works, everything is fantastic. However quite often, across all kinds of models I've used, it has generated hallucinated user replies, and has caused the response to be silently scrapped, which makes it ultimately unreliable.
Is there some specific models recommended to be used that play well with oobabot's instructions to avoid this? I've tried quite a few, but found issues in nearly all cases. Should the model prompting be tweaked to more follow the instruction format the model is trained off of?
And I guess, is anyone else getting this issue?
Thanks, and apologies if I didn't include any details needed.
The text was updated successfully, but these errors were encountered: