Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Implement context overflow mitigation methods #36

Open
pedro-devv opened this issue Feb 26, 2024 · 3 comments · Fixed by #56
Open

Implement context overflow mitigation methods #36

pedro-devv opened this issue Feb 26, 2024 · 3 comments · Fixed by #56
Labels
bug Something isn't working enhancement New feature or request

Comments

@pedro-devv
Copy link
Contributor

At the moment when the max context size is reached, the program panics. Methods such as sliding context window should be implemented.

@pedro-devv pedro-devv added bug Something isn't working enhancement New feature or request labels Feb 26, 2024
@pedro-devv pedro-devv reopened this Mar 7, 2024
@ElhamAryanpur
Copy link

I didn't notice the panic on my end, using both streaming and non streaming methods.

@pedro-devv
Copy link
Contributor Author

I said panic, but that's not exactly correct, what happens is an error in llama.cpp's side, I forget if it's an exception or something else. Either way, this depends on how large you make your context/session, if you make it large enough and never use the same session for too long, you probably won't ever notice it, but this definitely still happens.

@ElhamAryanpur
Copy link

I see, I was using ~2048 context length so I didn't notice at all

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working enhancement New feature or request
Projects
None yet
Development

Successfully merging a pull request may close this issue.

2 participants