Skip to content
This repository has been archived by the owner on Jun 24, 2024. It is now read-only.

Perplexity can/will segfault over large input sizes #257

Open
philpax opened this issue May 20, 2023 · 3 comments
Open

Perplexity can/will segfault over large input sizes #257

philpax opened this issue May 20, 2023 · 3 comments
Labels
app:cli App: the `llm` CLI issue:bug Something isn't working

Comments

@philpax
Copy link
Collaborator

philpax commented May 20, 2023

The newly-added perplexity will segfault with large enough input. This is likely because the current implementation of the LLaMA model isn't really set up to handle going over context, and updating it (#210) should fix this.

The other models segfault too (from a quick test), but perplexity is primarily useful for comparing against llama.cpp, so I'm not as fussed. It would be nice to figure this out at some point, though.

@philpax philpax added issue:bug Something isn't working app:cli App: the `llm` CLI labels May 20, 2023
@philpax
Copy link
Collaborator Author

philpax commented Jun 18, 2023

Just a heads-up: updating LLaMA did not in fact fix this. We have other bugs we need to fix here, but the segfault is within ggml, so it's harder to track down.

@steventrouble
Copy link
Contributor

Note: #362 (context size change) affects this issue due to this line:

let n_ctx = model.context_size();

@philpax
Copy link
Collaborator Author

philpax commented Jul 12, 2023

Good catch! That almost certainly fixes this. Need to retest...

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
app:cli App: the `llm` CLI issue:bug Something isn't working
Projects
None yet
Development

No branches or pull requests

2 participants