Convert GGML to expect GGUF format #581
Merged
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
As of
llama.cpp
version 1046 the model format expected for GGML based tooling is now GGUF.This revision improves initialization to validate model file is in
GUFF
format and enhances error handling for subprocess execution.Changes take the approach that first
_call_model()
invocation will raise an exception if asubprocess.run()
raises an error however subsequent invocations will log the exception and returnNone
allowing the run to continue. Any other exception will be logged and returnNone
.Updates to
requirements.txt
andpyproject.toml
document that typing#573 impacts loadinggarak
when loaded in a debugger such as debugpy. Since the project requires python >= 3.10 and thetyping
library is included as a system util since 3.6 in theory the change has zero impact on requirements.Impact of this change:
Consider the case where user error presented the wrong filename for the model.
Garak's output was unclear as to cause.
When executing the command directly however the user error is more obvious:
Once working encoding test later errors with:
By expanding the error handling the testing can now complete.