Skip to content

No labels!

There aren’t any labels for this repository quite yet.

linux
linux
Issues specific to Linux
llava
llava
LLaVa and multimodal
low severity
low severity
Used to report low severity bugs in llama.cpp (e.g. cosmetic issues, non critical UI glitches)
macos
macos
Issues specific to macOS
medium severity
medium severity
Used to report medium severity bugs in llama.cpp (e.g. Malfunctioning Features but still useable)
merge ready
merge ready
indicates that this may be ready to merge soon and is just holding out in case of objections
model
model
Model specific
need feedback
need feedback
Testing and feedback with results are needed
need more info
need more info
The OP should provide more details about the issue
nix
nix
Issues specific to consuming flake.nix, or generally concerned with ❄ Nix-based llama.cpp deployment
Nvidia GPU
Nvidia GPU
Issues specific to Nvidia GPUs
obsolete?
obsolete?
Marker for potentially obsolete PR
performance
performance
Speed related topics
porting
porting
python
python
python script changes
Qualcomm QNN
Qualcomm QNN
Qualcomm's QNN(AI Direct Engine) SDK
question
question
Further information is requested
refactoring
refactoring
Refactoring
research 🔬
research 🔬
Review Complexity : High
Review Complexity : High
Generally require indepth knowledge of LLMs or GPUs
Review Complexity : Low
Review Complexity : Low
Trivial changes to code that most beginner devs (or those who want a break) can tackle. e.g. UI fix
Review Complexity : Medium
Review Complexity : Medium
Generally require more time to grok but manageable by beginner to medium expertise level
script
script
Script related
server/api
server/api
server/webui
server/webui
split
split
GGUF split model sharding
SYCL
SYCL
https://en.wikipedia.org/wiki/SYCL - GPU programming language
sync
sync
Requires sync with the ggml repo after merging