Is there high-level API for LLama.cpp on C++? #8074
Replies: 4 comments 8 replies
-
We actually planned to add one since a long time ago: #5215 |
Beta Was this translation helpful? Give feedback.
-
In the meantime, looking at the simple example also helps (helped me, at least). |
Beta Was this translation helpful? Give feedback.
-
why do you integrate, instead of running service independently then call the service from you application? trying to learn from other folks' best practice |
Beta Was this translation helpful? Give feedback.
-
I'm guessing C-like C++ is kinda necessary for bindings to other languges. I'm only just taking a first look through the source code, but if the details are somewhat low-level and based on some of the C-like style I'm seeing I'd assume that's part of the reason. But I'm also interested because C++ should in general be more optimisable than C, especially with some of the modern features since say C++14. Personally, I generally write in modular C++ and then try to wrap and bind in for C if I need to than the other way round. Out of interest, has this ever been considered, both from performance and maintainability? |
Beta Was this translation helpful? Give feedback.
-
I am trying to integrate LLama.cpp in a C++ project that I have and I am looking for high level APIs to interact with it, the closest thing that I got was in examples/main, however it contains a lot of low level details, and it outputs a lot of debug and hardware information and that I can't find a way to disable.
Python have these nice high level API:
Rust has one too:
Is there one for C++ as well that I missed?
Beta Was this translation helpful? Give feedback.
All reactions