Releases: LlamaEdge/rag-api-server
Releases · LlamaEdge/rag-api-server
LlamaEdge-RAG 0.7.1
Major changes:
- Update deps
endpoints v0.9.1
chat-prompts v0.8.2
llama-core v0.11.4
LlamaEdge-RAG 0.7.0
Major change:
- (BREAKING) Update the response data of the
/v1/info
endpoint
LlamaEdge-RAG 0.6.6
Major change:
- Work around DNS failure in some containers by updating the default value of
qdrant-url
CLI option fromhttp://localhost:6333
tohttp://127.0.0.1:6333
LlamaEdge-RAG 0.6.5
Major change:
- Log the buffer of rag query request body while triggering bad request errors.
LlamaEdge-RAG 0.6.4
Major changes:
- Improve the rag context retrieval
- Upgrade the
endpoints
dependency tov0.8.1
LlamaEdge-RAG 0.6.3
Major changes:
- Add logging
LlamaEdge-RAG 0.6.2
Major change:
- Update to
llama-core v0.11.1
LlamaEdge-RAG 0.6.1
Major changes:
- Support
embedding
prompt template for embedding model
LlamaEdge-RAG 0.6.0
Major changes:
- Improve
--batch-size
CLI option: set batch size for chat and embedding models, respectively
LlamaEdge-RAG 0.5.3
Major changes:
- Add
user
header in chat completion responses - Support
PLUGIN_DEBUG
wasm environment variable for debugging the low-level ggml plugin