Releases: LlamaEdge/rag-api-server
Releases · LlamaEdge/rag-api-server
LlamaEdge-RAG 0.13.15
Major changes:
- Upgrade
llama-coredep to0.30.0 - Upgrade
chat-promptsdep to0.26.1
LlamaEdge-RAG 0.13.14
Major changes:
- Support Llama-4 (text-only)
- Verified on
ggml plugin b5074
LlamaEdge-RAG 0.13.13
Major changes:
- Support Tessa-T1 react reasoning models
LlamaEdge-RAG 0.13.12
Major changes:
- Support LGAI-EXAONE/EXAONE-3.5-Instruct models
LlamaEdge-RAG 0.13.11
Major changes:
-
Support LGAI-EXAONE/EXAONE-Deep models
-
Verified on
ggml plugin b4917
LlamaEdge-RAG 0.13.10
Major changes:
- Automatically switch
rag_policyfromsystem-messagetolast-user-messagewhen the model does not support system prompt.
LlamaEdge-RAG 0.13.9
Major changes:
- Upgrade
llama-coredep to0.28.2 - Verified on
ggml plugin b4875
LlamaEdge-RAG 0.13.8
Major changes:
- Upgrade
llama-coredep to0.28.0 - Upgrade
chat-promptsdep to0.21.2 - Upgrade
endpointsdep to0.25.1
LlamaEdge-RAG 0.13.7
Major change:
- Bug fix
LlamaEdge-RAG 0.13.6
Major changes:
- (NEW) Add
--include-usageCLI option. Note that theinclude_usageparam in requests has higher priority than the--include-usageCLI option.