Replies: 1 comment 4 replies
-
@Koesn Sorry for the late response, I didn't get a notification from GH discussions. llama.cpp server is OpenAI compatible, and you can use OpenAI-like endpoints with Paddler the same way you use llama.cpp specific ones: https://github.com/ggerganov/llama.cpp/tree/master/examples/server#post-v1chatcompletions-openai-compatible-chat-completions-api From what I see, llama.cpp already supports everything that exllamav2 does. Do you have any benefits from using TabbyAPI in mind over llama.cpp? Overall I am all for supporting runners other than llama.cpp. |
Beta Was this translation helpful? Give feedback.
4 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
This is great works. Is paddler works specifically with llamacpp server or it's OpenAI compatible? Is there any plan to support exllamav2 server like tabbyAPI?
Beta Was this translation helpful? Give feedback.
All reactions