Beyond simple inference #7
Replies: 5 comments 4 replies
-
Hi, thanks! This feature is next on my list, and I’m excited to dive into it within the next week or two. Stay tuned! |
Beta Was this translation helpful? Give feedback.
-
I sent you a msg on discord. |
Beta Was this translation helpful? Give feedback.
-
Hi Jarrod. jiAddMessage(jiROLE_USER, 'A quale città sto pensando ora?');
jiAddMessage(jiROLE_ASSISTANT, 'Tu stai pensando alla città di Ancona');
jiAddMessage(jiROLE_USER, 'E a quale città sto pensando adesso?');
jiAddMessage(jiROLE_ASSISTANT, 'Adesso invece stai pensando alla città di Roma');
jiAddMessage(jiROLE_USER, 'Indicami un tragitto per arrivare alla seconda città partendo dalla prima'); and the result was this: So I would say that everything works correctly. |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
https://github.com/ggerganov/llama.cpp/wiki/Templates-supported-by-llama_chat_apply_template This is the way I do it in Lumina, but llama.cpp does not get those templates from the GGUF file, they are hard coded in llama.cpp and based on the model, it will look it up and if not found just default to chatML (OpenAI format). That template can have logic it and as stated in the link, llama.cpp does not currently support this due to its complexity. Anyway, all ideas are welcome, and I will take everything into consideration. Since it's not released yet and is still WIP, can I ask you to communicate with me on discord. I've already sent you an updated build. I will be setting up a repo for it soon. The new build I sent you has a pretty much complete API now as well as docs. |
Beta Was this translation helpful? Give feedback.
-
Hi Jarrod.
Do you have a date in mind to implement inference via the chat template paradigm, with the possibility to specify the various roles of individual messages, as was done in Infero?
Bye and thanks again for your great work.
Sergio
Beta Was this translation helpful? Give feedback.
All reactions