Adaptive output and contextual dialogue capabilities of text-generation-inference #424
Open
1 of 4 tasks
Labels
bug
Something isn't working
System Info
Information
Tasks
examples
folder (such as GLUE/SQuAD, ...)Reproduction
Deploy the Llama-2-7b-chat-hf model through text-generation-inference, but there is no adaptive output when using the following command, instead the input and output size are max_new_tokens.
Also, how to implement chat functionality with context? Similar to GPT4, it can adaptively output appropriate content and has the ability to dialogue with context.
Expected behavior
The text was updated successfully, but these errors were encountered: