-
Notifications
You must be signed in to change notification settings - Fork 892
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fine-tuning on conversations (format of conversations) #248
Comments
@Eichhof
I wish you the best of luck and if possible, I wish you could share a part of your results if it doesn't mess this thread. |
Hello, I'm still relatively new here in gpt-j. I tried to run the Colab Demo to do some inferences, especially for a chatbot use case. I don't have any idea how to stop the models from generating a new tokens after bot end up answering. In GPT3 we can easily insert a Stop Sequences or the model already good enough to know when to stop. Looks like set the "gen_len" parameter also not works. Do you have any idea for this? |
Hello
I have a dataset consisting of dialogues between two people which I would like to use for fine-tuning GPT-J. Please see below for two example dialogues. The dialogues vary in length and can be longer than the examples.
Is the format of the conversations ok? For fine-tuning, should I just concatenate all conversations into one big file or do I have to use a separator between the conversations (if yes, which separator)?
First Dialogue:
Second Dialgoue:
The text was updated successfully, but these errors were encountered: