Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Question about the metric reported in the paper? #1091

Open
dsj96 opened this issue Apr 4, 2023 · 0 comments
Open

Question about the metric reported in the paper? #1091

dsj96 opened this issue Apr 4, 2023 · 0 comments

Comments

@dsj96
Copy link

dsj96 commented Apr 4, 2023

Question about the metric reported in the paper?.
HELLO! I am a new NLPer. I am confused about the pipline(pretrain->fineturn->test) of pre-training large language models.

  1. I would like to know which stage of the model was used for unlabeled dataset (e.g., c4), labeled dataset (e.g., glue, superGLUE WMT), respectively?
    In paper, section 2.4, I find that
    We instead allow for separately fine-tuning the model on each individual task and use short task prefixes instead of an explicit question-answer format.
    As shown in Table 1 of the paper, dose T5 model pre-trained on dataset C4, then fine-tuned on GLUE, CNNDM, SQuAD, SGLUE and WMT dataset, respectively? Finally, reported the score in Table 1.
  2. Other Large Language Models, like GPT, GPT2, have these models been fine-tuned on labeled dataset before reporting the scores?

Thank you!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant