Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Some errors when doing evaluation with hf models #352

Open
JiaQiSJTU opened this issue Sep 16, 2024 · 1 comment
Open

Some errors when doing evaluation with hf models #352

JiaQiSJTU opened this issue Sep 16, 2024 · 1 comment

Comments

@JiaQiSJTU
Copy link

JiaQiSJTU commented Sep 16, 2024

I would like to express my appreciation for this repository, particularly for its robust evaluation component.
However, during re-implementation, I encountered some unexpected behaviors in model generations. I have a few suggestions for improving the source code:

  1. In this line, it might be more appropriate to replace temperature=0 with sample=False.

  2. For the function generate_completions at this location, consider adding "add_special_tokens=False if args.use_chat_format else True".

  3. "Because many tokenizers will treat the word after space differently from the original word alone, to be consistent, we add a space before tokenization and remove it after tokenization." However, when using the llama-3 tokenizer, adding a space before "```" results in it being tokenized as a single token, so nothing should be removed?

@hamishivi
Copy link
Collaborator

Thanks for your comments! Actually, we use vllm for generation these days wherever possible, so I think we don't bump into 1 and 2 that often. I'll leave this issue open to track these.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants