Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

questions on mistral 7B with CoT #102

Open
twni2016 opened this issue Oct 29, 2024 · 1 comment
Open

questions on mistral 7B with CoT #102

twni2016 opened this issue Oct 29, 2024 · 1 comment

Comments

@twni2016
Copy link

Hi authors,

Thank you for your nice library! I am trying to use your library to run mistral 7B with CoT on gsm8k. I have several questions on the code when using HFModel:

  • Which mistral 7B model was used in your paper?
  • I tried mistralai/Mistral-7B-v0.3 and found the eos_token_id is no longer 13 Based on this, I think using eos_token_id = ["\n\n", ".\n", "\n", ".\n\n"] like Llama3Model is better?
  • Is this a typo that max_batch_size=batch_sizeis not added when calling HFModel?

Thank you for your time!

@Ber666
Copy link
Collaborator

Ber666 commented Nov 12, 2024

Hi! Thanks for your questions.

It was v0.1. See the command line here. Sorry that the library was refactored and this information was lost.

We will fix other issues you mentioned.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants