You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Thank you for your nice library! I am trying to use your library to run mistral 7B with CoT on gsm8k. I have several questions on the code when using HFModel:
Which mistral 7B model was used in your paper?
I tried mistralai/Mistral-7B-v0.3 and found the eos_token_id is no longer 13 Based on this, I think using eos_token_id = ["\n\n", ".\n", "\n", ".\n\n"] like Llama3Model is better?
Is this a typo that max_batch_size=batch_sizeis not added when calling HFModel?
Thank you for your time!
The text was updated successfully, but these errors were encountered:
Hi authors,
Thank you for your nice library! I am trying to use your library to run mistral 7B with CoT on gsm8k. I have several questions on the code when using
HFModel
:mistralai/Mistral-7B-v0.3
and found theeos_token_id
is no longer13
Based on this, I think usingeos_token_id = ["\n\n", ".\n", "\n", ".\n\n"]
like Llama3Model is better?max_batch_size=batch_size
is not added when callingHFModel
?Thank you for your time!
The text was updated successfully, but these errors were encountered: