Skip to content

Commit

Permalink
[BUG] bckward compatibility vllm do_sample -> use_beam_search (#373)
Browse files Browse the repository at this point in the history
  • Loading branch information
YannDubs authored Jul 18, 2024
1 parent 19866ff commit 6c8014c
Show file tree
Hide file tree
Showing 21 changed files with 21 additions and 21 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -9,7 +9,7 @@ LMCocktail-10.7B-v1:
max_new_tokens: 2048
temperature: 0.7
top_p: 1.0
do_sample: True
use_beam_search: True
length_penalty: 2.0
earlth_stopping: True
pretty_name: "LMCocktail-10.7B-v1"
Expand Down
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
Llama-3-Instruct-8B-SimPO:
completions_kwargs:
batch_size: 900
do_sample: true
use_beam_search: true
max_new_tokens: 4096
model_kwargs:
dtype: bfloat16
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,7 @@ Starling-LM-7B-alpha-ExPO:
dtype: 'bfloat16'
tokenizer_mode: 'auto'
max_new_tokens: 2048
do_sample: True
use_beam_search: True
seed: 42
temperature: 0.7
top_k: 50
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,7 @@ Starling-LM-7B-beta-ExPO:
dtype: 'bfloat16'
tokenizer_mode: 'auto'
max_new_tokens: 2048
do_sample: True
use_beam_search: True
seed: 42
temperature: 0.7
top_k: 50
Expand Down
2 changes: 1 addition & 1 deletion src/alpaca_eval/models_configs/causallm-14b/configs.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ causallm-14b:
max_new_tokens: 2048
temperature: 0.7
top_p: 1.0
do_sample: False
use_beam_search: False
batch_size: 8
pretty_name: "CausalLM-14B"
link: "https://huggingface.co/CausalLM/14B"
2 changes: 1 addition & 1 deletion src/alpaca_eval/models_configs/deita-7b-v1.0/configs.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -8,6 +8,6 @@ deita-7b-v1.0:
max_new_tokens: 2048
temperature: 0.7
top_p: 1.0
do_sample: True
use_beam_search: True
pretty_name: "DEITA 7B v1.0"
link: "https://github.com/hkust-nlp/deita"
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ internlm2-chat-20b-ExPO:
trust_remote_code: True
tokenizer_mode: 'auto'
max_new_tokens: 2048
do_sample: True
use_beam_search: True
seed: 42
temperature: 0.7
top_k: 50
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ internlm2-chat-7b-ExPO:
trust_remote_code: True
tokenizer_mode: 'auto'
max_new_tokens: 2048
do_sample: True
use_beam_search: True
seed: 42
temperature: 0.7
top_k: 50
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@ merlinite-7B-AOT:
model_name: "merlinite-7B-AOT"
model_kwargs:
torch_dtype: 'bfloat16'
do_sample: True
use_beam_search: True
temperature: 0.7
max_tokens: 2048
top_p: 0.9
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -9,7 +9,7 @@ openbuddy-llama2-13b-v11.1:
max_new_tokens: 2000 # on vllm, prompt_tokens+max_new_tokens must be <= model_max_length otherwise it will raise an error and nothing will be generated
temperature: 0.7
top_p: 1.0
do_sample: False
use_beam_search: False
batch_size: 16
pretty_name: "OpenBudddy-LLaMA2-13B-v11.1"
link: "https://huggingface.co/OpenBuddy/openbuddy-llama2-13b-v11.1-bf16"
Original file line number Diff line number Diff line change
Expand Up @@ -9,7 +9,7 @@ pairrm-tulu-2-13b:
max_new_tokens: 7500
temperature: 0.0
top_p: 1.0
do_sample: False
use_beam_search: False
best_of: 16 # number of completions to generate, using PairRM to select the best one
batch_size: 800
pretty_name: "PairRM 0.4B+Tulu 2+DPO 13B (best-of-16)"
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -9,7 +9,7 @@ pairrm-tulu-2-70b:
max_new_tokens: 7500
temperature: 0.0
top_p: 1.0
do_sample: False
use_beam_search: False
best_of: 16 # number of completions to generate, using PairRM to select the best one
batch_size: 800
pretty_name: "PairRM 0.4B+Tulu 2+DPO 70B (best-of-16)"
Expand Down
2 changes: 1 addition & 1 deletion src/alpaca_eval/models_configs/phi-2/configs.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -11,6 +11,6 @@ phi-2:
max_new_tokens: 2048
temperature: 0.7
top_p: 1.0
do_sample: True
use_beam_search: True
pretty_name: "Phi 2"
link: "https://huggingface.co/microsoft/phi-2"
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,7 @@ tulu-2-dpo-13b-ExPO:
dtype: 'bfloat16'
tokenizer_mode: 'auto'
max_new_tokens: 2048
do_sample: True
use_beam_search: True
seed: 42
temperature: 0.7
top_k: 50
Expand Down
2 changes: 1 addition & 1 deletion src/alpaca_eval/models_configs/tulu-2-dpo-13b/configs.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -9,7 +9,7 @@ tulu-2-dpo-13b:
max_new_tokens: 7500
temperature: 0.0
top_p: 1.0
do_sample: False
use_beam_search: False
batch_size: 800
pretty_name: "Tulu 2+DPO 13B"
link: "https://huggingface.co/allenai/tulu-2-dpo-13b"
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ tulu-2-dpo-70b-ExPO:
tokenizer_mode: 'auto'
tensor_parallel_size: 2 # you need at least 2 A100 80GB GPUs to run this model
max_new_tokens: 2048
do_sample: True
use_beam_search: True
seed: 42
temperature: 0.7
top_k: 50
Expand Down
2 changes: 1 addition & 1 deletion src/alpaca_eval/models_configs/tulu-2-dpo-70b/configs.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -9,7 +9,7 @@ tulu-2-dpo-70b:
max_new_tokens: 7500
temperature: 0.0
top_p: 1.0
do_sample: False
use_beam_search: False
batch_size: 800
pretty_name: "Tulu 2+DPO 70B"
link: "https://huggingface.co/allenai/tulu-2-dpo-70b"
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,7 @@ tulu-2-dpo-7b-ExPO:
dtype: 'bfloat16'
tokenizer_mode: 'auto'
max_new_tokens: 2048
do_sample: True
use_beam_search: True
seed: 42
temperature: 0.7
top_k: 50
Expand Down
2 changes: 1 addition & 1 deletion src/alpaca_eval/models_configs/tulu-2-dpo-7b/configs.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -9,7 +9,7 @@ tulu-2-dpo-7b:
max_new_tokens: 7500
temperature: 0.0
top_p: 1.0
do_sample: False
use_beam_search: False
batch_size: 800
pretty_name: "Tulu 2+DPO 7B"
link: "https://huggingface.co/allenai/tulu-2-dpo-7b"
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,7 @@ zephyr-7b-alpha-ExPO:
dtype: 'bfloat16'
tokenizer_mode: 'auto'
max_new_tokens: 2048
do_sample: True
use_beam_search: True
seed: 42
temperature: 0.7
top_k: 50
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,7 @@ zephyr-7b-beta-ExPO:
dtype: 'bfloat16'
tokenizer_mode: 'auto'
max_new_tokens: 2048
do_sample: True
use_beam_search: True
seed: 42
temperature: 0.7
top_k: 50
Expand Down

0 comments on commit 6c8014c

Please sign in to comment.