Skip to content

Commit

Permalink
sft trainer auto concats multiple datasets
Browse files Browse the repository at this point in the history
  • Loading branch information
lucidrains committed Feb 1, 2024
1 parent e9a582c commit bded2cc
Show file tree
Hide file tree
Showing 2 changed files with 3 additions and 3 deletions.
4 changes: 2 additions & 2 deletions self_rewarding_lm_pytorch/self_rewarding_lm_pytorch.py
Original file line number Diff line number Diff line change
Expand Up @@ -623,7 +623,7 @@ class FinetuneConfig:

@dataclass
class SFTConfig(FinetuneConfig):
train_dataset: Dataset
train_dataset: Union[Dataset, List[Dataset]]
valid_dataset: Optional[Dataset] = None
dropout: float = 0.1
trainer_kwargs: dict = field(default_factory = dict)
Expand Down Expand Up @@ -677,7 +677,7 @@ class SelfPlayConfig(FinetuneConfig):
@beartype
def create_default_paper_config(
*,
train_sft_dataset: Dataset,
train_sft_dataset: Union[Dataset, List[Dataset]],
self_reward_prompt_dataset: Union[Dataset, Tuple[Dataset, Dataset]],
valid_sft_dataset: Optional[Dataset] = None,
num_generated_preference_pairs = (3964, 6942),
Expand Down
2 changes: 1 addition & 1 deletion setup.py
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@
setup(
name = 'self-rewarding-lm-pytorch',
packages = find_packages(exclude=[]),
version = '0.2.1',
version = '0.2.2',
license='MIT',
description = 'Self Rewarding LM - Pytorch',
author = 'Phil Wang',
Expand Down

0 comments on commit bded2cc

Please sign in to comment.