Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

关于文本生成 #219

Open
yuzhang112 opened this issue Nov 2, 2021 · 2 comments
Open

关于文本生成 #219

yuzhang112 opened this issue Nov 2, 2021 · 2 comments

Comments

@yuzhang112
Copy link

你好,我想用自己的语料,对预训练GPT2模型进行微调,然后进行文本生成任务,请问可以用增量预训练的命令进行微调吗?

这是我现在使用的预训练命令:
python3 pretrain.py --dataset_path dataset.pt --pretrained_model_path models/cluecorpussmall_gpt2_seq1024_model.bin-250000 --vocab_path models/google_zh_vocab.txt --config_path models/gpt2/distil_config.json --output_model_path models/my_gpt2_model.bin --world_size 1 --gpu_ranks 0 --total_steps 10000 --save_checkpoint_steps 5000 --report_steps 1000 --learning_rate 5e-5 --batch_size 32 --embedding word_pos --remove_embedding_layernorm --encoder transformer --mask causal --layernorm_positioning pre --target lm --tie_weights

@ydli-ai
Copy link
Collaborator

ydli-ai commented Nov 2, 2021 via email

@yuzhang112
Copy link
Author

好的,感谢回复

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants