v1.7.0
New Features:
- Added support for swift export, enabling awq-int4 quantization and gpt-int2,3,4,8 quantization. Models can be pushed to the Modelscope Hub. You can view the documentation here.
- Enabled fine-tuning of awq quantized models.
- Enabled fine-tuning of aqlm quantized models.
- Added support for deploying LLM with infer_backend='pt'.
- Added web-ui with task management and visualization of training loss, eval loss, etc. Inference is accelerated using VLLM.
New Tuners:
- Lora+.
- LlamaPro.
New Models:
- qwen1.5 awq series.
- gemma series.
- yi-9b.
- deepseek-math series.
- internlm2-1_8b series.
- openbuddy-mixtral-moe-7b-chat.
- llama2 aqlm series.
New Datasets:
- ms-bench-mini.
- hh-rlhf-cn series.
- disc-law-sft-zh, disc-med-sft-zh.
- pileval.
What's Changed
- Fix vllm==0.3 deploy bug by @Jintao-Huang in #412
- Support deepseek math by @Jintao-Huang in #413
- update support_vllm by @Jintao-Huang in #415
- fix zero3 & swift lora by @Jintao-Huang in #416
- Support peft0.8.0 by @tastelikefeet in #423
- update readme by @Jintao-Huang in #426
- fix pai open with 'a' by @Jintao-Huang in #430
- default load_best_model_at_end=False by @Jintao-Huang in #432
- support openbuddy mixtral by @Jintao-Huang in #437
- support gemma by @Jintao-Huang in #441
- Support ms bench mini by @Jintao-Huang in #442
- Add roadmap and contributing doc by @tastelikefeet in #431
- support peft format by @tastelikefeet in #438
- update contributing.md by @Jintao-Huang in #446
- fix link by @tastelikefeet in #447
- Fix rlhf dataset by @tastelikefeet in #451
- Add task management for webui by @tastelikefeet in #457
- Support swift export by @Jintao-Huang in #455
- Fix llm quantization docs by @Jintao-Huang in #458
- fix get_vllm_engine bug by @Jintao-Huang in #463
- use cpu export by @Jintao-Huang in #462
- Fix llama2 generation config by @Jintao-Huang in #468
- Support editing model_id_or_path by @tastelikefeet in #469
- Support pt deploy by @Jintao-Huang in #467
- Fix swift deploy bug by @Jintao-Huang in #470
- fix deploy dep by @Jintao-Huang in #471
- Support LLaMAPRO and LoRA+ by @tastelikefeet in #472
- Support internlm2 1.8b by @Jintao-Huang in #473
- fix deepseek moe device_map by @Jintao-Huang in #476
- fix peft compatible bug by @tastelikefeet in #482
- Fix deepspeed init bug by @Jintao-Huang in #481
- fix scripts docs by @Jintao-Huang in #483
- Update
swift export
and update docs by @Jintao-Huang in #484 - support gptq export by @Jintao-Huang in #485
- fix docs & readme by @Jintao-Huang in #486
- fix app-ui bug by @Jintao-Huang in #488
- Support peft0.9 by @tastelikefeet in #490
- support torchrun_args for dpo cli and support web_ui model deployment by @slin000111 in #496
- Support transformers 4.33.0 by @tastelikefeet in #498
- Update deepspeed config by @Jintao-Huang in #500
- move docs to classroom by @tastelikefeet in #503
- Support yi 9b by @Jintao-Huang in #504
- Update yi sh by @Jintao-Huang in #506
Full Changelog: v1.6.0...v1.7.0