Skip to content

V0.1.2 Released!

Compare
Choose a tag to compare
@ver217 ver217 released this 06 Apr 05:45
· 3400 commits to main since this release
03e1d35

Overview

Here are the main improvements of this release:

  1. MOE and BERT models can be trained with ZeRO.
  2. Provide a uniform checkpoint for all kinds of parallelism.
  3. Optimize ZeRO-offload, and improve model scaling.
  4. Design a uniform model memory tracer.
  5. Implement an efficient hybrid Adam (CPU and CUDA kernels).
  6. Improve activation offloading.
  7. Profiler TensorBoard plugin of Beta version.
  8. Refactor pipeline module for closer integration with engine.
  9. Chinese tutorials, WeChat and Slack user groups.

What's Changed

Features

Bug Fix

Unit Testing

Documentation

Model Zoo

  • [model zoo] add activation offload for gpt model by @Gy-Lu in #582

Miscellaneous

  • [logging] polish logger format by @feifeibear in #543
  • [profiler] add MemProfiler by @raejaf in #356
  • [Bot] Synchronize Submodule References by @github-actions in #501
  • [tool] create .clang-format for pre-commit by @BoxiangW in #578
  • [GitHub] Add prefix and label in issue template by @binmakeswell in #652

Full Changelog: v0.1.1...v0.1.2