github hpcaitech/ColossalAI v0.1.2
V0.1.2 Released!

latest releases: v0.4.6, v0.4.5, v0.4.4...
2 years ago

Overview

Here are the main improvements of this release:

  1. MOE and BERT models can be trained with ZeRO.
  2. Provide a uniform checkpoint for all kinds of parallelism.
  3. Optimize ZeRO-offload, and improve model scaling.
  4. Design a uniform model memory tracer.
  5. Implement an efficient hybrid Adam (CPU and CUDA kernels).
  6. Improve activation offloading.
  7. Profiler TensorBoard plugin of Beta version.
  8. Refactor pipeline module for closer integration with engine.
  9. Chinese tutorials, WeChat and Slack user groups.

What's Changed

Features

Bug Fix

Unit Testing

Documentation

Model Zoo

  • [model zoo] add activation offload for gpt model by @Gy-Lu in #582

Miscellaneous

  • [logging] polish logger format by @feifeibear in #543
  • [profiler] add MemProfiler by @raejaf in #356
  • [Bot] Synchronize Submodule References by @github-actions in #501
  • [tool] create .clang-format for pre-commit by @BoxiangW in #578
  • [GitHub] Add prefix and label in issue template by @binmakeswell in #652

Full Changelog: v0.1.1...v0.1.2

Don't miss a new ColossalAI release

NewReleases is sending notifications on new releases.