What's Changed
Release
- [release] update version (#4775) by Hongxin Liu
Inference
Feature
- [feature] add gptq for inference (#4754) by Xu Kai
- [Feature] The first PR to Add TP inference engine, kv-cache manager and related kernels for our inference system (#4577) by Cuiqing Li
Bug
- [bug] Fix the version check bug in colossalai run when generating the cmd. (#4713) by littsk
- [bug] fix get_default_parser in examples (#4764) by Baizhou Zhang
Lazy
- [lazy] support torch 2.0 (#4763) by Hongxin Liu
Chat
- [chat]: add lora merge weights config (#4766) by Wenhao Chen
- [chat]: update rm, add wandb and fix bugs (#4471) by Wenhao Chen
Doc
- [doc] add shardformer doc to sidebar (#4768) by Baizhou Zhang
- [doc] clean up outdated docs (#4765) by Hongxin Liu
- Merge pull request #4757 from ppt0011/main by ppt0011
- [doc] put native colossalai plugins first in description section by Pengtai Xu
- [doc] add model examples for each plugin by Pengtai Xu
- [doc] put individual plugin explanation in front by Pengtai Xu
- [doc] explain suitable use case for each plugin by Pengtai Xu
- [doc] explaination of loading large pretrained models (#4741) by Baizhou Zhang
- [doc] polish shardformer doc (#4735) by Baizhou Zhang
- [doc] add shardformer support matrix/update tensor parallel documents (#4728) by Baizhou Zhang
- [doc] Add user document for Shardformer (#4702) by Baizhou Zhang
- [doc] fix llama2 code link (#4726) by binmakeswell
- [doc] add potential solution for OOM in llama2 example (#4699) by Baizhou Zhang
- [doc] Update booster user documents. (#4669) by Baizhou Zhang
Shardformer
- [shardformer] fix master param sync for hybrid plugin/rewrite unwrapping logic (#4758) by Baizhou Zhang
- [shardformer] add custom policy in hybrid parallel plugin (#4718) by Xuanlei Zhao
- [shardformer] update seq parallel document (#4730) by Bin Jia
- [shardformer] update pipeline parallel document (#4725) by flybird11111
- [shardformer] to fix whisper test failed due to significant accuracy differences. (#4710) by flybird11111
- [shardformer] fix GPT2DoubleHeadsModel (#4703) by flybird11111
- [shardformer] update shardformer readme (#4689) by flybird11111
- [shardformer]fix gpt2 double head (#4663) by flybird11111
- [shardformer] update llama2/opt finetune example and fix llama2 policy (#4645) by flybird11111
- [shardformer] Support customized policy for llamav2 based model with HybridParallelPlugin (#4624) by eric8607242
Misc
- [misc] update pre-commit and run all files (#4752) by Hongxin Liu
Format
- [format] applied code formatting on changed files in pull request 4743 (#4750) by github-actions[bot]
- [format] applied code formatting on changed files in pull request 4726 (#4727) by github-actions[bot]
Legacy
- [legacy] clean up legacy code (#4743) by Hongxin Liu
- Merge pull request #4738 from ppt0011/main by ppt0011
- [legacy] remove deterministic data loader test by Pengtai Xu
- [legacy] move communication and nn to legacy and refactor logger (#4671) by Hongxin Liu
Kernel
- [kernel] update triton init #4740 (#4740) by Xuanlei Zhao
Example
- [example] llama2 add fine-tune example (#4673) by flybird11111
- [example] add gpt2 HybridParallelPlugin example (#4653) by Bin Jia
- [example] update vit example for hybrid parallel plugin (#4641) by Baizhou Zhang
Hotfix
- [hotfix] Fix import error: colossal.kernel without triton installed (#4722) by Yuanheng Zhao
- [hotfix] fix typo in hybrid parallel io (#4697) by Baizhou Zhang
Devops
- [devops] fix concurrency group (#4667) by Hongxin Liu
- [devops] fix concurrency group and compatibility test (#4665) by Hongxin Liu
Pipeline
- [pipeline] set optimizer to optional in execute_pipeline (#4630) by Baizhou Zhang
Full Changelog: v0.3.3...v0.3.2