Version v0.3.9 Release Today!
What's Changed
Release
- [release] update version (#5833) by Hongxin Liu
Fix
- [Fix] Fix spec-dec Glide LlamaModel for compatibility with transformers (#5837) by Yuanheng Zhao
Shardformer
- [shardformer] Change atol in test command-r weight-check to pass pytest (#5835) by Guangyao Zhang
- Merge pull request #5818 from GuangyaoZhang/command-r by Guangyao Zhang
- [shardformer] upgrade transformers to 4.39.3 (#5815) by flybird11111
- [shardformer] fix modeling of bloom and falcon (#5796) by Hongxin Liu
- [shardformer] fix import (#5788) by Hongxin Liu
Devops
- [devops] Remove building on PR when edited to avoid skip issue (#5836) by Guangyao Zhang
- [devops] fix docker ci (#5780) by Hongxin Liu
Launch
Misc
- [misc] Add dist optim to doc sidebar (#5806) by Edenzzzz
- [misc] update requirements (#5787) by Hongxin Liu
- [misc] fix dist logger (#5782) by Hongxin Liu
- [misc] Accelerate CI for zero and dist optim (#5758) by Edenzzzz
- [misc] update dockerfile (#5776) by Hongxin Liu
Pre-commit.ci
- [pre-commit.ci] auto fixes from pre-commit.com hooks by pre-commit-ci[bot]
- [pre-commit.ci] auto fixes from pre-commit.com hooks by pre-commit-ci[bot]
- [pre-commit.ci] auto fixes from pre-commit.com hooks by pre-commit-ci[bot]
Gemini
- [gemini] quick fix on possible async operation (#5803) by botbw
- [Gemini] Use async stream to prefetch and h2d data moving (#5781) by Haze188
- [gemini] optimize reduce scatter d2h copy (#5760) by botbw
Inference
- [Inference] Fix flash-attn import and add model test (#5794) by Li Xingjian
- [Inference]refactor baichuan (#5791) by Runyu Lu
- Merge pull request #5771 from char-1ee/refactor/modeling by Li Xingjian
- [Inference]Add Streaming LLM (#5745) by yuehuayingxueluo
Test
- [test] fix qwen2 pytest distLarge (#5797) by Guangyao Zhang
- [test] fix chatglm test kit (#5793) by Hongxin Liu
- [test] Fix/fix testcase (#5770) by duanjunwen
Colossalchat
Install
- [install]fix setup (#5786) by flybird11111
Hotfix
- [hotfix] fix testcase in test_fx/test_tracer (#5779) by duanjunwen
- [hotfix] fix llama flash attention forward (#5777) by flybird11111
- [Hotfix] Add missing init file in inference.executor (#5774) by Yuanheng Zhao
Test/ci
Ci/tests
Full Changelog: v0.3.9...v0.3.8