Internals:
- Allow for a
data_seed
argument in #3150 - Trigger
weights_only=True
by default for all compatible objects when checkpointing and saving withtorch.save
in #3036 - Handle negative values for
dim
input inpad_across_processes
in #3114 - Enable cpu bnb distributed lora finetune in #3159
DeepSpeed
- Support torch dynamo for deepspeed>=0.14.4 in #3069
Megatron
- update Megatron-LM plugin code to version 0.8.0 or higher in #3174
Big Model Inference
- New
has_offloaded_params
utility added in #3188
Examples
- Florence2 distributed inference example in #3123
Full Changelog
- Handle negative values for
dim
input inpad_across_processes
by @mariusarvinte in #3114 - Fixup DS issue with weakref by @muellerzr in #3143
- Refactor scaler to util by @muellerzr in #3142
- DS fix, continued by @muellerzr in #3145
- Florence2 distributed inference example by @hlky in #3123
- POC: Allow for a
data_seed
by @muellerzr in #3150 - Adding multi gpu speech generation by @dame-cell in #3149
- support torch dynamo for deepspeed>=0.14.4 by @oraluben in #3069
- Fixup Zero3 +
save_model
by @muellerzr in #3146 - Trigger
weights_only=True
by default for all compatible objects by @muellerzr in #3036 - Remove broken dynamo test by @oraluben in #3155
- fix version check bug in
get_xpu_available_memory
by @faaany in #3165 - enable cpu bnb distributed lora finetune by @jiqing-feng in #3159
- [Utils]
has_offloaded_params
by @kylesayrs in #3188 - fix bnb by @eljandoubi in #3186
- [docs] update neptune API by @faaany in #3181
- docs: fix a wrong word in comment in src/accelerate/accelerate.py:1255 by @Rebornix-zero in #3183
- [docs] use nn.module instead of tensor as model by @faaany in #3157
- Fix typo by @kylesayrs in #3191
- MLU devices : Checks if mlu is available via an cndev-based check which won't trigger the drivers and leave mlu by @huismiling in #3187
- update Megatron-LM plugin code to version 0.8.0 or higher. by @eljandoubi in #3174
- 🚨 🚨 🚨 Goodbye Python 3.8! 🚨 🚨 🚨 by @muellerzr in #3194
- Update transformers.deepspeed references from transformers 4.46.0 release by @loadams in #3196
- eliminate dead code by @statelesshz in #3198
- take
torch.nn.Module
model into account when moving to device by @faaany in #3167 - [docs] add xpu part and fix bug in
torchrun
by @faaany in #3166 - Models With Tied Weights Need Re-Tieing After FSDP Param Init by @fabianlim in #3154
- add the missing xpu for local sgd by @faaany in #3163
- typo fix in big_modeling.py by @a-r-r-o-w in #3207
- [Utils]
align_module_device
by @kylesayrs in #3204
New Contributors
- @mariusarvinte made their first contribution in #3114
- @hlky made their first contribution in #3123
- @dame-cell made their first contribution in #3149
- @kylesayrs made their first contribution in #3188
- @eljandoubi made their first contribution in #3186
- @Rebornix-zero made their first contribution in #3183
- @loadams made their first contribution in #3196
Full Changelog: v1.0.1...v1.1.0