v0.7.7: Patch release PPO & DDPO tags
A fix has been introduce to fix a breaking change with PPOTrainer.push_to_hub()
and DDPOTrainer.push_to_hub()
- [
PPOTrainer
/DDPOTrainer
] Fix ppo & ddpo push to Hub by @younesbelkada in #1141
What's Changed
- Release: v0.7.6 by @younesbelkada in #1134
- set dev version by @younesbelkada in #1135
- clear up the parameters of supervised_finetuning.py by @sywangyi in #1126
- Add type hints to core.py by @zachschillaci27 in #1097
- fix_ddpo_demo by @zhangsibo1129 in #1129
- Add npu support for ppo example by @zhangsibo1129 in #1128
New Contributors
- @zachschillaci27 made their first contribution in #1097
- @zhangsibo1129 made their first contribution in #1129
Full Changelog: v0.7.6...v0.7.7