pypi pytorch-lightning 1.5.2
Standard weekly patch release

latest releases: 2.4.0, 2.3.3, 2.3.2...
2 years ago

[1.5.2] - 2021-11-16

Fixed

  • Fixed CombinedLoader and max_size_cycle didn't receive a DistributedSampler (#10374)
  • Fixed an issue where class or init-only variables of dataclasses were passed to the dataclass constructor in utilities.apply_to_collection (#9702)
  • Fixed isinstance not working with init_meta_context, materialized model not being moved to the device (#10493)
  • Fixed an issue that prevented the Trainer to shutdown workers when execution is interrupted due to failure(#10463)
  • Squeeze the early stopping monitor to remove empty tensor dimensions (#10461)
  • Fixed sampler replacement logic with overfit_batches to only replace the sample when SequentialSampler is not used (#10486)
  • Fixed scripting causing false positive deprecation warnings (#10470, #10555)
  • Do not fail if batch size could not be inferred for logging when using DeepSpeed (#10438)
  • Fixed propagation of device and dtype information to submodules of LightningLite when they inherit from DeviceDtypeModuleMixin (#10559)

Contributors

@a-gardner1 @awaelchli @carmocca @justusschock @Raahul-Singh @rohitgr7 @SeanNaren @tchaton

If we forgot someone due to not matching commit email with GitHub account, let us know :]

Don't miss a new pytorch-lightning release

NewReleases is sending notifications on new releases.