pypi accelerate 0.8.0
v0.8.0: Big model inference

latest releases: 1.0.0rc1, 1.0.0rc0, 0.34.2...
2 years ago

v0.8.0: Big model inference

Big model inference

To handle very large models, new functionality has been added in Accelerate:

  • a context manager to initalize empty models
  • a function to load a sharded checkpoint directly on the right devices
  • a set of custom hooks that allow execution of a model split on different devices, as well as CPU or disk offload
  • a magic method that auto-determines a device map for a given model, maximizing the GPU spaces, available RAM before using disk offload as a last resort.
  • a function that wraps the last three blocks in one simple call (load_checkpoint_and_dispatch)

See more in the documentation

What's new

Don't miss a new accelerate release

NewReleases is sending notifications on new releases.