github vllm-project/vllm v0.5.0.post1

15 days ago

Highlights

  • Add initial TPU integration (#5292)
  • Fix crashes when using FlashAttention backend (#5478)
  • Fix issues when using num_devices < num_available_devices (#5473)

What's Changed

New Contributors

Full Changelog: v0.5.0...v0.5.0.post1

Don't miss a new vllm release

NewReleases is sending notifications on new releases.