pypi vllm 0.3.0
v0.3.0

latest releases: 0.6.1.post2, 0.6.1.post1, 0.6.1...
7 months ago

Major Changes

  • Experimental multi-lora support
  • Experimental prefix caching support
  • FP8 KV Cache support
  • Optimized MoE performance and Deepseek MoE support
  • CI tested PRs
  • Support batch completion in server

What's Changed

New Contributors

Full Changelog: v0.2.7...v0.3.0

Don't miss a new vllm release

NewReleases is sending notifications on new releases.