pypi vllm 0.1.3
vLLM v0.1.3

latest releases: 0.4.2, 0.4.1, 0.4.0.post1...
9 months ago

What's Changed

Major changes

  • More model support: LLaMA 2, Falcon, GPT-J, Baichuan, etc.
  • Efficient support for MQA and GQA.
  • Changes in the scheduling algorithm: vLLM now uses a TGI-style continuous batching.
  • And many bug fixes.

All changes

New Contributors

Full Changelog: v0.1.2...v0.1.3

Don't miss a new vllm release

NewReleases is sending notifications on new releases.