github InternLM/lmdeploy v0.0.8
LMDeploy Release V0.0.8

latest releases: v0.10.2, v0.10.1, v0.10.0...
2 years ago

Highlight

  • Support Baichuan2-7B-Base and Baichuan2-7B-Chat
  • Support all features of Code Llama: code completion, infilling, chat / instruct, and python specialist

What's Changed

🚀 Features

🐞 Bug fixes

  • [Fix] when using stream is False, continuous batching doesn't work by @sleepwalker2017 in #346
  • [Fix] Set max dynamic smem size for decoder MHA to support context length > 8k by @lvhan028 in #377
  • Fix exceed session len core dump for chat and generate by @AllentDan in #366
  • [Fix] update puyu model by @Harold-lkk in #399

📚 Documentations

New Contributors

Full Changelog: v0.0.7...v0.0.8

Don't miss a new lmdeploy release

NewReleases is sending notifications on new releases.