What's new in 0.7.1 (2023-12-12)
These are the changes in inference v0.7.1.
Enhancements
- ENH: [UI] Supports
model_uid
input when launching models by @ChengjieLi28 in #746 - ENH: Add more vllm supported models by @aresnow1 in #756
Bug fixes
- BUG: Fix
cached
tag on UI by @ChengjieLi28 in #748 - BUG: Fix stream arg for vllm backend by @aresnow1 in #758
Documentation
Others
- Bugs: Fixing issue with emote encoding in streaming chat, Fixing issue with missing pad_token for pytorch tokenizers, allowing system message as latest message in chat by @AndiMajore in #747
New Contributors
- @AndiMajore made their first contribution in #747
Full Changelog: v0.7.0...v0.7.1