What's new in 0.7.0 (2023-12-08)
These are the changes in inference v0.7.0.
Enhancements
- ENH: upgrade insecure requests when necessary by @waltcow in #712
- ENH: [UI] Using tab in running models by @ChengjieLi28 in #714
- ENH: [UI] supports launching rerank models by @ChengjieLi28 in #711
- ENH: [UI] Error can be shown on web UI directly via Snackbar by @ChengjieLi28 in #721
- ENH: [UI] Supports
n_gpu
config when launching LLM models on web ui by @ChengjieLi28 in #730 - ENH: [UI]
n_gpu
default valueauto
by @ChengjieLi28 in #738 - ENH: [UI] Support unregistering custom model on web UI by @ChengjieLi28 in #735
- ENH: Auto recover model actor by @codingl2k1 in #694
- ENH: allow rerank models run with LLM models on same device by @aresnow1 in #741
Bug fixes
- BUG: Auto patch trust remote code for embedding model by @codingl2k1 in #710
- BUG: Fix vLLM backend by @codingl2k1 in #728
Others
- Update builtin model list by @onesuper in #709
- Revert "ENH: upgrade insecure requests when necessary" by @qinxuye in #716
- CHORE: Format js file and check js code style by @ChengjieLi28 in #727
New Contributors
Full Changelog: v0.6.5...v0.7.0