What's new in 0.10.0 (2024-03-29)
These are the changes in inference v0.10.0.
New features
- FEAT: launch UI of audio model. by @hainaweiben in #1102
- FEAT: Supports
OmniLMM
chat model by @hainaweiben in #1171 - FEAT: Added vllm support for deepseek models by @ivanzfb in #1200
- FEAT: force to specify worker ip and gpu idx when launching models by @ChengjieLi28 in #1195
- FEAT: OAuth system supports api-key by @Ago327 in #1168
- FEAT: Support deepseek vl by @codingl2k1 in #1175
- FEAT: support some builtin new models by @mujin2 in #1204
Enhancements
- BLD: add autoawq in setup by @utopia2077 in #1190
Bug fixes
- BUG: Fix the incorrect model interface address caused a 307 redirect to HTTP, blocking the request and preventing the display of the model list. by @wertycn in #1182
- BUG: fix doc fail introduced by #1171 & update readme by @qinxuye in #1203
- BUG: Increase validator types for thie 'input' parameter of embeddings to match OpenAI API by @Minamiyama in #1201
Documentation
- DOC: internal design by @1572161937 in #1178
- Doc: update readme and models doc by @qinxuye in #1176
- DOC: Doc for oauth system with api-key by @ChengjieLi28 in #1210
New Contributors
- @utopia2077 made their first contribution in #1190
- @ivanzfb made their first contribution in #1200
Full Changelog: v0.9.4...v0.10.0