What's new in 0.7.5 (2024-01-05)
These are the changes in inference v0.7.5.
New features
- FEAT: text2vec by @ChengjieLi28 in #857
Enhancements
- ENH: Offload all response serialization to ModelActor by @codingl2k1 in #837
- ENH: Custom model uses vLLM by @ChengjieLi28 in #861
- BLD: Docker image by @ChengjieLi28 in #855
Bug fixes
- BUG: Fix typing_extension version problem in notebook by @onesuper in #856
- BUG: Fix multimodal cmdline by @codingl2k1 in #850
- BUG: Fix generate of chatglm3 by @aresnow1 in #858
Documentation
- DOC: CUDA Version recommendation by @ChengjieLi28 in #841
- DOC: new doc cover by @onesuper in #843
- DOC: Autogen modelhub info by @onesuper in #845
- DOC: Add multimodal feature in README by @onesuper in #846
- DOC: Chinese doc for user guide by @aresnow1 in #847
- DOC: add notebook for quickstart by @onesuper in #854
- DOC: Add docs about environments by @aresnow1 in #853
- DOC: Add jupyter notebook quick start tutorial by @onesuper in #851
Others
- CHORE: Add docker image with
latest
tag by @ChengjieLi28 in #862
Full Changelog: v0.7.4.1...v0.7.5