What's new in 0.11.1 (2024-05-17)
These are the changes in inference v0.11.1.
New features
- FEAT: support Yi-1.5 series by @qinxuye in #1489
- FEAT: [UI] embedding and rerank support the specified GPU and CPU. by @yiboyasss in #1491
Enhancements
- ENH: Refactoring the LoRa adaptation method for the LLM model. by @hainaweiben in #1470
- ENH: Add stream_options support by @amumu96 in #1508
Bug fixes
- BUG: fix top_k for vllm backend by @sixsun10 in #1461
- BUG: Docker image issue due to
torchvision
by @ChengjieLi28 in #1485 - BUG: Docker image crash during startup due to
llama-cpp-python
by @ChengjieLi28 in #1507 - BUG: Fix prompt is needed when docker image builds by @ChengjieLi28 in #1512
- BUG:
llama.cpp
model failed when chat due tolora
by @ChengjieLi28 in #1513
Documentation
- DOC: update quick start ipynb by @qinxuye in #1482
- DOC: Update readme for being integrated by RAGFlow by @JinHai-CN in #1493
- DOC: Lora usage by @ChengjieLi28 in #1506
New Contributors
- @sixsun10 made their first contribution in #1461
- @JinHai-CN made their first contribution in #1493
Full Changelog: v0.11.0...v0.11.1