What's new in 0.2.0 (2023-08-19)
These are the changes in inference v0.2.0.
New features
- FEAT:Support Starchat-Beta and StarCoderPlus with Pytorch. by @RayJi01 in #333
- FEAT: Support Ctransformers by @RayJi01 in #289
- FEAT: internlm by @UranusSeven in #352
- FEAT: Support Vicuna-v1.5 and Vicuna-v1.5-16k by @RayJi01 in #343
- FEAT: wizardmath by @UranusSeven in #351
- FEAT: support generate/chat/create_embedding/register/unregister/registrations method in cmdline by @pangyoki in #363
Enhancements
- ENH: Use Llama 2 chat for inference in LangChain QA demo by @jiayini1119 in #324
- ENH: cache from URI by @UranusSeven in #350
- ENH: Update System Prompt for llama-2-chat by @Bojun-Feng in #359
- ENH: RESTful client supports custom model APIs by @jiayini1119 in #360
- BLD: fix readthedocs by @UranusSeven in #340
- BLD: fix readthedocs by @UranusSeven in #342
Bug fixes
- BUG: Chatglm max_length doesn't work by @Bojun-Feng in #349
- BUG: builtin stop_token_ids changes by @UranusSeven in #353
- BUG: custom model related bugs by @UranusSeven in #364
Documentation
- DOC: framework by @UranusSeven in #332
- DOC: models by @UranusSeven in #338
- DOC: fix README.md by @UranusSeven in #354
- DOC: update builtin models by @UranusSeven in #365
Others
- FEAT : Add Model Dashboard by @Bojun-Feng in #334
- Revert "FEAT : Add Model Dashboard" by @UranusSeven in #362
Full Changelog: v0.1.3...v0.2.0