Major features
- Add OpenAI-compatible API server (Completion and ChatCompletion)
- Fix
sgl.select
All PRs
- Support v1/chat/completions by @comaniac in #50
- Fix select and normalized logprobs by @merrymercy in #67
- Bump version to 0.1.5 by @merrymercy in #33
- Use HTTP link in 3rdparty module by @comaniac in #42
- Document sampling parameters by @merrymercy in #45
- Increase interpreter parallelism by @merrymercy in #46
- Add a llava example by @merrymercy in #47
- Support stream=True in v1/completions by @comaniac in #49
- Format code & Improve readme by @merrymercy in #52
- Fix the possible bug of decode out of memory by @hnyls2002 in #36
- Improve error message & Add vicuna template by @merrymercy in #57
- Update README.md by @eltociear in #58
- Disk FSM cache and adjust code. by @hnyls2002 in #63
- Fix select by @merrymercy in #64
- Bump version to 0.1.6 by @merrymercy in #68
New Contributors
- @comaniac made their first contribution in #42
- @eltociear made their first contribution in #58
Full Changelog: v0.1.5...v0.1.6