github mudler/LocalAI v2.6.1

latest releases: v2.23.0, v2.22.1, v2.22.0...
9 months ago

This is a patch release containing bug-fixes around parallel request support with llama.cpp models.

What's Changed

Bug fixes 🐛

  • fix(llama.cpp): Enable parallel requests by @tauven in #1616
  • fix(llama.cpp): enable cont batching when parallel is set by @mudler in #1622

Exciting New Features 🎉

  • feat(grpc): backend SPI pluggable in embedding mode by @coyzeng in #1621

👒 Dependencies

Other Changes

New Contributors

Full Changelog: v2.6.0...v2.6.1

Don't miss a new LocalAI release

NewReleases is sending notifications on new releases.