github mudler/LocalAI v2.6.1

latest releases: v2.20.1, v2.20.0, v2.19.4...
8 months ago

This is a patch release containing bug-fixes around parallel request support with llama.cpp models.

What's Changed

Bug fixes 🐛

  • fix(llama.cpp): Enable parallel requests by @tauven in #1616
  • fix(llama.cpp): enable cont batching when parallel is set by @mudler in #1622

Exciting New Features 🎉

  • feat(grpc): backend SPI pluggable in embedding mode by @coyzeng in #1621

👒 Dependencies

Other Changes

New Contributors

Full Changelog: v2.6.0...v2.6.1

Don't miss a new LocalAI release

NewReleases is sending notifications on new releases.