github mudler/LocalAI v2.8.1

latest releases: v2.21.1, v2.21.0, v2.20.1...
7 months ago

This is a patch release, mostly containing minor patches and bugfixes from 2.8.0.

Most importantly it contains a bugfix for #1333 which made the llama.cpp backend to get stuck in some cases where the model starts to hallucinate and fixes to the python-based backends.

Spread the word!

First off, a massive thank you to each and every one of you who've chipped in to squash bugs and suggest cool new features for LocalAI. Your help, kind words, and brilliant ideas are truly appreciated - more than words can say!

And to those of you who've been heros, giving up your own time to help out fellow users on Discord and in our repo, you're absolutely amazing. We couldn't have asked for a better community.

Just so you know, LocalAI doesn't have the luxury of big corporate sponsors behind it. It's all us, folks. So, if you've found value in what we're building together and want to keep the momentum going, consider showing your support. A little shoutout on your favorite social platforms using @LocalAI_OSS and @mudler_it or joining our sponsorship program can make a big difference.

Also, if you haven't yet joined our Discord, come on over! Here's the link: https://discord.gg/uJAeKSAGDy

Every bit of support, every mention, and every star adds up and helps us keep this ship sailing. Let's keep making LocalAI awesome together!

Thanks a ton, and here's to more exciting times ahead with LocalAI! 🚀

What's Changed

Bug fixes 🐛

Exciting New Features 🎉

  • feat(tts): respect YAMLs config file, add sycl docs/examples by @mudler in #1692
  • ci: add cuda builds to release by @sozercan in #1702

Other Changes

Full Changelog: v2.8.0...v2.8.1

Don't miss a new LocalAI release

NewReleases is sending notifications on new releases.