This is a patch release to address few regressions from the last release and the upcoming Gemma4, most importantly:
- Fixes Gemma 4 tokenization with llama.cpp
- Show login in api key only mode
- Small fixes to improve Anthropic API compatibility
What's Changed
Other Changes
- docs: Update Home Assistant integrations list by @loryanstrant in #9206
- chore: ⬆️ Update ggml-org/llama.cpp to
a1cfb645307edc61a89e41557f290f441043d3c2by @localai-bot in #9203 - chore(model gallery): 🤖 add 1 new models via gallery agent by @localai-bot in #9210
- chore: bump inference defaults from unsloth by @github-actions[bot] in #9219
- docs: ⬆️ update docs version mudler/LocalAI by @localai-bot in #9214
- chore: ⬆️ Update ggml-org/llama.cpp to
d006858316d4650bb4da0c6923294ccd741caefdby @localai-bot in #9215 - fix(ui): pass by staticApiKeyRequired to show login when only api key is configured by @mudler in #9220
- feat(gemma4): add thinking support by @mudler in #9221
- fix(nats): improve error handling by @mudler in #9222
- feat(autoparser): prefer chat deltas from backends when emitted by @mudler in #9224
- fix(anthropic): show null index when not present, default to 0 by @mudler in #9225
- feat(api): Allow coding agents to interactively discover how to control and configure LocalAI by @richiejp in #9084
- chore(refactor): use interface by @mudler in #9226
- fix(reasoning): accumulate and strip reasoning tags from autoparser results by @mudler in #9227
- chore(model-gallery): ⬆️ update checksum by @localai-bot in #9233
- chore: ⬆️ Update ggml-org/llama.cpp to
b8635075ffe27b135c49afb9a8b5c434bd42c502by @localai-bot in #9231
New Contributors
- @github-actions[bot] made their first contribution in #9219
Full Changelog: v4.1.0...v4.1.1