What's Changed
Summary:
- Added support for vulkan dylibs
- Added windows build scripts -> we now have cuda, rocm, vulkan both as .so and as .dll libraries
- Updated llama.cpp submodule to 5e9c63546 -> we now have llama.cpp support for new models (e.g. gemma-4, bonsai, Qwen3.6) and functionalities (e.g. llama.cpp internal tools for agents)
Details:
- Fix README_0.10.0 by @aittalam in #918
- Add support for vulkan dylibs by @aittalam in #892
- Add tinyblasStrsmBatched kernel by @aittalam in #923
- Fix: GGUF Q5_1 quant crashes llamafile on aarch64 cpu by @aittalam in #928
- fix broken link 'feel free to choose' -> example llamafiles by @bquast in #927
- Fix block-size assumption by @aittalam in #935
- Add windows build scripts for CUDA by @aittalam in #924
- Fix whisperfile documentation link in README by @martin0258 in #939
- Update llama.cpp submodule to 5e9c63546 by @aittalam in #941
- Migrate docs from MkDocs/GitHub Pages to GitBook by @angpt in #946
- Win build improvements by @aittalam in #940
- Fix
cuda: /lib/x86_64-linux-gnu/libstdc++.so.6: versionGLIBCXX_3.4.32' not found` by @aittalam in #948 - Docs Updates by @angpt in #949
- New release version: update version.h and docs by @aittalam in #952
New Contributors
- @bquast made their first contribution in #927
- @martin0258 made their first contribution in #939
- @angpt made their first contribution in #946
Full Changelog: 0.10.0...0.10.1