What's Changed
- Print error when converting from an OCI Image by @rhatdan in #932
- Make compatible with the macOS system python3 by @ericcurtin in #933
- Bugfixes noticed while installing on Raspberry Pi by @ericcurtin in #935
- Add note about updating nvidia.yaml file by @rhatdan in #938
- Fix docker handling of GPUs. by @rhatdan in #941
- macOS detection fix by @ericcurtin in #942
- Add chat template support by @engelmi in #917
- Consolidate gpu detection by @ericcurtin in #943
- Implement RamaLama shell by @ericcurtin in #915
- Add Linux x86-64 support for Ascend NPU accelerator in llama.cpp backend by @leo-pony in #950
- Handle CNAI annotation deprecation by @s3rj1k in #939
- Fix install.sh for OSTree system by @ericcurtin in #951
- Lets run container in all tests, to make sure it does not explode. by @rhatdan in #946
- Added --chat-template-file support to ramalama serve by @engelmi in #952
- Update registry.access.redhat.com/ubi9/ubi Docker tag to v9.5-1741850090 by @renovate in #956
- Add specified nvidia-oci runtime by @rhatdan in #953
- python3 validator by @ericcurtin in #959
- There must be at least one CDI device present to use CUDA by @ericcurtin in #954
- [NPU][Fix] only specify device num, but without ascend-docker-runtime installed, running ramalama/cann container image will failing by @leo-pony in #962
- Fix port rendering in README by @andreadecorte in #963
- Update docker.io/nvidia/cuda Docker tag to v12.8.1 by @renovate in #960
- Update llama.cpp to contain threads features by @ericcurtin in #967
- Fix ENTRYPOINTS of whisper-server and llama-server by @rhatdan in #965
- Add software to support using rag in RamaLama by @rhatdan in #968
- Update llama.cpp for some Gemma features by @ericcurtin in #973
- Only set this environment variable if we can resolve CDI by @ericcurtin in #971
- feat(cpu): add --threads option to specify number of cpu threads by @antheas in #966
- Asashi build is failing because of no python3-devel package by @rhatdan in #974
- GPG Check is failing on the Intel Repo by @cgruver in #976
- Add --runtime-arg option for run and serve by @edmcman in #949
- Fix handling of whisper-server and llama-server entrypoints by @rhatdan in #975
- Bump to v0.6.4 by @rhatdan in #978
New Contributors
- @s3rj1k made their first contribution in #939
- @antheas made their first contribution in #966
- @edmcman made their first contribution in #949
Full Changelog: v0.6.3...v0.6.4