github containers/ramalama v0.6.4

latest releases: v0.12.2, v0.12.1, v0.12.0...
5 months ago

What's Changed

  • Print error when converting from an OCI Image by @rhatdan in #932
  • Make compatible with the macOS system python3 by @ericcurtin in #933
  • Bugfixes noticed while installing on Raspberry Pi by @ericcurtin in #935
  • Add note about updating nvidia.yaml file by @rhatdan in #938
  • Fix docker handling of GPUs. by @rhatdan in #941
  • macOS detection fix by @ericcurtin in #942
  • Add chat template support by @engelmi in #917
  • Consolidate gpu detection by @ericcurtin in #943
  • Implement RamaLama shell by @ericcurtin in #915
  • Add Linux x86-64 support for Ascend NPU accelerator in llama.cpp backend by @leo-pony in #950
  • Handle CNAI annotation deprecation by @s3rj1k in #939
  • Fix install.sh for OSTree system by @ericcurtin in #951
  • Lets run container in all tests, to make sure it does not explode. by @rhatdan in #946
  • Added --chat-template-file support to ramalama serve by @engelmi in #952
  • Update registry.access.redhat.com/ubi9/ubi Docker tag to v9.5-1741850090 by @renovate in #956
  • Add specified nvidia-oci runtime by @rhatdan in #953
  • python3 validator by @ericcurtin in #959
  • There must be at least one CDI device present to use CUDA by @ericcurtin in #954
  • [NPU][Fix] only specify device num, but without ascend-docker-runtime installed, running ramalama/cann container image will failing by @leo-pony in #962
  • Fix port rendering in README by @andreadecorte in #963
  • Update docker.io/nvidia/cuda Docker tag to v12.8.1 by @renovate in #960
  • Update llama.cpp to contain threads features by @ericcurtin in #967
  • Fix ENTRYPOINTS of whisper-server and llama-server by @rhatdan in #965
  • Add software to support using rag in RamaLama by @rhatdan in #968
  • Update llama.cpp for some Gemma features by @ericcurtin in #973
  • Only set this environment variable if we can resolve CDI by @ericcurtin in #971
  • feat(cpu): add --threads option to specify number of cpu threads by @antheas in #966
  • Asashi build is failing because of no python3-devel package by @rhatdan in #974
  • GPG Check is failing on the Intel Repo by @cgruver in #976
  • Add --runtime-arg option for run and serve by @edmcman in #949
  • Fix handling of whisper-server and llama-server entrypoints by @rhatdan in #975
  • Bump to v0.6.4 by @rhatdan in #978

New Contributors

Full Changelog: v0.6.3...v0.6.4

Don't miss a new ramalama release

NewReleases is sending notifications on new releases.