github rjmalagon/ollama-linux-amd-apu v0.11.5

latest releases: v0.15.6, v0.15.0, v0.14.1...
pre-release6 months ago

What's Changed (this repo branch)

  • Sync to Ollama main v0.11.5
  • Apply @progval linux kernel version detection patches
  • "accommodate" the new Ollama memory allocation updates to the AMD APUs (may be somehow broken)

What's Changed (from Ollama)

What's changed

  • Performance improvements for the gpt-oss models
  • Improved multi-GPU scheduling and reduced VRAM allocation when using more than 2 GPUs
  • Fix error when parsing bad harmony tool calls
  • OLLAMA_FLASH_ATTENTION=1 will also enable flash attention for pure-CPU models
  • Fixed OpenAI-compatible API not supporting reasoning_effort
  • Reduced size of installation on Windows and Linux
  • New Memory Management by @jessegross in ollama#11090
  • openai: allow for content and tool calls in the same message by @drifkin in ollama#11759
  • openai: when converting role=tool messages, propagate the tool name by @drifkin in ollama#11761
  • openai: always provide reasoning by @drifkin in ollama#11765

New ContributorsNew Contributors

Full Changelog: v0.11.3...v0.11.5

Don't miss a new ollama-linux-amd-apu release

NewReleases is sending notifications on new releases.