github ggml-org/llama.cpp b8308

latest releases: b8322, b8320, b8318...
9 hours ago
Details

vulkan: Fix ErrorOutOfHostMemory on Intel GPU when loading large models with --no-mmap (#20059)

  • Changed to reuse command buffers to fix crashing on Intel GPU

  • Removed unused parameter

  • Fixed compile error and minor mistake

  • Fix logging

  • Changing to use usage flag per command buffer

  • fixed style

  • added buffer reset

  • Removed cmd_buffer_idx for reuse consistency

  • Fixed style

macOS/iOS:

Linux:

Windows:

openEuler:

Don't miss a new llama.cpp release

NewReleases is sending notifications on new releases.