github ggml-org/llama.cpp b7991

3 hours ago
Details

[WebGPU] Plug memory leaks and free resources on shutdown (#19315)

  • Fix memory leaks in shader lib, backend, backend_context, buffer_context, and webgpu_buf_pool

  • Free pools

  • Cleanup

  • More cleanup

  • Run clang-format

  • Fix arg-parser and tokenizer test errors that free an unallocated buffer

  • Fix device lost callback to not print on device teardown

  • Fix include and run clang-format

  • remove unused unused

  • Update binary ops


Co-authored-by: Reese Levine reeselevine1@gmail.com

macOS/iOS:

Linux:

Windows:

openEuler:

Don't miss a new llama.cpp release

NewReleases is sending notifications on new releases.