github ggml-org/llama.cpp b7795

latest releases: b7801, b7798
2 hours ago
Details

vulkan: Remove transfer_ctx, do everything in compute_ctx. (#18945)

  • vulkan: Remove transfer_ctx, do everything in compute_ctx.

We had a bug where a set_tensor_async (using transfer_ctx) didn't get
submitted before the graph_compute (using compute_ctx) that came after
it. To avoid this sort of issue, just do everything in compute_ctx.

Remove transfer_cmd_pool, which was already unused.

  • fix crash with perf logger

macOS/iOS:

Linux:

Windows:

openEuler:

Don't miss a new llama.cpp release

NewReleases is sending notifications on new releases.