github ggml-org/llama.cpp b9156

latest release: b9158
2 hours ago
Details

ggml-webgpu: Enable NVIDIA self-hosted CI (#22976)

  • Enabel nvidia ci for webgpu

  • Address precision issues

  • fix placement

  • Relax more set_rows and div

  • Try relaxing all f16

  • formatting and naming

  • Add comment explaining max_nmse_err logic

Added comment referencing pull request for clarification.

macOS/iOS:

Linux:

Android:

Windows:

openEuler:

Don't miss a new llama.cpp release

NewReleases is sending notifications on new releases.