github ggml-org/llama.cpp b8339

latest release: b8340
2 hours ago
Details

Use fp32 in cuBLAS V100 to avoid overflows, env variables to override cuBLAS compute type (#19959)

  • Update ggml-cuda.cu

  • Update ggml-cuda.cu

  • Update build.md

  • Update build.md

  • Update ggml/src/ggml-cuda/ggml-cuda.cu

Co-authored-by: Johannes Gäßler johannesg@5d6.de

  • Update ggml-cuda.cu

  • Update build.md

  • Update ggml/src/ggml-cuda/ggml-cuda.cu

Co-authored-by: Johannes Gäßler johannesg@5d6.de

  • Update build.md

  • Update ggml-cuda.cu

  • Update ggml-cuda.cu


Co-authored-by: Johannes Gäßler johannesg@5d6.de

macOS/iOS:

Linux:

Windows:

openEuler:

Don't miss a new llama.cpp release

NewReleases is sending notifications on new releases.