github ggml-org/llama.cpp b8030

latest releases: b8033, b8032
2 hours ago
Details

CUDA: Do not mutate cgraph for fused ADDs (#19566)

  • Do not mutate cgraph for fused ADDs
  1. We should try to minimize in-place changes to the incoming
    ggml_cgraph where possible (those should happen in graph_optimize)
  2. Modifying in-place leads to an additional, unnecessary graph capture
    step as we store the properties before modifying the graph in-place
    in the cuda-backend
  • Assert ggml_tensor is trivially copyable

  • Update ggml/src/ggml-cuda/ggml-cuda.cu

Co-authored-by: Aman Gupta amangupta052@gmail.com


Co-authored-by: Aman Gupta amangupta052@gmail.com

macOS/iOS:

Linux:

Windows:

openEuler:

Don't miss a new llama.cpp release

NewReleases is sending notifications on new releases.