github ggml-org/llama.cpp b8117

6 hours ago
Details

ggml-cpu: add RVV vec dot kernels for quantization types (#18784)

  • ggml-cpu: add rvv vec_dot for iq2_s

Co-authored-by: Rehan Qasim rehan.qasim@10xengineers.ai

  • ggml-cpu: add rvv vec_dot for iq3_s

Co-authored-by: Rehan Qasim rehan.qasim@10xengineers.ai

  • ggml-cpu: add rvv vec_dot for tq1_0, tq2_0

Co-authored-by: Rehan Qasim rehan.qasim@10xengineers.ai

ggml-cpu: add rvv vec_dot for tq1_0, tq2_0

  • ggml-cpu: add rvv vec_dot for iq1_s, iq1_m

Co-authored-by: Rehan Qasim rehan.qasim@10xengineers.ai

  • ggml-cpu: add vlen switch for rvv vec_dot

Co-authored-by: Rehan Qasim rehan.qasim@10xengineers.ai

macOS/iOS:

Linux:

Windows:

openEuler:

Don't miss a new llama.cpp release

NewReleases is sending notifications on new releases.