github ggml-org/llama.cpp b8003

latest releases: b8005, b8004
2 hours ago
Details

model: Add Kimi-K2.5 support (#19170)

  • Move dequant_model to after the text_config merge
    Add new kimi-k2.5 keys to mtmd convert
    Update V_MMPROJ tensor mapping for new mm_projector.proj keys
    Update V_M_IMP_NORM for new mm_projector.pre_norm key

  • Fix a couple of oversights

  • Add image support for Kimi-K2.5

  • Revert changes to KimiVLForConditionalGeneration

  • Fix an assert crash

  • Fix permute swapping w / h on accident

  • Kimi-K2.5: Use merged QKV for vision

  • Kimi-K2.5: pre-convert vision QK to use build_rope_2d

  • Kimi-K2.5: support non-interleaved rope for vision

  • Kimi-K2.5: fix min / max pixel

  • Kimi-K2.5: remove v/o permutes, unnecessary

  • Kimi-K2.5: update permute name to match

  • Update convert_hf_to_gguf.py

Co-authored-by: Sigbjørn Skjæret sigbjorn.skjaeret@scala.com

  • Kimi-K2.5: replace build_rope_2d ggml_cont with ggml_view_3d pointers

Co-authored-by: Sigbjørn Skjæret sigbjorn.skjaeret@scala.com

macOS/iOS:

Linux:

Windows:

openEuler:

Don't miss a new llama.cpp release

NewReleases is sending notifications on new releases.