github ggml-org/llama.cpp b7429

latest releases: b7442, b7441, b7444...
12 hours ago

Warning

Release Format Update: Linux releases will soon use .tar.gz archives instead of .zip. Please make the necessary changes to your deployment scripts.

Details

model: support GLM4V vision encoder (#18042)

  • convert ok

  • no deepstack

  • less new tensors

  • cgraph ok

  • add mrope for text model

  • faster patch merger

  • add GGML_ROPE_TYPE_MRNORM

  • add support for metal

  • move glm4v do dedicated graph

  • convert: add norm_embd

  • clip: add debugging fn

  • working correctly

  • fix style

  • use bicubic

  • fix mrope metal

  • improve cpu

  • convert to neox ordering on conversion

  • revert backend changes

  • force stop if using old weight

  • support moe variant

  • fix conversion

  • fix convert (2)

  • Update tools/mtmd/clip-graph.h

Co-authored-by: Georgi Gerganov ggerganov@gmail.com

  • process mrope_section on TextModel base class

  • resolve conflict merge


Co-authored-by: Georgi Gerganov ggerganov@gmail.com

macOS/iOS:

Linux:

Windows:

openEuler:

Don't miss a new llama.cpp release

NewReleases is sending notifications on new releases.