github ggml-org/llama.cpp b8152

latest releases: b8157, b8156, b8155...
8 hours ago
Details

server : support multi-modal context checkpoints (#19849)

  • Modify llama-memory-hybrid-iswa.cpp

  • Modify llama-memory-recurrent.cpp

  • Modify server-common.cpp

  • Modify server-common.h

  • Modify server-context.cpp

  • Modify server-task.h

  • Added comment to llama-memory-hybrid-iswa.cpp

  • Remove comment from server-context.cpp

  • Stylistic fix server-context.cpp

  • Fix an issue when seqrm isn't called in server-context.cpp

  • cont : alternative impl

  • cont : cleanup

  • cont : n_tokens -> int64_t


Co-authored-by: timkhronos timkhronos@gmail.com

macOS/iOS:

Linux:

Windows:

openEuler:

Don't miss a new llama.cpp release

NewReleases is sending notifications on new releases.