github ggml-org/llama.cpp b7747

latest releases: b7751, b7750, b7749...
7 hours ago
Details

server: improve slots scheduling for n_cmpl (#18789)

  • server : make sure children tasks are scheduled to launch with parent

  • fix

  • add comment pointing to this PR

  • fix

  • clean up

  • more debug messages

  • add pop_deferred_task with specific ID version

  • improve the logic

  • simple approach

  • no double move

  • correct return type of launch_slots_with_parent_task

macOS/iOS:

Linux:

Windows:

openEuler:

Don't miss a new llama.cpp release

NewReleases is sending notifications on new releases.