github ggml-org/llama.cpp b8748

latest releases: b8757, b8756, b8755...
9 hours ago
Details

server : ignore --alias when using --models-preset (#21380)

I'm not sure what the purpose of keeping --alias was when using
--models-preset, but the result is really weird, as shown in the
following logs:

$ build/bin/llama-server --models-preset preset.ini --alias "Gemma 4 E4B UD Q8_K_XL"
...
init: using 31 threads for HTTP server
srv   load_models: Loaded 2 cached model presets
srv   load_models: Loaded 1 custom model presets from preset.ini
main: failed to initialize router models: alias 'Gemma 4 E4B UD Q8_K_XL' for model 'angt/test-split-model-stories260K:F32' conflicts with existing model name

So I propose to simply ignore --alias too in this case. With this
commit, the server starts in routing mode correctly.

Signed-off-by: Adrien Gallouët angt@huggingface.co

macOS/iOS:

Linux:

Windows:

openEuler:

Don't miss a new llama.cpp release

NewReleases is sending notifications on new releases.