github mostlygeek/llama-swap v131

latest releases: v157, v156, v155...
2 months ago

Worth writing a bit more about this release and #174.

In this release llama-swap model's gain the ability to prevent clients from changing server default sampling parameters like temperature, top_p, top_k, etc in requests.

This is done through a new filter configuration on models, it looks like this:

models: 
  model1: 
    cmd:  ... 

    # filters: a dictionary of filter settings
    # - optional, default: empty dictionary
    filters:
      # strip_params: a comma separated list of parameters to remove from the request
      # - optional, default: ""
      # - useful for preventing overriding of default server params by requests
      # - `model` parameter is never removed
      # - can be any JSON key in the request body
      # - recommended to stick to sampling parameters
      strip_params: "temperature, top_p, top_k"

Changelog

Don't miss a new llama-swap release

NewReleases is sending notifications on new releases.