- Initial support for async models. Plugins can now provide an
AsyncModelsubclass that can be accessed in the Python API using the newllm.get_async_model(model_id)method. See async models in the Python API docs and implementing async models in plugins. #507 - OpenAI models all now include async models, so function calls such as
llm.get_async_model("gpt-4o-mini")will return an async model. gpt-4o-audio-previewmodel can be used to send audio attachments to the GPT-4o audio model. #608- Attachments can now be sent without requiring a prompt. #611
llm models --optionsnow includes information on whether a model supports attachments. #612llm models --asyncshows available async models.- Custom OpenAI-compatible models can now be marked as
can_stream: falsein the YAML if they do not support streaming. Thanks, Chris Mungall. #600 - Fixed bug where OpenAI usage data was incorrectly serialized to JSON. #614
- Standardized on
audio/wavMIME type for audio attachments rather thanaudio/wave. [#603](#603