What's Changed
- Completely revamp model loading to support explicit unload/reload (#1969)
- We no longer load a model by default on application start
- We no longer load a model by default on chat context switch
- Save and restore of window geometry across application starts (#1989)
- Update to latest llama.cpp as of 2/21/2024 and add CPU/GPU support for Gemma (#1992)
- Also enable Vulkan GPU support for Phi and Phi-2, Qwen2, and StableLM
Fixes
- Fix visual artifact in update reminder dialog (16927d9)
- Blacklist Intel GPUs as they are still not supported (a1471be, nomic-ai/llama.cpp#14)
- Improve chat save/load speed (excluding startup/shutdown with defaults) (6fdec80, nomic-ai/llama.cpp#15)
- Significantly improve handling of chat-style prompt templates, and reupload Mistral OpenOrca (#1970, #1993)
New Contributors
Full Changelog: v2.7.0...v2.7.1