Changes
- Characters can now think in
chat-instructmode! This was possible thanks to many simplifications and improvements to jinja2 template handling:
- Add support for the Seed-OSS-36B-Instruct template.
- Better handle the growth of the chat input textarea:
| Before | After |
|---|---|
|
|
- Make the
--modelflag work with absolute paths for gguf models, like--model /tmp/gemma-3-270m-it-IQ4_NL.gguf - Make venv portable installs work with Python 3.13
- Optimize LaTeX rendering during streaming for long replies
- Give streaming instruct messages more vertical space
- Preload the instruct and chat fonts for smoother startup
- Improve right sidebar borders in light mode
- Remove the
--flash-attnflag (it's always on now in llama.cpp) - Suppress "Attempted to select a non-interactive or hidden tab" console warnings, reducing the UI CPU usage during streaming
- Statically link MSVC runtime to remove the Visual C++ Redistributable dependency on Windows for the llama.cpp binaries
- Make the llama.cpp terminal output with
--verboseless verbose
Bug fixes
- llama.cpp: Fix stderr deadlock while loading some models
- llama.cpp: Fix obtaining the maximum sequence length for GPT-OSS
- Fix the UI failing to launch if the Notebook prompt is too long
- Fix LaTeX rendering for equations with asterisks
- Fix italic and quote colors in headings
Backend updates
- Update llama.cpp to https://github.com/ggml-org/llama.cpp/tree/9961d244f2df6baf40af2f1ddc0927f8d91578c8
Portable builds
Below you can find self-contained packages that work with GGUF models (llama.cpp) and require no installation! Just download the right version for your system, unzip, and run.
Which version to download:
-
Windows/Linux:
- NVIDIA GPU: Use
cuda12.4for newer GPUs orcuda11.7for older GPUs and systems with older drivers. - AMD/Intel GPU: Use
vulkanbuilds. - CPU only: Use
cpubuilds.
- NVIDIA GPU: Use
-
Mac:
- Apple Silicon: Use
macos-arm64. - Intel CPU: Use
macos-x86_64.
- Apple Silicon: Use
Updating a portable install:
- Download and unzip the latest version.
- Replace the
user_datafolder with the one in your existing install. All your settings and models will be moved.

