v0.8.3-beta5
BREAKING FEATURE: Llama.cpp no longer supports Q4_0_4_4, Q4_0_4_8 and Q4_0_8_8. Instead, Q4_0 will automatically be aligned for ARM optimized kernels, use Q4_0 from now on.
Features:
- Reimplemented Horde for the new API system.
Changes:
- Updated Sliders which should fix buggy behavior.
- This should also fix tapping out of the slider textbox not updating the value.
- Sync llama.cpp:
- This removes support for Q4_0_4_4, Q4_0_4_8 and Q4_0_8_8. These models should now remain on Q4_0 only.
- ChatML is now the default instruct format on first install
- Editing chat should automatically scroll to bottom
Fixes:
- Legacy API system not working correctly.
- Incorrect Gemma2 Instruct format.