Overview
Minor maintenance release:
- Fix CUDA issues where the transcription produces garbage
- FIX quantized models to work with CUDA backend
- Allow to use
whisper.cpp
andllama.cpp
together in SwiftUI projects
What's Changed
- Update bench.py by @ForkedInTime in #1655
- cmake : Resolve quantized model issue when CUBLAS enabled by @bobqianic in #1667
- examples : Revert CMakeLists.txt for talk-llama by @bobqianic in #1669
- CI : Add coverage for talk-llama when WHISPER_CUBLAS=1 by @bobqianic in #1672
- ci: build and push docker image by @OpenWaygate in #1674
- sync : ggml (ggml_scale, ggml_row_size, etc.) by @ggerganov in #1677
- Replace
WHISPER_PRINT_DEBUG
withWHISPER_LOG_DEBUG
by @bobqianic in #1681 - download: Fix large q5 model name by @dimopep in #1695
- sync : ggml (VMM, sync-ggml-am.sh, dotprod ARM fixes) by @ggerganov in #1691
- whisper : replace
tensor->n_dims
withggml_n_dims(tensor)
by @bobqianic in #1694 - Build with CLBlast by @tamo in #1576
- docker : Fix the Publishing of the CUDA Docker Image by @bobqianic in #1704
- emscripten: fix "Stack Overflow!" by @Huguet57 in #1713
- sync : ggml by @ggerganov in #1717
- Add error handling to graph_compute by @finnvoor in #1714
- Updates Package.swift to use ggml as package dependency by @1-ashraful-islam in #1701
New Contributors
- @ForkedInTime made their first contribution in #1655
- @OpenWaygate made their first contribution in #1674
- @dimopep made their first contribution in #1695
- @Huguet57 made their first contribution in #1713
- @1-ashraful-islam made their first contribution in #1701
Full Changelog: v1.5.2...v1.5.3