What's Changed
- Add Distill local-model onboarding with local as the default provider.
- Add automatic local server startup and reuse for local LLM calls.
- Add MLX backend for Apple Silicon and llama.cpp backend for Linux, Windows, and Intel macOS.
- Add runtime auto-install paths and Hugging Face/model download warmup with percentage progress in onboarding.
- Add local concurrency controls, config/env overrides, README documentation, and test coverage.
Verification
- npm run verify
- 114 pass, 1 skip, 0 fail
- build and release check passed