3.0.0-beta.13 (2024-03-03)
Bug Fixes
- adapt to
llama.cppbreaking change (#175) (5a70576) - return user-defined llama tokens (#175) (5a70576)
Features
- gguf parser (#168) (bcaab4f)
- use the best compute layer available by default (#175) (5a70576)
- more guardrails to prevent loading an incompatible prebuilt binary (#175) (5a70576)
inspectcommand (#175) (5a70576)GemmaChatWrapper(#175) (5a70576)TemplateChatWrapper(#175) (5a70576)
Shipped with llama.cpp release b2329
To use the latest
llama.cpprelease available, runnpx --no node-llama-cpp download --release latest. (learn more)