v3.0.0-beta.18
Pre-release
Pre-release
3.0.0-beta.18 (2024-05-09)
Bug Fixes
- more efficient max context size finding algorithm (#214) (453c162)
- make embedding-only models work correctly (#214) (453c162)
- perform context shift on the correct token index on generation (#214) (453c162)
- make context loading work for all models on Electron (#214) (453c162)
Features
- split gguf files support (#214) (453c162)
pull
command (#214) (453c162)stopOnAbortSignal
andcustomStopTriggers
onLlamaChat
andLlamaChatSession
(#214) (453c162)checkTensors
parameter onloadModel
(#214) (453c162)- improve Electron support (#214) (453c162)
Shipped with llama.cpp
release b2834
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)