Releases: withcatai/node-llama-cpp
v3.0.0-beta.23
3.0.0-beta.23 (2024-06-09)
Bug Fixes
Features
- parallel function calling (#225) (95f4645)
- preload prompt (#225) (95f4645)
- prompt completion engine (#225) (95f4645)
- chat wrapper based system message support (#225) (95f4645)
- add prompt completion to the Electron example (#225) (95f4645)
- model compatibility warnings (#225) (95f4645)
- Functionary
v2.llama3
support (#225) (95f4645) - parallel function calling with plain Llama 3 Instruct (#225) (95f4645)
- improve function calling support for default chat wrapper (#225) (95f4645)
- parallel model downloads (#225) (95f4645)
- improve the electron example (#225) (95f4645)
customStopTriggers
forLlamaCompletion
(#225) (95f4645)- improve loading status in the Electron example (#226) (4ea0c3c)
Shipped with llama.cpp
release b3091
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)
v2.8.11
v3.0.0-beta.22
3.0.0-beta.22 (2024-05-19)
Bug Fixes
Shipped with llama.cpp
release b2929
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)
v3.0.0-beta.21
3.0.0-beta.21 (2024-05-19)
Bug Fixes
Shipped with llama.cpp
release b2929
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)
v3.0.0-beta.20
3.0.0-beta.20 (2024-05-19)
Bug Fixes
Features
init
command to scaffold a new project from a template (withnode-typescript
andelectron-typescript-react
templates) (#217) (d6a0f43)- debug mode (#217) (d6a0f43)
- load LoRA adapters (#217) (d6a0f43)
- improve Electron support (#217) (d6a0f43)
Shipped with llama.cpp
release b2928
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)
v3.0.0-beta.19
3.0.0-beta.19 (2024-05-12)
Bug Fixes
Features
Shipped with llama.cpp
release b2861
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)
v3.0.0-beta.18
3.0.0-beta.18 (2024-05-09)
Bug Fixes
- more efficient max context size finding algorithm (#214) (453c162)
- make embedding-only models work correctly (#214) (453c162)
- perform context shift on the correct token index on generation (#214) (453c162)
- make context loading work for all models on Electron (#214) (453c162)
Features
- split gguf files support (#214) (453c162)
pull
command (#214) (453c162)stopOnAbortSignal
andcustomStopTriggers
onLlamaChat
andLlamaChatSession
(#214) (453c162)checkTensors
parameter onloadModel
(#214) (453c162)- improve Electron support (#214) (453c162)
Shipped with llama.cpp
release b2834
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)
v2.8.10
v3.0.0-beta.17
3.0.0-beta.17 (2024-04-24)
Bug Fixes
FunctionaryChatWrapper
bugs (#205) (ef501f9)- function calling syntax bugs (#205) ([ef501f9]
- show
GPU layers
in theModel
line in CLI commands (#205) ([ef501f9] - refactor: rename
LlamaChatWrapper
toLlama2ChatWrapper
Features
- Llama 3 support (#205) (ef501f9)
--gpu
flag in generation CLI commands (#205) (ef501f9)specialTokens
parameter onmodel.detokenize
(#205) (ef501f9)
Shipped with llama.cpp
release b2717
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)
v3.0.0-beta.16
3.0.0-beta.16 (2024-04-13)
Bug Fixes
Features
inspect gpu
command: print device names (#198) (5ca33c7)inspect gpu
command: print env info (#202) (d332b77)- download models using the CLI (#191) (b542b53)
- interactively select a model from CLI commands (#191) (b542b53)
- change the default log level to warn (#191) (b542b53)
- token biases (#196) (3ad4494)
Shipped with llama.cpp
release b2665
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)