Skip to content

v3.0.0-beta.18

Pre-release
Pre-release
Compare
Choose a tag to compare
@github-actions github-actions released this 09 May 23:28
453c162

3.0.0-beta.18 (2024-05-09)

Bug Fixes

  • more efficient max context size finding algorithm (#214) (453c162)
  • make embedding-only models work correctly (#214) (453c162)
  • perform context shift on the correct token index on generation (#214) (453c162)
  • make context loading work for all models on Electron (#214) (453c162)

Features


Shipped with llama.cpp release b2834

To use the latest llama.cpp release available, run npx --no node-llama-cpp download --release latest. (learn more)