Releases: withcatai/node-llama-cpp
v3.0.0-beta.43
3.0.0-beta.43 (2024-08-09)
Bug Fixes
Shipped with llama.cpp
release b3560
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)
v3.0.0-beta.42
3.0.0-beta.42 (2024-08-07)
Bug Fixes
Shipped with llama.cpp
release b3541
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)
v2.8.15
v3.0.0-beta.41
3.0.0-beta.41 (2024-08-02)
Bug Fixes
Shipped with llama.cpp
release b3504
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)
v3.0.0-beta.40
3.0.0-beta.40 (2024-07-30)
Bug Fixes
Features
Shipped with llama.cpp
release b3488
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)
v3.0.0-beta.39
3.0.0-beta.39 (2024-07-28)
Bug Fixes
- Gemma chat wrapper bug (#273) (e3e0994)
- GGUF metadata nested key conflicts (#273) (e3e0994)
- adapt to
llama.cpp
breaking changes (#273) (e3e0994) - preserve function calling chunks (#273) (e3e0994)
- format JSON objects like models expect (#273) (e3e0994)
Features
- Llama 3.1 support (#273) (e3e0994)
- Phi-3 support (#273) (e3e0994)
- model metadata overrides (#273) (e3e0994)
- use LoRA on a context instead of on a model (#273) (e3e0994)
onTextChunk
option (#273) (e3e0994)
Shipped with llama.cpp
release b3479
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)
v2.8.14
v2.8.13
v3.0.0-beta.38
3.0.0-beta.38 (2024-07-09)
Bug Fixes
- adapt to
llama.cpp
breaking changes (#266) (c35ff5a) - Llama 3 Instruct function calling (#266) (c35ff5a)
Features
Shipped with llama.cpp
release b3347
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)
v3.0.0-beta.37
3.0.0-beta.37 (2024-07-05)
Features
- flash attention (#264) (c2e322c)
- manually exclude specific GPU types from
gpu: "auto"
(#264) (c2e322c)
Shipped with llama.cpp
release b3322
To use the latest
llama.cpp
release available, runnpx --no node-llama-cpp download --release latest
. (learn more)