diff --git a/Cargo.lock b/Cargo.lock index 8ca8430..b216f41 100644 --- a/Cargo.lock +++ b/Cargo.lock @@ -592,7 +592,7 @@ checksum = "da2479e8c062e40bf0066ffa0bc823de0a9368974af99c9f6df941d2c231e03f" [[package]] name = "llama_cpp" -version = "0.1.2" +version = "0.1.3" dependencies = [ "ctor", "derive_more", diff --git a/crates/llama_cpp/CHANGELOG.md b/crates/llama_cpp/CHANGELOG.md index 841297e..3a2a062 100644 --- a/crates/llama_cpp/CHANGELOG.md +++ b/crates/llama_cpp/CHANGELOG.md @@ -5,6 +5,39 @@ All notable changes to this project will be documented in this file. The format is based on [Keep a Changelog](https://keepachangelog.com/en/1.0.0/), and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0.html). +## v0.1.3 (2023-11-08) + +### New Features + + - more `async` function variants + - add `LlamaSession.model` + +### Other + + - typo + +### Commit Statistics + + + + - 5 commits contributed to the release. + - 3 commits were understood as [conventional](https://www.conventionalcommits.org). + - 0 issues like '(#ID)' were seen in commit messages + +### Commit Details + + + +
view details + + * **Uncategorized** + - Typo ([`0a0d5f3`](https://github.com/binedge/llama_cpp-rs/commit/0a0d5f3fce1c46f914b5f48802241f200538c4f7)) + - Release llama_cpp v0.1.2 ([`4d0b130`](https://github.com/binedge/llama_cpp-rs/commit/4d0b130be8f250e599908bab042431db8aa2f553)) + - More `async` function variants ([`1019402`](https://github.com/binedge/llama_cpp-rs/commit/1019402eeaa6bff176a228b477486105d16d36ef)) + - Add `LlamaSession.model` ([`c190df6`](https://github.com/binedge/llama_cpp-rs/commit/c190df6ebfd02ef5f3e0fd50d82a456ef426e6e6)) + - Release llama_cpp_sys v0.2.1, llama_cpp v0.1.1 ([`a9e5813`](https://github.com/binedge/llama_cpp-rs/commit/a9e58133cb1c1d4d45f99a7746e0af7da1a099e1)) +
+ ## v0.1.2 (2023-11-08) ### New Features @@ -16,7 +49,7 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0 - - 2 commits contributed to the release. + - 3 commits contributed to the release. - 2 commits were understood as [conventional](https://www.conventionalcommits.org). - 0 issues like '(#ID)' were seen in commit messages @@ -27,6 +60,7 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0
view details * **Uncategorized** + - Release llama_cpp v0.1.2 ([`368a5de`](https://github.com/binedge/llama_cpp-rs/commit/368a5dec4379ccdbe7b68c40535f30e13f23d8c2)) - More `async` function variants ([`dcfccdf`](https://github.com/binedge/llama_cpp-rs/commit/dcfccdf721eb47a364cce5b1c7a54bcf94335ac0)) - Add `LlamaSession.model` ([`56285a1`](https://github.com/binedge/llama_cpp-rs/commit/56285a119633682951f8748e85c6b8988e514232))
@@ -39,9 +73,14 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0 - Remove debug binary from Cargo.toml +### Chore + + - Remove debug binary from Cargo.toml + ### New Features - add `LlamaModel::load_from_file_async` + - add `LlamaModel::load_from_file_async` ### Bug Fixes @@ -49,14 +88,18 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0 This solves a race condition when several `get_completions` threads are spawned at the same time - `start_completing` should not be invoked on a per-iteration basis There's still some UB that can be triggered due to llama.cpp's threading model, which needs patching up. + - require `llama_context` is accessed from behind a mutex + This solves a race condition when several `get_completions` threads are spawned at the same time + - `start_completing` should not be invoked on a per-iteration basis + There's still some UB that can be triggered due to llama.cpp's threading model, which needs patching up. ### Commit Statistics - - 6 commits contributed to the release. + - 11 commits contributed to the release. - 13 days passed between releases. - - 4 commits were understood as [conventional](https://www.conventionalcommits.org). + - 8 commits were understood as [conventional](https://www.conventionalcommits.org). - 0 issues like '(#ID)' were seen in commit messages ### Commit Details @@ -67,6 +110,11 @@ and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0 * **Uncategorized** - Release llama_cpp_sys v0.2.1, llama_cpp v0.1.1 ([`ef4e3f7`](https://github.com/binedge/llama_cpp-rs/commit/ef4e3f7a3c868a892f26acfae2a5211de4900d1c)) + - Add `LlamaModel::load_from_file_async` ([`bbf9f69`](https://github.com/binedge/llama_cpp-rs/commit/bbf9f69a2dd068a3a20199ffce44d3c8a25b64d5)) + - Remove debug binary from Cargo.toml ([`dbdd9a4`](https://github.com/binedge/llama_cpp-rs/commit/dbdd9a4a2d813d990e5829a09fc5c8df75d9d54b)) + - Require `llama_context` is accessed from behind a mutex ([`81e5de9`](https://github.com/binedge/llama_cpp-rs/commit/81e5de901a3da88a97ba00c6a36e303d8708380d)) + - `start_completing` should not be invoked on a per-iteration basis ([`27706de`](https://github.com/binedge/llama_cpp-rs/commit/27706de1a471b317e4b7b4fdd4c5bbabfbd95ed6)) + - Update to llama.cpp 0a7c980 ([`eb8f627`](https://github.com/binedge/llama_cpp-rs/commit/eb8f62777aa63787004771d86d34a8862b3a4157)) - Add `LlamaModel::load_from_file_async` ([`3bada65`](https://github.com/binedge/llama_cpp-rs/commit/3bada658c9139af1c3dcdb32c60c222efb87a9f6)) - Remove debug binary from Cargo.toml ([`3eddbab`](https://github.com/binedge/llama_cpp-rs/commit/3eddbab3cc35a59acbe66fa4f5333a9ca0edb326)) - Require `llama_context` is accessed from behind a mutex ([`b676baa`](https://github.com/binedge/llama_cpp-rs/commit/b676baa3c1a6863c7afd7a88b6f7e8ddd2a1b9bd)) diff --git a/crates/llama_cpp/Cargo.toml b/crates/llama_cpp/Cargo.toml index c5c2cad..847ece8 100644 --- a/crates/llama_cpp/Cargo.toml +++ b/crates/llama_cpp/Cargo.toml @@ -1,6 +1,6 @@ [package] name = "llama_cpp" -version = "0.1.2" +version = "0.1.3" description = "High-level bindings to llama.cpp with a focus on just being really, really easy to use" edition = "2021" authors = ["Dakota Thompson "]