llama_cpp 0.20.0 → 0.20.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
checksums.yaml CHANGED
@@ -1,7 +1,7 @@
1
1
  ---
2
2
  SHA256:
3
- metadata.gz: 33d3b46593a1ee0950c7f86ffe802f5841e8e065bc2bdf7d9679dcff37dbe06b
4
- data.tar.gz: 117eba5fa85437e8cd0bb8767090bf63b4e72e523f96e8cf539a9ddf2cd15195
3
+ metadata.gz: 93de918fd9c6ed1391c9d192c4c5af15c24011e72ed5537303cd276455de264c
4
+ data.tar.gz: fe77379e15cef0a4c2d9fccca22646beb026bfa5339d2dad5ac4a74165328264
5
5
  SHA512:
6
- metadata.gz: '0825dfa498b5a6616bebacb1b7aeec03db58b2ea992cafeebf4fc3e7553fbd2d30c6d14fda80caee8d303b84dfe447429a720add1c028eac17a16778c311d218'
7
- data.tar.gz: 3b29add60fe63985974daa34b2e1d6a198688d83b495e6fe5381ca97074752f526a2900ae0e85c298a89d80f723c662d685fdefd74814ef956bd54567ef83c5b
6
+ metadata.gz: 12c8406b316c7031af7b840aaa79516a42c911ae789b94b7e88e37af0e1162178fa6917a090d9f25d5555bf768d08cbcb528338d403ecbc659cd5fabd58ac761
7
+ data.tar.gz: b8d2a389395f866d9e975eba22b1d4c2746f785ff0694da4b93c5be7b98608e8979efee7010df8c83e54b1348cef26b573a76344d44bd926a4371414cc515856
data/CHANGELOG.md CHANGED
@@ -1,3 +1,8 @@
1
+ ## [[0.20.1](https://github.com/yoshoku/llama_cpp.rb/compare/v0.20.0...v0.20.1)] - 2025-05-31
2
+
3
+ - Change supported llama.cpp version to b5530
4
+ - Add `llama_max_parallel_sequences` module function.
5
+
1
6
  ## [[0.20.0](https://github.com/yoshoku/llama_cpp.rb/compare/v0.19.6...v0.20.0)] - 2025-05-23
2
7
 
3
8
  - Change supported llama.cpp version to b5460
@@ -1322,6 +1322,14 @@ static VALUE rb_llama_max_devices(VALUE self) {
1322
1322
  return SIZET2NUM(llama_max_devices());
1323
1323
  }
1324
1324
 
1325
+ /**
1326
+ * @overload llama_max_parallel_sequences
1327
+ * @return [Integer]
1328
+ */
1329
+ static VALUE rb_llama_max_parallel_sequences(VALUE self) {
1330
+ return SIZET2NUM(llama_max_parallel_sequences());
1331
+ }
1332
+
1325
1333
  /**
1326
1334
  * @overload llama_supports_mmap?
1327
1335
  * @return [Boolean]
@@ -4684,6 +4692,9 @@ void Init_llama_cpp(void) {
4684
4692
  /* llama_max_devices */
4685
4693
  rb_define_module_function(rb_mLlamaCpp, "llama_max_devices", rb_llama_max_devices, 0);
4686
4694
 
4695
+ /* llama_max_parallel_sequences */
4696
+ rb_define_module_function(rb_mLlamaCpp, "llama_max_parallel_sequences", rb_llama_max_parallel_sequences, 0);
4697
+
4687
4698
  /* llama_supports_mmap */
4688
4699
  rb_define_module_function(rb_mLlamaCpp, "llama_supports_mmap?", rb_llama_supports_mmap, 0);
4689
4700
 
@@ -3,8 +3,8 @@
3
3
  # llama_cpp.rb provides Ruby bindings for the llama.cpp.
4
4
  module LlamaCpp
5
5
  # The version of llama_cpp.rb you install.
6
- VERSION = '0.20.0'
6
+ VERSION = '0.20.1'
7
7
 
8
8
  # The supported version of llama.cpp.
9
- LLAMA_CPP_VERSION = 'b5460'
9
+ LLAMA_CPP_VERSION = 'b5530'
10
10
  end
metadata CHANGED
@@ -1,7 +1,7 @@
1
1
  --- !ruby/object:Gem::Specification
2
2
  name: llama_cpp
3
3
  version: !ruby/object:Gem::Version
4
- version: 0.20.0
4
+ version: 0.20.1
5
5
  platform: ruby
6
6
  authors:
7
7
  - yoshoku
@@ -49,7 +49,7 @@ required_rubygems_version: !ruby/object:Gem::Requirement
49
49
  - !ruby/object:Gem::Version
50
50
  version: '0'
51
51
  requirements: []
52
- rubygems_version: 3.6.9
52
+ rubygems_version: 3.6.7
53
53
  specification_version: 4
54
54
  summary: Ruby bindings for the llama.cpp.
55
55
  test_files: []