llama_cpp 0.17.7 → 0.17.8

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
checksums.yaml CHANGED
@@ -1,7 +1,7 @@
1
1
  ---
2
2
  SHA256:
3
- metadata.gz: ad273fdefc8934e4fdb11f18f20379cd77e89ed9d7e1db84ea4d6d847a676540
4
- data.tar.gz: ad91a3d8589e66c8ba3fb7e12a354d917fb0aa9596fd75ae773db79eace59cb5
3
+ metadata.gz: d4283cb48aef0a09e98f7e065a4e1f7fdd9532a334c7e184149d4b9ff4593927
4
+ data.tar.gz: 92db5cafd75f507c8748548ecf12697356ba4b8de08a0f6f9fdb9ee790757883
5
5
  SHA512:
6
- metadata.gz: 733d18dc4a3c7a72a7de1a4df331f5861d0d2321baf7d439c0f581ceec6ad7246b5dda45739b5154b9ec94c61a094915230506292b48de232f01840b29515f1f
7
- data.tar.gz: 745e2d6f265f72871c258a52af907ecc616bdc2bf1f9dc8fc10d1d81597c930c0a999d4a518fe7f3a6d807df90fee8e559e1ef31e19f03f3d81cfa84d6105b9a
6
+ metadata.gz: 9d18d4bbfe834cfb4c16a86bc231421fc038a5560cb7a2f5fb436db89a418029373057da6a2ad2331830b970df0a90453d4a8da9181ce39dadcb6957172f0d1a
7
+ data.tar.gz: 3976c8a29227e9cd37e91dee00b9a9ad7fcad3716ba3cda9c6f15773de2ec8ee112d00149717ac89c10c0a198f06e046d9be6cc2cbb1ef45e8fac5403ed4dcab
data/CHANGELOG.md CHANGED
@@ -1,4 +1,10 @@
1
- ## [[0.17.7](https://github.com/yoshoku/llama_cpp.rb/compare/v0.17.5...v0.17.6)] - 2024-08-17
1
+ ## [[0.17.8](https://github.com/yoshoku/llama_cpp.rb/compare/v0.17.7...v0.17.8)] - 2024-08-25
2
+
3
+ - Change supported llama.cpp version to b3614.
4
+ - Add `LLAMA_VOCAB_PRE_TYPE_EXAONE` constant.
5
+ - Add `is_recurrent?` method to `Model`.
6
+
7
+ ## [[0.17.7](https://github.com/yoshoku/llama_cpp.rb/compare/v0.17.6...v0.17.7)] - 2024-08-17
2
8
 
3
9
  - Change supported llama.cpp version to b3590.
4
10
  - Add `LLAMA_VOCAB_PRE_TYPE_BLOOM` and `LLAMA_VOCAB_PRE_TYPE_GPT3_FINNISH` constants
@@ -1555,6 +1555,7 @@ public:
1555
1555
  rb_define_method(rb_cLLaMAModel, "has_encoder?", RUBY_METHOD_FUNC(_llama_model_has_encoder), 0);
1556
1556
  rb_define_method(rb_cLLaMAModel, "has_decoder?", RUBY_METHOD_FUNC(_llama_model_has_decoder), 0);
1557
1557
  rb_define_method(rb_cLLaMAModel, "decoder_start_token", RUBY_METHOD_FUNC(_llama_model_decoder_start_token), 0);
1558
+ rb_define_method(rb_cLLaMAModel, "is_recurrent?", RUBY_METHOD_FUNC(_llama_model_is_recurrent), 0);
1558
1559
  rb_define_method(rb_cLLaMAModel, "detokenize", RUBY_METHOD_FUNC(_llama_model_detokenize), -1);
1559
1560
  }
1560
1561
 
@@ -1914,6 +1915,11 @@ private:
1914
1915
  return INT2NUM(llama_model_decoder_start_token(ptr->model));
1915
1916
  }
1916
1917
 
1918
+ static VALUE _llama_model_is_recurrent(VALUE self) {
1919
+ LLaMAModelWrapper* ptr = get_llama_model(self);
1920
+ return llama_model_is_recurrent(ptr->model) ? Qtrue : Qfalse;
1921
+ }
1922
+
1917
1923
  static VALUE _llama_model_detokenize(int argc, VALUE* argv, VALUE self) {
1918
1924
  VALUE kw_args = Qnil;
1919
1925
  ID kw_table[2] = { rb_intern("remove_special"), rb_intern("unparse_special") };
@@ -3631,6 +3637,7 @@ extern "C" void Init_llama_cpp(void) {
3631
3637
  rb_define_const(rb_mLLaMACpp, "LLAMA_VOCAB_PRE_TYPE_CODESHELL", INT2NUM(LLAMA_VOCAB_PRE_TYPE_CODESHELL));
3632
3638
  rb_define_const(rb_mLLaMACpp, "LLAMA_VOCAB_PRE_TYPE_BLOOM", INT2NUM(LLAMA_VOCAB_PRE_TYPE_BLOOM));
3633
3639
  rb_define_const(rb_mLLaMACpp, "LLAMA_VOCAB_PRE_TYPE_GPT3_FINNISH", INT2NUM(LLAMA_VOCAB_PRE_TYPE_GPT3_FINNISH));
3640
+ rb_define_const(rb_mLLaMACpp, "LLAMA_VOCAB_PRE_TYPE_EXAONE", INT2NUM(LLAMA_VOCAB_PRE_TYPE_EXAONE));
3634
3641
 
3635
3642
  rb_define_const(rb_mLLaMACpp, "LLAMA_TOKEN_TYPE_UNDEFINED", INT2NUM(LLAMA_TOKEN_TYPE_UNDEFINED));
3636
3643
  rb_define_const(rb_mLLaMACpp, "LLAMA_TOKEN_TYPE_NORMAL", INT2NUM(LLAMA_TOKEN_TYPE_NORMAL));
@@ -3,8 +3,8 @@
3
3
  # llama_cpp.rb provides Ruby bindings for the llama.cpp.
4
4
  module LLaMACpp
5
5
  # The version of llama_cpp.rb you install.
6
- VERSION = '0.17.7'
6
+ VERSION = '0.17.8'
7
7
 
8
8
  # The supported version of llama.cpp.
9
- LLAMA_CPP_VERSION = 'b3590'
9
+ LLAMA_CPP_VERSION = 'b3614'
10
10
  end
data/sig/llama_cpp.rbs CHANGED
@@ -42,6 +42,7 @@ module LLaMACpp
42
42
  LLAMA_VOCAB_PRE_TYPE_CODESHELL: Integer
43
43
  LLAMA_VOCAB_PRE_TYPE_BLOOM: Integer
44
44
  LLAMA_VOCAB_PRE_TYPE_GPT3_FINNISH: Integer
45
+ LLAMA_VOCAB_PRE_TYPE_EXAONE: Integer
45
46
 
46
47
  LLAMA_TOKEN_ATTR_UNDEFINED: Integer
47
48
  LLAMA_TOKEN_ATTR_UNKNOWN: Integer
@@ -197,6 +198,7 @@ module LLaMACpp
197
198
  def has_encoder?: () -> bool
198
199
  def has_decoder?: () -> bool
199
200
  def decoder_start_token: () -> Integer
201
+ def is_recurrent?: () -> bool
200
202
  def detokenize: (Array[Integer], ?remove_special: bool, ?unparse_special: bool) -> String
201
203
  end
202
204
 
metadata CHANGED
@@ -1,14 +1,14 @@
1
1
  --- !ruby/object:Gem::Specification
2
2
  name: llama_cpp
3
3
  version: !ruby/object:Gem::Version
4
- version: 0.17.7
4
+ version: 0.17.8
5
5
  platform: ruby
6
6
  authors:
7
7
  - yoshoku
8
8
  autorequire:
9
9
  bindir: exe
10
10
  cert_chain: []
11
- date: 2024-08-17 00:00:00.000000000 Z
11
+ date: 2024-08-25 00:00:00.000000000 Z
12
12
  dependencies: []
13
13
  description: llama_cpp.rb provides Ruby bindings for the llama.cpp.
14
14
  email: