llama_cpp 0.20.3 → 0.20.4

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
checksums.yaml CHANGED
@@ -1,7 +1,7 @@
1
1
  ---
2
2
  SHA256:
3
- metadata.gz: 23b27d58fb2f7b948a0862df53c9c523981ebde95a1193bfb9b575c5717daa37
4
- data.tar.gz: fc3c5a7d01d96a55cd36f55f730742517f101f3593e2d0d4249b30f8a08c6053
3
+ metadata.gz: 6acca29477a43c9703c7035e53acd69450de7103b2d7f242506c7e2016f1a261
4
+ data.tar.gz: 77c108b2f1ea33588a9dbe6c0538e289f90ea5be2090adbf8f663fff8b8b0221
5
5
  SHA512:
6
- metadata.gz: 0cb14dc2cf2d82d15256efa76bc1c89021f152845ecfdea3e3a2aaa4ca01e497c8c2621fa3263415ab988225e9c5e7f87b1658ef6800db66c499b2b73b228162
7
- data.tar.gz: 8597474f6f983b0514c6d9644d7d6d68775b39533b4c0488960acbecbfc4d0ef6d98f205dc331623c7bf1e6b4e5083fec3d2f879d4b6c2fbc154fb10f9b0ecd1
6
+ metadata.gz: 69d453a3cf9c23ab3aaa60c6a20d80a7fc75424cb762c631a06712a2134fc7cf6830168241a3d689fd0b7b621804e27b6461415977fb51c096c10fbb2aa0e922
7
+ data.tar.gz: bc8de61663616ffd40c90e34df71095a7c85e3aa3373fc0c395fe101ded4f38e1670af42aeb11c6380c6932d2939d137d299d377db3ce9dc0510fd56e9d8b7a5
data/CHANGELOG.md CHANGED
@@ -1,5 +1,11 @@
1
+ ## [[0.20.4](https://github.com/yoshoku/llama_cpp.rb/compare/v0.20.3...v0.20.4)] - 2025-06-21
2
+
3
+ - Change supported llama.cpp version to b5720.
4
+ - Add `llama_vocab_get_add_sep` module function.
5
+
1
6
  ## [[0.20.3](https://github.com/yoshoku/llama_cpp.rb/compare/v0.20.2...v0.20.3)] - 2025-06-14
2
7
 
8
+
3
9
  - Change supported llama.cpp version to b5650
4
10
  - Add `data` argument to `llama_memory_clear` module function.
5
11
  - Fix llama_memory_t wrapper by removing unnecessary struct keyword and pointer symbol.
@@ -2972,6 +2972,22 @@ static VALUE rb_llama_vocab_get_add_eos(VALUE self, VALUE vocab) {
2972
2972
  return flag ? Qtrue : Qfalse;
2973
2973
  }
2974
2974
 
2975
+ /**
2976
+ * @overload llama_vocab_get_add_sep(vocab)
2977
+ * @param [LlamaVocab] vocab
2978
+ * @return [Boolean]
2979
+ */
2980
+ static VALUE rb_llama_vocab_get_add_sep(VALUE self, VALUE vocab) {
2981
+ if (!rb_obj_is_kind_of(vocab, rb_cLlamaVocab)) {
2982
+ rb_raise(rb_eArgError, "vocab must be a LlamaVocab");
2983
+ return Qnil;
2984
+ }
2985
+ llama_vocab_wrapper* vocab_wrapper = get_llama_vocab_wrapper(vocab);
2986
+ const bool flag = llama_vocab_get_add_sep(vocab_wrapper->vocab);
2987
+ RB_GC_GUARD(vocab);
2988
+ return flag ? Qtrue : Qfalse;
2989
+ }
2990
+
2975
2991
  /**
2976
2992
  * @overload llama_vocab_fim_pre(vocab)
2977
2993
  * @param [LlamaVocab] vocab
@@ -5275,6 +5291,9 @@ void Init_llama_cpp(void) {
5275
5291
  /* llama_vocab_get_add_eos */
5276
5292
  rb_define_module_function(rb_mLlamaCpp, "llama_vocab_get_add_eos", rb_llama_vocab_get_add_eos, 1);
5277
5293
 
5294
+ /* llama_vocab_get_add_sep */
5295
+ rb_define_module_function(rb_mLlamaCpp, "llama_vocab_get_add_sep", rb_llama_vocab_get_add_sep, 1);
5296
+
5278
5297
  /* llama_vocab_fim_pre */
5279
5298
  rb_define_module_function(rb_mLlamaCpp, "llama_vocab_fim_pre", rb_llama_vocab_fim_pre, 1);
5280
5299
 
@@ -3,8 +3,8 @@
3
3
  # llama_cpp.rb provides Ruby bindings for the llama.cpp.
4
4
  module LlamaCpp
5
5
  # The version of llama_cpp.rb you install.
6
- VERSION = '0.20.3'
6
+ VERSION = '0.20.4'
7
7
 
8
8
  # The supported version of llama.cpp.
9
- LLAMA_CPP_VERSION = 'b5650'
9
+ LLAMA_CPP_VERSION = 'b5720'
10
10
  end
metadata CHANGED
@@ -1,7 +1,7 @@
1
1
  --- !ruby/object:Gem::Specification
2
2
  name: llama_cpp
3
3
  version: !ruby/object:Gem::Version
4
- version: 0.20.3
4
+ version: 0.20.4
5
5
  platform: ruby
6
6
  authors:
7
7
  - yoshoku