llama_cpp 0.5.3 → 0.7.0

Sign up to get free protection for your applications and to get access to all the features.
data/sig/llama_cpp.rbs CHANGED
@@ -67,16 +67,16 @@ module LLaMACpp
67
67
  class Model
68
68
  public
69
69
 
70
- def initialize: (model_path: String, params: ::LLaMACpp::ContextParams) -> void
70
+ def initialize: (model_path: String, params: ::LLaMACpp::ModelParams) -> void
71
71
  | () -> void
72
72
  def empty?: () -> bool
73
73
  def free: () -> void
74
- def load: (model_path: String, params: ::LLaMACpp::ContextParams) -> void
75
- def apply_lora_from_file: (lora_path: String, ?base_model_path: String, ?n_threads: Integer) -> void
74
+ def load: (model_path: String, params: ::LLaMACpp::ModelParams) -> void
75
+ def apply_lora_from_file: (lora_path: String, ?scale: Float, ?base_model_path: String, ?n_threads: Integer) -> void
76
76
  def n_vocab: () -> Integer
77
- def n_ctx: () -> Integer
78
77
  def n_ctx_train: () -> Integer
79
78
  def n_embd: () -> Integer
79
+ def rope_freq_scale_train: () -> Float
80
80
  def token_to_piece: (Integer) -> String
81
81
  def tokenize: (text: String, ?n_max_tokens: Integer, ?add_bos: bool) -> Array[Integer]
82
82
  def desc: () -> String
@@ -98,10 +98,50 @@ module LLaMACpp
98
98
  def n_eval: () -> Integer
99
99
  end
100
100
 
101
+ class ModelParams
102
+ public
103
+
104
+ def n_gpu_layers: () -> Integer
105
+ def n_gpu_layers=: (Integer) -> Integer
106
+ def main_gpu: () -> Integer
107
+ def main_gpu=: (Integer) -> Integer
108
+ def tensor_split: () -> Array[Float]
109
+ def vocab_only: () -> bool
110
+ def vocab_only=: (bool) -> bool
111
+ def use_mmap: () -> bool
112
+ def use_mmap=: (bool) -> bool
113
+ def use_mlock: () -> bool
114
+ def use_mlock=: (bool) -> bool
115
+ end
116
+
117
+ class Batch
118
+ public
119
+
120
+ def initialize: (n_tokens: Integer, embd: Integer) -> void
121
+ def n_tokens=: (Integer) -> Integer
122
+ def n_tokens: () -> Integer
123
+ def all_pos_zero=: (Integer) -> Integer
124
+ def all_pos_zero: () -> Integer
125
+ def all_pos_one=: (Integer) -> Integer
126
+ def all_pos_one: () -> Integer
127
+ def all_seq_id=: (Integer) -> Integer
128
+ def all_seq_id: () -> Integer
129
+ def set_token: (Integer, Integer) -> Integer
130
+ def get_token: (Integer) -> Integer
131
+ def set_pos: (Integer, Integer) -> Integer
132
+ def get_pos: (Integer) -> Integer
133
+ def set_seq_id: (Integer, Integer) -> Integer
134
+ def get_seq_id: (Integer) -> Integer
135
+ def set_logit: (Integer, bool) -> bool
136
+ def get_logit: (Integer) -> bool
137
+ end
138
+
101
139
  class Context
102
140
  public
103
141
 
104
- def initialize: (model: ::LLaMACpp::Model) -> void
142
+ attr_reader model: ::LLaMACpp::Model
143
+
144
+ def initialize: (model: ::LLaMACpp::Model, params: ::LLaMACpp::ContextParams) -> void
105
145
  def embeddings: () -> Array[Float]
106
146
  def text: (Integer) -> String
107
147
  def score: (Integer) -> Float
@@ -109,20 +149,24 @@ module LLaMACpp
109
149
  def token_bos: () -> Integer
110
150
  def token_eos: () -> Integer
111
151
  def token_nl: () -> Integer
112
- def eval: (tokens: Array[Integer], n_past: Integer, ?n_tokens: Integer, ?n_threads: Integer) -> void
113
- def eval_embd: (tokens: Array[Float], n_past: Integer, ?n_tokens: Integer, ?n_threads: Integer) -> void
114
- def eval_export: (String) -> bool
152
+ def token_prefix: () -> Integer
153
+ def token_middle: () -> Integer
154
+ def token_suffix: () -> Integer
155
+ def token_eot: () -> Integer
156
+ def eval: (tokens: Array[Integer], n_past: Integer, ?n_tokens: Integer) -> void
157
+ def eval_embd: (tokens: Array[Float], n_past: Integer, ?n_tokens: Integer) -> void
158
+ def decode: (::LLaMACpp::Batch) -> void
115
159
  def logits: () -> Array[Float]
116
160
  def n_ctx: () -> Integer
117
- def n_ctx_train: () -> Integer
118
- def n_embd: () -> Integer
119
- def n_vocab: () -> Integer
120
161
  def timings: () -> ::LLaMACpp::Timings
121
162
  def print_timings: () -> void
122
163
  def reset_timings: () -> void
123
- def token_to_piece: (Integer) -> String
124
- def tokenize: (text: String, ?n_max_tokens: Integer, ?add_bos: bool) -> Array[Integer]
125
164
  def kv_cache_token_count: () -> Integer
165
+ def kv_cache_tokens_rm: (Integer, Integer) -> void
166
+ def kv_cache_seq_rm: (Integer, Integer,Integer) -> void
167
+ def kv_cache_seq_cp: (Integer, Integer,Integer, Integer) -> void
168
+ def kv_cache_seq_keep: (Integer) -> void
169
+ def kv_cache_seq_shift: (Integer, Integer, Ingteger, Integer) -> void
126
170
  def set_rng_seed: (Integer) -> void
127
171
  def load_session_file: (session_path: String) -> void
128
172
  def save_session_file: (session_path: String, session_tokens: Array[Integer]) -> void
@@ -134,6 +178,7 @@ module LLaMACpp
134
178
  def sample_top_p: (::LLaMACpp::TokenDataArray, prob: Float, ?min_keep: Integer) -> void
135
179
  def sample_tail_free: (::LLaMACpp::TokenDataArray, z: Float, ?min_keep: Integer) -> void
136
180
  def sample_typical: (::LLaMACpp::TokenDataArray, prob: Float, ?min_keep: Integer) -> void
181
+ def sample_temp: (::LLaMACpp::TokenDataArray, temp: Float) -> void
137
182
  def sample_temperature: (::LLaMACpp::TokenDataArray, temperature: Float) -> void
138
183
  def sample_token_mirostat: (::LLaMACpp::TokenDataArray, tau: Float, eta: Float, m: Integer, mu: Float) -> [Integer, Float]
139
184
  def sample_token_mirostat_v2: (::LLaMACpp::TokenDataArray, tau: Float, eta: Float, mu: Float) -> [Integer, Float]
@@ -146,37 +191,28 @@ module LLaMACpp
146
191
  class ContextParams
147
192
  public
148
193
 
149
- def embedding: () -> bool
150
- def embedding=: (bool) -> bool
151
- def f16_kv: () -> bool
152
- def f16_kv=: (bool) -> bool
153
- def logits_all: () -> bool
154
- def logits_all=: (bool) -> bool
194
+ def seed: () -> Integer
195
+ def seed=: (Integer) -> Integer
155
196
  def n_ctx: () -> Integer
156
197
  def n_ctx=: (Integer) -> Integer
157
198
  def n_batch: () -> Integer
158
199
  def n_batch=: (Integer) -> Integer
159
- def n_gpu_layers: () -> Integer
160
- def n_gpu_layers=: (Integer) -> Integer
161
- def main_gpu: () -> Integer
162
- def main_gpu=: (Integer) -> Integer
163
- def tensor_split: () -> Array[Float]
200
+ def n_threads: () -> Integer
201
+ def n_threads=: (Integer) -> Integer
202
+ def n_threads_batch: () -> Integer
203
+ def n_threads_batch=: (Integer) -> Integer
164
204
  def rope_freq_base=: (Float) -> Float
165
205
  def rope_freq_base: () -> Float
166
206
  def rope_freq_scale=: (Float) -> Float
167
207
  def rope_freq_scale: () -> Float
168
- def low_vram: () -> bool
169
- def low_vram=: (bool) -> bool
170
208
  def mul_mat_q: () -> bool
171
209
  def mul_mat_q=: (bool) -> bool
172
- def seed: () -> Integer
173
- def seed=: (Integer) -> Integer
174
- def use_mlock: () -> bool
175
- def use_mlock=: (bool) -> bool
176
- def use_mmap: () -> bool
177
- def use_mmap=: (bool) -> bool
178
- def vocab_only: () -> bool
179
- def vocab_only=: (bool) -> bool
210
+ def f16_kv: () -> bool
211
+ def f16_kv=: (bool) -> bool
212
+ def logits_all: () -> bool
213
+ def logits_all=: (bool) -> bool
214
+ def embedding: () -> bool
215
+ def embedding=: (bool) -> bool
180
216
  end
181
217
 
182
218
  class ModelQuantizeParams
metadata CHANGED
@@ -1,14 +1,14 @@
1
1
  --- !ruby/object:Gem::Specification
2
2
  name: llama_cpp
3
3
  version: !ruby/object:Gem::Version
4
- version: 0.5.3
4
+ version: 0.7.0
5
5
  platform: ruby
6
6
  authors:
7
7
  - yoshoku
8
8
  autorequire:
9
9
  bindir: exe
10
10
  cert_chain: []
11
- date: 2023-09-22 00:00:00.000000000 Z
11
+ date: 2023-10-07 00:00:00.000000000 Z
12
12
  dependencies: []
13
13
  description: llama_cpp.rb provides Ruby bindings for the llama.cpp.
14
14
  email:
@@ -48,6 +48,7 @@ files:
48
48
  - ext/llama_cpp/src/llama-util.h
49
49
  - ext/llama_cpp/src/llama.cpp
50
50
  - ext/llama_cpp/src/llama.h
51
+ - ext/llama_cpp/src/unicode.h
51
52
  - lib/llama_cpp.rb
52
53
  - lib/llama_cpp/version.rb
53
54
  - sig/llama_cpp.rbs
@@ -75,7 +76,7 @@ required_rubygems_version: !ruby/object:Gem::Requirement
75
76
  - !ruby/object:Gem::Version
76
77
  version: '0'
77
78
  requirements: []
78
- rubygems_version: 3.3.26
79
+ rubygems_version: 3.4.19
79
80
  signing_key:
80
81
  specification_version: 4
81
82
  summary: Ruby bindings for the llama.cpp.