llama_cpp 0.0.6 → 0.0.7

Sign up to get free protection for your applications and to get access to all the features.
data/sig/llama_cpp.rbs CHANGED
@@ -12,9 +12,12 @@ module LLaMACpp
12
12
  LLAMA_FTYPE_MOSTLY_Q4_1_SOME_F16: Integer
13
13
  LLAMA_FTYPE_MOSTLY_Q4_2: Integer
14
14
  LLAMA_FTYPE_MOSTLY_Q4_3: Integer
15
+ LLAMA_FTYPE_MOSTLY_Q8_0: Integer
16
+ LLAMA_FTYPE_MOSTLY_Q5_0: Integer
17
+ LLAMA_FTYPE_MOSTLY_Q5_1: Integer
15
18
 
16
19
  def self?.model_quantize: (input_path: String, output_path: String, ftype: Integer, ?n_threads: Integer) -> void
17
- def self?.generate: (::LLaMACpp::Context, String, ?n_threads: Integer) -> String
20
+ def self?.generate: (::LLaMACpp::Context, String, ?n_predict: Integer, ?n_threads: Integer) -> String
18
21
  def self?.print_system_info: () -> void
19
22
  def self?.token_bos: () -> Integer
20
23
  def self?.token_eos: () -> Integer
@@ -27,6 +30,7 @@ module LLaMACpp
27
30
  def initialize: (model_path: String, params: ::LLaMACpp::ContextParams) -> void
28
31
  | () -> void
29
32
  def embeddings: () -> Array[Float]
33
+ def empty?: () -> bool
30
34
  def eval: (tokens: Array[Integer], n_past: Integer, ?n_tokens: Integer, ?n_threads: Integer) -> void
31
35
  def free: () -> void
32
36
  def load: (model_path: String, params: ::LLaMACpp::ContextParams) -> void
@@ -59,9 +63,20 @@ module LLaMACpp
59
63
  def seed=: (Integer) -> Integer
60
64
  def use_mlock: () -> bool
61
65
  def use_mlock=: (bool) -> bool
66
+ def use_mmap: () -> bool
67
+ def use_mmap=: (bool) -> bool
62
68
  def vocab_only: () -> bool
63
69
  def vocab_only=: (bool) -> bool
64
70
  end
65
71
 
66
72
  class Params = ContextParams
73
+
74
+ class Client
75
+ def initialize(model_path: String, ?lora_adapter_path: String, ?lora_base_path: String,
76
+ ?n_ctx: Integer, ?n_parts: Integer, ?memory_f16: bool, ?use_mmap: bool, ?use_mlock: bool,
77
+ ?embedding: bool, ?n_threads: Integer, ?seed: Integer) -> void
78
+ def completions(String, ?max_tokens: Integer, ?n_keep: Integer, ?repeat_last_n: Integer, ?n_batch: Integer,
79
+ ?top_k: Integer, ?top_p: Float, ?temperature: Float, ?repeat_penalty: Float) -> String
80
+ def embeddings(String) -> Array[Float]
81
+ end
67
82
  end
metadata CHANGED
@@ -1,14 +1,14 @@
1
1
  --- !ruby/object:Gem::Specification
2
2
  name: llama_cpp
3
3
  version: !ruby/object:Gem::Version
4
- version: 0.0.6
4
+ version: 0.0.7
5
5
  platform: ruby
6
6
  authors:
7
7
  - yoshoku
8
8
  autorequire:
9
9
  bindir: exe
10
10
  cert_chain: []
11
- date: 2023-04-22 00:00:00.000000000 Z
11
+ date: 2023-04-29 00:00:00.000000000 Z
12
12
  dependencies: []
13
13
  description: llama_cpp.rb provides Ruby bindings for the llama.cpp.
14
14
  email:
@@ -27,12 +27,15 @@ files:
27
27
  - ext/llama_cpp/llama_cpp.h
28
28
  - ext/llama_cpp/src/LICENSE
29
29
  - ext/llama_cpp/src/ggml-cuda.h
30
+ - ext/llama_cpp/src/ggml-opencl.c
31
+ - ext/llama_cpp/src/ggml-opencl.h
30
32
  - ext/llama_cpp/src/ggml.c
31
33
  - ext/llama_cpp/src/ggml.h
32
34
  - ext/llama_cpp/src/llama.cpp
33
35
  - ext/llama_cpp/src/llama.h
34
36
  - ext/llama_cpp/src/llama_util.h
35
37
  - lib/llama_cpp.rb
38
+ - lib/llama_cpp/client.rb
36
39
  - lib/llama_cpp/version.rb
37
40
  - sig/llama_cpp.rbs
38
41
  homepage: https://github.com/yoshoku/llama_cpp.rb