llama-rb 0.2.0 → 0.2.1
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/README.md +2 -8
- data/lib/llama/version.rb +1 -1
- metadata +1 -1
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: '03801e4f99933be9c0e8d559008626991535c2167af88c8cb31defb31c88d0f6'
|
4
|
+
data.tar.gz: 6f17e50818de906f33de2686cf1b75c0e17aa052f0fba60889bad85df0591f59
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: 40602fc8c253087a78fd4e5edf5fbae24f3a4ad0d9a3bb2f6730ef701753f6815e8716303220e8edcb1984484d5ffbd20c6adb7e07690244cd738ec6918c80e8
|
7
|
+
data.tar.gz: 9cbf6bed4fa4359bd007d083f99976a885b1557b0bf01c4d22a55e231515adf7f66e58e951e01bf731e827b893bf6fc278a306f8a566be3e133039f210214bc2
|
data/README.md
CHANGED
@@ -42,21 +42,15 @@ m.predict('hello world')
|
|
42
42
|
```ruby
|
43
43
|
def self.new(
|
44
44
|
model, # path to model file, e.g. "models/7B/ggml-model-q4_0.bin"
|
45
|
-
|
46
|
-
n_parts: -1, # amount of model parts (-1 = determine from model dimensions)
|
45
|
+
n_predict: 128 # number of tokens to predict
|
47
46
|
seed: Time.now.to_i, # RNG seed
|
48
|
-
memory_f16: true, # use f16 instead of f32 for memory kv
|
49
|
-
use_mlock: false # use mlock to keep model in memory
|
50
47
|
)
|
51
48
|
```
|
52
49
|
|
53
50
|
#### Llama::Model#predict
|
54
51
|
|
55
52
|
```ruby
|
56
|
-
def predict(
|
57
|
-
prompt, # string used as prompt
|
58
|
-
n_predict: 128 # number of tokens to predict
|
59
|
-
)
|
53
|
+
def predict(prompt)
|
60
54
|
```
|
61
55
|
|
62
56
|
## Development
|
data/lib/llama/version.rb
CHANGED