llama_cpp 0.0.6 → 0.1.0

Sign up to get free protection for your applications and to get access to all the features.
checksums.yaml CHANGED
@@ -1,7 +1,7 @@
1
1
  ---
2
2
  SHA256:
3
- metadata.gz: 00d1390f6d91449d27050952aa9dd80572fb18d6809dc64098e2e4fce79bc91e
4
- data.tar.gz: 49f4422e5c8903bb83fbd69a7901ff8ed0cbfae248d6652bb93b43feac331821
3
+ metadata.gz: c4058abcb7afa897554fc75bb368caeea0e77429e01fb5f3a1949191c50f4de5
4
+ data.tar.gz: 9929e94c02b5d9c21379a9275f08668e835f91d3d7be3570a2da9ab4ecbe6ad1
5
5
  SHA512:
6
- metadata.gz: 0633f8565d940b618eb1637fa095a7bc940618100698986971274ffb943fca9fab32eb04e5a0e22bd930f18f998b7f4ca59fb0e4f7ac6210efdf7d4c44e8cc9c
7
- data.tar.gz: 2e79f037ec38c415cbe9a64485b986c64e3de73c26580b0e6e8be577cec38abcc4eb9506de2bd1d0b84646f4f44b7422aafe83c0ee867028c2a223f345a294ce
6
+ metadata.gz: ab267defd1769e7bf4599da199f50a7c5cc2355d2281ab7fd2ccd1a5ef196b716350cf8df9522a9185d02c8c3ad6a5d0f46f271fad0951440ab9b3fab4019932
7
+ data.tar.gz: 16727a2ac2c68f7913749b656c26523e9eee0118b69ff06bbc0935f899eac1874f16395d9e72ed2caa853e9c61fb9f614ad5913fca623e356aa249308b2f3dda
data/CHANGELOG.md CHANGED
@@ -1,9 +1,28 @@
1
1
  ## [Unreleased]
2
2
 
3
+ ## [[0.1.0](https://github.com/yoshoku/llama_cpp.rb/compare/v0.0.7...v0.1.0)] - 2023-05-20
4
+
5
+ **Breaking Changes**
6
+
7
+ - Bump bundled llama.cpp from master-11d9023 to master-173d0e6.
8
+ - Support new API.
9
+
10
+ ## [[0.0.7](https://github.com/yoshoku/llama_cpp.rb/compare/v0.0.6...v0.0.7)] - 2023-04-29
11
+
12
+ - Bump bundled llama.cpp from master-12b5900 to master-11d9023.
13
+ - Add Client class.
14
+ - Add model file type constants.
15
+ - Add getter and setter methods of use_mmap to ContextParams.
16
+ - Add empty? method to Context.
17
+ - Add clblast config option:
18
+ ```
19
+ $ gem install llama_cpp -- --with-clblast
20
+ ```
21
+
3
22
  ## [[0.0.6](https://github.com/yoshoku/llama_cpp.rb/compare/v0.0.5...v0.0.6)] - 2023-04-22
4
23
 
5
24
  - Bump bundled llama.cpp from master-315a95a to master-12b5900.
6
- - Add model file type constants
25
+ - Add model file type constants.
7
26
  - Add `model_quantize` module function to LLaMACpp.
8
27
  - Add cublas config option:
9
28
  ```
@@ -5,6 +5,8 @@ require 'mkmf'
5
5
  abort 'libstdc++ is not found.' unless have_library('stdc++')
6
6
 
7
7
  $srcs = %w[ggml.c llama.cpp llama_cpp.cpp]
8
+ $srcs << 'ggml-opencl.c' if with_config('clblast')
9
+
8
10
  $CFLAGS << ' -w'
9
11
  $CXXFLAGS << ' -std=c++11'
10
12
  $INCFLAGS << ' -I$(srcdir)/src'
@@ -34,6 +36,13 @@ if with_config('cublas')
34
36
  $objs = %w[ggml-cuda.o ggml.o llama.o llama_cpp.o]
35
37
  end
36
38
 
39
+ if with_config('clblast')
40
+ abort 'libclblast is not found.' unless have_library('clblast')
41
+ abort 'libOpenCL is not found.' unless have_library('OpenCL')
42
+
43
+ $CFLAGS << ' -DGGML_USE_CLBLAST'
44
+ end
45
+
37
46
  UNAME_M = RbConfig::CONFIG['build_cpu'] || RbConfig::CONFIG['host_cpu'] || RbConfig::CONFIG['target_cpu']
38
47
 
39
48
  # rubocop:disable Layout/LineLength