llama_cpp 0.0.6 → 0.1.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
checksums.yaml CHANGED
@@ -1,7 +1,7 @@
1
1
  ---
2
2
  SHA256:
3
- metadata.gz: 00d1390f6d91449d27050952aa9dd80572fb18d6809dc64098e2e4fce79bc91e
4
- data.tar.gz: 49f4422e5c8903bb83fbd69a7901ff8ed0cbfae248d6652bb93b43feac331821
3
+ metadata.gz: c4058abcb7afa897554fc75bb368caeea0e77429e01fb5f3a1949191c50f4de5
4
+ data.tar.gz: 9929e94c02b5d9c21379a9275f08668e835f91d3d7be3570a2da9ab4ecbe6ad1
5
5
  SHA512:
6
- metadata.gz: 0633f8565d940b618eb1637fa095a7bc940618100698986971274ffb943fca9fab32eb04e5a0e22bd930f18f998b7f4ca59fb0e4f7ac6210efdf7d4c44e8cc9c
7
- data.tar.gz: 2e79f037ec38c415cbe9a64485b986c64e3de73c26580b0e6e8be577cec38abcc4eb9506de2bd1d0b84646f4f44b7422aafe83c0ee867028c2a223f345a294ce
6
+ metadata.gz: ab267defd1769e7bf4599da199f50a7c5cc2355d2281ab7fd2ccd1a5ef196b716350cf8df9522a9185d02c8c3ad6a5d0f46f271fad0951440ab9b3fab4019932
7
+ data.tar.gz: 16727a2ac2c68f7913749b656c26523e9eee0118b69ff06bbc0935f899eac1874f16395d9e72ed2caa853e9c61fb9f614ad5913fca623e356aa249308b2f3dda
data/CHANGELOG.md CHANGED
@@ -1,9 +1,28 @@
1
1
  ## [Unreleased]
2
2
 
3
+ ## [[0.1.0](https://github.com/yoshoku/llama_cpp.rb/compare/v0.0.7...v0.1.0)] - 2023-05-20
4
+
5
+ **Breaking Changes**
6
+
7
+ - Bump bundled llama.cpp from master-11d9023 to master-173d0e6.
8
+ - Support new API.
9
+
10
+ ## [[0.0.7](https://github.com/yoshoku/llama_cpp.rb/compare/v0.0.6...v0.0.7)] - 2023-04-29
11
+
12
+ - Bump bundled llama.cpp from master-12b5900 to master-11d9023.
13
+ - Add Client class.
14
+ - Add model file type constants.
15
+ - Add getter and setter methods of use_mmap to ContextParams.
16
+ - Add empty? method to Context.
17
+ - Add clblast config option:
18
+ ```
19
+ $ gem install llama_cpp -- --with-clblast
20
+ ```
21
+
3
22
  ## [[0.0.6](https://github.com/yoshoku/llama_cpp.rb/compare/v0.0.5...v0.0.6)] - 2023-04-22
4
23
 
5
24
  - Bump bundled llama.cpp from master-315a95a to master-12b5900.
6
- - Add model file type constants
25
+ - Add model file type constants.
7
26
  - Add `model_quantize` module function to LLaMACpp.
8
27
  - Add cublas config option:
9
28
  ```
@@ -5,6 +5,8 @@ require 'mkmf'
5
5
  abort 'libstdc++ is not found.' unless have_library('stdc++')
6
6
 
7
7
  $srcs = %w[ggml.c llama.cpp llama_cpp.cpp]
8
+ $srcs << 'ggml-opencl.c' if with_config('clblast')
9
+
8
10
  $CFLAGS << ' -w'
9
11
  $CXXFLAGS << ' -std=c++11'
10
12
  $INCFLAGS << ' -I$(srcdir)/src'
@@ -34,6 +36,13 @@ if with_config('cublas')
34
36
  $objs = %w[ggml-cuda.o ggml.o llama.o llama_cpp.o]
35
37
  end
36
38
 
39
+ if with_config('clblast')
40
+ abort 'libclblast is not found.' unless have_library('clblast')
41
+ abort 'libOpenCL is not found.' unless have_library('OpenCL')
42
+
43
+ $CFLAGS << ' -DGGML_USE_CLBLAST'
44
+ end
45
+
37
46
  UNAME_M = RbConfig::CONFIG['build_cpu'] || RbConfig::CONFIG['host_cpu'] || RbConfig::CONFIG['target_cpu']
38
47
 
39
48
  # rubocop:disable Layout/LineLength