bigdl-core-cpp 2.5.0b20240723__py3-none-manylinux2010_x86_64.whl → 2.5.0b20240725__py3-none-manylinux2010_x86_64.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -2342,6 +2342,53 @@ class GemmaModel(Model):
2342
2342
  return [(self.map_tensor_name(name), data_torch)]
2343
2343
 
2344
2344
 
2345
+ @Model.register("Gemma2ForCausalLM")
2346
+ class Gemma2Model(Model):
2347
+ model_arch = gguf.MODEL_ARCH.GEMMA2
2348
+
2349
+ def set_vocab(self):
2350
+ self._set_vocab_llama_hf()
2351
+ self.gguf_writer.add_add_space_prefix(False)
2352
+
2353
+ def set_gguf_parameters(self):
2354
+ hparams = self.hparams
2355
+ block_count = hparams["num_hidden_layers"]
2356
+
2357
+ self.gguf_writer.add_name(self.dir_model.name if self.model_name is None else self.model_name)
2358
+ self.gguf_writer.add_context_length(hparams["max_position_embeddings"])
2359
+ self.gguf_writer.add_embedding_length(hparams["hidden_size"])
2360
+ self.gguf_writer.add_block_count(block_count)
2361
+ self.gguf_writer.add_feed_forward_length(hparams["intermediate_size"])
2362
+ self.gguf_writer.add_head_count(hparams["num_attention_heads"])
2363
+ self.gguf_writer.add_head_count_kv(self.hparams["num_key_value_heads"] if "num_key_value_heads" in hparams else hparams["num_attention_heads"])
2364
+ self.gguf_writer.add_layer_norm_rms_eps(self.hparams["rms_norm_eps"])
2365
+ self.gguf_writer.add_key_length(hparams["head_dim"])
2366
+ self.gguf_writer.add_value_length(hparams["head_dim"])
2367
+ self.gguf_writer.add_file_type(self.ftype)
2368
+ self.gguf_writer.add_attn_logit_softcapping(
2369
+ self.hparams["attn_logit_softcapping"]
2370
+ )
2371
+ self.gguf_writer.add_final_logit_softcapping(
2372
+ self.hparams["final_logit_softcapping"]
2373
+ )
2374
+ self.gguf_writer.add_sliding_window(self.hparams["sliding_window"])
2375
+
2376
+ def modify_tensors(self, data_torch: Tensor, name: str, bid: int | None) -> Iterable[tuple[str, Tensor]]:
2377
+ del bid # unusem
2378
+
2379
+ # lm_head is not used in llama.cpp, while autoawq will include this tensor in model
2380
+ # To prevent errors, skip loading lm_head.weight.
2381
+ if name == "lm_head.weight":
2382
+ logger.debug(f"Skipping get tensor {name!r} in safetensors so that convert can end normally.")
2383
+ return []
2384
+
2385
+ # ref: https://github.com/huggingface/transformers/blob/fc37f38915372c15992b540dfcbbe00a916d4fc6/src/transformers/models/gemma/modeling_gemma.py#L89
2386
+ if name.endswith("norm.weight"):
2387
+ data_torch = data_torch + 1
2388
+
2389
+ return [(self.map_tensor_name(name), data_torch)]
2390
+
2391
+
2345
2392
  @Model.register("Starcoder2ForCausalLM")
2346
2393
  class StarCoder2Model(Model):
2347
2394
  model_arch = gguf.MODEL_ARCH.STARCODER2
@@ -55,6 +55,10 @@ class Keys:
55
55
  LAYERNORM_EPS = "{arch}.attention.layer_norm_epsilon"
56
56
  LAYERNORM_RMS_EPS = "{arch}.attention.layer_norm_rms_epsilon"
57
57
  CAUSAL = "{arch}.attention.causal"
58
+ Q_LORA_RANK = "{arch}.attention.q_lora_rank"
59
+ KV_LORA_RANK = "{arch}.attention.kv_lora_rank"
60
+ REL_BUCKETS_COUNT = "{arch}.attention.relative_buckets_count"
61
+ SLIDING_WINDOW = "{arch}.attention.sliding_window"
58
62
 
59
63
  class Rope:
60
64
  DIMENSION_COUNT = "{arch}.rope.dimension_count"
@@ -133,6 +137,7 @@ class MODEL_ARCH(IntEnum):
133
137
  INTERNLM2 = auto()
134
138
  MINICPM = auto()
135
139
  GEMMA = auto()
140
+ GEMMA2 = auto()
136
141
  STARCODER2 = auto()
137
142
  MAMBA = auto()
138
143
  XVERSE = auto()
@@ -160,10 +165,13 @@ class MODEL_TENSOR(IntEnum):
160
165
  ATTN_NORM = auto()
161
166
  ATTN_NORM_2 = auto()
162
167
  ATTN_OUT_NORM = auto()
168
+ ATTN_POST_NORM = auto()
163
169
  ATTN_ROT_EMBD = auto()
164
170
  FFN_GATE_INP = auto()
165
171
  FFN_GATE_INP_SHEXP = auto()
166
172
  FFN_NORM = auto()
173
+ FFN_PRE_NORM = auto()
174
+ FFN_POST_NORM = auto()
167
175
  FFN_GATE = auto()
168
176
  FFN_DOWN = auto()
169
177
  FFN_UP = auto()
@@ -214,6 +222,7 @@ MODEL_ARCH_NAMES: dict[MODEL_ARCH, str] = {
214
222
  MODEL_ARCH.INTERNLM2: "internlm2",
215
223
  MODEL_ARCH.MINICPM: "minicpm",
216
224
  MODEL_ARCH.GEMMA: "gemma",
225
+ MODEL_ARCH.GEMMA2: "gemma2",
217
226
  MODEL_ARCH.STARCODER2: "starcoder2",
218
227
  MODEL_ARCH.MAMBA: "mamba",
219
228
  MODEL_ARCH.XVERSE: "xverse",
@@ -244,9 +253,12 @@ TENSOR_NAMES: dict[MODEL_TENSOR, str] = {
244
253
  MODEL_TENSOR.ATTN_Q_NORM: "blk.{bid}.attn_q_norm",
245
254
  MODEL_TENSOR.ATTN_K_NORM: "blk.{bid}.attn_k_norm",
246
255
  MODEL_TENSOR.ATTN_OUT_NORM: "blk.{bid}.attn_output_norm",
256
+ MODEL_TENSOR.ATTN_POST_NORM: "blk.{bid}.post_attention_norm",
247
257
  MODEL_TENSOR.FFN_GATE_INP: "blk.{bid}.ffn_gate_inp",
248
258
  MODEL_TENSOR.FFN_GATE_INP_SHEXP: "blk.{bid}.ffn_gate_inp_shexp",
249
259
  MODEL_TENSOR.FFN_NORM: "blk.{bid}.ffn_norm",
260
+ MODEL_TENSOR.FFN_PRE_NORM: "blk.{bid}.ffn_norm",
261
+ MODEL_TENSOR.FFN_POST_NORM: "blk.{bid}.post_ffw_norm",
250
262
  MODEL_TENSOR.FFN_GATE: "blk.{bid}.ffn_gate",
251
263
  MODEL_TENSOR.FFN_DOWN: "blk.{bid}.ffn_down",
252
264
  MODEL_TENSOR.FFN_UP: "blk.{bid}.ffn_up",
@@ -654,6 +666,21 @@ MODEL_TENSORS: dict[MODEL_ARCH, list[MODEL_TENSOR]] = {
654
666
  MODEL_TENSOR.FFN_UP,
655
667
  MODEL_TENSOR.FFN_NORM,
656
668
  ],
669
+ MODEL_ARCH.GEMMA2: [
670
+ MODEL_TENSOR.TOKEN_EMBD,
671
+ MODEL_TENSOR.OUTPUT_NORM,
672
+ MODEL_TENSOR.ATTN_Q,
673
+ MODEL_TENSOR.ATTN_K,
674
+ MODEL_TENSOR.ATTN_V,
675
+ MODEL_TENSOR.ATTN_OUT,
676
+ MODEL_TENSOR.FFN_GATE,
677
+ MODEL_TENSOR.FFN_DOWN,
678
+ MODEL_TENSOR.FFN_UP,
679
+ MODEL_TENSOR.ATTN_NORM,
680
+ MODEL_TENSOR.ATTN_POST_NORM,
681
+ MODEL_TENSOR.FFN_PRE_NORM,
682
+ MODEL_TENSOR.FFN_POST_NORM,
683
+ ],
657
684
  MODEL_ARCH.STARCODER2: [
658
685
  MODEL_TENSOR.TOKEN_EMBD,
659
686
  MODEL_TENSOR.OUTPUT_NORM,
@@ -416,6 +416,9 @@ class GGUFWriter:
416
416
  def add_causal_attention(self, value: bool) -> None:
417
417
  self.add_bool(Keys.Attention.CAUSAL.format(arch=self.arch), value)
418
418
 
419
+ def add_sliding_window(self, value: int) -> None:
420
+ self.add_uint32(Keys.Attention.SLIDING_WINDOW.format(arch=self.arch), value)
421
+
419
422
  def add_pooling_type(self, value: PoolingType) -> None:
420
423
  self.add_uint32(Keys.LLM.POOLING_TYPE.format(arch=self.arch), value.value)
421
424
 
@@ -185,6 +185,10 @@ class TensorNameMap:
185
185
  "transformer.blocks.{bid}.norm_attn_norm.norm_2", # dbrx
186
186
  ),
187
187
 
188
+ MODEL_TENSOR.ATTN_POST_NORM: (
189
+ "model.layers.{bid}.post_attention_layernorm", # gemma2
190
+ ),
191
+
188
192
  # Rotary embeddings
189
193
  MODEL_TENSOR.ATTN_ROT_EMBD: (
190
194
  "model.layers.{bid}.self_attn.rotary_emb.inv_freq", # llama-hf
@@ -208,6 +212,16 @@ class TensorNameMap:
208
212
  "transformer.decoder_layer.{bid}.rms_norm_2", # Grok
209
213
  ),
210
214
 
215
+ # Post feed-forward norm
216
+ MODEL_TENSOR.FFN_PRE_NORM: (
217
+ "model.layers.{bid}.pre_feedforward_layernorm", # gemma2
218
+ ),
219
+
220
+ # Post feed-forward norm
221
+ MODEL_TENSOR.FFN_POST_NORM: (
222
+ "model.layers.{bid}.post_feedforward_layernorm", # gemma2
223
+ ),
224
+
211
225
  MODEL_TENSOR.FFN_GATE_INP: (
212
226
  "layers.{bid}.feed_forward.gate", # mixtral
213
227
  "model.layers.{bid}.block_sparse_moe.gate", # mixtral
bigdl/cpp/libs/baby-llama CHANGED
Binary file
bigdl/cpp/libs/batched CHANGED
Binary file
Binary file
bigdl/cpp/libs/benchmark CHANGED
Binary file
bigdl/cpp/libs/embedding CHANGED
Binary file
bigdl/cpp/libs/gguf CHANGED
Binary file
bigdl/cpp/libs/imatrix CHANGED
Binary file
Binary file
bigdl/cpp/libs/llava-cli CHANGED
Binary file
bigdl/cpp/libs/lookahead CHANGED
Binary file
bigdl/cpp/libs/lookup CHANGED
Binary file
Binary file
bigdl/cpp/libs/main CHANGED
Binary file
bigdl/cpp/libs/ollama CHANGED
Binary file
bigdl/cpp/libs/perplexity CHANGED
Binary file
bigdl/cpp/libs/quantize CHANGED
Binary file
Binary file
Binary file
bigdl/cpp/libs/server CHANGED
Binary file
Binary file
bigdl/cpp/libs/tokenize CHANGED
Binary file
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: bigdl-core-cpp
3
- Version: 2.5.0b20240723
3
+ Version: 2.5.0b20240725
4
4
  Summary: Large Language Model Develop Toolkit
5
5
  Author: BigDL Authors
6
6
  License: Apache License, Version 2.0
@@ -0,0 +1,43 @@
1
+ bigdl/cpp/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
2
+ bigdl/cpp/convert-hf-to-gguf.py,sha256=ZQWPwgfgf-BCQMyZUHWbGuFNx4XP7KRhBTflr4ZfKO0,129245
3
+ bigdl/cpp/convert.py,sha256=XMMcpfWHwEAAWzwLXe9mmJTU7cMvcyw8g2BFctfZnvI,69417
4
+ bigdl/cpp/cli/init-llama-cpp,sha256=qpIwMQc2Vb11fmovXFJKefp5iD1tQMfL8Ma7Y3Ebguk,400
5
+ bigdl/cpp/cli/init-ollama,sha256=5eg4DsWg87iysFRh7leawr6H47F_sHydwL5we6fUG6o,193
6
+ bigdl/cpp/gguf-py/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
7
+ bigdl/cpp/gguf-py/gguf/__init__.py,sha256=Xr2OoA3yQ8Z5GcbGcppNQ4Tw9zKMzPxTD0iI6qm-SRc,172
8
+ bigdl/cpp/gguf-py/gguf/constants.py,sha256=2PStnJUUAoAuc0Se_boiGgpC2r0YG7Eai_gPEvQq8k0,35416
9
+ bigdl/cpp/gguf-py/gguf/gguf.py,sha256=8MDu7a0JEXhLUv_tjhYqDrWubVNc41cFvBYZbkZZenI,478
10
+ bigdl/cpp/gguf-py/gguf/gguf_reader.py,sha256=ReWAZwL589DW_bKWix0LK2qaDWOj0r4jXPKBymoJ-B4,12028
11
+ bigdl/cpp/gguf-py/gguf/gguf_writer.py,sha256=S-MxxZWJ-Vs_aCIEDVuyHo4W8bUtKwkUL3CTEzmNpR0,21085
12
+ bigdl/cpp/gguf-py/gguf/lazy.py,sha256=H7YCO73KFK_s1EZtO8-Cj98NJa7JaNPDaD6k1KeKBuA,9788
13
+ bigdl/cpp/gguf-py/gguf/py.typed,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
14
+ bigdl/cpp/gguf-py/gguf/quants.py,sha256=dI6JsSpaSzfdz7OwfqH1kvK98M_Ax8d6NMNYY5pyoQM,4354
15
+ bigdl/cpp/gguf-py/gguf/tensor_mapping.py,sha256=fFeDJ7HQQm-hEyq1698sUVxrCf01hTaKR3nXnR6cJcc,23852
16
+ bigdl/cpp/gguf-py/gguf/vocab.py,sha256=vsRrPQ6RafirhBmMp9h3YGZh_tqEjeg7S2hiLvMv7Ys,6847
17
+ bigdl/cpp/libs/baby-llama,sha256=Se4hMft0RmV0hQSdOy48Q_-FqN0y9tgdCI6xR4LDmnI,14209296
18
+ bigdl/cpp/libs/batched,sha256=DUvxfARun7uyUfKuVPtJ3SexJx9nZhvMeefmUyneiyo,14133856
19
+ bigdl/cpp/libs/batched-bench,sha256=uNp3FNIgLqg0vU7-AuSiyrEoy9gJTZybR30taIwHy1o,14134016
20
+ bigdl/cpp/libs/benchmark,sha256=P1dPH6y_ishJS6R4XpRfeUJCArRPTE4Z_voYbSLddsQ,1452656
21
+ bigdl/cpp/libs/embedding,sha256=FM5UWjOmprMdcqHSgRElUixLB0YLP78EdFWDhZpQz4o,14129896
22
+ bigdl/cpp/libs/gguf,sha256=zQn08SRpEPmJKO0_7sE3TIb3oTUVmruBOeaOlYYvbnA,12389616
23
+ bigdl/cpp/libs/imatrix,sha256=LN7Im5SJRKQBaI4NN6V55gyfH92PrnT56l4Z097hGhw,14157760
24
+ bigdl/cpp/libs/llama-bench,sha256=1rGYsGgstEJ0twUoSdxdAgwv1kLt0z5tHOrjoDSA9dw,14227512
25
+ bigdl/cpp/libs/llava-cli,sha256=stSrjvMvJ2MJEZPZXe54ugMjzJjNznmOhe98DosTObo,14448016
26
+ bigdl/cpp/libs/lookahead,sha256=bXWstxm3PMy0creOik1urGV__C1gtdsY7l7K7iBZjT4,14142224
27
+ bigdl/cpp/libs/lookup,sha256=5v1I-e6w_rt8v1TGOb9vs5uk8YZm7KYhTuaPKH3uPq8,14162000
28
+ bigdl/cpp/libs/ls-sycl-device,sha256=LSeaS1Nz_vlNZ-qo9G1hAwRWtZdBaDo1rSxBswqjtIA,12381264
29
+ bigdl/cpp/libs/main,sha256=L1Htc0j6aCj5PMTEN5CEQET-nIQDuWdYOozGAgK-dRI,14189344
30
+ bigdl/cpp/libs/ollama,sha256=2OtpLw2orQp8P5WFX3hlIQLIbDFr4ZLiO0xcfpRlV48,47268720
31
+ bigdl/cpp/libs/perplexity,sha256=a6omCm_5xdV-Wzl9xtEhuVKALB8Edh9Ou5Tiv9_aCXk,14230840
32
+ bigdl/cpp/libs/quantize,sha256=FbrXcPywU-i7jxaIsr3X2Nb1nzAFGVVGBSbh--CrBE4,14160232
33
+ bigdl/cpp/libs/quantize-stats,sha256=U4xnQP2gdjKzeRXM-Zd4na9rEfSxRSAeoZOV6N9auH4,13674472
34
+ bigdl/cpp/libs/save-load-state,sha256=7WTHFx-XtBuI5j97NKUDwAcT93tF8PQP6DGiSIOM5uI,14130032
35
+ bigdl/cpp/libs/server,sha256=zbJhr3Q1JDcab8xIhPnQNdqEhyuOKIfpw3fADDHZvC4,14987480
36
+ bigdl/cpp/libs/speculative,sha256=PhgB_7Sr2KgbkqB72y_vEr68wQHB-odkzBJ7genQWco,14168240
37
+ bigdl/cpp/libs/tokenize,sha256=cZvKK_zhyZGLqby3X_yt3NlMWSS_Faex2X16fPZK1qQ,14129944
38
+ bigdl_core_cpp-2.5.0b20240725.data/scripts/init-llama-cpp,sha256=qpIwMQc2Vb11fmovXFJKefp5iD1tQMfL8Ma7Y3Ebguk,400
39
+ bigdl_core_cpp-2.5.0b20240725.data/scripts/init-ollama,sha256=5eg4DsWg87iysFRh7leawr6H47F_sHydwL5we6fUG6o,193
40
+ bigdl_core_cpp-2.5.0b20240725.dist-info/METADATA,sha256=mGkEEI168LdjL_ypEiUAPUiD17EW5GuGFHMaHiZnU3g,650
41
+ bigdl_core_cpp-2.5.0b20240725.dist-info/WHEEL,sha256=GHycjhgos9e5ojAyWZiFxz9HM8k84hAls6WqrsH8row,109
42
+ bigdl_core_cpp-2.5.0b20240725.dist-info/top_level.txt,sha256=iGuLfZARD_qANcIMfy0tbbrC3EtCg6BSiH8icc3dLWs,6
43
+ bigdl_core_cpp-2.5.0b20240725.dist-info/RECORD,,
@@ -1,43 +0,0 @@
1
- bigdl/cpp/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
2
- bigdl/cpp/convert-hf-to-gguf.py,sha256=VbKd1bvq0MX12aIv27-tySszWIHw1Mj4et8I_XsLOFM,126986
3
- bigdl/cpp/convert.py,sha256=XMMcpfWHwEAAWzwLXe9mmJTU7cMvcyw8g2BFctfZnvI,69417
4
- bigdl/cpp/cli/init-llama-cpp,sha256=qpIwMQc2Vb11fmovXFJKefp5iD1tQMfL8Ma7Y3Ebguk,400
5
- bigdl/cpp/cli/init-ollama,sha256=5eg4DsWg87iysFRh7leawr6H47F_sHydwL5we6fUG6o,193
6
- bigdl/cpp/gguf-py/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
7
- bigdl/cpp/gguf-py/gguf/__init__.py,sha256=Xr2OoA3yQ8Z5GcbGcppNQ4Tw9zKMzPxTD0iI6qm-SRc,172
8
- bigdl/cpp/gguf-py/gguf/constants.py,sha256=WKonDaR_T0NlibX3p3mE4aDj0ohTjLVyGDnWdMHvOME,34361
9
- bigdl/cpp/gguf-py/gguf/gguf.py,sha256=8MDu7a0JEXhLUv_tjhYqDrWubVNc41cFvBYZbkZZenI,478
10
- bigdl/cpp/gguf-py/gguf/gguf_reader.py,sha256=ReWAZwL589DW_bKWix0LK2qaDWOj0r4jXPKBymoJ-B4,12028
11
- bigdl/cpp/gguf-py/gguf/gguf_writer.py,sha256=kr3biGPsSq3nBEQ4uDIrKJ9zfYOdIRMLOQI4fS6lcpk,20945
12
- bigdl/cpp/gguf-py/gguf/lazy.py,sha256=H7YCO73KFK_s1EZtO8-Cj98NJa7JaNPDaD6k1KeKBuA,9788
13
- bigdl/cpp/gguf-py/gguf/py.typed,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
14
- bigdl/cpp/gguf-py/gguf/quants.py,sha256=dI6JsSpaSzfdz7OwfqH1kvK98M_Ax8d6NMNYY5pyoQM,4354
15
- bigdl/cpp/gguf-py/gguf/tensor_mapping.py,sha256=WmyVE4nXKyQbfX3_qW4vyFWYX-5xgK_FL7V9oKIf8lY,23425
16
- bigdl/cpp/gguf-py/gguf/vocab.py,sha256=vsRrPQ6RafirhBmMp9h3YGZh_tqEjeg7S2hiLvMv7Ys,6847
17
- bigdl/cpp/libs/baby-llama,sha256=zbS1Jemem07ZyRa6AqrtI6L6UArS7Sma46TyvhG9zY8,13995576
18
- bigdl/cpp/libs/batched,sha256=xX0OrhX5Fp6g2V7BVoiO5kfWfY8pceDJHy4_m4Xy6R0,13920144
19
- bigdl/cpp/libs/batched-bench,sha256=Jd55uBgyHqCpjXtes8xdwK9_gAYcprOGAlZ1u_4vbJo,13920304
20
- bigdl/cpp/libs/benchmark,sha256=HW0ts7Q_ubjHRV6-eLZO65SLccVQDgykH_8hixJrM38,1452608
21
- bigdl/cpp/libs/embedding,sha256=LZAQKRzPCFSZcElbkXk_WHGYDe0P8R4K24rH-CTTc0I,13920288
22
- bigdl/cpp/libs/gguf,sha256=Xuni6618gX3JpgBLKByIpuFOmtp4KRLLp_odTzuueeA,12192232
23
- bigdl/cpp/libs/imatrix,sha256=kR3-tbfXtSOlBmcKVL1fyt-yS2NdhwxPG5ooY5KrJ00,13948144
24
- bigdl/cpp/libs/llama-bench,sha256=_6uyrwYRm1Y2RL3y67vpj7fRpm9g86neym35ZQrNkoA,14013784
25
- bigdl/cpp/libs/llava-cli,sha256=BOw5S9N0J5rBEe5K8ZbLBiCLML2HAklMLfLP-sUnS_I,14234296
26
- bigdl/cpp/libs/lookahead,sha256=h-LExzgThFUZrUoLgaQxe3zLvkrgIbRC9vf1BNm2ITU,13928512
27
- bigdl/cpp/libs/lookup,sha256=O73rdTq5Wsl_HR1CK7CSc3RcEetl2iAwY9AH0uXrTew,13952392
28
- bigdl/cpp/libs/ls-sycl-device,sha256=hnCBumqKkKl2UXWarux4dM9ZO0PbnrvV2uitSOADWHw,12183880
29
- bigdl/cpp/libs/main,sha256=5jtAQe_shtxGfwUHGILkFjovcLa_iKHICerKaxyPty4,13979736
30
- bigdl/cpp/libs/ollama,sha256=kEoF3x7YOyHaaIcrtGvauSlADQNBoeN9X8qd1C_QCqc,47108800
31
- bigdl/cpp/libs/perplexity,sha256=SbXNvHjtelbexfJ4YJp7XV16ErTD7Z7MKbk6xjMnZlQ,14021232
32
- bigdl/cpp/libs/quantize,sha256=cq5KL3K2XDK-8CaD313hPQrfLl3bFF_9-C2A2I9vUcg,13946520
33
- bigdl/cpp/libs/quantize-stats,sha256=s-uRhZtj6pivQgylGdza_ub7S3RlmQ7ZPCsk6ZkZsJ0,13460688
34
- bigdl/cpp/libs/save-load-state,sha256=ObWyyzsTs66ZxvoiXlig0mX8DeEsrUF2zxEzsdhZJyY,13920416
35
- bigdl/cpp/libs/server,sha256=s9Y0ILostMaYrufjw1cRJ9SfFxdrUjz148FLkgWoeFI,14773616
36
- bigdl/cpp/libs/speculative,sha256=tPYOVigtH5HwsZXeix-eCSNsIJ-dz3wvYQMffOscwE8,13954528
37
- bigdl/cpp/libs/tokenize,sha256=iCNhv897hGQAjyTrtQWklam5r-vhCr38_9925KLgRFc,13920336
38
- bigdl_core_cpp-2.5.0b20240723.data/scripts/init-llama-cpp,sha256=qpIwMQc2Vb11fmovXFJKefp5iD1tQMfL8Ma7Y3Ebguk,400
39
- bigdl_core_cpp-2.5.0b20240723.data/scripts/init-ollama,sha256=5eg4DsWg87iysFRh7leawr6H47F_sHydwL5we6fUG6o,193
40
- bigdl_core_cpp-2.5.0b20240723.dist-info/METADATA,sha256=Ml6lH2aQxFeaxgclAfg2oG56jX7Pvmw8LnLjP0i__uk,650
41
- bigdl_core_cpp-2.5.0b20240723.dist-info/WHEEL,sha256=GHycjhgos9e5ojAyWZiFxz9HM8k84hAls6WqrsH8row,109
42
- bigdl_core_cpp-2.5.0b20240723.dist-info/top_level.txt,sha256=iGuLfZARD_qANcIMfy0tbbrC3EtCg6BSiH8icc3dLWs,6
43
- bigdl_core_cpp-2.5.0b20240723.dist-info/RECORD,,