bigdl-core-cpp 2.5.0b20240724__py3-none-manylinux2010_x86_64.whl → 2.5.0b20240725__py3-none-manylinux2010_x86_64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- bigdl/cpp/convert-hf-to-gguf.py +47 -0
- bigdl/cpp/gguf-py/gguf/constants.py +27 -0
- bigdl/cpp/gguf-py/gguf/gguf_writer.py +3 -0
- bigdl/cpp/gguf-py/gguf/tensor_mapping.py +14 -0
- bigdl/cpp/libs/baby-llama +0 -0
- bigdl/cpp/libs/batched +0 -0
- bigdl/cpp/libs/batched-bench +0 -0
- bigdl/cpp/libs/benchmark +0 -0
- bigdl/cpp/libs/embedding +0 -0
- bigdl/cpp/libs/gguf +0 -0
- bigdl/cpp/libs/imatrix +0 -0
- bigdl/cpp/libs/llama-bench +0 -0
- bigdl/cpp/libs/llava-cli +0 -0
- bigdl/cpp/libs/lookahead +0 -0
- bigdl/cpp/libs/lookup +0 -0
- bigdl/cpp/libs/ls-sycl-device +0 -0
- bigdl/cpp/libs/main +0 -0
- bigdl/cpp/libs/ollama +0 -0
- bigdl/cpp/libs/perplexity +0 -0
- bigdl/cpp/libs/quantize +0 -0
- bigdl/cpp/libs/quantize-stats +0 -0
- bigdl/cpp/libs/save-load-state +0 -0
- bigdl/cpp/libs/server +0 -0
- bigdl/cpp/libs/speculative +0 -0
- bigdl/cpp/libs/tokenize +0 -0
- {bigdl_core_cpp-2.5.0b20240724.dist-info → bigdl_core_cpp-2.5.0b20240725.dist-info}/METADATA +1 -1
- bigdl_core_cpp-2.5.0b20240725.dist-info/RECORD +43 -0
- bigdl_core_cpp-2.5.0b20240724.dist-info/RECORD +0 -43
- {bigdl_core_cpp-2.5.0b20240724.data → bigdl_core_cpp-2.5.0b20240725.data}/scripts/init-llama-cpp +0 -0
- {bigdl_core_cpp-2.5.0b20240724.data → bigdl_core_cpp-2.5.0b20240725.data}/scripts/init-ollama +0 -0
- {bigdl_core_cpp-2.5.0b20240724.dist-info → bigdl_core_cpp-2.5.0b20240725.dist-info}/WHEEL +0 -0
- {bigdl_core_cpp-2.5.0b20240724.dist-info → bigdl_core_cpp-2.5.0b20240725.dist-info}/top_level.txt +0 -0
bigdl/cpp/convert-hf-to-gguf.py
CHANGED
@@ -2342,6 +2342,53 @@ class GemmaModel(Model):
|
|
2342
2342
|
return [(self.map_tensor_name(name), data_torch)]
|
2343
2343
|
|
2344
2344
|
|
2345
|
+
@Model.register("Gemma2ForCausalLM")
|
2346
|
+
class Gemma2Model(Model):
|
2347
|
+
model_arch = gguf.MODEL_ARCH.GEMMA2
|
2348
|
+
|
2349
|
+
def set_vocab(self):
|
2350
|
+
self._set_vocab_llama_hf()
|
2351
|
+
self.gguf_writer.add_add_space_prefix(False)
|
2352
|
+
|
2353
|
+
def set_gguf_parameters(self):
|
2354
|
+
hparams = self.hparams
|
2355
|
+
block_count = hparams["num_hidden_layers"]
|
2356
|
+
|
2357
|
+
self.gguf_writer.add_name(self.dir_model.name if self.model_name is None else self.model_name)
|
2358
|
+
self.gguf_writer.add_context_length(hparams["max_position_embeddings"])
|
2359
|
+
self.gguf_writer.add_embedding_length(hparams["hidden_size"])
|
2360
|
+
self.gguf_writer.add_block_count(block_count)
|
2361
|
+
self.gguf_writer.add_feed_forward_length(hparams["intermediate_size"])
|
2362
|
+
self.gguf_writer.add_head_count(hparams["num_attention_heads"])
|
2363
|
+
self.gguf_writer.add_head_count_kv(self.hparams["num_key_value_heads"] if "num_key_value_heads" in hparams else hparams["num_attention_heads"])
|
2364
|
+
self.gguf_writer.add_layer_norm_rms_eps(self.hparams["rms_norm_eps"])
|
2365
|
+
self.gguf_writer.add_key_length(hparams["head_dim"])
|
2366
|
+
self.gguf_writer.add_value_length(hparams["head_dim"])
|
2367
|
+
self.gguf_writer.add_file_type(self.ftype)
|
2368
|
+
self.gguf_writer.add_attn_logit_softcapping(
|
2369
|
+
self.hparams["attn_logit_softcapping"]
|
2370
|
+
)
|
2371
|
+
self.gguf_writer.add_final_logit_softcapping(
|
2372
|
+
self.hparams["final_logit_softcapping"]
|
2373
|
+
)
|
2374
|
+
self.gguf_writer.add_sliding_window(self.hparams["sliding_window"])
|
2375
|
+
|
2376
|
+
def modify_tensors(self, data_torch: Tensor, name: str, bid: int | None) -> Iterable[tuple[str, Tensor]]:
|
2377
|
+
del bid # unusem
|
2378
|
+
|
2379
|
+
# lm_head is not used in llama.cpp, while autoawq will include this tensor in model
|
2380
|
+
# To prevent errors, skip loading lm_head.weight.
|
2381
|
+
if name == "lm_head.weight":
|
2382
|
+
logger.debug(f"Skipping get tensor {name!r} in safetensors so that convert can end normally.")
|
2383
|
+
return []
|
2384
|
+
|
2385
|
+
# ref: https://github.com/huggingface/transformers/blob/fc37f38915372c15992b540dfcbbe00a916d4fc6/src/transformers/models/gemma/modeling_gemma.py#L89
|
2386
|
+
if name.endswith("norm.weight"):
|
2387
|
+
data_torch = data_torch + 1
|
2388
|
+
|
2389
|
+
return [(self.map_tensor_name(name), data_torch)]
|
2390
|
+
|
2391
|
+
|
2345
2392
|
@Model.register("Starcoder2ForCausalLM")
|
2346
2393
|
class StarCoder2Model(Model):
|
2347
2394
|
model_arch = gguf.MODEL_ARCH.STARCODER2
|
@@ -55,6 +55,10 @@ class Keys:
|
|
55
55
|
LAYERNORM_EPS = "{arch}.attention.layer_norm_epsilon"
|
56
56
|
LAYERNORM_RMS_EPS = "{arch}.attention.layer_norm_rms_epsilon"
|
57
57
|
CAUSAL = "{arch}.attention.causal"
|
58
|
+
Q_LORA_RANK = "{arch}.attention.q_lora_rank"
|
59
|
+
KV_LORA_RANK = "{arch}.attention.kv_lora_rank"
|
60
|
+
REL_BUCKETS_COUNT = "{arch}.attention.relative_buckets_count"
|
61
|
+
SLIDING_WINDOW = "{arch}.attention.sliding_window"
|
58
62
|
|
59
63
|
class Rope:
|
60
64
|
DIMENSION_COUNT = "{arch}.rope.dimension_count"
|
@@ -133,6 +137,7 @@ class MODEL_ARCH(IntEnum):
|
|
133
137
|
INTERNLM2 = auto()
|
134
138
|
MINICPM = auto()
|
135
139
|
GEMMA = auto()
|
140
|
+
GEMMA2 = auto()
|
136
141
|
STARCODER2 = auto()
|
137
142
|
MAMBA = auto()
|
138
143
|
XVERSE = auto()
|
@@ -160,10 +165,13 @@ class MODEL_TENSOR(IntEnum):
|
|
160
165
|
ATTN_NORM = auto()
|
161
166
|
ATTN_NORM_2 = auto()
|
162
167
|
ATTN_OUT_NORM = auto()
|
168
|
+
ATTN_POST_NORM = auto()
|
163
169
|
ATTN_ROT_EMBD = auto()
|
164
170
|
FFN_GATE_INP = auto()
|
165
171
|
FFN_GATE_INP_SHEXP = auto()
|
166
172
|
FFN_NORM = auto()
|
173
|
+
FFN_PRE_NORM = auto()
|
174
|
+
FFN_POST_NORM = auto()
|
167
175
|
FFN_GATE = auto()
|
168
176
|
FFN_DOWN = auto()
|
169
177
|
FFN_UP = auto()
|
@@ -214,6 +222,7 @@ MODEL_ARCH_NAMES: dict[MODEL_ARCH, str] = {
|
|
214
222
|
MODEL_ARCH.INTERNLM2: "internlm2",
|
215
223
|
MODEL_ARCH.MINICPM: "minicpm",
|
216
224
|
MODEL_ARCH.GEMMA: "gemma",
|
225
|
+
MODEL_ARCH.GEMMA2: "gemma2",
|
217
226
|
MODEL_ARCH.STARCODER2: "starcoder2",
|
218
227
|
MODEL_ARCH.MAMBA: "mamba",
|
219
228
|
MODEL_ARCH.XVERSE: "xverse",
|
@@ -244,9 +253,12 @@ TENSOR_NAMES: dict[MODEL_TENSOR, str] = {
|
|
244
253
|
MODEL_TENSOR.ATTN_Q_NORM: "blk.{bid}.attn_q_norm",
|
245
254
|
MODEL_TENSOR.ATTN_K_NORM: "blk.{bid}.attn_k_norm",
|
246
255
|
MODEL_TENSOR.ATTN_OUT_NORM: "blk.{bid}.attn_output_norm",
|
256
|
+
MODEL_TENSOR.ATTN_POST_NORM: "blk.{bid}.post_attention_norm",
|
247
257
|
MODEL_TENSOR.FFN_GATE_INP: "blk.{bid}.ffn_gate_inp",
|
248
258
|
MODEL_TENSOR.FFN_GATE_INP_SHEXP: "blk.{bid}.ffn_gate_inp_shexp",
|
249
259
|
MODEL_TENSOR.FFN_NORM: "blk.{bid}.ffn_norm",
|
260
|
+
MODEL_TENSOR.FFN_PRE_NORM: "blk.{bid}.ffn_norm",
|
261
|
+
MODEL_TENSOR.FFN_POST_NORM: "blk.{bid}.post_ffw_norm",
|
250
262
|
MODEL_TENSOR.FFN_GATE: "blk.{bid}.ffn_gate",
|
251
263
|
MODEL_TENSOR.FFN_DOWN: "blk.{bid}.ffn_down",
|
252
264
|
MODEL_TENSOR.FFN_UP: "blk.{bid}.ffn_up",
|
@@ -654,6 +666,21 @@ MODEL_TENSORS: dict[MODEL_ARCH, list[MODEL_TENSOR]] = {
|
|
654
666
|
MODEL_TENSOR.FFN_UP,
|
655
667
|
MODEL_TENSOR.FFN_NORM,
|
656
668
|
],
|
669
|
+
MODEL_ARCH.GEMMA2: [
|
670
|
+
MODEL_TENSOR.TOKEN_EMBD,
|
671
|
+
MODEL_TENSOR.OUTPUT_NORM,
|
672
|
+
MODEL_TENSOR.ATTN_Q,
|
673
|
+
MODEL_TENSOR.ATTN_K,
|
674
|
+
MODEL_TENSOR.ATTN_V,
|
675
|
+
MODEL_TENSOR.ATTN_OUT,
|
676
|
+
MODEL_TENSOR.FFN_GATE,
|
677
|
+
MODEL_TENSOR.FFN_DOWN,
|
678
|
+
MODEL_TENSOR.FFN_UP,
|
679
|
+
MODEL_TENSOR.ATTN_NORM,
|
680
|
+
MODEL_TENSOR.ATTN_POST_NORM,
|
681
|
+
MODEL_TENSOR.FFN_PRE_NORM,
|
682
|
+
MODEL_TENSOR.FFN_POST_NORM,
|
683
|
+
],
|
657
684
|
MODEL_ARCH.STARCODER2: [
|
658
685
|
MODEL_TENSOR.TOKEN_EMBD,
|
659
686
|
MODEL_TENSOR.OUTPUT_NORM,
|
@@ -416,6 +416,9 @@ class GGUFWriter:
|
|
416
416
|
def add_causal_attention(self, value: bool) -> None:
|
417
417
|
self.add_bool(Keys.Attention.CAUSAL.format(arch=self.arch), value)
|
418
418
|
|
419
|
+
def add_sliding_window(self, value: int) -> None:
|
420
|
+
self.add_uint32(Keys.Attention.SLIDING_WINDOW.format(arch=self.arch), value)
|
421
|
+
|
419
422
|
def add_pooling_type(self, value: PoolingType) -> None:
|
420
423
|
self.add_uint32(Keys.LLM.POOLING_TYPE.format(arch=self.arch), value.value)
|
421
424
|
|
@@ -185,6 +185,10 @@ class TensorNameMap:
|
|
185
185
|
"transformer.blocks.{bid}.norm_attn_norm.norm_2", # dbrx
|
186
186
|
),
|
187
187
|
|
188
|
+
MODEL_TENSOR.ATTN_POST_NORM: (
|
189
|
+
"model.layers.{bid}.post_attention_layernorm", # gemma2
|
190
|
+
),
|
191
|
+
|
188
192
|
# Rotary embeddings
|
189
193
|
MODEL_TENSOR.ATTN_ROT_EMBD: (
|
190
194
|
"model.layers.{bid}.self_attn.rotary_emb.inv_freq", # llama-hf
|
@@ -208,6 +212,16 @@ class TensorNameMap:
|
|
208
212
|
"transformer.decoder_layer.{bid}.rms_norm_2", # Grok
|
209
213
|
),
|
210
214
|
|
215
|
+
# Post feed-forward norm
|
216
|
+
MODEL_TENSOR.FFN_PRE_NORM: (
|
217
|
+
"model.layers.{bid}.pre_feedforward_layernorm", # gemma2
|
218
|
+
),
|
219
|
+
|
220
|
+
# Post feed-forward norm
|
221
|
+
MODEL_TENSOR.FFN_POST_NORM: (
|
222
|
+
"model.layers.{bid}.post_feedforward_layernorm", # gemma2
|
223
|
+
),
|
224
|
+
|
211
225
|
MODEL_TENSOR.FFN_GATE_INP: (
|
212
226
|
"layers.{bid}.feed_forward.gate", # mixtral
|
213
227
|
"model.layers.{bid}.block_sparse_moe.gate", # mixtral
|
bigdl/cpp/libs/baby-llama
CHANGED
Binary file
|
bigdl/cpp/libs/batched
CHANGED
Binary file
|
bigdl/cpp/libs/batched-bench
CHANGED
Binary file
|
bigdl/cpp/libs/benchmark
CHANGED
Binary file
|
bigdl/cpp/libs/embedding
CHANGED
Binary file
|
bigdl/cpp/libs/gguf
CHANGED
Binary file
|
bigdl/cpp/libs/imatrix
CHANGED
Binary file
|
bigdl/cpp/libs/llama-bench
CHANGED
Binary file
|
bigdl/cpp/libs/llava-cli
CHANGED
Binary file
|
bigdl/cpp/libs/lookahead
CHANGED
Binary file
|
bigdl/cpp/libs/lookup
CHANGED
Binary file
|
bigdl/cpp/libs/ls-sycl-device
CHANGED
Binary file
|
bigdl/cpp/libs/main
CHANGED
Binary file
|
bigdl/cpp/libs/ollama
CHANGED
Binary file
|
bigdl/cpp/libs/perplexity
CHANGED
Binary file
|
bigdl/cpp/libs/quantize
CHANGED
Binary file
|
bigdl/cpp/libs/quantize-stats
CHANGED
Binary file
|
bigdl/cpp/libs/save-load-state
CHANGED
Binary file
|
bigdl/cpp/libs/server
CHANGED
Binary file
|
bigdl/cpp/libs/speculative
CHANGED
Binary file
|
bigdl/cpp/libs/tokenize
CHANGED
Binary file
|
@@ -0,0 +1,43 @@
|
|
1
|
+
bigdl/cpp/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
2
|
+
bigdl/cpp/convert-hf-to-gguf.py,sha256=ZQWPwgfgf-BCQMyZUHWbGuFNx4XP7KRhBTflr4ZfKO0,129245
|
3
|
+
bigdl/cpp/convert.py,sha256=XMMcpfWHwEAAWzwLXe9mmJTU7cMvcyw8g2BFctfZnvI,69417
|
4
|
+
bigdl/cpp/cli/init-llama-cpp,sha256=qpIwMQc2Vb11fmovXFJKefp5iD1tQMfL8Ma7Y3Ebguk,400
|
5
|
+
bigdl/cpp/cli/init-ollama,sha256=5eg4DsWg87iysFRh7leawr6H47F_sHydwL5we6fUG6o,193
|
6
|
+
bigdl/cpp/gguf-py/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
7
|
+
bigdl/cpp/gguf-py/gguf/__init__.py,sha256=Xr2OoA3yQ8Z5GcbGcppNQ4Tw9zKMzPxTD0iI6qm-SRc,172
|
8
|
+
bigdl/cpp/gguf-py/gguf/constants.py,sha256=2PStnJUUAoAuc0Se_boiGgpC2r0YG7Eai_gPEvQq8k0,35416
|
9
|
+
bigdl/cpp/gguf-py/gguf/gguf.py,sha256=8MDu7a0JEXhLUv_tjhYqDrWubVNc41cFvBYZbkZZenI,478
|
10
|
+
bigdl/cpp/gguf-py/gguf/gguf_reader.py,sha256=ReWAZwL589DW_bKWix0LK2qaDWOj0r4jXPKBymoJ-B4,12028
|
11
|
+
bigdl/cpp/gguf-py/gguf/gguf_writer.py,sha256=S-MxxZWJ-Vs_aCIEDVuyHo4W8bUtKwkUL3CTEzmNpR0,21085
|
12
|
+
bigdl/cpp/gguf-py/gguf/lazy.py,sha256=H7YCO73KFK_s1EZtO8-Cj98NJa7JaNPDaD6k1KeKBuA,9788
|
13
|
+
bigdl/cpp/gguf-py/gguf/py.typed,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
14
|
+
bigdl/cpp/gguf-py/gguf/quants.py,sha256=dI6JsSpaSzfdz7OwfqH1kvK98M_Ax8d6NMNYY5pyoQM,4354
|
15
|
+
bigdl/cpp/gguf-py/gguf/tensor_mapping.py,sha256=fFeDJ7HQQm-hEyq1698sUVxrCf01hTaKR3nXnR6cJcc,23852
|
16
|
+
bigdl/cpp/gguf-py/gguf/vocab.py,sha256=vsRrPQ6RafirhBmMp9h3YGZh_tqEjeg7S2hiLvMv7Ys,6847
|
17
|
+
bigdl/cpp/libs/baby-llama,sha256=Se4hMft0RmV0hQSdOy48Q_-FqN0y9tgdCI6xR4LDmnI,14209296
|
18
|
+
bigdl/cpp/libs/batched,sha256=DUvxfARun7uyUfKuVPtJ3SexJx9nZhvMeefmUyneiyo,14133856
|
19
|
+
bigdl/cpp/libs/batched-bench,sha256=uNp3FNIgLqg0vU7-AuSiyrEoy9gJTZybR30taIwHy1o,14134016
|
20
|
+
bigdl/cpp/libs/benchmark,sha256=P1dPH6y_ishJS6R4XpRfeUJCArRPTE4Z_voYbSLddsQ,1452656
|
21
|
+
bigdl/cpp/libs/embedding,sha256=FM5UWjOmprMdcqHSgRElUixLB0YLP78EdFWDhZpQz4o,14129896
|
22
|
+
bigdl/cpp/libs/gguf,sha256=zQn08SRpEPmJKO0_7sE3TIb3oTUVmruBOeaOlYYvbnA,12389616
|
23
|
+
bigdl/cpp/libs/imatrix,sha256=LN7Im5SJRKQBaI4NN6V55gyfH92PrnT56l4Z097hGhw,14157760
|
24
|
+
bigdl/cpp/libs/llama-bench,sha256=1rGYsGgstEJ0twUoSdxdAgwv1kLt0z5tHOrjoDSA9dw,14227512
|
25
|
+
bigdl/cpp/libs/llava-cli,sha256=stSrjvMvJ2MJEZPZXe54ugMjzJjNznmOhe98DosTObo,14448016
|
26
|
+
bigdl/cpp/libs/lookahead,sha256=bXWstxm3PMy0creOik1urGV__C1gtdsY7l7K7iBZjT4,14142224
|
27
|
+
bigdl/cpp/libs/lookup,sha256=5v1I-e6w_rt8v1TGOb9vs5uk8YZm7KYhTuaPKH3uPq8,14162000
|
28
|
+
bigdl/cpp/libs/ls-sycl-device,sha256=LSeaS1Nz_vlNZ-qo9G1hAwRWtZdBaDo1rSxBswqjtIA,12381264
|
29
|
+
bigdl/cpp/libs/main,sha256=L1Htc0j6aCj5PMTEN5CEQET-nIQDuWdYOozGAgK-dRI,14189344
|
30
|
+
bigdl/cpp/libs/ollama,sha256=2OtpLw2orQp8P5WFX3hlIQLIbDFr4ZLiO0xcfpRlV48,47268720
|
31
|
+
bigdl/cpp/libs/perplexity,sha256=a6omCm_5xdV-Wzl9xtEhuVKALB8Edh9Ou5Tiv9_aCXk,14230840
|
32
|
+
bigdl/cpp/libs/quantize,sha256=FbrXcPywU-i7jxaIsr3X2Nb1nzAFGVVGBSbh--CrBE4,14160232
|
33
|
+
bigdl/cpp/libs/quantize-stats,sha256=U4xnQP2gdjKzeRXM-Zd4na9rEfSxRSAeoZOV6N9auH4,13674472
|
34
|
+
bigdl/cpp/libs/save-load-state,sha256=7WTHFx-XtBuI5j97NKUDwAcT93tF8PQP6DGiSIOM5uI,14130032
|
35
|
+
bigdl/cpp/libs/server,sha256=zbJhr3Q1JDcab8xIhPnQNdqEhyuOKIfpw3fADDHZvC4,14987480
|
36
|
+
bigdl/cpp/libs/speculative,sha256=PhgB_7Sr2KgbkqB72y_vEr68wQHB-odkzBJ7genQWco,14168240
|
37
|
+
bigdl/cpp/libs/tokenize,sha256=cZvKK_zhyZGLqby3X_yt3NlMWSS_Faex2X16fPZK1qQ,14129944
|
38
|
+
bigdl_core_cpp-2.5.0b20240725.data/scripts/init-llama-cpp,sha256=qpIwMQc2Vb11fmovXFJKefp5iD1tQMfL8Ma7Y3Ebguk,400
|
39
|
+
bigdl_core_cpp-2.5.0b20240725.data/scripts/init-ollama,sha256=5eg4DsWg87iysFRh7leawr6H47F_sHydwL5we6fUG6o,193
|
40
|
+
bigdl_core_cpp-2.5.0b20240725.dist-info/METADATA,sha256=mGkEEI168LdjL_ypEiUAPUiD17EW5GuGFHMaHiZnU3g,650
|
41
|
+
bigdl_core_cpp-2.5.0b20240725.dist-info/WHEEL,sha256=GHycjhgos9e5ojAyWZiFxz9HM8k84hAls6WqrsH8row,109
|
42
|
+
bigdl_core_cpp-2.5.0b20240725.dist-info/top_level.txt,sha256=iGuLfZARD_qANcIMfy0tbbrC3EtCg6BSiH8icc3dLWs,6
|
43
|
+
bigdl_core_cpp-2.5.0b20240725.dist-info/RECORD,,
|
@@ -1,43 +0,0 @@
|
|
1
|
-
bigdl/cpp/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
2
|
-
bigdl/cpp/convert-hf-to-gguf.py,sha256=VbKd1bvq0MX12aIv27-tySszWIHw1Mj4et8I_XsLOFM,126986
|
3
|
-
bigdl/cpp/convert.py,sha256=XMMcpfWHwEAAWzwLXe9mmJTU7cMvcyw8g2BFctfZnvI,69417
|
4
|
-
bigdl/cpp/cli/init-llama-cpp,sha256=qpIwMQc2Vb11fmovXFJKefp5iD1tQMfL8Ma7Y3Ebguk,400
|
5
|
-
bigdl/cpp/cli/init-ollama,sha256=5eg4DsWg87iysFRh7leawr6H47F_sHydwL5we6fUG6o,193
|
6
|
-
bigdl/cpp/gguf-py/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
7
|
-
bigdl/cpp/gguf-py/gguf/__init__.py,sha256=Xr2OoA3yQ8Z5GcbGcppNQ4Tw9zKMzPxTD0iI6qm-SRc,172
|
8
|
-
bigdl/cpp/gguf-py/gguf/constants.py,sha256=WKonDaR_T0NlibX3p3mE4aDj0ohTjLVyGDnWdMHvOME,34361
|
9
|
-
bigdl/cpp/gguf-py/gguf/gguf.py,sha256=8MDu7a0JEXhLUv_tjhYqDrWubVNc41cFvBYZbkZZenI,478
|
10
|
-
bigdl/cpp/gguf-py/gguf/gguf_reader.py,sha256=ReWAZwL589DW_bKWix0LK2qaDWOj0r4jXPKBymoJ-B4,12028
|
11
|
-
bigdl/cpp/gguf-py/gguf/gguf_writer.py,sha256=kr3biGPsSq3nBEQ4uDIrKJ9zfYOdIRMLOQI4fS6lcpk,20945
|
12
|
-
bigdl/cpp/gguf-py/gguf/lazy.py,sha256=H7YCO73KFK_s1EZtO8-Cj98NJa7JaNPDaD6k1KeKBuA,9788
|
13
|
-
bigdl/cpp/gguf-py/gguf/py.typed,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
14
|
-
bigdl/cpp/gguf-py/gguf/quants.py,sha256=dI6JsSpaSzfdz7OwfqH1kvK98M_Ax8d6NMNYY5pyoQM,4354
|
15
|
-
bigdl/cpp/gguf-py/gguf/tensor_mapping.py,sha256=WmyVE4nXKyQbfX3_qW4vyFWYX-5xgK_FL7V9oKIf8lY,23425
|
16
|
-
bigdl/cpp/gguf-py/gguf/vocab.py,sha256=vsRrPQ6RafirhBmMp9h3YGZh_tqEjeg7S2hiLvMv7Ys,6847
|
17
|
-
bigdl/cpp/libs/baby-llama,sha256=S0YJ5ivOQPJTzUk_5p593W_j4-GbUCiS2qmVpuiKe2w,14197008
|
18
|
-
bigdl/cpp/libs/batched,sha256=RJPZVTZ7oJRvm4hL-HlyLlMJn0scecQRLAafd9CJxYU,14121568
|
19
|
-
bigdl/cpp/libs/batched-bench,sha256=rT5RDmwyJV4IC4MeZOGibw2vdUqFtcPKp-0OYeiEQDM,14121728
|
20
|
-
bigdl/cpp/libs/benchmark,sha256=BIwIVKlTx9JVLprtiyMp4hc8ThkSnR8u7unh1Yn0Hxs,1452656
|
21
|
-
bigdl/cpp/libs/embedding,sha256=tK6obvYoZS2hbafeLPxKRa-nrEl4ZH1JYX_UcYqFcro,14117608
|
22
|
-
bigdl/cpp/libs/gguf,sha256=6pF3exBdfUHBddai4Q_k9WBSpHFICvKimZZon6zygdw,12389616
|
23
|
-
bigdl/cpp/libs/imatrix,sha256=oe78J3U2cngeptMYKCZc9T9E0w3HKUI9ak7OT53WfMs,14145472
|
24
|
-
bigdl/cpp/libs/llama-bench,sha256=m5zjcg6qMHCx2-ABjiLkujBSDx0RvGxl3VdXZPqHasc,14211128
|
25
|
-
bigdl/cpp/libs/llava-cli,sha256=5xM3GPn_t38mDZ7xZeAQ-pkovqiIxC4nW81bmVNFDjc,14435728
|
26
|
-
bigdl/cpp/libs/lookahead,sha256=9FOKKO3oOIxqHDxqVsGnLIN9wrVL5w8SkXbsBqYZxhM,14129936
|
27
|
-
bigdl/cpp/libs/lookup,sha256=2-XOdZS3BB5ur_12a20YFh-EMnNOywqVLr5qz-ePxrc,14149712
|
28
|
-
bigdl/cpp/libs/ls-sycl-device,sha256=qfXUiLZRG6X8SG2eHMBp_xR8rlyqake8PzVOb0tICQU,12381264
|
29
|
-
bigdl/cpp/libs/main,sha256=OhTISrEzJp0OOXhaVtEzaqbVuUwsSk5IMQKZxfFLPU0,14177056
|
30
|
-
bigdl/cpp/libs/ollama,sha256=jdQhFh7AtGaTvQm-8uWhWtiGo_qNQJxj7nY7jWb-qjk,47248128
|
31
|
-
bigdl/cpp/libs/perplexity,sha256=I5-Kyd9f2Q-co6mgToxTvEew1BnddMtPc5Nab8U12ro,14218552
|
32
|
-
bigdl/cpp/libs/quantize,sha256=mF34TE3caGRz9bgkwV4EcpzeGezntFPr2lnoRs6T7Ug,14147944
|
33
|
-
bigdl/cpp/libs/quantize-stats,sha256=ZMhmJP4NaoXnIrbfOZSRZmhjQYiBNu_Fw7JOYdxCJ2Q,13658088
|
34
|
-
bigdl/cpp/libs/save-load-state,sha256=MPJwwctUA6kSNyG03Vzjx0QykDvyfAfjAbaKmn_NBss,14117744
|
35
|
-
bigdl/cpp/libs/server,sha256=ITkVqfpRSzni8A3YtGgIHIRqi-yYzY2ZBNhpXr1vMTk,14975192
|
36
|
-
bigdl/cpp/libs/speculative,sha256=vbXKQIzpWKo5vhFWzjkPpmhLTTo_E18RTpilhoADvg8,14155952
|
37
|
-
bigdl/cpp/libs/tokenize,sha256=3awbHB_pizhlUkQKRpnCYhJxvKpH2O1OwxFv9ETtM-o,14117656
|
38
|
-
bigdl_core_cpp-2.5.0b20240724.data/scripts/init-llama-cpp,sha256=qpIwMQc2Vb11fmovXFJKefp5iD1tQMfL8Ma7Y3Ebguk,400
|
39
|
-
bigdl_core_cpp-2.5.0b20240724.data/scripts/init-ollama,sha256=5eg4DsWg87iysFRh7leawr6H47F_sHydwL5we6fUG6o,193
|
40
|
-
bigdl_core_cpp-2.5.0b20240724.dist-info/METADATA,sha256=Wp0x0u-qDRxef2bEHMBl3yhY3uPTrNWypNFzT4tqUQ8,650
|
41
|
-
bigdl_core_cpp-2.5.0b20240724.dist-info/WHEEL,sha256=GHycjhgos9e5ojAyWZiFxz9HM8k84hAls6WqrsH8row,109
|
42
|
-
bigdl_core_cpp-2.5.0b20240724.dist-info/top_level.txt,sha256=iGuLfZARD_qANcIMfy0tbbrC3EtCg6BSiH8icc3dLWs,6
|
43
|
-
bigdl_core_cpp-2.5.0b20240724.dist-info/RECORD,,
|
{bigdl_core_cpp-2.5.0b20240724.data → bigdl_core_cpp-2.5.0b20240725.data}/scripts/init-llama-cpp
RENAMED
File without changes
|
{bigdl_core_cpp-2.5.0b20240724.data → bigdl_core_cpp-2.5.0b20240725.data}/scripts/init-ollama
RENAMED
File without changes
|
File without changes
|
{bigdl_core_cpp-2.5.0b20240724.dist-info → bigdl_core_cpp-2.5.0b20240725.dist-info}/top_level.txt
RENAMED
File without changes
|