ipex-llm 2.2.0b20250106__py3-none-win_amd64.whl → 2.2.0b20250107__py3-none-win_amd64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- ipex_llm/libs/bloom-api.dll +0 -0
- ipex_llm/libs/bloom.dll +0 -0
- ipex_llm/libs/gptneox-api.dll +0 -0
- ipex_llm/libs/gptneox.dll +0 -0
- ipex_llm/libs/libbloom_avx.dll +0 -0
- ipex_llm/libs/libbloom_vnni.dll +0 -0
- ipex_llm/libs/libgptneox_avx.dll +0 -0
- ipex_llm/libs/libgptneox_vnni.dll +0 -0
- ipex_llm/libs/libllama_avx.dll +0 -0
- ipex_llm/libs/libllama_vnni.dll +0 -0
- ipex_llm/libs/libstarcoder_avx.dll +0 -0
- ipex_llm/libs/libstarcoder_vnni.dll +0 -0
- ipex_llm/libs/llama-api.dll +0 -0
- ipex_llm/libs/llama.dll +0 -0
- ipex_llm/libs/main-bloom.exe +0 -0
- ipex_llm/libs/main-gptneox.exe +0 -0
- ipex_llm/libs/main-llama.exe +0 -0
- ipex_llm/libs/main-starcoder.exe +0 -0
- ipex_llm/libs/pipeline.dll +0 -0
- ipex_llm/libs/quantize-bloom.exe +0 -0
- ipex_llm/libs/quantize-bloom_vnni.exe +0 -0
- ipex_llm/libs/quantize-gptneox.exe +0 -0
- ipex_llm/libs/quantize-gptneox_vnni.exe +0 -0
- ipex_llm/libs/quantize-llama.exe +0 -0
- ipex_llm/libs/quantize-llama_vnni.exe +0 -0
- ipex_llm/libs/quantize-starcoder.exe +0 -0
- ipex_llm/libs/quantize-starcoder_vnni.exe +0 -0
- ipex_llm/libs/starcoder-api.dll +0 -0
- ipex_llm/libs/starcoder.dll +0 -0
- ipex_llm/transformers/convert.py +17 -132
- ipex_llm/transformers/lookup.py +2 -2
- ipex_llm/transformers/low_bit_linear.py +8 -8
- ipex_llm/transformers/models/chatglm2.py +1 -192
- ipex_llm/transformers/models/minicpmv.py +2 -2
- ipex_llm/transformers/models/sd.py +2 -2
- ipex_llm/transformers/models/utils.py +14 -89
- ipex_llm/transformers/npu_pipeline_model/convert_pipeline.py +5 -8
- ipex_llm/transformers/utils.py +5 -20
- {ipex_llm-2.2.0b20250106.dist-info → ipex_llm-2.2.0b20250107.dist-info}/METADATA +40 -19
- {ipex_llm-2.2.0b20250106.dist-info → ipex_llm-2.2.0b20250107.dist-info}/RECORD +46 -49
- ipex_llm/transformers/models/cohere.py +0 -589
- ipex_llm/transformers/models/falcon.py +0 -829
- ipex_llm/transformers/models/mixtral.py +0 -576
- {ipex_llm-2.2.0b20250106.data → ipex_llm-2.2.0b20250107.data}/scripts/ipex-llm-init.bat +0 -0
- {ipex_llm-2.2.0b20250106.data → ipex_llm-2.2.0b20250107.data}/scripts/llm-chat.ps1 +0 -0
- {ipex_llm-2.2.0b20250106.data → ipex_llm-2.2.0b20250107.data}/scripts/llm-cli.ps1 +0 -0
- {ipex_llm-2.2.0b20250106.dist-info → ipex_llm-2.2.0b20250107.dist-info}/WHEEL +0 -0
- {ipex_llm-2.2.0b20250106.dist-info → ipex_llm-2.2.0b20250107.dist-info}/entry_points.txt +0 -0
- {ipex_llm-2.2.0b20250106.dist-info → ipex_llm-2.2.0b20250107.dist-info}/top_level.txt +0 -0
@@ -19,7 +19,7 @@ import torch
|
|
19
19
|
import warnings
|
20
20
|
from ipex_llm.utils.common import invalidInputError
|
21
21
|
from ipex_llm.ggml.quantize import ggml_tensor_qtype
|
22
|
-
from ipex_llm.transformers.utils import get_ipex_version,
|
22
|
+
from ipex_llm.transformers.utils import get_ipex_version, get_xpu_device_name
|
23
23
|
from ipex_llm.transformers.low_bit_linear import SYM_INT4, SYM_INT8, FP8E5, IQ2_XXS, FP4, FP8E4,\
|
24
24
|
FP6, ASYM_INT4
|
25
25
|
|
@@ -85,16 +85,14 @@ def use_quantize_kv_cache(linear: torch.nn.Module, x: torch.Tensor, kv_group: in
|
|
85
85
|
return os.environ["IPEX_LLM_QUANTIZE_KV_CACHE"] == "1"
|
86
86
|
elif os.environ.get("IPEX_LLM_LOW_MEM", None) is not None:
|
87
87
|
return os.environ["IPEX_LLM_LOW_MEM"] == "1"
|
88
|
+
elif linear.qtype in [ggml_tensor_qtype["fp16"], ggml_tensor_qtype["bf16"]]:
|
89
|
+
return False
|
88
90
|
else:
|
89
|
-
|
90
|
-
|
91
|
-
|
92
|
-
|
93
|
-
|
94
|
-
def kv_cache_device_check(x: torch.Tensor, kv_group: int) -> bool:
|
95
|
-
return (get_xpu_device_type(x) in ["mtl", "lnl"] and kv_group <= 1) or \
|
96
|
-
((get_xpu_device_type(x) == "arc" or get_xpu_device_type(x) == "flex") and
|
97
|
-
1 < x.size(0) and x.size(0) <= 8)
|
91
|
+
device_name = get_xpu_device_name(x.device)
|
92
|
+
return (
|
93
|
+
device_name in ["mtl", "lnl", "arl"] and kv_group == 1
|
94
|
+
or device_name in ["arc", "bmg"] and x.size(0) > 1
|
95
|
+
)
|
98
96
|
|
99
97
|
|
100
98
|
def init_fp8_kv_cache(batch_size, num_heads, current_length, head_dim, device):
|
@@ -226,57 +224,6 @@ def is_enough_kv_cache_room_4_31(past_key_value, seq_len=1):
|
|
226
224
|
(past_key_value[0].size(2) + seq_len) * past_key_value[0].size(3)
|
227
225
|
|
228
226
|
|
229
|
-
def use_flash_attention(query, key, attention_mask=None):
|
230
|
-
# here we support query's shape is always [batch_size, head_num, q_len, head_dim],
|
231
|
-
# key's shape is always [batch_size, head_num, k_len, head_dim]
|
232
|
-
invalidInputError(query.dim() == 4,
|
233
|
-
"Here query input of use_flash_attention should be [batch_size, "
|
234
|
-
"head_num, q_len, head_dim]")
|
235
|
-
invalidInputError(key.dim() == 4,
|
236
|
-
"Here key input of use_flash_attention should be [batch_size, "
|
237
|
-
"head_num, k_len, head_dim]")
|
238
|
-
bsz, _, q_len, _ = query.size()
|
239
|
-
k_len = key.size()[2]
|
240
|
-
# check whether ipex flash attention can be used
|
241
|
-
if q_len != k_len:
|
242
|
-
# now only use flash attention for first token
|
243
|
-
# as it seems have no performance benifit for rest token now
|
244
|
-
return False
|
245
|
-
if query.device.type != "xpu":
|
246
|
-
# ipex flash attention only support for xpu
|
247
|
-
return False
|
248
|
-
ipex_version = get_ipex_version()
|
249
|
-
if ipex_version <= "2.0.110+xpu":
|
250
|
-
# ipex flash attention is supported from ipex 2.1
|
251
|
-
return False
|
252
|
-
if not torch.xpu.has_xetla():
|
253
|
-
# ipex flash attention is only supported for xetla
|
254
|
-
# may update this later
|
255
|
-
return False
|
256
|
-
elif get_xpu_device_type(query) != "pvc":
|
257
|
-
return False
|
258
|
-
if query.dtype not in [torch.float32, torch.float16]:
|
259
|
-
# only use flash attention for fp32/fp16 input
|
260
|
-
return False
|
261
|
-
if bsz > 1:
|
262
|
-
# as flash attention doesn't support attn_mask in ipex 2.1,
|
263
|
-
# so it will cause output error for padded batch input
|
264
|
-
if attention_mask is None:
|
265
|
-
return True
|
266
|
-
else:
|
267
|
-
# TODO: below logic may change for different model
|
268
|
-
# attention mask shape : [bsz, 1, q_len, k_len]
|
269
|
-
if attention_mask[0].squeeze()[0, 0].item() != 0:
|
270
|
-
# first batch contains padding
|
271
|
-
# otherwise we suppose it should be a upper triangular matrix
|
272
|
-
# at the same time, the diagonal is also 0
|
273
|
-
return False
|
274
|
-
elif not attention_mask.equal(attention_mask[0].repeat(bsz, 1, 1, 1)):
|
275
|
-
# check whether mask of every batch is the same
|
276
|
-
return False
|
277
|
-
return True
|
278
|
-
|
279
|
-
|
280
227
|
def use_sdp(q_len, kv_len, head_dim, query_states):
|
281
228
|
return (
|
282
229
|
query_states.device.type == "xpu"
|
@@ -315,38 +262,16 @@ def mlp_fusion_check(x, qtype, training):
|
|
315
262
|
if training or x.requires_grad:
|
316
263
|
return False
|
317
264
|
if qtype == FP6:
|
318
|
-
device =
|
319
|
-
if device in ["mtl", "lnl"]:
|
265
|
+
device = get_xpu_device_name(x.device)
|
266
|
+
if device in ["mtl", "lnl", "arl"]:
|
320
267
|
return False
|
321
268
|
return True
|
322
269
|
|
323
270
|
|
324
|
-
def use_decoding_fast_path(proj,
|
325
|
-
use_fuse_rope,
|
326
|
-
enough_kv_room,
|
327
|
-
bs,
|
328
|
-
qtype_check=decoding_fast_path_qtype_check):
|
329
|
-
if proj is None:
|
330
|
-
return False
|
331
|
-
device = get_xpu_device_type(proj.weight)
|
332
|
-
if not qtype_check(proj):
|
333
|
-
return False
|
334
|
-
if not use_fuse_rope:
|
335
|
-
return False
|
336
|
-
if not enough_kv_room:
|
337
|
-
return False
|
338
|
-
if bs != 1:
|
339
|
-
return False
|
340
|
-
|
341
|
-
if device in ["uhd"]:
|
342
|
-
return False
|
343
|
-
return True
|
344
|
-
|
345
|
-
|
346
271
|
def use_xmx(x: torch.Tensor, qtype: int):
|
347
|
-
device =
|
272
|
+
device = get_xpu_device_name(x.device)
|
348
273
|
return (
|
349
|
-
device in ["arc", "
|
274
|
+
device in ["arc", "pvc"]
|
350
275
|
and qtype in [SYM_INT4, SYM_INT8, FP8E4, FP8E5]
|
351
276
|
and (
|
352
277
|
(device == "pvc" and 1 < x.size(0) <= 16)
|
@@ -370,7 +295,7 @@ def fp16_fusion_check(proj, x, training):
|
|
370
295
|
return False
|
371
296
|
if x.requires_grad:
|
372
297
|
return False
|
373
|
-
device_type =
|
298
|
+
device_type = get_xpu_device_name(x.device)
|
374
299
|
if device_type != "pvc":
|
375
300
|
return False
|
376
301
|
return True
|
@@ -439,7 +364,7 @@ def should_use_compresskv(x: torch.Tensor, prompt_len: int):
|
|
439
364
|
else:
|
440
365
|
if use_compress_kv is None:
|
441
366
|
return (
|
442
|
-
|
367
|
+
get_xpu_device_name(x.device) in ["mtl", "lnl", "arl"]
|
443
368
|
and prompt_len >= 1800
|
444
369
|
and prompt_len <= 4500
|
445
370
|
)
|
@@ -473,10 +473,6 @@ def convert_llm_for_deploy(model: torch.nn.Module,
|
|
473
473
|
"n_splits_linear": n_splits_linear,
|
474
474
|
"n_splits_down_proj": n_splits_down_proj,
|
475
475
|
"lm_head_low_bit": lm_head_low_bit}
|
476
|
-
model.config.update(update_dict)
|
477
|
-
model.config.save_pretrained(save_directory)
|
478
|
-
if model.can_generate():
|
479
|
-
model.generation_config.save_pretrained(save_directory)
|
480
476
|
|
481
477
|
from .qwen import convert_qwen_layer, convert_fused_qwen_layer
|
482
478
|
from .qwen import convert_lm_head_and_embedding
|
@@ -537,8 +533,6 @@ def convert_llm_for_deploy(model: torch.nn.Module,
|
|
537
533
|
"n_splits_linear": n_splits_linear,
|
538
534
|
"n_splits_down_proj": n_splits_down_proj,
|
539
535
|
"lm_head_low_bit": lm_head_low_bit}
|
540
|
-
model.config.update(update_dict)
|
541
|
-
model.config.save_pretrained(save_directory)
|
542
536
|
|
543
537
|
from .llama import convert_llama_layer, convert_fused_llama_layer
|
544
538
|
from .llama import convert_lm_head_and_embedding
|
@@ -577,8 +571,6 @@ def convert_llm_for_deploy(model: torch.nn.Module,
|
|
577
571
|
"n_splits_linear": n_splits_linear,
|
578
572
|
"n_splits_down_proj": n_splits_down_proj,
|
579
573
|
"lm_head_low_bit": lm_head_low_bit}
|
580
|
-
model.config.update(update_dict)
|
581
|
-
model.config.save_pretrained(save_directory)
|
582
574
|
|
583
575
|
from .minicpm import convert_minicpm_layer, convert_fused_minicpm_layer
|
584
576
|
from .minicpm import convert_lm_head_and_embedding
|
@@ -595,3 +587,8 @@ def convert_llm_for_deploy(model: torch.nn.Module,
|
|
595
587
|
save_directory, weight_dir,
|
596
588
|
convert_model=True,
|
597
589
|
max_prompt_len=max_prompt_len)
|
590
|
+
|
591
|
+
model.config.update(update_dict)
|
592
|
+
model.config.save_pretrained(save_directory)
|
593
|
+
if model.can_generate():
|
594
|
+
model.generation_config.save_pretrained(save_directory)
|
ipex_llm/transformers/utils.py
CHANGED
@@ -168,27 +168,12 @@ def get_ipex_version():
|
|
168
168
|
return _ipex_version
|
169
169
|
|
170
170
|
|
171
|
-
def
|
172
|
-
if
|
173
|
-
return
|
174
|
-
name = torch.xpu.get_device_name(x.device.index)
|
175
|
-
if name.startswith("Intel(R) Arc(TM) A"):
|
176
|
-
return "arc"
|
177
|
-
elif name.startswith("Intel(R) Graphics [0xe20b]"):
|
178
|
-
return "bmg"
|
179
|
-
elif name.startswith("Intel(R) Arc(TM)"):
|
180
|
-
if 'V' in name:
|
181
|
-
return "lnl"
|
182
|
-
else:
|
183
|
-
return "mtl"
|
184
|
-
elif name.startswith("Intel(R) Data Center GPU Flex"):
|
185
|
-
return "flex"
|
186
|
-
elif name.startswith("Intel(R) Data Center GPU Max"):
|
187
|
-
return "pvc"
|
188
|
-
elif name.startswith("Intel(R) UHD"):
|
189
|
-
return "uhd"
|
171
|
+
def get_xpu_device_name(device: torch.device):
|
172
|
+
if device.type != "xpu":
|
173
|
+
return device.type
|
190
174
|
else:
|
191
|
-
|
175
|
+
import xe_linear
|
176
|
+
return xe_linear.get_xpu_device_name(device)
|
192
177
|
|
193
178
|
|
194
179
|
def load_imatrix_data(imatrix_file):
|
@@ -1,6 +1,6 @@
|
|
1
1
|
Metadata-Version: 2.1
|
2
2
|
Name: ipex-llm
|
3
|
-
Version: 2.2.
|
3
|
+
Version: 2.2.0b20250107
|
4
4
|
Summary: Large Language Model Develop Toolkit
|
5
5
|
Home-page: https://github.com/intel-analytics/ipex-llm
|
6
6
|
Author: BigDL Authors
|
@@ -27,15 +27,17 @@ Requires-Dist: intel-openmp ; (platform_machine == "x86_64" or platform_machine
|
|
27
27
|
Requires-Dist: torch ==2.1.2+cpu ; (platform_system == "Linux") and extra == 'all'
|
28
28
|
Requires-Dist: torch ==2.1.2 ; (platform_system == "Windows") and extra == 'all'
|
29
29
|
Provides-Extra: cpp
|
30
|
-
Requires-Dist: bigdl-core-cpp ==2.6.
|
30
|
+
Requires-Dist: bigdl-core-cpp ==2.6.0b20250107 ; extra == 'cpp'
|
31
31
|
Requires-Dist: setuptools ; extra == 'cpp'
|
32
32
|
Provides-Extra: cpp-arl
|
33
|
-
Requires-Dist: bigdl-core-cpp ==2.6.
|
33
|
+
Requires-Dist: bigdl-core-cpp ==2.6.0b20250107 ; extra == 'cpp-arl'
|
34
34
|
Requires-Dist: setuptools ; extra == 'cpp-arl'
|
35
35
|
Requires-Dist: onednn-devel ==2024.1.1 ; (platform_system == "Windows") and extra == 'cpp-arl'
|
36
|
+
Requires-Dist: onednn ==2024.1.1 ; (platform_system == "Windows") and extra == 'cpp-arl'
|
36
37
|
Requires-Dist: dpcpp-cpp-rt ==2024.2.1 ; (platform_system == "Windows") and extra == 'cpp-arl'
|
37
38
|
Requires-Dist: mkl-dpcpp ==2024.2.1 ; (platform_system == "Windows") and extra == 'cpp-arl'
|
38
39
|
Requires-Dist: onednn-devel ==2024.2.1 ; (platform_system == "Windows") and extra == 'cpp'
|
40
|
+
Requires-Dist: onednn ==2024.2.1 ; (platform_system == "Windows") and extra == 'cpp'
|
39
41
|
Requires-Dist: dpcpp-cpp-rt ==2024.2.1 ; (platform_system == "Windows") and extra == 'cpp'
|
40
42
|
Requires-Dist: mkl-dpcpp ==2024.2.1 ; (platform_system == "Windows") and extra == 'cpp'
|
41
43
|
Provides-Extra: llama-index
|
@@ -65,7 +67,7 @@ Requires-Dist: transformers ==4.40.0 ; extra == 'npu'
|
|
65
67
|
Requires-Dist: intel-openmp ; (platform_machine == "x86_64" or platform_machine == "AMD64") and extra == 'npu'
|
66
68
|
Requires-Dist: torch ==2.1.2+cpu ; (platform_system == "Linux") and extra == 'npu'
|
67
69
|
Requires-Dist: torch ==2.1.2 ; (platform_system == "Windows") and extra == 'npu'
|
68
|
-
Requires-Dist: bigdl-core-npu ==2.6.
|
70
|
+
Requires-Dist: bigdl-core-npu ==2.6.0b20250107 ; (platform_system == "Windows") and extra == 'npu'
|
69
71
|
Provides-Extra: serving
|
70
72
|
Requires-Dist: py-cpuinfo ; extra == 'serving'
|
71
73
|
Requires-Dist: fschat[model_worker,webui] ==0.2.36 ; extra == 'serving'
|
@@ -85,9 +87,9 @@ Requires-Dist: setuptools <70.0.0 ; extra == 'xpu'
|
|
85
87
|
Requires-Dist: torch ==2.1.0a0 ; extra == 'xpu'
|
86
88
|
Requires-Dist: torchvision ==0.16.0a0 ; extra == 'xpu'
|
87
89
|
Requires-Dist: intel-extension-for-pytorch ==2.1.10+xpu ; extra == 'xpu'
|
88
|
-
Requires-Dist: bigdl-core-xe-21 ==2.6.
|
89
|
-
Requires-Dist: bigdl-core-xe-batch-21 ==2.6.
|
90
|
-
Requires-Dist: bigdl-core-xe-addons-21 ==2.6.
|
90
|
+
Requires-Dist: bigdl-core-xe-21 ==2.6.0b20250107 ; extra == 'xpu'
|
91
|
+
Requires-Dist: bigdl-core-xe-batch-21 ==2.6.0b20250107 ; extra == 'xpu'
|
92
|
+
Requires-Dist: bigdl-core-xe-addons-21 ==2.6.0b20250107 ; extra == 'xpu'
|
91
93
|
Provides-Extra: xpu-2-1
|
92
94
|
Requires-Dist: py-cpuinfo ; extra == 'xpu-2-1'
|
93
95
|
Requires-Dist: protobuf ; extra == 'xpu-2-1'
|
@@ -102,12 +104,28 @@ Requires-Dist: setuptools <70.0.0 ; extra == 'xpu-2-1'
|
|
102
104
|
Requires-Dist: torch ==2.1.0a0 ; extra == 'xpu-2-1'
|
103
105
|
Requires-Dist: torchvision ==0.16.0a0 ; extra == 'xpu-2-1'
|
104
106
|
Requires-Dist: intel-extension-for-pytorch ==2.1.10+xpu ; extra == 'xpu-2-1'
|
105
|
-
Requires-Dist: bigdl-core-xe-21 ==2.6.
|
106
|
-
Requires-Dist: bigdl-core-xe-batch-21 ==2.6.
|
107
|
-
Requires-Dist: bigdl-core-xe-addons-21 ==2.6.
|
107
|
+
Requires-Dist: bigdl-core-xe-21 ==2.6.0b20250107 ; extra == 'xpu-2-1'
|
108
|
+
Requires-Dist: bigdl-core-xe-batch-21 ==2.6.0b20250107 ; extra == 'xpu-2-1'
|
109
|
+
Requires-Dist: bigdl-core-xe-addons-21 ==2.6.0b20250107 ; extra == 'xpu-2-1'
|
108
110
|
Requires-Dist: intel-openmp ; (platform_machine == "x86_64" or platform_machine == "AMD64") and extra == 'xpu-2-1'
|
109
111
|
Requires-Dist: dpcpp-cpp-rt ==2024.0.2 ; (platform_system == "Windows") and extra == 'xpu-2-1'
|
110
112
|
Requires-Dist: mkl-dpcpp ==2024.0.0 ; (platform_system == "Windows") and extra == 'xpu-2-1'
|
113
|
+
Provides-Extra: xpu-2-6
|
114
|
+
Requires-Dist: py-cpuinfo ; extra == 'xpu-2-6'
|
115
|
+
Requires-Dist: protobuf ; extra == 'xpu-2-6'
|
116
|
+
Requires-Dist: mpmath ==1.3.0 ; extra == 'xpu-2-6'
|
117
|
+
Requires-Dist: numpy ==1.26.4 ; extra == 'xpu-2-6'
|
118
|
+
Requires-Dist: transformers ==4.37.0 ; extra == 'xpu-2-6'
|
119
|
+
Requires-Dist: sentencepiece ; extra == 'xpu-2-6'
|
120
|
+
Requires-Dist: tokenizers ==0.15.2 ; extra == 'xpu-2-6'
|
121
|
+
Requires-Dist: accelerate ==0.23.0 ; extra == 'xpu-2-6'
|
122
|
+
Requires-Dist: tabulate ; extra == 'xpu-2-6'
|
123
|
+
Requires-Dist: setuptools ; extra == 'xpu-2-6'
|
124
|
+
Requires-Dist: torch ==2.6.0+xpu ; extra == 'xpu-2-6'
|
125
|
+
Requires-Dist: torchvision ==0.21.0+xpu ; extra == 'xpu-2-6'
|
126
|
+
Requires-Dist: torchaudio ==2.6.0+xpu ; extra == 'xpu-2-6'
|
127
|
+
Requires-Dist: bigdl-core-xe-all ==2.6.0b20250107 ; extra == 'xpu-2-6'
|
128
|
+
Requires-Dist: intel-openmp ; (platform_machine == "x86_64" or platform_machine == "AMD64") and extra == 'xpu-2-6'
|
111
129
|
Provides-Extra: xpu-arc
|
112
130
|
Requires-Dist: py-cpuinfo ; extra == 'xpu-arc'
|
113
131
|
Requires-Dist: protobuf ; extra == 'xpu-arc'
|
@@ -119,9 +137,9 @@ Requires-Dist: tokenizers ==0.15.2 ; extra == 'xpu-arc'
|
|
119
137
|
Requires-Dist: accelerate ==0.23.0 ; extra == 'xpu-arc'
|
120
138
|
Requires-Dist: tabulate ; extra == 'xpu-arc'
|
121
139
|
Requires-Dist: setuptools ; extra == 'xpu-arc'
|
122
|
-
Requires-Dist: bigdl-core-xe-23 ==2.6.
|
123
|
-
Requires-Dist: bigdl-core-xe-batch-23 ==2.6.
|
124
|
-
Requires-Dist: bigdl-core-xe-addons-23 ==2.6.
|
140
|
+
Requires-Dist: bigdl-core-xe-23 ==2.6.0b20250107 ; extra == 'xpu-arc'
|
141
|
+
Requires-Dist: bigdl-core-xe-batch-23 ==2.6.0b20250107 ; extra == 'xpu-arc'
|
142
|
+
Requires-Dist: bigdl-core-xe-addons-23 ==2.6.0b20250107 ; extra == 'xpu-arc'
|
125
143
|
Requires-Dist: intel-openmp ; (platform_machine == "x86_64" or platform_machine == "AMD64") and extra == 'xpu-arc'
|
126
144
|
Requires-Dist: torch ==2.3.1+cxx11.abi ; (platform_system == "Linux") and extra == 'xpu-arc'
|
127
145
|
Requires-Dist: torchvision ==0.18.1+cxx11.abi ; (platform_system == "Linux") and extra == 'xpu-arc'
|
@@ -130,6 +148,7 @@ Requires-Dist: torch ==2.3.1.post0+cxx11.abi ; (platform_system == "Windows") an
|
|
130
148
|
Requires-Dist: torchvision ==0.18.1.post0+cxx11.abi ; (platform_system == "Windows") and extra == 'xpu-arc'
|
131
149
|
Requires-Dist: intel-extension-for-pytorch ==2.3.110.post0+xpu ; (platform_system == "Windows") and extra == 'xpu-arc'
|
132
150
|
Requires-Dist: onednn-devel ==2024.1.1 ; (platform_system == "Windows") and extra == 'xpu-arc'
|
151
|
+
Requires-Dist: onednn ==2024.1.1 ; (platform_system == "Windows") and extra == 'xpu-arc'
|
133
152
|
Provides-Extra: xpu-arl
|
134
153
|
Requires-Dist: py-cpuinfo ; extra == 'xpu-arl'
|
135
154
|
Requires-Dist: protobuf ; extra == 'xpu-arl'
|
@@ -141,9 +160,9 @@ Requires-Dist: tokenizers ==0.15.2 ; extra == 'xpu-arl'
|
|
141
160
|
Requires-Dist: accelerate ==0.23.0 ; extra == 'xpu-arl'
|
142
161
|
Requires-Dist: tabulate ; extra == 'xpu-arl'
|
143
162
|
Requires-Dist: setuptools ; extra == 'xpu-arl'
|
144
|
-
Requires-Dist: bigdl-core-xe-23 ==2.6.
|
145
|
-
Requires-Dist: bigdl-core-xe-batch-23 ==2.6.
|
146
|
-
Requires-Dist: bigdl-core-xe-addons-23 ==2.6.
|
163
|
+
Requires-Dist: bigdl-core-xe-23 ==2.6.0b20250107 ; extra == 'xpu-arl'
|
164
|
+
Requires-Dist: bigdl-core-xe-batch-23 ==2.6.0b20250107 ; extra == 'xpu-arl'
|
165
|
+
Requires-Dist: bigdl-core-xe-addons-23 ==2.6.0b20250107 ; extra == 'xpu-arl'
|
147
166
|
Requires-Dist: intel-openmp ; (platform_machine == "x86_64" or platform_machine == "AMD64") and extra == 'xpu-arl'
|
148
167
|
Requires-Dist: torch ==2.3.1+cxx11.abi ; (platform_system == "Linux") and extra == 'xpu-arl'
|
149
168
|
Requires-Dist: torchvision ==0.18.1+cxx11.abi ; (platform_system == "Linux") and extra == 'xpu-arl'
|
@@ -152,6 +171,7 @@ Requires-Dist: torch ==2.3.1.post0+cxx11.abi ; (platform_system == "Windows") an
|
|
152
171
|
Requires-Dist: torchvision ==0.18.1.post0+cxx11.abi ; (platform_system == "Windows") and extra == 'xpu-arl'
|
153
172
|
Requires-Dist: intel-extension-for-pytorch ==2.3.110.post0+xpu ; (platform_system == "Windows") and extra == 'xpu-arl'
|
154
173
|
Requires-Dist: onednn-devel ==2024.1.1 ; (platform_system == "Windows") and extra == 'xpu-arl'
|
174
|
+
Requires-Dist: onednn ==2024.1.1 ; (platform_system == "Windows") and extra == 'xpu-arl'
|
155
175
|
Provides-Extra: xpu-lnl
|
156
176
|
Requires-Dist: py-cpuinfo ; extra == 'xpu-lnl'
|
157
177
|
Requires-Dist: protobuf ; extra == 'xpu-lnl'
|
@@ -163,9 +183,9 @@ Requires-Dist: tokenizers ==0.15.2 ; extra == 'xpu-lnl'
|
|
163
183
|
Requires-Dist: accelerate ==0.23.0 ; extra == 'xpu-lnl'
|
164
184
|
Requires-Dist: tabulate ; extra == 'xpu-lnl'
|
165
185
|
Requires-Dist: setuptools ; extra == 'xpu-lnl'
|
166
|
-
Requires-Dist: bigdl-core-xe-23 ==2.6.
|
167
|
-
Requires-Dist: bigdl-core-xe-batch-23 ==2.6.
|
168
|
-
Requires-Dist: bigdl-core-xe-addons-23 ==2.6.
|
186
|
+
Requires-Dist: bigdl-core-xe-23 ==2.6.0b20250107 ; extra == 'xpu-lnl'
|
187
|
+
Requires-Dist: bigdl-core-xe-batch-23 ==2.6.0b20250107 ; extra == 'xpu-lnl'
|
188
|
+
Requires-Dist: bigdl-core-xe-addons-23 ==2.6.0b20250107 ; extra == 'xpu-lnl'
|
169
189
|
Requires-Dist: intel-openmp ; (platform_machine == "x86_64" or platform_machine == "AMD64") and extra == 'xpu-lnl'
|
170
190
|
Requires-Dist: torch ==2.3.1+cxx11.abi ; (platform_system == "Linux") and extra == 'xpu-lnl'
|
171
191
|
Requires-Dist: torchvision ==0.18.1+cxx11.abi ; (platform_system == "Linux") and extra == 'xpu-lnl'
|
@@ -174,6 +194,7 @@ Requires-Dist: torch ==2.3.1.post0+cxx11.abi ; (platform_system == "Windows") an
|
|
174
194
|
Requires-Dist: torchvision ==0.18.1.post0+cxx11.abi ; (platform_system == "Windows") and extra == 'xpu-lnl'
|
175
195
|
Requires-Dist: intel-extension-for-pytorch ==2.3.110.post0+xpu ; (platform_system == "Windows") and extra == 'xpu-lnl'
|
176
196
|
Requires-Dist: onednn-devel ==2024.1.1 ; (platform_system == "Windows") and extra == 'xpu-lnl'
|
197
|
+
Requires-Dist: onednn ==2024.1.1 ; (platform_system == "Windows") and extra == 'xpu-lnl'
|
177
198
|
Requires-Dist: intel-openmp ; (platform_machine == "x86_64" or platform_machine == "AMD64") and extra == 'xpu'
|
178
199
|
Requires-Dist: dpcpp-cpp-rt ==2024.0.2 ; (platform_system == "Windows") and extra == 'xpu'
|
179
200
|
Requires-Dist: mkl-dpcpp ==2024.0.0 ; (platform_system == "Windows") and extra == 'xpu'
|
@@ -41,35 +41,35 @@ ipex_llm/langchain/llms/transformerspipelinellm.py,sha256=vm522YPPwWxxAPVvQBtxRf
|
|
41
41
|
ipex_llm/langchain/vllm/__init__.py,sha256=T-EbRT6GJ_8RCu-iLmSzcftOimXSPQf2d5X72AUAy2Y,874
|
42
42
|
ipex_llm/langchain/vllm/vllm.py,sha256=6dxc-ZISZQrJilEa_HA827l75Dv9rcHpY_G6FdJ8BVs,7793
|
43
43
|
ipex_llm/libs/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
44
|
-
ipex_llm/libs/bloom-api.dll,sha256=
|
45
|
-
ipex_llm/libs/bloom.dll,sha256=
|
46
|
-
ipex_llm/libs/gptneox-api.dll,sha256=
|
47
|
-
ipex_llm/libs/gptneox.dll,sha256=
|
48
|
-
ipex_llm/libs/libbloom_avx.dll,sha256=
|
49
|
-
ipex_llm/libs/libbloom_vnni.dll,sha256=
|
50
|
-
ipex_llm/libs/libgptneox_avx.dll,sha256=
|
51
|
-
ipex_llm/libs/libgptneox_vnni.dll,sha256=
|
52
|
-
ipex_llm/libs/libllama_avx.dll,sha256=
|
53
|
-
ipex_llm/libs/libllama_vnni.dll,sha256=
|
54
|
-
ipex_llm/libs/libstarcoder_avx.dll,sha256=
|
55
|
-
ipex_llm/libs/libstarcoder_vnni.dll,sha256=
|
56
|
-
ipex_llm/libs/llama-api.dll,sha256=-
|
57
|
-
ipex_llm/libs/llama.dll,sha256=
|
58
|
-
ipex_llm/libs/main-bloom.exe,sha256=
|
59
|
-
ipex_llm/libs/main-gptneox.exe,sha256=
|
60
|
-
ipex_llm/libs/main-llama.exe,sha256=
|
61
|
-
ipex_llm/libs/main-starcoder.exe,sha256=
|
62
|
-
ipex_llm/libs/pipeline.dll,sha256=
|
63
|
-
ipex_llm/libs/quantize-bloom.exe,sha256=
|
64
|
-
ipex_llm/libs/quantize-bloom_vnni.exe,sha256=
|
65
|
-
ipex_llm/libs/quantize-gptneox.exe,sha256=
|
66
|
-
ipex_llm/libs/quantize-gptneox_vnni.exe,sha256=
|
67
|
-
ipex_llm/libs/quantize-llama.exe,sha256=
|
68
|
-
ipex_llm/libs/quantize-llama_vnni.exe,sha256=
|
69
|
-
ipex_llm/libs/quantize-starcoder.exe,sha256=
|
70
|
-
ipex_llm/libs/quantize-starcoder_vnni.exe,sha256=
|
71
|
-
ipex_llm/libs/starcoder-api.dll,sha256=
|
72
|
-
ipex_llm/libs/starcoder.dll,sha256=
|
44
|
+
ipex_llm/libs/bloom-api.dll,sha256=aa3qzp0VqG1RzqCWC8uPfyzCsf3xC0qqgyj4WxDoOsM,36352
|
45
|
+
ipex_llm/libs/bloom.dll,sha256=kVhdgqqPvusN-aFfVCBnSZSGUWRw9dv9ICv-klmF0UE,506880
|
46
|
+
ipex_llm/libs/gptneox-api.dll,sha256=bjp6ZiVvmT6aWrDiqgV43pYeifLnGDlNnRwi3r1rTFo,24576
|
47
|
+
ipex_llm/libs/gptneox.dll,sha256=xKfJ9oNMoUXzHH43vXrap9wbm-KgKY-AhARrVvU_iLg,567296
|
48
|
+
ipex_llm/libs/libbloom_avx.dll,sha256=VGPb6z7Jh4pgStkfnzqEBW-mnzLTETVUsc20ym5c1eI,535040
|
49
|
+
ipex_llm/libs/libbloom_vnni.dll,sha256=_ibtsNyO3G9KV30mKJky49eKMm_8ZVgKPA0FdjjIvB0,506880
|
50
|
+
ipex_llm/libs/libgptneox_avx.dll,sha256=onI5sXdEsCAPL3JAIGH5r4jwPbevYi6bw15CmPyZL5s,595456
|
51
|
+
ipex_llm/libs/libgptneox_vnni.dll,sha256=6BvdfvGnikm78KlPcdGq6IiAecPkPwm6i38PotgB9FQ,567808
|
52
|
+
ipex_llm/libs/libllama_avx.dll,sha256=TRqBjWOmR7dCQtEq9c5l0wshBPM2e7klfca8eV-f79Q,589824
|
53
|
+
ipex_llm/libs/libllama_vnni.dll,sha256=F9va1YfNVe9XGfilbio3mEdbbGtmpfAR-9T8vWEqulI,561664
|
54
|
+
ipex_llm/libs/libstarcoder_avx.dll,sha256=IvM2V18nf_g9tsaLz8W30qgchg496V89PpCoKMHnQPY,626688
|
55
|
+
ipex_llm/libs/libstarcoder_vnni.dll,sha256=X9H9Tyy3DdQg1LIvc8ILP33ilj4amEUfwF_E9KPVWvE,598528
|
56
|
+
ipex_llm/libs/llama-api.dll,sha256=-XYElLfNrWEP63PStFxXX9wM0kP6PGYQJH6oePg5u-I,25600
|
57
|
+
ipex_llm/libs/llama.dll,sha256=zLW5de-ASaVBoPQbRYUMK6L6FoPwmt6II4bGFVeHE-M,561152
|
58
|
+
ipex_llm/libs/main-bloom.exe,sha256=cNaMadt2EpUgCu55XjQW0Vi1UUBQ5eA9CnfXZV6OGjc,103424
|
59
|
+
ipex_llm/libs/main-gptneox.exe,sha256=54va_bkWTOw0Gf_EMEF2MaqNsKEugEsxTxUszTAxY7Q,98816
|
60
|
+
ipex_llm/libs/main-llama.exe,sha256=banfh_rJMhBR4FERotQ6FBOKDEhz-KF49Zu6E86Wlso,99840
|
61
|
+
ipex_llm/libs/main-starcoder.exe,sha256=L-SIpoGn0vA8CD37MMm44i56CTWKXQjnC6WWZE1okPw,157696
|
62
|
+
ipex_llm/libs/pipeline.dll,sha256=rOzOT0lY3RMQlIDHCvaAx_bjAvQD8BBZNyqi3BssG2U,72704
|
63
|
+
ipex_llm/libs/quantize-bloom.exe,sha256=GvY5jFYGrK_wFa1RyAurX0vD_UAg7FCD9Ls1zAbYYKc,126464
|
64
|
+
ipex_llm/libs/quantize-bloom_vnni.exe,sha256=j3i2Q_ymqlnrPBUCskU3gqrwV198NkfLxmlz0KOsqoI,127488
|
65
|
+
ipex_llm/libs/quantize-gptneox.exe,sha256=he-NK5wy8u_t3S26Qb4Euenf0E6mw8O6lsNOcX-m2ts,104448
|
66
|
+
ipex_llm/libs/quantize-gptneox_vnni.exe,sha256=Oz7iHHL01QZYJpCBsPeYXhBCRBJ9kFklYZ2UhEw77rk,104960
|
67
|
+
ipex_llm/libs/quantize-llama.exe,sha256=Ns-98bA07AeTPaN7v0zs3cOb3PAvUmj7R_Xvsn-1bTM,109568
|
68
|
+
ipex_llm/libs/quantize-llama_vnni.exe,sha256=t4wts7FmUT4n5_Ii6wqtrVMv73pf3pOMMfrziRZAI5U,110592
|
69
|
+
ipex_llm/libs/quantize-starcoder.exe,sha256=xYLEIQ4gtU23ae7cN_1hZrxqDrSiQdCNu3EZRbffNLs,127488
|
70
|
+
ipex_llm/libs/quantize-starcoder_vnni.exe,sha256=5gmwDZTTqM9fFuj9UihVyA91TmcoBq95xxAofaBi4sM,128512
|
71
|
+
ipex_llm/libs/starcoder-api.dll,sha256=UhlNrAN44aZwkbjdmVFr9E2tV3iuQ8jR1OvEgiQzVXk,21504
|
72
|
+
ipex_llm/libs/starcoder.dll,sha256=1jGXZ-nQo3m1kMXeOVwd-fc68p7kjPGfzAu4ifngrlE,598016
|
73
73
|
ipex_llm/llamaindex/__init__.py,sha256=T-EbRT6GJ_8RCu-iLmSzcftOimXSPQf2d5X72AUAy2Y,874
|
74
74
|
ipex_llm/llamaindex/llms/__init__.py,sha256=KP1lEdGqDuxPoxL1ZSH25Pm2kKMPJBWUTLR0ckSLMIU,1139
|
75
75
|
ipex_llm/llamaindex/llms/bigdlllm.py,sha256=FQBzq1KOjfc6uofTXAha3O7TqpJkNfOFepXQmOVlbnI,26314
|
@@ -87,14 +87,14 @@ ipex_llm/serving/fastchat/tgi_api_protocol.py,sha256=brT3k3-V0NJrU4fRqUwWjC0O3iO
|
|
87
87
|
ipex_llm/serving/fastchat/tgi_api_server.py,sha256=agNTAEiZPSuj3dEdIdYKwkoY0cXOUDX06DiM9VP2knQ,24418
|
88
88
|
ipex_llm/serving/fastchat/vllm_worker.py,sha256=ZLz2Q9GxJO6r_LOiP6epgCRjBGk-K4EB1SNEWSJp5DA,11091
|
89
89
|
ipex_llm/transformers/__init__.py,sha256=l4KkMkLe-pRC7b_kj6LCfeifgE-Uo33_Av_FwN9HnFA,1074
|
90
|
-
ipex_llm/transformers/convert.py,sha256=
|
90
|
+
ipex_llm/transformers/convert.py,sha256=pFm6VlU84u_Llr2sp6-gRrEYDeNgIk2QPukolq4IE1s,99947
|
91
91
|
ipex_llm/transformers/convert_ipex.py,sha256=iKXo0n8fVFTOA2fNYYrByMFK0dovL-kLd2sVDk88AlQ,14334
|
92
92
|
ipex_llm/transformers/embedding.py,sha256=bdgk59DvD4ZZyxRzewXOR7g56nThgO6uhIwk8QL7f-s,9299
|
93
93
|
ipex_llm/transformers/kv.py,sha256=k4TU18LlA-Sbq9WNNQnfuzu3RSFBwFhmaV3BcGN5bAo,19191
|
94
94
|
ipex_llm/transformers/lisa.py,sha256=F5WxbtXQ7RdKulj83h_2DnEIgKiKGZf7zvOmg6QBl2s,3289
|
95
95
|
ipex_llm/transformers/loader.py,sha256=cOgX93xOC-4dt01GTJ5wyd7PjZ8S43r4mctkR2YxVuw,6893
|
96
|
-
ipex_llm/transformers/lookup.py,sha256=
|
97
|
-
ipex_llm/transformers/low_bit_linear.py,sha256=
|
96
|
+
ipex_llm/transformers/lookup.py,sha256=b6OlZ9OV10R9qeWw8mVryVpDxszkjwLkldvi7GPMJY8,19614
|
97
|
+
ipex_llm/transformers/low_bit_linear.py,sha256=nKraUvZJ7UdXP29HSE4CJPIVxmN-TvG8dpT4gpleuyQ,41688
|
98
98
|
ipex_llm/transformers/model.py,sha256=KcRjkauGg48BYrUBoUZaVMpg7Piuz5JrfIpVZd3EIjs,41105
|
99
99
|
ipex_llm/transformers/modelling_bigdl.py,sha256=7JpNVMuyq_OmtNUaMFMXdxPWZp2q0QHC02QeA-VTPOw,6709
|
100
100
|
ipex_llm/transformers/npu_model.py,sha256=YW02GeVz-9ZGqxAeSz0AOvciS-17bo9eK5ZOBrICwSQ,39508
|
@@ -105,7 +105,7 @@ ipex_llm/transformers/relora.py,sha256=-dYzUV0P-IhO2jFdnzN9-v_sFzJpRj3ZwN9eCJzOo
|
|
105
105
|
ipex_llm/transformers/speculative.py,sha256=Zf1nQb5GXpJQrUHBTL-H4RUBfdv3lGhfehzudHimhYk,64109
|
106
106
|
ipex_llm/transformers/streamer.py,sha256=RrVlLblzCOtABRUpaMXAyaMnCGgLUtAi_YesLumRbww,4842
|
107
107
|
ipex_llm/transformers/training_patch.py,sha256=oxMkUtqyvqJiprw6dE3skkYfD1HOmUlH9N0hBkbn0G0,10799
|
108
|
-
ipex_llm/transformers/utils.py,sha256=
|
108
|
+
ipex_llm/transformers/utils.py,sha256=fXLIlr9hoBr27p3w3xzczZGPk2cCTIRbUKBkiVCGYbc,16889
|
109
109
|
ipex_llm/transformers/xpu_customize_fwd.py,sha256=wFpIhs5F6tkNs8gBOrLxWdhLzO3EDHovVkERPIAoAvg,7611
|
110
110
|
ipex_llm/transformers/xpu_ops.py,sha256=H46-69pMRQhekbAEoDfNacCInLWycMHDqrgMGLvFYfI,4362
|
111
111
|
ipex_llm/transformers/awq/__init__.py,sha256=Du5gu3-eeAkeDO_dEMBTzrDBA66DSN3uL3-rn8WGXQw,875
|
@@ -141,13 +141,11 @@ ipex_llm/transformers/models/baichuan.py,sha256=oJCAEENSG8oQhJ-QPN2SiapARjAGdOM6
|
|
141
141
|
ipex_llm/transformers/models/bert.py,sha256=bJNic2pt1kph0kBwdK5MRGyWupFfx2Ts0V3D1L-5kWo,6085
|
142
142
|
ipex_llm/transformers/models/bloom.py,sha256=PxfzyYT-nFn3K5rZhTQjmcEjUUzAhUFzxIN4kzRlCuc,8103
|
143
143
|
ipex_llm/transformers/models/chatglm.py,sha256=UHai1t2AUtGmF765_eHF8LUMVQzp_oCBx8TJB21WrHk,12597
|
144
|
-
ipex_llm/transformers/models/chatglm2.py,sha256=
|
144
|
+
ipex_llm/transformers/models/chatglm2.py,sha256=SGCABJdYQLW0zDarEoWrEQLuWlbq9iQhYU8ZeR1-ptQ,15957
|
145
145
|
ipex_llm/transformers/models/chatglm4.py,sha256=AAhAFFDDas5DBQPfh2Mwl7a2v7taKf6xphoeeNNFaBI,16593
|
146
146
|
ipex_llm/transformers/models/chatglm4v.py,sha256=YRfuf9g1E0MQ_7wbHAOMvadFnO-j3LqI_k1SaRkDs0M,14055
|
147
|
-
ipex_llm/transformers/models/cohere.py,sha256=RCUBfhI_eNuZ1NyQC6vm28eYqPmgZn6pHSvQMwFyT2A,25884
|
148
147
|
ipex_llm/transformers/models/common.py,sha256=4obQMGF02FCiXrHnFle9Fsx7C33b1FDt37qJJ4YgxRc,11578
|
149
148
|
ipex_llm/transformers/models/decilm.py,sha256=P-PBuDPf07GvKggLwJx_wPwIn6esN3rX8ai2JxRuZmE,5246
|
150
|
-
ipex_llm/transformers/models/falcon.py,sha256=f5BzMbv4E-R5Pete8zBscbgiueXGIaWGs-5RbcMlUo4,33549
|
151
149
|
ipex_llm/transformers/models/gemma.py,sha256=_E3Yw8Y45xyNVeLqyVKcpr8kjuICtETeL82cJ-bWJuU,9424
|
152
150
|
ipex_llm/transformers/models/gemma2.py,sha256=2WZuv-FLzJyTJFaYxOuzJt47QE64M0lHnzAiO5T6ozI,8049
|
153
151
|
ipex_llm/transformers/models/glm.py,sha256=gHYgfn20jPRL-ElXy-rUqMh6_LQcc5x7DEXSZuRA4E0,7094
|
@@ -160,9 +158,8 @@ ipex_llm/transformers/models/internvl.py,sha256=Vx0vENIEQLX2M6P398mw5TOhpks0U8xf
|
|
160
158
|
ipex_llm/transformers/models/llama.py,sha256=ozwtdQ0MbanJEtW4LBFGxqs_QAq82EonhL2dL6tGyw0,8567
|
161
159
|
ipex_llm/transformers/models/minicpm.py,sha256=ib2rJTN7Tf7znBCtVrtXsF-_Uuk2aA7KVg02xzatLiI,10103
|
162
160
|
ipex_llm/transformers/models/minicpm3.py,sha256=FhNS6mi2rg7dSdF_QQGrao3g9EC6XLn1MTKd-kd0wF0,9191
|
163
|
-
ipex_llm/transformers/models/minicpmv.py,sha256=
|
161
|
+
ipex_llm/transformers/models/minicpmv.py,sha256=ZV4s48WNIyRoEkvENnlmopnx3ojZANBer0LI6bRtxrY,9826
|
164
162
|
ipex_llm/transformers/models/mistral.py,sha256=rE1GWQxXvF6aG-buPHDR13zeynDZEDIubPF4PiVhZbM,7451
|
165
|
-
ipex_llm/transformers/models/mixtral.py,sha256=zShaxabIoQaL3cV0Rptf7VCN4QuJsV4KBhanff1TASY,26601
|
166
163
|
ipex_llm/transformers/models/mllama.py,sha256=ogpLmmN_OwcFUyjYB-oDC-l3uw8urFvUEc5edkjWHAk,10939
|
167
164
|
ipex_llm/transformers/models/mpt.py,sha256=z02NwHogJZVh-Mk4sYoIzR90SFIKhoNN_-ifsD907TQ,9540
|
168
165
|
ipex_llm/transformers/models/phi.py,sha256=E6qz4EEuHIVGvaPo-wtLC5lz3iyMqTbAE_cRlcjQRKI,6670
|
@@ -175,10 +172,10 @@ ipex_llm/transformers/models/qwen2_vl.py,sha256=jIm4yZSd751BkRqgj3wR1QBkDIh-TMCL
|
|
175
172
|
ipex_llm/transformers/models/qwen_vl.py,sha256=j7Nzzz2Qvynu9yrCXmoEfERjw43hXof5TbXIs7Ms-oY,17105
|
176
173
|
ipex_llm/transformers/models/rwkv4.py,sha256=H4KMtxN0JA2ZTXnonHpsUUJ5xULemo-D1Jzl0ri_UY8,6123
|
177
174
|
ipex_llm/transformers/models/rwkv5.py,sha256=OkRNj1pCAZg1z2Fw-I0DEnxLEdZyPeRSQ6msrkxLOCs,10710
|
178
|
-
ipex_llm/transformers/models/sd.py,sha256=
|
175
|
+
ipex_llm/transformers/models/sd.py,sha256=VvHV5u-0k2MgHu3NL9113hPj7DgfxqctuKzEEeNfRDU,5981
|
179
176
|
ipex_llm/transformers/models/stablelm.py,sha256=RGQCYuQhYqtZ1j3RZkYi0_QvCRnUgUIPYxfBcLnElzg,6885
|
180
177
|
ipex_llm/transformers/models/starcoder2.py,sha256=4P3mhRYf2Kreb1ESjrQGfy1puLMmZXgV35zf-Tksvao,6462
|
181
|
-
ipex_llm/transformers/models/utils.py,sha256=
|
178
|
+
ipex_llm/transformers/models/utils.py,sha256=Qbz7UkYSbsM5bodH2445O0-JF50Mu3UEwW0j2ZNxHSU,15997
|
182
179
|
ipex_llm/transformers/models/yuan.py,sha256=1jRPebwAK2ENbyYokOmb4LSVo-szucWiygz9zTv-scs,7656
|
183
180
|
ipex_llm/transformers/npu_models/__init__.py,sha256=ulEUGLjaP48LCrVeury3UxLjXxKzRi0UpSG4bYu-7f8,585
|
184
181
|
ipex_llm/transformers/npu_models/baichuan.py,sha256=fJtd7fBrttySghRUgfZTAdxLjsSNC-XL08HISsXigLE,4685
|
@@ -211,7 +208,7 @@ ipex_llm/transformers/npu_models/xlm_mp.py,sha256=sj8OVun8xJprM7ZJp0XzWa55rqlSIz
|
|
211
208
|
ipex_llm/transformers/npu_pipeline_model/__init__.py,sha256=b2IXvVqQ5cItki021h8s3ymW12RPu8QNPprq4Mn3bDM,586
|
212
209
|
ipex_llm/transformers/npu_pipeline_model/baichuan.py,sha256=ICxRzFQ4OIANDkkVi2_4xOeQXmfFXYMx3H52KuE1xR4,6208
|
213
210
|
ipex_llm/transformers/npu_pipeline_model/common.py,sha256=QxJoJESpv0BpwO_FBeAT2wKA56wNFfen8iI37PrMKuA,7838
|
214
|
-
ipex_llm/transformers/npu_pipeline_model/convert_pipeline.py,sha256=
|
211
|
+
ipex_llm/transformers/npu_pipeline_model/convert_pipeline.py,sha256=wrRgmNT13RVtQRp5gFRBxNEPJHxFMLeGqb8a58YodPQ,28512
|
215
212
|
ipex_llm/transformers/npu_pipeline_model/llama.py,sha256=MnvHRytLt3oy5jIPUBe8AeEJ6PtPWLbhQ5a9WqjZ1TQ,19905
|
216
213
|
ipex_llm/transformers/npu_pipeline_model/minicpm.py,sha256=MDMesYlVbECKdK0xxkt1LwHgpkJOO7ZwBExYAwMGQa0,20637
|
217
214
|
ipex_llm/transformers/npu_pipeline_model/pipeline_cpp.py,sha256=JNmodAMg_NQvDILug3E_fGXEh6cd3wsj4bvAzcd-vaU,2749
|
@@ -247,11 +244,11 @@ ipex_llm/vllm/xpu/engine/__init__.py,sha256=pY_CpyuZd72fr6s32ejeKHKFW0K4vUU2rzZj
|
|
247
244
|
ipex_llm/vllm/xpu/engine/engine.py,sha256=k4-D27WS_Gk3mA--w3HWAjPjb4Aiu043MVPi0ZoAUBc,5984
|
248
245
|
ipex_llm/vllm/xpu/entrypoints/openai/api_server.py,sha256=GshTZFB8e4PWvqckfbmTOU6b0oLkNn7A-vzLuG9--j8,21544
|
249
246
|
ipex_llm/vllm/xpu/entrypoints/openai/cli_args.py,sha256=2rENA2ucynMaIjiZBEh2ez1o5vR32GaP514t39CD7KM,8676
|
250
|
-
ipex_llm-2.2.
|
251
|
-
ipex_llm-2.2.
|
252
|
-
ipex_llm-2.2.
|
253
|
-
ipex_llm-2.2.
|
254
|
-
ipex_llm-2.2.
|
255
|
-
ipex_llm-2.2.
|
256
|
-
ipex_llm-2.2.
|
257
|
-
ipex_llm-2.2.
|
247
|
+
ipex_llm-2.2.0b20250107.data/scripts/ipex-llm-init.bat,sha256=HPtCYuDYwEatq7dAwOvdfVcHYCpAVdbj75K1qh0vQek,2578
|
248
|
+
ipex_llm-2.2.0b20250107.data/scripts/llm-chat.ps1,sha256=6qrs-hGVAV8IKh7Jx8nq_XrnZcjd7qGU5wndArM7Yag,2769
|
249
|
+
ipex_llm-2.2.0b20250107.data/scripts/llm-cli.ps1,sha256=3qBtTLs_EjYDnM8YyCpJhzLnGCKTEGssu9UNqfkjVXs,3009
|
250
|
+
ipex_llm-2.2.0b20250107.dist-info/METADATA,sha256=rPJCuVvUndZ0XZBTZzlQEPi1y_W0fpmQTEGmogyRzRw,12705
|
251
|
+
ipex_llm-2.2.0b20250107.dist-info/WHEEL,sha256=6iYPr8vTHsyDK75jr9X0V3I9wPSVmtwr_8fdATBciGk,98
|
252
|
+
ipex_llm-2.2.0b20250107.dist-info/entry_points.txt,sha256=TiUyBB2MRmfF3ko-pyAEzqeBCRnyhu27bNOAsWPp3e8,61
|
253
|
+
ipex_llm-2.2.0b20250107.dist-info/top_level.txt,sha256=CGCMHM-SyqUabU4h8RqJ2KTYckQUO3LvIWwmUQ6Qbzw,9
|
254
|
+
ipex_llm-2.2.0b20250107.dist-info/RECORD,,
|