ipex-llm 2.2.0b20250106__py3-none-manylinux2010_x86_64.whl → 2.2.0b20250107__py3-none-manylinux2010_x86_64.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (44) hide show
  1. ipex_llm/libs/libbloom_amx.so +0 -0
  2. ipex_llm/libs/libbloom_avx.so +0 -0
  3. ipex_llm/libs/libbloom_avx2.so +0 -0
  4. ipex_llm/libs/libbloom_avx512.so +0 -0
  5. ipex_llm/libs/libbloom_avxvnni.so +0 -0
  6. ipex_llm/libs/libgptneox_amx.so +0 -0
  7. ipex_llm/libs/libgptneox_avx.so +0 -0
  8. ipex_llm/libs/libgptneox_avx2.so +0 -0
  9. ipex_llm/libs/libgptneox_avx512.so +0 -0
  10. ipex_llm/libs/libgptneox_avxvnni.so +0 -0
  11. ipex_llm/libs/libllama_amx.so +0 -0
  12. ipex_llm/libs/libllama_avx.so +0 -0
  13. ipex_llm/libs/libllama_avx2.so +0 -0
  14. ipex_llm/libs/libllama_avx512.so +0 -0
  15. ipex_llm/libs/libllama_avxvnni.so +0 -0
  16. ipex_llm/libs/libstarcoder_amx.so +0 -0
  17. ipex_llm/libs/libstarcoder_avx.so +0 -0
  18. ipex_llm/libs/libstarcoder_avx2.so +0 -0
  19. ipex_llm/libs/libstarcoder_avx512.so +0 -0
  20. ipex_llm/libs/libstarcoder_avxvnni.so +0 -0
  21. ipex_llm/libs/quantize-bloom +0 -0
  22. ipex_llm/libs/quantize-gptneox +0 -0
  23. ipex_llm/libs/quantize-llama +0 -0
  24. ipex_llm/libs/quantize-starcoder +0 -0
  25. ipex_llm/transformers/convert.py +17 -132
  26. ipex_llm/transformers/lookup.py +2 -2
  27. ipex_llm/transformers/low_bit_linear.py +8 -8
  28. ipex_llm/transformers/models/chatglm2.py +1 -192
  29. ipex_llm/transformers/models/minicpmv.py +2 -2
  30. ipex_llm/transformers/models/sd.py +2 -2
  31. ipex_llm/transformers/models/utils.py +14 -89
  32. ipex_llm/transformers/npu_pipeline_model/convert_pipeline.py +5 -8
  33. ipex_llm/transformers/utils.py +5 -20
  34. {ipex_llm-2.2.0b20250106.dist-info → ipex_llm-2.2.0b20250107.dist-info}/METADATA +40 -19
  35. {ipex_llm-2.2.0b20250106.dist-info → ipex_llm-2.2.0b20250107.dist-info}/RECORD +41 -44
  36. ipex_llm/transformers/models/cohere.py +0 -589
  37. ipex_llm/transformers/models/falcon.py +0 -829
  38. ipex_llm/transformers/models/mixtral.py +0 -576
  39. {ipex_llm-2.2.0b20250106.data → ipex_llm-2.2.0b20250107.data}/scripts/ipex-llm-init +0 -0
  40. {ipex_llm-2.2.0b20250106.data → ipex_llm-2.2.0b20250107.data}/scripts/llm-chat +0 -0
  41. {ipex_llm-2.2.0b20250106.data → ipex_llm-2.2.0b20250107.data}/scripts/llm-cli +0 -0
  42. {ipex_llm-2.2.0b20250106.dist-info → ipex_llm-2.2.0b20250107.dist-info}/WHEEL +0 -0
  43. {ipex_llm-2.2.0b20250106.dist-info → ipex_llm-2.2.0b20250107.dist-info}/entry_points.txt +0 -0
  44. {ipex_llm-2.2.0b20250106.dist-info → ipex_llm-2.2.0b20250107.dist-info}/top_level.txt +0 -0
@@ -19,7 +19,7 @@ import torch
19
19
  import warnings
20
20
  from ipex_llm.utils.common import invalidInputError
21
21
  from ipex_llm.ggml.quantize import ggml_tensor_qtype
22
- from ipex_llm.transformers.utils import get_ipex_version, get_xpu_device_type
22
+ from ipex_llm.transformers.utils import get_ipex_version, get_xpu_device_name
23
23
  from ipex_llm.transformers.low_bit_linear import SYM_INT4, SYM_INT8, FP8E5, IQ2_XXS, FP4, FP8E4,\
24
24
  FP6, ASYM_INT4
25
25
 
@@ -85,16 +85,14 @@ def use_quantize_kv_cache(linear: torch.nn.Module, x: torch.Tensor, kv_group: in
85
85
  return os.environ["IPEX_LLM_QUANTIZE_KV_CACHE"] == "1"
86
86
  elif os.environ.get("IPEX_LLM_LOW_MEM", None) is not None:
87
87
  return os.environ["IPEX_LLM_LOW_MEM"] == "1"
88
+ elif linear.qtype in [ggml_tensor_qtype["fp16"], ggml_tensor_qtype["bf16"]]:
89
+ return False
88
90
  else:
89
- return x.device.type == 'xpu' and kv_cache_device_check(x, kv_group) \
90
- and hasattr(linear, "qtype") and \
91
- linear.qtype != ggml_tensor_qtype["fp16"] and linear.qtype != ggml_tensor_qtype["bf16"]
92
-
93
-
94
- def kv_cache_device_check(x: torch.Tensor, kv_group: int) -> bool:
95
- return (get_xpu_device_type(x) in ["mtl", "lnl"] and kv_group <= 1) or \
96
- ((get_xpu_device_type(x) == "arc" or get_xpu_device_type(x) == "flex") and
97
- 1 < x.size(0) and x.size(0) <= 8)
91
+ device_name = get_xpu_device_name(x.device)
92
+ return (
93
+ device_name in ["mtl", "lnl", "arl"] and kv_group == 1
94
+ or device_name in ["arc", "bmg"] and x.size(0) > 1
95
+ )
98
96
 
99
97
 
100
98
  def init_fp8_kv_cache(batch_size, num_heads, current_length, head_dim, device):
@@ -226,57 +224,6 @@ def is_enough_kv_cache_room_4_31(past_key_value, seq_len=1):
226
224
  (past_key_value[0].size(2) + seq_len) * past_key_value[0].size(3)
227
225
 
228
226
 
229
- def use_flash_attention(query, key, attention_mask=None):
230
- # here we support query's shape is always [batch_size, head_num, q_len, head_dim],
231
- # key's shape is always [batch_size, head_num, k_len, head_dim]
232
- invalidInputError(query.dim() == 4,
233
- "Here query input of use_flash_attention should be [batch_size, "
234
- "head_num, q_len, head_dim]")
235
- invalidInputError(key.dim() == 4,
236
- "Here key input of use_flash_attention should be [batch_size, "
237
- "head_num, k_len, head_dim]")
238
- bsz, _, q_len, _ = query.size()
239
- k_len = key.size()[2]
240
- # check whether ipex flash attention can be used
241
- if q_len != k_len:
242
- # now only use flash attention for first token
243
- # as it seems have no performance benifit for rest token now
244
- return False
245
- if query.device.type != "xpu":
246
- # ipex flash attention only support for xpu
247
- return False
248
- ipex_version = get_ipex_version()
249
- if ipex_version <= "2.0.110+xpu":
250
- # ipex flash attention is supported from ipex 2.1
251
- return False
252
- if not torch.xpu.has_xetla():
253
- # ipex flash attention is only supported for xetla
254
- # may update this later
255
- return False
256
- elif get_xpu_device_type(query) != "pvc":
257
- return False
258
- if query.dtype not in [torch.float32, torch.float16]:
259
- # only use flash attention for fp32/fp16 input
260
- return False
261
- if bsz > 1:
262
- # as flash attention doesn't support attn_mask in ipex 2.1,
263
- # so it will cause output error for padded batch input
264
- if attention_mask is None:
265
- return True
266
- else:
267
- # TODO: below logic may change for different model
268
- # attention mask shape : [bsz, 1, q_len, k_len]
269
- if attention_mask[0].squeeze()[0, 0].item() != 0:
270
- # first batch contains padding
271
- # otherwise we suppose it should be a upper triangular matrix
272
- # at the same time, the diagonal is also 0
273
- return False
274
- elif not attention_mask.equal(attention_mask[0].repeat(bsz, 1, 1, 1)):
275
- # check whether mask of every batch is the same
276
- return False
277
- return True
278
-
279
-
280
227
  def use_sdp(q_len, kv_len, head_dim, query_states):
281
228
  return (
282
229
  query_states.device.type == "xpu"
@@ -315,38 +262,16 @@ def mlp_fusion_check(x, qtype, training):
315
262
  if training or x.requires_grad:
316
263
  return False
317
264
  if qtype == FP6:
318
- device = get_xpu_device_type(x)
319
- if device in ["mtl", "lnl"]:
265
+ device = get_xpu_device_name(x.device)
266
+ if device in ["mtl", "lnl", "arl"]:
320
267
  return False
321
268
  return True
322
269
 
323
270
 
324
- def use_decoding_fast_path(proj,
325
- use_fuse_rope,
326
- enough_kv_room,
327
- bs,
328
- qtype_check=decoding_fast_path_qtype_check):
329
- if proj is None:
330
- return False
331
- device = get_xpu_device_type(proj.weight)
332
- if not qtype_check(proj):
333
- return False
334
- if not use_fuse_rope:
335
- return False
336
- if not enough_kv_room:
337
- return False
338
- if bs != 1:
339
- return False
340
-
341
- if device in ["uhd"]:
342
- return False
343
- return True
344
-
345
-
346
271
  def use_xmx(x: torch.Tensor, qtype: int):
347
- device = get_xpu_device_type(x)
272
+ device = get_xpu_device_name(x.device)
348
273
  return (
349
- device in ["arc", "flex", "pvc"]
274
+ device in ["arc", "pvc"]
350
275
  and qtype in [SYM_INT4, SYM_INT8, FP8E4, FP8E5]
351
276
  and (
352
277
  (device == "pvc" and 1 < x.size(0) <= 16)
@@ -370,7 +295,7 @@ def fp16_fusion_check(proj, x, training):
370
295
  return False
371
296
  if x.requires_grad:
372
297
  return False
373
- device_type = get_xpu_device_type(x)
298
+ device_type = get_xpu_device_name(x.device)
374
299
  if device_type != "pvc":
375
300
  return False
376
301
  return True
@@ -439,7 +364,7 @@ def should_use_compresskv(x: torch.Tensor, prompt_len: int):
439
364
  else:
440
365
  if use_compress_kv is None:
441
366
  return (
442
- get_xpu_device_type(x) in ["mtl", "lnl"]
367
+ get_xpu_device_name(x.device) in ["mtl", "lnl", "arl"]
443
368
  and prompt_len >= 1800
444
369
  and prompt_len <= 4500
445
370
  )
@@ -473,10 +473,6 @@ def convert_llm_for_deploy(model: torch.nn.Module,
473
473
  "n_splits_linear": n_splits_linear,
474
474
  "n_splits_down_proj": n_splits_down_proj,
475
475
  "lm_head_low_bit": lm_head_low_bit}
476
- model.config.update(update_dict)
477
- model.config.save_pretrained(save_directory)
478
- if model.can_generate():
479
- model.generation_config.save_pretrained(save_directory)
480
476
 
481
477
  from .qwen import convert_qwen_layer, convert_fused_qwen_layer
482
478
  from .qwen import convert_lm_head_and_embedding
@@ -537,8 +533,6 @@ def convert_llm_for_deploy(model: torch.nn.Module,
537
533
  "n_splits_linear": n_splits_linear,
538
534
  "n_splits_down_proj": n_splits_down_proj,
539
535
  "lm_head_low_bit": lm_head_low_bit}
540
- model.config.update(update_dict)
541
- model.config.save_pretrained(save_directory)
542
536
 
543
537
  from .llama import convert_llama_layer, convert_fused_llama_layer
544
538
  from .llama import convert_lm_head_and_embedding
@@ -577,8 +571,6 @@ def convert_llm_for_deploy(model: torch.nn.Module,
577
571
  "n_splits_linear": n_splits_linear,
578
572
  "n_splits_down_proj": n_splits_down_proj,
579
573
  "lm_head_low_bit": lm_head_low_bit}
580
- model.config.update(update_dict)
581
- model.config.save_pretrained(save_directory)
582
574
 
583
575
  from .minicpm import convert_minicpm_layer, convert_fused_minicpm_layer
584
576
  from .minicpm import convert_lm_head_and_embedding
@@ -595,3 +587,8 @@ def convert_llm_for_deploy(model: torch.nn.Module,
595
587
  save_directory, weight_dir,
596
588
  convert_model=True,
597
589
  max_prompt_len=max_prompt_len)
590
+
591
+ model.config.update(update_dict)
592
+ model.config.save_pretrained(save_directory)
593
+ if model.can_generate():
594
+ model.generation_config.save_pretrained(save_directory)
@@ -168,27 +168,12 @@ def get_ipex_version():
168
168
  return _ipex_version
169
169
 
170
170
 
171
- def get_xpu_device_type(x):
172
- if x.device.type != "xpu":
173
- return x.device.type
174
- name = torch.xpu.get_device_name(x.device.index)
175
- if name.startswith("Intel(R) Arc(TM) A"):
176
- return "arc"
177
- elif name.startswith("Intel(R) Graphics [0xe20b]"):
178
- return "bmg"
179
- elif name.startswith("Intel(R) Arc(TM)"):
180
- if 'V' in name:
181
- return "lnl"
182
- else:
183
- return "mtl"
184
- elif name.startswith("Intel(R) Data Center GPU Flex"):
185
- return "flex"
186
- elif name.startswith("Intel(R) Data Center GPU Max"):
187
- return "pvc"
188
- elif name.startswith("Intel(R) UHD"):
189
- return "uhd"
171
+ def get_xpu_device_name(device: torch.device):
172
+ if device.type != "xpu":
173
+ return device.type
190
174
  else:
191
- return "others"
175
+ import xe_linear
176
+ return xe_linear.get_xpu_device_name(device)
192
177
 
193
178
 
194
179
  def load_imatrix_data(imatrix_file):
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: ipex-llm
3
- Version: 2.2.0b20250106
3
+ Version: 2.2.0b20250107
4
4
  Summary: Large Language Model Develop Toolkit
5
5
  Home-page: https://github.com/intel-analytics/ipex-llm
6
6
  Author: BigDL Authors
@@ -27,15 +27,17 @@ Requires-Dist: intel-openmp ; (platform_machine == "x86_64" or platform_machine
27
27
  Requires-Dist: torch ==2.1.2+cpu ; (platform_system == "Linux") and extra == 'all'
28
28
  Requires-Dist: torch ==2.1.2 ; (platform_system == "Windows") and extra == 'all'
29
29
  Provides-Extra: cpp
30
- Requires-Dist: bigdl-core-cpp ==2.6.0b20250106 ; extra == 'cpp'
30
+ Requires-Dist: bigdl-core-cpp ==2.6.0b20250107 ; extra == 'cpp'
31
31
  Requires-Dist: setuptools ; extra == 'cpp'
32
32
  Provides-Extra: cpp-arl
33
- Requires-Dist: bigdl-core-cpp ==2.6.0b20250106 ; extra == 'cpp-arl'
33
+ Requires-Dist: bigdl-core-cpp ==2.6.0b20250107 ; extra == 'cpp-arl'
34
34
  Requires-Dist: setuptools ; extra == 'cpp-arl'
35
35
  Requires-Dist: onednn-devel ==2024.1.1 ; (platform_system == "Windows") and extra == 'cpp-arl'
36
+ Requires-Dist: onednn ==2024.1.1 ; (platform_system == "Windows") and extra == 'cpp-arl'
36
37
  Requires-Dist: dpcpp-cpp-rt ==2024.2.1 ; (platform_system == "Windows") and extra == 'cpp-arl'
37
38
  Requires-Dist: mkl-dpcpp ==2024.2.1 ; (platform_system == "Windows") and extra == 'cpp-arl'
38
39
  Requires-Dist: onednn-devel ==2024.2.1 ; (platform_system == "Windows") and extra == 'cpp'
40
+ Requires-Dist: onednn ==2024.2.1 ; (platform_system == "Windows") and extra == 'cpp'
39
41
  Requires-Dist: dpcpp-cpp-rt ==2024.2.1 ; (platform_system == "Windows") and extra == 'cpp'
40
42
  Requires-Dist: mkl-dpcpp ==2024.2.1 ; (platform_system == "Windows") and extra == 'cpp'
41
43
  Provides-Extra: llama-index
@@ -65,7 +67,7 @@ Requires-Dist: transformers ==4.40.0 ; extra == 'npu'
65
67
  Requires-Dist: intel-openmp ; (platform_machine == "x86_64" or platform_machine == "AMD64") and extra == 'npu'
66
68
  Requires-Dist: torch ==2.1.2+cpu ; (platform_system == "Linux") and extra == 'npu'
67
69
  Requires-Dist: torch ==2.1.2 ; (platform_system == "Windows") and extra == 'npu'
68
- Requires-Dist: bigdl-core-npu ==2.6.0b20250106 ; (platform_system == "Windows") and extra == 'npu'
70
+ Requires-Dist: bigdl-core-npu ==2.6.0b20250107 ; (platform_system == "Windows") and extra == 'npu'
69
71
  Provides-Extra: serving
70
72
  Requires-Dist: py-cpuinfo ; extra == 'serving'
71
73
  Requires-Dist: fschat[model_worker,webui] ==0.2.36 ; extra == 'serving'
@@ -85,9 +87,9 @@ Requires-Dist: setuptools <70.0.0 ; extra == 'xpu'
85
87
  Requires-Dist: torch ==2.1.0a0 ; extra == 'xpu'
86
88
  Requires-Dist: torchvision ==0.16.0a0 ; extra == 'xpu'
87
89
  Requires-Dist: intel-extension-for-pytorch ==2.1.10+xpu ; extra == 'xpu'
88
- Requires-Dist: bigdl-core-xe-21 ==2.6.0b20250106 ; extra == 'xpu'
89
- Requires-Dist: bigdl-core-xe-batch-21 ==2.6.0b20250106 ; extra == 'xpu'
90
- Requires-Dist: bigdl-core-xe-addons-21 ==2.6.0b20250106 ; extra == 'xpu'
90
+ Requires-Dist: bigdl-core-xe-21 ==2.6.0b20250107 ; extra == 'xpu'
91
+ Requires-Dist: bigdl-core-xe-batch-21 ==2.6.0b20250107 ; extra == 'xpu'
92
+ Requires-Dist: bigdl-core-xe-addons-21 ==2.6.0b20250107 ; extra == 'xpu'
91
93
  Provides-Extra: xpu-2-1
92
94
  Requires-Dist: py-cpuinfo ; extra == 'xpu-2-1'
93
95
  Requires-Dist: protobuf ; extra == 'xpu-2-1'
@@ -102,12 +104,28 @@ Requires-Dist: setuptools <70.0.0 ; extra == 'xpu-2-1'
102
104
  Requires-Dist: torch ==2.1.0a0 ; extra == 'xpu-2-1'
103
105
  Requires-Dist: torchvision ==0.16.0a0 ; extra == 'xpu-2-1'
104
106
  Requires-Dist: intel-extension-for-pytorch ==2.1.10+xpu ; extra == 'xpu-2-1'
105
- Requires-Dist: bigdl-core-xe-21 ==2.6.0b20250106 ; extra == 'xpu-2-1'
106
- Requires-Dist: bigdl-core-xe-batch-21 ==2.6.0b20250106 ; extra == 'xpu-2-1'
107
- Requires-Dist: bigdl-core-xe-addons-21 ==2.6.0b20250106 ; extra == 'xpu-2-1'
107
+ Requires-Dist: bigdl-core-xe-21 ==2.6.0b20250107 ; extra == 'xpu-2-1'
108
+ Requires-Dist: bigdl-core-xe-batch-21 ==2.6.0b20250107 ; extra == 'xpu-2-1'
109
+ Requires-Dist: bigdl-core-xe-addons-21 ==2.6.0b20250107 ; extra == 'xpu-2-1'
108
110
  Requires-Dist: intel-openmp ; (platform_machine == "x86_64" or platform_machine == "AMD64") and extra == 'xpu-2-1'
109
111
  Requires-Dist: dpcpp-cpp-rt ==2024.0.2 ; (platform_system == "Windows") and extra == 'xpu-2-1'
110
112
  Requires-Dist: mkl-dpcpp ==2024.0.0 ; (platform_system == "Windows") and extra == 'xpu-2-1'
113
+ Provides-Extra: xpu-2-6
114
+ Requires-Dist: py-cpuinfo ; extra == 'xpu-2-6'
115
+ Requires-Dist: protobuf ; extra == 'xpu-2-6'
116
+ Requires-Dist: mpmath ==1.3.0 ; extra == 'xpu-2-6'
117
+ Requires-Dist: numpy ==1.26.4 ; extra == 'xpu-2-6'
118
+ Requires-Dist: transformers ==4.37.0 ; extra == 'xpu-2-6'
119
+ Requires-Dist: sentencepiece ; extra == 'xpu-2-6'
120
+ Requires-Dist: tokenizers ==0.15.2 ; extra == 'xpu-2-6'
121
+ Requires-Dist: accelerate ==0.23.0 ; extra == 'xpu-2-6'
122
+ Requires-Dist: tabulate ; extra == 'xpu-2-6'
123
+ Requires-Dist: setuptools ; extra == 'xpu-2-6'
124
+ Requires-Dist: torch ==2.6.0+xpu ; extra == 'xpu-2-6'
125
+ Requires-Dist: torchvision ==0.21.0+xpu ; extra == 'xpu-2-6'
126
+ Requires-Dist: torchaudio ==2.6.0+xpu ; extra == 'xpu-2-6'
127
+ Requires-Dist: bigdl-core-xe-all ==2.6.0b20250107 ; extra == 'xpu-2-6'
128
+ Requires-Dist: intel-openmp ; (platform_machine == "x86_64" or platform_machine == "AMD64") and extra == 'xpu-2-6'
111
129
  Provides-Extra: xpu-arc
112
130
  Requires-Dist: py-cpuinfo ; extra == 'xpu-arc'
113
131
  Requires-Dist: protobuf ; extra == 'xpu-arc'
@@ -119,9 +137,9 @@ Requires-Dist: tokenizers ==0.15.2 ; extra == 'xpu-arc'
119
137
  Requires-Dist: accelerate ==0.23.0 ; extra == 'xpu-arc'
120
138
  Requires-Dist: tabulate ; extra == 'xpu-arc'
121
139
  Requires-Dist: setuptools ; extra == 'xpu-arc'
122
- Requires-Dist: bigdl-core-xe-23 ==2.6.0b20250106 ; extra == 'xpu-arc'
123
- Requires-Dist: bigdl-core-xe-batch-23 ==2.6.0b20250106 ; extra == 'xpu-arc'
124
- Requires-Dist: bigdl-core-xe-addons-23 ==2.6.0b20250106 ; extra == 'xpu-arc'
140
+ Requires-Dist: bigdl-core-xe-23 ==2.6.0b20250107 ; extra == 'xpu-arc'
141
+ Requires-Dist: bigdl-core-xe-batch-23 ==2.6.0b20250107 ; extra == 'xpu-arc'
142
+ Requires-Dist: bigdl-core-xe-addons-23 ==2.6.0b20250107 ; extra == 'xpu-arc'
125
143
  Requires-Dist: intel-openmp ; (platform_machine == "x86_64" or platform_machine == "AMD64") and extra == 'xpu-arc'
126
144
  Requires-Dist: torch ==2.3.1+cxx11.abi ; (platform_system == "Linux") and extra == 'xpu-arc'
127
145
  Requires-Dist: torchvision ==0.18.1+cxx11.abi ; (platform_system == "Linux") and extra == 'xpu-arc'
@@ -130,6 +148,7 @@ Requires-Dist: torch ==2.3.1.post0+cxx11.abi ; (platform_system == "Windows") an
130
148
  Requires-Dist: torchvision ==0.18.1.post0+cxx11.abi ; (platform_system == "Windows") and extra == 'xpu-arc'
131
149
  Requires-Dist: intel-extension-for-pytorch ==2.3.110.post0+xpu ; (platform_system == "Windows") and extra == 'xpu-arc'
132
150
  Requires-Dist: onednn-devel ==2024.1.1 ; (platform_system == "Windows") and extra == 'xpu-arc'
151
+ Requires-Dist: onednn ==2024.1.1 ; (platform_system == "Windows") and extra == 'xpu-arc'
133
152
  Provides-Extra: xpu-arl
134
153
  Requires-Dist: py-cpuinfo ; extra == 'xpu-arl'
135
154
  Requires-Dist: protobuf ; extra == 'xpu-arl'
@@ -141,9 +160,9 @@ Requires-Dist: tokenizers ==0.15.2 ; extra == 'xpu-arl'
141
160
  Requires-Dist: accelerate ==0.23.0 ; extra == 'xpu-arl'
142
161
  Requires-Dist: tabulate ; extra == 'xpu-arl'
143
162
  Requires-Dist: setuptools ; extra == 'xpu-arl'
144
- Requires-Dist: bigdl-core-xe-23 ==2.6.0b20250106 ; extra == 'xpu-arl'
145
- Requires-Dist: bigdl-core-xe-batch-23 ==2.6.0b20250106 ; extra == 'xpu-arl'
146
- Requires-Dist: bigdl-core-xe-addons-23 ==2.6.0b20250106 ; extra == 'xpu-arl'
163
+ Requires-Dist: bigdl-core-xe-23 ==2.6.0b20250107 ; extra == 'xpu-arl'
164
+ Requires-Dist: bigdl-core-xe-batch-23 ==2.6.0b20250107 ; extra == 'xpu-arl'
165
+ Requires-Dist: bigdl-core-xe-addons-23 ==2.6.0b20250107 ; extra == 'xpu-arl'
147
166
  Requires-Dist: intel-openmp ; (platform_machine == "x86_64" or platform_machine == "AMD64") and extra == 'xpu-arl'
148
167
  Requires-Dist: torch ==2.3.1+cxx11.abi ; (platform_system == "Linux") and extra == 'xpu-arl'
149
168
  Requires-Dist: torchvision ==0.18.1+cxx11.abi ; (platform_system == "Linux") and extra == 'xpu-arl'
@@ -152,6 +171,7 @@ Requires-Dist: torch ==2.3.1.post0+cxx11.abi ; (platform_system == "Windows") an
152
171
  Requires-Dist: torchvision ==0.18.1.post0+cxx11.abi ; (platform_system == "Windows") and extra == 'xpu-arl'
153
172
  Requires-Dist: intel-extension-for-pytorch ==2.3.110.post0+xpu ; (platform_system == "Windows") and extra == 'xpu-arl'
154
173
  Requires-Dist: onednn-devel ==2024.1.1 ; (platform_system == "Windows") and extra == 'xpu-arl'
174
+ Requires-Dist: onednn ==2024.1.1 ; (platform_system == "Windows") and extra == 'xpu-arl'
155
175
  Provides-Extra: xpu-lnl
156
176
  Requires-Dist: py-cpuinfo ; extra == 'xpu-lnl'
157
177
  Requires-Dist: protobuf ; extra == 'xpu-lnl'
@@ -163,9 +183,9 @@ Requires-Dist: tokenizers ==0.15.2 ; extra == 'xpu-lnl'
163
183
  Requires-Dist: accelerate ==0.23.0 ; extra == 'xpu-lnl'
164
184
  Requires-Dist: tabulate ; extra == 'xpu-lnl'
165
185
  Requires-Dist: setuptools ; extra == 'xpu-lnl'
166
- Requires-Dist: bigdl-core-xe-23 ==2.6.0b20250106 ; extra == 'xpu-lnl'
167
- Requires-Dist: bigdl-core-xe-batch-23 ==2.6.0b20250106 ; extra == 'xpu-lnl'
168
- Requires-Dist: bigdl-core-xe-addons-23 ==2.6.0b20250106 ; extra == 'xpu-lnl'
186
+ Requires-Dist: bigdl-core-xe-23 ==2.6.0b20250107 ; extra == 'xpu-lnl'
187
+ Requires-Dist: bigdl-core-xe-batch-23 ==2.6.0b20250107 ; extra == 'xpu-lnl'
188
+ Requires-Dist: bigdl-core-xe-addons-23 ==2.6.0b20250107 ; extra == 'xpu-lnl'
169
189
  Requires-Dist: intel-openmp ; (platform_machine == "x86_64" or platform_machine == "AMD64") and extra == 'xpu-lnl'
170
190
  Requires-Dist: torch ==2.3.1+cxx11.abi ; (platform_system == "Linux") and extra == 'xpu-lnl'
171
191
  Requires-Dist: torchvision ==0.18.1+cxx11.abi ; (platform_system == "Linux") and extra == 'xpu-lnl'
@@ -174,6 +194,7 @@ Requires-Dist: torch ==2.3.1.post0+cxx11.abi ; (platform_system == "Windows") an
174
194
  Requires-Dist: torchvision ==0.18.1.post0+cxx11.abi ; (platform_system == "Windows") and extra == 'xpu-lnl'
175
195
  Requires-Dist: intel-extension-for-pytorch ==2.3.110.post0+xpu ; (platform_system == "Windows") and extra == 'xpu-lnl'
176
196
  Requires-Dist: onednn-devel ==2024.1.1 ; (platform_system == "Windows") and extra == 'xpu-lnl'
197
+ Requires-Dist: onednn ==2024.1.1 ; (platform_system == "Windows") and extra == 'xpu-lnl'
177
198
  Requires-Dist: intel-openmp ; (platform_machine == "x86_64" or platform_machine == "AMD64") and extra == 'xpu'
178
199
  Requires-Dist: dpcpp-cpp-rt ==2024.0.2 ; (platform_system == "Windows") and extra == 'xpu'
179
200
  Requires-Dist: mkl-dpcpp ==2024.0.0 ; (platform_system == "Windows") and extra == 'xpu'
@@ -44,39 +44,39 @@ ipex_llm/langchain/vllm/__init__.py,sha256=T-EbRT6GJ_8RCu-iLmSzcftOimXSPQf2d5X72
44
44
  ipex_llm/langchain/vllm/vllm.py,sha256=6dxc-ZISZQrJilEa_HA827l75Dv9rcHpY_G6FdJ8BVs,7793
45
45
  ipex_llm/libs/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
46
46
  ipex_llm/libs/libbloom-api.so,sha256=ylQsq2MM8zLHgypOPNGvQnUN2xYraaVbzimhJTKGF9s,33576
47
- ipex_llm/libs/libbloom_amx.so,sha256=vyxQmiMnVKu3dWXwVsiWwN69LK1WBf7DT8yc4ujxjCE,528856
48
- ipex_llm/libs/libbloom_avx.so,sha256=rfBudWvF7TITnEqDoDo9Cjwkr5_aWGQ3WuXA8104-jI,536152
49
- ipex_llm/libs/libbloom_avx2.so,sha256=dqqgS5Pic5-sdkbhjDmOqtNGPYBLzvUX-jeL6lgmLHM,519680
50
- ipex_llm/libs/libbloom_avx512.so,sha256=oUNIE-QA4axkWfomxfCDSIVVk2wl8YrslgQMeEfq5iY,515584
51
- ipex_llm/libs/libbloom_avxvnni.so,sha256=HXAjpFmWCwQFuN3-aodDskNRi4t4vLeaYpKLubViAXo,519680
47
+ ipex_llm/libs/libbloom_amx.so,sha256=9eotMu7MqGcSsmhV8zvMwmQi_uj0h7iQws4SJE_IM6Y,528904
48
+ ipex_llm/libs/libbloom_avx.so,sha256=Sw7nGyCouy6Oz-EaPtYtzZj5naRWqjGt_wUHsnqwwZk,536208
49
+ ipex_llm/libs/libbloom_avx2.so,sha256=Vyz0PEhqtFAQNz01lCvikBMcOgyy_WGLjKQ4BYaL8Us,519736
50
+ ipex_llm/libs/libbloom_avx512.so,sha256=Y7ClRp_BqKqtTTgspykrO7jwpInNxvf5mEQZ682X9ns,519736
51
+ ipex_llm/libs/libbloom_avxvnni.so,sha256=IwD7iRyguGtUcfeD16VYeYMKkddKsGE-aJIz4YJeV-I,519736
52
52
  ipex_llm/libs/libgptneox-api.so,sha256=X6glOWTpK-bw9hmP2gGLcQpwqThWwyDuPqxdNigbp4w,28240
53
- ipex_llm/libs/libgptneox_amx.so,sha256=uc5spwmksH_MeZaw46jQiHci2D1wuB0YXpJ7GC9HEmU,581952
54
- ipex_llm/libs/libgptneox_avx.so,sha256=pA0H2xKMXxww1vDzvxedyTKBQNVAU167zZnynmFSSak,589208
55
- ipex_llm/libs/libgptneox_avx2.so,sha256=HklIdD-26xokiLRFZ0-rTtV5VKxUx7Tomn6zN629-Og,576832
56
- ipex_llm/libs/libgptneox_avx512.so,sha256=pINXzvghq0B2IrEtMudHJiug2E0swgjFvqXRbtCNKUM,572736
57
- ipex_llm/libs/libgptneox_avxvnni.so,sha256=NNlvVW0FJG4tynJAsvtpwStw_Ez_aE2Z8JAAQtG-BCA,572736
53
+ ipex_llm/libs/libgptneox_amx.so,sha256=gS61QqlSl-hAYEbSRquuAZrlzSSMlSYh7Gs56Qk5MjA,582000
54
+ ipex_llm/libs/libgptneox_avx.so,sha256=6GA2dnHSSFjSnv0ZYo27cJ1sBLvTH-xX9kT1DV7AF3Y,589264
55
+ ipex_llm/libs/libgptneox_avx2.so,sha256=GxZ3Al19J3Iddir9xBWaOOZDt9Hqc12eIv8yRBj7KXc,576888
56
+ ipex_llm/libs/libgptneox_avx512.so,sha256=FKms4rOGCzeNBstcaIYvh0_v6S1QN-ABlVJ3dsFW19k,572792
57
+ ipex_llm/libs/libgptneox_avxvnni.so,sha256=1djfToDCw-x-NO4aXdHNAcaI9lFhFIFDA41eqjhgJmk,572792
58
58
  ipex_llm/libs/libjemalloc.so,sha256=gBgVkr0vI-5w7KdT38ez485J6RvRlZE1-Xe3JnLe4h8,5460184
59
59
  ipex_llm/libs/libllama-api.so,sha256=ffrgtLhkdGf2dv3jrTf1J06pwLQMYU57VXcX3LuU400,32688
60
- ipex_llm/libs/libllama_amx.so,sha256=vjViSXQJqy6zPG_fGe0YWrD00xb1UgLDBCzWJRaDvy8,581424
61
- ipex_llm/libs/libllama_avx.so,sha256=pySdsGHAg4ANLigbuy4HibiIPFUc6LZKkSxwLzUQlY0,592776
62
- ipex_llm/libs/libllama_avx2.so,sha256=Pk13Hy9SmX7EOow7mvTT_hz0FViK0qPKsM29orpiEtg,576304
63
- ipex_llm/libs/libllama_avx512.so,sha256=mXGgBKVEsukWVAs09fHjvRRURtYaH1wIvEDANjgAI84,576304
64
- ipex_llm/libs/libllama_avxvnni.so,sha256=TONV4V-sJMwU2QEGUyozTAzDOhLc40eR8Ck-F26hbrU,576304
60
+ ipex_llm/libs/libllama_amx.so,sha256=cS-ylndBti79VeW5QRc-sijqOlB94zuEjEXEFUBwbwA,581472
61
+ ipex_llm/libs/libllama_avx.so,sha256=PFg4oD28nRrJR8dAlrH-VMSnaBb-bjjcd5tpSwSsQPw,592832
62
+ ipex_llm/libs/libllama_avx2.so,sha256=xBLX-8Rgiy0ZBA-X41PRW4Wba7uB-mJMN3gwS7lj-Hw,576360
63
+ ipex_llm/libs/libllama_avx512.so,sha256=jFtLlTiayfi1V-RPnwI53aZftySJ1D0kHSg4OFdKa1g,576360
64
+ ipex_llm/libs/libllama_avxvnni.so,sha256=qWCPBASh04sZSzUE9hD4ctkn0B1nHMvKUfNfuBgwLLI,576360
65
65
  ipex_llm/libs/libstarcoder-api.so,sha256=C126ZZ0Yr80mpQx77U4egdbijpTJEa6gvzK5ASP6CcY,18368
66
- ipex_llm/libs/libstarcoder_amx.so,sha256=msBmnqcB3sqf2IYPBFU9A0M3mkTYeMrCi52cxSloW_c,721424
67
- ipex_llm/libs/libstarcoder_avx.so,sha256=otmtdNDgTa9iQL7GiJ9J3ruhqS2aD1LrtoKTWAGcPfc,728720
68
- ipex_llm/libs/libstarcoder_avx2.so,sha256=FAjKK0eGEdaGWIJI8i3Io1KsLuAaYR-GzQRmpFsenR4,716344
69
- ipex_llm/libs/libstarcoder_avx512.so,sha256=la8_1GyZv2dPMM0-GeY3MXPJiEVyC6WUghNKAd3s648,716344
70
- ipex_llm/libs/libstarcoder_avxvnni.so,sha256=3AdgMHBpTdLYdPr_utmNp0VnRyXeq0N8ri3O-QrtCYQ,716344
66
+ ipex_llm/libs/libstarcoder_amx.so,sha256=hWD_NoOfXVhaqAiscIqBTpxOwTJoC89su-aEqazwL5U,721472
67
+ ipex_llm/libs/libstarcoder_avx.so,sha256=abVPGV8VFfQJRO-57YrPvt1GACGn3L9TT8Ch3gkojfU,732864
68
+ ipex_llm/libs/libstarcoder_avx2.so,sha256=RoTEOUIBifaJ2F_mVlkLaCABW38FblBSkD9aA3OrNrk,716392
69
+ ipex_llm/libs/libstarcoder_avx512.so,sha256=b7pD2zug1gR57u9j8tqz9Gi888jjtNk0dZpSyd90vw4,716392
70
+ ipex_llm/libs/libstarcoder_avxvnni.so,sha256=eC4hTX627LhKQJ5vsVhKG4-BcHroOfS4kjdiUR7LMms,716392
71
71
  ipex_llm/libs/libtcmalloc.so,sha256=sGmZ-oxCTzm9O7gBFY_Q5tInlih1tBIFfkOhvrQu2zU,393904
72
72
  ipex_llm/libs/main-bloom,sha256=TeVWpZ60CdFW5yN3ejZ6bHCnWxqG5tZo9pXw13ncTPc,130416
73
73
  ipex_llm/libs/main-gptneox,sha256=3Y0viffGzc0Gh3XTXHY4tvt6M4nkjHAbIJ1oDFqFqm0,116208
74
74
  ipex_llm/libs/main-llama,sha256=5Ml3WWow-tMXCbUQXqtXTnKqfbUhzMNHDo4v0DVA5ks,116560
75
75
  ipex_llm/libs/main-starcoder,sha256=ykjZ9Me3Av2FirUamVMl5rrBloLxdW_GAYQwGab5FXU,263336
76
- ipex_llm/libs/quantize-bloom,sha256=d7Fg26B1J8QUvc7zqhCQ8kmpEUdG2gQiHezp0sq0Bu4,560656
77
- ipex_llm/libs/quantize-gptneox,sha256=pzuFlKbdQBJ4Tv3CH35cfa2790h5m1nxYrZF2AATxnA,545432
78
- ipex_llm/libs/quantize-llama,sha256=VmORvJULYcjXkExOXTByXuU-OS-rtd1XyDjibXJNFd8,549624
79
- ipex_llm/libs/quantize-starcoder,sha256=D0AzAIzOeMXqxyZQ3XPh8k3pEINfBEw1gG3UBC7WUFQ,583656
76
+ ipex_llm/libs/quantize-bloom,sha256=hJollIrE2WtID438DlsFg3SqMb6vlJfpunVkuyJmKCo,560712
77
+ ipex_llm/libs/quantize-gptneox,sha256=Ozwniigz59nbOLJmsjb0E-SIU7BH5gQt5szwU4XvQgY,545480
78
+ ipex_llm/libs/quantize-llama,sha256=IXKt7TRkK_nFcIHIUzped-7sZ_y6251ppp0lenqOFm8,553776
79
+ ipex_llm/libs/quantize-starcoder,sha256=Fg6ssZfO2oHPjrmKIo4US9fVF_3sqqClrGcMyY1AkQo,583704
80
80
  ipex_llm/llamaindex/__init__.py,sha256=T-EbRT6GJ_8RCu-iLmSzcftOimXSPQf2d5X72AUAy2Y,874
81
81
  ipex_llm/llamaindex/llms/__init__.py,sha256=KP1lEdGqDuxPoxL1ZSH25Pm2kKMPJBWUTLR0ckSLMIU,1139
82
82
  ipex_llm/llamaindex/llms/bigdlllm.py,sha256=FQBzq1KOjfc6uofTXAha3O7TqpJkNfOFepXQmOVlbnI,26314
@@ -94,14 +94,14 @@ ipex_llm/serving/fastchat/tgi_api_protocol.py,sha256=brT3k3-V0NJrU4fRqUwWjC0O3iO
94
94
  ipex_llm/serving/fastchat/tgi_api_server.py,sha256=agNTAEiZPSuj3dEdIdYKwkoY0cXOUDX06DiM9VP2knQ,24418
95
95
  ipex_llm/serving/fastchat/vllm_worker.py,sha256=ZLz2Q9GxJO6r_LOiP6epgCRjBGk-K4EB1SNEWSJp5DA,11091
96
96
  ipex_llm/transformers/__init__.py,sha256=l4KkMkLe-pRC7b_kj6LCfeifgE-Uo33_Av_FwN9HnFA,1074
97
- ipex_llm/transformers/convert.py,sha256=2kcaxal7x1ltv_d-qE1r0FXlYNEwlQ2LwAMXLuz-X9Y,106252
97
+ ipex_llm/transformers/convert.py,sha256=pFm6VlU84u_Llr2sp6-gRrEYDeNgIk2QPukolq4IE1s,99947
98
98
  ipex_llm/transformers/convert_ipex.py,sha256=iKXo0n8fVFTOA2fNYYrByMFK0dovL-kLd2sVDk88AlQ,14334
99
99
  ipex_llm/transformers/embedding.py,sha256=bdgk59DvD4ZZyxRzewXOR7g56nThgO6uhIwk8QL7f-s,9299
100
100
  ipex_llm/transformers/kv.py,sha256=k4TU18LlA-Sbq9WNNQnfuzu3RSFBwFhmaV3BcGN5bAo,19191
101
101
  ipex_llm/transformers/lisa.py,sha256=F5WxbtXQ7RdKulj83h_2DnEIgKiKGZf7zvOmg6QBl2s,3289
102
102
  ipex_llm/transformers/loader.py,sha256=cOgX93xOC-4dt01GTJ5wyd7PjZ8S43r4mctkR2YxVuw,6893
103
- ipex_llm/transformers/lookup.py,sha256=c4ETIha6ZLbWvhcclSKRDdi5Ipuet4mfUnOkBa0E8kk,19607
104
- ipex_llm/transformers/low_bit_linear.py,sha256=dyyYyCqw0GK8hzaUGanrg-uIhU1HTLEEbvbxXMlm-80,41668
103
+ ipex_llm/transformers/lookup.py,sha256=b6OlZ9OV10R9qeWw8mVryVpDxszkjwLkldvi7GPMJY8,19614
104
+ ipex_llm/transformers/low_bit_linear.py,sha256=nKraUvZJ7UdXP29HSE4CJPIVxmN-TvG8dpT4gpleuyQ,41688
105
105
  ipex_llm/transformers/model.py,sha256=KcRjkauGg48BYrUBoUZaVMpg7Piuz5JrfIpVZd3EIjs,41105
106
106
  ipex_llm/transformers/modelling_bigdl.py,sha256=7JpNVMuyq_OmtNUaMFMXdxPWZp2q0QHC02QeA-VTPOw,6709
107
107
  ipex_llm/transformers/npu_model.py,sha256=YW02GeVz-9ZGqxAeSz0AOvciS-17bo9eK5ZOBrICwSQ,39508
@@ -112,7 +112,7 @@ ipex_llm/transformers/relora.py,sha256=-dYzUV0P-IhO2jFdnzN9-v_sFzJpRj3ZwN9eCJzOo
112
112
  ipex_llm/transformers/speculative.py,sha256=Zf1nQb5GXpJQrUHBTL-H4RUBfdv3lGhfehzudHimhYk,64109
113
113
  ipex_llm/transformers/streamer.py,sha256=RrVlLblzCOtABRUpaMXAyaMnCGgLUtAi_YesLumRbww,4842
114
114
  ipex_llm/transformers/training_patch.py,sha256=oxMkUtqyvqJiprw6dE3skkYfD1HOmUlH9N0hBkbn0G0,10799
115
- ipex_llm/transformers/utils.py,sha256=IRzmoRwgCwnOaiDroSyqc2pvlj-ipNvcW8RubvVT_rY,17374
115
+ ipex_llm/transformers/utils.py,sha256=fXLIlr9hoBr27p3w3xzczZGPk2cCTIRbUKBkiVCGYbc,16889
116
116
  ipex_llm/transformers/xpu_customize_fwd.py,sha256=wFpIhs5F6tkNs8gBOrLxWdhLzO3EDHovVkERPIAoAvg,7611
117
117
  ipex_llm/transformers/xpu_ops.py,sha256=H46-69pMRQhekbAEoDfNacCInLWycMHDqrgMGLvFYfI,4362
118
118
  ipex_llm/transformers/awq/__init__.py,sha256=Du5gu3-eeAkeDO_dEMBTzrDBA66DSN3uL3-rn8WGXQw,875
@@ -148,13 +148,11 @@ ipex_llm/transformers/models/baichuan.py,sha256=oJCAEENSG8oQhJ-QPN2SiapARjAGdOM6
148
148
  ipex_llm/transformers/models/bert.py,sha256=bJNic2pt1kph0kBwdK5MRGyWupFfx2Ts0V3D1L-5kWo,6085
149
149
  ipex_llm/transformers/models/bloom.py,sha256=PxfzyYT-nFn3K5rZhTQjmcEjUUzAhUFzxIN4kzRlCuc,8103
150
150
  ipex_llm/transformers/models/chatglm.py,sha256=UHai1t2AUtGmF765_eHF8LUMVQzp_oCBx8TJB21WrHk,12597
151
- ipex_llm/transformers/models/chatglm2.py,sha256=kfJThuKYb3unAB1XCzfop1iDW1gOkyFOjSr-lEjUdS0,24781
151
+ ipex_llm/transformers/models/chatglm2.py,sha256=SGCABJdYQLW0zDarEoWrEQLuWlbq9iQhYU8ZeR1-ptQ,15957
152
152
  ipex_llm/transformers/models/chatglm4.py,sha256=AAhAFFDDas5DBQPfh2Mwl7a2v7taKf6xphoeeNNFaBI,16593
153
153
  ipex_llm/transformers/models/chatglm4v.py,sha256=YRfuf9g1E0MQ_7wbHAOMvadFnO-j3LqI_k1SaRkDs0M,14055
154
- ipex_llm/transformers/models/cohere.py,sha256=RCUBfhI_eNuZ1NyQC6vm28eYqPmgZn6pHSvQMwFyT2A,25884
155
154
  ipex_llm/transformers/models/common.py,sha256=4obQMGF02FCiXrHnFle9Fsx7C33b1FDt37qJJ4YgxRc,11578
156
155
  ipex_llm/transformers/models/decilm.py,sha256=P-PBuDPf07GvKggLwJx_wPwIn6esN3rX8ai2JxRuZmE,5246
157
- ipex_llm/transformers/models/falcon.py,sha256=f5BzMbv4E-R5Pete8zBscbgiueXGIaWGs-5RbcMlUo4,33549
158
156
  ipex_llm/transformers/models/gemma.py,sha256=_E3Yw8Y45xyNVeLqyVKcpr8kjuICtETeL82cJ-bWJuU,9424
159
157
  ipex_llm/transformers/models/gemma2.py,sha256=2WZuv-FLzJyTJFaYxOuzJt47QE64M0lHnzAiO5T6ozI,8049
160
158
  ipex_llm/transformers/models/glm.py,sha256=gHYgfn20jPRL-ElXy-rUqMh6_LQcc5x7DEXSZuRA4E0,7094
@@ -167,9 +165,8 @@ ipex_llm/transformers/models/internvl.py,sha256=Vx0vENIEQLX2M6P398mw5TOhpks0U8xf
167
165
  ipex_llm/transformers/models/llama.py,sha256=ozwtdQ0MbanJEtW4LBFGxqs_QAq82EonhL2dL6tGyw0,8567
168
166
  ipex_llm/transformers/models/minicpm.py,sha256=ib2rJTN7Tf7znBCtVrtXsF-_Uuk2aA7KVg02xzatLiI,10103
169
167
  ipex_llm/transformers/models/minicpm3.py,sha256=FhNS6mi2rg7dSdF_QQGrao3g9EC6XLn1MTKd-kd0wF0,9191
170
- ipex_llm/transformers/models/minicpmv.py,sha256=igfugwyP3Nu3DcLhYXpe9W9Sk3q2YSrzjj7ukAwuJkg,9829
168
+ ipex_llm/transformers/models/minicpmv.py,sha256=ZV4s48WNIyRoEkvENnlmopnx3ojZANBer0LI6bRtxrY,9826
171
169
  ipex_llm/transformers/models/mistral.py,sha256=rE1GWQxXvF6aG-buPHDR13zeynDZEDIubPF4PiVhZbM,7451
172
- ipex_llm/transformers/models/mixtral.py,sha256=zShaxabIoQaL3cV0Rptf7VCN4QuJsV4KBhanff1TASY,26601
173
170
  ipex_llm/transformers/models/mllama.py,sha256=ogpLmmN_OwcFUyjYB-oDC-l3uw8urFvUEc5edkjWHAk,10939
174
171
  ipex_llm/transformers/models/mpt.py,sha256=z02NwHogJZVh-Mk4sYoIzR90SFIKhoNN_-ifsD907TQ,9540
175
172
  ipex_llm/transformers/models/phi.py,sha256=E6qz4EEuHIVGvaPo-wtLC5lz3iyMqTbAE_cRlcjQRKI,6670
@@ -182,10 +179,10 @@ ipex_llm/transformers/models/qwen2_vl.py,sha256=jIm4yZSd751BkRqgj3wR1QBkDIh-TMCL
182
179
  ipex_llm/transformers/models/qwen_vl.py,sha256=j7Nzzz2Qvynu9yrCXmoEfERjw43hXof5TbXIs7Ms-oY,17105
183
180
  ipex_llm/transformers/models/rwkv4.py,sha256=H4KMtxN0JA2ZTXnonHpsUUJ5xULemo-D1Jzl0ri_UY8,6123
184
181
  ipex_llm/transformers/models/rwkv5.py,sha256=OkRNj1pCAZg1z2Fw-I0DEnxLEdZyPeRSQ6msrkxLOCs,10710
185
- ipex_llm/transformers/models/sd.py,sha256=7qkti_5jf3KmB-W4hDVeYGJvipuwGrxPox_4scGmnIc,5991
182
+ ipex_llm/transformers/models/sd.py,sha256=VvHV5u-0k2MgHu3NL9113hPj7DgfxqctuKzEEeNfRDU,5981
186
183
  ipex_llm/transformers/models/stablelm.py,sha256=RGQCYuQhYqtZ1j3RZkYi0_QvCRnUgUIPYxfBcLnElzg,6885
187
184
  ipex_llm/transformers/models/starcoder2.py,sha256=4P3mhRYf2Kreb1ESjrQGfy1puLMmZXgV35zf-Tksvao,6462
188
- ipex_llm/transformers/models/utils.py,sha256=hpTT9X5zCi2CVlFRR-nApn_1Bsh1S8SHlFJwc-tKxf0,18985
185
+ ipex_llm/transformers/models/utils.py,sha256=Qbz7UkYSbsM5bodH2445O0-JF50Mu3UEwW0j2ZNxHSU,15997
189
186
  ipex_llm/transformers/models/yuan.py,sha256=1jRPebwAK2ENbyYokOmb4LSVo-szucWiygz9zTv-scs,7656
190
187
  ipex_llm/transformers/npu_models/__init__.py,sha256=ulEUGLjaP48LCrVeury3UxLjXxKzRi0UpSG4bYu-7f8,585
191
188
  ipex_llm/transformers/npu_models/baichuan.py,sha256=fJtd7fBrttySghRUgfZTAdxLjsSNC-XL08HISsXigLE,4685
@@ -218,7 +215,7 @@ ipex_llm/transformers/npu_models/xlm_mp.py,sha256=sj8OVun8xJprM7ZJp0XzWa55rqlSIz
218
215
  ipex_llm/transformers/npu_pipeline_model/__init__.py,sha256=b2IXvVqQ5cItki021h8s3ymW12RPu8QNPprq4Mn3bDM,586
219
216
  ipex_llm/transformers/npu_pipeline_model/baichuan.py,sha256=ICxRzFQ4OIANDkkVi2_4xOeQXmfFXYMx3H52KuE1xR4,6208
220
217
  ipex_llm/transformers/npu_pipeline_model/common.py,sha256=QxJoJESpv0BpwO_FBeAT2wKA56wNFfen8iI37PrMKuA,7838
221
- ipex_llm/transformers/npu_pipeline_model/convert_pipeline.py,sha256=953Gua2tFKLICpbmCBm-Lxnur85Ce7jNMeRTNT4DKZE,28715
218
+ ipex_llm/transformers/npu_pipeline_model/convert_pipeline.py,sha256=wrRgmNT13RVtQRp5gFRBxNEPJHxFMLeGqb8a58YodPQ,28512
222
219
  ipex_llm/transformers/npu_pipeline_model/llama.py,sha256=MnvHRytLt3oy5jIPUBe8AeEJ6PtPWLbhQ5a9WqjZ1TQ,19905
223
220
  ipex_llm/transformers/npu_pipeline_model/minicpm.py,sha256=MDMesYlVbECKdK0xxkt1LwHgpkJOO7ZwBExYAwMGQa0,20637
224
221
  ipex_llm/transformers/npu_pipeline_model/pipeline_cpp.py,sha256=JNmodAMg_NQvDILug3E_fGXEh6cd3wsj4bvAzcd-vaU,2749
@@ -254,11 +251,11 @@ ipex_llm/vllm/xpu/engine/__init__.py,sha256=pY_CpyuZd72fr6s32ejeKHKFW0K4vUU2rzZj
254
251
  ipex_llm/vllm/xpu/engine/engine.py,sha256=k4-D27WS_Gk3mA--w3HWAjPjb4Aiu043MVPi0ZoAUBc,5984
255
252
  ipex_llm/vllm/xpu/entrypoints/openai/api_server.py,sha256=GshTZFB8e4PWvqckfbmTOU6b0oLkNn7A-vzLuG9--j8,21544
256
253
  ipex_llm/vllm/xpu/entrypoints/openai/cli_args.py,sha256=2rENA2ucynMaIjiZBEh2ez1o5vR32GaP514t39CD7KM,8676
257
- ipex_llm-2.2.0b20250106.data/scripts/ipex-llm-init,sha256=fLQsT2dRL6H5bThb4GuIWotAuqoLsIxFwA-0c2qmaO8,6672
258
- ipex_llm-2.2.0b20250106.data/scripts/llm-chat,sha256=TdUnUmNapzuoe1c8IzrdVOQwWEg8IqsMSBRlOD3daZM,2249
259
- ipex_llm-2.2.0b20250106.data/scripts/llm-cli,sha256=RXGPlLElHxcKzoUxljEMBIAXbzCDysXL-Nxw-xF-7LU,2457
260
- ipex_llm-2.2.0b20250106.dist-info/METADATA,sha256=RVDr0pwoPE6J0yPUZ9k7t6_jQn01wTwAXkU5ViqE-c8,11374
261
- ipex_llm-2.2.0b20250106.dist-info/WHEEL,sha256=PPJcBMAZibF_2GFE9NmOJGqiaSMPiNFbJd6QaJjdA6Y,109
262
- ipex_llm-2.2.0b20250106.dist-info/entry_points.txt,sha256=TiUyBB2MRmfF3ko-pyAEzqeBCRnyhu27bNOAsWPp3e8,61
263
- ipex_llm-2.2.0b20250106.dist-info/top_level.txt,sha256=CGCMHM-SyqUabU4h8RqJ2KTYckQUO3LvIWwmUQ6Qbzw,9
264
- ipex_llm-2.2.0b20250106.dist-info/RECORD,,
254
+ ipex_llm-2.2.0b20250107.data/scripts/ipex-llm-init,sha256=fLQsT2dRL6H5bThb4GuIWotAuqoLsIxFwA-0c2qmaO8,6672
255
+ ipex_llm-2.2.0b20250107.data/scripts/llm-chat,sha256=TdUnUmNapzuoe1c8IzrdVOQwWEg8IqsMSBRlOD3daZM,2249
256
+ ipex_llm-2.2.0b20250107.data/scripts/llm-cli,sha256=RXGPlLElHxcKzoUxljEMBIAXbzCDysXL-Nxw-xF-7LU,2457
257
+ ipex_llm-2.2.0b20250107.dist-info/METADATA,sha256=rPJCuVvUndZ0XZBTZzlQEPi1y_W0fpmQTEGmogyRzRw,12705
258
+ ipex_llm-2.2.0b20250107.dist-info/WHEEL,sha256=PPJcBMAZibF_2GFE9NmOJGqiaSMPiNFbJd6QaJjdA6Y,109
259
+ ipex_llm-2.2.0b20250107.dist-info/entry_points.txt,sha256=TiUyBB2MRmfF3ko-pyAEzqeBCRnyhu27bNOAsWPp3e8,61
260
+ ipex_llm-2.2.0b20250107.dist-info/top_level.txt,sha256=CGCMHM-SyqUabU4h8RqJ2KTYckQUO3LvIWwmUQ6Qbzw,9
261
+ ipex_llm-2.2.0b20250107.dist-info/RECORD,,