ipex-llm 2.2.0b20250406__py3-none-win_amd64.whl → 2.3.0b20250410__py3-none-win_amd64.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (39) hide show
  1. ipex_llm/libs/bloom-api.dll +0 -0
  2. ipex_llm/libs/bloom.dll +0 -0
  3. ipex_llm/libs/gptneox-api.dll +0 -0
  4. ipex_llm/libs/gptneox.dll +0 -0
  5. ipex_llm/libs/libbloom_avx.dll +0 -0
  6. ipex_llm/libs/libbloom_vnni.dll +0 -0
  7. ipex_llm/libs/libgptneox_avx.dll +0 -0
  8. ipex_llm/libs/libgptneox_vnni.dll +0 -0
  9. ipex_llm/libs/libllama_avx.dll +0 -0
  10. ipex_llm/libs/libllama_vnni.dll +0 -0
  11. ipex_llm/libs/libstarcoder_avx.dll +0 -0
  12. ipex_llm/libs/libstarcoder_vnni.dll +0 -0
  13. ipex_llm/libs/llama-api.dll +0 -0
  14. ipex_llm/libs/llama.dll +0 -0
  15. ipex_llm/libs/main-bloom.exe +0 -0
  16. ipex_llm/libs/main-gptneox.exe +0 -0
  17. ipex_llm/libs/main-llama.exe +0 -0
  18. ipex_llm/libs/main-starcoder.exe +0 -0
  19. ipex_llm/libs/pipeline.dll +0 -0
  20. ipex_llm/libs/quantize-bloom.exe +0 -0
  21. ipex_llm/libs/quantize-bloom_vnni.exe +0 -0
  22. ipex_llm/libs/quantize-gptneox.exe +0 -0
  23. ipex_llm/libs/quantize-gptneox_vnni.exe +0 -0
  24. ipex_llm/libs/quantize-llama.exe +0 -0
  25. ipex_llm/libs/quantize-llama_vnni.exe +0 -0
  26. ipex_llm/libs/quantize-starcoder.exe +0 -0
  27. ipex_llm/libs/quantize-starcoder_vnni.exe +0 -0
  28. ipex_llm/libs/starcoder-api.dll +0 -0
  29. ipex_llm/libs/starcoder.dll +0 -0
  30. ipex_llm/transformers/convert.py +22 -3
  31. ipex_llm/transformers/models/qwen2_5_omni.py +160 -1
  32. {ipex_llm-2.2.0b20250406.dist-info → ipex_llm-2.3.0b20250410.dist-info}/METADATA +20 -20
  33. {ipex_llm-2.2.0b20250406.dist-info → ipex_llm-2.3.0b20250410.dist-info}/RECORD +39 -39
  34. {ipex_llm-2.2.0b20250406.data → ipex_llm-2.3.0b20250410.data}/scripts/ipex-llm-init.bat +0 -0
  35. {ipex_llm-2.2.0b20250406.data → ipex_llm-2.3.0b20250410.data}/scripts/llm-chat.ps1 +0 -0
  36. {ipex_llm-2.2.0b20250406.data → ipex_llm-2.3.0b20250410.data}/scripts/llm-cli.ps1 +0 -0
  37. {ipex_llm-2.2.0b20250406.dist-info → ipex_llm-2.3.0b20250410.dist-info}/WHEEL +0 -0
  38. {ipex_llm-2.2.0b20250406.dist-info → ipex_llm-2.3.0b20250410.dist-info}/entry_points.txt +0 -0
  39. {ipex_llm-2.2.0b20250406.dist-info → ipex_llm-2.3.0b20250410.dist-info}/top_level.txt +0 -0
Binary file
ipex_llm/libs/bloom.dll CHANGED
Binary file
Binary file
ipex_llm/libs/gptneox.dll CHANGED
Binary file
Binary file
Binary file
Binary file
Binary file
Binary file
Binary file
Binary file
Binary file
Binary file
ipex_llm/libs/llama.dll CHANGED
Binary file
Binary file
Binary file
Binary file
Binary file
Binary file
Binary file
Binary file
Binary file
Binary file
Binary file
Binary file
Binary file
Binary file
Binary file
Binary file
@@ -2072,12 +2072,31 @@ def _optimize_post(model):
2072
2072
  convert_forward(model.thinker.visual, module.Qwen2_5OmniVisionSdpaAttention,
2073
2073
  qwen2_5_omni_vision_attention_forward)
2074
2074
 
2075
+ # audio opt
2076
+ from ipex_llm.transformers.models.qwen2_5_omni import qwen2_5_omni_audio_attention_forward
2077
+ convert_forward(model.thinker.audio_tower, module.Qwen2_5OmniAudioAttention,
2078
+ qwen2_5_omni_audio_attention_forward)
2079
+ convert_forward(model.thinker.audio_tower, module.Qwen2_5OmniAudioSdpaAttention,
2080
+ qwen2_5_omni_audio_attention_forward)
2081
+
2075
2082
  # tts opt
2076
- if hasattr(model, "talker"):
2077
- convert_forward(model.talker, module.Qwen2_5OmniAttention,
2083
+ if model.has_talker:
2084
+ # talker part
2085
+ convert_forward(model.talker.model, module.Qwen2_5OmniAttention,
2086
+ qwen2_5_omni_attention_forward)
2087
+ convert_forward(model.talker.model, module.Qwen2_5OmniSdpaAttention,
2078
2088
  qwen2_5_omni_attention_forward)
2079
- convert_forward(model.talker, module.Qwen2_5OmniThinkerModel,
2089
+ convert_forward(model.talker.model, module.Qwen2_5OmniTalkerModel,
2080
2090
  qwen2_5_omni_thinker_model_forward)
2091
+ convert_forward(model.talker.model, module.Qwen2MLP, qwen2_mlp_forward)
2092
+
2093
+ # token2wav part
2094
+ from ipex_llm.transformers.models.qwen2_5_omni import dit_attention_forward
2095
+ from ipex_llm.transformers.models.qwen2_5_omni import _create_block_diff
2096
+ convert_forward(model.token2wav, module.DiTAttention, dit_attention_forward)
2097
+ dit_model = model.token2wav.code2wav_dit_model
2098
+ dit_model._create_block_diff = MethodType(_create_block_diff, dit_model)
2099
+
2081
2100
  return model
2082
2101
 
2083
2102
 
@@ -20,9 +20,11 @@
20
20
  import math
21
21
  import torch
22
22
  from typing import Optional, Tuple, List, Union
23
- from transformers.cache_utils import Cache
23
+ from transformers.cache_utils import Cache, EncoderDecoderCache
24
+ from transformers.modeling_utils import ALL_ATTENTION_FUNCTIONS
24
25
  from transformers.modeling_outputs import BaseModelOutputWithPast
25
26
  from transformers.models.qwen2_5_omni.modeling_qwen2_5_omni import Qwen2_5OmniAttention
27
+ from transformers.models.qwen2_5_omni.modeling_qwen2_5_omni import apply_rotary_pos_emb
26
28
  from transformers.models.qwen2_5_omni.modeling_qwen2_5_omni import apply_rotary_pos_emb_vision
27
29
  from transformers.models.qwen2_5_omni.modeling_qwen2_5_omni import apply_multimodal_rotary_pos_emb
28
30
 
@@ -284,3 +286,160 @@ def qwen2_5_omni_vision_attention_forward(
284
286
  attn_output = attn_output.reshape(seq_length, -1)
285
287
  attn_output = self.proj(attn_output)
286
288
  return attn_output
289
+
290
+
291
+ def qwen2_5_omni_audio_attention_forward(
292
+ self,
293
+ hidden_states: torch.Tensor,
294
+ key_value_states: Optional[torch.Tensor] = None,
295
+ past_key_value: Optional[EncoderDecoderCache] = None,
296
+ cu_seqlens: Optional[torch.Tensor] = None,
297
+ layer_head_mask: Optional[torch.Tensor] = None,
298
+ output_attentions: bool = False,
299
+ cache_position: Optional[torch.LongTensor] = None,
300
+ ) -> Tuple[torch.Tensor, Optional[torch.Tensor], Optional[Tuple[torch.Tensor]]]:
301
+ """Input shape: Batch x Time x Channel"""
302
+
303
+ # if key_value_states are provided this layer is used as a cross-attention layer
304
+ # for the decoder
305
+ is_cross_attention = key_value_states is not None
306
+ seq_length, _ = hidden_states.size()
307
+
308
+ # get query proj
309
+ query_states = self.q_proj(hidden_states)
310
+ query_states = query_states.reshape(seq_length, self.num_heads, -1)
311
+
312
+ seq_lens = cu_seqlens.tolist()
313
+ invalidInputError(seq_lens[0] == 0 and seq_lens[-1] == seq_length,
314
+ "unexpected input")
315
+
316
+ if past_key_value is not None:
317
+ is_updated = past_key_value.is_updated.get(self.layer_idx)
318
+ if is_cross_attention:
319
+ # after the first generated id,
320
+ # we can subsequently re-use all key/value_states from cache
321
+ past_key_value.is_updated[self.layer_idx] = True
322
+ past_key_value = past_key_value.cross_attention_cache
323
+ else:
324
+ past_key_value = past_key_value.self_attention_cache
325
+
326
+ # use key_value_states if cross attention
327
+ current_states = key_value_states if key_value_states is not None else hidden_states
328
+ if is_cross_attention and past_key_value and is_updated:
329
+ # reuse k,v, cross_attentions
330
+ key_states = past_key_value.key_cache[self.layer_idx]
331
+ value_states = past_key_value.value_cache[self.layer_idx]
332
+ else:
333
+ key_states = self.k_proj(current_states).reshape(seq_length, self.num_heads, -1)
334
+ value_states = self.v_proj(current_states).reshape(seq_length, self.num_heads, -1)
335
+ if past_key_value is not None:
336
+ # save all key/value_states to cache to be re-used for fast auto-regressive generation
337
+ cache_position = cache_position if not is_cross_attention else None
338
+ key_states, value_states = past_key_value.update(
339
+ key_states, value_states, self.layer_idx, {"cache_position": cache_position}
340
+ )
341
+
342
+ if layer_head_mask is None and use_sdp_non_causal(query_states.size(-1),
343
+ query_states.device, query_states.dtype):
344
+ kv_length = key_states.size(0)
345
+ padding_kv_length = (kv_length + 128 - 1) // 128 * 128
346
+ attention_mask = torch.full(
347
+ [1, 1, seq_length, padding_kv_length], torch.finfo(query_states.dtype).min,
348
+ device=query_states.device, dtype=query_states.dtype,
349
+ )
350
+ for i in range(1, len(cu_seqlens)):
351
+ attention_mask[..., seq_lens[i - 1]:seq_lens[i], seq_lens[i - 1]:seq_lens[i]] = 0
352
+
353
+ q = query_states.transpose(0, 1).unsqueeze(0)
354
+ k = key_states.transpose(0, 1).unsqueeze(0).contiguous()
355
+ v = value_states.transpose(0, 1).unsqueeze(0).contiguous()
356
+ # q, k, v: [1, num_heads, seq_length, head_dim]
357
+
358
+ attn_weights = None
359
+ attn_output = scaled_dot_product_attention(q, k, v, attention_mask, False)
360
+ attn_output = attn_output.permute(0, 2, 1, 3).squeeze(0)
361
+ # attn_output: [seq_length, num_heads, head_dim]
362
+ else:
363
+ attention_mask = torch.full(
364
+ [1, seq_length, key_states.size(0)], torch.finfo(query_states.dtype).min,
365
+ device=query_states.device, dtype=query_states.dtype,
366
+ )
367
+ for i in range(1, len(cu_seqlens)):
368
+ attention_mask[..., seq_lens[i - 1]:seq_lens[i], seq_lens[i - 1]:seq_lens[i]] = 0
369
+
370
+ query_states = query_states.transpose(0, 1)
371
+ key_states = key_states.transpose(0, 1)
372
+ value_states = value_states.transpose(0, 1)
373
+
374
+ attn_weights = torch.matmul(query_states,
375
+ key_states.transpose(1, 2)) / math.sqrt(self.head_dim)
376
+ attn_weights = attn_weights + attention_mask
377
+ attn_weights = attention_softmax(attn_weights)
378
+
379
+ if layer_head_mask is not None:
380
+ attn_weights = layer_head_mask.view(1, -1, 1, 1) * attn_weights
381
+
382
+ attn_output = torch.matmul(attn_weights, value_states).transpose(0, 1)
383
+
384
+ # Use the `embed_dim` from the config (stored in the class) rather than `hidden_state`s
385
+ # because `attn_output` can be partitioned across GPUs when using tensor-parallelism.
386
+ attn_output = attn_output.reshape(seq_length, self.embed_dim)
387
+ attn_output = self.out_proj(attn_output)
388
+
389
+ return attn_output, attn_weights, past_key_value
390
+
391
+
392
+ def dit_attention_forward(
393
+ self,
394
+ x,
395
+ rope=None,
396
+ mask=None,
397
+ ) -> torch.Tensor:
398
+ batch_size = x.shape[0]
399
+
400
+ # `sample` projections.
401
+ query = self.to_q(x)
402
+ key = self.to_k(x)
403
+ value = self.to_v(x)
404
+
405
+ # attention
406
+ inner_dim = key.shape[-1]
407
+ head_dim = inner_dim // self.heads
408
+ query = query.view(batch_size, -1, self.heads, head_dim).transpose(1, 2)
409
+ key = key.view(batch_size, -1, self.heads, head_dim).transpose(1, 2)
410
+ value = value.view(batch_size, -1, self.heads, head_dim).transpose(1, 2)
411
+
412
+ # apply rotary position embedding
413
+ # Due to training process, only first head is applied with RoPE, will be fixed at next release
414
+ cos, sin = rope
415
+ query[:, :1], key[:, :1] = apply_rotary_pos_emb(query[:, :1], key[:, :1], cos, sin)
416
+
417
+ if use_sdp_non_causal(head_dim, query.device, query.dtype):
418
+ mask = torch.where(mask, 0, torch.finfo(query.dtype).min)
419
+ x = scaled_dot_product_attention(query, key.contiguous(), value.contiguous(), mask, False)
420
+ x = x.transpose(1, 2)
421
+ else:
422
+ attention_interface = ALL_ATTENTION_FUNCTIONS[self._attn_implementation]
423
+ x, _ = attention_interface(self, query, key, value, attention_mask=mask, is_causal=False)
424
+
425
+ # mask
426
+ x = x.reshape(batch_size, -1, self.heads * head_dim)
427
+ x = x.to(query.dtype)
428
+
429
+ # linear proj
430
+ x = self.to_out[0](x)
431
+ # dropout
432
+ x = self.to_out[1](x)
433
+
434
+ return x
435
+
436
+
437
+ def _create_block_diff(self, x):
438
+ batch, seq_len = x.shape[0], x.shape[1]
439
+ block_indices = torch.arange(seq_len, device=x.device) // self.block_size
440
+
441
+ block_i = block_indices.unsqueeze(1) # [seq_length, 1]
442
+ block_j = block_indices.unsqueeze(0) # [1, seq_length]
443
+
444
+ block_diff = block_j - block_i # (n, n)
445
+ return block_diff.unsqueeze(0).unsqueeze(0)
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: ipex-llm
3
- Version: 2.2.0b20250406
3
+ Version: 2.3.0b20250410
4
4
  Summary: Large Language Model Develop Toolkit
5
5
  Home-page: https://github.com/intel-analytics/ipex-llm
6
6
  Author: BigDL Authors
@@ -27,7 +27,7 @@ Requires-Dist: intel-openmp ; (platform_machine == "x86_64" or platform_machine
27
27
  Requires-Dist: torch ==2.1.2+cpu ; (platform_system == "Linux") and extra == 'all'
28
28
  Requires-Dist: torch ==2.1.2 ; (platform_system == "Windows") and extra == 'all'
29
29
  Provides-Extra: cpp
30
- Requires-Dist: bigdl-core-cpp ==2.6.0b20250406 ; extra == 'cpp'
30
+ Requires-Dist: bigdl-core-cpp ==2.7.0b20250410 ; extra == 'cpp'
31
31
  Requires-Dist: setuptools ; extra == 'cpp'
32
32
  Requires-Dist: onednn-devel ==2025.0.1 ; (platform_system == "Windows") and extra == 'cpp'
33
33
  Requires-Dist: onednn ==2025.0.1 ; (platform_system == "Windows") and extra == 'cpp'
@@ -60,7 +60,7 @@ Requires-Dist: transformers ==4.40.0 ; extra == 'npu'
60
60
  Requires-Dist: intel-openmp ; (platform_machine == "x86_64" or platform_machine == "AMD64") and extra == 'npu'
61
61
  Requires-Dist: torch ==2.1.2+cpu ; (platform_system == "Linux") and extra == 'npu'
62
62
  Requires-Dist: torch ==2.1.2 ; (platform_system == "Windows") and extra == 'npu'
63
- Requires-Dist: bigdl-core-npu ==2.6.0b20250406 ; (platform_system == "Windows") and extra == 'npu'
63
+ Requires-Dist: bigdl-core-npu ==2.7.0b20250410 ; (platform_system == "Windows") and extra == 'npu'
64
64
  Provides-Extra: serving
65
65
  Requires-Dist: py-cpuinfo ; extra == 'serving'
66
66
  Requires-Dist: fschat[model_worker,webui] ==0.2.36 ; extra == 'serving'
@@ -80,9 +80,9 @@ Requires-Dist: setuptools <70.0.0 ; extra == 'xpu'
80
80
  Requires-Dist: torch ==2.1.0a0 ; extra == 'xpu'
81
81
  Requires-Dist: torchvision ==0.16.0a0 ; extra == 'xpu'
82
82
  Requires-Dist: intel-extension-for-pytorch ==2.1.10+xpu ; extra == 'xpu'
83
- Requires-Dist: bigdl-core-xe-21 ==2.6.0b20250406 ; extra == 'xpu'
84
- Requires-Dist: bigdl-core-xe-batch-21 ==2.6.0b20250406 ; extra == 'xpu'
85
- Requires-Dist: bigdl-core-xe-addons-21 ==2.6.0b20250406 ; extra == 'xpu'
83
+ Requires-Dist: bigdl-core-xe-21 ==2.7.0b20250410 ; extra == 'xpu'
84
+ Requires-Dist: bigdl-core-xe-batch-21 ==2.7.0b20250410 ; extra == 'xpu'
85
+ Requires-Dist: bigdl-core-xe-addons-21 ==2.7.0b20250410 ; extra == 'xpu'
86
86
  Provides-Extra: xpu-2-1
87
87
  Requires-Dist: py-cpuinfo ; extra == 'xpu-2-1'
88
88
  Requires-Dist: protobuf ; extra == 'xpu-2-1'
@@ -97,9 +97,9 @@ Requires-Dist: setuptools <70.0.0 ; extra == 'xpu-2-1'
97
97
  Requires-Dist: torch ==2.1.0a0 ; extra == 'xpu-2-1'
98
98
  Requires-Dist: torchvision ==0.16.0a0 ; extra == 'xpu-2-1'
99
99
  Requires-Dist: intel-extension-for-pytorch ==2.1.10+xpu ; extra == 'xpu-2-1'
100
- Requires-Dist: bigdl-core-xe-21 ==2.6.0b20250406 ; extra == 'xpu-2-1'
101
- Requires-Dist: bigdl-core-xe-batch-21 ==2.6.0b20250406 ; extra == 'xpu-2-1'
102
- Requires-Dist: bigdl-core-xe-addons-21 ==2.6.0b20250406 ; extra == 'xpu-2-1'
100
+ Requires-Dist: bigdl-core-xe-21 ==2.7.0b20250410 ; extra == 'xpu-2-1'
101
+ Requires-Dist: bigdl-core-xe-batch-21 ==2.7.0b20250410 ; extra == 'xpu-2-1'
102
+ Requires-Dist: bigdl-core-xe-addons-21 ==2.7.0b20250410 ; extra == 'xpu-2-1'
103
103
  Requires-Dist: intel-openmp ; (platform_machine == "x86_64" or platform_machine == "AMD64") and extra == 'xpu-2-1'
104
104
  Requires-Dist: dpcpp-cpp-rt ==2024.0.2 ; (platform_system == "Windows") and extra == 'xpu-2-1'
105
105
  Requires-Dist: mkl-dpcpp ==2024.0.0 ; (platform_system == "Windows") and extra == 'xpu-2-1'
@@ -117,7 +117,7 @@ Requires-Dist: setuptools ; extra == 'xpu-2-6'
117
117
  Requires-Dist: torch ==2.6.0+xpu ; extra == 'xpu-2-6'
118
118
  Requires-Dist: torchvision ==0.21.0+xpu ; extra == 'xpu-2-6'
119
119
  Requires-Dist: torchaudio ==2.6.0+xpu ; extra == 'xpu-2-6'
120
- Requires-Dist: bigdl-core-xe-all ==2.6.0b20250406 ; extra == 'xpu-2-6'
120
+ Requires-Dist: bigdl-core-xe-all ==2.7.0b20250410 ; extra == 'xpu-2-6'
121
121
  Requires-Dist: onednn-devel ==2025.0.1 ; extra == 'xpu-2-6'
122
122
  Requires-Dist: onednn ==2025.0.1 ; extra == 'xpu-2-6'
123
123
  Requires-Dist: dpcpp-cpp-rt ==2025.0.2 ; extra == 'xpu-2-6'
@@ -132,7 +132,7 @@ Requires-Dist: tokenizers ==0.15.2 ; extra == 'xpu-2-6-arl'
132
132
  Requires-Dist: accelerate ==0.23.0 ; extra == 'xpu-2-6-arl'
133
133
  Requires-Dist: tabulate ; extra == 'xpu-2-6-arl'
134
134
  Requires-Dist: setuptools ; extra == 'xpu-2-6-arl'
135
- Requires-Dist: bigdl-core-xe-all ==2.6.0b20250406 ; extra == 'xpu-2-6-arl'
135
+ Requires-Dist: bigdl-core-xe-all ==2.7.0b20250410 ; extra == 'xpu-2-6-arl'
136
136
  Requires-Dist: onednn-devel ==2025.0.1 ; extra == 'xpu-2-6-arl'
137
137
  Requires-Dist: onednn ==2025.0.1 ; extra == 'xpu-2-6-arl'
138
138
  Requires-Dist: dpcpp-cpp-rt ==2025.0.2 ; extra == 'xpu-2-6-arl'
@@ -155,9 +155,9 @@ Requires-Dist: tokenizers ==0.15.2 ; extra == 'xpu-arc'
155
155
  Requires-Dist: accelerate ==0.23.0 ; extra == 'xpu-arc'
156
156
  Requires-Dist: tabulate ; extra == 'xpu-arc'
157
157
  Requires-Dist: setuptools ; extra == 'xpu-arc'
158
- Requires-Dist: bigdl-core-xe-23 ==2.6.0b20250406 ; extra == 'xpu-arc'
159
- Requires-Dist: bigdl-core-xe-batch-23 ==2.6.0b20250406 ; extra == 'xpu-arc'
160
- Requires-Dist: bigdl-core-xe-addons-23 ==2.6.0b20250406 ; extra == 'xpu-arc'
158
+ Requires-Dist: bigdl-core-xe-23 ==2.7.0b20250410 ; extra == 'xpu-arc'
159
+ Requires-Dist: bigdl-core-xe-batch-23 ==2.7.0b20250410 ; extra == 'xpu-arc'
160
+ Requires-Dist: bigdl-core-xe-addons-23 ==2.7.0b20250410 ; extra == 'xpu-arc'
161
161
  Requires-Dist: intel-openmp ; (platform_machine == "x86_64" or platform_machine == "AMD64") and extra == 'xpu-arc'
162
162
  Requires-Dist: torch ==2.3.1+cxx11.abi ; (platform_system == "Linux") and extra == 'xpu-arc'
163
163
  Requires-Dist: torchvision ==0.18.1+cxx11.abi ; (platform_system == "Linux") and extra == 'xpu-arc'
@@ -178,9 +178,9 @@ Requires-Dist: tokenizers ==0.15.2 ; extra == 'xpu-arl'
178
178
  Requires-Dist: accelerate ==0.23.0 ; extra == 'xpu-arl'
179
179
  Requires-Dist: tabulate ; extra == 'xpu-arl'
180
180
  Requires-Dist: setuptools ; extra == 'xpu-arl'
181
- Requires-Dist: bigdl-core-xe-23 ==2.6.0b20250406 ; extra == 'xpu-arl'
182
- Requires-Dist: bigdl-core-xe-batch-23 ==2.6.0b20250406 ; extra == 'xpu-arl'
183
- Requires-Dist: bigdl-core-xe-addons-23 ==2.6.0b20250406 ; extra == 'xpu-arl'
181
+ Requires-Dist: bigdl-core-xe-23 ==2.7.0b20250410 ; extra == 'xpu-arl'
182
+ Requires-Dist: bigdl-core-xe-batch-23 ==2.7.0b20250410 ; extra == 'xpu-arl'
183
+ Requires-Dist: bigdl-core-xe-addons-23 ==2.7.0b20250410 ; extra == 'xpu-arl'
184
184
  Requires-Dist: intel-openmp ; (platform_machine == "x86_64" or platform_machine == "AMD64") and extra == 'xpu-arl'
185
185
  Requires-Dist: torch ==2.3.1+cxx11.abi ; (platform_system == "Linux") and extra == 'xpu-arl'
186
186
  Requires-Dist: torchvision ==0.18.1+cxx11.abi ; (platform_system == "Linux") and extra == 'xpu-arl'
@@ -201,9 +201,9 @@ Requires-Dist: tokenizers ==0.15.2 ; extra == 'xpu-lnl'
201
201
  Requires-Dist: accelerate ==0.23.0 ; extra == 'xpu-lnl'
202
202
  Requires-Dist: tabulate ; extra == 'xpu-lnl'
203
203
  Requires-Dist: setuptools ; extra == 'xpu-lnl'
204
- Requires-Dist: bigdl-core-xe-23 ==2.6.0b20250406 ; extra == 'xpu-lnl'
205
- Requires-Dist: bigdl-core-xe-batch-23 ==2.6.0b20250406 ; extra == 'xpu-lnl'
206
- Requires-Dist: bigdl-core-xe-addons-23 ==2.6.0b20250406 ; extra == 'xpu-lnl'
204
+ Requires-Dist: bigdl-core-xe-23 ==2.7.0b20250410 ; extra == 'xpu-lnl'
205
+ Requires-Dist: bigdl-core-xe-batch-23 ==2.7.0b20250410 ; extra == 'xpu-lnl'
206
+ Requires-Dist: bigdl-core-xe-addons-23 ==2.7.0b20250410 ; extra == 'xpu-lnl'
207
207
  Requires-Dist: intel-openmp ; (platform_machine == "x86_64" or platform_machine == "AMD64") and extra == 'xpu-lnl'
208
208
  Requires-Dist: torch ==2.3.1+cxx11.abi ; (platform_system == "Linux") and extra == 'xpu-lnl'
209
209
  Requires-Dist: torchvision ==0.18.1+cxx11.abi ; (platform_system == "Linux") and extra == 'xpu-lnl'
@@ -41,35 +41,35 @@ ipex_llm/langchain/llms/transformerspipelinellm.py,sha256=vm522YPPwWxxAPVvQBtxRf
41
41
  ipex_llm/langchain/vllm/__init__.py,sha256=T-EbRT6GJ_8RCu-iLmSzcftOimXSPQf2d5X72AUAy2Y,874
42
42
  ipex_llm/langchain/vllm/vllm.py,sha256=6dxc-ZISZQrJilEa_HA827l75Dv9rcHpY_G6FdJ8BVs,7793
43
43
  ipex_llm/libs/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
44
- ipex_llm/libs/bloom-api.dll,sha256=Ludwn52Jx_t1WdrVXtAL5xpG5KRgwEOfRbBzjbwM2os,36352
45
- ipex_llm/libs/bloom.dll,sha256=YBm6l9Xot59G9SIC54u1LqWaTKk2zOhFlgZKLgrACjE,507904
46
- ipex_llm/libs/gptneox-api.dll,sha256=f0SMuljGde616lo8tI7aMqBwAd7LC6WCn3_9B8xmhKU,24576
47
- ipex_llm/libs/gptneox.dll,sha256=qHxRNBGan7wsNLMg0ObOGVm14tZ70S_nF6nRbqXMr-0,568320
48
- ipex_llm/libs/libbloom_avx.dll,sha256=TVnHuMQfSMU7xE082TeGPAeZm2DX-UGQhE5fYqqvl_Q,536576
49
- ipex_llm/libs/libbloom_vnni.dll,sha256=eqXJjMNMG8g_qvT91LOVo8ZmnrMKOVteesGNvECEGvs,508416
50
- ipex_llm/libs/libgptneox_avx.dll,sha256=ptBspg6y__MixWsO8XaLarPuUQsSrYGWoA4f8PLSSUI,596992
51
- ipex_llm/libs/libgptneox_vnni.dll,sha256=MTXjPHO96eUYMp57RdmeawyLYnMZjvTLSGMUiqo4dLI,568832
52
- ipex_llm/libs/libllama_avx.dll,sha256=xDtaTAbSxXaimO8kbpPsf1nfAPKc7VhV-z3oaRzjH2g,591360
53
- ipex_llm/libs/libllama_vnni.dll,sha256=60fJNPU9-nwSaFHXia5afwPbZoimNAcfrLE6XqDa8V4,563200
54
- ipex_llm/libs/libstarcoder_avx.dll,sha256=LD3kkWiwogUv726z7mY0SiBic9o7qjZ5yauDNn8Q5Yg,627712
55
- ipex_llm/libs/libstarcoder_vnni.dll,sha256=yX4TjzQzxFM3ncsz8W6J73jD4bhgMjIjw_4UmRJGGiM,599552
56
- ipex_llm/libs/llama-api.dll,sha256=zhlUaEmt1LqsZVIXERnR3z-K0nNbWYr6KC-G-jfprKI,25600
57
- ipex_llm/libs/llama.dll,sha256=9d0I61ZvgBxOJONYwalztS1FFTTgAfMtJXovUa6_p8Y,562688
58
- ipex_llm/libs/main-bloom.exe,sha256=r_if-mP36-0XivRHY6_yh9zTl-rC99AHk0wtv4g0I-A,103424
59
- ipex_llm/libs/main-gptneox.exe,sha256=2e8UYQaptyDb_gZgD1Z-wZ8Tvdkbd_uoO13f5r0NxWo,98816
60
- ipex_llm/libs/main-llama.exe,sha256=Ch4cIo66XOg5Ok6Dcisre5wbOVY_Kb4bCPcQPMJI7q4,99840
61
- ipex_llm/libs/main-starcoder.exe,sha256=OVCzIte_WVZnViLBs7WideS2SVVkL_9HTuEEKOi3yvc,157696
62
- ipex_llm/libs/pipeline.dll,sha256=VevD7YeLPT8TKPpeYd3a77sVN6x_t6k9AJKRmNCMcWM,73216
63
- ipex_llm/libs/quantize-bloom.exe,sha256=Y9h7CRzCfLjJHQ87w3eIMFdI815irBVIe--p3D708z8,126464
64
- ipex_llm/libs/quantize-bloom_vnni.exe,sha256=bSTqkSxuwBrPLR_eF8c_7LBY6V5Xvypvvel-rzqCZYs,128000
65
- ipex_llm/libs/quantize-gptneox.exe,sha256=OynK1Y9AqJT4SL_5sEjJiCTSnio3fX2YW8ZCYjcJ3Dk,104448
66
- ipex_llm/libs/quantize-gptneox_vnni.exe,sha256=xqRrEozA3-XKv5nS7S0A6xRK4UNI3TncSI0IJIdRr6I,104960
67
- ipex_llm/libs/quantize-llama.exe,sha256=vW2YhvHEGXelf8Xt7p1igiHI3tezHooq-4yYW9dJSf8,110080
68
- ipex_llm/libs/quantize-llama_vnni.exe,sha256=B871TelJcTtrPVbCZLg745WnuJdcFXUo1bHURK7aN5g,110592
69
- ipex_llm/libs/quantize-starcoder.exe,sha256=TKqctPS53ycP3ag0jh_rP3ehVRtIxAW8MNhhOxLQFmU,127488
70
- ipex_llm/libs/quantize-starcoder_vnni.exe,sha256=omVd1xPypzpgSEM2A7_sw5tHhstwiUTVxvQGA2lz2ko,128512
71
- ipex_llm/libs/starcoder-api.dll,sha256=qqq9Pjn-hv6GYM9i0WBXxXVFwn0-vqpLBP_5lGwP4-U,21504
72
- ipex_llm/libs/starcoder.dll,sha256=D7Ennr2_foo1g5QwTfePV9Yl_QCm869FY5VzBkrR0CE,599040
44
+ ipex_llm/libs/bloom-api.dll,sha256=WfAd-vNHNBYmooU8EnEoQA2JkWpyBuIejPaTfyY_SZk,36352
45
+ ipex_llm/libs/bloom.dll,sha256=9fM_JsT4LUA9XM26LvJ9ZiZdZw0GY_UftjdVsPbC-7M,507904
46
+ ipex_llm/libs/gptneox-api.dll,sha256=ngWpc5XL-yQYfBsxN0QtqgGiWjoZUG-yUW7S_BCNpb4,24576
47
+ ipex_llm/libs/gptneox.dll,sha256=yy1fkaCLDdIEIkAA0pfq8kbq81oMEk3Mm-Rkl0VX2yg,568320
48
+ ipex_llm/libs/libbloom_avx.dll,sha256=Q7kv0z-0GyCEp5nj5Ern_NxFEWl7eDV6wmRgauaC5GA,536576
49
+ ipex_llm/libs/libbloom_vnni.dll,sha256=cgx0LDge7gATzfF7NkIL4yLXFRtFm_DT3r72hge3V40,508416
50
+ ipex_llm/libs/libgptneox_avx.dll,sha256=IBl-RDmNw0Qs2RHJzjX7jofILYQKL9IAWV-suC8m62A,596992
51
+ ipex_llm/libs/libgptneox_vnni.dll,sha256=CvcoD4UfSIhOTO1WG4EFN6Vkv2XdZq5J5f_pYh4Nt7o,568832
52
+ ipex_llm/libs/libllama_avx.dll,sha256=xjko1uyWQFGOJ7h_VEMjTrcsCNeZUUrunoFX5YnTfzM,591360
53
+ ipex_llm/libs/libllama_vnni.dll,sha256=RcvAYB4CPffb-YlS8VMfliFGhL4CrIVvf-4bZb-r_2M,563200
54
+ ipex_llm/libs/libstarcoder_avx.dll,sha256=OKtbH_1hJZ4jlkl3gZLapf7clikeC3rVV8RuZ_hWLhQ,627712
55
+ ipex_llm/libs/libstarcoder_vnni.dll,sha256=kSAIo8O3YtBBgUwfXCeadECJK9nAxs1YvRa8SfIxE-Q,599552
56
+ ipex_llm/libs/llama-api.dll,sha256=vxKuiZix6fx5ydDeWOdT1sPb1wV0ouIXuD7elzk8kpc,25600
57
+ ipex_llm/libs/llama.dll,sha256=wtbG9_50kOTFAxef8Tb8P8FIzZ5JYqkmig9A9JxVp0A,562688
58
+ ipex_llm/libs/main-bloom.exe,sha256=hvo6htyCEyZ8xZ2LOt90xbx43v5cDQtt7YrfskzncSU,103424
59
+ ipex_llm/libs/main-gptneox.exe,sha256=qvp8UlpkBU09W29bUJkckfpN5hFjmN4HGL2i7G7KhpE,98816
60
+ ipex_llm/libs/main-llama.exe,sha256=yB2z4ddSKWZnECsn5jm70gMyOq4t-QK1QRT7SlM8d_E,99840
61
+ ipex_llm/libs/main-starcoder.exe,sha256=wGcRZBMqWTK64kmR36f94Jq1R1ElF8JH-xYl4UkxUxg,157696
62
+ ipex_llm/libs/pipeline.dll,sha256=3x6sBRn72IP6CyhFdsq9dyFDJra8WRGZs0dstl5SYHo,73216
63
+ ipex_llm/libs/quantize-bloom.exe,sha256=2au0XxXiYkWmg-z3RtqH958_2hAsMDsj_HP69DuykGI,126464
64
+ ipex_llm/libs/quantize-bloom_vnni.exe,sha256=Pygds4P-hMVyc9vqCNHIKrRcETNwDlIU-sHGiKwDUBU,128000
65
+ ipex_llm/libs/quantize-gptneox.exe,sha256=2F8d_iMGeWxtmoN5WpMnf2qhosPqm1aZd4FFlunqb7k,104448
66
+ ipex_llm/libs/quantize-gptneox_vnni.exe,sha256=u-_BVBPRj5edcTzjQj_jd_dJ3ANDER73I8pnA_fI0v8,104960
67
+ ipex_llm/libs/quantize-llama.exe,sha256=MdSzxiyjFEgNooqgUM8DMVNniiDRwnzuQ63IEu0b4As,110080
68
+ ipex_llm/libs/quantize-llama_vnni.exe,sha256=fMg9YKVbJoB2GgdZLjpuPkOvecqF2joR2JovboQLzhg,110592
69
+ ipex_llm/libs/quantize-starcoder.exe,sha256=cgHF7r8oMpxIDCoKYQfcVRQ5RLDA2n9ttqklvsfBTj0,127488
70
+ ipex_llm/libs/quantize-starcoder_vnni.exe,sha256=BywuKBiNyJovLQ8uQxvmRgK5Mvr-VZMXh6ZkwIGXjDQ,128512
71
+ ipex_llm/libs/starcoder-api.dll,sha256=tYqccLQiJkRgMST2UlkhYLGJQOYPott3mnwGZZxkuGU,21504
72
+ ipex_llm/libs/starcoder.dll,sha256=Xt6gimIDqhNGkNnKkeOGiEgXU9kAJKrnMlJlbppzHrY,599040
73
73
  ipex_llm/llamaindex/__init__.py,sha256=T-EbRT6GJ_8RCu-iLmSzcftOimXSPQf2d5X72AUAy2Y,874
74
74
  ipex_llm/llamaindex/llms/__init__.py,sha256=KP1lEdGqDuxPoxL1ZSH25Pm2kKMPJBWUTLR0ckSLMIU,1139
75
75
  ipex_llm/llamaindex/llms/bigdlllm.py,sha256=FQBzq1KOjfc6uofTXAha3O7TqpJkNfOFepXQmOVlbnI,26314
@@ -87,7 +87,7 @@ ipex_llm/serving/fastchat/tgi_api_protocol.py,sha256=brT3k3-V0NJrU4fRqUwWjC0O3iO
87
87
  ipex_llm/serving/fastchat/tgi_api_server.py,sha256=agNTAEiZPSuj3dEdIdYKwkoY0cXOUDX06DiM9VP2knQ,24418
88
88
  ipex_llm/serving/fastchat/vllm_worker.py,sha256=ZLz2Q9GxJO6r_LOiP6epgCRjBGk-K4EB1SNEWSJp5DA,11091
89
89
  ipex_llm/transformers/__init__.py,sha256=BreA3EY6hkNq0rVixb_sUuTLzMrcWXTt3yRsshCPHHQ,1214
90
- ipex_llm/transformers/convert.py,sha256=D3QSoD48NzOKz9rKIsgrI0sNzdmJvb0sP4k_0f8Lcmo,105824
90
+ ipex_llm/transformers/convert.py,sha256=xqqZFGcdDRko2IYgfSgDRs8ef4THUR25IAhSyDV0VUs,106933
91
91
  ipex_llm/transformers/convert_ipex.py,sha256=_nSnUTQy-yfkKaqGdqnBdWztZf3NGmnbZ0TKaDrF4X4,14617
92
92
  ipex_llm/transformers/embedding.py,sha256=bdgk59DvD4ZZyxRzewXOR7g56nThgO6uhIwk8QL7f-s,9299
93
93
  ipex_llm/transformers/kv.py,sha256=k4TU18LlA-Sbq9WNNQnfuzu3RSFBwFhmaV3BcGN5bAo,19191
@@ -171,7 +171,7 @@ ipex_llm/transformers/models/phi3.py,sha256=AaWB7TPQdrDYgpcVHglG0Q0480bxNOw1mFeP
171
171
  ipex_llm/transformers/models/phixtral.py,sha256=MDTMghcu7qAmZmRcUGqXXDXhSU3y_N59HRIXmlcjp5g,4890
172
172
  ipex_llm/transformers/models/qwen.py,sha256=A3WiVCzA7NLkcjp4zhFkZvKZzZWZlg0WFuVV_556TAI,19543
173
173
  ipex_llm/transformers/models/qwen2.py,sha256=zK-FpUaxEhjD4gZa1ZvArodAilz29T_cpeAqfCGosc0,14317
174
- ipex_llm/transformers/models/qwen2_5_omni.py,sha256=uAm_dZBAf53nYt3d1bFitFcIWJV80wqU4q4NJRbwxIE,12015
174
+ ipex_llm/transformers/models/qwen2_5_omni.py,sha256=gNnWvGiPntn6GlyIUtCVYCSIMPU6FG9TCS7fhWbwerY,18779
175
175
  ipex_llm/transformers/models/qwen2_moe.py,sha256=a0gYo-ngf8SxaEnBdZUJDnPS6Mkn_poDd8xqhx50icI,19516
176
176
  ipex_llm/transformers/models/qwen2_vl.py,sha256=G-9e2oN4f5p5IWQ-zsBZuONxTura3BjlgyT2meigbHQ,13579
177
177
  ipex_llm/transformers/models/qwen_vl.py,sha256=j7Nzzz2Qvynu9yrCXmoEfERjw43hXof5TbXIs7Ms-oY,17105
@@ -256,11 +256,11 @@ ipex_llm/vllm/xpu/engine/__init__.py,sha256=pY_CpyuZd72fr6s32ejeKHKFW0K4vUU2rzZj
256
256
  ipex_llm/vllm/xpu/engine/engine.py,sha256=NvCMbp0X8NVrOqbwm4FTvXOptTRLzu9jQsy37ZHnTk8,9493
257
257
  ipex_llm/vllm/xpu/entrypoints/openai/api_server.py,sha256=IjiSze9vzBCAkLu_VwIcJwuO1jyFna7DLrj6aSL7RaQ,35220
258
258
  ipex_llm/vllm/xpu/entrypoints/openai/cli_args.py,sha256=hB398yYtKauASRzevctScdbFIjiiSGMAe1bwEuIHrhY,10893
259
- ipex_llm-2.2.0b20250406.data/scripts/ipex-llm-init.bat,sha256=HPtCYuDYwEatq7dAwOvdfVcHYCpAVdbj75K1qh0vQek,2578
260
- ipex_llm-2.2.0b20250406.data/scripts/llm-chat.ps1,sha256=6qrs-hGVAV8IKh7Jx8nq_XrnZcjd7qGU5wndArM7Yag,2769
261
- ipex_llm-2.2.0b20250406.data/scripts/llm-cli.ps1,sha256=3qBtTLs_EjYDnM8YyCpJhzLnGCKTEGssu9UNqfkjVXs,3009
262
- ipex_llm-2.2.0b20250406.dist-info/METADATA,sha256=F4HquWjK1bonrhub9LdMVh-XXx4TjrgaUiyAMIPBdoo,13917
263
- ipex_llm-2.2.0b20250406.dist-info/WHEEL,sha256=6iYPr8vTHsyDK75jr9X0V3I9wPSVmtwr_8fdATBciGk,98
264
- ipex_llm-2.2.0b20250406.dist-info/entry_points.txt,sha256=TiUyBB2MRmfF3ko-pyAEzqeBCRnyhu27bNOAsWPp3e8,61
265
- ipex_llm-2.2.0b20250406.dist-info/top_level.txt,sha256=CGCMHM-SyqUabU4h8RqJ2KTYckQUO3LvIWwmUQ6Qbzw,9
266
- ipex_llm-2.2.0b20250406.dist-info/RECORD,,
259
+ ipex_llm-2.3.0b20250410.data/scripts/ipex-llm-init.bat,sha256=HPtCYuDYwEatq7dAwOvdfVcHYCpAVdbj75K1qh0vQek,2578
260
+ ipex_llm-2.3.0b20250410.data/scripts/llm-chat.ps1,sha256=6qrs-hGVAV8IKh7Jx8nq_XrnZcjd7qGU5wndArM7Yag,2769
261
+ ipex_llm-2.3.0b20250410.data/scripts/llm-cli.ps1,sha256=3qBtTLs_EjYDnM8YyCpJhzLnGCKTEGssu9UNqfkjVXs,3009
262
+ ipex_llm-2.3.0b20250410.dist-info/METADATA,sha256=Tr1wuth1NPT_yf_bvKcaIS2ZlYLd0mR_A2uZ7m3tJb0,13917
263
+ ipex_llm-2.3.0b20250410.dist-info/WHEEL,sha256=6iYPr8vTHsyDK75jr9X0V3I9wPSVmtwr_8fdATBciGk,98
264
+ ipex_llm-2.3.0b20250410.dist-info/entry_points.txt,sha256=TiUyBB2MRmfF3ko-pyAEzqeBCRnyhu27bNOAsWPp3e8,61
265
+ ipex_llm-2.3.0b20250410.dist-info/top_level.txt,sha256=CGCMHM-SyqUabU4h8RqJ2KTYckQUO3LvIWwmUQ6Qbzw,9
266
+ ipex_llm-2.3.0b20250410.dist-info/RECORD,,