optimum-rbln 0.9.2a3__py3-none-any.whl → 0.9.2a5__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of optimum-rbln might be problematic. Click here for more details.
- optimum/rbln/__init__.py +4 -0
- optimum/rbln/__version__.py +2 -2
- optimum/rbln/configuration_utils.py +3 -0
- optimum/rbln/modeling.py +71 -1
- optimum/rbln/transformers/__init__.py +4 -0
- optimum/rbln/transformers/modeling_generic.py +23 -1
- optimum/rbln/transformers/models/__init__.py +4 -0
- optimum/rbln/transformers/models/blip_2/modeling_blip_2.py +65 -1
- optimum/rbln/transformers/models/decoderonly/__init__.py +1 -0
- optimum/rbln/transformers/models/decoderonly/configuration_decoderonly.py +34 -0
- optimum/rbln/transformers/models/decoderonly/configuration_lora.py +411 -0
- optimum/rbln/transformers/models/decoderonly/decoderonly_architecture.py +100 -20
- optimum/rbln/transformers/models/decoderonly/decoderonly_runtime_utils.py +33 -0
- optimum/rbln/transformers/models/decoderonly/lora_architecture.py +204 -0
- optimum/rbln/transformers/models/decoderonly/modeling_decoderonly.py +79 -4
- optimum/rbln/transformers/models/gemma3/gemma3_architecture.py +12 -2
- optimum/rbln/transformers/models/gemma3/gemma3_runtime_utils.py +31 -3
- optimum/rbln/transformers/models/gemma3/modeling_gemma3.py +9 -1
- optimum/rbln/transformers/models/gpt2/gpt2_architecture.py +4 -1
- optimum/rbln/transformers/models/idefics3/modeling_idefics3.py +2 -4
- optimum/rbln/transformers/models/llava/modeling_llava.py +2 -1
- optimum/rbln/transformers/models/llava_next/modeling_llava_next.py +2 -1
- optimum/rbln/transformers/models/midm/midm_architecture.py +4 -1
- optimum/rbln/transformers/models/phi/phi_architecture.py +5 -1
- optimum/rbln/transformers/models/qwen2_5_vl/modeling_qwen2_5_vl.py +0 -9
- optimum/rbln/transformers/models/qwen2_5_vl/qwen2_5_vl_architecture.py +2 -0
- optimum/rbln/transformers/models/qwen2_vl/qwen2_vl_architecture.py +2 -0
- optimum/rbln/transformers/models/seq2seq/modeling_seq2seq.py +3 -1
- optimum/rbln/transformers/models/whisper/generation_whisper.py +15 -5
- optimum/rbln/transformers/models/whisper/modeling_whisper.py +2 -1
- {optimum_rbln-0.9.2a3.dist-info → optimum_rbln-0.9.2a5.dist-info}/METADATA +5 -5
- {optimum_rbln-0.9.2a3.dist-info → optimum_rbln-0.9.2a5.dist-info}/RECORD +34 -32
- {optimum_rbln-0.9.2a3.dist-info → optimum_rbln-0.9.2a5.dist-info}/WHEEL +0 -0
- {optimum_rbln-0.9.2a3.dist-info → optimum_rbln-0.9.2a5.dist-info}/licenses/LICENSE +0 -0
|
@@ -123,7 +123,10 @@ class MidmAttention(DecoderOnlyAttention):
|
|
|
123
123
|
self.split_size = self._original_mod.split_size
|
|
124
124
|
self.num_key_value_heads = self._original_mod.num_heads
|
|
125
125
|
|
|
126
|
-
def projection(self, hidden_states) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor]:
|
|
126
|
+
def projection(self, hidden_states, lora_int_id) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor]:
|
|
127
|
+
if lora_int_id is not None:
|
|
128
|
+
raise NotImplementedError("LoRA is not supported for MidmAttention")
|
|
129
|
+
|
|
127
130
|
query_states, key_states, value_states = self.c_attn(hidden_states).split(self.split_size, dim=2)
|
|
128
131
|
return query_states, key_states, value_states
|
|
129
132
|
|
|
@@ -56,7 +56,10 @@ class PhiAttention(DecoderOnlyAttention):
|
|
|
56
56
|
self.qk_layernorm = self._original_mod.qk_layernorm
|
|
57
57
|
self.rotary_ndims = self._original_mod.rotary_ndims
|
|
58
58
|
|
|
59
|
-
def projection(self, hidden_states) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor]:
|
|
59
|
+
def projection(self, hidden_states, lora_int_id) -> Tuple[torch.Tensor, torch.Tensor, torch.Tensor]:
|
|
60
|
+
if lora_int_id is not None:
|
|
61
|
+
raise NotImplementedError("LoRA is not supported for PhiAttention")
|
|
62
|
+
|
|
60
63
|
query_states = self.q_proj(hidden_states)
|
|
61
64
|
key_states = self.k_proj(hidden_states)
|
|
62
65
|
value_states = self.v_proj(hidden_states)
|
|
@@ -84,6 +87,7 @@ class PhiLayer(DecoderOnlyLayer):
|
|
|
84
87
|
cos: Optional[torch.Tensor] = None,
|
|
85
88
|
sin: Optional[torch.Tensor] = None,
|
|
86
89
|
block_tables: Optional[torch.Tensor] = None,
|
|
90
|
+
lora_int_id: Optional[torch.Tensor] = None,
|
|
87
91
|
):
|
|
88
92
|
residual = hidden_states
|
|
89
93
|
|
|
@@ -400,15 +400,6 @@ class RBLNQwen2_5_VLForConditionalGeneration(RBLNDecoderOnlyModelForCausalLM):
|
|
|
400
400
|
del model.lm_head
|
|
401
401
|
return model
|
|
402
402
|
|
|
403
|
-
@classmethod
|
|
404
|
-
def update_kwargs(cls, kwargs):
|
|
405
|
-
kwargs.update(
|
|
406
|
-
{
|
|
407
|
-
"_attn_implementation": "eager",
|
|
408
|
-
}
|
|
409
|
-
)
|
|
410
|
-
return super().update_kwargs(kwargs)
|
|
411
|
-
|
|
412
403
|
@classmethod
|
|
413
404
|
def get_input_info(
|
|
414
405
|
cls,
|
|
@@ -165,6 +165,7 @@ class Qwen2_5_VL_LanguageModelWrapper(DecoderOnlyWrapper):
|
|
|
165
165
|
position_embeds = args.pop(0)
|
|
166
166
|
query_position = args.pop(0) if self.phase == "prefill" else None
|
|
167
167
|
position_ids = None
|
|
168
|
+
lora_int_id = None
|
|
168
169
|
attention_mask = args.pop(0) if self.rbln_config.use_attention_mask else None
|
|
169
170
|
past_key_values = args
|
|
170
171
|
|
|
@@ -192,6 +193,7 @@ class Qwen2_5_VL_LanguageModelWrapper(DecoderOnlyWrapper):
|
|
|
192
193
|
query_position,
|
|
193
194
|
attention_mask,
|
|
194
195
|
position_ids,
|
|
196
|
+
lora_int_id,
|
|
195
197
|
past_key_values,
|
|
196
198
|
position_embeds,
|
|
197
199
|
)
|
|
@@ -111,6 +111,7 @@ class Qwen2VL_LanguageModelWrapper(DecoderOnlyWrapper):
|
|
|
111
111
|
query_position = args.pop(0) if self.phase == "prefill" else None
|
|
112
112
|
position_ids = None
|
|
113
113
|
attention_mask = args.pop(0) if self.rbln_config.use_attention_mask else None
|
|
114
|
+
lora_int_id = args.pop(0) if self.rbln_config.lora_config else None
|
|
114
115
|
past_key_values = args
|
|
115
116
|
|
|
116
117
|
if len(past_key_values) != 2 * self.num_hidden_layers:
|
|
@@ -137,6 +138,7 @@ class Qwen2VL_LanguageModelWrapper(DecoderOnlyWrapper):
|
|
|
137
138
|
query_position,
|
|
138
139
|
attention_mask,
|
|
139
140
|
position_ids,
|
|
141
|
+
lora_int_id,
|
|
140
142
|
past_key_values,
|
|
141
143
|
position_embeds,
|
|
142
144
|
)
|
|
@@ -20,6 +20,7 @@ import rebel
|
|
|
20
20
|
import torch
|
|
21
21
|
from rebel.compile_context import CompileContext
|
|
22
22
|
from transformers import AutoModelForSeq2SeqLM, PretrainedConfig, PreTrainedModel
|
|
23
|
+
from transformers.generation.utils import GenerationMixin
|
|
23
24
|
from transformers.modeling_outputs import BaseModelOutput, Seq2SeqLMOutput
|
|
24
25
|
|
|
25
26
|
from ....configuration_utils import RBLNCompileConfig
|
|
@@ -101,7 +102,7 @@ class RBLNRuntimeDecoder(RBLNPytorchRuntime):
|
|
|
101
102
|
return Seq2SeqLMOutput(logits=lm_logits)
|
|
102
103
|
|
|
103
104
|
|
|
104
|
-
class RBLNModelForSeq2SeqLM(RBLNModel, ABC):
|
|
105
|
+
class RBLNModelForSeq2SeqLM(RBLNModel, GenerationMixin, ABC):
|
|
105
106
|
"""
|
|
106
107
|
This is a generic model class that will be instantiated as one of the model classes of the library (with a sequence-to-sequence language modeling head) when created with the from_pretrained() class method.
|
|
107
108
|
This model inherits from [`RBLNModel`]. Check the superclass documentation for the generic methods the library implements for all its models.
|
|
@@ -117,6 +118,7 @@ class RBLNModelForSeq2SeqLM(RBLNModel, ABC):
|
|
|
117
118
|
main_input_name = "input_ids"
|
|
118
119
|
auto_model_class = AutoModelForSeq2SeqLM
|
|
119
120
|
support_causal_attn = None
|
|
121
|
+
_is_stateful = False
|
|
120
122
|
|
|
121
123
|
def __post_init__(self, **kwargs):
|
|
122
124
|
batch_size = self.rbln_config.batch_size
|
|
@@ -39,11 +39,21 @@ from transformers.models.whisper.generation_whisper import WhisperGenerationMixi
|
|
|
39
39
|
|
|
40
40
|
|
|
41
41
|
class RBLNWhisperGenerationMixin(WhisperGenerationMixin, GenerationMixin):
|
|
42
|
-
|
|
43
|
-
|
|
44
|
-
|
|
45
|
-
|
|
46
|
-
|
|
42
|
+
def generate(self, *args, generation_config=None, **kwargs):
|
|
43
|
+
num_beams = kwargs.get(
|
|
44
|
+
"num_beams",
|
|
45
|
+
generation_config.num_beams
|
|
46
|
+
if hasattr(generation_config, "num_beams") and generation_config.num_beams is not None
|
|
47
|
+
else 1,
|
|
48
|
+
)
|
|
49
|
+
if num_beams > 1:
|
|
50
|
+
raise ValueError(
|
|
51
|
+
f"Beam search is not supported in RBLNWhisperGenerationMixin. "
|
|
52
|
+
f"Received num_beams={num_beams}, but only num_beams=1 is allowed. "
|
|
53
|
+
f"Please set num_beams=1 for greedy search or adjust your configuration."
|
|
54
|
+
)
|
|
55
|
+
|
|
56
|
+
return super().generate(*args, **kwargs)
|
|
47
57
|
|
|
48
58
|
def _postprocess_outputs(
|
|
49
59
|
self, seek_outputs, decoder_input_ids, return_token_timestamps, generation_config, *args, **kwargs
|
|
@@ -150,7 +150,8 @@ class RBLNWhisperForConditionalGeneration(RBLNModel, RBLNWhisperGenerationMixin)
|
|
|
150
150
|
"""
|
|
151
151
|
|
|
152
152
|
auto_model_class = AutoModelForSpeechSeq2Seq
|
|
153
|
-
main_input_name = "
|
|
153
|
+
main_input_name = "input_features"
|
|
154
|
+
_is_stateful = False
|
|
154
155
|
|
|
155
156
|
def __post_init__(self, **kwargs):
|
|
156
157
|
super().__post_init__(**kwargs)
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.4
|
|
2
2
|
Name: optimum-rbln
|
|
3
|
-
Version: 0.9.
|
|
3
|
+
Version: 0.9.2a5
|
|
4
4
|
Summary: Optimum RBLN is the interface between the HuggingFace Transformers and Diffusers libraries and RBLN accelerators. It provides a set of tools enabling easy model loading and inference on single and multiple rbln device settings for different downstream tasks.
|
|
5
5
|
Project-URL: Homepage, https://rebellions.ai
|
|
6
6
|
Project-URL: Documentation, https://docs.rbln.ai
|
|
@@ -26,10 +26,10 @@ Requires-Python: <3.14,>=3.9
|
|
|
26
26
|
Requires-Dist: accelerate>=1.0.1
|
|
27
27
|
Requires-Dist: diffusers==0.35.1
|
|
28
28
|
Requires-Dist: packaging>=24.1
|
|
29
|
-
Requires-Dist: torch==2.
|
|
30
|
-
Requires-Dist: torchaudio<=2.
|
|
31
|
-
Requires-Dist: torchvision<=0.
|
|
32
|
-
Requires-Dist: transformers==4.
|
|
29
|
+
Requires-Dist: torch==2.8.0
|
|
30
|
+
Requires-Dist: torchaudio<=2.8.0
|
|
31
|
+
Requires-Dist: torchvision<=0.23.0
|
|
32
|
+
Requires-Dist: transformers==4.57.1
|
|
33
33
|
Description-Content-Type: text/markdown
|
|
34
34
|
|
|
35
35
|
|
|
@@ -1,7 +1,7 @@
|
|
|
1
|
-
optimum/rbln/__init__.py,sha256=
|
|
2
|
-
optimum/rbln/__version__.py,sha256=
|
|
3
|
-
optimum/rbln/configuration_utils.py,sha256=
|
|
4
|
-
optimum/rbln/modeling.py,sha256=
|
|
1
|
+
optimum/rbln/__init__.py,sha256=AZ-7X3ZCjMNcz4mkC_98y-HWRw38ijh8gETJet9tiyM,18828
|
|
2
|
+
optimum/rbln/__version__.py,sha256=SknjPrWnnK8RN1n8X1WB74WmhEFBtAnbp5Zb4GrhDlA,712
|
|
3
|
+
optimum/rbln/configuration_utils.py,sha256=KFibQ8IYcapw3M1GpgNLgQ1ZolRIip0_bOlwfU0OYac,38193
|
|
4
|
+
optimum/rbln/modeling.py,sha256=jfYPOl0FmYq6B3et3rmoCbiWV4939LlVgs-PrZ7IWIk,16928
|
|
5
5
|
optimum/rbln/modeling_base.py,sha256=poXfHZCAlFd28MY9dvMi7tC2RytLx77Lee2XGS_KeZg,27684
|
|
6
6
|
optimum/rbln/diffusers/__init__.py,sha256=1tgU_xWA42BmInqu9bBz_5R_E9TGhhK3mI06YlaiTLg,7232
|
|
7
7
|
optimum/rbln/diffusers/modeling_diffusers.py,sha256=egx137ECmNA0bK2JPRUOpxCl1Wb3qOpE-xSaPaJOs-g,20549
|
|
@@ -72,13 +72,13 @@ optimum/rbln/ops/flash_attn.py,sha256=yTCdYQVqm_1rHMHWjrMQaIR8WTuG_xA6t033x1IVvT
|
|
|
72
72
|
optimum/rbln/ops/kv_cache_update.py,sha256=aIvK2Sp7M3EfJzJgNvIvAJv4emoN6QOhmgaWj-VboLs,1440
|
|
73
73
|
optimum/rbln/ops/linear.py,sha256=5K3pcrrUHu_p8LrMIU-jX2TnafksveFjjZSCsYSp_yw,1328
|
|
74
74
|
optimum/rbln/ops/sliding_window_attn.py,sha256=EQrV_yRGc5z6kvwEsAcLP028bJWkQg2UPI3xubt9skU,3487
|
|
75
|
-
optimum/rbln/transformers/__init__.py,sha256=
|
|
75
|
+
optimum/rbln/transformers/__init__.py,sha256=hyv53b_d_IJ9KYsDogTmKHDNuXIGNCFkrJI21RHE5ak,12573
|
|
76
76
|
optimum/rbln/transformers/configuration_generic.py,sha256=rM4XY1a_UlRf3ZCZkCav59JKRuvqiEEUUgnqNlgdcv8,5207
|
|
77
77
|
optimum/rbln/transformers/modeling_attention_utils.py,sha256=aLyOaq4me1m-JMmnKbuyNQageDxNU2jjEhGE_ew2P5o,11465
|
|
78
|
-
optimum/rbln/transformers/modeling_generic.py,sha256=
|
|
78
|
+
optimum/rbln/transformers/modeling_generic.py,sha256=csViS2KrCr4qaID2jZ1vdHn_W-3wdPFEAtsheEB_LFc,13619
|
|
79
79
|
optimum/rbln/transformers/modeling_outputs.py,sha256=cd8ZlhHAGq7S6i5-QK6TJCxgORvoPMnZpqPBlUc_pMY,1177
|
|
80
80
|
optimum/rbln/transformers/modeling_rope_utils.py,sha256=6Zg3r-TeUk4WQAlr95pqfhuoAD_RQ4njT1rbO9uPL0Q,14379
|
|
81
|
-
optimum/rbln/transformers/models/__init__.py,sha256=
|
|
81
|
+
optimum/rbln/transformers/models/__init__.py,sha256=yzcjyHCHH4-Mi26N34HzNs7Tl5HjjT1rrwQ8f_W2_nc,13532
|
|
82
82
|
optimum/rbln/transformers/models/audio_spectrogram_transformer/__init__.py,sha256=I2vL4lrzbT5p4eJcH-EKHzEfcPkj_XVsie7jb9q6yic,775
|
|
83
83
|
optimum/rbln/transformers/models/audio_spectrogram_transformer/configuration_audio_spectrogram_transformer.py,sha256=z7LJiVJPmnlCM3mcyhPJP8AufSrxO_dsPeJ51onq-Nc,833
|
|
84
84
|
optimum/rbln/transformers/models/audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py,sha256=FIKEVWpIt6-JQX9B_rAfCrAPqdUHtR2i8D_X2k7639E,1498
|
|
@@ -95,7 +95,7 @@ optimum/rbln/transformers/models/bert/configuration_bert.py,sha256=nEZnX6LXpLKWa
|
|
|
95
95
|
optimum/rbln/transformers/models/bert/modeling_bert.py,sha256=7MQZS11k4__oyeni5ek2SzRf-gtD3_hMKl_oOzN7_XQ,2263
|
|
96
96
|
optimum/rbln/transformers/models/blip_2/__init__.py,sha256=L01gPXcUCa8Vg-bcng20vZvBIN_jlqCzwUSFuq0QOag,855
|
|
97
97
|
optimum/rbln/transformers/models/blip_2/configuration_blip_2.py,sha256=8eSilBwcPWQhBg-oilCmDPo-DN6V5lpLMlTB7WPknII,4630
|
|
98
|
-
optimum/rbln/transformers/models/blip_2/modeling_blip_2.py,sha256=
|
|
98
|
+
optimum/rbln/transformers/models/blip_2/modeling_blip_2.py,sha256=MUDwSboH8gdIaJxbPUJsBPuhQf8ViNbVAViU2DASm1g,19308
|
|
99
99
|
optimum/rbln/transformers/models/clip/__init__.py,sha256=TLeXDqcFK6M6v9x7Xr64kBbqGu3hFHM7p754dQ8UVQc,938
|
|
100
100
|
optimum/rbln/transformers/models/clip/configuration_clip.py,sha256=Ea8TCVmMayydfw9p4kTP3UdtvoaPWf4Z4claB61JuE4,4175
|
|
101
101
|
optimum/rbln/transformers/models/clip/modeling_clip.py,sha256=BLAYJAtv_2ZnKOlZ8iDBr2Su3bKM_eMWeUSK9MOaj7I,13198
|
|
@@ -103,12 +103,14 @@ optimum/rbln/transformers/models/colpali/__init__.py,sha256=n3rueXT_oC0N8myoZiic
|
|
|
103
103
|
optimum/rbln/transformers/models/colpali/colpali_architecture.py,sha256=TCOW3v5l9fIt1uIFtWa8ZAxq1cdCER8gXWjmbLQD20M,8079
|
|
104
104
|
optimum/rbln/transformers/models/colpali/configuration_colpali.py,sha256=_HuZBVV-ponml95UapkYpRhffZy53-9jSZknx7hID7o,3348
|
|
105
105
|
optimum/rbln/transformers/models/colpali/modeling_colpali.py,sha256=2lHxvtrK3x2GOv7r-5nZelmjezm3ehe6Qf28cMdNmoQ,17961
|
|
106
|
-
optimum/rbln/transformers/models/decoderonly/__init__.py,sha256=
|
|
107
|
-
optimum/rbln/transformers/models/decoderonly/configuration_decoderonly.py,sha256=
|
|
108
|
-
optimum/rbln/transformers/models/decoderonly/
|
|
109
|
-
optimum/rbln/transformers/models/decoderonly/
|
|
106
|
+
optimum/rbln/transformers/models/decoderonly/__init__.py,sha256=pKBXAtE3y_6nnwYfQJjdPmWqUwxuJ0lr8rrqkgyH07M,1126
|
|
107
|
+
optimum/rbln/transformers/models/decoderonly/configuration_decoderonly.py,sha256=GX-IwTe6ywM9hmyquIu66y0YgIVZS5JNIz8LKAb4Ow8,17003
|
|
108
|
+
optimum/rbln/transformers/models/decoderonly/configuration_lora.py,sha256=5DuTs2vy7jF7MLy161QD_KvCTaNW-5Mok7hBH0yK44U,17356
|
|
109
|
+
optimum/rbln/transformers/models/decoderonly/decoderonly_architecture.py,sha256=h1n9vSHgQc7D0ds1C9SAzWxmIdTaqnDL7auDU_VJNXg,46813
|
|
110
|
+
optimum/rbln/transformers/models/decoderonly/decoderonly_runtime_utils.py,sha256=2yGU1JfxCepG37DbyY8oWq3gMIKhMnPyBkLH8VniKfA,21530
|
|
110
111
|
optimum/rbln/transformers/models/decoderonly/generation_decoderonly.py,sha256=zabSgQd2VzHhkpbhUFW5Z-CjYB1JvSJOb5yXKjXCQV0,4326
|
|
111
|
-
optimum/rbln/transformers/models/decoderonly/
|
|
112
|
+
optimum/rbln/transformers/models/decoderonly/lora_architecture.py,sha256=jo-jYy95JhdvOsX1UTCXeYTNer37wBbtY578C0QQpZo,8306
|
|
113
|
+
optimum/rbln/transformers/models/decoderonly/modeling_decoderonly.py,sha256=myXdyYLivoPVJVGa2vmeaGLFxmjw-vUyYYsNTAX41yk,35907
|
|
112
114
|
optimum/rbln/transformers/models/depth_anything/__init__.py,sha256=xvPSIriMJWyNeVYoVB1Z7YqB4kkHOIkaHq7loNps-dk,756
|
|
113
115
|
optimum/rbln/transformers/models/depth_anything/configuration_depth_anything.py,sha256=JujBVEUa_zZDXNPr1y-B_PhK5SgFFcY8Ib4EoGjjtmE,989
|
|
114
116
|
optimum/rbln/transformers/models/depth_anything/modeling_depth_anything.py,sha256=tTmsVaW9Wb2WD3nKRLwp7swn3hbMvgwUEJwwVIfNYEc,1008
|
|
@@ -128,12 +130,12 @@ optimum/rbln/transformers/models/gemma/gemma_architecture.py,sha256=4Ry2pFfWg0sV
|
|
|
128
130
|
optimum/rbln/transformers/models/gemma/modeling_gemma.py,sha256=AsuFnrwZcRCKtF39BpHHNea0S34N2lNWKV4qZimmY8I,4170
|
|
129
131
|
optimum/rbln/transformers/models/gemma3/__init__.py,sha256=6rugk3615SEt4lh7gduo_J9VyGiSReuEIvL0Uno0eaI,790
|
|
130
132
|
optimum/rbln/transformers/models/gemma3/configuration_gemma3.py,sha256=NJJfarzbWJc3pm0XvICN7D0FFF9nqidagIEoOvYLixQ,4696
|
|
131
|
-
optimum/rbln/transformers/models/gemma3/gemma3_architecture.py,sha256=
|
|
132
|
-
optimum/rbln/transformers/models/gemma3/gemma3_runtime_utils.py,sha256=
|
|
133
|
-
optimum/rbln/transformers/models/gemma3/modeling_gemma3.py,sha256=
|
|
133
|
+
optimum/rbln/transformers/models/gemma3/gemma3_architecture.py,sha256=TkGt2g313hXbB8vPFz8-oDBEsuR3HJI6LjSFgqec_Sc,6533
|
|
134
|
+
optimum/rbln/transformers/models/gemma3/gemma3_runtime_utils.py,sha256=ZhWgecT4v4Ewd1hmrlJH47QUZuQweVB1qAaK-Qw24-Q,11127
|
|
135
|
+
optimum/rbln/transformers/models/gemma3/modeling_gemma3.py,sha256=Yx1rUxTgGvaCsNldggL3rFc2zxsndDKkQovjCmmNf28,25868
|
|
134
136
|
optimum/rbln/transformers/models/gpt2/__init__.py,sha256=SsawHMStE3wYRtqkH5EvdTFkCdX0LLmp-QSKFhEBrHo,740
|
|
135
137
|
optimum/rbln/transformers/models/gpt2/configuration_gpt2.py,sha256=iGdHfzG7plekZcIz-Z5U8lRE4SB8gbJJNcFQJ9l8Myg,1533
|
|
136
|
-
optimum/rbln/transformers/models/gpt2/gpt2_architecture.py,sha256=
|
|
138
|
+
optimum/rbln/transformers/models/gpt2/gpt2_architecture.py,sha256=ul87zvaLkqsuNJirvl6QtGXM147taNEbnb9qPulR1Ps,2933
|
|
137
139
|
optimum/rbln/transformers/models/gpt2/modeling_gpt2.py,sha256=DhF6hU3oCYGbZ7UijKCsRfTx-VCkTqqqNwqqMSrjqRE,2230
|
|
138
140
|
optimum/rbln/transformers/models/grounding_dino/__init__.py,sha256=DE7DipZGvrKC6b1T77k4I4X3G70ss8mlr-PrZCaohto,307
|
|
139
141
|
optimum/rbln/transformers/models/grounding_dino/configuration_grounding_dino.py,sha256=s-5MjEEle0zDBhskeYZQiPbbNsFvpTNcqcz21-kl6Gk,3820
|
|
@@ -141,20 +143,20 @@ optimum/rbln/transformers/models/grounding_dino/grounding_dino_architecture.py,s
|
|
|
141
143
|
optimum/rbln/transformers/models/grounding_dino/modeling_grounding_dino.py,sha256=bXAOs2QH4sy2UFoFLUSM6u1_VHouUT5COERLQX20F6Y,46897
|
|
142
144
|
optimum/rbln/transformers/models/idefics3/__init__.py,sha256=ulxE7HEfXsNJhd25J9Fvi6vggo9aZH9sLKJjWB6LlzQ,814
|
|
143
145
|
optimum/rbln/transformers/models/idefics3/configuration_idefics3.py,sha256=7IENNxflZL8ZH3YRqtCXfYdKs-RdUeGiPzq-C03te_s,3679
|
|
144
|
-
optimum/rbln/transformers/models/idefics3/modeling_idefics3.py,sha256=
|
|
146
|
+
optimum/rbln/transformers/models/idefics3/modeling_idefics3.py,sha256=LEFJu9JsoiS3ZJoG2J3QkwDAyTa75xJQmMtvpomYwsw,19918
|
|
145
147
|
optimum/rbln/transformers/models/llama/__init__.py,sha256=6tgx9-qlM5r9ouoeZEouVRNLs3r6Sku-cuXNkyfeFHc,746
|
|
146
148
|
optimum/rbln/transformers/models/llama/configuration_llama.py,sha256=_uxfH5kaGbeJTMJfESYn0Vg3OEkINS2ShGtVQTeOcs4,1578
|
|
147
149
|
optimum/rbln/transformers/models/llama/llama_architecture.py,sha256=S7MCPfyjG5eUqgaS-QNBB0ApUD6wnb5fR0RHq7k7-pA,728
|
|
148
150
|
optimum/rbln/transformers/models/llama/modeling_llama.py,sha256=uRxEXYhHOuEwPjBo_Ps3eFU1uwScasla6P8HwsQgAu0,4214
|
|
149
151
|
optimum/rbln/transformers/models/llava/__init__.py,sha256=FaVLgBIqKGjT_nvwYO9k9BVqrzH_Ym3DfjGRCSUhG2s,734
|
|
150
152
|
optimum/rbln/transformers/models/llava/configuration_llava.py,sha256=c1rie8LCypxlsT7SNjZJE07_xCLAasV4EBs97o1757Q,2998
|
|
151
|
-
optimum/rbln/transformers/models/llava/modeling_llava.py,sha256=
|
|
153
|
+
optimum/rbln/transformers/models/llava/modeling_llava.py,sha256=MaszTboXRr-PTWZKxhbw5w5rBZ95ES2_fHEW280T2LU,20351
|
|
152
154
|
optimum/rbln/transformers/models/llava_next/__init__.py,sha256=kDXKr7wMkp1XqE__DER2B8kQF_NYMxhzsQS5ytGg56I,752
|
|
153
155
|
optimum/rbln/transformers/models/llava_next/configuration_llava_next.py,sha256=Sz8L8p_23T7xw7pkUmW5pyK_wZclph1p_kQYbslc8m8,2708
|
|
154
|
-
optimum/rbln/transformers/models/llava_next/modeling_llava_next.py,sha256=
|
|
156
|
+
optimum/rbln/transformers/models/llava_next/modeling_llava_next.py,sha256=0aooyMG7ElVIa52MB0ysVKqB4Pdxyl4tbeD1QdehiZk,21342
|
|
155
157
|
optimum/rbln/transformers/models/midm/__init__.py,sha256=IC3FETwgYinbp3wDj7tp4zIHJhbqM-c6GfTRdYcMNj8,913
|
|
156
158
|
optimum/rbln/transformers/models/midm/configuration_midm.py,sha256=DxhcSJlApxfi00XxYmSkKZ6bY9vfLXT0zh-oMKkZot0,1365
|
|
157
|
-
optimum/rbln/transformers/models/midm/midm_architecture.py,sha256=
|
|
159
|
+
optimum/rbln/transformers/models/midm/midm_architecture.py,sha256=f9IwLLyYErliWJhkRj880QByMEYs_XVwm2Yh6r-Y_ik,5186
|
|
158
160
|
optimum/rbln/transformers/models/midm/modeling_midm.py,sha256=EXTBFaOWco2h3NL8jUACeDmPFJuAjFrtQcgO_BTMuCk,6004
|
|
159
161
|
optimum/rbln/transformers/models/mistral/__init__.py,sha256=bYPqrkmqXmhNDqRgKFaL9iH7piGLSHKzsVrGl_0qs1Q,758
|
|
160
162
|
optimum/rbln/transformers/models/mistral/configuration_mistral.py,sha256=mIfz8J8GZV9ojCMuNj9Zeky_PNu1Ir34DQ7FDZrGkP8,1595
|
|
@@ -171,7 +173,7 @@ optimum/rbln/transformers/models/pegasus/pegasus_architecture.py,sha256=Hk4N7LDA
|
|
|
171
173
|
optimum/rbln/transformers/models/phi/__init__.py,sha256=M5Sh4AtIhJYegl-yAKPggAU3DtJtQOa8MrIQypZ6N7U,734
|
|
172
174
|
optimum/rbln/transformers/models/phi/configuration_phi.py,sha256=CXHIG3xlBdr628oDu_u4OGsu_QZLx5EUSqu3zfmfEnk,1553
|
|
173
175
|
optimum/rbln/transformers/models/phi/modeling_phi.py,sha256=r7B0NlqwIGjm-MmE-h5_xeRJPzs4O2OotgbjI-FYA2o,3403
|
|
174
|
-
optimum/rbln/transformers/models/phi/phi_architecture.py,sha256=
|
|
176
|
+
optimum/rbln/transformers/models/phi/phi_architecture.py,sha256=bzK7Qhd1FAC049fdDhzzPYK7HtlHTjBqVmuFAhTX80Q,3866
|
|
175
177
|
optimum/rbln/transformers/models/pixtral/__init__.py,sha256=fhclVAWnIDsfMfC-TW6mYrJXxgyehlLaadK64LOShH4,716
|
|
176
178
|
optimum/rbln/transformers/models/pixtral/configuration_pixtral.py,sha256=b79zkJB1jzHx4S1wTe-Ju_Yel_PS5Q8bfmlQPzkchKU,1677
|
|
177
179
|
optimum/rbln/transformers/models/pixtral/modeling_pixtral.py,sha256=P1lzi6JOTB43nBfCOonUDYhIXoMq6DnQpcvGfOO7ZP8,12252
|
|
@@ -182,12 +184,12 @@ optimum/rbln/transformers/models/qwen2/modeling_qwen2.py,sha256=VOboPJF1rvvSVWkH
|
|
|
182
184
|
optimum/rbln/transformers/models/qwen2/qwen2_architecture.py,sha256=XlNAMYAcDLohnSAhIFGKOPuCB5XLgzYs5ABWdeQSaZs,720
|
|
183
185
|
optimum/rbln/transformers/models/qwen2_5_vl/__init__.py,sha256=rAW3DKQUzGL6EMwa5r1iLu94yhpiZpk6zfoD7TtYXrc,865
|
|
184
186
|
optimum/rbln/transformers/models/qwen2_5_vl/configuration_qwen2_5_vl.py,sha256=WHLH72i7Pe16Ee1waMixMsR3eD6TsMGN08QD82qdVvw,6162
|
|
185
|
-
optimum/rbln/transformers/models/qwen2_5_vl/modeling_qwen2_5_vl.py,sha256=
|
|
186
|
-
optimum/rbln/transformers/models/qwen2_5_vl/qwen2_5_vl_architecture.py,sha256=
|
|
187
|
+
optimum/rbln/transformers/models/qwen2_5_vl/modeling_qwen2_5_vl.py,sha256=ey6uZqf9ULr5LOugf-KrHGKMkdPyZ5XOPt8I-tBBXOc,26730
|
|
188
|
+
optimum/rbln/transformers/models/qwen2_5_vl/qwen2_5_vl_architecture.py,sha256=hlx9Tt9n9m-fL4m21QFKgsN719CDhwhgfOMjnhde4RE,8392
|
|
187
189
|
optimum/rbln/transformers/models/qwen2_vl/__init__.py,sha256=O3t6zKda92CnZDzEnz_dcisMOQ71-OOJxElXzKCH5e0,849
|
|
188
190
|
optimum/rbln/transformers/models/qwen2_vl/configuration_qwen2_vl.py,sha256=mi5CqSKZ77G5Fib3g8a86_4CEB6lb-qJOhDnSqslvNk,4714
|
|
189
191
|
optimum/rbln/transformers/models/qwen2_vl/modeling_qwen2_vl.py,sha256=YO8cKBEb7dU9D--gidYsPyhS2arOwgVqDe3tLlGHdx4,20424
|
|
190
|
-
optimum/rbln/transformers/models/qwen2_vl/qwen2_vl_architecture.py,sha256=
|
|
192
|
+
optimum/rbln/transformers/models/qwen2_vl/qwen2_vl_architecture.py,sha256=xjp52RXqKA_BiyZ5CqwFAosav7ysvOJxeRxbPnTVIjM,5829
|
|
191
193
|
optimum/rbln/transformers/models/qwen3/__init__.py,sha256=tI4KwvXpD35dUUaa8aLUXpWoU9gJGcmKXeywOlH14ZE,746
|
|
192
194
|
optimum/rbln/transformers/models/qwen3/configuration_qwen3.py,sha256=BFRPggnH4VlsXlOa19C6KAID-bPgQ8ooQ29dvogh5zk,2102
|
|
193
195
|
optimum/rbln/transformers/models/qwen3/modeling_qwen3.py,sha256=S05efusxjXJhMMYztstGes6ZbqkSr5I4fHFaLSYVG8c,5760
|
|
@@ -200,7 +202,7 @@ optimum/rbln/transformers/models/roberta/configuration_roberta.py,sha256=6KhO-xB
|
|
|
200
202
|
optimum/rbln/transformers/models/roberta/modeling_roberta.py,sha256=74Pswb5JJNtctvrQHlo2zYocKZN0npWhjAaKMUDVBUU,1535
|
|
201
203
|
optimum/rbln/transformers/models/seq2seq/__init__.py,sha256=HiSyWFcKeZ8okfo-s-_Mf_upyvAoZwraUIJyGNLNurY,714
|
|
202
204
|
optimum/rbln/transformers/models/seq2seq/configuration_seq2seq.py,sha256=pFnt382URDduIpeNb7z_xmONCSOt_2mKssro5xe8y3E,3121
|
|
203
|
-
optimum/rbln/transformers/models/seq2seq/modeling_seq2seq.py,sha256=
|
|
205
|
+
optimum/rbln/transformers/models/seq2seq/modeling_seq2seq.py,sha256=nb511JHbi1wLCr5dOTClItuScx1fb-PCaNxKXjFTVQs,18395
|
|
204
206
|
optimum/rbln/transformers/models/seq2seq/seq2seq_architecture.py,sha256=jmBgj7BkUS_S-T-9DI53rE3KXUHSCoIofr7k5JDVPrU,20024
|
|
205
207
|
optimum/rbln/transformers/models/siglip/__init__.py,sha256=X1Fc1GUnJ2EIxFx45nbeoW-T2t0OyP3W73C0HD8Vowo,712
|
|
206
208
|
optimum/rbln/transformers/models/siglip/configuration_siglip.py,sha256=Fy-ANF91bQno_QVd4ZpyRs-uNgC_XRyBRScBg2uKM6w,3029
|
|
@@ -224,8 +226,8 @@ optimum/rbln/transformers/models/wav2vec2/configuration_wav2vec2.py,sha256=24sXi
|
|
|
224
226
|
optimum/rbln/transformers/models/wav2vec2/modeling_wav2vec2.py,sha256=bMKHdUDHgzu1pXH0yrrOFCiA_T9xqb9B19kljCQ9yUU,1945
|
|
225
227
|
optimum/rbln/transformers/models/whisper/__init__.py,sha256=ErquiUlYycSYPsDcq9IwwmbZXoYLn1MVZ8VikWY5gQo,792
|
|
226
228
|
optimum/rbln/transformers/models/whisper/configuration_whisper.py,sha256=bSwDN7VLuk1aVXvfrQIgb9SLdFBDhO5q8ZFaPQPJal0,3077
|
|
227
|
-
optimum/rbln/transformers/models/whisper/generation_whisper.py,sha256=
|
|
228
|
-
optimum/rbln/transformers/models/whisper/modeling_whisper.py,sha256=
|
|
229
|
+
optimum/rbln/transformers/models/whisper/generation_whisper.py,sha256=0MYzMTZwTHFcJV_ZEtCm2AZbKN9RHgAr9jefuOAouVI,5017
|
|
230
|
+
optimum/rbln/transformers/models/whisper/modeling_whisper.py,sha256=MFKYAqO1ep3teYumMY5E_jjyCU4552GKZacSNFyjVQM,19323
|
|
229
231
|
optimum/rbln/transformers/models/whisper/whisper_architecture.py,sha256=fKUbAMIl20o6EBMVcLg9TDSsJ1FDp8NKcl4jT9RWCEM,13981
|
|
230
232
|
optimum/rbln/transformers/models/xlm_roberta/__init__.py,sha256=O3o2KzJ8Li3QhB7GHdRQASc93SYO2jz00Rx4pxYRuDg,982
|
|
231
233
|
optimum/rbln/transformers/models/xlm_roberta/configuration_xlm_roberta.py,sha256=wHRpGTXL9khYqSkKL1IgA7__6_lt9QpOz9tHumjK7fo,1260
|
|
@@ -243,7 +245,7 @@ optimum/rbln/utils/model_utils.py,sha256=4k5879Kh75m3x_vS4-qOGfqsOiAvc2kdNFFfvsF
|
|
|
243
245
|
optimum/rbln/utils/runtime_utils.py,sha256=R6uXDbeJP03-FWdd4vthNe2D4aCra5n12E3WB1ifiGM,7933
|
|
244
246
|
optimum/rbln/utils/save_utils.py,sha256=hG5uOtYmecSXZuGTvCXsTM-SiyZpr5q3InUGCCq_jzQ,3619
|
|
245
247
|
optimum/rbln/utils/submodule.py,sha256=SKLnM3KsX8_rv3HauO4oB2-JSjzuadQjRwo_BhMUzLI,6362
|
|
246
|
-
optimum_rbln-0.9.
|
|
247
|
-
optimum_rbln-0.9.
|
|
248
|
-
optimum_rbln-0.9.
|
|
249
|
-
optimum_rbln-0.9.
|
|
248
|
+
optimum_rbln-0.9.2a5.dist-info/METADATA,sha256=8gm204ZrCL7yjPNG5vEGsSE5RhIDtPk2H-eWGFxX8yw,5350
|
|
249
|
+
optimum_rbln-0.9.2a5.dist-info/WHEEL,sha256=qtCwoSJWgHk21S1Kb4ihdzI2rlJ1ZKaIurTj_ngOhyQ,87
|
|
250
|
+
optimum_rbln-0.9.2a5.dist-info/licenses/LICENSE,sha256=QwcOLU5TJoTeUhuIXzhdCEEDDvorGiC6-3YTOl4TecE,11356
|
|
251
|
+
optimum_rbln-0.9.2a5.dist-info/RECORD,,
|
|
File without changes
|
|
File without changes
|