optimum-rbln 0.9.4a2__py3-none-any.whl → 0.9.5a4__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- optimum/rbln/__init__.py +36 -0
- optimum/rbln/__version__.py +2 -2
- optimum/rbln/configuration_utils.py +35 -16
- optimum/rbln/modeling_base.py +6 -6
- optimum/rbln/ops/__init__.py +1 -0
- optimum/rbln/ops/attn.py +10 -0
- optimum/rbln/ops/flash_attn.py +8 -0
- optimum/rbln/ops/moe.py +180 -0
- optimum/rbln/ops/sliding_window_attn.py +9 -0
- optimum/rbln/transformers/__init__.py +36 -0
- optimum/rbln/transformers/modeling_attention_utils.py +118 -222
- optimum/rbln/transformers/modeling_outputs.py +25 -0
- optimum/rbln/transformers/modeling_rope_utils.py +78 -42
- optimum/rbln/transformers/models/__init__.py +28 -0
- optimum/rbln/transformers/models/bart/bart_architecture.py +24 -24
- optimum/rbln/transformers/models/colpali/colpali_architecture.py +14 -20
- optimum/rbln/transformers/models/colpali/configuration_colpali.py +12 -17
- optimum/rbln/transformers/models/colpali/modeling_colpali.py +66 -182
- optimum/rbln/transformers/models/colqwen2/configuration_colqwen2.py +38 -21
- optimum/rbln/transformers/models/colqwen2/modeling_colqwen2.py +107 -371
- optimum/rbln/transformers/models/decoderonly/__init__.py +2 -0
- optimum/rbln/transformers/models/decoderonly/configuration_decoderonly.py +118 -16
- optimum/rbln/transformers/models/decoderonly/configuration_lora.py +1 -1
- optimum/rbln/transformers/models/decoderonly/decoderonly_architecture.py +121 -48
- optimum/rbln/transformers/models/decoderonly/decoderonly_runtime_utils.py +5 -7
- optimum/rbln/transformers/models/decoderonly/modeling_decoderonly.py +75 -107
- optimum/rbln/transformers/models/exaone/exaone_architecture.py +0 -36
- optimum/rbln/transformers/models/gemma/gemma_architecture.py +1 -1
- optimum/rbln/transformers/models/gemma2/__init__.py +16 -0
- optimum/rbln/transformers/models/gemma2/configuration_gemma2.py +45 -0
- optimum/rbln/transformers/models/gemma2/gemma2_architecture.py +83 -0
- optimum/rbln/transformers/models/gemma2/modeling_gemma2.py +101 -0
- optimum/rbln/transformers/models/gemma3/gemma3_architecture.py +16 -18
- optimum/rbln/transformers/models/gemma3/modeling_gemma3.py +1 -1
- optimum/rbln/transformers/models/gpt2/gpt2_architecture.py +8 -34
- optimum/rbln/transformers/models/gpt_oss/__init__.py +16 -0
- optimum/rbln/transformers/models/gpt_oss/configuration_gpt_oss.py +41 -0
- optimum/rbln/transformers/models/gpt_oss/gpt_oss_architecture.py +122 -0
- optimum/rbln/transformers/models/gpt_oss/modeling_gpt_oss.py +165 -0
- optimum/rbln/transformers/models/grounding_dino/configuration_grounding_dino.py +8 -5
- optimum/rbln/transformers/models/grounding_dino/grounding_dino_architecture.py +6 -4
- optimum/rbln/transformers/models/llava/modeling_llava.py +0 -1
- optimum/rbln/transformers/models/midm/midm_architecture.py +29 -22
- optimum/rbln/transformers/models/opt/opt_architecture.py +1 -44
- optimum/rbln/transformers/models/paligemma/__init__.py +16 -0
- optimum/rbln/transformers/models/paligemma/configuration_paligemma.py +129 -0
- optimum/rbln/transformers/models/paligemma/modeling_paligemma.py +564 -0
- optimum/rbln/transformers/models/pegasus/pegasus_architecture.py +24 -24
- optimum/rbln/transformers/models/phi/phi_architecture.py +13 -21
- optimum/rbln/transformers/models/qwen2_5_vl/__init__.py +6 -1
- optimum/rbln/transformers/models/qwen2_5_vl/configuration_qwen2_5_vl.py +11 -1
- optimum/rbln/transformers/models/qwen2_5_vl/modeling_qwen2_5_vl.py +271 -122
- optimum/rbln/transformers/models/qwen2_5_vl/qwen2_5_vl_architecture.py +43 -39
- optimum/rbln/transformers/models/qwen2_moe/__init__.py +16 -0
- optimum/rbln/transformers/models/qwen2_moe/configuration_qwen2_moe.py +38 -0
- optimum/rbln/transformers/models/qwen2_moe/modeling_qwen2_moe.py +68 -0
- optimum/rbln/transformers/models/qwen2_moe/qwen2_moe_architecture.py +94 -0
- optimum/rbln/transformers/models/qwen2_vl/__init__.py +6 -1
- optimum/rbln/transformers/models/qwen2_vl/configuration_qwen2_vl.py +11 -1
- optimum/rbln/transformers/models/qwen2_vl/modeling_qwen2_vl.py +263 -105
- optimum/rbln/transformers/models/qwen2_vl/qwen2_vl_architecture.py +26 -34
- optimum/rbln/transformers/models/qwen3/qwen3_architecture.py +7 -7
- optimum/rbln/transformers/models/qwen3_moe/__init__.py +16 -0
- optimum/rbln/transformers/models/qwen3_moe/configuration_qwen3_moe.py +38 -0
- optimum/rbln/transformers/models/qwen3_moe/modeling_qwen3_moe.py +68 -0
- optimum/rbln/transformers/models/qwen3_moe/qwen3_moe_architecture.py +100 -0
- optimum/rbln/transformers/models/seq2seq/seq2seq_architecture.py +14 -12
- optimum/rbln/transformers/models/siglip/modeling_siglip.py +4 -18
- optimum/rbln/transformers/models/swin/configuration_swin.py +1 -6
- optimum/rbln/transformers/models/t5/t5_architecture.py +15 -16
- optimum/rbln/transformers/models/time_series_transformer/time_series_transformers_architecture.py +0 -3
- optimum/rbln/transformers/models/whisper/whisper_architecture.py +0 -3
- optimum/rbln/transformers/utils/rbln_quantization.py +20 -12
- optimum/rbln/utils/import_utils.py +16 -1
- optimum/rbln/utils/runtime_utils.py +10 -6
- optimum/rbln/utils/submodule.py +24 -0
- {optimum_rbln-0.9.4a2.dist-info → optimum_rbln-0.9.5a4.dist-info}/METADATA +6 -6
- {optimum_rbln-0.9.4a2.dist-info → optimum_rbln-0.9.5a4.dist-info}/RECORD +81 -62
- optimum/rbln/transformers/models/colqwen2/colqwen2_architecture.py +0 -233
- {optimum_rbln-0.9.4a2.dist-info → optimum_rbln-0.9.5a4.dist-info}/WHEEL +0 -0
- {optimum_rbln-0.9.4a2.dist-info → optimum_rbln-0.9.5a4.dist-info}/entry_points.txt +0 -0
- {optimum_rbln-0.9.4a2.dist-info → optimum_rbln-0.9.5a4.dist-info}/licenses/LICENSE +0 -0
|
@@ -1,9 +1,9 @@
|
|
|
1
|
-
optimum/rbln/__init__.py,sha256=
|
|
2
|
-
optimum/rbln/__version__.py,sha256=
|
|
1
|
+
optimum/rbln/__init__.py,sha256=8beQjjRL9J4-o1Bpg051HXil3CQOp3YX8UZ6qvZdqIg,20824
|
|
2
|
+
optimum/rbln/__version__.py,sha256=J_joXgk9sJs4d8LV1-4tzsQl2YEoPN8AtINjd9GCXGc,712
|
|
3
3
|
optimum/rbln/cli.py,sha256=944P_f9btDyFryHfHzxUKQvwXWYD1hrceDuK6SWNQcQ,22832
|
|
4
|
-
optimum/rbln/configuration_utils.py,sha256=
|
|
4
|
+
optimum/rbln/configuration_utils.py,sha256=IwGETb_43ILHbfm9PouMDUND2Ym4lpgesefjNwnMU_k,38470
|
|
5
5
|
optimum/rbln/modeling.py,sha256=M9kEqbAqVZIeFxOF9dyfS8i7loJz3LV67zf1wzeVcxM,15218
|
|
6
|
-
optimum/rbln/modeling_base.py,sha256=
|
|
6
|
+
optimum/rbln/modeling_base.py,sha256=dwvORoAQSCBzAvz7nM8HRP6_jrc0oAMHs3gYvH4ZkkE,27760
|
|
7
7
|
optimum/rbln/diffusers/__init__.py,sha256=dISoQ-mylK-n9DM0doqo3oeQFA2SWu9BZcbrcr4vO0I,7800
|
|
8
8
|
optimum/rbln/diffusers/modeling_diffusers.py,sha256=M4lpr5fdbL9fAGVPc_1IY9evxDA47rV68WyvK9jhnO0,20423
|
|
9
9
|
optimum/rbln/diffusers/configurations/__init__.py,sha256=8xhIEEa9HTbIg-9khGvxqr6kHPaMnnHZc-BNsEADO1o,1458
|
|
@@ -74,19 +74,20 @@ optimum/rbln/diffusers/pipelines/stable_diffusion_xl/pipeline_stable_diffusion_x
|
|
|
74
74
|
optimum/rbln/diffusers/pipelines/stable_diffusion_xl/pipeline_stable_diffusion_xl_inpaint.py,sha256=XUchrMMrm70KKOheSUM6mmzoGWtTyCKd9vkAcd5kvEo,1407
|
|
75
75
|
optimum/rbln/diffusers/pipelines/stable_video_diffusion/__init__.py,sha256=8Mqu_saC_Wp67Zx5UKq90eeKv8tzv8bIm5z9lUdoa3o,677
|
|
76
76
|
optimum/rbln/diffusers/pipelines/stable_video_diffusion/pipeline_stable_video_diffusion.py,sha256=ApXTZZB_tzOBBCY97xHc6tg60-eY3zWXCtEWd3Ou3i4,1864
|
|
77
|
-
optimum/rbln/ops/__init__.py,sha256=
|
|
78
|
-
optimum/rbln/ops/attn.py,sha256=
|
|
79
|
-
optimum/rbln/ops/flash_attn.py,sha256=
|
|
77
|
+
optimum/rbln/ops/__init__.py,sha256=9d4zXHpf4EUYxUBt84_TFqpLcBVDHrBtuQ16vL35iPQ,757
|
|
78
|
+
optimum/rbln/ops/attn.py,sha256=fMSwLi0hSyI9StFGmdcyJKBcPVPzSYgo77e3asvT3og,13189
|
|
79
|
+
optimum/rbln/ops/flash_attn.py,sha256=IOQMF6NmYJjxxhvwo25Y2vDjMKivPSGjBX0W_9Xa1XM,8154
|
|
80
80
|
optimum/rbln/ops/kv_cache_update.py,sha256=aIvK2Sp7M3EfJzJgNvIvAJv4emoN6QOhmgaWj-VboLs,1440
|
|
81
81
|
optimum/rbln/ops/linear.py,sha256=5K3pcrrUHu_p8LrMIU-jX2TnafksveFjjZSCsYSp_yw,1328
|
|
82
|
-
optimum/rbln/ops/
|
|
83
|
-
optimum/rbln/
|
|
82
|
+
optimum/rbln/ops/moe.py,sha256=WJBFt88e3dEaLzs45PTD1vcnYhWTs6yHhNdKMROl9_E,5435
|
|
83
|
+
optimum/rbln/ops/sliding_window_attn.py,sha256=1geQsTLGzoOusTF9Uw-2QzWI8h8MsmD5qYEYG0gTvXw,3784
|
|
84
|
+
optimum/rbln/transformers/__init__.py,sha256=ciwG4UKHOKRPewrMrlBji5s6oPK8ole1tGIz85TLr2Q,14001
|
|
84
85
|
optimum/rbln/transformers/configuration_generic.py,sha256=5_KWSqcpsEoAHXhMzEpLV62m-0DlWqCY_zwgi9kzjIs,4161
|
|
85
|
-
optimum/rbln/transformers/modeling_attention_utils.py,sha256=
|
|
86
|
+
optimum/rbln/transformers/modeling_attention_utils.py,sha256=0qZdigdk66Y9f21q3suxMvTZXBN0uBsGCsKgFHTLO5c,12275
|
|
86
87
|
optimum/rbln/transformers/modeling_generic.py,sha256=Z_1m5d_hsmQC2qnNkargjMIqlIm3FzBFTNUKqdYdaOc,11499
|
|
87
|
-
optimum/rbln/transformers/modeling_outputs.py,sha256=
|
|
88
|
-
optimum/rbln/transformers/modeling_rope_utils.py,sha256=
|
|
89
|
-
optimum/rbln/transformers/models/__init__.py,sha256=
|
|
88
|
+
optimum/rbln/transformers/modeling_outputs.py,sha256=7ANZ-i-pnY1iBHwD64Gizqdk5RyiAS1htO7EInTMNU0,2411
|
|
89
|
+
optimum/rbln/transformers/modeling_rope_utils.py,sha256=vKs2b1jI1PVxi7dlcLKZvm-T-9OdRPCRt_QZxZg7IyY,16383
|
|
90
|
+
optimum/rbln/transformers/models/__init__.py,sha256=DhM3UcXIxFGFIEFDqdwIhXMAmMoRuTsXKngsR7TylFs,15065
|
|
90
91
|
optimum/rbln/transformers/models/audio_spectrogram_transformer/__init__.py,sha256=I2vL4lrzbT5p4eJcH-EKHzEfcPkj_XVsie7jb9q6yic,775
|
|
91
92
|
optimum/rbln/transformers/models/audio_spectrogram_transformer/configuration_audio_spectrogram_transformer.py,sha256=biFBo1twaWScF2CmNYoF_PW-RvJBUfVgBVpdfI_igBY,1741
|
|
92
93
|
optimum/rbln/transformers/models/audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py,sha256=Kzya5XlU15FkEQlWj1HYXBAGrJleEyF9wiTSiHXSIqo,4124
|
|
@@ -94,7 +95,7 @@ optimum/rbln/transformers/models/auto/__init__.py,sha256=tdYqXkg9xBGNr4fZjH7_O3q
|
|
|
94
95
|
optimum/rbln/transformers/models/auto/auto_factory.py,sha256=xUDgwrcVwe1NtiMreZ17AYOwHFR950WURleS2PxTC4U,11801
|
|
95
96
|
optimum/rbln/transformers/models/auto/modeling_auto.py,sha256=Iu5vl60CHX1XNbn445woZhSUkNSf0qHPW0DzbcAnDYE,5797
|
|
96
97
|
optimum/rbln/transformers/models/bart/__init__.py,sha256=fVo-gZEmJ0yxkIxEX6ciuRAGgXNyuvaXE2s88bhbjAE,830
|
|
97
|
-
optimum/rbln/transformers/models/bart/bart_architecture.py,sha256=
|
|
98
|
+
optimum/rbln/transformers/models/bart/bart_architecture.py,sha256=XaUy0fV-HMwywS_ncO6B3q5IHFApWR6QqyMm9kzYO7I,6077
|
|
98
99
|
optimum/rbln/transformers/models/bart/configuration_bart.py,sha256=PrRA7OwPTegPamd_mmVnwNygRbNG7pZrsrXdKyfZ6Bo,1351
|
|
99
100
|
optimum/rbln/transformers/models/bart/modeling_bart.py,sha256=FCSZFmq0bhPHkne-1tmsxiQH3E427VfkWtLU9Nct3a0,3276
|
|
100
101
|
optimum/rbln/transformers/models/bert/__init__.py,sha256=86FuGRBLw315_Roa9D5OUx6Ku2PM0DqSPZ-YSqbF-io,806
|
|
@@ -108,21 +109,20 @@ optimum/rbln/transformers/models/clip/__init__.py,sha256=TLeXDqcFK6M6v9x7Xr64kBb
|
|
|
108
109
|
optimum/rbln/transformers/models/clip/configuration_clip.py,sha256=Ea8TCVmMayydfw9p4kTP3UdtvoaPWf4Z4claB61JuE4,4175
|
|
109
110
|
optimum/rbln/transformers/models/clip/modeling_clip.py,sha256=CeHl52UVr2UVKUeWTyT8OcRWXsZzrLnQpjzK_neu304,14835
|
|
110
111
|
optimum/rbln/transformers/models/colpali/__init__.py,sha256=n3rueXT_oC0N8myoZiic0YkVK24CW5hZBPa-0L8so6Y,119
|
|
111
|
-
optimum/rbln/transformers/models/colpali/colpali_architecture.py,sha256=
|
|
112
|
-
optimum/rbln/transformers/models/colpali/configuration_colpali.py,sha256=
|
|
113
|
-
optimum/rbln/transformers/models/colpali/modeling_colpali.py,sha256=
|
|
112
|
+
optimum/rbln/transformers/models/colpali/colpali_architecture.py,sha256=qeAdCmYHcDLNmFyIwYSIo1hI9K_8dyVgD76wDShdolo,7816
|
|
113
|
+
optimum/rbln/transformers/models/colpali/configuration_colpali.py,sha256=9qFi2N1T-3crCYzxG6JVmIr2eFh_SbT1wDIDF4B-Nck,2893
|
|
114
|
+
optimum/rbln/transformers/models/colpali/modeling_colpali.py,sha256=sKf0k5h5AGkOl9kVyYQ_Cj77_37otVA6A5Q3sPtZSRU,10479
|
|
114
115
|
optimum/rbln/transformers/models/colqwen2/__init__.py,sha256=gEKc5X4uGME4XKySDD1H6JlT89jaMvZ00HqbDVXNHU8,123
|
|
115
|
-
optimum/rbln/transformers/models/colqwen2/
|
|
116
|
-
optimum/rbln/transformers/models/colqwen2/
|
|
117
|
-
optimum/rbln/transformers/models/
|
|
118
|
-
optimum/rbln/transformers/models/decoderonly/
|
|
119
|
-
optimum/rbln/transformers/models/decoderonly/
|
|
120
|
-
optimum/rbln/transformers/models/decoderonly/
|
|
121
|
-
optimum/rbln/transformers/models/decoderonly/
|
|
122
|
-
optimum/rbln/transformers/models/decoderonly/decoderonly_runtime_utils.py,sha256=vY7S-4ms4eW3WIEGLfjDzX3dsQKcc6QHhOiUOXyWSWc,29268
|
|
116
|
+
optimum/rbln/transformers/models/colqwen2/configuration_colqwen2.py,sha256=mJ0bTbPAkQw3oP-klaEQK8a7ssglsktC9TngdgNpYYo,3189
|
|
117
|
+
optimum/rbln/transformers/models/colqwen2/modeling_colqwen2.py,sha256=bZD6uvJ0lv8AjifG7lzsdR7q5F7NaQibQhsBlJm_sII,7946
|
|
118
|
+
optimum/rbln/transformers/models/decoderonly/__init__.py,sha256=uVUvtkIZ1ewX-9fHsPlmJSfAQTuT1WTFUT1rBLuXSHQ,1165
|
|
119
|
+
optimum/rbln/transformers/models/decoderonly/configuration_decoderonly.py,sha256=jO2Kk_Vuo1LitNl7UGy83EBPetJ875iJcBpDNWvoOQY,21533
|
|
120
|
+
optimum/rbln/transformers/models/decoderonly/configuration_lora.py,sha256=DAKpiqN1F53p3gImcUDsapUkRvOq3x3Igi9iAC4rNHo,17357
|
|
121
|
+
optimum/rbln/transformers/models/decoderonly/decoderonly_architecture.py,sha256=Jj_hk0Sg8xVqbdvkBsCMw3xr8bve3iyXyQ-NIWJ5sB4,51932
|
|
122
|
+
optimum/rbln/transformers/models/decoderonly/decoderonly_runtime_utils.py,sha256=FfPQuNMn7fQfEjWpFKvfkI4mUpRtcm20tUd17QrF4Cw,29216
|
|
123
123
|
optimum/rbln/transformers/models/decoderonly/generation_decoderonly.py,sha256=_Rp1vtGow4quWHnIKpHtZFGMxrLjIN-FCc6gz0XL1Sc,5539
|
|
124
124
|
optimum/rbln/transformers/models/decoderonly/lora_architecture.py,sha256=c4O5N56Y2uS-gxsXITt19qmqyV2rqSlPEHxkmOnWE2s,8306
|
|
125
|
-
optimum/rbln/transformers/models/decoderonly/modeling_decoderonly.py,sha256=
|
|
125
|
+
optimum/rbln/transformers/models/decoderonly/modeling_decoderonly.py,sha256=flq6ItDGu2gO5xnQx2dfBPXIIIDnWvOSu_wlYXOM7AI,38302
|
|
126
126
|
optimum/rbln/transformers/models/depth_anything/__init__.py,sha256=xvPSIriMJWyNeVYoVB1Z7YqB4kkHOIkaHq7loNps-dk,756
|
|
127
127
|
optimum/rbln/transformers/models/depth_anything/configuration_depth_anything.py,sha256=JujBVEUa_zZDXNPr1y-B_PhK5SgFFcY8Ib4EoGjjtmE,989
|
|
128
128
|
optimum/rbln/transformers/models/depth_anything/modeling_depth_anything.py,sha256=RxscJiKp7PDmbQTDUy2R_Ryxf_0YZ0TieRS5bg53dyQ,1698
|
|
@@ -134,24 +134,32 @@ optimum/rbln/transformers/models/dpt/configuration_dpt.py,sha256=3Bb_K0sKI6TKeoH
|
|
|
134
134
|
optimum/rbln/transformers/models/dpt/modeling_dpt.py,sha256=2XnXwCTeJ7Vj28yK7kvqWxrViqikteX-7l-Ys7ubJpY,1649
|
|
135
135
|
optimum/rbln/transformers/models/exaone/__init__.py,sha256=eUL0mq3yGVzCQfjLlOtVF2MecIN3DQWm07EmXubGSTs,921
|
|
136
136
|
optimum/rbln/transformers/models/exaone/configuration_exaone.py,sha256=S4s4kJemPbmn-otYv-XNHE40DJaEYY6cmzaWV6MTGsY,1388
|
|
137
|
-
optimum/rbln/transformers/models/exaone/exaone_architecture.py,sha256=
|
|
137
|
+
optimum/rbln/transformers/models/exaone/exaone_architecture.py,sha256=jdl18rSIokNJDQDw2dZzPOYymfORSHXABqVUqhIMFUI,1322
|
|
138
138
|
optimum/rbln/transformers/models/exaone/modeling_exaone.py,sha256=yW0Hws35v_70OGn0nmDNMXAsnAzniyxUZ78VmjIbpJg,6060
|
|
139
139
|
optimum/rbln/transformers/models/gemma/__init__.py,sha256=HQISZaSrhwGtsvGuHqkiyMhoS1QAVhd0tXXCaCruq4U,746
|
|
140
140
|
optimum/rbln/transformers/models/gemma/configuration_gemma.py,sha256=H1nVp8HBJxxN__VFyDpBuhoqf_RZxgvfE2cVOSvxBIg,1569
|
|
141
|
-
optimum/rbln/transformers/models/gemma/gemma_architecture.py,sha256=
|
|
141
|
+
optimum/rbln/transformers/models/gemma/gemma_architecture.py,sha256=dDsmoiPippuQVOZefXN7cBXrDWb_1maRYpIUQ9iZe6k,929
|
|
142
142
|
optimum/rbln/transformers/models/gemma/modeling_gemma.py,sha256=AsuFnrwZcRCKtF39BpHHNea0S34N2lNWKV4qZimmY8I,4170
|
|
143
|
+
optimum/rbln/transformers/models/gemma2/__init__.py,sha256=AhpaC9zAb8yeNJyCspN1s-GnNr3ag5nAQa5Wta9OPLs,752
|
|
144
|
+
optimum/rbln/transformers/models/gemma2/configuration_gemma2.py,sha256=kJeo3dz4Ba-YqLEgIvxeRpEUa3-l0zG_x0IWdDXA5_M,1574
|
|
145
|
+
optimum/rbln/transformers/models/gemma2/gemma2_architecture.py,sha256=fjNr66rnN7JP3C2If09OXYwVbhLPKqZ7I7ng7YK3MF8,2896
|
|
146
|
+
optimum/rbln/transformers/models/gemma2/modeling_gemma2.py,sha256=cQUzthGnVQn3wWX2hrCc4idCEfEg4snbQ3i8LOuH1Tg,4209
|
|
143
147
|
optimum/rbln/transformers/models/gemma3/__init__.py,sha256=6rugk3615SEt4lh7gduo_J9VyGiSReuEIvL0Uno0eaI,790
|
|
144
148
|
optimum/rbln/transformers/models/gemma3/configuration_gemma3.py,sha256=NJJfarzbWJc3pm0XvICN7D0FFF9nqidagIEoOvYLixQ,4696
|
|
145
|
-
optimum/rbln/transformers/models/gemma3/gemma3_architecture.py,sha256=
|
|
149
|
+
optimum/rbln/transformers/models/gemma3/gemma3_architecture.py,sha256=qdYHvn8KY93Is1W2PcOydnPRnzwD7OS4YhFCBrJukHg,6799
|
|
146
150
|
optimum/rbln/transformers/models/gemma3/gemma3_runtime_utils.py,sha256=1f8LVqtF3Tr6ITVC43QpiSXefFsVeZ7jStoR4SlNTfk,9640
|
|
147
|
-
optimum/rbln/transformers/models/gemma3/modeling_gemma3.py,sha256
|
|
151
|
+
optimum/rbln/transformers/models/gemma3/modeling_gemma3.py,sha256=-AclIXvUZt83QsJQqyjxNEKoiv9OZAtRI3Wi2WuYjuA,26557
|
|
148
152
|
optimum/rbln/transformers/models/gpt2/__init__.py,sha256=SsawHMStE3wYRtqkH5EvdTFkCdX0LLmp-QSKFhEBrHo,740
|
|
149
153
|
optimum/rbln/transformers/models/gpt2/configuration_gpt2.py,sha256=iGdHfzG7plekZcIz-Z5U8lRE4SB8gbJJNcFQJ9l8Myg,1533
|
|
150
|
-
optimum/rbln/transformers/models/gpt2/gpt2_architecture.py,sha256=
|
|
154
|
+
optimum/rbln/transformers/models/gpt2/gpt2_architecture.py,sha256=HidhcrVrZddMGdUMIspAER7NvSNyeDeihcSuP_9Lrn8,2264
|
|
151
155
|
optimum/rbln/transformers/models/gpt2/modeling_gpt2.py,sha256=DhF6hU3oCYGbZ7UijKCsRfTx-VCkTqqqNwqqMSrjqRE,2230
|
|
156
|
+
optimum/rbln/transformers/models/gpt_oss/__init__.py,sha256=pUUqYnDf9-Lc1GgKW5eRCxqQ7p9dUJMYzJSw1MDiq40,714
|
|
157
|
+
optimum/rbln/transformers/models/gpt_oss/configuration_gpt_oss.py,sha256=_Z4nAyayFuai8WqIGDZq6aB4ZxtY2TTvukQ31Hfwy9c,1345
|
|
158
|
+
optimum/rbln/transformers/models/gpt_oss/gpt_oss_architecture.py,sha256=3bqFPSM8QJ1yGRnmcrqfQFV_joP0ewbCQ-0pO02wWos,4519
|
|
159
|
+
optimum/rbln/transformers/models/gpt_oss/modeling_gpt_oss.py,sha256=-mFmV933EGpysGnK8M6HYGAC694nEH8x7gvMgBUVllo,6235
|
|
152
160
|
optimum/rbln/transformers/models/grounding_dino/__init__.py,sha256=DE7DipZGvrKC6b1T77k4I4X3G70ss8mlr-PrZCaohto,307
|
|
153
|
-
optimum/rbln/transformers/models/grounding_dino/configuration_grounding_dino.py,sha256=
|
|
154
|
-
optimum/rbln/transformers/models/grounding_dino/grounding_dino_architecture.py,sha256=
|
|
161
|
+
optimum/rbln/transformers/models/grounding_dino/configuration_grounding_dino.py,sha256=ktQD85690lD2R9HUYzYsORHZe9vUaKMqYK4f6RcRhuM,4182
|
|
162
|
+
optimum/rbln/transformers/models/grounding_dino/grounding_dino_architecture.py,sha256=dOvLHD7w4mLA325i1uJW65l3R2eaImYkLgrvnJkFTSM,26794
|
|
155
163
|
optimum/rbln/transformers/models/grounding_dino/modeling_grounding_dino.py,sha256=8LKysGbqrheFI2nWZpv4ErUfWmUuca3UE2lQo0RIYhc,48429
|
|
156
164
|
optimum/rbln/transformers/models/idefics3/__init__.py,sha256=ulxE7HEfXsNJhd25J9Fvi6vggo9aZH9sLKJjWB6LlzQ,814
|
|
157
165
|
optimum/rbln/transformers/models/idefics3/configuration_idefics3.py,sha256=7IENNxflZL8ZH3YRqtCXfYdKs-RdUeGiPzq-C03te_s,3679
|
|
@@ -162,13 +170,13 @@ optimum/rbln/transformers/models/llama/llama_architecture.py,sha256=S7MCPfyjG5eU
|
|
|
162
170
|
optimum/rbln/transformers/models/llama/modeling_llama.py,sha256=uRxEXYhHOuEwPjBo_Ps3eFU1uwScasla6P8HwsQgAu0,4214
|
|
163
171
|
optimum/rbln/transformers/models/llava/__init__.py,sha256=FaVLgBIqKGjT_nvwYO9k9BVqrzH_Ym3DfjGRCSUhG2s,734
|
|
164
172
|
optimum/rbln/transformers/models/llava/configuration_llava.py,sha256=c1rie8LCypxlsT7SNjZJE07_xCLAasV4EBs97o1757Q,2998
|
|
165
|
-
optimum/rbln/transformers/models/llava/modeling_llava.py,sha256=
|
|
173
|
+
optimum/rbln/transformers/models/llava/modeling_llava.py,sha256=9m8R6gappLYbkFvQKRFod1pmgKF1ejdlg6XI4TjHdI8,21069
|
|
166
174
|
optimum/rbln/transformers/models/llava_next/__init__.py,sha256=kDXKr7wMkp1XqE__DER2B8kQF_NYMxhzsQS5ytGg56I,752
|
|
167
175
|
optimum/rbln/transformers/models/llava_next/configuration_llava_next.py,sha256=Sz8L8p_23T7xw7pkUmW5pyK_wZclph1p_kQYbslc8m8,2708
|
|
168
176
|
optimum/rbln/transformers/models/llava_next/modeling_llava_next.py,sha256=aVpM0SYSSZ6jClMvpO5J83qrUg3GkDBAFkGC5ylSOvw,21303
|
|
169
177
|
optimum/rbln/transformers/models/midm/__init__.py,sha256=IC3FETwgYinbp3wDj7tp4zIHJhbqM-c6GfTRdYcMNj8,913
|
|
170
178
|
optimum/rbln/transformers/models/midm/configuration_midm.py,sha256=DxhcSJlApxfi00XxYmSkKZ6bY9vfLXT0zh-oMKkZot0,1365
|
|
171
|
-
optimum/rbln/transformers/models/midm/midm_architecture.py,sha256=
|
|
179
|
+
optimum/rbln/transformers/models/midm/midm_architecture.py,sha256=sWorUfJjUk3-19flA4_7FEOG_uhgXli4A9QCyIpTR4Y,5518
|
|
172
180
|
optimum/rbln/transformers/models/midm/modeling_midm.py,sha256=EXTBFaOWco2h3NL8jUACeDmPFJuAjFrtQcgO_BTMuCk,6004
|
|
173
181
|
optimum/rbln/transformers/models/mistral/__init__.py,sha256=bYPqrkmqXmhNDqRgKFaL9iH7piGLSHKzsVrGl_0qs1Q,758
|
|
174
182
|
optimum/rbln/transformers/models/mistral/configuration_mistral.py,sha256=mIfz8J8GZV9ojCMuNj9Zeky_PNu1Ir34DQ7FDZrGkP8,1595
|
|
@@ -177,15 +185,18 @@ optimum/rbln/transformers/models/mistral/modeling_mistral.py,sha256=TdOPjF1eUDWa
|
|
|
177
185
|
optimum/rbln/transformers/models/opt/__init__.py,sha256=mkSmAUr_ezMtlMK77f48T0THTFddf0HThH1lp6y5Pfw,734
|
|
178
186
|
optimum/rbln/transformers/models/opt/configuration_opt.py,sha256=aP7cyEuBF4DrQxVERPdP3fXYkuqIUcGxEK2fc8ezh7I,1135
|
|
179
187
|
optimum/rbln/transformers/models/opt/modeling_opt.py,sha256=4KZlCnKwDIOMbltPxvO7FX-lIRmI2auC-NVTWqkIPmc,4002
|
|
180
|
-
optimum/rbln/transformers/models/opt/opt_architecture.py,sha256=
|
|
188
|
+
optimum/rbln/transformers/models/opt/opt_architecture.py,sha256=m2vdc9zveTvgXvvfNd9CRHsD4YTDKmc9wKXCu_ZHNRo,1128
|
|
189
|
+
optimum/rbln/transformers/models/paligemma/__init__.py,sha256=Q5hks3HpJ-V9RDHyf66EMKMku8UiZ69jx41g1YRqboI,796
|
|
190
|
+
optimum/rbln/transformers/models/paligemma/configuration_paligemma.py,sha256=z7sGydwLkTxK0MCE2NbaF6ATFNJGqFUo5aYQ7rpGDHE,5837
|
|
191
|
+
optimum/rbln/transformers/models/paligemma/modeling_paligemma.py,sha256=1guP-TV67DCvEXw3iyefStvg9HReCMMT7QZjWNDLglU,23947
|
|
181
192
|
optimum/rbln/transformers/models/pegasus/__init__.py,sha256=hXKIvrY0OMe7o2qiOFdolim7yyMuVQrUma_C39YsTqM,848
|
|
182
193
|
optimum/rbln/transformers/models/pegasus/configuration_pegasus.py,sha256=FufYVCN_MrMH24RRCh6MJWlS2e380tG13Oua-dyginY,1478
|
|
183
194
|
optimum/rbln/transformers/models/pegasus/modeling_pegasus.py,sha256=L8xvVwecRqXiF7xoj-1WgSgLBSCSllw1YRdgZhg6kp0,2571
|
|
184
|
-
optimum/rbln/transformers/models/pegasus/pegasus_architecture.py,sha256=
|
|
195
|
+
optimum/rbln/transformers/models/pegasus/pegasus_architecture.py,sha256=GV647yxMP7awSks-nrtyvg2sMg9FKGU9N8RPDsYZMRk,6059
|
|
185
196
|
optimum/rbln/transformers/models/phi/__init__.py,sha256=M5Sh4AtIhJYegl-yAKPggAU3DtJtQOa8MrIQypZ6N7U,734
|
|
186
197
|
optimum/rbln/transformers/models/phi/configuration_phi.py,sha256=CXHIG3xlBdr628oDu_u4OGsu_QZLx5EUSqu3zfmfEnk,1553
|
|
187
198
|
optimum/rbln/transformers/models/phi/modeling_phi.py,sha256=r7B0NlqwIGjm-MmE-h5_xeRJPzs4O2OotgbjI-FYA2o,3403
|
|
188
|
-
optimum/rbln/transformers/models/phi/phi_architecture.py,sha256=
|
|
199
|
+
optimum/rbln/transformers/models/phi/phi_architecture.py,sha256=nGnTNHzzJCku7tsCMyNmyDTGAuHYIHbaarBRWLnJf_8,3666
|
|
189
200
|
optimum/rbln/transformers/models/pixtral/__init__.py,sha256=fhclVAWnIDsfMfC-TW6mYrJXxgyehlLaadK64LOShH4,716
|
|
190
201
|
optimum/rbln/transformers/models/pixtral/configuration_pixtral.py,sha256=b79zkJB1jzHx4S1wTe-Ju_Yel_PS5Q8bfmlQPzkchKU,1677
|
|
191
202
|
optimum/rbln/transformers/models/pixtral/modeling_pixtral.py,sha256=2zIm5zFbuEi-O0QCawzv0AOeukXo3JWN3YKuj6zlUWU,13189
|
|
@@ -194,18 +205,26 @@ optimum/rbln/transformers/models/qwen2/__init__.py,sha256=h9dWJ3HX4xspMLt44g7r3U
|
|
|
194
205
|
optimum/rbln/transformers/models/qwen2/configuration_qwen2.py,sha256=tTWcPOk_ycZvdSPlal9S5elTmWZAX2BbpZP5Ok2ySwI,1567
|
|
195
206
|
optimum/rbln/transformers/models/qwen2/modeling_qwen2.py,sha256=z0tegK-B7tC0ocTLUp_V4eNmwREa12wJ7qQpbnRLEFo,3477
|
|
196
207
|
optimum/rbln/transformers/models/qwen2/qwen2_architecture.py,sha256=XlNAMYAcDLohnSAhIFGKOPuCB5XLgzYs5ABWdeQSaZs,720
|
|
197
|
-
optimum/rbln/transformers/models/qwen2_5_vl/__init__.py,sha256=
|
|
198
|
-
optimum/rbln/transformers/models/qwen2_5_vl/configuration_qwen2_5_vl.py,sha256=
|
|
199
|
-
optimum/rbln/transformers/models/qwen2_5_vl/modeling_qwen2_5_vl.py,sha256=
|
|
200
|
-
optimum/rbln/transformers/models/qwen2_5_vl/qwen2_5_vl_architecture.py,sha256=
|
|
201
|
-
optimum/rbln/transformers/models/
|
|
202
|
-
optimum/rbln/transformers/models/
|
|
203
|
-
optimum/rbln/transformers/models/
|
|
204
|
-
optimum/rbln/transformers/models/
|
|
208
|
+
optimum/rbln/transformers/models/qwen2_5_vl/__init__.py,sha256=_YnG4LxYdKLW61tSM3__DNB1PfHov8TgR8cTCJ285F4,934
|
|
209
|
+
optimum/rbln/transformers/models/qwen2_5_vl/configuration_qwen2_5_vl.py,sha256=kz9_k5W9LOOBpfk88COuCCzNPESMQYTLcuX5r855NFg,6515
|
|
210
|
+
optimum/rbln/transformers/models/qwen2_5_vl/modeling_qwen2_5_vl.py,sha256=fyDVCXlWbuVZtHvGkWdpHNJkH0REvNqVM-I3uKQ8Mas,32708
|
|
211
|
+
optimum/rbln/transformers/models/qwen2_5_vl/qwen2_5_vl_architecture.py,sha256=v4ejtIrbcCIHXBfkh7LD41z6Q_kzdtWpQSwYKkkxjks,8784
|
|
212
|
+
optimum/rbln/transformers/models/qwen2_moe/__init__.py,sha256=2a_ww3_9yBJnni9Sw_AudPsok0qPT37L4R-fdMWzHvI,722
|
|
213
|
+
optimum/rbln/transformers/models/qwen2_moe/configuration_qwen2_moe.py,sha256=SDkmCC7n8K14cPTW8Mkoc2z2Ac-hNnh4PFSYVrNC34o,1383
|
|
214
|
+
optimum/rbln/transformers/models/qwen2_moe/modeling_qwen2_moe.py,sha256=XmAkP1TjaO1QbZsCIHLjgXH_eCeLITSoY97eQ05zCoA,3114
|
|
215
|
+
optimum/rbln/transformers/models/qwen2_moe/qwen2_moe_architecture.py,sha256=Pux0PzBkTQuvQVgqDGsXp6RG0sFNQxTK_co8-Adyzl4,3985
|
|
216
|
+
optimum/rbln/transformers/models/qwen2_vl/__init__.py,sha256=KPL7ciTaA9n_6PFuOcvMwUnZLTLF_hDYGVGf3q8LNRI,912
|
|
217
|
+
optimum/rbln/transformers/models/qwen2_vl/configuration_qwen2_vl.py,sha256=FXsGTfKW7VV1ekAIvVxYUZ9iza06bUWGz7k_plyJaSE,5072
|
|
218
|
+
optimum/rbln/transformers/models/qwen2_vl/modeling_qwen2_vl.py,sha256=5N6PWVgK-3ahtPfpmF-AwtpSVQJCbLaRYnokneezXdo,26838
|
|
219
|
+
optimum/rbln/transformers/models/qwen2_vl/qwen2_vl_architecture.py,sha256=B0EitNdFhT5CO6TK1uLyPMLUoa99b7vM6pGhi2Rmf-c,5879
|
|
205
220
|
optimum/rbln/transformers/models/qwen3/__init__.py,sha256=tI4KwvXpD35dUUaa8aLUXpWoU9gJGcmKXeywOlH14ZE,746
|
|
206
221
|
optimum/rbln/transformers/models/qwen3/configuration_qwen3.py,sha256=BFRPggnH4VlsXlOa19C6KAID-bPgQ8ooQ29dvogh5zk,2102
|
|
207
222
|
optimum/rbln/transformers/models/qwen3/modeling_qwen3.py,sha256=jOg1Oqefi88rpcn6P2GXL7JDnYl_AjfI63xMXwuWttQ,4888
|
|
208
|
-
optimum/rbln/transformers/models/qwen3/qwen3_architecture.py,sha256=
|
|
223
|
+
optimum/rbln/transformers/models/qwen3/qwen3_architecture.py,sha256=m_pZ1wnh6ib8PhabWbo_n8x-G5hVP9h9iUC1JgEabNA,1119
|
|
224
|
+
optimum/rbln/transformers/models/qwen3_moe/__init__.py,sha256=3Hh4LDyNM388a8VZ68IJJcBWrvJ7TL8Akdax1RE0RBM,722
|
|
225
|
+
optimum/rbln/transformers/models/qwen3_moe/configuration_qwen3_moe.py,sha256=0UJ-xtgFzjyQUj96O43LAdWDLy2M44Blb9jdQ-ROETY,1395
|
|
226
|
+
optimum/rbln/transformers/models/qwen3_moe/modeling_qwen3_moe.py,sha256=AV1EGDrvMFqM_xgj7tFe9ZoAJFsGdmg9eP0wg7m6L4A,3149
|
|
227
|
+
optimum/rbln/transformers/models/qwen3_moe/qwen3_moe_architecture.py,sha256=4KiG9uqWDF8ljSbeq1qQHIZzEs9z05dg1dztGHhpxis,3985
|
|
209
228
|
optimum/rbln/transformers/models/resnet/__init__.py,sha256=0QqtEQF1IMYgEmmfXMGarCDS8kJB5tzODfwTEzDVZRg,837
|
|
210
229
|
optimum/rbln/transformers/models/resnet/configuration_resnet.py,sha256=T2CDlq-oGmT2LYf0J80X_h4WNxdWrNIgGufGDV55Pf0,1750
|
|
211
230
|
optimum/rbln/transformers/models/resnet/modeling_resnet.py,sha256=tAVa1r-yZMw56xEBtQUKMaUM0u1Zr5mg1uqXtr5u4gE,4419
|
|
@@ -215,21 +234,21 @@ optimum/rbln/transformers/models/roberta/modeling_roberta.py,sha256=1ybyReE9EB--
|
|
|
215
234
|
optimum/rbln/transformers/models/seq2seq/__init__.py,sha256=HiSyWFcKeZ8okfo-s-_Mf_upyvAoZwraUIJyGNLNurY,714
|
|
216
235
|
optimum/rbln/transformers/models/seq2seq/configuration_seq2seq.py,sha256=SBIFHxsDce2_s3laDBLa21l7minrTh6ZWSyhq1vXLa0,3060
|
|
217
236
|
optimum/rbln/transformers/models/seq2seq/modeling_seq2seq.py,sha256=9sVR5IqEoN9H0nfG8dk89PDhuaPTRhinO2ZswWyQbn4,20128
|
|
218
|
-
optimum/rbln/transformers/models/seq2seq/seq2seq_architecture.py,sha256=
|
|
237
|
+
optimum/rbln/transformers/models/seq2seq/seq2seq_architecture.py,sha256=bKhh8hSQ4q9qo0yi7zFsEvIlxpdhEcAUGcuCo2DORBM,20193
|
|
219
238
|
optimum/rbln/transformers/models/siglip/__init__.py,sha256=X1Fc1GUnJ2EIxFx45nbeoW-T2t0OyP3W73C0HD8Vowo,712
|
|
220
239
|
optimum/rbln/transformers/models/siglip/configuration_siglip.py,sha256=Fy-ANF91bQno_QVd4ZpyRs-uNgC_XRyBRScBg2uKM6w,3029
|
|
221
|
-
optimum/rbln/transformers/models/siglip/modeling_siglip.py,sha256=
|
|
240
|
+
optimum/rbln/transformers/models/siglip/modeling_siglip.py,sha256=2Fyf8u-iWmJZmRGB9cFPIQhud_QKBd5oRRm5Zl2T6W4,7988
|
|
222
241
|
optimum/rbln/transformers/models/swin/__init__.py,sha256=gUsLDB8ceNxt53Cf69OT32JuZoRdmmIsRfjRdHTLDd0,698
|
|
223
|
-
optimum/rbln/transformers/models/swin/configuration_swin.py,sha256=
|
|
242
|
+
optimum/rbln/transformers/models/swin/configuration_swin.py,sha256=16PNp37KVemOGyuFaVXNWBi-JwEnRNpvfTKS7HmAot0,1490
|
|
224
243
|
optimum/rbln/transformers/models/swin/modeling_swin.py,sha256=SAsHPme9izpKUKrVbxN8HSbOID2jRo3DhEthoV2jIyI,14999
|
|
225
244
|
optimum/rbln/transformers/models/t5/__init__.py,sha256=R1Q8Z1vaIdx4rDjeCmm_ZMSgewWaqaI0l93AHwewtew,818
|
|
226
245
|
optimum/rbln/transformers/models/t5/configuration_t5.py,sha256=nqDbibqykeeWn1TlKk6LmCn-DawTVudMMuBn2c2jds8,1362
|
|
227
246
|
optimum/rbln/transformers/models/t5/modeling_t5.py,sha256=lP__icG548arC9N4FHKfV7PQTpaqT7RpaHO1Tuvq3Ds,5125
|
|
228
|
-
optimum/rbln/transformers/models/t5/t5_architecture.py,sha256=
|
|
247
|
+
optimum/rbln/transformers/models/t5/t5_architecture.py,sha256=_iXpRd7i2coO0pJbcMKJ7eGdwgbTHeZP0VJ8KAagI-Y,10061
|
|
229
248
|
optimum/rbln/transformers/models/time_series_transformer/__init__.py,sha256=xJaFWQawlwtv4H5tVFcY1pxLYzjHtMAlLq6nXysdkN8,1243
|
|
230
249
|
optimum/rbln/transformers/models/time_series_transformer/configuration_time_series_transformer.py,sha256=EUBXE_10W0wtuoAl2OVuQakBpsC7kSpRo3VokXI8Pdo,1619
|
|
231
250
|
optimum/rbln/transformers/models/time_series_transformer/modeling_time_series_transformer.py,sha256=4rrivdPuuyg2MnfLaey_inai_sYE83GLg0nqN4Y85dg,18783
|
|
232
|
-
optimum/rbln/transformers/models/time_series_transformer/time_series_transformers_architecture.py,sha256=
|
|
251
|
+
optimum/rbln/transformers/models/time_series_transformer/time_series_transformers_architecture.py,sha256=SG9K7Pk7YNHWOX2CruWe5IWmBHBU88zuUJ87iZg1WpE,13926
|
|
233
252
|
optimum/rbln/transformers/models/vit/__init__.py,sha256=CrrkHehfCe3U-_rUS00aMBY7Tncdeh43sNUgVI9Dt_g,807
|
|
234
253
|
optimum/rbln/transformers/models/vit/configuration_vit.py,sha256=x98CxKR1cpKAG7Eh43uuPeGeGn4gS3HcKLPoDL3SWJo,994
|
|
235
254
|
optimum/rbln/transformers/models/vit/modeling_vit.py,sha256=49P3b8Q7qhfYxVyJt3XzjE0UNSODqHZQTjFhw5rbVzM,1777
|
|
@@ -240,25 +259,25 @@ optimum/rbln/transformers/models/whisper/__init__.py,sha256=ErquiUlYycSYPsDcq9Iw
|
|
|
240
259
|
optimum/rbln/transformers/models/whisper/configuration_whisper.py,sha256=bSwDN7VLuk1aVXvfrQIgb9SLdFBDhO5q8ZFaPQPJal0,3077
|
|
241
260
|
optimum/rbln/transformers/models/whisper/generation_whisper.py,sha256=Ts9g_i2oiWJ_eQAhYF9fW84_T-HwsWidhcI0Qpwq6aw,7827
|
|
242
261
|
optimum/rbln/transformers/models/whisper/modeling_whisper.py,sha256=k3kiy5EtDAzoVRVhWVjRbcgk1K4-MFzgZLkWDxI1fZw,19325
|
|
243
|
-
optimum/rbln/transformers/models/whisper/whisper_architecture.py,sha256=
|
|
262
|
+
optimum/rbln/transformers/models/whisper/whisper_architecture.py,sha256=mmnZtj_cg6fn1FAP6UgsdcAX4ugsdrgI5bNxNd3UUc0,13869
|
|
244
263
|
optimum/rbln/transformers/models/xlm_roberta/__init__.py,sha256=O3o2KzJ8Li3QhB7GHdRQASc93SYO2jz00Rx4pxYRuDg,982
|
|
245
264
|
optimum/rbln/transformers/models/xlm_roberta/configuration_xlm_roberta.py,sha256=wHRpGTXL9khYqSkKL1IgA7__6_lt9QpOz9tHumjK7fo,1260
|
|
246
265
|
optimum/rbln/transformers/models/xlm_roberta/modeling_xlm_roberta.py,sha256=Pjqvfp0V0kFBW7U7VD3kthZkpWfzOKsRS_2-z6StTnI,3710
|
|
247
266
|
optimum/rbln/transformers/utils/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
248
|
-
optimum/rbln/transformers/utils/rbln_quantization.py,sha256=
|
|
267
|
+
optimum/rbln/transformers/utils/rbln_quantization.py,sha256=Qdz3jwl4tsocA6chKbuiG45_veeuU-HM_9mHeu8w4LQ,22323
|
|
249
268
|
optimum/rbln/transformers/utils/rbln_runtime_wrapper.py,sha256=l_-zWpRrp6hp-tDANTrEbspIZH-AUSi_jNJICns_QgE,2672
|
|
250
269
|
optimum/rbln/utils/__init__.py,sha256=ieDBT2VFTt2E0M4v_POLBpuGW9LxSydpb_DuPd6PQqc,712
|
|
251
270
|
optimum/rbln/utils/decorator_utils.py,sha256=xu-TrsNi33SRC2a7DBsyoo6-pEQxWKZPZSmM9QlDe2Y,3745
|
|
252
271
|
optimum/rbln/utils/deprecation.py,sha256=qO6xlrT_GNCOCJx4i28t8Q-1hDGwp-cJMC5OrD7lUOQ,13226
|
|
253
272
|
optimum/rbln/utils/hub.py,sha256=EI2ZsD71jhmPaA1imJ2_7P6y8i2uoX5l6wya5fICdQA,3119
|
|
254
|
-
optimum/rbln/utils/import_utils.py,sha256=
|
|
273
|
+
optimum/rbln/utils/import_utils.py,sha256=ZAp6t5vwUVS9yiFxQYarTLO0v-WqPj75QG5QodzCa2E,6306
|
|
255
274
|
optimum/rbln/utils/logging.py,sha256=VKKBmlQSdg6iZCGmAXaWYiW67K84jyp1QJhLQSSjPPE,3453
|
|
256
275
|
optimum/rbln/utils/model_utils.py,sha256=4k5879Kh75m3x_vS4-qOGfqsOiAvc2kdNFFfvsFvz3k,1748
|
|
257
|
-
optimum/rbln/utils/runtime_utils.py,sha256=
|
|
276
|
+
optimum/rbln/utils/runtime_utils.py,sha256=p8ZMqIjQsMPncA-_qI4RjRo8Hp7CBB2lMGx_HWb7kxg,9602
|
|
258
277
|
optimum/rbln/utils/save_utils.py,sha256=hG5uOtYmecSXZuGTvCXsTM-SiyZpr5q3InUGCCq_jzQ,3619
|
|
259
|
-
optimum/rbln/utils/submodule.py,sha256=
|
|
260
|
-
optimum_rbln-0.9.
|
|
261
|
-
optimum_rbln-0.9.
|
|
262
|
-
optimum_rbln-0.9.
|
|
263
|
-
optimum_rbln-0.9.
|
|
264
|
-
optimum_rbln-0.9.
|
|
278
|
+
optimum/rbln/utils/submodule.py,sha256=j-LxEbFpjixFWLRnSBJC-IC3bGSIl7L_hs3gwaqjK1Q,7503
|
|
279
|
+
optimum_rbln-0.9.5a4.dist-info/METADATA,sha256=bixbeCg-vxyYrHd3Gv0hzmpzsuBmwj98yYYug0F7LFo,5326
|
|
280
|
+
optimum_rbln-0.9.5a4.dist-info/WHEEL,sha256=WLgqFyCfm_KASv4WHyYy0P3pM_m7J5L9k2skdKLirC8,87
|
|
281
|
+
optimum_rbln-0.9.5a4.dist-info/entry_points.txt,sha256=-orKDGKfLypxlPlTz8-ZkmdKULNvax9yeCCCn-q89n4,59
|
|
282
|
+
optimum_rbln-0.9.5a4.dist-info/licenses/LICENSE,sha256=QwcOLU5TJoTeUhuIXzhdCEEDDvorGiC6-3YTOl4TecE,11356
|
|
283
|
+
optimum_rbln-0.9.5a4.dist-info/RECORD,,
|
|
@@ -1,233 +0,0 @@
|
|
|
1
|
-
# Copyright 2025 Rebellions Inc. All rights reserved.
|
|
2
|
-
|
|
3
|
-
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
4
|
-
# you may not use this file except in compliance with the License.
|
|
5
|
-
# You may obtain a copy of the License at:
|
|
6
|
-
|
|
7
|
-
# http://www.apache.org/licenses/LICENSE-2.0
|
|
8
|
-
|
|
9
|
-
# Unless required by applicable law or agreed to in writing, software
|
|
10
|
-
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
11
|
-
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
12
|
-
# See the License for the specific language governing permissions and
|
|
13
|
-
# limitations under the License.
|
|
14
|
-
|
|
15
|
-
from typing import List, Optional, Tuple, Union
|
|
16
|
-
|
|
17
|
-
import torch
|
|
18
|
-
import torch.nn as nn
|
|
19
|
-
from transformers import PreTrainedModel
|
|
20
|
-
|
|
21
|
-
from optimum.rbln.transformers.models.decoderonly.decoderonly_architecture import (
|
|
22
|
-
DecoderOnlyLayer,
|
|
23
|
-
DecoderOnlyModel,
|
|
24
|
-
DecoderOnlyWrapper,
|
|
25
|
-
)
|
|
26
|
-
|
|
27
|
-
from .configuration_colqwen2 import (
|
|
28
|
-
RBLNColQwen2ForRetrievalConfig,
|
|
29
|
-
)
|
|
30
|
-
|
|
31
|
-
|
|
32
|
-
def slice_and_unsqueeze_cos_sin(cos, sin, position_ids):
|
|
33
|
-
"""Slice cos[cache_position], sin[cache_position] vector for the query."""
|
|
34
|
-
cos = cos[position_ids[0]][None, None, None, :, :]
|
|
35
|
-
sin = sin[position_ids[0]][None, None, None, :, :]
|
|
36
|
-
|
|
37
|
-
return cos, sin
|
|
38
|
-
|
|
39
|
-
|
|
40
|
-
class ColQwen2LanguageModelWrapper(DecoderOnlyWrapper):
|
|
41
|
-
def __init__(
|
|
42
|
-
self, model: PreTrainedModel, rbln_config: "RBLNColQwen2ForRetrievalConfig", use_rotary_emb: bool = True
|
|
43
|
-
):
|
|
44
|
-
model.config = (
|
|
45
|
-
model.config.vlm_config.text_config if hasattr(model.config, "vlm_config") else model.config.text_config
|
|
46
|
-
)
|
|
47
|
-
super().__init__(model, rbln_config, use_rotary_emb)
|
|
48
|
-
|
|
49
|
-
def get_decoder_layers(self, model: PreTrainedModel):
|
|
50
|
-
return model.language_model.layers
|
|
51
|
-
|
|
52
|
-
def convert_to_rbln_class(self, model: PreTrainedModel, max_seq_len: int):
|
|
53
|
-
new_layers = []
|
|
54
|
-
for layer_idx, layer in enumerate(self.get_decoder_layers(model)):
|
|
55
|
-
is_sliding = layer_idx in self.rbln_config.sliding_window_layers
|
|
56
|
-
new_self_attn = self.get_rbln_attn_class()(
|
|
57
|
-
self.get_attn_layer(layer),
|
|
58
|
-
self.rbln_config,
|
|
59
|
-
is_sliding=is_sliding,
|
|
60
|
-
)
|
|
61
|
-
new_layer = self.get_rbln_layer_class()(layer, new_self_attn)
|
|
62
|
-
new_layers.append(new_layer)
|
|
63
|
-
|
|
64
|
-
new_model = self.get_rbln_model_class()(
|
|
65
|
-
model.language_model,
|
|
66
|
-
new_layers,
|
|
67
|
-
self.rbln_config,
|
|
68
|
-
use_learned_pos_emb=self.__class__._use_learned_pos_emb,
|
|
69
|
-
)
|
|
70
|
-
|
|
71
|
-
# text_projection layer from model
|
|
72
|
-
self.embedding_proj_layer = (
|
|
73
|
-
model.embedding_proj_layer if hasattr(model, "embedding_proj_layer") else model.custom_text_proj
|
|
74
|
-
)
|
|
75
|
-
return new_model
|
|
76
|
-
|
|
77
|
-
def get_rbln_model_class(self):
|
|
78
|
-
return RBLNColQwen2LanguageModel
|
|
79
|
-
|
|
80
|
-
def prepare_forward_args(self, *args):
|
|
81
|
-
args = list(args)
|
|
82
|
-
input_ids = None if self.rbln_config.use_inputs_embeds else args.pop(0)
|
|
83
|
-
inputs_embeds = args.pop(0) if self.rbln_config.use_inputs_embeds else None
|
|
84
|
-
cache_position = args.pop(0)
|
|
85
|
-
global_block_tables = args.pop(0)
|
|
86
|
-
local_block_tables = None
|
|
87
|
-
position_embeds = args.pop(0)
|
|
88
|
-
position_ids = None
|
|
89
|
-
attention_mask = args.pop(0) if self.rbln_config.use_attention_mask else None
|
|
90
|
-
past_key_values = args
|
|
91
|
-
|
|
92
|
-
if len(past_key_values) != 2 * self.num_hidden_layers:
|
|
93
|
-
raise ValueError(
|
|
94
|
-
f"Different past_key_values to model's config. {len(past_key_values)} != {2 * self.num_hidden_layers}"
|
|
95
|
-
)
|
|
96
|
-
|
|
97
|
-
_past_key_values = []
|
|
98
|
-
for i in range(self.config.num_hidden_layers):
|
|
99
|
-
key_states = past_key_values[i * 2]
|
|
100
|
-
value_states = past_key_values[i * 2 + 1]
|
|
101
|
-
past_key_value = [key_states, value_states]
|
|
102
|
-
_past_key_values.append(past_key_value)
|
|
103
|
-
past_key_values = _past_key_values
|
|
104
|
-
|
|
105
|
-
return (
|
|
106
|
-
input_ids,
|
|
107
|
-
inputs_embeds,
|
|
108
|
-
cache_position,
|
|
109
|
-
global_block_tables,
|
|
110
|
-
local_block_tables,
|
|
111
|
-
attention_mask,
|
|
112
|
-
position_ids,
|
|
113
|
-
past_key_values,
|
|
114
|
-
position_embeds,
|
|
115
|
-
)
|
|
116
|
-
|
|
117
|
-
def forward(self, *args):
|
|
118
|
-
(
|
|
119
|
-
input_ids,
|
|
120
|
-
inputs_embeds,
|
|
121
|
-
cache_position,
|
|
122
|
-
global_block_tables,
|
|
123
|
-
local_block_tables,
|
|
124
|
-
attention_mask,
|
|
125
|
-
position_ids,
|
|
126
|
-
past_key_values,
|
|
127
|
-
rotary_emb,
|
|
128
|
-
) = self.prepare_forward_args(*args)
|
|
129
|
-
|
|
130
|
-
last_hidden_states = self.model(
|
|
131
|
-
input_ids=input_ids,
|
|
132
|
-
inputs_embeds=inputs_embeds,
|
|
133
|
-
attention_mask=attention_mask,
|
|
134
|
-
cache_position=cache_position,
|
|
135
|
-
position_ids=position_ids,
|
|
136
|
-
past_key_values=past_key_values,
|
|
137
|
-
rotary_emb=rotary_emb,
|
|
138
|
-
global_block_tables=global_block_tables,
|
|
139
|
-
local_block_tables=local_block_tables,
|
|
140
|
-
)
|
|
141
|
-
|
|
142
|
-
proj = self.embedding_proj_layer(last_hidden_states[0])
|
|
143
|
-
all_hidden_states = last_hidden_states[1] if self.rbln_config.output_hidden_states else None
|
|
144
|
-
|
|
145
|
-
if self.rbln_config.output_hidden_states:
|
|
146
|
-
return proj, all_hidden_states
|
|
147
|
-
else:
|
|
148
|
-
return proj
|
|
149
|
-
|
|
150
|
-
|
|
151
|
-
class RBLNColQwen2LanguageModel(DecoderOnlyModel):
|
|
152
|
-
def __init__(
|
|
153
|
-
self,
|
|
154
|
-
model,
|
|
155
|
-
layers: List["DecoderOnlyLayer"],
|
|
156
|
-
rbln_config: "RBLNColQwen2ForRetrievalConfig",
|
|
157
|
-
use_learned_pos_emb=None,
|
|
158
|
-
):
|
|
159
|
-
super().__init__(model, layers, rbln_config, use_learned_pos_emb)
|
|
160
|
-
|
|
161
|
-
self.output_hidden_states = rbln_config.output_hidden_states
|
|
162
|
-
|
|
163
|
-
def forward(
|
|
164
|
-
self,
|
|
165
|
-
input_ids: torch.Tensor = None,
|
|
166
|
-
inputs_embeds: Optional[torch.Tensor] = None,
|
|
167
|
-
attention_mask: torch.Tensor = None,
|
|
168
|
-
cache_position: torch.Tensor = None,
|
|
169
|
-
position_ids: torch.Tensor = None,
|
|
170
|
-
query_position: torch.Tensor = None,
|
|
171
|
-
past_key_values: Tuple[Tuple[torch.Tensor]] = None,
|
|
172
|
-
rotary_emb: Optional[Union[nn.Module, torch.Tensor]] = None,
|
|
173
|
-
global_block_tables: Optional[torch.Tensor] = None,
|
|
174
|
-
local_block_tables: Optional[torch.Tensor] = None,
|
|
175
|
-
lora_int_id: Optional[torch.Tensor] = None,
|
|
176
|
-
):
|
|
177
|
-
# retrieve input_ids and inputs_embeds
|
|
178
|
-
if (input_ids is None) ^ (inputs_embeds is not None):
|
|
179
|
-
raise ValueError(
|
|
180
|
-
"You cannot specify both input_ids and inputs_embeds at the same time, and must specify either one"
|
|
181
|
-
)
|
|
182
|
-
|
|
183
|
-
# embed positions
|
|
184
|
-
if inputs_embeds is None:
|
|
185
|
-
inputs_embeds = self.get_embedding()(input_ids)
|
|
186
|
-
|
|
187
|
-
hidden_states = inputs_embeds * self.hidden_multiplier
|
|
188
|
-
|
|
189
|
-
# get cos,sin vector if needed
|
|
190
|
-
position_ids = position_ids if position_ids is not None else cache_position
|
|
191
|
-
if rotary_emb is not None:
|
|
192
|
-
if isinstance(rotary_emb, torch.Tensor):
|
|
193
|
-
cos = rotary_emb[0]
|
|
194
|
-
sin = rotary_emb[1]
|
|
195
|
-
else:
|
|
196
|
-
cos, sin = rotary_emb(hidden_states, self.max_seq_len) # dtype carrier, max_seq_len
|
|
197
|
-
cos, sin = slice_and_unsqueeze_cos_sin(cos, sin, position_ids)
|
|
198
|
-
|
|
199
|
-
# Get sequence positions for flash attention
|
|
200
|
-
if self.attn_impl == "flash_attn":
|
|
201
|
-
seq_positions = cache_position[:, 0]
|
|
202
|
-
seq_positions = self.convert_sequence_positions_for_flash_attn(
|
|
203
|
-
seq_positions=seq_positions, max_seq_len=self.max_seq_len
|
|
204
|
-
)
|
|
205
|
-
else:
|
|
206
|
-
seq_positions = cache_position[:, :1]
|
|
207
|
-
|
|
208
|
-
# Get local cache positions for sliding window layers
|
|
209
|
-
if len(self.sliding_window_layers) > 0:
|
|
210
|
-
sliding_cache_pos = self.get_local_cache_positions(position_ids, query_position)
|
|
211
|
-
|
|
212
|
-
all_hidden_states = () if self.output_hidden_states else None
|
|
213
|
-
for layer_idx, layer in enumerate(self.layers):
|
|
214
|
-
if self.output_hidden_states:
|
|
215
|
-
all_hidden_states += (hidden_states,)
|
|
216
|
-
|
|
217
|
-
is_sliding = True if layer_idx in self.sliding_window_layers else False
|
|
218
|
-
hidden_states = layer(
|
|
219
|
-
hidden_states=hidden_states,
|
|
220
|
-
attention_mask=attention_mask,
|
|
221
|
-
seq_positions=sliding_cache_pos if is_sliding else seq_positions,
|
|
222
|
-
past_key_values=past_key_values,
|
|
223
|
-
cos=cos,
|
|
224
|
-
sin=sin,
|
|
225
|
-
block_tables=local_block_tables if is_sliding else global_block_tables,
|
|
226
|
-
lora_int_id=lora_int_id,
|
|
227
|
-
)
|
|
228
|
-
|
|
229
|
-
hidden_states = self.get_last_layernorm()(hidden_states)
|
|
230
|
-
if self.output_hidden_states:
|
|
231
|
-
all_hidden_states += (hidden_states,)
|
|
232
|
-
|
|
233
|
-
return hidden_states, all_hidden_states
|
|
File without changes
|
|
File without changes
|
|
File without changes
|