optimum-rbln 0.7.5rc2__py3-none-any.whl → 0.8.0__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -17,5 +17,5 @@ __version__: str
17
17
  __version_tuple__: VERSION_TUPLE
18
18
  version_tuple: VERSION_TUPLE
19
19
 
20
- __version__ = version = '0.7.5rc2'
21
- __version_tuple__ = version_tuple = (0, 7, 5, 'rc2')
20
+ __version__ = version = '0.8.0'
21
+ __version_tuple__ = version_tuple = (0, 8, 0)
@@ -11,9 +11,10 @@
11
11
  # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12
12
  # See the License for the specific language governing permissions and
13
13
  # limitations under the License.
14
-
15
14
  from typing import Optional
16
15
 
16
+ import rebel
17
+
17
18
  from ....configuration_utils import RBLNModelConfig
18
19
  from ..decoderonly.configuration_decoderonly import RBLNDecoderOnlyModelForCausalLMConfig
19
20
  from ..siglip.configuration_siglip import RBLNSiglipVisionModelConfig
@@ -39,6 +40,10 @@ class RBLNGemma3ForCausalLMConfig(RBLNDecoderOnlyModelForCausalLMConfig):
39
40
  **kwargs,
40
41
  )
41
42
 
43
+ npu = self.npu or rebel.get_npu_name()
44
+ if npu == "RBLN-CA02":
45
+ raise NotImplementedError("Gemma3 is currently not supported on RBLN-CA02")
46
+
42
47
 
43
48
  class RBLNGemma3ForConditionalGenerationConfig(RBLNModelConfig):
44
49
  submodules = ["vision_tower", "language_model"]
@@ -858,11 +858,6 @@ class RBLNGemma3ForCausalLM(RBLNDecoderOnlyModelForCausalLM):
858
858
  max_seq_len=rbln_config.max_seq_len,
859
859
  )
860
860
 
861
- if rbln_config.attn_impl == "eager":
862
- raise ValueError(
863
- "Eager attention is not supported for Gemma3. Please use flash attention by setting `rbln_attn_impl='flash_attn'`. Stay tuned for future updates!"
864
- )
865
-
866
861
  validate_attention_method(
867
862
  attn_impl=rbln_config.attn_impl,
868
863
  kvcache_partition_len=rbln_config.kvcache_partition_len,
@@ -595,9 +595,8 @@ class RBLNQwen2_5_VLForConditionalGeneration(RBLNDecoderOnlyModelForCausalLM):
595
595
  )
596
596
  logits.append(output.logits)
597
597
  logits = torch.cat(logits, dim=0)
598
- # Decoder
598
+ # Decoder
599
599
  else:
600
- print(input_ids[0], cache_position[0])
601
600
  inputs_embeds, position_embed = self._preprocess_decoder(input_ids, cache_position)
602
601
  output = self.decoder(
603
602
  inputs_embeds=inputs_embeds,
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: optimum-rbln
3
- Version: 0.7.5rc2
3
+ Version: 0.8.0
4
4
  Summary: Optimum RBLN is the interface between the HuggingFace Transformers and Diffusers libraries and RBLN accelerators. It provides a set of tools enabling easy model loading and inference on single and multiple rbln device settings for different downstream tasks.
5
5
  Project-URL: Homepage, https://rebellions.ai
6
6
  Project-URL: Documentation, https://docs.rbln.ai
@@ -1,5 +1,5 @@
1
1
  optimum/rbln/__init__.py,sha256=oAnsJSMrPYwBGEttUt3CMXTIESVNe15ftTWRTShwhZI,14386
2
- optimum/rbln/__version__.py,sha256=fpFaQLT4vGQYujVJTSb1WZo3X-GKEGeYInrc_bpJrpQ,521
2
+ optimum/rbln/__version__.py,sha256=fSm5pLlwHxfTD7vBTVEqChJUua9ilUsdQYNN_V3u3iE,511
3
3
  optimum/rbln/configuration_utils.py,sha256=gvAjRFEGw5rnSoH0IoyuLrE4fkxtk3DN1pikqrN_Rpk,31277
4
4
  optimum/rbln/modeling.py,sha256=BpydF-bLBF60NnRMbtZwn5odOUjU4Awu9azqGeSufTI,11462
5
5
  optimum/rbln/modeling_base.py,sha256=HQgscr5jpUEtuXU1ACJHSLIntX-kq6Ef0SQ_W2-rp5A,25341
@@ -100,9 +100,9 @@ optimum/rbln/transformers/models/gemma/configuration_gemma.py,sha256=ek7PiMWx4LD
100
100
  optimum/rbln/transformers/models/gemma/gemma_architecture.py,sha256=_Ip8J4IKEkXwEiynKvoj7NEYDR0eJ4IEWXqgjXibBy4,2447
101
101
  optimum/rbln/transformers/models/gemma/modeling_gemma.py,sha256=-U3w3cEOv3ps1S8aL7uOq6Kq2siCPZz7Z8MXhDQgQqo,1530
102
102
  optimum/rbln/transformers/models/gemma3/__init__.py,sha256=6rugk3615SEt4lh7gduo_J9VyGiSReuEIvL0Uno0eaI,790
103
- optimum/rbln/transformers/models/gemma3/configuration_gemma3.py,sha256=nndcYVwDYkOige_qO4td-YwLNtUz6aLiSQDIfPdGG9A,2840
103
+ optimum/rbln/transformers/models/gemma3/configuration_gemma3.py,sha256=RsowCftLZyluzuBwXsMMbHWOABslznMR7WY3rP2sAYw,3020
104
104
  optimum/rbln/transformers/models/gemma3/gemma3_architecture.py,sha256=Uer27wG06hgV1WNf92x1ZeUpl4Q0zskfCqzlLhtgtNU,17348
105
- optimum/rbln/transformers/models/gemma3/modeling_gemma3.py,sha256=mx3upghkboeyRGYxwPfA1fzRNPWj5MzX8bIy0kszHWY,45235
105
+ optimum/rbln/transformers/models/gemma3/modeling_gemma3.py,sha256=OUGsML-H6FOKldld7KRW9l0nRoT4DojWBDl8ZHpfXVA,44982
106
106
  optimum/rbln/transformers/models/gpt2/__init__.py,sha256=socBMIBZSiLbrVN12rQ4nL9gFeT0axMgz6SWaCaD4Ac,704
107
107
  optimum/rbln/transformers/models/gpt2/configuration_gpt2.py,sha256=vKvJD8P9Li4W9wdVoQcqMEr1MwEXojPBnF2NE85VXAo,772
108
108
  optimum/rbln/transformers/models/gpt2/gpt2_architecture.py,sha256=kf5jk7Djv9XXX3Q83oTosiMpt9g44TF_gCT-vMiWDJk,3097
@@ -139,7 +139,7 @@ optimum/rbln/transformers/models/qwen2/modeling_qwen2.py,sha256=9-aFDvjMzPNUyGOz
139
139
  optimum/rbln/transformers/models/qwen2/qwen2_architecture.py,sha256=XlNAMYAcDLohnSAhIFGKOPuCB5XLgzYs5ABWdeQSaZs,720
140
140
  optimum/rbln/transformers/models/qwen2_5_vl/__init__.py,sha256=rAW3DKQUzGL6EMwa5r1iLu94yhpiZpk6zfoD7TtYXrc,865
141
141
  optimum/rbln/transformers/models/qwen2_5_vl/configuration_qwen2_5_vl.py,sha256=dPcGNaLwJf61PIvVbyt-lvBflp_dvK0hubhNoA3len0,3123
142
- optimum/rbln/transformers/models/qwen2_5_vl/modeling_qwen2_5_vl.py,sha256=DUlTQc24IJwWMTIi3ou3F7eraYJy9WERjH7yAZo80u4,25172
142
+ optimum/rbln/transformers/models/qwen2_5_vl/modeling_qwen2_5_vl.py,sha256=pb4OC5rRG-Y-a8WMBGqrV0owgjeGH83Lqx2fXztnlJg,25117
143
143
  optimum/rbln/transformers/models/qwen2_5_vl/qwen2_5_vl_architecture.py,sha256=OR-tTu8uzeFryenohoqRwgIVqw5zM7dcsZHwKL_jD0A,7232
144
144
  optimum/rbln/transformers/models/seq2seq/__init__.py,sha256=6WKstWiS1kW0oFDn_jyrKMW5QEJAWkmsSRAaadNedDM,715
145
145
  optimum/rbln/transformers/models/seq2seq/configuration_seq2seq2.py,sha256=vSNP1eILfL32cbiLOAD58Ocz6lk3hYFnhIRLDVqlSoI,2624
@@ -178,7 +178,7 @@ optimum/rbln/utils/model_utils.py,sha256=V2kFpUe2aqVzLwbpztD8JOVFQqRHncvIWwJbgnU
178
178
  optimum/rbln/utils/runtime_utils.py,sha256=LoKNK3AQNV_BSScstIZWjICkJf265MnUgy360BOocVI,5454
179
179
  optimum/rbln/utils/save_utils.py,sha256=hG5uOtYmecSXZuGTvCXsTM-SiyZpr5q3InUGCCq_jzQ,3619
180
180
  optimum/rbln/utils/submodule.py,sha256=ZfI7e3YzbjbbBW4Yjfucj8NygEsukfIkaJi3PtwHrhc,5105
181
- optimum_rbln-0.7.5rc2.dist-info/METADATA,sha256=WltiXJxC_INRxeoxhT4xyvLOO8FM6cqJOqbpnTQ6xhg,5300
182
- optimum_rbln-0.7.5rc2.dist-info/WHEEL,sha256=qtCwoSJWgHk21S1Kb4ihdzI2rlJ1ZKaIurTj_ngOhyQ,87
183
- optimum_rbln-0.7.5rc2.dist-info/licenses/LICENSE,sha256=QwcOLU5TJoTeUhuIXzhdCEEDDvorGiC6-3YTOl4TecE,11356
184
- optimum_rbln-0.7.5rc2.dist-info/RECORD,,
181
+ optimum_rbln-0.8.0.dist-info/METADATA,sha256=o162FwSFWUcaXw4SoFtdnPtAp14nkmE4wjRbY4O3HEU,5297
182
+ optimum_rbln-0.8.0.dist-info/WHEEL,sha256=qtCwoSJWgHk21S1Kb4ihdzI2rlJ1ZKaIurTj_ngOhyQ,87
183
+ optimum_rbln-0.8.0.dist-info/licenses/LICENSE,sha256=QwcOLU5TJoTeUhuIXzhdCEEDDvorGiC6-3YTOl4TecE,11356
184
+ optimum_rbln-0.8.0.dist-info/RECORD,,