optimum-rbln 0.7.5rc2__py3-none-any.whl → 0.8.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- optimum/rbln/__version__.py +2 -2
- optimum/rbln/transformers/models/gemma3/configuration_gemma3.py +6 -1
- optimum/rbln/transformers/models/gemma3/modeling_gemma3.py +0 -5
- optimum/rbln/transformers/models/qwen2_5_vl/modeling_qwen2_5_vl.py +1 -2
- {optimum_rbln-0.7.5rc2.dist-info → optimum_rbln-0.8.0.dist-info}/METADATA +1 -1
- {optimum_rbln-0.7.5rc2.dist-info → optimum_rbln-0.8.0.dist-info}/RECORD +8 -8
- {optimum_rbln-0.7.5rc2.dist-info → optimum_rbln-0.8.0.dist-info}/WHEEL +0 -0
- {optimum_rbln-0.7.5rc2.dist-info → optimum_rbln-0.8.0.dist-info}/licenses/LICENSE +0 -0
optimum/rbln/__version__.py
CHANGED
@@ -17,5 +17,5 @@ __version__: str
|
|
17
17
|
__version_tuple__: VERSION_TUPLE
|
18
18
|
version_tuple: VERSION_TUPLE
|
19
19
|
|
20
|
-
__version__ = version = '0.
|
21
|
-
__version_tuple__ = version_tuple = (0,
|
20
|
+
__version__ = version = '0.8.0'
|
21
|
+
__version_tuple__ = version_tuple = (0, 8, 0)
|
@@ -11,9 +11,10 @@
|
|
11
11
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
12
12
|
# See the License for the specific language governing permissions and
|
13
13
|
# limitations under the License.
|
14
|
-
|
15
14
|
from typing import Optional
|
16
15
|
|
16
|
+
import rebel
|
17
|
+
|
17
18
|
from ....configuration_utils import RBLNModelConfig
|
18
19
|
from ..decoderonly.configuration_decoderonly import RBLNDecoderOnlyModelForCausalLMConfig
|
19
20
|
from ..siglip.configuration_siglip import RBLNSiglipVisionModelConfig
|
@@ -39,6 +40,10 @@ class RBLNGemma3ForCausalLMConfig(RBLNDecoderOnlyModelForCausalLMConfig):
|
|
39
40
|
**kwargs,
|
40
41
|
)
|
41
42
|
|
43
|
+
npu = self.npu or rebel.get_npu_name()
|
44
|
+
if npu == "RBLN-CA02":
|
45
|
+
raise NotImplementedError("Gemma3 is currently not supported on RBLN-CA02")
|
46
|
+
|
42
47
|
|
43
48
|
class RBLNGemma3ForConditionalGenerationConfig(RBLNModelConfig):
|
44
49
|
submodules = ["vision_tower", "language_model"]
|
@@ -858,11 +858,6 @@ class RBLNGemma3ForCausalLM(RBLNDecoderOnlyModelForCausalLM):
|
|
858
858
|
max_seq_len=rbln_config.max_seq_len,
|
859
859
|
)
|
860
860
|
|
861
|
-
if rbln_config.attn_impl == "eager":
|
862
|
-
raise ValueError(
|
863
|
-
"Eager attention is not supported for Gemma3. Please use flash attention by setting `rbln_attn_impl='flash_attn'`. Stay tuned for future updates!"
|
864
|
-
)
|
865
|
-
|
866
861
|
validate_attention_method(
|
867
862
|
attn_impl=rbln_config.attn_impl,
|
868
863
|
kvcache_partition_len=rbln_config.kvcache_partition_len,
|
@@ -595,9 +595,8 @@ class RBLNQwen2_5_VLForConditionalGeneration(RBLNDecoderOnlyModelForCausalLM):
|
|
595
595
|
)
|
596
596
|
logits.append(output.logits)
|
597
597
|
logits = torch.cat(logits, dim=0)
|
598
|
-
|
598
|
+
# Decoder
|
599
599
|
else:
|
600
|
-
print(input_ids[0], cache_position[0])
|
601
600
|
inputs_embeds, position_embed = self._preprocess_decoder(input_ids, cache_position)
|
602
601
|
output = self.decoder(
|
603
602
|
inputs_embeds=inputs_embeds,
|
@@ -1,6 +1,6 @@
|
|
1
1
|
Metadata-Version: 2.4
|
2
2
|
Name: optimum-rbln
|
3
|
-
Version: 0.
|
3
|
+
Version: 0.8.0
|
4
4
|
Summary: Optimum RBLN is the interface between the HuggingFace Transformers and Diffusers libraries and RBLN accelerators. It provides a set of tools enabling easy model loading and inference on single and multiple rbln device settings for different downstream tasks.
|
5
5
|
Project-URL: Homepage, https://rebellions.ai
|
6
6
|
Project-URL: Documentation, https://docs.rbln.ai
|
@@ -1,5 +1,5 @@
|
|
1
1
|
optimum/rbln/__init__.py,sha256=oAnsJSMrPYwBGEttUt3CMXTIESVNe15ftTWRTShwhZI,14386
|
2
|
-
optimum/rbln/__version__.py,sha256=
|
2
|
+
optimum/rbln/__version__.py,sha256=fSm5pLlwHxfTD7vBTVEqChJUua9ilUsdQYNN_V3u3iE,511
|
3
3
|
optimum/rbln/configuration_utils.py,sha256=gvAjRFEGw5rnSoH0IoyuLrE4fkxtk3DN1pikqrN_Rpk,31277
|
4
4
|
optimum/rbln/modeling.py,sha256=BpydF-bLBF60NnRMbtZwn5odOUjU4Awu9azqGeSufTI,11462
|
5
5
|
optimum/rbln/modeling_base.py,sha256=HQgscr5jpUEtuXU1ACJHSLIntX-kq6Ef0SQ_W2-rp5A,25341
|
@@ -100,9 +100,9 @@ optimum/rbln/transformers/models/gemma/configuration_gemma.py,sha256=ek7PiMWx4LD
|
|
100
100
|
optimum/rbln/transformers/models/gemma/gemma_architecture.py,sha256=_Ip8J4IKEkXwEiynKvoj7NEYDR0eJ4IEWXqgjXibBy4,2447
|
101
101
|
optimum/rbln/transformers/models/gemma/modeling_gemma.py,sha256=-U3w3cEOv3ps1S8aL7uOq6Kq2siCPZz7Z8MXhDQgQqo,1530
|
102
102
|
optimum/rbln/transformers/models/gemma3/__init__.py,sha256=6rugk3615SEt4lh7gduo_J9VyGiSReuEIvL0Uno0eaI,790
|
103
|
-
optimum/rbln/transformers/models/gemma3/configuration_gemma3.py,sha256=
|
103
|
+
optimum/rbln/transformers/models/gemma3/configuration_gemma3.py,sha256=RsowCftLZyluzuBwXsMMbHWOABslznMR7WY3rP2sAYw,3020
|
104
104
|
optimum/rbln/transformers/models/gemma3/gemma3_architecture.py,sha256=Uer27wG06hgV1WNf92x1ZeUpl4Q0zskfCqzlLhtgtNU,17348
|
105
|
-
optimum/rbln/transformers/models/gemma3/modeling_gemma3.py,sha256=
|
105
|
+
optimum/rbln/transformers/models/gemma3/modeling_gemma3.py,sha256=OUGsML-H6FOKldld7KRW9l0nRoT4DojWBDl8ZHpfXVA,44982
|
106
106
|
optimum/rbln/transformers/models/gpt2/__init__.py,sha256=socBMIBZSiLbrVN12rQ4nL9gFeT0axMgz6SWaCaD4Ac,704
|
107
107
|
optimum/rbln/transformers/models/gpt2/configuration_gpt2.py,sha256=vKvJD8P9Li4W9wdVoQcqMEr1MwEXojPBnF2NE85VXAo,772
|
108
108
|
optimum/rbln/transformers/models/gpt2/gpt2_architecture.py,sha256=kf5jk7Djv9XXX3Q83oTosiMpt9g44TF_gCT-vMiWDJk,3097
|
@@ -139,7 +139,7 @@ optimum/rbln/transformers/models/qwen2/modeling_qwen2.py,sha256=9-aFDvjMzPNUyGOz
|
|
139
139
|
optimum/rbln/transformers/models/qwen2/qwen2_architecture.py,sha256=XlNAMYAcDLohnSAhIFGKOPuCB5XLgzYs5ABWdeQSaZs,720
|
140
140
|
optimum/rbln/transformers/models/qwen2_5_vl/__init__.py,sha256=rAW3DKQUzGL6EMwa5r1iLu94yhpiZpk6zfoD7TtYXrc,865
|
141
141
|
optimum/rbln/transformers/models/qwen2_5_vl/configuration_qwen2_5_vl.py,sha256=dPcGNaLwJf61PIvVbyt-lvBflp_dvK0hubhNoA3len0,3123
|
142
|
-
optimum/rbln/transformers/models/qwen2_5_vl/modeling_qwen2_5_vl.py,sha256=
|
142
|
+
optimum/rbln/transformers/models/qwen2_5_vl/modeling_qwen2_5_vl.py,sha256=pb4OC5rRG-Y-a8WMBGqrV0owgjeGH83Lqx2fXztnlJg,25117
|
143
143
|
optimum/rbln/transformers/models/qwen2_5_vl/qwen2_5_vl_architecture.py,sha256=OR-tTu8uzeFryenohoqRwgIVqw5zM7dcsZHwKL_jD0A,7232
|
144
144
|
optimum/rbln/transformers/models/seq2seq/__init__.py,sha256=6WKstWiS1kW0oFDn_jyrKMW5QEJAWkmsSRAaadNedDM,715
|
145
145
|
optimum/rbln/transformers/models/seq2seq/configuration_seq2seq2.py,sha256=vSNP1eILfL32cbiLOAD58Ocz6lk3hYFnhIRLDVqlSoI,2624
|
@@ -178,7 +178,7 @@ optimum/rbln/utils/model_utils.py,sha256=V2kFpUe2aqVzLwbpztD8JOVFQqRHncvIWwJbgnU
|
|
178
178
|
optimum/rbln/utils/runtime_utils.py,sha256=LoKNK3AQNV_BSScstIZWjICkJf265MnUgy360BOocVI,5454
|
179
179
|
optimum/rbln/utils/save_utils.py,sha256=hG5uOtYmecSXZuGTvCXsTM-SiyZpr5q3InUGCCq_jzQ,3619
|
180
180
|
optimum/rbln/utils/submodule.py,sha256=ZfI7e3YzbjbbBW4Yjfucj8NygEsukfIkaJi3PtwHrhc,5105
|
181
|
-
optimum_rbln-0.
|
182
|
-
optimum_rbln-0.
|
183
|
-
optimum_rbln-0.
|
184
|
-
optimum_rbln-0.
|
181
|
+
optimum_rbln-0.8.0.dist-info/METADATA,sha256=o162FwSFWUcaXw4SoFtdnPtAp14nkmE4wjRbY4O3HEU,5297
|
182
|
+
optimum_rbln-0.8.0.dist-info/WHEEL,sha256=qtCwoSJWgHk21S1Kb4ihdzI2rlJ1ZKaIurTj_ngOhyQ,87
|
183
|
+
optimum_rbln-0.8.0.dist-info/licenses/LICENSE,sha256=QwcOLU5TJoTeUhuIXzhdCEEDDvorGiC6-3YTOl4TecE,11356
|
184
|
+
optimum_rbln-0.8.0.dist-info/RECORD,,
|
File without changes
|
File without changes
|