optimum-rbln 0.8.1a2__tar.gz → 0.8.1a4__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/PKG-INFO +2 -2
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/pyproject.toml +1 -1
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/__version__.py +2 -2
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/models/unets/unet_2d_condition.py +4 -1
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/decoderonly/modeling_decoderonly.py +2 -2
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/gemma3/modeling_gemma3.py +4 -30
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/qwen2_5_vl/modeling_qwen2_5_vl.py +1 -11
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/qwen2_5_vl/qwen2_5_vl_architecture.py +0 -43
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/tests/test_llm.py +2 -2
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/uv.lock +964 -964
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/.github/ISSUE_TEMPLATE/bug_report.md +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/.github/ISSUE_TEMPLATE/config.yml +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/.github/ISSUE_TEMPLATE/feature_request.md +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/.github/ISSUE_TEMPLATE/model_request.md +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/.github/pull_request_template.md +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/.github/scripts/auto_code_review.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/.github/scripts/validate_pr_checklist.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/.github/version.yaml +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/.github/workflows/auto_code_review.yml +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/.github/workflows/check_code_quality.yml +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/.github/workflows/deploy-on-tag.yaml +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/.github/workflows/deploy.yaml +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/.github/workflows/pr-title-check.yaml +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/.github/workflows/pr_checklist_validator.yml +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/.github/workflows/rbln_check_compiler.yaml +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/.github/workflows/rbln_dispatch_pytest.yaml +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/.github/workflows/rbln_optimum_inference_test.yaml +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/.github/workflows/rbln_optimum_pytest.yaml +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/.github/workflows/rbln_scheduled_test.yaml +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/.github/workflows/rbln_trigger_on_pr.yaml +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/.gitignore +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/CODE_OF_CONDUCT.md +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/CONTRIBUTING.md +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/LICENSE +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/README.md +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/assets/rbln_logo.png +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/advanced/custom_class.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/audio-classification/run_ast_audio_classification.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/depth-estimation/run_dpt.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/image-classification/run_image_classification.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/image-classification/run_vit_image_classification.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/image-to-text/run_idefics3.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/image-to-text/run_llava_next_image_to_text.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/kandinsky2_2/run_kandinsky2_2.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/kandinsky2_2/run_kandinsky2_2_combined.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/kandinsky2_2/run_kandinsky2_2_img2img.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/kandinsky2_2/run_kandinsky2_2_img2img_combined.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/kandinsky2_2/run_kandinsky2_2_inpaint.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/kandinsky2_2/run_kandinsky2_2_inpaint_combined.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/kandinsky2_2/run_kandinsky2_2_prior_interpolate.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/question-answering/run_question_answering.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/speech-recognition/run_wav2vec2.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/speech-recognition/run_whisper.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/stable-diffusion/run_stable_diffusion.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/stable-diffusion/run_stable_diffusion_controlnet.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/stable-diffusion/run_stable_diffusion_img2img.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/stable-diffusion/run_stable_diffusion_img2img_controlnet.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/stable-diffusion/run_stable_diffusion_inpaint.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/stable-diffusion/run_stable_diffusion_lora.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/stable-diffusion/run_stable_diffusion_multicontrolnet.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/text-classification/run_bge_m3_text_classification.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/text-classification/run_bge_reranker_v2_m3_text_classification.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/text-classification/run_secureBERT.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/text-classification/run_t5_classification.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/text-classification/run_twitter_roberta_text_classification.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/text2text-generation/run_bart_text2text_generation.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/text2text-generation/run_llama_peft.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/text2text-generation/run_llama_text2text_generation.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/examples/time-series-forecasting/run_time_series_forecasting.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/scripts/uv-lock.sh +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/scripts/uv-sync.sh +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/configuration_utils.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/configurations/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/configurations/models/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/configurations/models/configuration_autoencoder_kl.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/configurations/models/configuration_controlnet.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/configurations/models/configuration_prior_transformer.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/configurations/models/configuration_transformer_sd3.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/configurations/models/configuration_unet_2d_condition.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/configurations/models/configuration_vq_model.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/configurations/pipelines/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/configurations/pipelines/configuration_controlnet.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/configurations/pipelines/configuration_kandinsky2_2.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/configurations/pipelines/configuration_stable_diffusion.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/configurations/pipelines/configuration_stable_diffusion_3.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/configurations/pipelines/configuration_stable_diffusion_xl.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/modeling_diffusers.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/models/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/models/autoencoders/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/models/autoencoders/autoencoder_kl.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/models/autoencoders/vae.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/models/autoencoders/vq_model.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/models/controlnet.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/models/transformers/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/models/transformers/prior_transformer.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/models/transformers/transformer_sd3.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/models/unets/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/pipelines/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/pipelines/controlnet/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/pipelines/controlnet/multicontrolnet.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/pipelines/controlnet/pipeline_controlnet.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/pipelines/controlnet/pipeline_controlnet_img2img.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/pipelines/controlnet/pipeline_controlnet_sd_xl.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/pipelines/controlnet/pipeline_controlnet_sd_xl_img2img.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/pipelines/kandinsky2_2/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2_combined.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2_img2img.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2_inpaint.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2_prior.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/pipelines/stable_diffusion/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/pipelines/stable_diffusion/pipeline_stable_diffusion.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/pipelines/stable_diffusion/pipeline_stable_diffusion_img2img.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/pipelines/stable_diffusion/pipeline_stable_diffusion_inpaint.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/pipelines/stable_diffusion_3/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/pipelines/stable_diffusion_3/pipeline_stable_diffusion_3.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/pipelines/stable_diffusion_3/pipeline_stable_diffusion_3_img2img.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/pipelines/stable_diffusion_3/pipeline_stable_diffusion_3_inpaint.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/pipelines/stable_diffusion_xl/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/pipelines/stable_diffusion_xl/pipeline_stable_diffusion_xl.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/pipelines/stable_diffusion_xl/pipeline_stable_diffusion_xl_img2img.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/diffusers/pipelines/stable_diffusion_xl/pipeline_stable_diffusion_xl_inpaint.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/modeling.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/modeling_base.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/ops/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/ops/attn.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/ops/flash_attn.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/ops/kv_cache_update.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/ops/linear.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/ops/sliding_window_attn.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/configuration_generic.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/modeling_generic.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/modeling_rope_utils.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/audio_spectrogram_transformer/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/audio_spectrogram_transformer/configuration_audio_spectrogram_transformer.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/auto/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/auto/auto_factory.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/auto/modeling_auto.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/bart/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/bart/bart_architecture.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/bart/configuration_bart.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/bart/modeling_bart.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/bert/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/bert/configuration_bert.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/bert/modeling_bert.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/blip_2/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/blip_2/configuration_blip_2.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/blip_2/modeling_blip_2.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/clip/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/clip/configuration_clip.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/clip/modeling_clip.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/decoderonly/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/decoderonly/configuration_decoderonly.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/decoderonly/decoderonly_architecture.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/distilbert/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/distilbert/configuration_distilbert.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/distilbert/modeling_distilbert.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/dpt/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/dpt/configuration_dpt.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/dpt/modeling_dpt.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/exaone/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/exaone/configuration_exaone.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/exaone/exaone_architecture.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/exaone/modeling_exaone.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/gemma/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/gemma/configuration_gemma.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/gemma/gemma_architecture.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/gemma/modeling_gemma.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/gemma3/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/gemma3/configuration_gemma3.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/gemma3/gemma3_architecture.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/gpt2/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/gpt2/configuration_gpt2.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/gpt2/gpt2_architecture.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/gpt2/modeling_gpt2.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/idefics3/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/idefics3/configuration_idefics3.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/idefics3/modeling_idefics3.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/llama/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/llama/configuration_llama.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/llama/llama_architecture.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/llama/modeling_llama.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/llava_next/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/llava_next/configuration_llava_next.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/llava_next/modeling_llava_next.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/midm/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/midm/configuration_midm.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/midm/midm_architecture.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/midm/modeling_midm.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/mistral/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/mistral/configuration_mistral.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/mistral/mistral_architecture.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/mistral/modeling_mistral.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/opt/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/opt/configuration_opt.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/opt/modeling_opt.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/opt/opt_architecture.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/phi/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/phi/configuration_phi.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/phi/modeling_phi.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/phi/phi_architecture.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/qwen2/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/qwen2/configuration_qwen2.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/qwen2/modeling_qwen2.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/qwen2/qwen2_architecture.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/qwen2_5_vl/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/qwen2_5_vl/configuration_qwen2_5_vl.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/resnet/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/resnet/configuration_resnet.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/resnet/modeling_resnet.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/roberta/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/roberta/configuration_roberta.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/roberta/modeling_roberta.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/seq2seq/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/seq2seq/configuration_seq2seq.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/seq2seq/modeling_seq2seq.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/seq2seq/seq2seq_architecture.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/siglip/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/siglip/configuration_siglip.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/siglip/modeling_siglip.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/t5/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/t5/configuration_t5.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/t5/modeling_t5.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/t5/t5_architecture.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/time_series_transformer/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/time_series_transformer/configuration_time_series_transformer.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/time_series_transformer/modeling_time_series_transformer.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/time_series_transformer/time_series_transformers_architecture.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/vit/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/vit/configuration_vit.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/vit/modeling_vit.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/wav2vec2/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/wav2vec2/configuration_wav2vec2.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/wav2vec2/modeling_wav2vec2.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/whisper/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/whisper/configuration_whisper.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/whisper/generation_whisper.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/whisper/modeling_whisper.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/whisper/whisper_architecture.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/xlm_roberta/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/xlm_roberta/configuration_xlm_roberta.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/models/xlm_roberta/modeling_xlm_roberta.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/utils/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/transformers/utils/rbln_quantization.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/utils/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/utils/decorator_utils.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/utils/hub.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/utils/import_utils.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/utils/logging.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/utils/model_utils.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/utils/runtime_utils.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/utils/save_utils.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/src/optimum/rbln/utils/submodule.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/tests/__init__.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/tests/psnr.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/tests/requirements_sdxl.txt +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/tests/run_stable_diffusion_xl_base.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/tests/test_base.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/tests/test_config.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/tests/test_diffusers.py +0 -0
- {optimum_rbln-0.8.1a2 → optimum_rbln-0.8.1a4}/tests/test_transformers.py +0 -0
@@ -1,6 +1,6 @@
|
|
1
1
|
Metadata-Version: 2.4
|
2
2
|
Name: optimum-rbln
|
3
|
-
Version: 0.8.
|
3
|
+
Version: 0.8.1a4
|
4
4
|
Summary: Optimum RBLN is the interface between the HuggingFace Transformers and Diffusers libraries and RBLN accelerators. It provides a set of tools enabling easy model loading and inference on single and multiple rbln device settings for different downstream tasks.
|
5
5
|
Project-URL: Homepage, https://rebellions.ai
|
6
6
|
Project-URL: Documentation, https://docs.rbln.ai
|
@@ -23,7 +23,7 @@ Classifier: Programming Language :: Python :: 3.12
|
|
23
23
|
Classifier: Topic :: Scientific/Engineering :: Artificial Intelligence
|
24
24
|
Requires-Python: <3.13,>=3.9
|
25
25
|
Requires-Dist: accelerate>=1.0.1
|
26
|
-
Requires-Dist: diffusers
|
26
|
+
Requires-Dist: diffusers==0.34.0
|
27
27
|
Requires-Dist: packaging>=24.1
|
28
28
|
Requires-Dist: torch==2.6.0
|
29
29
|
Requires-Dist: torchaudio<=2.6.0
|
@@ -17,5 +17,5 @@ __version__: str
|
|
17
17
|
__version_tuple__: VERSION_TUPLE
|
18
18
|
version_tuple: VERSION_TUPLE
|
19
19
|
|
20
|
-
__version__ = version = '0.8.
|
21
|
-
__version_tuple__ = version_tuple = (0, 8, 1, '
|
20
|
+
__version__ = version = '0.8.1a4'
|
21
|
+
__version_tuple__ = version_tuple = (0, 8, 1, 'a4')
|
@@ -185,7 +185,10 @@ class RBLNUNet2DConditionModel(RBLNModel):
|
|
185
185
|
rbln_config: RBLNUNet2DConditionModelConfig,
|
186
186
|
image_size: Optional[Tuple[int, int]] = None,
|
187
187
|
) -> Tuple[int, int]:
|
188
|
-
|
188
|
+
if hasattr(pipe, "movq"):
|
189
|
+
scale_factor = 2 ** (len(pipe.movq.config.block_out_channels) - 1)
|
190
|
+
else:
|
191
|
+
scale_factor = pipe.vae_scale_factor
|
189
192
|
|
190
193
|
if image_size is None:
|
191
194
|
if "Img2Img" in pipe.__class__.__name__:
|
@@ -11,7 +11,6 @@
|
|
11
11
|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
12
12
|
# See the License for the specific language governing permissions and
|
13
13
|
# limitations under the License.
|
14
|
-
import importlib
|
15
14
|
import inspect
|
16
15
|
from collections import deque
|
17
16
|
from dataclasses import dataclass
|
@@ -124,23 +123,6 @@ class RBLNGemma3ForConditionalGeneration(RBLNModel):
|
|
124
123
|
def can_generate(self):
|
125
124
|
return True
|
126
125
|
|
127
|
-
@classmethod
|
128
|
-
def get_pytorch_model(cls, *args, **kwargs):
|
129
|
-
model = super().get_pytorch_model(*args, **kwargs)
|
130
|
-
|
131
|
-
with no_init_weights():
|
132
|
-
model_cls_name = model.model.language_model.__class__.__name__
|
133
|
-
causal_model_cls_name = model_cls_name.replace("TextModel", "ForCausalLM")
|
134
|
-
causal_model_cls = getattr(importlib.import_module("transformers"), causal_model_cls_name)
|
135
|
-
new_language_model = causal_model_cls(model.model.language_model.config)
|
136
|
-
|
137
|
-
new_language_model.lm_head = model.lm_head
|
138
|
-
new_language_model.model = model.model.language_model
|
139
|
-
model.model.language_model = new_language_model
|
140
|
-
model.lm_head = None
|
141
|
-
del model.lm_head
|
142
|
-
return model
|
143
|
-
|
144
126
|
def __post_init__(self, **kwargs):
|
145
127
|
self.vision_tower = LoopVisionTower(self.rbln_submodules[0])
|
146
128
|
self.language_model = self.rbln_submodules[1]
|
@@ -559,7 +541,7 @@ class RBLNGemma3RuntimeModel(RBLNRuntimeModel):
|
|
559
541
|
(
|
560
542
|
inputs,
|
561
543
|
cache_position,
|
562
|
-
|
544
|
+
padded_attention_mask,
|
563
545
|
out_buffers,
|
564
546
|
position_ids,
|
565
547
|
position_embed,
|
@@ -571,7 +553,7 @@ class RBLNGemma3RuntimeModel(RBLNRuntimeModel):
|
|
571
553
|
)
|
572
554
|
if not is_external_block_tables:
|
573
555
|
local_block_tables = torch.tensor([batch_idx], dtype=torch.int16)
|
574
|
-
self.dec_attn_mask[batch_idx : batch_idx + 1] =
|
556
|
+
self.dec_attn_mask[batch_idx : batch_idx + 1] = padded_attention_mask[:1]
|
575
557
|
|
576
558
|
if self.rbln_config.use_attention_mask and self.rbln_config.use_position_ids:
|
577
559
|
chunked_attention_mask = torch.zeros(1, self.rbln_config.max_seq_len, dtype=torch.float32)
|
@@ -587,18 +569,10 @@ class RBLNGemma3RuntimeModel(RBLNRuntimeModel):
|
|
587
569
|
else None
|
588
570
|
)
|
589
571
|
|
590
|
-
# Not used in Gemma3 yet.
|
591
572
|
if self.rbln_config.use_attention_mask:
|
592
573
|
if self.rbln_config.use_position_ids:
|
593
|
-
chunked_attention_mask[0, step : step + self.rbln_config.prefill_chunk_size] =
|
594
|
-
|
595
|
-
]
|
596
|
-
else:
|
597
|
-
# Update attention mask to ensure proper causal behavior
|
598
|
-
if step >= self.rbln_config.prefill_chunk_size:
|
599
|
-
chunked_attention_mask[:, :, :, step - self.rbln_config.prefill_chunk_size : step] = 1
|
600
|
-
chunked_attention_mask[:, :, :, step : step + self.rbln_config.prefill_chunk_size] = (
|
601
|
-
self.causal_mask
|
574
|
+
chunked_attention_mask[0, step : step + self.rbln_config.prefill_chunk_size] = (
|
575
|
+
padded_attention_mask[0, step : step + self.rbln_config.prefill_chunk_size]
|
602
576
|
)
|
603
577
|
|
604
578
|
# Define query position
|
@@ -28,7 +28,6 @@ from transformers.models.qwen2_5_vl.modeling_qwen2_5_vl import (
|
|
28
28
|
Qwen2_5_VisionPatchEmbed,
|
29
29
|
Qwen2_5_VisionRotaryEmbedding,
|
30
30
|
Qwen2_5_VisionTransformerPretrainedModel,
|
31
|
-
Qwen2_5_VLModel,
|
32
31
|
Qwen2_5_VLRotaryEmbedding,
|
33
32
|
)
|
34
33
|
|
@@ -391,14 +390,6 @@ class RBLNQwen2_5_VLForConditionalGeneration(RBLNDecoderOnlyModelForCausalLM):
|
|
391
390
|
def can_generate(self):
|
392
391
|
return True
|
393
392
|
|
394
|
-
@classmethod
|
395
|
-
def get_pytorch_model(cls, *args, **kwargs):
|
396
|
-
model = super().get_pytorch_model(*args, **kwargs)
|
397
|
-
model.model.lm_head = model.lm_head
|
398
|
-
model.lm_head = None
|
399
|
-
del model.lm_head
|
400
|
-
return model
|
401
|
-
|
402
393
|
@classmethod
|
403
394
|
def update_kwargs(cls, kwargs):
|
404
395
|
kwargs.update(
|
@@ -540,8 +531,7 @@ class RBLNQwen2_5_VLForConditionalGeneration(RBLNDecoderOnlyModelForCausalLM):
|
|
540
531
|
vision_tokens = input_id[0][vision_start_indices + 1]
|
541
532
|
image_nums = (vision_tokens == image_token_id).sum()
|
542
533
|
video_nums = (vision_tokens == video_token_id).sum()
|
543
|
-
position_ids, rope_deltas =
|
544
|
-
self,
|
534
|
+
position_ids, rope_deltas = self.get_rope_index(
|
545
535
|
input_id,
|
546
536
|
image_grid_thw[image_idx : image_idx + image_nums] if image_grid_thw is not None else None,
|
547
537
|
video_grid_thw[video_idx : video_idx + video_nums] if video_grid_thw is not None else None,
|
@@ -3,14 +3,8 @@ from typing import Tuple
|
|
3
3
|
|
4
4
|
import torch
|
5
5
|
import torch.nn as nn
|
6
|
-
from transformers import PreTrainedModel
|
7
6
|
|
8
7
|
from ..decoderonly.decoderonly_architecture import (
|
9
|
-
DecoderOnlyAttention,
|
10
|
-
DecoderOnlyFlashAttention,
|
11
|
-
DecoderOnlyForCausalLM,
|
12
|
-
DecoderOnlyLayer,
|
13
|
-
DecoderOnlyModel,
|
14
8
|
DecoderOnlyWrapper,
|
15
9
|
apply_rotary_pos_emb,
|
16
10
|
)
|
@@ -203,40 +197,3 @@ class Qwen2_5_VL_LanguageModelWrapper(DecoderOnlyWrapper):
|
|
203
197
|
past_key_values,
|
204
198
|
position_embeds,
|
205
199
|
)
|
206
|
-
|
207
|
-
def convert_to_rbln_causal_lm(self, causal_lm: PreTrainedModel, max_seq_len: int):
|
208
|
-
new_layers = []
|
209
|
-
|
210
|
-
for layer in causal_lm.model.language_model.layers:
|
211
|
-
if self.attn_impl == "eager":
|
212
|
-
new_self_attn = DecoderOnlyAttention(
|
213
|
-
layer.self_attn,
|
214
|
-
self.use_attention_mask,
|
215
|
-
self.use_position_ids,
|
216
|
-
kvcache_block_size=self.kvcache_block_size,
|
217
|
-
)
|
218
|
-
elif self.attn_impl == "flash_attn":
|
219
|
-
new_self_attn = DecoderOnlyFlashAttention(
|
220
|
-
layer.self_attn,
|
221
|
-
kvcache_partition_len=self.kvcache_partition_len,
|
222
|
-
kvcache_block_size=self.kvcache_block_size,
|
223
|
-
use_attention_mask=self.use_attention_mask,
|
224
|
-
use_position_ids=self.use_position_ids,
|
225
|
-
)
|
226
|
-
else:
|
227
|
-
raise NotImplementedError(f"Unknwon attn : {self.attn_impl}")
|
228
|
-
|
229
|
-
new_layer = DecoderOnlyLayer(layer, new_self_attn)
|
230
|
-
new_layers.append(new_layer)
|
231
|
-
|
232
|
-
new_model = DecoderOnlyModel(
|
233
|
-
causal_lm.model.language_model,
|
234
|
-
new_layers,
|
235
|
-
partition_len=self.kvcache_partition_len,
|
236
|
-
max_seq_len=max_seq_len,
|
237
|
-
kvcache_block_size=self.kvcache_block_size,
|
238
|
-
use_learned_pos_emb=self.use_learned_pos_emb,
|
239
|
-
sliding_window_layers=self.sliding_window_layers,
|
240
|
-
)
|
241
|
-
new_causal_lm = DecoderOnlyForCausalLM(causal_lm.model, new_model)
|
242
|
-
return new_causal_lm
|
@@ -67,7 +67,7 @@ class LLMTest:
|
|
67
67
|
class TestQwen2Model(LLMTest.TestLLM):
|
68
68
|
RBLN_CLASS = RBLNQwen2ForCausalLM
|
69
69
|
HF_MODEL_ID = "Qwen/Qwen2-0.5B-Instruct"
|
70
|
-
EXPECTED_OUTPUT = " I am a
|
70
|
+
EXPECTED_OUTPUT = " I am a 20 year old girl from the United States. I have been studying English for"
|
71
71
|
HF_CONFIG_KWARGS = {"max_position_embeddings": 1024}
|
72
72
|
|
73
73
|
|
@@ -108,7 +108,7 @@ class TestLlamaForCausalLM_Flash(LLMTest.TestLLM):
|
|
108
108
|
class TestLlamaForCausalLM_Multibatch(TestLlamaForCausalLM):
|
109
109
|
PROMPT = ["Who are you?", "What is the capital of France?", "What is the capital of Germany?"]
|
110
110
|
EXPECTED_OUTPUT = [
|
111
|
-
"reress makefable R����
|
111
|
+
"reress makefable R���� noethetss0oss invetetet",
|
112
112
|
"resget makeget makeichget makeichualichual#choolchool accngngngng",
|
113
113
|
"resget makeget makeichget makeichualichual#choolchool accngngngng",
|
114
114
|
]
|