optimum-rbln 0.9.2a6__tar.gz → 0.9.2a8__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of optimum-rbln might be problematic. Click here for more details.
- optimum_rbln-0.9.2a8/.github/version.yaml +1 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/.github/workflows/check_code_quality.yml +1 -1
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/.github/workflows/rbln_check_compiler.yaml +5 -2
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/.github/workflows/rbln_optimum_inference_test.yaml +1 -1
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/.github/workflows/rbln_optimum_pytest.yaml +1 -1
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/PKG-INFO +1 -1
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/pyproject.toml +1 -1
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/__version__.py +2 -2
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/configuration_utils.py +6 -25
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/modeling_diffusers.py +5 -7
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/modeling_base.py +1 -2
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/decoderonly/decoderonly_runtime_utils.py +17 -2
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/decoderonly/modeling_decoderonly.py +7 -28
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/whisper/generation_whisper.py +13 -1
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/utils/runtime_utils.py +0 -5
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/tests/test_config.py +1 -6
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/uv.lock +7 -7
- optimum_rbln-0.9.2a6/.github/version.yaml +0 -1
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/.github/ISSUE_TEMPLATE/bug_report.md +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/.github/ISSUE_TEMPLATE/config.yml +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/.github/ISSUE_TEMPLATE/feature_request.md +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/.github/ISSUE_TEMPLATE/model_request.md +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/.github/pull_request_template.md +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/.github/renovate.json +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/.github/scripts/auto_code_review.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/.github/scripts/validate_docstrings.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/.github/scripts/validate_pr_checklist.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/.github/workflows/auto_code_review.yml +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/.github/workflows/auto_dependency_bot.yml +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/.github/workflows/deploy-on-tag.yaml +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/.github/workflows/deploy.yaml +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/.github/workflows/pr-title-check.yaml +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/.github/workflows/pr_checklist_validator.yml +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/.github/workflows/rbln_dispatch_pytest.yaml +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/.github/workflows/rbln_scheduled_test.yaml +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/.github/workflows/rbln_trigger_on_pr.yaml +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/.github/workflows/test-docstrings.yml +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/.gitignore +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/CODE_OF_CONDUCT.md +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/CONTRIBUTING.md +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/LICENSE +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/README.md +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/assets/rbln_logo.png +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/advanced/custom_class.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/audio-classification/run_ast_audio_classification.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/depth-estimation/run_dpt.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/image-classification/run_image_classification.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/image-classification/run_vit_image_classification.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/image-to-text/run_idefics3.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/image-to-text/run_llava_next_image_to_text.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/kandinsky2_2/run_kandinsky2_2.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/kandinsky2_2/run_kandinsky2_2_combined.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/kandinsky2_2/run_kandinsky2_2_img2img.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/kandinsky2_2/run_kandinsky2_2_img2img_combined.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/kandinsky2_2/run_kandinsky2_2_inpaint.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/kandinsky2_2/run_kandinsky2_2_inpaint_combined.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/kandinsky2_2/run_kandinsky2_2_prior_interpolate.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/question-answering/run_question_answering.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/speech-recognition/run_wav2vec2.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/speech-recognition/run_whisper.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/stable-diffusion/run_stable_diffusion.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/stable-diffusion/run_stable_diffusion_controlnet.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/stable-diffusion/run_stable_diffusion_img2img.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/stable-diffusion/run_stable_diffusion_img2img_controlnet.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/stable-diffusion/run_stable_diffusion_inpaint.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/stable-diffusion/run_stable_diffusion_lora.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/stable-diffusion/run_stable_diffusion_multicontrolnet.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/text-classification/run_bge_m3_text_classification.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/text-classification/run_bge_reranker_v2_m3_text_classification.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/text-classification/run_secureBERT.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/text-classification/run_t5_classification.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/text-classification/run_twitter_roberta_text_classification.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/text2text-generation/run_bart_text2text_generation.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/text2text-generation/run_llama_peft.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/text2text-generation/run_llama_text2text_generation.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/time-series-forecasting/run_time_series_forecasting.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/scripts/uv-lock.sh +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/scripts/uv-sync.sh +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/configurations/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/configurations/models/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/configurations/models/configuration_autoencoder_kl.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/configurations/models/configuration_autoencoder_kl_cosmos.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/configurations/models/configuration_controlnet.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/configurations/models/configuration_prior_transformer.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/configurations/models/configuration_transformer_cosmos.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/configurations/models/configuration_transformer_sd3.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/configurations/models/configuration_unet_2d_condition.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/configurations/models/configuration_vq_model.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/configurations/pipelines/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/configurations/pipelines/configuration_controlnet.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/configurations/pipelines/configuration_cosmos.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/configurations/pipelines/configuration_kandinsky2_2.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/configurations/pipelines/configuration_stable_diffusion.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/configurations/pipelines/configuration_stable_diffusion_3.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/configurations/pipelines/configuration_stable_diffusion_xl.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/models/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/models/autoencoders/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/models/autoencoders/autoencoder_kl.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/models/autoencoders/autoencoder_kl_cosmos.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/models/autoencoders/vae.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/models/autoencoders/vq_model.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/models/controlnet.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/models/transformers/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/models/transformers/prior_transformer.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/models/transformers/transformer_cosmos.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/models/transformers/transformer_sd3.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/models/unets/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/models/unets/unet_2d_condition.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/auto_pipeline.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/controlnet/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/controlnet/multicontrolnet.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/controlnet/pipeline_controlnet.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/controlnet/pipeline_controlnet_img2img.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/controlnet/pipeline_controlnet_sd_xl.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/controlnet/pipeline_controlnet_sd_xl_img2img.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/cosmos/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/cosmos/configuration_cosmos_guardrail.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/cosmos/cosmos_guardrail.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/cosmos/pipeline_cosmos_text2world.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/cosmos/pipeline_cosmos_video2world.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/kandinsky2_2/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2_combined.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2_img2img.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2_inpaint.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2_prior.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/stable_diffusion/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/stable_diffusion/pipeline_stable_diffusion.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/stable_diffusion/pipeline_stable_diffusion_img2img.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/stable_diffusion/pipeline_stable_diffusion_inpaint.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/stable_diffusion_3/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/stable_diffusion_3/pipeline_stable_diffusion_3.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/stable_diffusion_3/pipeline_stable_diffusion_3_img2img.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/stable_diffusion_3/pipeline_stable_diffusion_3_inpaint.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/stable_diffusion_xl/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/stable_diffusion_xl/pipeline_stable_diffusion_xl.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/stable_diffusion_xl/pipeline_stable_diffusion_xl_img2img.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/pipelines/stable_diffusion_xl/pipeline_stable_diffusion_xl_inpaint.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/modeling.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/ops/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/ops/attn.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/ops/flash_attn.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/ops/kv_cache_update.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/ops/linear.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/ops/sliding_window_attn.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/configuration_generic.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/modeling_attention_utils.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/modeling_generic.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/modeling_outputs.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/modeling_rope_utils.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/audio_spectrogram_transformer/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/audio_spectrogram_transformer/configuration_audio_spectrogram_transformer.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/auto/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/auto/auto_factory.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/auto/modeling_auto.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/bart/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/bart/bart_architecture.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/bart/configuration_bart.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/bart/modeling_bart.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/bert/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/bert/bert_architecture.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/bert/configuration_bert.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/bert/modeling_bert.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/blip_2/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/blip_2/configuration_blip_2.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/blip_2/modeling_blip_2.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/clip/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/clip/configuration_clip.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/clip/modeling_clip.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/colpali/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/colpali/colpali_architecture.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/colpali/configuration_colpali.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/colpali/modeling_colpali.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/decoderonly/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/decoderonly/configuration_decoderonly.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/decoderonly/configuration_lora.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/decoderonly/decoderonly_architecture.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/decoderonly/generation_decoderonly.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/decoderonly/lora_architecture.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/depth_anything/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/depth_anything/configuration_depth_anything.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/depth_anything/modeling_depth_anything.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/distilbert/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/distilbert/configuration_distilbert.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/distilbert/modeling_distilbert.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/dpt/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/dpt/configuration_dpt.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/dpt/modeling_dpt.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/exaone/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/exaone/configuration_exaone.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/exaone/exaone_architecture.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/exaone/modeling_exaone.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/gemma/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/gemma/configuration_gemma.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/gemma/gemma_architecture.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/gemma/modeling_gemma.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/gemma3/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/gemma3/configuration_gemma3.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/gemma3/gemma3_architecture.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/gemma3/gemma3_runtime_utils.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/gemma3/modeling_gemma3.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/gpt2/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/gpt2/configuration_gpt2.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/gpt2/gpt2_architecture.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/gpt2/modeling_gpt2.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/grounding_dino/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/grounding_dino/configuration_grounding_dino.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/grounding_dino/grounding_dino_architecture.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/grounding_dino/modeling_grounding_dino.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/idefics3/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/idefics3/configuration_idefics3.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/idefics3/modeling_idefics3.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/llama/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/llama/configuration_llama.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/llama/llama_architecture.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/llama/modeling_llama.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/llava/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/llava/configuration_llava.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/llava/modeling_llava.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/llava_next/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/llava_next/configuration_llava_next.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/llava_next/modeling_llava_next.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/midm/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/midm/configuration_midm.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/midm/midm_architecture.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/midm/modeling_midm.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/mistral/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/mistral/configuration_mistral.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/mistral/mistral_architecture.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/mistral/modeling_mistral.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/opt/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/opt/configuration_opt.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/opt/modeling_opt.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/opt/opt_architecture.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/pegasus/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/pegasus/configuration_pegasus.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/pegasus/modeling_pegasus.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/pegasus/pegasus_architecture.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/phi/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/phi/configuration_phi.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/phi/modeling_phi.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/phi/phi_architecture.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/pixtral/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/pixtral/configuration_pixtral.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/pixtral/modeling_pixtral.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/pixtral/pixtral_architecture.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/qwen2/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/qwen2/configuration_qwen2.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/qwen2/modeling_qwen2.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/qwen2/qwen2_architecture.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/qwen2_5_vl/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/qwen2_5_vl/configuration_qwen2_5_vl.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/qwen2_5_vl/modeling_qwen2_5_vl.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/qwen2_5_vl/qwen2_5_vl_architecture.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/qwen2_vl/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/qwen2_vl/configuration_qwen2_vl.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/qwen2_vl/modeling_qwen2_vl.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/qwen2_vl/qwen2_vl_architecture.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/qwen3/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/qwen3/configuration_qwen3.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/qwen3/modeling_qwen3.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/qwen3/qwen3_architecture.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/resnet/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/resnet/configuration_resnet.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/resnet/modeling_resnet.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/roberta/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/roberta/configuration_roberta.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/roberta/modeling_roberta.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/seq2seq/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/seq2seq/configuration_seq2seq.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/seq2seq/modeling_seq2seq.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/seq2seq/seq2seq_architecture.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/siglip/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/siglip/configuration_siglip.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/siglip/modeling_siglip.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/swin/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/swin/configuration_swin.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/swin/modeling_swin.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/t5/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/t5/configuration_t5.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/t5/modeling_t5.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/t5/t5_architecture.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/time_series_transformer/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/time_series_transformer/configuration_time_series_transformer.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/time_series_transformer/modeling_time_series_transformer.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/time_series_transformer/time_series_transformers_architecture.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/vit/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/vit/configuration_vit.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/vit/modeling_vit.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/wav2vec2/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/wav2vec2/configuration_wav2vec2.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/wav2vec2/modeling_wav2vec2.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/whisper/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/whisper/configuration_whisper.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/whisper/modeling_whisper.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/whisper/whisper_architecture.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/xlm_roberta/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/xlm_roberta/configuration_xlm_roberta.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/models/xlm_roberta/modeling_xlm_roberta.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/utils/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/utils/rbln_quantization.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/transformers/utils/rbln_runtime_wrapper.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/utils/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/utils/decorator_utils.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/utils/depreacate_utils.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/utils/hub.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/utils/import_utils.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/utils/logging.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/utils/model_utils.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/utils/save_utils.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/utils/submodule.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/tests/__init__.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/tests/psnr.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/tests/requirements_sdxl.txt +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/tests/run_stable_diffusion_xl_base.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/tests/test_base.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/tests/test_diffusers.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/tests/test_llm.py +0 -0
- {optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/tests/test_transformers.py +0 -0
|
@@ -0,0 +1 @@
|
|
|
1
|
+
rebel_compiler_version: 0.9.2.dev164+g0896157e.prod
|
|
@@ -25,7 +25,7 @@ jobs:
|
|
|
25
25
|
- name: Set up Python
|
|
26
26
|
uses: actions/setup-python@v5
|
|
27
27
|
with:
|
|
28
|
-
python-version: "3.
|
|
28
|
+
python-version: "3.12"
|
|
29
29
|
|
|
30
30
|
- name: Install rebel-compiler
|
|
31
31
|
id: install_compiler
|
|
@@ -62,7 +62,10 @@ jobs:
|
|
|
62
62
|
- name: Check pyarmor
|
|
63
63
|
id: check_pyarmor
|
|
64
64
|
run: |
|
|
65
|
-
|
|
65
|
+
PKG_PATH=$(python3 -c "import rebel; print(rebel.__file__)")
|
|
66
|
+
PKG_DIR=$(dirname "$PKG_PATH")
|
|
67
|
+
|
|
68
|
+
if find "$PKG_DIR" -type f | grep -E '__pyarmor__|pytransform|pyarmor_runtime|license\.lic'; then
|
|
66
69
|
echo "PYARMOR_ENC=true" >> $GITHUB_ENV
|
|
67
70
|
echo "pyarmor check pass(✅)!!"
|
|
68
71
|
else
|
|
@@ -99,7 +99,7 @@ jobs:
|
|
|
99
99
|
|
|
100
100
|
- name: Setup Python
|
|
101
101
|
if: steps.should_skip.outputs.skip != 'true'
|
|
102
|
-
run: uv python install 3.
|
|
102
|
+
run: uv python install 3.12
|
|
103
103
|
|
|
104
104
|
- name: Install optimum-rbln with tests group dependencies
|
|
105
105
|
if: steps.should_skip.outputs.skip != 'true'
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.4
|
|
2
2
|
Name: optimum-rbln
|
|
3
|
-
Version: 0.9.
|
|
3
|
+
Version: 0.9.2a8
|
|
4
4
|
Summary: Optimum RBLN is the interface between the HuggingFace Transformers and Diffusers libraries and RBLN accelerators. It provides a set of tools enabling easy model loading and inference on single and multiple rbln device settings for different downstream tasks.
|
|
5
5
|
Project-URL: Homepage, https://rebellions.ai
|
|
6
6
|
Project-URL: Documentation, https://docs.rbln.ai
|
|
@@ -28,7 +28,7 @@ version_tuple: VERSION_TUPLE
|
|
|
28
28
|
commit_id: COMMIT_ID
|
|
29
29
|
__commit_id__: COMMIT_ID
|
|
30
30
|
|
|
31
|
-
__version__ = version = '0.9.
|
|
32
|
-
__version_tuple__ = version_tuple = (0, 9, 2, '
|
|
31
|
+
__version__ = version = '0.9.2a8'
|
|
32
|
+
__version_tuple__ = version_tuple = (0, 9, 2, 'a8')
|
|
33
33
|
|
|
34
34
|
__commit_id__ = commit_id = None
|
|
@@ -33,7 +33,6 @@ logger = get_logger(__name__)
|
|
|
33
33
|
|
|
34
34
|
|
|
35
35
|
DEFAULT_COMPILED_MODEL_NAME = "compiled_model"
|
|
36
|
-
DEFAULT_MOD_NAME = "default"
|
|
37
36
|
TypeInputInfo = List[Tuple[str, Tuple[int], str]]
|
|
38
37
|
|
|
39
38
|
|
|
@@ -52,17 +51,13 @@ class RBLNCompileConfig:
|
|
|
52
51
|
|
|
53
52
|
Attributes:
|
|
54
53
|
compiled_model_name (str): Name of the compiled model.
|
|
55
|
-
mod_name (str): Name of the RBLN module.
|
|
56
54
|
input_info (Union[List[TypeInputInfo], TypeInputInfo]): Information about input tensors.
|
|
57
|
-
fusion (Optional[bool]): Whether to use fusion optimization.
|
|
58
55
|
npu (Optional[str]): NPU configuration.
|
|
59
56
|
tensor_parallel_size (Optional[int]): Size for tensor parallelism.
|
|
60
57
|
"""
|
|
61
58
|
|
|
62
59
|
compiled_model_name: str = DEFAULT_COMPILED_MODEL_NAME
|
|
63
|
-
mod_name: str = DEFAULT_MOD_NAME
|
|
64
60
|
input_info: Union[List[TypeInputInfo], TypeInputInfo] = None
|
|
65
|
-
fusion: Optional[bool] = None
|
|
66
61
|
npu: Optional[str] = None
|
|
67
62
|
tensor_parallel_size: Optional[int] = None
|
|
68
63
|
|
|
@@ -116,9 +111,7 @@ class RBLNCompileConfig:
|
|
|
116
111
|
|
|
117
112
|
def update(self, kwargs: Dict[str, Any]):
|
|
118
113
|
self.compiled_model_name = kwargs.get("compiled_model_name", self.compiled_model_name)
|
|
119
|
-
self.mod_name = kwargs.get("mod_name", self.mod_name)
|
|
120
114
|
self.input_info = kwargs.get("input_info", self.input_info)
|
|
121
|
-
self.fusion = kwargs.get("fusion", self.fusion)
|
|
122
115
|
self.npu = kwargs.get("npu", self.npu)
|
|
123
116
|
self.tensor_parallel_size = kwargs.get("tensor_parallel_size", self.tensor_parallel_size)
|
|
124
117
|
return self
|
|
@@ -152,7 +145,7 @@ class RBLNCompileConfig:
|
|
|
152
145
|
return asdict(self)
|
|
153
146
|
|
|
154
147
|
|
|
155
|
-
RUNTIME_KEYWORDS = ["create_runtimes", "
|
|
148
|
+
RUNTIME_KEYWORDS = ["create_runtimes", "device", "device_map", "activate_profiler", "timeout"]
|
|
156
149
|
CONFIG_MAPPING: Dict[str, Type["RBLNModelConfig"]] = {}
|
|
157
150
|
|
|
158
151
|
|
|
@@ -528,7 +521,6 @@ class RBLNModelConfig(RBLNSerializableConfigProtocol):
|
|
|
528
521
|
"npu",
|
|
529
522
|
"tensor_parallel_size",
|
|
530
523
|
"create_runtimes",
|
|
531
|
-
"optimize_host_memory",
|
|
532
524
|
"device",
|
|
533
525
|
"device_map",
|
|
534
526
|
"activate_profiler",
|
|
@@ -654,7 +646,6 @@ class RBLNModelConfig(RBLNSerializableConfigProtocol):
|
|
|
654
646
|
self,
|
|
655
647
|
cls_name: Optional[str] = None,
|
|
656
648
|
create_runtimes: Optional[bool] = None,
|
|
657
|
-
optimize_host_memory: Optional[bool] = None,
|
|
658
649
|
device: Optional[Union[int, List[int]]] = None,
|
|
659
650
|
device_map: Optional[Dict[str, Union[int, List[int]]]] = None,
|
|
660
651
|
activate_profiler: Optional[bool] = None,
|
|
@@ -664,6 +655,8 @@ class RBLNModelConfig(RBLNSerializableConfigProtocol):
|
|
|
664
655
|
optimum_rbln_version: Optional[str] = None,
|
|
665
656
|
_torch_dtype: Optional[str] = None,
|
|
666
657
|
_compile_cfgs: List[RBLNCompileConfig] = [],
|
|
658
|
+
*,
|
|
659
|
+
optimize_host_memory: Optional[bool] = None,
|
|
667
660
|
**kwargs: Any,
|
|
668
661
|
):
|
|
669
662
|
"""
|
|
@@ -672,7 +665,6 @@ class RBLNModelConfig(RBLNSerializableConfigProtocol):
|
|
|
672
665
|
Args:
|
|
673
666
|
cls_name (Optional[str]): The class name of the configuration. Defaults to the current class name.
|
|
674
667
|
create_runtimes (Optional[bool]): Whether to create RBLN runtimes. Defaults to True.
|
|
675
|
-
optimize_host_memory (Optional[bool]): Whether to optimize host memory usage. Defaults to True.
|
|
676
668
|
device (Optional[Union[int, List[int]]]): The device(s) to load the model onto. Can be a single device ID or a list.
|
|
677
669
|
device_map (Optional[Dict[str, Union[int, List[int]]]]): Mapping from compiled model names to device IDs.
|
|
678
670
|
activate_profiler (Optional[bool]): Whether to activate the profiler for performance analysis.
|
|
@@ -698,12 +690,14 @@ class RBLNModelConfig(RBLNSerializableConfigProtocol):
|
|
|
698
690
|
|
|
699
691
|
self._runtime_options = {}
|
|
700
692
|
self._runtime_options["create_runtimes"] = create_runtimes
|
|
701
|
-
self._runtime_options["optimize_host_memory"] = optimize_host_memory
|
|
702
693
|
self._runtime_options["device"] = device
|
|
703
694
|
self._runtime_options["device_map"] = device_map
|
|
704
695
|
self._runtime_options["activate_profiler"] = activate_profiler
|
|
705
696
|
self._runtime_options["timeout"] = timeout
|
|
706
697
|
|
|
698
|
+
if optimize_host_memory is not None:
|
|
699
|
+
logger.warning("`optimize_host_memory` is deprecated and will be removed in future versions.")
|
|
700
|
+
|
|
707
701
|
# Automatically pass npu, tensor_parallel_size to compile_cfgs
|
|
708
702
|
self.npu = npu
|
|
709
703
|
self.tensor_parallel_size = tensor_parallel_size
|
|
@@ -921,19 +915,6 @@ class RBLNModelConfig(RBLNSerializableConfigProtocol):
|
|
|
921
915
|
def create_runtimes(self, create_runtimes: bool):
|
|
922
916
|
self._runtime_options["create_runtimes"] = create_runtimes
|
|
923
917
|
|
|
924
|
-
@property
|
|
925
|
-
def optimize_host_memory(self):
|
|
926
|
-
context = ContextRblnConfig.get_current_context()["optimize_host_memory"]
|
|
927
|
-
if context is not None:
|
|
928
|
-
return context
|
|
929
|
-
elif self._runtime_options["optimize_host_memory"] is None:
|
|
930
|
-
return True
|
|
931
|
-
return self._runtime_options["optimize_host_memory"]
|
|
932
|
-
|
|
933
|
-
@optimize_host_memory.setter
|
|
934
|
-
def optimize_host_memory(self, optimize_host_memory: bool):
|
|
935
|
-
self._runtime_options["optimize_host_memory"] = optimize_host_memory
|
|
936
|
-
|
|
937
918
|
@property
|
|
938
919
|
def device(self):
|
|
939
920
|
context = ContextRblnConfig.get_current_context()["device"]
|
{optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/src/optimum/rbln/diffusers/modeling_diffusers.py
RENAMED
|
@@ -244,7 +244,6 @@ class RBLNDiffusionMixin:
|
|
|
244
244
|
device=rbln_config.device,
|
|
245
245
|
device_map=rbln_config.device_map,
|
|
246
246
|
create_runtimes=rbln_config.create_runtimes,
|
|
247
|
-
optimize_host_mem=rbln_config.optimize_host_memory,
|
|
248
247
|
activate_profiler=rbln_config.activate_profiler,
|
|
249
248
|
timeout=rbln_config.timeout,
|
|
250
249
|
):
|
|
@@ -412,12 +411,11 @@ class RBLNDiffusionMixin:
|
|
|
412
411
|
# overwrite to replace incorrect config
|
|
413
412
|
model.save_config(model_save_dir)
|
|
414
413
|
|
|
415
|
-
|
|
416
|
-
|
|
417
|
-
|
|
418
|
-
|
|
419
|
-
|
|
420
|
-
model.compiled_models.extend(submodule.compiled_models)
|
|
414
|
+
# Keep compiled_model objs to further analysis. -> TODO: remove soon...
|
|
415
|
+
model.compiled_models = []
|
|
416
|
+
for name in cls._submodules:
|
|
417
|
+
submodule = getattr(model, name)
|
|
418
|
+
model.compiled_models.extend(submodule.compiled_models)
|
|
421
419
|
|
|
422
420
|
return model
|
|
423
421
|
|
|
@@ -315,7 +315,7 @@ class RBLNBaseModel(SubModulesMixin, PushToHubMixin, PreTrainedModel):
|
|
|
315
315
|
rbln_config,
|
|
316
316
|
model_save_dir=model_save_dir,
|
|
317
317
|
subfolder=subfolder,
|
|
318
|
-
rbln_compiled_models=
|
|
318
|
+
rbln_compiled_models=rbln_compiled_models,
|
|
319
319
|
rbln_submodules=rbln_submodules,
|
|
320
320
|
**kwargs,
|
|
321
321
|
)
|
|
@@ -433,7 +433,6 @@ class RBLNBaseModel(SubModulesMixin, PushToHubMixin, PreTrainedModel):
|
|
|
433
433
|
compiled_model = rebel.compile_from_torch(
|
|
434
434
|
model,
|
|
435
435
|
input_info=rbln_compile_config.input_info,
|
|
436
|
-
fusion=rbln_compile_config.fusion,
|
|
437
436
|
npu=rbln_compile_config.npu,
|
|
438
437
|
tensor_parallel_size=rbln_compile_config.tensor_parallel_size,
|
|
439
438
|
**kwargs,
|
|
@@ -420,6 +420,16 @@ class RBLNRuntimeModel(RBLNPytorchRuntime):
|
|
|
420
420
|
inputs, cache_position, attention_mask, position_embed, token_type_ids=token_type_ids
|
|
421
421
|
)
|
|
422
422
|
|
|
423
|
+
# Assumed that prefix caching was performed externally if cache_position doesn't start from 0.
|
|
424
|
+
prefix_cached_len = cache_position[0][0].item()
|
|
425
|
+
if prefix_cached_len > 0:
|
|
426
|
+
if prefix_cached_len % self.rbln_config.prefill_chunk_size != 0:
|
|
427
|
+
raise NotImplementedError(
|
|
428
|
+
"Prefix Caching is not supported yet for non-multiple of prefill_chunk_size."
|
|
429
|
+
)
|
|
430
|
+
if self.rbln_config.use_attention_mask:
|
|
431
|
+
chunked_attention_mask[:, :, :, :prefix_cached_len] = 1
|
|
432
|
+
|
|
423
433
|
# Process input in chunks of size `prefill_chunk_size`
|
|
424
434
|
output_logits = []
|
|
425
435
|
for step in range(0, query_length, self.rbln_config.prefill_chunk_size):
|
|
@@ -434,9 +444,14 @@ class RBLNRuntimeModel(RBLNPytorchRuntime):
|
|
|
434
444
|
if self.rbln_config.use_attention_mask and not self.rbln_config.use_position_ids:
|
|
435
445
|
if step > 0: # update previous chunk
|
|
436
446
|
chunked_attention_mask[
|
|
437
|
-
:,
|
|
447
|
+
:,
|
|
448
|
+
:,
|
|
449
|
+
:,
|
|
450
|
+
s - self.rbln_config.prefill_chunk_size + prefix_cached_len : e
|
|
451
|
+
- self.rbln_config.prefill_chunk_size
|
|
452
|
+
+ prefix_cached_len,
|
|
438
453
|
] = 1
|
|
439
|
-
chunked_attention_mask[:, :, :, s:e] = self.causal_mask
|
|
454
|
+
chunked_attention_mask[:, :, :, s + prefix_cached_len : e + prefix_cached_len] = self.causal_mask
|
|
440
455
|
|
|
441
456
|
# Calculate query position if needed
|
|
442
457
|
if self.rbln_config.use_local_attention or self.rbln_config.logits_to_keep > 0:
|
|
@@ -19,7 +19,7 @@ from typing import TYPE_CHECKING, Any, Callable, List, Optional, Tuple, Union
|
|
|
19
19
|
import rebel
|
|
20
20
|
import torch
|
|
21
21
|
from rebel.compile_context import CompileContext
|
|
22
|
-
from transformers import
|
|
22
|
+
from transformers import AutoModel, AutoModelForCausalLM, PretrainedConfig, PreTrainedModel
|
|
23
23
|
from transformers.modeling_outputs import BaseModelOutputWithPast
|
|
24
24
|
from transformers.modeling_utils import no_init_weights
|
|
25
25
|
|
|
@@ -260,10 +260,12 @@ class RBLNDecoderOnlyModel(RBLNModel, RBLNDecoderOnlyFlashAttentionMixin):
|
|
|
260
260
|
|
|
261
261
|
# Mark static tensors (self kv states)
|
|
262
262
|
static_tensors = {}
|
|
263
|
+
idx = 0
|
|
263
264
|
for (name, _, _), tensor in zip(compile_config.input_info, example_inputs):
|
|
264
265
|
if "past_key_values" in name:
|
|
265
266
|
static_tensors[name] = tensor
|
|
266
|
-
context.mark_static_address(tensor)
|
|
267
|
+
context.mark_static_address(tensor, f"kv_cache_{idx}")
|
|
268
|
+
idx += 1
|
|
267
269
|
|
|
268
270
|
return context, static_tensors
|
|
269
271
|
|
|
@@ -317,35 +319,12 @@ class RBLNDecoderOnlyModel(RBLNModel, RBLNDecoderOnlyFlashAttentionMixin):
|
|
|
317
319
|
|
|
318
320
|
@classmethod
|
|
319
321
|
def get_pytorch_model(
|
|
320
|
-
cls,
|
|
321
|
-
model_id: str,
|
|
322
|
-
*args,
|
|
323
|
-
rbln_config: Optional[RBLNDecoderOnlyModelConfig] = None,
|
|
324
|
-
num_hidden_layers: Optional[int] = None,
|
|
325
|
-
trust_remote_code: Optional[bool] = None,
|
|
326
|
-
torch_dtype: Optional[torch.dtype] = None,
|
|
327
|
-
dtype: Optional[torch.dtype] = None,
|
|
328
|
-
**kwargs,
|
|
322
|
+
cls, *args, rbln_config: Optional[RBLNDecoderOnlyModelConfig] = None, **kwargs
|
|
329
323
|
) -> PreTrainedModel:
|
|
330
324
|
if rbln_config and rbln_config.quantization:
|
|
331
|
-
model = cls.get_quantized_model(
|
|
325
|
+
model = cls.get_quantized_model(*args, rbln_config=rbln_config, **kwargs)
|
|
332
326
|
else:
|
|
333
|
-
|
|
334
|
-
if num_hidden_layers is not None:
|
|
335
|
-
config, kwargs = AutoConfig.from_pretrained(
|
|
336
|
-
model_id,
|
|
337
|
-
return_unused_kwargs=True,
|
|
338
|
-
trust_remote_code=trust_remote_code,
|
|
339
|
-
num_hidden_layers=num_hidden_layers,
|
|
340
|
-
**kwargs,
|
|
341
|
-
)
|
|
342
|
-
if hasattr(config, "layer_types"):
|
|
343
|
-
config.layer_types = config.layer_types[:num_hidden_layers]
|
|
344
|
-
kwargs["config"] = config
|
|
345
|
-
|
|
346
|
-
model = super().get_pytorch_model(
|
|
347
|
-
model_id, *args, trust_remote_code=trust_remote_code, torch_dtype=torch_dtype, dtype=dtype, **kwargs
|
|
348
|
-
)
|
|
327
|
+
model = super().get_pytorch_model(*args, **kwargs)
|
|
349
328
|
|
|
350
329
|
return model
|
|
351
330
|
|
|
@@ -56,7 +56,14 @@ class RBLNWhisperGenerationMixin(WhisperGenerationMixin, GenerationMixin):
|
|
|
56
56
|
return super().generate(*args, **kwargs)
|
|
57
57
|
|
|
58
58
|
def _postprocess_outputs(
|
|
59
|
-
self,
|
|
59
|
+
self,
|
|
60
|
+
seek_outputs,
|
|
61
|
+
decoder_input_ids,
|
|
62
|
+
return_token_timestamps,
|
|
63
|
+
generation_config,
|
|
64
|
+
is_shortform,
|
|
65
|
+
seek,
|
|
66
|
+
batch_idx_map,
|
|
60
67
|
):
|
|
61
68
|
# remove all previously passed decoder input ids
|
|
62
69
|
# should happen only if it is the first generated segment
|
|
@@ -74,6 +81,11 @@ class RBLNWhisperGenerationMixin(WhisperGenerationMixin, GenerationMixin):
|
|
|
74
81
|
|
|
75
82
|
if return_token_timestamps and hasattr(generation_config, "alignment_heads"):
|
|
76
83
|
num_frames = getattr(generation_config, "num_frames", None)
|
|
84
|
+
|
|
85
|
+
if num_frames is not None:
|
|
86
|
+
num_frames = num_frames - seek
|
|
87
|
+
num_frames = num_frames[batch_idx_map]
|
|
88
|
+
|
|
77
89
|
if version.parse(transformers.__version__) >= version.parse("4.46.0"):
|
|
78
90
|
seek_outputs["token_timestamps"] = self._extract_token_timestamps(
|
|
79
91
|
seek_outputs,
|
|
@@ -167,14 +167,12 @@ class ContextRblnConfig:
|
|
|
167
167
|
device=None,
|
|
168
168
|
device_map=None,
|
|
169
169
|
create_runtimes=None,
|
|
170
|
-
optimize_host_mem=None,
|
|
171
170
|
activate_profiler=None,
|
|
172
171
|
timeout=None,
|
|
173
172
|
):
|
|
174
173
|
self.device = device
|
|
175
174
|
self.device_map = device_map
|
|
176
175
|
self.create_runtimes = create_runtimes
|
|
177
|
-
self.optimize_host_mem = optimize_host_mem
|
|
178
176
|
self.activate_profiler = activate_profiler
|
|
179
177
|
self.timeout = timeout
|
|
180
178
|
|
|
@@ -182,7 +180,6 @@ class ContextRblnConfig:
|
|
|
182
180
|
self._local.device = self.device
|
|
183
181
|
self._local.device_map = self.device_map
|
|
184
182
|
self._local.create_runtimes = self.create_runtimes
|
|
185
|
-
self._local.optimize_host_memory = self.optimize_host_mem
|
|
186
183
|
self._local.activate_profiler = self.activate_profiler
|
|
187
184
|
self._local.timeout = self.timeout
|
|
188
185
|
return self
|
|
@@ -191,7 +188,6 @@ class ContextRblnConfig:
|
|
|
191
188
|
self._local.device = None
|
|
192
189
|
self._local.device_map = None
|
|
193
190
|
self._local.create_runtimes = None
|
|
194
|
-
self._local.optimize_host_memory = None
|
|
195
191
|
self._local.activate_profiler = None
|
|
196
192
|
self._local.timeout = None
|
|
197
193
|
|
|
@@ -201,7 +197,6 @@ class ContextRblnConfig:
|
|
|
201
197
|
"device": getattr(cls._local, "device", None),
|
|
202
198
|
"device_map": getattr(cls._local, "device_map", None),
|
|
203
199
|
"create_runtimes": getattr(cls._local, "create_runtimes", None),
|
|
204
|
-
"optimize_host_memory": getattr(cls._local, "optimize_host_memory", None),
|
|
205
200
|
"activate_profiler": getattr(cls._local, "activate_profiler", None),
|
|
206
201
|
"timeout": getattr(cls._local, "timeout", None),
|
|
207
202
|
}
|
|
@@ -34,10 +34,6 @@ def stable_diffusion_model():
|
|
|
34
34
|
"batch_size": 1,
|
|
35
35
|
"npu": "RBLN-CA22",
|
|
36
36
|
"create_runtimes": False,
|
|
37
|
-
"optimize_host_memory": False,
|
|
38
|
-
},
|
|
39
|
-
"text_encoder": {
|
|
40
|
-
"optimize_host_memory": False,
|
|
41
37
|
},
|
|
42
38
|
},
|
|
43
39
|
)
|
|
@@ -50,7 +46,6 @@ def test_stable_diffusion_config(stable_diffusion_model):
|
|
|
50
46
|
assert model.unet.rbln_config.batch_size == 1
|
|
51
47
|
assert model.unet.rbln_config.npu == "RBLN-CA22"
|
|
52
48
|
assert model.unet.rbln_config.create_runtimes is False
|
|
53
|
-
assert model.unet.rbln_config.optimize_host_memory is False
|
|
54
49
|
assert model.unet.compiled_models[0]._meta["npu"] == "RBLN-CA22"
|
|
55
50
|
|
|
56
51
|
npu = rebel.get_npu_name()
|
|
@@ -69,7 +64,7 @@ def test_explicit_config_parameters(model_id):
|
|
|
69
64
|
|
|
70
65
|
def test_config_dict(model_id):
|
|
71
66
|
"""Test loading model with configuration passed as a dictionary."""
|
|
72
|
-
rbln_config = {"create_runtimes": False, "
|
|
67
|
+
rbln_config = {"create_runtimes": False, "image_size": 64}
|
|
73
68
|
|
|
74
69
|
model = RBLNResNetForImageClassification.from_pretrained(model_id, rbln_config=rbln_config)
|
|
75
70
|
assert model is not None
|
|
@@ -1507,7 +1507,7 @@ tests = [
|
|
|
1507
1507
|
{ name = "sentencepiece", specifier = ">=0.2.0" },
|
|
1508
1508
|
{ name = "simphile", specifier = ">=1.0.2" },
|
|
1509
1509
|
{ name = "soundfile", specifier = ">=0.13.1" },
|
|
1510
|
-
{ name = "torchcodec", specifier = "==0.
|
|
1510
|
+
{ name = "torchcodec", specifier = "==0.7" },
|
|
1511
1511
|
]
|
|
1512
1512
|
|
|
1513
1513
|
[[package]]
|
|
@@ -2667,14 +2667,14 @@ wheels = [
|
|
|
2667
2667
|
|
|
2668
2668
|
[[package]]
|
|
2669
2669
|
name = "torchcodec"
|
|
2670
|
-
version = "0.
|
|
2670
|
+
version = "0.7.0"
|
|
2671
2671
|
source = { registry = "https://pypi.org/simple" }
|
|
2672
2672
|
wheels = [
|
|
2673
|
-
{ url = "https://files.pythonhosted.org/packages/
|
|
2674
|
-
{ url = "https://files.pythonhosted.org/packages/
|
|
2675
|
-
{ url = "https://files.pythonhosted.org/packages/
|
|
2676
|
-
{ url = "https://files.pythonhosted.org/packages/
|
|
2677
|
-
{ url = "https://files.pythonhosted.org/packages/
|
|
2673
|
+
{ url = "https://files.pythonhosted.org/packages/0d/81/cff42793544b7d3e2ff9a4912542c6d1c7a617aabe8404f8fd3d52453f20/torchcodec-0.7.0-cp310-cp310-manylinux_2_28_x86_64.whl", hash = "sha256:a0071096724e8ded6a171457ce4680f646499b4a4d285cdb46e130983f965ce4", size = 1411823, upload-time = "2025-09-08T14:17:39.405Z" },
|
|
2674
|
+
{ url = "https://files.pythonhosted.org/packages/06/14/8ff28247988365fc47e8471e28cdfd8d037232fcf73abb67ee815ac80f1d/torchcodec-0.7.0-cp311-cp311-manylinux_2_28_x86_64.whl", hash = "sha256:afb1c48b52bd4ee8f485f5a427bb4e82380590255a26b8e9e3fe099e0779287f", size = 1419444, upload-time = "2025-09-08T14:17:41.479Z" },
|
|
2675
|
+
{ url = "https://files.pythonhosted.org/packages/41/10/4a1a8407d0fad37cb43d1f749e7b422e5a0f6def17f3b90ab9ab9a105e32/torchcodec-0.7.0-cp312-cp312-manylinux_2_28_x86_64.whl", hash = "sha256:3cd23c3296c9b071d56bb2c534a6a98275d65c1a6a7213cdb72a26ec9f9d2fd8", size = 1421871, upload-time = "2025-09-08T14:17:43.419Z" },
|
|
2676
|
+
{ url = "https://files.pythonhosted.org/packages/01/25/177ea01d138598ab68d5e3b000789e8617bf97874bd8f761d89093f419ba/torchcodec-0.7.0-cp313-cp313-manylinux_2_28_x86_64.whl", hash = "sha256:9c70f910f9f48e6625aacaed534f766e13d447b895dc7299e96d4db9a93f1514", size = 1422493, upload-time = "2025-09-08T14:17:45.41Z" },
|
|
2677
|
+
{ url = "https://files.pythonhosted.org/packages/e1/85/d4d3ac080325cdae3c8144de41afbf98bbf0f5e15e6c78644a70c5721ee6/torchcodec-0.7.0-cp39-cp39-manylinux_2_28_x86_64.whl", hash = "sha256:d2b10c1ad9bfdc21c75b1794566a0d68131f087dd609c27ee707efb035a45198", size = 1408250, upload-time = "2025-09-08T14:17:47.095Z" },
|
|
2678
2678
|
]
|
|
2679
2679
|
|
|
2680
2680
|
[[package]]
|
|
@@ -1 +0,0 @@
|
|
|
1
|
-
rebel_compiler_version: 0.9.2.dev155+g1ab49983.prod
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
{optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/image-to-text/run_llava_next_image_to_text.py
RENAMED
|
File without changes
|
|
File without changes
|
{optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/kandinsky2_2/run_kandinsky2_2_combined.py
RENAMED
|
File without changes
|
{optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/kandinsky2_2/run_kandinsky2_2_img2img.py
RENAMED
|
File without changes
|
|
File without changes
|
{optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/kandinsky2_2/run_kandinsky2_2_inpaint.py
RENAMED
|
File without changes
|
|
File without changes
|
|
File without changes
|
{optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/question-answering/run_question_answering.py
RENAMED
|
File without changes
|
|
File without changes
|
|
File without changes
|
{optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/stable-diffusion/run_stable_diffusion.py
RENAMED
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
{optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/stable-diffusion/run_stable_diffusion_lora.py
RENAMED
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
{optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/text-classification/run_secureBERT.py
RENAMED
|
File without changes
|
{optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/text-classification/run_t5_classification.py
RENAMED
|
File without changes
|
|
File without changes
|
|
File without changes
|
{optimum_rbln-0.9.2a6 → optimum_rbln-0.9.2a8}/examples/text2text-generation/run_llama_peft.py
RENAMED
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|