optimum-rbln 0.8.2a6__tar.gz → 0.8.2rc0__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of optimum-rbln might be problematic. Click here for more details.

Files changed (300) hide show
  1. optimum_rbln-0.8.2rc0/.github/version.yaml +1 -0
  2. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/PKG-INFO +1 -1
  3. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/__init__.py +4 -0
  4. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/__version__.py +2 -2
  5. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/__init__.py +4 -0
  6. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/__init__.py +4 -0
  7. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/decoderonly/decoderonly_architecture.py +1 -2
  8. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/decoderonly/modeling_decoderonly.py +1 -1
  9. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/gemma3/configuration_gemma3.py +11 -0
  10. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/gemma3/modeling_gemma3.py +66 -34
  11. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/llava/modeling_llava.py +6 -5
  12. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/tests/test_base.py +7 -2
  13. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/tests/test_llm.py +1 -3
  14. optimum_rbln-0.8.2a6/.github/version.yaml +0 -1
  15. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/.github/ISSUE_TEMPLATE/bug_report.md +0 -0
  16. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/.github/ISSUE_TEMPLATE/config.yml +0 -0
  17. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/.github/ISSUE_TEMPLATE/feature_request.md +0 -0
  18. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/.github/ISSUE_TEMPLATE/model_request.md +0 -0
  19. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/.github/pull_request_template.md +0 -0
  20. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/.github/scripts/auto_code_review.py +0 -0
  21. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/.github/scripts/validate_docstrings.py +0 -0
  22. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/.github/scripts/validate_pr_checklist.py +0 -0
  23. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/.github/workflows/auto_code_review.yml +0 -0
  24. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/.github/workflows/check_code_quality.yml +0 -0
  25. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/.github/workflows/deploy-on-tag.yaml +0 -0
  26. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/.github/workflows/deploy.yaml +0 -0
  27. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/.github/workflows/pr-title-check.yaml +0 -0
  28. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/.github/workflows/pr_checklist_validator.yml +0 -0
  29. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/.github/workflows/rbln_check_compiler.yaml +0 -0
  30. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/.github/workflows/rbln_dispatch_pytest.yaml +0 -0
  31. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/.github/workflows/rbln_optimum_inference_test.yaml +0 -0
  32. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/.github/workflows/rbln_optimum_pytest.yaml +0 -0
  33. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/.github/workflows/rbln_scheduled_test.yaml +0 -0
  34. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/.github/workflows/rbln_trigger_on_pr.yaml +0 -0
  35. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/.github/workflows/test-docstrings.yml +0 -0
  36. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/.gitignore +0 -0
  37. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/CODE_OF_CONDUCT.md +0 -0
  38. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/CONTRIBUTING.md +0 -0
  39. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/LICENSE +0 -0
  40. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/README.md +0 -0
  41. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/assets/rbln_logo.png +0 -0
  42. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/advanced/custom_class.py +0 -0
  43. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/audio-classification/run_ast_audio_classification.py +0 -0
  44. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/depth-estimation/run_dpt.py +0 -0
  45. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/image-classification/run_image_classification.py +0 -0
  46. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/image-classification/run_vit_image_classification.py +0 -0
  47. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/image-to-text/run_idefics3.py +0 -0
  48. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/image-to-text/run_llava_next_image_to_text.py +0 -0
  49. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/kandinsky2_2/run_kandinsky2_2.py +0 -0
  50. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/kandinsky2_2/run_kandinsky2_2_combined.py +0 -0
  51. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/kandinsky2_2/run_kandinsky2_2_img2img.py +0 -0
  52. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/kandinsky2_2/run_kandinsky2_2_img2img_combined.py +0 -0
  53. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/kandinsky2_2/run_kandinsky2_2_inpaint.py +0 -0
  54. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/kandinsky2_2/run_kandinsky2_2_inpaint_combined.py +0 -0
  55. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/kandinsky2_2/run_kandinsky2_2_prior_interpolate.py +0 -0
  56. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/question-answering/run_question_answering.py +0 -0
  57. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/speech-recognition/run_wav2vec2.py +0 -0
  58. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/speech-recognition/run_whisper.py +0 -0
  59. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/stable-diffusion/run_stable_diffusion.py +0 -0
  60. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/stable-diffusion/run_stable_diffusion_controlnet.py +0 -0
  61. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/stable-diffusion/run_stable_diffusion_img2img.py +0 -0
  62. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/stable-diffusion/run_stable_diffusion_img2img_controlnet.py +0 -0
  63. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/stable-diffusion/run_stable_diffusion_inpaint.py +0 -0
  64. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/stable-diffusion/run_stable_diffusion_lora.py +0 -0
  65. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/stable-diffusion/run_stable_diffusion_multicontrolnet.py +0 -0
  66. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/text-classification/run_bge_m3_text_classification.py +0 -0
  67. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/text-classification/run_bge_reranker_v2_m3_text_classification.py +0 -0
  68. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/text-classification/run_secureBERT.py +0 -0
  69. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/text-classification/run_t5_classification.py +0 -0
  70. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/text-classification/run_twitter_roberta_text_classification.py +0 -0
  71. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/text2text-generation/run_bart_text2text_generation.py +0 -0
  72. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/text2text-generation/run_llama_peft.py +0 -0
  73. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/text2text-generation/run_llama_text2text_generation.py +0 -0
  74. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/examples/time-series-forecasting/run_time_series_forecasting.py +0 -0
  75. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/pyproject.toml +0 -0
  76. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/scripts/uv-lock.sh +0 -0
  77. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/scripts/uv-sync.sh +0 -0
  78. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/configuration_utils.py +0 -0
  79. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/__init__.py +0 -0
  80. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/configurations/__init__.py +0 -0
  81. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/configurations/models/__init__.py +0 -0
  82. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/configurations/models/configuration_autoencoder_kl.py +0 -0
  83. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/configurations/models/configuration_autoencoder_kl_cosmos.py +0 -0
  84. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/configurations/models/configuration_controlnet.py +0 -0
  85. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/configurations/models/configuration_prior_transformer.py +0 -0
  86. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/configurations/models/configuration_transformer_cosmos.py +0 -0
  87. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/configurations/models/configuration_transformer_sd3.py +0 -0
  88. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/configurations/models/configuration_unet_2d_condition.py +0 -0
  89. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/configurations/models/configuration_vq_model.py +0 -0
  90. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/configurations/pipelines/__init__.py +0 -0
  91. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/configurations/pipelines/configuration_controlnet.py +0 -0
  92. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/configurations/pipelines/configuration_cosmos.py +0 -0
  93. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/configurations/pipelines/configuration_kandinsky2_2.py +0 -0
  94. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/configurations/pipelines/configuration_stable_diffusion.py +0 -0
  95. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/configurations/pipelines/configuration_stable_diffusion_3.py +0 -0
  96. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/configurations/pipelines/configuration_stable_diffusion_xl.py +0 -0
  97. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/modeling_diffusers.py +0 -0
  98. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/models/__init__.py +0 -0
  99. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/models/autoencoders/__init__.py +0 -0
  100. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/models/autoencoders/autoencoder_kl.py +0 -0
  101. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/models/autoencoders/autoencoder_kl_cosmos.py +0 -0
  102. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/models/autoencoders/vae.py +0 -0
  103. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/models/autoencoders/vq_model.py +0 -0
  104. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/models/controlnet.py +0 -0
  105. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/models/transformers/__init__.py +0 -0
  106. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/models/transformers/prior_transformer.py +0 -0
  107. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/models/transformers/transformer_cosmos.py +0 -0
  108. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/models/transformers/transformer_sd3.py +0 -0
  109. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/models/unets/__init__.py +0 -0
  110. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/models/unets/unet_2d_condition.py +0 -0
  111. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/pipelines/__init__.py +0 -0
  112. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/pipelines/controlnet/__init__.py +0 -0
  113. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/pipelines/controlnet/multicontrolnet.py +0 -0
  114. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/pipelines/controlnet/pipeline_controlnet.py +0 -0
  115. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/pipelines/controlnet/pipeline_controlnet_img2img.py +0 -0
  116. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/pipelines/controlnet/pipeline_controlnet_sd_xl.py +0 -0
  117. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/pipelines/controlnet/pipeline_controlnet_sd_xl_img2img.py +0 -0
  118. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/pipelines/cosmos/__init__.py +0 -0
  119. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/pipelines/cosmos/configuration_cosmos_guardrail.py +0 -0
  120. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/pipelines/cosmos/cosmos_guardrail.py +0 -0
  121. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/pipelines/cosmos/pipeline_cosmos_text2world.py +0 -0
  122. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/pipelines/cosmos/pipeline_cosmos_video2world.py +0 -0
  123. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/pipelines/kandinsky2_2/__init__.py +0 -0
  124. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2.py +0 -0
  125. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2_combined.py +0 -0
  126. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2_img2img.py +0 -0
  127. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2_inpaint.py +0 -0
  128. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/pipelines/kandinsky2_2/pipeline_kandinsky2_2_prior.py +0 -0
  129. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/pipelines/stable_diffusion/__init__.py +0 -0
  130. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/pipelines/stable_diffusion/pipeline_stable_diffusion.py +0 -0
  131. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/pipelines/stable_diffusion/pipeline_stable_diffusion_img2img.py +0 -0
  132. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/pipelines/stable_diffusion/pipeline_stable_diffusion_inpaint.py +0 -0
  133. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/pipelines/stable_diffusion_3/__init__.py +0 -0
  134. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/pipelines/stable_diffusion_3/pipeline_stable_diffusion_3.py +0 -0
  135. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/pipelines/stable_diffusion_3/pipeline_stable_diffusion_3_img2img.py +0 -0
  136. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/pipelines/stable_diffusion_3/pipeline_stable_diffusion_3_inpaint.py +0 -0
  137. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/pipelines/stable_diffusion_xl/__init__.py +0 -0
  138. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/pipelines/stable_diffusion_xl/pipeline_stable_diffusion_xl.py +0 -0
  139. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/pipelines/stable_diffusion_xl/pipeline_stable_diffusion_xl_img2img.py +0 -0
  140. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/diffusers/pipelines/stable_diffusion_xl/pipeline_stable_diffusion_xl_inpaint.py +0 -0
  141. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/modeling.py +0 -0
  142. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/modeling_base.py +0 -0
  143. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/ops/__init__.py +0 -0
  144. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/ops/attn.py +0 -0
  145. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/ops/flash_attn.py +0 -0
  146. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/ops/kv_cache_update.py +0 -0
  147. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/ops/linear.py +0 -0
  148. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/ops/sliding_window_attn.py +0 -0
  149. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/configuration_generic.py +0 -0
  150. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/modeling_attention_utils.py +0 -0
  151. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/modeling_generic.py +0 -0
  152. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/modeling_rope_utils.py +0 -0
  153. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/audio_spectrogram_transformer/__init__.py +0 -0
  154. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/audio_spectrogram_transformer/configuration_audio_spectrogram_transformer.py +0 -0
  155. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/audio_spectrogram_transformer/modeling_audio_spectrogram_transformer.py +0 -0
  156. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/auto/__init__.py +0 -0
  157. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/auto/auto_factory.py +0 -0
  158. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/auto/modeling_auto.py +0 -0
  159. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/bart/__init__.py +0 -0
  160. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/bart/bart_architecture.py +0 -0
  161. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/bart/configuration_bart.py +0 -0
  162. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/bart/modeling_bart.py +0 -0
  163. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/bert/__init__.py +0 -0
  164. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/bert/configuration_bert.py +0 -0
  165. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/bert/modeling_bert.py +0 -0
  166. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/blip_2/__init__.py +0 -0
  167. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/blip_2/configuration_blip_2.py +0 -0
  168. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/blip_2/modeling_blip_2.py +0 -0
  169. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/clip/__init__.py +0 -0
  170. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/clip/configuration_clip.py +0 -0
  171. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/clip/modeling_clip.py +0 -0
  172. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/colpali/__init__.py +0 -0
  173. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/colpali/colpali_architecture.py +0 -0
  174. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/colpali/configuration_colpali.py +0 -0
  175. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/colpali/modeling_colpali.py +0 -0
  176. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/decoderonly/__init__.py +0 -0
  177. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/decoderonly/configuration_decoderonly.py +0 -0
  178. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/distilbert/__init__.py +0 -0
  179. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/distilbert/configuration_distilbert.py +0 -0
  180. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/distilbert/modeling_distilbert.py +0 -0
  181. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/dpt/__init__.py +0 -0
  182. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/dpt/configuration_dpt.py +0 -0
  183. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/dpt/modeling_dpt.py +0 -0
  184. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/exaone/__init__.py +0 -0
  185. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/exaone/configuration_exaone.py +0 -0
  186. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/exaone/exaone_architecture.py +0 -0
  187. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/exaone/modeling_exaone.py +0 -0
  188. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/gemma/__init__.py +0 -0
  189. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/gemma/configuration_gemma.py +0 -0
  190. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/gemma/gemma_architecture.py +0 -0
  191. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/gemma/modeling_gemma.py +0 -0
  192. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/gemma3/__init__.py +0 -0
  193. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/gemma3/gemma3_architecture.py +0 -0
  194. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/gpt2/__init__.py +0 -0
  195. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/gpt2/configuration_gpt2.py +0 -0
  196. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/gpt2/gpt2_architecture.py +0 -0
  197. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/gpt2/modeling_gpt2.py +0 -0
  198. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/idefics3/__init__.py +0 -0
  199. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/idefics3/configuration_idefics3.py +0 -0
  200. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/idefics3/modeling_idefics3.py +0 -0
  201. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/llama/__init__.py +0 -0
  202. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/llama/configuration_llama.py +0 -0
  203. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/llama/llama_architecture.py +0 -0
  204. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/llama/modeling_llama.py +0 -0
  205. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/llava/__init__.py +0 -0
  206. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/llava/configuration_llava.py +0 -0
  207. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/llava_next/__init__.py +0 -0
  208. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/llava_next/configuration_llava_next.py +0 -0
  209. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/llava_next/modeling_llava_next.py +0 -0
  210. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/midm/__init__.py +0 -0
  211. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/midm/configuration_midm.py +0 -0
  212. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/midm/midm_architecture.py +0 -0
  213. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/midm/modeling_midm.py +0 -0
  214. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/mistral/__init__.py +0 -0
  215. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/mistral/configuration_mistral.py +0 -0
  216. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/mistral/mistral_architecture.py +0 -0
  217. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/mistral/modeling_mistral.py +0 -0
  218. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/opt/__init__.py +0 -0
  219. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/opt/configuration_opt.py +0 -0
  220. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/opt/modeling_opt.py +0 -0
  221. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/opt/opt_architecture.py +0 -0
  222. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/pegasus/__init__.py +0 -0
  223. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/pegasus/configuration_pegasus.py +0 -0
  224. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/pegasus/modeling_pegasus.py +0 -0
  225. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/pegasus/pegasus_architecture.py +0 -0
  226. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/phi/__init__.py +0 -0
  227. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/phi/configuration_phi.py +0 -0
  228. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/phi/modeling_phi.py +0 -0
  229. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/phi/phi_architecture.py +0 -0
  230. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/pixtral/__init__.py +0 -0
  231. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/pixtral/configuration_pixtral.py +0 -0
  232. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/pixtral/modeling_pixtral.py +0 -0
  233. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/pixtral/pixtral_architecture.py +0 -0
  234. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/qwen2/__init__.py +0 -0
  235. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/qwen2/configuration_qwen2.py +0 -0
  236. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/qwen2/modeling_qwen2.py +0 -0
  237. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/qwen2/qwen2_architecture.py +0 -0
  238. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/qwen2_5_vl/__init__.py +0 -0
  239. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/qwen2_5_vl/configuration_qwen2_5_vl.py +0 -0
  240. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/qwen2_5_vl/modeling_qwen2_5_vl.py +0 -0
  241. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/qwen2_5_vl/qwen2_5_vl_architecture.py +0 -0
  242. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/qwen3/__init__.py +0 -0
  243. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/qwen3/configuration_qwen3.py +0 -0
  244. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/qwen3/modeling_qwen3.py +0 -0
  245. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/qwen3/qwen3_architecture.py +0 -0
  246. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/resnet/__init__.py +0 -0
  247. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/resnet/configuration_resnet.py +0 -0
  248. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/resnet/modeling_resnet.py +0 -0
  249. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/roberta/__init__.py +0 -0
  250. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/roberta/configuration_roberta.py +0 -0
  251. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/roberta/modeling_roberta.py +0 -0
  252. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/seq2seq/__init__.py +0 -0
  253. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/seq2seq/configuration_seq2seq.py +0 -0
  254. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/seq2seq/modeling_seq2seq.py +0 -0
  255. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/seq2seq/seq2seq_architecture.py +0 -0
  256. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/siglip/__init__.py +0 -0
  257. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/siglip/configuration_siglip.py +0 -0
  258. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/siglip/modeling_siglip.py +0 -0
  259. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/t5/__init__.py +0 -0
  260. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/t5/configuration_t5.py +0 -0
  261. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/t5/modeling_t5.py +0 -0
  262. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/t5/t5_architecture.py +0 -0
  263. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/time_series_transformer/__init__.py +0 -0
  264. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/time_series_transformer/configuration_time_series_transformer.py +0 -0
  265. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/time_series_transformer/modeling_time_series_transformer.py +0 -0
  266. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/time_series_transformer/time_series_transformers_architecture.py +0 -0
  267. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/vit/__init__.py +0 -0
  268. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/vit/configuration_vit.py +0 -0
  269. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/vit/modeling_vit.py +0 -0
  270. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/wav2vec2/__init__.py +0 -0
  271. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/wav2vec2/configuration_wav2vec2.py +0 -0
  272. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/wav2vec2/modeling_wav2vec2.py +0 -0
  273. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/whisper/__init__.py +0 -0
  274. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/whisper/configuration_whisper.py +0 -0
  275. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/whisper/generation_whisper.py +0 -0
  276. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/whisper/modeling_whisper.py +0 -0
  277. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/whisper/whisper_architecture.py +0 -0
  278. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/xlm_roberta/__init__.py +0 -0
  279. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/xlm_roberta/configuration_xlm_roberta.py +0 -0
  280. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/models/xlm_roberta/modeling_xlm_roberta.py +0 -0
  281. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/utils/__init__.py +0 -0
  282. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/transformers/utils/rbln_quantization.py +0 -0
  283. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/utils/__init__.py +0 -0
  284. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/utils/decorator_utils.py +0 -0
  285. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/utils/depreacate_utils.py +0 -0
  286. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/utils/hub.py +0 -0
  287. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/utils/import_utils.py +0 -0
  288. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/utils/logging.py +0 -0
  289. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/utils/model_utils.py +0 -0
  290. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/utils/runtime_utils.py +0 -0
  291. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/utils/save_utils.py +0 -0
  292. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/src/optimum/rbln/utils/submodule.py +0 -0
  293. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/tests/__init__.py +0 -0
  294. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/tests/psnr.py +0 -0
  295. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/tests/requirements_sdxl.txt +0 -0
  296. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/tests/run_stable_diffusion_xl_base.py +0 -0
  297. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/tests/test_config.py +0 -0
  298. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/tests/test_diffusers.py +0 -0
  299. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/tests/test_transformers.py +0 -0
  300. {optimum_rbln-0.8.2a6 → optimum_rbln-0.8.2rc0}/uv.lock +0 -0
@@ -0,0 +1 @@
1
+ rebel_compiler_version: 0.8.2.dev187+g9f5b6c9b
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: optimum-rbln
3
- Version: 0.8.2a6
3
+ Version: 0.8.2rc0
4
4
  Summary: Optimum RBLN is the interface between the HuggingFace Transformers and Diffusers libraries and RBLN accelerators. It provides a set of tools enabling easy model loading and inference on single and multiple rbln device settings for different downstream tasks.
5
5
  Project-URL: Homepage, https://rebellions.ai
6
6
  Project-URL: Documentation, https://docs.rbln.ai
@@ -72,6 +72,8 @@ _import_structure = {
72
72
  "RBLNCLIPVisionModelWithProjectionConfig",
73
73
  "RBLNColPaliForRetrieval",
74
74
  "RBLNColPaliForRetrievalConfig",
75
+ "RBLNDecoderOnlyModelConfig",
76
+ "RBLNDecoderOnlyModel",
75
77
  "RBLNDecoderOnlyModelForCausalLM",
76
78
  "RBLNDecoderOnlyModelForCausalLMConfig",
77
79
  "RBLNDistilBertForQuestionAnswering",
@@ -345,6 +347,8 @@ if TYPE_CHECKING:
345
347
  RBLNCLIPVisionModelWithProjectionConfig,
346
348
  RBLNColPaliForRetrieval,
347
349
  RBLNColPaliForRetrievalConfig,
350
+ RBLNDecoderOnlyModel,
351
+ RBLNDecoderOnlyModelConfig,
348
352
  RBLNDecoderOnlyModelForCausalLM,
349
353
  RBLNDecoderOnlyModelForCausalLMConfig,
350
354
  RBLNDistilBertForQuestionAnswering,
@@ -17,5 +17,5 @@ __version__: str
17
17
  __version_tuple__: VERSION_TUPLE
18
18
  version_tuple: VERSION_TUPLE
19
19
 
20
- __version__ = version = '0.8.2a6'
21
- __version_tuple__ = version_tuple = (0, 8, 2, 'a6')
20
+ __version__ = version = '0.8.2rc0'
21
+ __version_tuple__ = version_tuple = (0, 8, 2, 'rc0')
@@ -62,6 +62,8 @@ _import_structure = {
62
62
  "RBLNCLIPVisionModelWithProjectionConfig",
63
63
  "RBLNDecoderOnlyModelForCausalLM",
64
64
  "RBLNDecoderOnlyModelForCausalLMConfig",
65
+ "RBLNDecoderOnlyModelConfig",
66
+ "RBLNDecoderOnlyModel",
65
67
  "RBLNDistilBertForQuestionAnswering",
66
68
  "RBLNDistilBertForQuestionAnsweringConfig",
67
69
  "RBLNDPTForDepthEstimation",
@@ -196,6 +198,8 @@ if TYPE_CHECKING:
196
198
  RBLNCLIPVisionModelWithProjectionConfig,
197
199
  RBLNColPaliForRetrieval,
198
200
  RBLNColPaliForRetrievalConfig,
201
+ RBLNDecoderOnlyModel,
202
+ RBLNDecoderOnlyModelConfig,
199
203
  RBLNDecoderOnlyModelForCausalLM,
200
204
  RBLNDecoderOnlyModelForCausalLMConfig,
201
205
  RBLNDistilBertForQuestionAnswering,
@@ -84,6 +84,8 @@ _import_structure = {
84
84
  "RBLNQwen2_5_VLForConditionalGenerationConfig",
85
85
  ],
86
86
  "decoderonly": [
87
+ "RBLNDecoderOnlyModelConfig",
88
+ "RBLNDecoderOnlyModel",
87
89
  "RBLNDecoderOnlyModelForCausalLM",
88
90
  "RBLNDecoderOnlyModelForCausalLMConfig",
89
91
  ],
@@ -216,6 +218,8 @@ if TYPE_CHECKING:
216
218
  RBLNColPaliForRetrievalConfig,
217
219
  )
218
220
  from .decoderonly import (
221
+ RBLNDecoderOnlyModel,
222
+ RBLNDecoderOnlyModelConfig,
219
223
  RBLNDecoderOnlyModelForCausalLM,
220
224
  RBLNDecoderOnlyModelForCausalLMConfig,
221
225
  )
@@ -1025,8 +1025,7 @@ class SlidingWindowAttentionOp(AttentionOp):
1025
1025
  }
1026
1026
 
1027
1027
  if self.phase == "prefill" or self.phase == "image_prefill":
1028
- if not self.use_attention_mask or self.use_position_ids:
1029
- op_args["is_bidirectional"] = self.phase == "image_prefill" # FIXME, Hard-coded for Gemma3.
1028
+ op_args["is_bidirectional"] = self.phase == "image_prefill" # FIXME, Hard-coded for Gemma3.
1030
1029
 
1031
1030
  attn_op_name = self.get_attn_op_name()
1032
1031
  attn_op = getattr(torch.ops.rbln_custom_ops, attn_op_name, None)
@@ -1403,7 +1403,7 @@ class RBLNDecoderOnlyModelForCausalLM(RBLNDecoderOnlyModel):
1403
1403
  )
1404
1404
  padded_cache_lengths = torch.zeros_like(generate_idx)
1405
1405
 
1406
- # Prefll
1406
+ # Prefill
1407
1407
  if cache_position is None:
1408
1408
  logits = []
1409
1409
  inputs = inputs_embeds if inputs_embeds is not None else input_ids
@@ -23,14 +23,17 @@ class RBLNGemma3ForCausalLMConfig(RBLNDecoderOnlyModelForCausalLMConfig):
23
23
  self,
24
24
  use_position_ids: Optional[bool] = None,
25
25
  use_attention_mask: Optional[bool] = None,
26
+ prefill_chunk_size: Optional[int] = None,
26
27
  image_prefill_chunk_size: Optional[int] = None,
27
28
  **kwargs: Dict[str, Any],
28
29
  ):
29
30
  # use_attention_mask and use_position_ids are always True for Gemma3
30
31
  use_attention_mask = use_attention_mask or True
31
32
  use_position_ids = use_position_ids or True
33
+ prefill_chunk_size = prefill_chunk_size or 256
32
34
 
33
35
  super().__init__(
36
+ prefill_chunk_size=prefill_chunk_size,
34
37
  use_attention_mask=use_attention_mask,
35
38
  use_position_ids=use_position_ids,
36
39
  **kwargs,
@@ -73,3 +76,11 @@ class RBLNGemma3ForConditionalGenerationConfig(RBLNModelConfig):
73
76
 
74
77
  self.vision_tower = self.init_submodule_config(RBLNSiglipVisionModelConfig, vision_tower)
75
78
  self.language_model = self.init_submodule_config(RBLNGemma3ForCausalLMConfig, language_model)
79
+
80
+ @property
81
+ def image_prefill_chunk_size(self):
82
+ return self.language_model.image_prefill_chunk_size
83
+
84
+ @property
85
+ def prefill_chunk_size(self):
86
+ return self.language_model.prefill_chunk_size
@@ -258,17 +258,45 @@ class RBLNGemma3ForConditionalGeneration(RBLNModel):
258
258
 
259
259
  return inputs_embeds
260
260
 
261
+ def get_padded_cache_position(
262
+ self,
263
+ cache_position: torch.Tensor, # shape: [1, seq_len]
264
+ token_type_ids: torch.Tensor, # shape: [1, seq_len]
265
+ ) -> torch.Tensor:
266
+ seq_len = cache_position[0][-1].item() + 1
267
+
268
+ # Find image start positions
269
+ image_starts = [
270
+ s
271
+ for s in torch.where(token_type_ids == 1)[1]
272
+ if torch.all(token_type_ids[:, s : s + self.rbln_config.image_prefill_chunk_size] == 1)
273
+ ]
274
+
275
+ # Initialize padded tensors
276
+ padded_input_len = seq_len
277
+ for image_start in image_starts:
278
+ pad_needed = (
279
+ self.rbln_config.image_prefill_chunk_size
280
+ - (image_start + padded_input_len - seq_len) % self.rbln_config.image_prefill_chunk_size
281
+ ) % self.rbln_config.image_prefill_chunk_size
282
+ padded_input_len += pad_needed
283
+
284
+ return torch.cat(
285
+ [cache_position, torch.arange(seq_len, padded_input_len, dtype=torch.int32).unsqueeze(0)],
286
+ dim=1,
287
+ )
288
+
261
289
  def forward(
262
290
  self,
263
291
  input_ids: torch.LongTensor = None,
292
+ attention_mask: torch.Tensor = None,
293
+ token_type_ids: torch.Tensor = None,
264
294
  pixel_values: torch.FloatTensor = None,
265
- attention_mask: Optional[torch.Tensor] = None,
266
295
  cache_position: Optional[torch.LongTensor] = None,
267
296
  inputs_embeds: Optional[torch.FloatTensor] = None,
268
297
  generate_idx: Optional[torch.Tensor] = None,
269
298
  padded_cache_lengths: Optional[torch.Tensor] = None,
270
299
  position_ids: Optional[torch.Tensor] = None,
271
- token_type_ids: Optional[torch.Tensor] = None,
272
300
  **lm_kwargs: Dict[str, Any],
273
301
  ) -> Union[Tuple, RBLNDecoderOnlyForCausalLMOutput]:
274
302
  # prefill
@@ -279,12 +307,15 @@ class RBLNGemma3ForConditionalGeneration(RBLNModel):
279
307
 
280
308
  for b_idx in range(batch_size):
281
309
  cache_position = torch.arange(0, generate_idx[b_idx].item(), dtype=torch.int32).unsqueeze(0)
310
+ token_type_id = token_type_ids[b_idx : b_idx + 1, attention_mask[b_idx].bool()]
311
+ cache_position = self.get_padded_cache_position(cache_position, token_type_id)
312
+
282
313
  output = self.language_model.prefill_decoder(
283
314
  inputs_embeds=inputs_embeds[b_idx : b_idx + 1],
284
315
  attention_mask=attention_mask[b_idx],
285
316
  cache_position=cache_position,
286
317
  batch_idx=b_idx,
287
- token_type_ids=token_type_ids[b_idx : b_idx + 1] if token_type_ids is not None else None,
318
+ token_type_ids=token_type_ids[b_idx : b_idx + 1], # do not pass token_type_id
288
319
  )
289
320
  padded_cache_lengths[b_idx] += output.padded_cache_lengths
290
321
  logits.append(output.logits)
@@ -336,9 +367,10 @@ class RBLNGemma3RuntimeModel(RBLNRuntimeModel):
336
367
  # chunked_attention_mask shape
337
368
  chunked_attention_mask = torch.zeros(1, chunked_attention_mask.shape[-1], dtype=torch.float32)
338
369
 
339
- # as gemma3 has different prefill chunk size for image and text, we need to pad the inputs to the max of the two.
370
+ # In case of Gemma3ForConditionalGeneration, the loop counter may not be a prefill_chunk_size,
371
+ # so we cannot guarantee that the last chunk starts at a position that is a multiple of prefill_chunk_size.
340
372
  if self.rbln_config.use_image_prefill:
341
- padding_size = max(self.rbln_config.prefill_chunk_size, self.rbln_config.image_prefill_chunk_size)
373
+ padding_size = self.rbln_config.image_prefill_chunk_size
342
374
  inputs = torch.nn.functional.pad(inputs, (0, 0, 0, padding_size))
343
375
  cache_position = torch.nn.functional.pad(cache_position, (0, padding_size))
344
376
  position_ids = torch.nn.functional.pad(position_ids, (0, padding_size))
@@ -389,45 +421,39 @@ class RBLNGemma3RuntimeModel(RBLNRuntimeModel):
389
421
 
390
422
  step = 0
391
423
  while step < query_length:
392
- # Check if the prefill chunk is an image prefill
393
- is_image_prefill = self.rbln_config.use_image_prefill and torch.all(
394
- token_type_ids[:, step : step + self.rbln_config.image_prefill_chunk_size] == 1
395
- )
396
- prefill_chunk_size = (
397
- self.rbln_config.image_prefill_chunk_size if is_image_prefill else self.rbln_config.prefill_chunk_size
398
- )
399
-
400
- # Check if the prefill chunk is a text prefill which have image_tokens in it.
401
- is_text_prefill_with_image_tokens = (
402
- self.rbln_config.use_image_prefill
403
- and not is_image_prefill
404
- and torch.any(token_type_ids[:, step : step + prefill_chunk_size] == 1)
405
- )
406
-
407
- # Check if the prefill chunk crosses a block boundary, requiring padding to align with block boundaries
408
- is_cross_block_boundary = (
409
- step // self.rbln_config.kvcache_block_size
410
- != (step + prefill_chunk_size) // self.rbln_config.kvcache_block_size
411
- )
424
+ if self.rbln_config.use_image_prefill:
425
+ # Check if the prefill chunk is an image prefill
426
+ is_image_prefill = torch.all(
427
+ token_type_ids[:, step : step + self.rbln_config.image_prefill_chunk_size] == 1
428
+ )
429
+ # Check if the prefill chunk is a text prefill which have image_tokens in it.
430
+ is_text_prefill_with_image_tokens = not is_image_prefill and torch.any(
431
+ token_type_ids[:, step : step + self.rbln_config.prefill_chunk_size] == 1
432
+ )
433
+ else:
434
+ is_image_prefill, is_text_prefill_with_image_tokens = False, False
412
435
 
413
436
  # Check if the prefill chunk is the last chunk
414
- is_last_chunk = step + prefill_chunk_size >= query_length
437
+ is_last_chunk = step + self.rbln_config.prefill_chunk_size >= query_length
415
438
 
416
- if is_cross_block_boundary:
417
- padding_size = prefill_chunk_size - (step + prefill_chunk_size) % self.rbln_config.kvcache_block_size
418
- padded_cache_lengths += padding_size
439
+ input_chunk = inputs[:, step : step + self.rbln_config.prefill_chunk_size]
440
+ cache_pos_chunk = (
441
+ cache_position[:, step : step + self.rbln_config.prefill_chunk_size] + padded_cache_lengths
442
+ )
443
+ position_ids_chunk = position_ids[:, step : step + self.rbln_config.prefill_chunk_size]
419
444
 
420
445
  # if text_prefill end with image_tokens, we only treat the text part.
421
- num_processed_tokens = prefill_chunk_size
446
+ num_processed_tokens = self.rbln_config.prefill_chunk_size
447
+ current_padded_cache_lengths = 0
422
448
  if is_text_prefill_with_image_tokens:
423
- first_image_token_idx = torch.where(token_type_ids[:, step : step + prefill_chunk_size] == 1)[1][0]
449
+ first_image_token_idx = torch.where(
450
+ token_type_ids[:, step : step + self.rbln_config.prefill_chunk_size] == 1
451
+ )[1][0]
424
452
  num_processed_tokens = first_image_token_idx.item()
453
+ current_padded_cache_lengths = self.rbln_config.prefill_chunk_size - num_processed_tokens
425
454
  if is_last_chunk:
426
455
  num_processed_tokens = query_length - step
427
456
 
428
- input_chunk = inputs[:, step : step + prefill_chunk_size]
429
- cache_pos_chunk = cache_position[:, step : step + prefill_chunk_size].clone() + padded_cache_lengths
430
- position_ids_chunk = position_ids[:, step : step + prefill_chunk_size].clone()
431
457
  chunked_attention_mask[
432
458
  :, step + padded_cache_lengths : step + num_processed_tokens + padded_cache_lengths
433
459
  ] = 1
@@ -456,6 +482,7 @@ class RBLNGemma3RuntimeModel(RBLNRuntimeModel):
456
482
  out=out_buffers,
457
483
  )
458
484
 
485
+ padded_cache_lengths += current_padded_cache_lengths
459
486
  step += num_processed_tokens
460
487
 
461
488
  if not is_external_block_tables:
@@ -633,6 +660,11 @@ class RBLNGemma3ForCausalLM(RBLNDecoderOnlyModelForCausalLM):
633
660
  raise ValueError("use_attention_mask and use_position_ids must be True for RBLNGemma3ForCausalLM")
634
661
 
635
662
  if rbln_config.use_image_prefill:
663
+ if rbln_config.prefill_chunk_size != rbln_config.image_prefill_chunk_size:
664
+ raise NotImplementedError(
665
+ "Not implemented for different prefill chunk sizes between text and image prefill."
666
+ )
667
+
636
668
  # Update image prefill compile config
637
669
  img_prefill_input_info = cls.get_input_info(
638
670
  batch_size=1,
@@ -46,10 +46,7 @@ class LoopVisionTower:
46
46
  def __init__(self, vision_tower: RBLNModel) -> None:
47
47
  self.vision_tower = vision_tower
48
48
 
49
- def forward(self, *args, **kwargs):
50
- pixel_values = args[0]
51
- image_sizes = kwargs.pop("image_sizes", None)
52
-
49
+ def forward(self, pixel_values, image_sizes: Optional[torch.Tensor] = None, **kwargs):
53
50
  outputs = []
54
51
  for i in range(pixel_values.shape[0]):
55
52
  outputs.append(
@@ -161,6 +158,7 @@ class RBLNLlavaForConditionalGeneration(RBLNModel):
161
158
  model_config: Optional["PretrainedConfig"] = None,
162
159
  rbln_config: Optional[RBLNModelConfig] = None,
163
160
  ) -> RBLNModelConfig:
161
+ # support for pixtral that needs padding
164
162
  if hasattr(rbln_config.vision_tower, "max_image_size"):
165
163
  num_positions = (
166
164
  rbln_config.vision_tower.batch_size
@@ -171,7 +169,10 @@ class RBLNLlavaForConditionalGeneration(RBLNModel):
171
169
 
172
170
  else:
173
171
  num_positions = (model_config.vision_config.image_size // model_config.vision_config.patch_size) ** 2 + 1
174
- selected_image_feature_dim = num_positions - 1
172
+ if model_config.vision_feature_select_strategy == "default":
173
+ selected_image_feature_dim = num_positions - 1
174
+ else:
175
+ selected_image_feature_dim = num_positions
175
176
 
176
177
  input_info = [
177
178
  (
@@ -200,8 +200,13 @@ class BaseTest:
200
200
  if self.EXPECTED_OUTPUT:
201
201
  from simphile import jaccard_similarity
202
202
 
203
- similarity = jaccard_similarity(output, self.EXPECTED_OUTPUT)
204
- self.assertGreater(similarity, 0.9)
203
+ if isinstance(self.EXPECTED_OUTPUT, str):
204
+ similarity = jaccard_similarity(output, self.EXPECTED_OUTPUT)
205
+ self.assertGreater(similarity, 0.9)
206
+ else:
207
+ for o, e_o in zip(output, self.EXPECTED_OUTPUT):
208
+ similarity = jaccard_similarity(o, e_o)
209
+ self.assertGreater(similarity, 0.9)
205
210
 
206
211
  def _inner_test_save_load(self, tmpdir):
207
212
  with ContextRblnConfig(create_runtimes=False):
@@ -91,9 +91,7 @@ class LLMTest:
91
91
  class TestMistralForCausalLM(LLMTest.TestLLM):
92
92
  RBLN_CLASS = RBLNMistralForCausalLM
93
93
  HF_MODEL_ID = "openaccess-ai-collective/tiny-mistral"
94
- EXPECTED_OUTPUT = (
95
- "Edge wat ComecidBusDonald=-Battle Orts html тиційsprintfвата Orts sect matches terrible occup"
96
- )
94
+ EXPECTED_OUTPUT = "watasurescid completionennen Brad completion жеULT ba completion影 Fin сво Regimentixon cabin影 provisions bland"
97
95
  HF_CONFIG_KWARGS = {"num_hidden_layers": 1, "max_position_embeddings": 1024, "sliding_window": 512}
98
96
 
99
97
 
@@ -1 +0,0 @@
1
- rebel_compiler_version: 0.8.1.dev142+gab6ad3c7
File without changes