semantic-link-labs 0.8.10__tar.gz → 0.8.11__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of semantic-link-labs might be problematic. Click here for more details.

Files changed (202) hide show
  1. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/PKG-INFO +3 -2
  2. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/README.md +2 -1
  3. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/docs/source/conf.py +1 -1
  4. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/pyproject.toml +1 -1
  5. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/semantic_link_labs.egg-info/PKG-INFO +3 -2
  6. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/semantic_link_labs.egg-info/SOURCES.txt +1 -0
  7. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/__init__.py +6 -2
  8. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_clear_cache.py +39 -37
  9. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_connections.py +13 -13
  10. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_data_pipelines.py +20 -20
  11. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_dataflows.py +27 -28
  12. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_dax.py +41 -47
  13. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_environments.py +26 -23
  14. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_eventhouses.py +16 -15
  15. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_eventstreams.py +16 -15
  16. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_external_data_shares.py +18 -20
  17. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_gateways.py +14 -14
  18. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_generate_semantic_model.py +99 -62
  19. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_git.py +105 -43
  20. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_helper_functions.py +148 -131
  21. semantic_link_labs-0.8.11/src/sempy_labs/_job_scheduler.py +92 -0
  22. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_kql_databases.py +16 -15
  23. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_kql_querysets.py +16 -15
  24. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_list_functions.py +114 -99
  25. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_managed_private_endpoints.py +19 -17
  26. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_mirrored_databases.py +51 -48
  27. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_mirrored_warehouses.py +5 -4
  28. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_ml_experiments.py +16 -15
  29. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_ml_models.py +15 -14
  30. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_model_bpa.py +3 -3
  31. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_model_dependencies.py +55 -29
  32. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_notebooks.py +27 -25
  33. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_one_lake_integration.py +23 -26
  34. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_query_scale_out.py +67 -64
  35. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_refresh_semantic_model.py +25 -26
  36. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_spark.py +33 -32
  37. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_sql.py +12 -9
  38. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_translations.py +10 -7
  39. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_vertipaq.py +34 -31
  40. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_warehouses.py +22 -21
  41. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_workspace_identity.py +11 -10
  42. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_workspaces.py +40 -33
  43. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/admin/_basic_functions.py +10 -12
  44. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/admin/_external_data_share.py +3 -3
  45. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/admin/_items.py +4 -4
  46. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/admin/_scanner.py +3 -1
  47. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/directlake/_directlake_schema_compare.py +18 -14
  48. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/directlake/_directlake_schema_sync.py +18 -12
  49. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/directlake/_dl_helper.py +25 -26
  50. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/directlake/_generate_shared_expression.py +10 -9
  51. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/directlake/_get_directlake_lakehouse.py +16 -13
  52. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/directlake/_get_shared_expression.py +4 -3
  53. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/directlake/_guardrails.py +12 -6
  54. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/directlake/_list_directlake_model_calc_tables.py +15 -9
  55. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/directlake/_show_unsupported_directlake_objects.py +16 -10
  56. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/directlake/_update_directlake_model_lakehouse_connection.py +35 -31
  57. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/directlake/_update_directlake_partition_entity.py +34 -31
  58. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/directlake/_warm_cache.py +87 -65
  59. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/lakehouse/_get_lakehouse_columns.py +10 -8
  60. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/lakehouse/_get_lakehouse_tables.py +10 -9
  61. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/lakehouse/_lakehouse.py +17 -13
  62. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/lakehouse/_shortcuts.py +42 -23
  63. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/migration/_create_pqt_file.py +16 -11
  64. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/migration/_refresh_calc_tables.py +16 -10
  65. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_download_report.py +9 -8
  66. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_generate_report.py +40 -44
  67. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_paginated.py +9 -9
  68. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_report_bpa.py +13 -9
  69. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_report_functions.py +80 -91
  70. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_report_helper.py +8 -4
  71. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_report_list_functions.py +24 -13
  72. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_report_rebind.py +17 -16
  73. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_reportwrapper.py +41 -33
  74. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/tom/_model.py +43 -6
  75. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/tests/test_shortcuts.py +6 -5
  76. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/.github/ISSUE_TEMPLATE/bug_report.md +0 -0
  77. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/.github/ISSUE_TEMPLATE/feature_request.md +0 -0
  78. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/.github/ISSUE_TEMPLATE/issue--question---advice-needed.md +0 -0
  79. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/.github/workflows/build.yaml +0 -0
  80. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/.github/workflows/codeql.yaml +0 -0
  81. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/.gitignore +0 -0
  82. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/.readthedocs.yaml +0 -0
  83. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/.vscode/settings.json +0 -0
  84. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/CODE_OF_CONDUCT.md +0 -0
  85. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/LICENSE +0 -0
  86. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/SECURITY.md +0 -0
  87. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/SUPPORT.md +0 -0
  88. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/docs/Makefile +0 -0
  89. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/docs/make.bat +0 -0
  90. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/docs/requirements.txt +0 -0
  91. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/docs/source/index.rst +0 -0
  92. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/docs/source/modules.rst +0 -0
  93. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/environment.yml +0 -0
  94. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/notebooks/Best Practice Analyzer Report.ipynb +0 -0
  95. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/notebooks/Capacity Migration.ipynb +0 -0
  96. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/notebooks/Migration to Direct Lake.ipynb +0 -0
  97. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/notebooks/Model Optimization.ipynb +0 -0
  98. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/notebooks/Query Scale Out.ipynb +0 -0
  99. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/notebooks/Report Analysis.ipynb +0 -0
  100. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/notebooks/SQL.ipynb +0 -0
  101. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/notebooks/Semantic Model Refresh.ipynb +0 -0
  102. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/notebooks/Tabular Object Model.ipynb +0 -0
  103. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/setup.cfg +0 -0
  104. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/semantic_link_labs.egg-info/dependency_links.txt +0 -0
  105. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/semantic_link_labs.egg-info/requires.txt +0 -0
  106. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/semantic_link_labs.egg-info/top_level.txt +0 -0
  107. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_ai.py +0 -0
  108. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_authentication.py +0 -0
  109. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_am-ET.po +0 -0
  110. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_ar-AE.po +0 -0
  111. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_bg-BG.po +0 -0
  112. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_ca-ES.po +0 -0
  113. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_cs-CZ.po +0 -0
  114. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_da-DK.po +0 -0
  115. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_de-DE.po +0 -0
  116. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_el-GR.po +0 -0
  117. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_es-ES.po +0 -0
  118. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_fa-IR.po +0 -0
  119. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_fi-FI.po +0 -0
  120. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_fr-FR.po +0 -0
  121. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_ga-IE.po +0 -0
  122. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_he-IL.po +0 -0
  123. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_hi-IN.po +0 -0
  124. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_hu-HU.po +0 -0
  125. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_id-ID.po +0 -0
  126. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_is-IS.po +0 -0
  127. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_it-IT.po +0 -0
  128. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_ja-JP.po +0 -0
  129. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_ko-KR.po +0 -0
  130. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_mt-MT.po +0 -0
  131. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_nl-NL.po +0 -0
  132. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_pl-PL.po +0 -0
  133. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_pt-BR.po +0 -0
  134. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_pt-PT.po +0 -0
  135. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_ro-RO.po +0 -0
  136. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_ru-RU.po +0 -0
  137. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_sk-SK.po +0 -0
  138. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_sl-SL.po +0 -0
  139. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_sv-SE.po +0 -0
  140. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_ta-IN.po +0 -0
  141. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_te-IN.po +0 -0
  142. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_th-TH.po +0 -0
  143. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_tr-TR.po +0 -0
  144. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_uk-UA.po +0 -0
  145. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_zh-CN.po +0 -0
  146. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_bpa_translation/_model/_translations_zu-ZA.po +0 -0
  147. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_capacities.py +0 -0
  148. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_capacity_migration.py +0 -0
  149. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_deployment_pipelines.py +0 -0
  150. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_documentation.py +0 -0
  151. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_icons.py +0 -0
  152. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_model_auto_build.py +0 -0
  153. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_model_bpa_bulk.py +0 -0
  154. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_model_bpa_rules.py +0 -0
  155. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/_workloads.py +0 -0
  156. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/admin/__init__.py +0 -0
  157. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/admin/_domains.py +0 -0
  158. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/admin/_git.py +0 -0
  159. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/directlake/__init__.py +0 -0
  160. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/lakehouse/__init__.py +0 -0
  161. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/migration/__init__.py +0 -0
  162. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/migration/_migrate_calctables_to_lakehouse.py +0 -0
  163. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/migration/_migrate_calctables_to_semantic_model.py +0 -0
  164. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/migration/_migrate_model_objects_to_semantic_model.py +0 -0
  165. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/migration/_migrate_tables_columns_to_semantic_model.py +0 -0
  166. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/migration/_migration_validation.py +0 -0
  167. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_BPAReportTemplate.json +0 -0
  168. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/__init__.py +0 -0
  169. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_bpareporttemplate/.pbi/localSettings.json +0 -0
  170. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_bpareporttemplate/.platform +0 -0
  171. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_bpareporttemplate/StaticResources/SharedResources/BaseThemes/CY24SU06.json +0 -0
  172. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_bpareporttemplate/definition/pages/01d72098bda5055bd500/page.json +0 -0
  173. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_bpareporttemplate/definition/pages/01d72098bda5055bd500/visuals/1b08bce3bebabb0a27a8/visual.json +0 -0
  174. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_bpareporttemplate/definition/pages/01d72098bda5055bd500/visuals/2f22ddb70c301693c165/visual.json +0 -0
  175. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_bpareporttemplate/definition/pages/01d72098bda5055bd500/visuals/3b1182230aa6c600b43a/visual.json +0 -0
  176. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_bpareporttemplate/definition/pages/01d72098bda5055bd500/visuals/58577ba6380c69891500/visual.json +0 -0
  177. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_bpareporttemplate/definition/pages/01d72098bda5055bd500/visuals/a2a8fa5028b3b776c96c/visual.json +0 -0
  178. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_bpareporttemplate/definition/pages/01d72098bda5055bd500/visuals/adfd47ef30652707b987/visual.json +0 -0
  179. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_bpareporttemplate/definition/pages/01d72098bda5055bd500/visuals/b6a80ee459e716e170b1/visual.json +0 -0
  180. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_bpareporttemplate/definition/pages/01d72098bda5055bd500/visuals/ce3130a721c020cc3d81/visual.json +0 -0
  181. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_bpareporttemplate/definition/pages/92735ae19b31712208ad/page.json +0 -0
  182. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_bpareporttemplate/definition/pages/92735ae19b31712208ad/visuals/66e60dfb526437cd78d1/visual.json +0 -0
  183. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_bpareporttemplate/definition/pages/c597da16dc7e63222a82/page.json +0 -0
  184. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_bpareporttemplate/definition/pages/c597da16dc7e63222a82/visuals/07deb8bce824e1be37d7/visual.json +0 -0
  185. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_bpareporttemplate/definition/pages/c597da16dc7e63222a82/visuals/0b1c68838818b32ad03b/visual.json +0 -0
  186. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_bpareporttemplate/definition/pages/c597da16dc7e63222a82/visuals/0c171de9d2683d10b930/visual.json +0 -0
  187. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_bpareporttemplate/definition/pages/c597da16dc7e63222a82/visuals/0efa01be0510e40a645e/visual.json +0 -0
  188. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_bpareporttemplate/definition/pages/c597da16dc7e63222a82/visuals/6bf2f0eb830ab53cc668/visual.json +0 -0
  189. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_bpareporttemplate/definition/pages/c597da16dc7e63222a82/visuals/88d8141cb8500b60030c/visual.json +0 -0
  190. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_bpareporttemplate/definition/pages/c597da16dc7e63222a82/visuals/a753273590beed656a03/visual.json +0 -0
  191. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_bpareporttemplate/definition/pages/c597da16dc7e63222a82/visuals/b8fdc82cddd61ac447bc/visual.json +0 -0
  192. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_bpareporttemplate/definition/pages/d37dce724a0ccc30044b/page.json +0 -0
  193. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_bpareporttemplate/definition/pages/d37dce724a0ccc30044b/visuals/ce8532a7e25020271077/visual.json +0 -0
  194. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_bpareporttemplate/definition/pages/pages.json +0 -0
  195. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_bpareporttemplate/definition/report.json +0 -0
  196. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_bpareporttemplate/definition/version.json +0 -0
  197. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_bpareporttemplate/definition.pbir +0 -0
  198. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/report/_report_bpa_rules.py +0 -0
  199. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/src/sempy_labs/tom/__init__.py +0 -0
  200. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/tests/__init__.py +0 -0
  201. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/tests/test_friendly_case.py +0 -0
  202. {semantic_link_labs-0.8.10 → semantic_link_labs-0.8.11}/tests/test_tom.py +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: semantic-link-labs
3
- Version: 0.8.10
3
+ Version: 0.8.11
4
4
  Summary: Semantic Link Labs for Microsoft Fabric
5
5
  Author: Microsoft Corporation
6
6
  License: MIT License
@@ -27,7 +27,7 @@ Requires-Dist: pytest>=8.2.1; extra == "test"
27
27
  # Semantic Link Labs
28
28
 
29
29
  [![PyPI version](https://badge.fury.io/py/semantic-link-labs.svg)](https://badge.fury.io/py/semantic-link-labs)
30
- [![Read The Docs](https://readthedocs.org/projects/semantic-link-labs/badge/?version=0.8.10&style=flat)](https://readthedocs.org/projects/semantic-link-labs/)
30
+ [![Read The Docs](https://readthedocs.org/projects/semantic-link-labs/badge/?version=0.8.11&style=flat)](https://readthedocs.org/projects/semantic-link-labs/)
31
31
  [![Code style: black](https://img.shields.io/badge/code%20style-black-000000.svg)](https://github.com/psf/black)
32
32
  [![Downloads](https://static.pepy.tech/badge/semantic-link-labs)](https://pepy.tech/project/semantic-link-labs)
33
33
 
@@ -142,6 +142,7 @@ An even better way to ensure the semantic-link-labs library is available in your
142
142
  2. Select your newly created environment within the 'Environment' drop down in the navigation bar at the top of the notebook
143
143
 
144
144
  ## Version History
145
+ * [0.8.11](https://github.com/microsoft/semantic-link-labs/releases/tag/0.8.11) (December 19, 2024)
145
146
  * [0.8.10](https://github.com/microsoft/semantic-link-labs/releases/tag/0.8.10) (December 16, 2024)
146
147
  * [0.8.9](https://github.com/microsoft/semantic-link-labs/releases/tag/0.8.9) (December 4, 2024)
147
148
  * [0.8.8](https://github.com/microsoft/semantic-link-labs/releases/tag/0.8.8) (November 28, 2024)
@@ -1,7 +1,7 @@
1
1
  # Semantic Link Labs
2
2
 
3
3
  [![PyPI version](https://badge.fury.io/py/semantic-link-labs.svg)](https://badge.fury.io/py/semantic-link-labs)
4
- [![Read The Docs](https://readthedocs.org/projects/semantic-link-labs/badge/?version=0.8.10&style=flat)](https://readthedocs.org/projects/semantic-link-labs/)
4
+ [![Read The Docs](https://readthedocs.org/projects/semantic-link-labs/badge/?version=0.8.11&style=flat)](https://readthedocs.org/projects/semantic-link-labs/)
5
5
  [![Code style: black](https://img.shields.io/badge/code%20style-black-000000.svg)](https://github.com/psf/black)
6
6
  [![Downloads](https://static.pepy.tech/badge/semantic-link-labs)](https://pepy.tech/project/semantic-link-labs)
7
7
 
@@ -116,6 +116,7 @@ An even better way to ensure the semantic-link-labs library is available in your
116
116
  2. Select your newly created environment within the 'Environment' drop down in the navigation bar at the top of the notebook
117
117
 
118
118
  ## Version History
119
+ * [0.8.11](https://github.com/microsoft/semantic-link-labs/releases/tag/0.8.11) (December 19, 2024)
119
120
  * [0.8.10](https://github.com/microsoft/semantic-link-labs/releases/tag/0.8.10) (December 16, 2024)
120
121
  * [0.8.9](https://github.com/microsoft/semantic-link-labs/releases/tag/0.8.9) (December 4, 2024)
121
122
  * [0.8.8](https://github.com/microsoft/semantic-link-labs/releases/tag/0.8.8) (November 28, 2024)
@@ -13,7 +13,7 @@ sys.path.insert(0, os.path.abspath('../../src/'))
13
13
  project = 'semantic-link-labs'
14
14
  copyright = '2024, Microsoft and community'
15
15
  author = 'Microsoft and community'
16
- release = '0.8.10'
16
+ release = '0.8.11'
17
17
 
18
18
  # -- General configuration ---------------------------------------------------
19
19
  # https://www.sphinx-doc.org/en/master/usage/configuration.html#general-configuration
@@ -7,7 +7,7 @@ name="semantic-link-labs"
7
7
  authors = [
8
8
  { name = "Microsoft Corporation" },
9
9
  ]
10
- version="0.8.10"
10
+ version="0.8.11"
11
11
  description="Semantic Link Labs for Microsoft Fabric"
12
12
  readme="README.md"
13
13
  requires-python=">=3.10,<3.12"
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: semantic-link-labs
3
- Version: 0.8.10
3
+ Version: 0.8.11
4
4
  Summary: Semantic Link Labs for Microsoft Fabric
5
5
  Author: Microsoft Corporation
6
6
  License: MIT License
@@ -27,7 +27,7 @@ Requires-Dist: pytest>=8.2.1; extra == "test"
27
27
  # Semantic Link Labs
28
28
 
29
29
  [![PyPI version](https://badge.fury.io/py/semantic-link-labs.svg)](https://badge.fury.io/py/semantic-link-labs)
30
- [![Read The Docs](https://readthedocs.org/projects/semantic-link-labs/badge/?version=0.8.10&style=flat)](https://readthedocs.org/projects/semantic-link-labs/)
30
+ [![Read The Docs](https://readthedocs.org/projects/semantic-link-labs/badge/?version=0.8.11&style=flat)](https://readthedocs.org/projects/semantic-link-labs/)
31
31
  [![Code style: black](https://img.shields.io/badge/code%20style-black-000000.svg)](https://github.com/psf/black)
32
32
  [![Downloads](https://static.pepy.tech/badge/semantic-link-labs)](https://pepy.tech/project/semantic-link-labs)
33
33
 
@@ -142,6 +142,7 @@ An even better way to ensure the semantic-link-labs library is available in your
142
142
  2. Select your newly created environment within the 'Environment' drop down in the navigation bar at the top of the notebook
143
143
 
144
144
  ## Version History
145
+ * [0.8.11](https://github.com/microsoft/semantic-link-labs/releases/tag/0.8.11) (December 19, 2024)
145
146
  * [0.8.10](https://github.com/microsoft/semantic-link-labs/releases/tag/0.8.10) (December 16, 2024)
146
147
  * [0.8.9](https://github.com/microsoft/semantic-link-labs/releases/tag/0.8.9) (December 4, 2024)
147
148
  * [0.8.8](https://github.com/microsoft/semantic-link-labs/releases/tag/0.8.8) (November 28, 2024)
@@ -54,6 +54,7 @@ src/sempy_labs/_generate_semantic_model.py
54
54
  src/sempy_labs/_git.py
55
55
  src/sempy_labs/_helper_functions.py
56
56
  src/sempy_labs/_icons.py
57
+ src/sempy_labs/_job_scheduler.py
57
58
  src/sempy_labs/_kql_databases.py
58
59
  src/sempy_labs/_kql_querysets.py
59
60
  src/sempy_labs/_list_functions.py
@@ -1,3 +1,4 @@
1
+ from sempy_labs._job_scheduler import list_item_job_instances
1
2
  from sempy_labs._gateways import (
2
3
  list_gateway_members,
3
4
  list_gateway_role_assigments,
@@ -160,7 +161,8 @@ from sempy_labs._git import (
160
161
  commit_to_git,
161
162
  initialize_git_connection,
162
163
  update_from_git,
163
- connect_workspace_to_git,
164
+ connect_workspace_to_azure_dev_ops,
165
+ connect_workspace_to_github,
164
166
  disconnect_workspace_from_git,
165
167
  )
166
168
  from sempy_labs._dataflows import (
@@ -368,7 +370,8 @@ __all__ = [
368
370
  "commit_to_git",
369
371
  "initialize_git_connection",
370
372
  "update_from_git",
371
- "connect_workspace_to_git",
373
+ "connect_workspace_to_azure_dev_ops",
374
+ "connect_workspace_to_github",
372
375
  "disconnect_workspace_from_git",
373
376
  "create_environment",
374
377
  "delete_environment",
@@ -466,4 +469,5 @@ __all__ = [
466
469
  "list_server_properties",
467
470
  "bind_semantic_model_to_gateway",
468
471
  "list_semantic_model_errors",
472
+ "list_item_job_instances",
469
473
  ]
@@ -1,39 +1,41 @@
1
1
  import sempy.fabric as fabric
2
2
  from sempy_labs._helper_functions import (
3
- resolve_dataset_id,
4
3
  is_default_semantic_model,
5
4
  _get_adls_client,
5
+ resolve_workspace_name_and_id,
6
+ resolve_dataset_name_and_id,
6
7
  )
7
8
  from typing import Optional
8
9
  import sempy_labs._icons as icons
9
10
  from sempy._utils._log import log
10
11
  import pandas as pd
11
12
  from sempy.fabric.exceptions import FabricHTTPException
13
+ from uuid import UUID
12
14
 
13
15
 
14
- def clear_cache(dataset: str, workspace: Optional[str] = None):
16
+ def clear_cache(dataset: str | UUID, workspace: Optional[str | UUID] = None):
15
17
  """
16
18
  Clears the cache of a semantic model.
17
19
  See `here <https://learn.microsoft.com/analysis-services/instances/clear-the-analysis-services-caches?view=asallproducts-allversions>`_ for documentation.
18
20
 
19
21
  Parameters
20
22
  ----------
21
- dataset : str
22
- Name of the semantic model.
23
- workspace : str, default=None
24
- The Fabric workspace name.
23
+ dataset : str | uuid.UUID
24
+ Name or ID of the semantic model.
25
+ workspace : str | uuid.UUID, default=None
26
+ The Fabric workspace name or ID.
25
27
  Defaults to None which resolves to the workspace of the attached lakehouse
26
28
  or if no lakehouse attached, resolves to the workspace of the notebook.
27
29
  """
28
30
 
29
- workspace = fabric.resolve_workspace_name(workspace)
31
+ (workspace_name, workspace_id) = resolve_workspace_name_and_id(workspace)
30
32
  if is_default_semantic_model(dataset=dataset, workspace=workspace):
31
33
  raise ValueError(
32
34
  f"{icons.red_dot} Cannot run XMLA operations against a default semantic model. Please choose a different semantic model. "
33
35
  "See here for more information: https://learn.microsoft.com/fabric/data-warehouse/semantic-models"
34
36
  )
35
37
 
36
- dataset_id = resolve_dataset_id(dataset=dataset, workspace=workspace)
38
+ (dataset_name, dataset_id) = resolve_dataset_name_and_id(dataset, workspace_id)
37
39
 
38
40
  xmla = f"""
39
41
  <ClearCache xmlns="http://schemas.microsoft.com/analysisservices/2003/engine">
@@ -42,27 +44,27 @@ def clear_cache(dataset: str, workspace: Optional[str] = None):
42
44
  </Object>
43
45
  </ClearCache>
44
46
  """
45
- fabric.execute_xmla(dataset=dataset, xmla_command=xmla, workspace=workspace)
47
+ fabric.execute_xmla(dataset=dataset_id, xmla_command=xmla, workspace=workspace_id)
46
48
  print(
47
- f"{icons.green_dot} Cache cleared for the '{dataset}' semantic model within the '{workspace}' workspace."
49
+ f"{icons.green_dot} Cache cleared for the '{dataset_name}' semantic model within the '{workspace_name}' workspace."
48
50
  )
49
51
 
50
52
 
51
53
  @log
52
54
  def backup_semantic_model(
53
- dataset: str,
55
+ dataset: str | UUID,
54
56
  file_path: str,
55
57
  allow_overwrite: bool = True,
56
58
  apply_compression: bool = True,
57
- workspace: Optional[str] = None,
59
+ workspace: Optional[str | UUID] = None,
58
60
  ):
59
61
  """
60
62
  `Backs up <https://learn.microsoft.com/azure/analysis-services/analysis-services-backup>`_ a semantic model to the ADLS Gen2 storage account connected to the workspace.
61
63
 
62
64
  Parameters
63
65
  ----------
64
- dataset : str
65
- Name of the semantic model.
66
+ dataset : str | uuid.UUID
67
+ Name or ID of the semantic model.
66
68
  file_path : str
67
69
  The ADLS Gen2 storage account location in which to backup the semantic model. Always saves within the 'power-bi-backup/<workspace name>' folder.
68
70
  Must end in '.abf'.
@@ -72,8 +74,8 @@ def backup_semantic_model(
72
74
  If True, overwrites backup files of the same name. If False, the file you are saving cannot have the same name as a file that already exists in the same location.
73
75
  apply_compression : bool, default=True
74
76
  If True, compresses the backup file. Compressed backup files save disk space, but require slightly higher CPU utilization.
75
- workspace : str, default=None
76
- The Fabric workspace name.
77
+ workspace : str | uuid.UUID, default=None
78
+ The Fabric workspace name or ID.
77
79
  Defaults to None which resolves to the workspace of the attached lakehouse
78
80
  or if no lakehouse attached, resolves to the workspace of the notebook.
79
81
  """
@@ -83,31 +85,32 @@ def backup_semantic_model(
83
85
  f"{icons.red_dot} The backup file for restoring must be in the .abf format."
84
86
  )
85
87
 
86
- workspace = fabric.resolve_workspace_name(workspace)
88
+ (workspace_name, workspace_id) = resolve_workspace_name_and_id(workspace)
89
+ (dataset_name, dataset_id) = resolve_dataset_name_and_id(dataset, workspace_id)
87
90
 
88
91
  tmsl = {
89
92
  "backup": {
90
- "database": dataset,
93
+ "database": dataset_name,
91
94
  "file": file_path,
92
95
  "allowOverwrite": allow_overwrite,
93
96
  "applyCompression": apply_compression,
94
97
  }
95
98
  }
96
99
 
97
- fabric.execute_tmsl(script=tmsl, workspace=workspace)
100
+ fabric.execute_tmsl(script=tmsl, workspace=workspace_id)
98
101
  print(
99
- f"{icons.green_dot} The '{dataset}' semantic model within the '{workspace}' workspace has been backed up to the '{file_path}' location."
102
+ f"{icons.green_dot} The '{dataset_name}' semantic model within the '{workspace_name}' workspace has been backed up to the '{file_path}' location."
100
103
  )
101
104
 
102
105
 
103
106
  @log
104
107
  def restore_semantic_model(
105
- dataset: str,
108
+ dataset: str | UUID,
106
109
  file_path: str,
107
110
  allow_overwrite: bool = True,
108
111
  ignore_incompatibilities: bool = True,
109
112
  force_restore: bool = False,
110
- workspace: Optional[str] = None,
113
+ workspace: Optional[str | UUID] = None,
111
114
  ):
112
115
  """
113
116
  `Restores <https://learn.microsoft.com/power-bi/enterprise/service-premium-backup-restore-dataset>`_ a semantic model based on a backup (.abf) file
@@ -115,8 +118,8 @@ def restore_semantic_model(
115
118
 
116
119
  Parameters
117
120
  ----------
118
- dataset : str
119
- Name of the semantic model.
121
+ dataset : str | uuid.UUID
122
+ Name or ID of the semantic model.
120
123
  file_path : str
121
124
  The location in which to backup the semantic model. Must end in '.abf'.
122
125
  Example 1: file_path = 'MyModel.abf'
@@ -127,23 +130,23 @@ def restore_semantic_model(
127
130
  If True, ignores incompatibilities between Azure Analysis Services and Power BI Premium.
128
131
  force_restore: bool, default=False
129
132
  If True, restores the semantic model with the existing semantic model unloaded and offline.
130
- workspace : str, default=None
131
- The Fabric workspace name.
133
+ workspace : str | uuid.UUID, default=None
134
+ The Fabric workspace name or ID.
132
135
  Defaults to None which resolves to the workspace of the attached lakehouse
133
136
  or if no lakehouse attached, resolves to the workspace of the notebook.
134
137
  """
135
- # https://learn.microsoft.com/en-us/power-bi/enterprise/service-premium-backup-restore-dataset
136
138
 
137
139
  if not file_path.endswith(".abf"):
138
140
  raise ValueError(
139
141
  f"{icons.red_dot} The backup file for restoring must be in the .abf format."
140
142
  )
141
143
 
142
- workspace = fabric.resolve_workspace_name(workspace)
144
+ (workspace_name, workspace_id) = resolve_workspace_name_and_id(workspace)
145
+ (dataset_name, dataset_id) = resolve_dataset_name_and_id(dataset, workspace_id)
143
146
 
144
147
  tmsl = {
145
148
  "restore": {
146
- "database": dataset,
149
+ "database": dataset_name,
147
150
  "file": file_path,
148
151
  "allowOverwrite": allow_overwrite,
149
152
  "security": "copyAll",
@@ -154,10 +157,10 @@ def restore_semantic_model(
154
157
  if force_restore:
155
158
  tmsl["restore"]["forceRestore"] = force_restore
156
159
 
157
- fabric.execute_tmsl(script=tmsl, workspace=workspace)
160
+ fabric.execute_tmsl(script=tmsl, workspace=workspace_id)
158
161
 
159
162
  print(
160
- f"{icons.green_dot} The '{dataset}' semantic model has been restored to the '{workspace}' workspace based on teh '{file_path}' backup file."
163
+ f"{icons.green_dot} The '{dataset_name}' semantic model has been restored to the '{workspace_name}' workspace based on the '{file_path}' backup file."
161
164
  )
162
165
 
163
166
 
@@ -243,15 +246,15 @@ def copy_semantic_model_backup_file(
243
246
 
244
247
 
245
248
  @log
246
- def list_backups(workspace: Optional[str] = None) -> pd.DataFrame:
249
+ def list_backups(workspace: Optional[str | UUID] = None) -> pd.DataFrame:
247
250
  """
248
251
  Shows a list of backup files contained within a workspace's ADLS Gen2 storage account.
249
252
  Requirement: An ADLS Gen2 storage account must be `connected to the workspace <https://learn.microsoft.com/power-bi/transform-model/dataflows/dataflows-azure-data-lake-storage-integration#connect-to-an-azure-data-lake-gen-2-at-a-workspace-level>`_.
250
253
 
251
254
  Parameters
252
255
  ----------
253
- workspace : str, default=None
254
- The Fabric workspace name.
256
+ workspace : str | uuid.UUID, default=None
257
+ The Fabric workspace name or ID.
255
258
  Defaults to None which resolves to the workspace of the attached lakehouse
256
259
  or if no lakehouse attached, resolves to the workspace of the notebook.
257
260
 
@@ -262,8 +265,7 @@ def list_backups(workspace: Optional[str] = None) -> pd.DataFrame:
262
265
  """
263
266
 
264
267
  client = fabric.PowerBIRestClient()
265
- workspace = fabric.resolve_workspace_name(workspace)
266
- workspace_id = fabric.resolve_workspace_id(workspace)
268
+ (workspace_name, workspace_id) = resolve_workspace_name_and_id(workspace)
267
269
  response = client.get(
268
270
  f"/v1.0/myorg/resources?resourceType=StorageAccount&folderObjectId={workspace_id}"
269
271
  )
@@ -274,7 +276,7 @@ def list_backups(workspace: Optional[str] = None) -> pd.DataFrame:
274
276
  v = response.json().get("value", [])
275
277
  if not v:
276
278
  raise ValueError(
277
- f"{icons.red_dot} A storage account is not associated with the '{workspace}' workspace."
279
+ f"{icons.red_dot} A storage account is not associated with the '{workspace_name}' workspace."
278
280
  )
279
281
  storage_account = v[0]["resourceName"]
280
282
 
@@ -5,6 +5,7 @@ from typing import Optional
5
5
  from sempy_labs._helper_functions import (
6
6
  pagination,
7
7
  _is_valid_uuid,
8
+ resolve_workspace_name_and_id,
8
9
  )
9
10
  from uuid import UUID
10
11
  import sempy_labs._icons as icons
@@ -19,7 +20,7 @@ def delete_connection(connection: str | UUID):
19
20
 
20
21
  Parameters
21
22
  ----------
22
- connection : str | UUID
23
+ connection : str | uuid.UUID
23
24
  The connection name or ID.
24
25
  """
25
26
 
@@ -42,9 +43,9 @@ def delete_connection_role_assignment(connection: str | UUID, role_assignment_id
42
43
 
43
44
  Parameters
44
45
  ----------
45
- connection : str | UUID
46
+ connection : str | uuid.UUID
46
47
  The connection name or ID.
47
- role_assignment_id : UUID
48
+ role_assignment_id : uuid.UUID
48
49
  The role assignment ID.
49
50
  """
50
51
 
@@ -87,7 +88,7 @@ def list_connection_role_assignments(connection: str | UUID) -> pd.DataFrame:
87
88
 
88
89
  Parameters
89
90
  ----------
90
- connection : str | UUID
91
+ connection : str | uuid.UUID
91
92
  The connection name or ID.
92
93
 
93
94
  Returns
@@ -205,7 +206,7 @@ def list_connections() -> pd.DataFrame:
205
206
 
206
207
 
207
208
  def list_item_connections(
208
- item_name: str, item_type: str, workspace: Optional[str] = None
209
+ item_name: str, item_type: str, workspace: Optional[str | UUID] = None
209
210
  ) -> pd.DataFrame:
210
211
  """
211
212
  Shows the list of connections that the specified item is connected to.
@@ -218,8 +219,8 @@ def list_item_connections(
218
219
  The item name.
219
220
  item_type : str
220
221
  The `item type <https://learn.microsoft.com/rest/api/fabric/core/items/update-item?tabs=HTTP#itemtype>`_.
221
- workspace : str, default=None
222
- The Fabric workspace name.
222
+ workspace : str | uuid.UUID, default=None
223
+ The Fabric workspace name or ID.
223
224
  Defaults to None which resolves to the workspace of the attached lakehouse
224
225
  or if no lakehouse attached, resolves to the workspace of the notebook.
225
226
 
@@ -229,11 +230,10 @@ def list_item_connections(
229
230
  A pandas dataframe showing the list of connections that the specified item is connected to.
230
231
  """
231
232
 
232
- workspace = fabric.resolve_workspace_name(workspace)
233
- workspace_id = fabric.resolve_workspace_id(workspace)
233
+ (workspace_name, workspace_id) = resolve_workspace_name_and_id(workspace)
234
234
  item_type = item_type[0].upper() + item_type[1:]
235
235
  item_id = fabric.resolve_item_id(
236
- item_name=item_name, type=item_type, workspace=workspace
236
+ item_name=item_name, type=item_type, workspace=workspace_id
237
237
  )
238
238
 
239
239
  client = fabric.FabricRestClient()
@@ -416,7 +416,7 @@ def create_on_prem_connection(
416
416
  ----------
417
417
  name : str
418
418
  The name of the connection.
419
- gateway : str | UUID
419
+ gateway : str | uuid.UUID
420
420
  The name or Id of the gateway.
421
421
  server_name : str
422
422
  The name of the server.
@@ -485,7 +485,7 @@ def create_vnet_connection(
485
485
  user_name: str,
486
486
  password: str,
487
487
  privacy_level: str,
488
- connection_encryption: Optional[str] = "NotEncrypted",
488
+ connection_encryption: str = "NotEncrypted",
489
489
  skip_test_connection: bool = False,
490
490
  ):
491
491
  """
@@ -497,7 +497,7 @@ def create_vnet_connection(
497
497
  ----------
498
498
  name : str
499
499
  The name of the connection.
500
- gateway : str | UUID
500
+ gateway : str | uuid.UUID
501
501
  The name or Id of the gateway.
502
502
  server_name : str
503
503
  The name of the server.
@@ -9,9 +9,10 @@ from sempy_labs._helper_functions import (
9
9
  _decode_b64,
10
10
  )
11
11
  from sempy.fabric.exceptions import FabricHTTPException
12
+ from uuid import UUID
12
13
 
13
14
 
14
- def list_data_pipelines(workspace: Optional[str] = None) -> pd.DataFrame:
15
+ def list_data_pipelines(workspace: Optional[str | UUID] = None) -> pd.DataFrame:
15
16
  """
16
17
  Shows the data pipelines within a workspace.
17
18
 
@@ -19,8 +20,8 @@ def list_data_pipelines(workspace: Optional[str] = None) -> pd.DataFrame:
19
20
 
20
21
  Parameters
21
22
  ----------
22
- workspace : str, default=None
23
- The Fabric workspace name.
23
+ workspace : str | uuid.UUID, default=None
24
+ The Fabric workspace name or ID.
24
25
  Defaults to None which resolves to the workspace of the attached lakehouse
25
26
  or if no lakehouse attached, resolves to the workspace of the notebook.
26
27
 
@@ -32,7 +33,7 @@ def list_data_pipelines(workspace: Optional[str] = None) -> pd.DataFrame:
32
33
 
33
34
  df = pd.DataFrame(columns=["Data Pipeline Name", "Data Pipeline ID", "Description"])
34
35
 
35
- (workspace, workspace_id) = resolve_workspace_name_and_id(workspace)
36
+ (workspace_name, workspace_id) = resolve_workspace_name_and_id(workspace)
36
37
 
37
38
  client = fabric.FabricRestClient()
38
39
  response = client.get(f"/v1/workspaces/{workspace_id}/dataPipelines")
@@ -54,7 +55,7 @@ def list_data_pipelines(workspace: Optional[str] = None) -> pd.DataFrame:
54
55
 
55
56
 
56
57
  def create_data_pipeline(
57
- name: str, description: Optional[str] = None, workspace: Optional[str] = None
58
+ name: str, description: Optional[str] = None, workspace: Optional[str | UUID] = None
58
59
  ):
59
60
  """
60
61
  Creates a Fabric data pipeline.
@@ -67,13 +68,13 @@ def create_data_pipeline(
67
68
  Name of the data pipeline.
68
69
  description : str, default=None
69
70
  A description of the environment.
70
- workspace : str, default=None
71
- The Fabric workspace name.
71
+ workspace : str | uuid.UUID, default=None
72
+ The Fabric workspace name or ID.
72
73
  Defaults to None which resolves to the workspace of the attached lakehouse
73
74
  or if no lakehouse attached, resolves to the workspace of the notebook.
74
75
  """
75
76
 
76
- (workspace, workspace_id) = resolve_workspace_name_and_id(workspace)
77
+ (workspace_name, workspace_id) = resolve_workspace_name_and_id(workspace)
77
78
 
78
79
  request_body = {"displayName": name}
79
80
 
@@ -88,11 +89,11 @@ def create_data_pipeline(
88
89
  lro(client, response, status_codes=[201, 202])
89
90
 
90
91
  print(
91
- f"{icons.green_dot} The '{name}' data pipeline has been created within the '{workspace}' workspace."
92
+ f"{icons.green_dot} The '{name}' data pipeline has been created within the '{workspace_name}' workspace."
92
93
  )
93
94
 
94
95
 
95
- def delete_data_pipeline(name: str, workspace: Optional[str] = None):
96
+ def delete_data_pipeline(name: str, workspace: Optional[str | UUID] = None):
96
97
  """
97
98
  Deletes a Fabric data pipeline.
98
99
 
@@ -102,16 +103,16 @@ def delete_data_pipeline(name: str, workspace: Optional[str] = None):
102
103
  ----------
103
104
  name: str
104
105
  Name of the data pipeline.
105
- workspace : str, default=None
106
+ workspace : str | uuid.UUID, default=None
106
107
  The Fabric workspace name.
107
108
  Defaults to None which resolves to the workspace of the attached lakehouse
108
109
  or if no lakehouse attached, resolves to the workspace of the notebook.
109
110
  """
110
111
 
111
- (workspace, workspace_id) = resolve_workspace_name_and_id(workspace)
112
+ (workspace_name, workspace_id) = resolve_workspace_name_and_id(workspace)
112
113
 
113
114
  item_id = fabric.resolve_item_id(
114
- item_name=name, type="DataPipeline", workspace=workspace
115
+ item_name=name, type="DataPipeline", workspace=workspace_id
115
116
  )
116
117
 
117
118
  client = fabric.FabricRestClient()
@@ -121,12 +122,12 @@ def delete_data_pipeline(name: str, workspace: Optional[str] = None):
121
122
  raise FabricHTTPException(response)
122
123
 
123
124
  print(
124
- f"{icons.green_dot} The '{name}' data pipeline within the '{workspace}' workspace has been deleted."
125
+ f"{icons.green_dot} The '{name}' data pipeline within the '{workspace_name}' workspace has been deleted."
125
126
  )
126
127
 
127
128
 
128
129
  def get_data_pipeline_definition(
129
- name: str, workspace: Optional[str] = None, decode: bool = True
130
+ name: str, workspace: Optional[str | UUID] = None, decode: bool = True
130
131
  ) -> dict | pd.DataFrame:
131
132
  """
132
133
  Obtains the definition of a data pipeline.
@@ -135,8 +136,8 @@ def get_data_pipeline_definition(
135
136
  ----------
136
137
  name : str
137
138
  The name of the data pipeline.
138
- workspace : str, default=None
139
- The Fabric workspace name.
139
+ workspace : str | uuid.UUID, default=None
140
+ The Fabric workspace name or ID.
140
141
  Defaults to None which resolves to the workspace of the attached lakehouse
141
142
  or if no lakehouse attached, resolves to the workspace of the notebook.
142
143
  decode : bool, default=True
@@ -150,10 +151,9 @@ def get_data_pipeline_definition(
150
151
  A pandas dataframe showing the data pipelines within a workspace.
151
152
  """
152
153
 
153
- workspace = fabric.resolve_workspace_name(workspace)
154
- workspace_id = fabric.resolve_workspace_id(workspace)
154
+ (workspace_name, workspace_id) = resolve_workspace_name_and_id(workspace)
155
155
  item_id = fabric.resolve_item_id(
156
- item_name=name, type="DataPipeline", workspace=workspace
156
+ item_name=name, type="DataPipeline", workspace=workspace_id
157
157
  )
158
158
 
159
159
  client = fabric.FabricRestClient()