databricks-labs-lakebridge 0.10.1__tar.gz → 0.10.2__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/PKG-INFO +3 -3
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/__about__.py +1 -1
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/helpers/file_utils.py +0 -11
- databricks_labs_lakebridge-0.10.2/databricks/labs/lakebridge/helpers/string_utils.py +28 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/intermediate/root_tables.py +5 -7
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/transpiler/execute.py +7 -6
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/transpiler/lsp/lsp_engine.py +1 -3
- databricks_labs_lakebridge-0.10.2/docs/lakebridge/src/theme/DocSidebarItems/index.tsx +42 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/pyproject.toml +2 -2
- databricks_labs_lakebridge-0.10.1/databricks/labs/lakebridge/helpers/string_utils.py +0 -62
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/.gitignore +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/LICENSE +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/NOTICE +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/README.md +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/__init__.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/__init__.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/__init__.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/assessments/configure_assessment.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/assessments/pipeline.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/assessments/profiler_config.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/base_install.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/cli.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/config.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/connections/__init__.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/connections/credential_manager.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/connections/database_manager.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/connections/env_getter.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/contexts/__init__.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/contexts/application.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/coverage/__init__.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/coverage/commons.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/coverage/lakebridge_snow_transpilation_coverage.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/coverage/local_report.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/coverage/sqlglot_snow_transpilation_coverage.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/coverage/sqlglot_tsql_transpilation_coverage.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/deployment/__init__.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/deployment/configurator.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/deployment/dashboard.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/deployment/installation.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/deployment/job.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/deployment/recon.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/deployment/table.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/deployment/upgrade_common.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/discovery/table.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/discovery/table_definition.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/discovery/tsql_table_definition.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/errors/exceptions.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/helpers/__init__.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/helpers/db_sql.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/helpers/execution_time.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/helpers/metastore.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/helpers/recon_config_utils.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/helpers/telemetry_utils.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/helpers/validation.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/install.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/intermediate/__init__.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/intermediate/dag.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/intermediate/engine_adapter.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/jvmproxy.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/lineage.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/reconcile/__init__.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/reconcile/compare.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/reconcile/connectors/__init__.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/reconcile/connectors/data_source.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/reconcile/connectors/databricks.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/reconcile/connectors/jdbc_reader.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/reconcile/connectors/oracle.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/reconcile/connectors/secrets.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/reconcile/connectors/snowflake.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/reconcile/connectors/source_adapter.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/reconcile/connectors/sql_server.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/reconcile/constants.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/reconcile/exception.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/reconcile/execute.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/reconcile/query_builder/__init__.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/reconcile/query_builder/aggregate_query.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/reconcile/query_builder/base.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/reconcile/query_builder/count_query.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/reconcile/query_builder/expression_generator.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/reconcile/query_builder/hash_query.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/reconcile/query_builder/sampling_query.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/reconcile/query_builder/threshold_query.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/reconcile/recon_capture.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/reconcile/recon_config.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/reconcile/recon_output_config.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/reconcile/runner.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/reconcile/sampler.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/reconcile/schema_compare.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/__init__.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/config/credentials.yml +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/__init__.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/__init__.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/aggregate_reconciliation_metrics/00_0_aggregate_recon_header.md +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/aggregate_reconciliation_metrics/01_0_recon_id.filter.yml +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/aggregate_reconciliation_metrics/01_1_executed_by.filter.yml +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/aggregate_reconciliation_metrics/01_2_started_at.filter.yml +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/aggregate_reconciliation_metrics/02_0_source_type.filter.yml +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/aggregate_reconciliation_metrics/02_1_source_table.filter.yml +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/aggregate_reconciliation_metrics/02_2_target_table.filter.yml +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/aggregate_reconciliation_metrics/04_0_aggregate_summary_table.sql +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/aggregate_reconciliation_metrics/05_0_aggregate_recon_drilldown_header.md +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/aggregate_reconciliation_metrics/06_0_recon_id.filter.yml +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/aggregate_reconciliation_metrics/06_1_category.filter.yml +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/aggregate_reconciliation_metrics/06_2_aggregate_type.filter.yml +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/aggregate_reconciliation_metrics/07_0_target_table.filter.yml +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/aggregate_reconciliation_metrics/07_1_source_table.filter.yml +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/aggregate_reconciliation_metrics/08_0_aggregate_details_table.sql +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/aggregate_reconciliation_metrics/09_0_aggregate_missing_mismatch_header.md +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/aggregate_reconciliation_metrics/10_0_aggr_mismatched_records.sql +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/aggregate_reconciliation_metrics/11_0_aggr_missing_in_databricks.sql +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/aggregate_reconciliation_metrics/11_1_aggr_missing_in_source.sql +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/aggregate_reconciliation_metrics/dashboard.yml +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/reconciliation_metrics/00_0_recon_main.md +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/reconciliation_metrics/01_0_recon_id.filter.yml +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/reconciliation_metrics/01_1_report_type.filter.yml +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/reconciliation_metrics/01_2_executed_by.filter.yml +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/reconciliation_metrics/02_0_source_type.filter.yml +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/reconciliation_metrics/02_1_source_table.filter.yml +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/reconciliation_metrics/02_2_target_table.filter.yml +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/reconciliation_metrics/03_0_started_at.filter.yml +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/reconciliation_metrics/05_0_summary_table.sql +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/reconciliation_metrics/06_0_schema_comparison_header.md +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/reconciliation_metrics/07_0_schema_details_table.sql +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/reconciliation_metrics/08_0_drill_down_header.md +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/reconciliation_metrics/09_0_recon_id.filter.yml +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/reconciliation_metrics/09_1_category.filter.yml +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/reconciliation_metrics/10_0_target_table.filter.yml +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/reconciliation_metrics/10_1_source_table.filter.yml +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/reconciliation_metrics/11_0_recon_details_pivot.sql +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/reconciliation_metrics/12_0_daily_data_validation_issue_header.md +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/reconciliation_metrics/13_0_success_fail_.filter.yml +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/reconciliation_metrics/14_0_failed_recon_ids.sql +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/reconciliation_metrics/15_0_total_failed_runs.sql +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/reconciliation_metrics/15_1_failed_targets.sql +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/reconciliation_metrics/15_2_successful_targets.sql +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/reconciliation_metrics/16_0_missing_mismatch_header.md +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/reconciliation_metrics/17_0_mismatched_records.sql +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/reconciliation_metrics/17_1_threshold_mismatches.sql +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/reconciliation_metrics/18_0_missing_in_databricks.sql +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/reconciliation_metrics/18_1_missing_in_source.sql +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/dashboards/reconciliation_metrics/dashboard.yml +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/queries/__init__.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/queries/installation/__init__.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/queries/installation/aggregate_details.sql +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/queries/installation/aggregate_metrics.sql +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/queries/installation/aggregate_rules.sql +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/queries/installation/details.sql +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/queries/installation/main.sql +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/resources/reconcile/queries/installation/metrics.sql +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/transpiler/__init__.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/transpiler/lsp/__init__.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/transpiler/sqlglot/__init__.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/transpiler/sqlglot/dialect_utils.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/transpiler/sqlglot/generator/__init__.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/transpiler/sqlglot/generator/databricks.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/transpiler/sqlglot/lca_utils.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/transpiler/sqlglot/local_expression.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/transpiler/sqlglot/parsers/__init__.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/transpiler/sqlglot/parsers/oracle.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/transpiler/sqlglot/parsers/presto.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/transpiler/sqlglot/parsers/snowflake.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/transpiler/sqlglot/sqlglot_engine.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/transpiler/transpile_engine.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/transpiler/transpile_status.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/uninstall.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/upgrades/v0.4.0_add_main_table_operation_name_column.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/lakebridge/upgrades/v0.6.0_alter_metrics_datatype.py +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/docs/lakebridge/src/components/Button.tsx +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/docs/lakebridge/src/css/custom.css +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/docs/lakebridge/src/css/table.css +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/docs/lakebridge/src/pages/index.tsx +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/docs/lakebridge/src/theme/Footer/index.tsx +0 -0
- {databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/docs/lakebridge/src/theme/Layout/index.tsx +0 -0
@@ -1,6 +1,6 @@
|
|
1
1
|
Metadata-Version: 2.4
|
2
2
|
Name: databricks-labs-lakebridge
|
3
|
-
Version: 0.10.
|
3
|
+
Version: 0.10.2
|
4
4
|
Summary: Fast and predictable migrations to Databricks Lakehouse Platform. This tool is designed to help you migrate your data and workloads to the Databricks Lakehouse Platform in a fast, predictable, and reliable way. It provides a set of tools and utilities to help you reconcile your data and workloads, assess your current state, and plan your migration.
|
5
5
|
Project-URL: Documentation, https://databrickslabs.github.io/lakebridge
|
6
6
|
Project-URL: Issues, https://github.com/databrickslabs/lakebridge/issues
|
@@ -25,8 +25,8 @@ Classifier: Topic :: Software Development :: Libraries
|
|
25
25
|
Classifier: Topic :: Utilities
|
26
26
|
Requires-Python: >=3.10
|
27
27
|
Requires-Dist: cryptography<45.1.0,>=44.0.2
|
28
|
-
Requires-Dist: databricks-bb-analyzer~=0.1.
|
29
|
-
Requires-Dist: databricks-labs-blueprint[yaml]<0.12.0,>=0.11.
|
28
|
+
Requires-Dist: databricks-bb-analyzer~=0.1.8
|
29
|
+
Requires-Dist: databricks-labs-blueprint[yaml]<0.12.0,>=0.11.1
|
30
30
|
Requires-Dist: databricks-labs-lsql==0.16.0
|
31
31
|
Requires-Dist: databricks-sdk~=0.51.0
|
32
32
|
Requires-Dist: duckdb~=1.2.2
|
@@ -1,2 +1,2 @@
|
|
1
1
|
# DO NOT MODIFY THIS FILE
|
2
|
-
__version__ = "0.10.
|
2
|
+
__version__ = "0.10.2"
|
@@ -51,14 +51,3 @@ def get_sql_file(input_path: str | Path) -> Generator[Path, None, None]:
|
|
51
51
|
for filename in files:
|
52
52
|
if is_sql_file(filename):
|
53
53
|
yield filename
|
54
|
-
|
55
|
-
|
56
|
-
def read_file(filename: str | Path) -> str:
|
57
|
-
"""
|
58
|
-
Reads the contents of the given file and returns it as a string.
|
59
|
-
:param filename: Input File Path
|
60
|
-
:return: File Contents as String
|
61
|
-
"""
|
62
|
-
# pylint: disable=unspecified-encoding
|
63
|
-
with Path(filename).open() as file:
|
64
|
-
return file.read()
|
@@ -0,0 +1,28 @@
|
|
1
|
+
def refactor_hexadecimal_chars(input_string: str) -> str:
|
2
|
+
"""
|
3
|
+
Updates the HexaDecimal characters ( \x1b[\\d+m ) in the given string as below.
|
4
|
+
:param input_string: String with HexaDecimal characters. ex: ( \x1b[4mWHERE\x1b[0m )
|
5
|
+
:return: String with HexaDecimal characters refactored to arrows. ex: ( --> WHERE <--)
|
6
|
+
"""
|
7
|
+
output_string = input_string
|
8
|
+
highlight = {"\x1b[4m": "--> ", "\x1b[0m": " <--"}
|
9
|
+
for key, value in highlight.items():
|
10
|
+
output_string = output_string.replace(key, value)
|
11
|
+
return output_string
|
12
|
+
|
13
|
+
|
14
|
+
def format_error_message(error_type: str, error_message: Exception, error_sql: str) -> str:
|
15
|
+
"""
|
16
|
+
Formats the error message with the error SQL.
|
17
|
+
:param error_type: Error Type
|
18
|
+
:param error_message: Error message
|
19
|
+
:param error_sql: Error SQL
|
20
|
+
:return: Formatted error message
|
21
|
+
"""
|
22
|
+
error_str = (
|
23
|
+
f"------------------------ {error_type} Start:------------------------\n"
|
24
|
+
f"/*\n{str(error_message)}\n*/\n\n"
|
25
|
+
f"/*\nOriginal Query:\n\n{str(error_sql)}\n*/\n"
|
26
|
+
f"------------------------- {error_type} End:-------------------------"
|
27
|
+
).strip()
|
28
|
+
return error_str
|
@@ -1,11 +1,9 @@
|
|
1
1
|
import logging
|
2
2
|
from pathlib import Path
|
3
3
|
|
4
|
-
from databricks.labs.
|
5
|
-
|
6
|
-
|
7
|
-
read_file,
|
8
|
-
)
|
4
|
+
from databricks.labs.blueprint.paths import read_text
|
5
|
+
|
6
|
+
from databricks.labs.lakebridge.helpers.file_utils import get_sql_file, is_sql_file
|
9
7
|
from databricks.labs.lakebridge.intermediate.dag import DAG
|
10
8
|
|
11
9
|
from databricks.labs.lakebridge.transpiler.sqlglot.sqlglot_engine import SqlglotEngine
|
@@ -26,14 +24,14 @@ class RootTableAnalyzer:
|
|
26
24
|
# when input is sql file then parse the file
|
27
25
|
if is_sql_file(self.input_path):
|
28
26
|
logger.debug(f"Generating Lineage file: {self.input_path}")
|
29
|
-
sql_content =
|
27
|
+
sql_content = read_text(self.input_path)
|
30
28
|
self._populate_dag(sql_content, self.input_path, dag)
|
31
29
|
return dag # return after processing the file
|
32
30
|
|
33
31
|
# when the input is a directory
|
34
32
|
for path in get_sql_file(self.input_path):
|
35
33
|
logger.debug(f"Generating Lineage file: {path}")
|
36
|
-
sql_content =
|
34
|
+
sql_content = read_text(path)
|
37
35
|
self._populate_dag(sql_content, path, dag)
|
38
36
|
|
39
37
|
return dag
|
@@ -9,6 +9,7 @@ from typing import cast
|
|
9
9
|
import itertools
|
10
10
|
|
11
11
|
from databricks.labs.blueprint.installation import JsonObject
|
12
|
+
from databricks.labs.blueprint.paths import read_text
|
12
13
|
from databricks.labs.lakebridge.__about__ import __version__
|
13
14
|
from databricks.labs.lakebridge.config import (
|
14
15
|
TranspileConfig,
|
@@ -28,7 +29,6 @@ from databricks.labs.lakebridge.transpiler.transpile_status import (
|
|
28
29
|
ErrorKind,
|
29
30
|
ErrorSeverity,
|
30
31
|
)
|
31
|
-
from databricks.labs.lakebridge.helpers.string_utils import remove_bom
|
32
32
|
from databricks.labs.lakebridge.helpers.validation import Validator
|
33
33
|
from databricks.labs.lakebridge.transpiler.sqlglot.sqlglot_engine import SqlglotEngine
|
34
34
|
from databricks.sdk import WorkspaceClient
|
@@ -62,15 +62,14 @@ async def _process_one_file(context: TranspilingContext) -> tuple[int, list[Tran
|
|
62
62
|
)
|
63
63
|
return 0, [error]
|
64
64
|
|
65
|
-
|
66
|
-
|
67
|
-
context = dataclasses.replace(context, source_code=source_code)
|
65
|
+
source_code = read_text(context.input_path)
|
66
|
+
context = dataclasses.replace(context, source_code=source_code)
|
68
67
|
|
69
68
|
transpile_result = await _transpile(
|
70
69
|
context.transpiler,
|
71
70
|
str(context.config.source_dialect),
|
72
71
|
context.config.target_dialect,
|
73
|
-
|
72
|
+
source_code,
|
74
73
|
context.input_path,
|
75
74
|
)
|
76
75
|
|
@@ -158,7 +157,9 @@ def _process_single_result(context: TranspilingContext, error_list: list[Transpi
|
|
158
157
|
|
159
158
|
output_path = cast(Path, context.output_path)
|
160
159
|
with output_path.open("w") as w:
|
161
|
-
|
160
|
+
# The above adds a java-style comment block at the top of the output file
|
161
|
+
# This would break .py or .json outputs so we disable it for now.
|
162
|
+
# w.write(make_header(context.input_path, error_list))
|
162
163
|
w.write(output_code)
|
163
164
|
|
164
165
|
logger.info(f"Processed file: {context.input_path} (errors: {len(error_list)})")
|
@@ -523,9 +523,7 @@ class LSPEngine(TranspileEngine):
|
|
523
523
|
self.close_document(file_path)
|
524
524
|
return ChangeManager.apply(source_code, response.changes, response.diagnostics, file_path)
|
525
525
|
|
526
|
-
def open_document(self, file_path: Path,
|
527
|
-
if source_code is None:
|
528
|
-
source_code = file_path.read_text(encoding)
|
526
|
+
def open_document(self, file_path: Path, source_code: str) -> None:
|
529
527
|
text_document = TextDocumentItem(
|
530
528
|
uri=file_path.as_uri(), language_id=LanguageKind.Sql, version=1, text=source_code
|
531
529
|
)
|
@@ -0,0 +1,42 @@
|
|
1
|
+
import React, {memo, type ReactNode} from 'react';
|
2
|
+
import {useLocation, useHistory} from 'react-router';
|
3
|
+
import type {PropSidebarItem} from '@docusaurus/plugin-content-docs';
|
4
|
+
import {
|
5
|
+
DocSidebarItemsExpandedStateProvider,
|
6
|
+
useVisibleSidebarItems,
|
7
|
+
} from '@docusaurus/plugin-content-docs/client';
|
8
|
+
import DocSidebarItem from '@theme/DocSidebarItem';
|
9
|
+
|
10
|
+
import type {Props} from '@theme/DocSidebarItems';
|
11
|
+
|
12
|
+
function DocSidebarItems({items, onItemClick, ...props}: Props): ReactNode {
|
13
|
+
const location = useLocation();
|
14
|
+
const history = useHistory();
|
15
|
+
const visibleItems = useVisibleSidebarItems(items, props.activePath);
|
16
|
+
|
17
|
+
/**
|
18
|
+
* Additional logic for handling custom UI scenarios
|
19
|
+
*/
|
20
|
+
const onClickHandler = (params: PropSidebarItem) => {
|
21
|
+
if (onItemClick) {
|
22
|
+
onItemClick(params);
|
23
|
+
}
|
24
|
+
|
25
|
+
// show initial page on menu collapse
|
26
|
+
if (params.type === "category") {
|
27
|
+
if (location.pathname !== params.href && location.pathname.includes(params.href)) {
|
28
|
+
history.push(params.href);
|
29
|
+
}
|
30
|
+
}
|
31
|
+
}
|
32
|
+
|
33
|
+
return (
|
34
|
+
<DocSidebarItemsExpandedStateProvider>
|
35
|
+
{visibleItems.map((item, index) => (
|
36
|
+
<DocSidebarItem key={index} item={item} index={index} {...props} onItemClick={onClickHandler} />
|
37
|
+
))}
|
38
|
+
</DocSidebarItemsExpandedStateProvider>
|
39
|
+
);
|
40
|
+
}
|
41
|
+
|
42
|
+
export default memo(DocSidebarItems);
|
@@ -29,9 +29,9 @@ classifiers = [
|
|
29
29
|
dependencies = [
|
30
30
|
"databricks-sdk~=0.51.0",
|
31
31
|
"standard-distutils~=3.11.9; python_version>='3.11'",
|
32
|
-
"databricks-bb-analyzer~=0.1.
|
32
|
+
"databricks-bb-analyzer~=0.1.8",
|
33
33
|
"sqlglot==26.1.3",
|
34
|
-
"databricks-labs-blueprint[yaml]>=0.11.
|
34
|
+
"databricks-labs-blueprint[yaml]>=0.11.1,<0.12.0",
|
35
35
|
"databricks-labs-lsql==0.16.0",
|
36
36
|
"cryptography>=44.0.2,<45.1.0",
|
37
37
|
"pyodbc~=5.2.0",
|
@@ -1,62 +0,0 @@
|
|
1
|
-
import codecs
|
2
|
-
|
3
|
-
|
4
|
-
# Optionally check to see if a string begins with a Byte Order Mark
|
5
|
-
# such a character will cause the transpiler to fail
|
6
|
-
def remove_bom(input_string: str) -> str:
|
7
|
-
"""
|
8
|
-
Removes the Byte Order Mark (BOM) from the given string if it exists.
|
9
|
-
:param input_string: String to remove BOM from
|
10
|
-
:return: String without BOM
|
11
|
-
"""
|
12
|
-
output_string = input_string
|
13
|
-
|
14
|
-
# Check and remove UTF-16 (LE and BE) BOM
|
15
|
-
if input_string.startswith(codecs.BOM_UTF16_BE.decode("utf-16-be")):
|
16
|
-
output_string = input_string[len(codecs.BOM_UTF16_BE.decode("utf-16-be")) :]
|
17
|
-
elif input_string.startswith(codecs.BOM_UTF16_LE.decode("utf-16-le")):
|
18
|
-
output_string = input_string[len(codecs.BOM_UTF16_LE.decode("utf-16-le")) :]
|
19
|
-
elif input_string.startswith(codecs.BOM_UTF16.decode("utf-16")):
|
20
|
-
output_string = input_string[len(codecs.BOM_UTF16.decode("utf-16")) :]
|
21
|
-
# Check and remove UTF-32 (LE and BE) BOM
|
22
|
-
elif input_string.startswith(codecs.BOM_UTF32_BE.decode("utf-32-be")):
|
23
|
-
output_string = input_string[len(codecs.BOM_UTF32_BE.decode("utf-32-be")) :]
|
24
|
-
elif input_string.startswith(codecs.BOM_UTF32_LE.decode("utf-32-le")):
|
25
|
-
output_string = input_string[len(codecs.BOM_UTF32_LE.decode("utf-32-le")) :]
|
26
|
-
elif input_string.startswith(codecs.BOM_UTF32.decode("utf-32")):
|
27
|
-
output_string = input_string[len(codecs.BOM_UTF32.decode("utf-32")) :]
|
28
|
-
# Check and remove UTF-8 BOM
|
29
|
-
elif input_string.startswith(codecs.BOM_UTF8.decode("utf-8")):
|
30
|
-
output_string = input_string[len(codecs.BOM_UTF8.decode("utf-8")) :]
|
31
|
-
|
32
|
-
return output_string
|
33
|
-
|
34
|
-
|
35
|
-
def refactor_hexadecimal_chars(input_string: str) -> str:
|
36
|
-
"""
|
37
|
-
Updates the HexaDecimal characters ( \x1b[\\d+m ) in the given string as below.
|
38
|
-
:param input_string: String with HexaDecimal characters. ex: ( \x1b[4mWHERE\x1b[0m )
|
39
|
-
:return: String with HexaDecimal characters refactored to arrows. ex: ( --> WHERE <--)
|
40
|
-
"""
|
41
|
-
output_string = input_string
|
42
|
-
highlight = {"\x1b[4m": "--> ", "\x1b[0m": " <--"}
|
43
|
-
for key, value in highlight.items():
|
44
|
-
output_string = output_string.replace(key, value)
|
45
|
-
return output_string
|
46
|
-
|
47
|
-
|
48
|
-
def format_error_message(error_type: str, error_message: Exception, error_sql: str) -> str:
|
49
|
-
"""
|
50
|
-
Formats the error message with the error SQL.
|
51
|
-
:param error_type: Error Type
|
52
|
-
:param error_message: Error message
|
53
|
-
:param error_sql: Error SQL
|
54
|
-
:return: Formatted error message
|
55
|
-
"""
|
56
|
-
error_str = (
|
57
|
-
f"------------------------ {error_type} Start:------------------------\n"
|
58
|
-
f"/*\n{str(error_message)}\n*/\n\n"
|
59
|
-
f"/*\nOriginal Query:\n\n{str(error_sql)}\n*/\n"
|
60
|
-
f"------------------------- {error_type} End:-------------------------"
|
61
|
-
).strip()
|
62
|
-
return error_str
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
{databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/__init__.py
RENAMED
File without changes
|
{databricks_labs_lakebridge-0.10.1 → databricks_labs_lakebridge-0.10.2}/databricks/labs/__init__.py
RENAMED
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|
File without changes
|