dagster-airbyte 0.26.9__py3-none-any.whl → 0.26.11rc0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of dagster-airbyte might be problematic. Click here for more details.
- dagster_airbyte/asset_defs.py +17 -17
- dagster_airbyte/managed/reconciliation.py +9 -9
- dagster_airbyte/resources.py +18 -16
- dagster_airbyte/version.py +1 -1
- {dagster_airbyte-0.26.9.dist-info → dagster_airbyte-0.26.11rc0.dist-info}/METADATA +3 -3
- {dagster_airbyte-0.26.9.dist-info → dagster_airbyte-0.26.11rc0.dist-info}/RECORD +10 -10
- {dagster_airbyte-0.26.9.dist-info → dagster_airbyte-0.26.11rc0.dist-info}/LICENSE +0 -0
- {dagster_airbyte-0.26.9.dist-info → dagster_airbyte-0.26.11rc0.dist-info}/WHEEL +0 -0
- {dagster_airbyte-0.26.9.dist-info → dagster_airbyte-0.26.11rc0.dist-info}/entry_points.txt +0 -0
- {dagster_airbyte-0.26.9.dist-info → dagster_airbyte-0.26.11rc0.dist-info}/top_level.txt +0 -0
dagster_airbyte/asset_defs.py
CHANGED
|
@@ -179,12 +179,12 @@ def _build_airbyte_assets_from_metadata(
|
|
|
179
179
|
assets_defn_meta: AssetsDefinitionCacheableData,
|
|
180
180
|
resource_defs: Optional[Mapping[str, ResourceDefinition]],
|
|
181
181
|
) -> AssetsDefinition:
|
|
182
|
-
metadata = cast(Mapping[str, Any], assets_defn_meta.extra_metadata)
|
|
183
|
-
connection_id = cast(str, metadata["connection_id"])
|
|
184
|
-
group_name = cast(Optional[str], metadata["group_name"])
|
|
185
|
-
destination_tables = cast(list[str], metadata["destination_tables"])
|
|
186
|
-
normalization_tables = cast(Mapping[str, list[str]], metadata["normalization_tables"])
|
|
187
|
-
io_manager_key = cast(Optional[str], metadata["io_manager_key"])
|
|
182
|
+
metadata = cast("Mapping[str, Any]", assets_defn_meta.extra_metadata)
|
|
183
|
+
connection_id = cast("str", metadata["connection_id"])
|
|
184
|
+
group_name = cast("Optional[str]", metadata["group_name"])
|
|
185
|
+
destination_tables = cast("list[str]", metadata["destination_tables"])
|
|
186
|
+
normalization_tables = cast("Mapping[str, list[str]]", metadata["normalization_tables"])
|
|
187
|
+
io_manager_key = cast("Optional[str]", metadata["io_manager_key"])
|
|
188
188
|
|
|
189
189
|
@multi_asset(
|
|
190
190
|
name=f"airbyte_sync_{connection_id.replace('-', '_')}",
|
|
@@ -501,7 +501,7 @@ class AirbyteConnectionMetadata(
|
|
|
501
501
|
def from_config(
|
|
502
502
|
cls, contents: Mapping[str, Any], destination: Mapping[str, Any]
|
|
503
503
|
) -> "AirbyteConnectionMetadata":
|
|
504
|
-
config_contents = cast(Mapping[str, Any], contents.get("configuration"))
|
|
504
|
+
config_contents = cast("Mapping[str, Any]", contents.get("configuration"))
|
|
505
505
|
check.invariant(
|
|
506
506
|
config_contents is not None, "Airbyte connection config is missing 'configuration' key"
|
|
507
507
|
)
|
|
@@ -531,7 +531,7 @@ class AirbyteConnectionMetadata(
|
|
|
531
531
|
]
|
|
532
532
|
|
|
533
533
|
for stream in enabled_streams:
|
|
534
|
-
name = cast(str, stream.get("stream", {}).get("name"))
|
|
534
|
+
name = cast("str", stream.get("stream", {}).get("name"))
|
|
535
535
|
prefixed_name = f"{self.stream_prefix}{name}"
|
|
536
536
|
|
|
537
537
|
schema = (
|
|
@@ -567,7 +567,7 @@ def _get_schema_by_table_name(
|
|
|
567
567
|
[
|
|
568
568
|
(k, v.schema)
|
|
569
569
|
for k, v in cast(
|
|
570
|
-
dict[str, AirbyteTableMetadata], meta.normalization_tables
|
|
570
|
+
"dict[str, AirbyteTableMetadata]", meta.normalization_tables
|
|
571
571
|
).items()
|
|
572
572
|
]
|
|
573
573
|
for meta in stream_table_metadata.values()
|
|
@@ -740,7 +740,7 @@ class AirbyteInstanceCacheableAssetsDefinition(AirbyteCoreCacheableAssetsDefinit
|
|
|
740
740
|
workspace_id = self._workspace_id
|
|
741
741
|
if not workspace_id:
|
|
742
742
|
workspaces = cast(
|
|
743
|
-
list[dict[str, Any]],
|
|
743
|
+
"list[dict[str, Any]]",
|
|
744
744
|
check.not_none(
|
|
745
745
|
self._airbyte_instance.make_request(endpoint="/workspaces/list", data={})
|
|
746
746
|
).get("workspaces", []),
|
|
@@ -752,7 +752,7 @@ class AirbyteInstanceCacheableAssetsDefinition(AirbyteCoreCacheableAssetsDefinit
|
|
|
752
752
|
workspace_id = workspaces[0].get("workspaceId")
|
|
753
753
|
|
|
754
754
|
connections = cast(
|
|
755
|
-
list[dict[str, Any]],
|
|
755
|
+
"list[dict[str, Any]]",
|
|
756
756
|
check.not_none(
|
|
757
757
|
self._airbyte_instance.make_request(
|
|
758
758
|
endpoint="/connections/list", data={"workspaceId": workspace_id}
|
|
@@ -762,10 +762,10 @@ class AirbyteInstanceCacheableAssetsDefinition(AirbyteCoreCacheableAssetsDefinit
|
|
|
762
762
|
|
|
763
763
|
output_connections: list[tuple[str, AirbyteConnectionMetadata]] = []
|
|
764
764
|
for connection_json in connections:
|
|
765
|
-
connection_id = cast(str, connection_json.get("connectionId"))
|
|
765
|
+
connection_id = cast("str", connection_json.get("connectionId"))
|
|
766
766
|
|
|
767
767
|
operations_json = cast(
|
|
768
|
-
dict[str, Any],
|
|
768
|
+
"dict[str, Any]",
|
|
769
769
|
check.not_none(
|
|
770
770
|
self._airbyte_instance.make_request(
|
|
771
771
|
endpoint="/operations/list",
|
|
@@ -774,9 +774,9 @@ class AirbyteInstanceCacheableAssetsDefinition(AirbyteCoreCacheableAssetsDefinit
|
|
|
774
774
|
),
|
|
775
775
|
)
|
|
776
776
|
|
|
777
|
-
destination_id = cast(str, connection_json.get("destinationId"))
|
|
777
|
+
destination_id = cast("str", connection_json.get("destinationId"))
|
|
778
778
|
destination_json = cast(
|
|
779
|
-
dict[str, Any],
|
|
779
|
+
"dict[str, Any]",
|
|
780
780
|
check.not_none(
|
|
781
781
|
self._airbyte_instance.make_request(
|
|
782
782
|
endpoint="/destinations/get",
|
|
@@ -850,7 +850,7 @@ class AirbyteYAMLCacheableAssetsDefinition(AirbyteCoreCacheableAssetsDefinition)
|
|
|
850
850
|
connection_data = yaml.safe_load(f.read())
|
|
851
851
|
|
|
852
852
|
destination_configuration_path = cast(
|
|
853
|
-
str, connection_data.get("destination_configuration_path")
|
|
853
|
+
"str", connection_data.get("destination_configuration_path")
|
|
854
854
|
)
|
|
855
855
|
with open(
|
|
856
856
|
os.path.join(self._project_dir, destination_configuration_path), encoding="utf-8"
|
|
@@ -886,7 +886,7 @@ class AirbyteYAMLCacheableAssetsDefinition(AirbyteCoreCacheableAssetsDefinition)
|
|
|
886
886
|
)
|
|
887
887
|
state_file = state_files[0]
|
|
888
888
|
|
|
889
|
-
with open(os.path.join(connection_dir, cast(str, state_file)), encoding="utf-8") as f:
|
|
889
|
+
with open(os.path.join(connection_dir, cast("str", state_file)), encoding="utf-8") as f:
|
|
890
890
|
state = yaml.safe_load(f.read())
|
|
891
891
|
connection_id = state.get("resource_id")
|
|
892
892
|
|
|
@@ -212,7 +212,7 @@ def reconcile_sources(
|
|
|
212
212
|
else:
|
|
213
213
|
if not dry_run:
|
|
214
214
|
create_result = cast(
|
|
215
|
-
dict[str, str],
|
|
215
|
+
"dict[str, str]",
|
|
216
216
|
check.not_none(
|
|
217
217
|
res.make_request(
|
|
218
218
|
endpoint="/sources/create",
|
|
@@ -300,7 +300,7 @@ def reconcile_destinations(
|
|
|
300
300
|
else:
|
|
301
301
|
if not dry_run:
|
|
302
302
|
create_result = cast(
|
|
303
|
-
dict[str, str],
|
|
303
|
+
"dict[str, str]",
|
|
304
304
|
check.not_none(
|
|
305
305
|
res.make_request(
|
|
306
306
|
endpoint="/destinations/create",
|
|
@@ -346,13 +346,13 @@ def reconcile_config(
|
|
|
346
346
|
workspace_id = res.get_default_workspace()
|
|
347
347
|
|
|
348
348
|
existing_sources_raw = cast(
|
|
349
|
-
dict[str, list[dict[str, Any]]],
|
|
349
|
+
"dict[str, list[dict[str, Any]]]",
|
|
350
350
|
check.not_none(
|
|
351
351
|
res.make_request(endpoint="/sources/list", data={"workspaceId": workspace_id})
|
|
352
352
|
),
|
|
353
353
|
)
|
|
354
354
|
existing_dests_raw = cast(
|
|
355
|
-
dict[str, list[dict[str, Any]]],
|
|
355
|
+
"dict[str, list[dict[str, Any]]]",
|
|
356
356
|
check.not_none(
|
|
357
357
|
res.make_request(endpoint="/destinations/list", data={"workspaceId": workspace_id})
|
|
358
358
|
),
|
|
@@ -422,7 +422,7 @@ def reconcile_normalization(
|
|
|
422
422
|
existing_basic_norm_op_id = None
|
|
423
423
|
if existing_connection_id:
|
|
424
424
|
operations = cast(
|
|
425
|
-
dict[str, list[dict[str, str]]],
|
|
425
|
+
"dict[str, list[dict[str, str]]]",
|
|
426
426
|
check.not_none(
|
|
427
427
|
res.make_request(
|
|
428
428
|
endpoint="/operations/list",
|
|
@@ -450,7 +450,7 @@ def reconcile_normalization(
|
|
|
450
450
|
return existing_basic_norm_op_id
|
|
451
451
|
else:
|
|
452
452
|
return cast(
|
|
453
|
-
dict[str, str],
|
|
453
|
+
"dict[str, str]",
|
|
454
454
|
check.not_none(
|
|
455
455
|
res.make_request(
|
|
456
456
|
endpoint="/operations/create",
|
|
@@ -492,7 +492,7 @@ def reconcile_connections_pre(
|
|
|
492
492
|
diff = ManagedElementDiff()
|
|
493
493
|
|
|
494
494
|
existing_connections_raw = cast(
|
|
495
|
-
dict[str, list[dict[str, Any]]],
|
|
495
|
+
"dict[str, list[dict[str, Any]]]",
|
|
496
496
|
check.not_none(
|
|
497
497
|
res.make_request(endpoint="/connections/list", data={"workspaceId": workspace_id})
|
|
498
498
|
),
|
|
@@ -537,7 +537,7 @@ def reconcile_connections_post(
|
|
|
537
537
|
) -> None:
|
|
538
538
|
"""Creates new and modifies existing connections based on the config if dry_run is False."""
|
|
539
539
|
existing_connections_raw = cast(
|
|
540
|
-
dict[str, list[dict[str, Any]]],
|
|
540
|
+
"dict[str, list[dict[str, Any]]]",
|
|
541
541
|
check.not_none(
|
|
542
542
|
res.make_request(endpoint="/connections/list", data={"workspaceId": workspace_id})
|
|
543
543
|
),
|
|
@@ -592,7 +592,7 @@ def reconcile_connections_post(
|
|
|
592
592
|
connection_base_json["namespaceDefinition"] = config_conn.destination_namespace.value
|
|
593
593
|
else:
|
|
594
594
|
connection_base_json["namespaceDefinition"] = "customformat"
|
|
595
|
-
connection_base_json["namespaceFormat"] = cast(str, config_conn.destination_namespace)
|
|
595
|
+
connection_base_json["namespaceFormat"] = cast("str", config_conn.destination_namespace)
|
|
596
596
|
|
|
597
597
|
if config_conn.prefix:
|
|
598
598
|
connection_base_json["prefix"] = config_conn.prefix
|
dagster_airbyte/resources.py
CHANGED
|
@@ -217,8 +217,8 @@ class BaseAirbyteResource(ConfigurableResource):
|
|
|
217
217
|
"""
|
|
218
218
|
connection_details = self.get_connection_details(connection_id)
|
|
219
219
|
job_details = self.start_sync(connection_id)
|
|
220
|
-
job_info = cast(dict[str, object], job_details.get("job", {}))
|
|
221
|
-
job_id = cast(int, job_info.get("id"))
|
|
220
|
+
job_info = cast("dict[str, object]", job_details.get("job", {}))
|
|
221
|
+
job_id = cast("int", job_info.get("id"))
|
|
222
222
|
|
|
223
223
|
self._log.info(f"Job {job_id} initialized for connection_id={connection_id}.")
|
|
224
224
|
start = time.monotonic()
|
|
@@ -235,7 +235,7 @@ class BaseAirbyteResource(ConfigurableResource):
|
|
|
235
235
|
)
|
|
236
236
|
time.sleep(poll_interval or self.poll_interval)
|
|
237
237
|
job_details = self.get_job_status(connection_id, job_id)
|
|
238
|
-
attempts = cast(list, job_details.get("attempts", []))
|
|
238
|
+
attempts = cast("list", job_details.get("attempts", []))
|
|
239
239
|
cur_attempt = len(attempts)
|
|
240
240
|
# spit out the available Airbyte log info
|
|
241
241
|
if cur_attempt:
|
|
@@ -252,7 +252,7 @@ class BaseAirbyteResource(ConfigurableResource):
|
|
|
252
252
|
logged_lines = 0
|
|
253
253
|
logged_attempts += 1
|
|
254
254
|
|
|
255
|
-
job_info = cast(dict[str, object], job_details.get("job", {}))
|
|
255
|
+
job_info = cast("dict[str, object]", job_details.get("job", {}))
|
|
256
256
|
state = job_info.get("status")
|
|
257
257
|
|
|
258
258
|
if state in (
|
|
@@ -577,7 +577,7 @@ class AirbyteResource(BaseAirbyteResource):
|
|
|
577
577
|
|
|
578
578
|
def get_default_workspace(self) -> str:
|
|
579
579
|
workspaces = cast(
|
|
580
|
-
list[dict[str, Any]],
|
|
580
|
+
"list[dict[str, Any]]",
|
|
581
581
|
check.not_none(self.make_request_cached(endpoint="/workspaces/list", data={})).get(
|
|
582
582
|
"workspaces", []
|
|
583
583
|
),
|
|
@@ -589,7 +589,7 @@ class AirbyteResource(BaseAirbyteResource):
|
|
|
589
589
|
definitions = check.not_none(
|
|
590
590
|
self.make_request_cached(endpoint="/source_definitions/list", data={})
|
|
591
591
|
)
|
|
592
|
-
source_definitions = cast(list[dict[str, Any]], definitions["sourceDefinitions"])
|
|
592
|
+
source_definitions = cast("list[dict[str, Any]]", definitions["sourceDefinitions"])
|
|
593
593
|
|
|
594
594
|
return next(
|
|
595
595
|
(
|
|
@@ -603,7 +603,7 @@ class AirbyteResource(BaseAirbyteResource):
|
|
|
603
603
|
def get_destination_definition_by_name(self, name: str):
|
|
604
604
|
name_lower = name.lower()
|
|
605
605
|
definitions = cast(
|
|
606
|
-
dict[str, list[dict[str, str]]],
|
|
606
|
+
"dict[str, list[dict[str, str]]]",
|
|
607
607
|
check.not_none(
|
|
608
608
|
self.make_request_cached(endpoint="/destination_definitions/list", data={})
|
|
609
609
|
),
|
|
@@ -619,7 +619,7 @@ class AirbyteResource(BaseAirbyteResource):
|
|
|
619
619
|
|
|
620
620
|
def get_source_catalog_id(self, source_id: str):
|
|
621
621
|
result = cast(
|
|
622
|
-
dict[str, Any],
|
|
622
|
+
"dict[str, Any]",
|
|
623
623
|
check.not_none(
|
|
624
624
|
self.make_request(endpoint="/sources/discover_schema", data={"sourceId": source_id})
|
|
625
625
|
),
|
|
@@ -628,7 +628,7 @@ class AirbyteResource(BaseAirbyteResource):
|
|
|
628
628
|
|
|
629
629
|
def get_source_schema(self, source_id: str) -> Mapping[str, Any]:
|
|
630
630
|
return cast(
|
|
631
|
-
dict[str, Any],
|
|
631
|
+
"dict[str, Any]",
|
|
632
632
|
check.not_none(
|
|
633
633
|
self.make_request(endpoint="/sources/discover_schema", data={"sourceId": source_id})
|
|
634
634
|
),
|
|
@@ -640,7 +640,7 @@ class AirbyteResource(BaseAirbyteResource):
|
|
|
640
640
|
# Airbyte API changed source of truth for normalization in PR
|
|
641
641
|
# https://github.com/airbytehq/airbyte/pull/21005
|
|
642
642
|
norm_dest_def_spec: bool = cast(
|
|
643
|
-
dict[str, Any],
|
|
643
|
+
"dict[str, Any]",
|
|
644
644
|
check.not_none(
|
|
645
645
|
self.make_request_cached(
|
|
646
646
|
endpoint="/destination_definition_specifications/get",
|
|
@@ -654,7 +654,7 @@ class AirbyteResource(BaseAirbyteResource):
|
|
|
654
654
|
|
|
655
655
|
norm_dest_def: bool = (
|
|
656
656
|
cast(
|
|
657
|
-
dict[str, Any],
|
|
657
|
+
"dict[str, Any]",
|
|
658
658
|
check.not_none(
|
|
659
659
|
self.make_request_cached(
|
|
660
660
|
endpoint="/destination_definitions/get",
|
|
@@ -687,7 +687,9 @@ class AirbyteResource(BaseAirbyteResource):
|
|
|
687
687
|
},
|
|
688
688
|
)
|
|
689
689
|
)
|
|
690
|
-
job = next(
|
|
690
|
+
job = next(
|
|
691
|
+
(job for job in cast("list", out["jobs"]) if job["job"]["id"] == job_id), None
|
|
692
|
+
)
|
|
691
693
|
|
|
692
694
|
return check.not_none(job)
|
|
693
695
|
|
|
@@ -722,8 +724,8 @@ class AirbyteResource(BaseAirbyteResource):
|
|
|
722
724
|
"""
|
|
723
725
|
connection_details = self.get_connection_details(connection_id)
|
|
724
726
|
job_details = self.start_sync(connection_id)
|
|
725
|
-
job_info = cast(dict[str, object], job_details.get("job", {}))
|
|
726
|
-
job_id = cast(int, job_info.get("id"))
|
|
727
|
+
job_info = cast("dict[str, object]", job_details.get("job", {}))
|
|
728
|
+
job_id = cast("int", job_info.get("id"))
|
|
727
729
|
|
|
728
730
|
self._log.info(f"Job {job_id} initialized for connection_id={connection_id}.")
|
|
729
731
|
start = time.monotonic()
|
|
@@ -740,7 +742,7 @@ class AirbyteResource(BaseAirbyteResource):
|
|
|
740
742
|
)
|
|
741
743
|
time.sleep(poll_interval or self.poll_interval)
|
|
742
744
|
job_details = self.get_job_status(connection_id, job_id)
|
|
743
|
-
attempts = cast(list, job_details.get("attempts", []))
|
|
745
|
+
attempts = cast("list", job_details.get("attempts", []))
|
|
744
746
|
cur_attempt = len(attempts)
|
|
745
747
|
# spit out the available Airbyte log info
|
|
746
748
|
if cur_attempt:
|
|
@@ -757,7 +759,7 @@ class AirbyteResource(BaseAirbyteResource):
|
|
|
757
759
|
logged_lines = 0
|
|
758
760
|
logged_attempts += 1
|
|
759
761
|
|
|
760
|
-
job_info = cast(dict[str, object], job_details.get("job", {}))
|
|
762
|
+
job_info = cast("dict[str, object]", job_details.get("job", {}))
|
|
761
763
|
state = job_info.get("status")
|
|
762
764
|
|
|
763
765
|
if state in (
|
dagster_airbyte/version.py
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
__version__ = "0.26.
|
|
1
|
+
__version__ = "0.26.11rc0"
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.1
|
|
2
2
|
Name: dagster-airbyte
|
|
3
|
-
Version: 0.26.
|
|
3
|
+
Version: 0.26.11rc0
|
|
4
4
|
Summary: Package for integrating Airbyte with Dagster.
|
|
5
5
|
Home-page: https://github.com/dagster-io/dagster/tree/master/python_modules/libraries/dagster-airbyte
|
|
6
6
|
Author: Dagster Labs
|
|
@@ -14,10 +14,10 @@ Classifier: License :: OSI Approved :: Apache Software License
|
|
|
14
14
|
Classifier: Operating System :: OS Independent
|
|
15
15
|
Requires-Python: >=3.9,<3.13
|
|
16
16
|
License-File: LICENSE
|
|
17
|
-
Requires-Dist: dagster ==1.10.
|
|
17
|
+
Requires-Dist: dagster ==1.10.11rc0
|
|
18
18
|
Requires-Dist: requests
|
|
19
19
|
Provides-Extra: managed
|
|
20
|
-
Requires-Dist: dagster-managed-elements ==0.26.
|
|
20
|
+
Requires-Dist: dagster-managed-elements ==0.26.11rc0 ; extra == 'managed'
|
|
21
21
|
Provides-Extra: test
|
|
22
22
|
Requires-Dist: requests-mock ; extra == 'test'
|
|
23
23
|
Requires-Dist: flaky ; extra == 'test'
|
|
@@ -1,23 +1,23 @@
|
|
|
1
1
|
dagster_airbyte/__init__.py,sha256=Cdn8CiAL5HSFeFjmKIZgXJqccj11xqQFzRfUQlD15b0,1720
|
|
2
2
|
dagster_airbyte/asset_decorator.py,sha256=8lvhY8OGk9KT2RS7Fwmm9ZxGUFmhNpGKPKq-lObaxRg,4193
|
|
3
|
-
dagster_airbyte/asset_defs.py,sha256=
|
|
3
|
+
dagster_airbyte/asset_defs.py,sha256=X-ong0InnxXe61_H9IZc6ATCACylKb-JQrPbtH-XTHo,50019
|
|
4
4
|
dagster_airbyte/cli.py,sha256=HErteP1MjfHozKKSrznh0yAreKETbXp5NDHzXGsdvvE,425
|
|
5
5
|
dagster_airbyte/ops.py,sha256=oOEczVYpqVRTc1-0osfpR5FZbPjNJDYihgRjk_qtOQA,4229
|
|
6
6
|
dagster_airbyte/py.typed,sha256=la67KBlbjXN-_-DfGNcdOcjYumVpKG_Tkw-8n5dnGB4,8
|
|
7
|
-
dagster_airbyte/resources.py,sha256=
|
|
7
|
+
dagster_airbyte/resources.py,sha256=5aGrPmH7E5MT3HKgnglxyqRA9l-mUVAj0H0pMj4ysMQ,54071
|
|
8
8
|
dagster_airbyte/translator.py,sha256=CL2EKeZIahgoFvtZpfEwVa2KmCuvUENXByM6F7369LQ,7529
|
|
9
9
|
dagster_airbyte/types.py,sha256=TYUjI3skjLYeANjesgJ-IAJNu8bAnL1ymsUfz5LsRTE,1565
|
|
10
10
|
dagster_airbyte/utils.py,sha256=wJtzx5cWq8Bp1f3T4XjLPaxjXZGstHlL4kRYP2c49bs,4157
|
|
11
|
-
dagster_airbyte/version.py,sha256=
|
|
11
|
+
dagster_airbyte/version.py,sha256=eS-xt-o3NPyxbPUwgn1WSzCSvbkXM3calY6j5hmh228,27
|
|
12
12
|
dagster_airbyte/managed/__init__.py,sha256=6SBtyNOMJ9Cu2UIwFExJHpL_ZVFo3rPMvyIxVOsKvWE,469
|
|
13
|
-
dagster_airbyte/managed/reconciliation.py,sha256=
|
|
13
|
+
dagster_airbyte/managed/reconciliation.py,sha256=LHtJc2zRwk1gAc86pKVSZ0rnfLNTT9r3usRFQilYCwY,34803
|
|
14
14
|
dagster_airbyte/managed/types.py,sha256=isPfX8L9YwtZAf9Vk4hhxBePLR00AEldsdK2TsM1H2o,14611
|
|
15
15
|
dagster_airbyte/managed/generated/__init__.py,sha256=eYq-yfXEeffuKAVFXY8plD0se1wHjFNVqklpbu9gljw,108
|
|
16
16
|
dagster_airbyte/managed/generated/destinations.py,sha256=x1wmWlXvOJHtfaZva3ErdKuVS--sDvfidSXR5ji9G5w,119692
|
|
17
17
|
dagster_airbyte/managed/generated/sources.py,sha256=y0TPNvcRd8c9mhje-NoXsHeKRPt1nXcpww8mNAtqCps,282685
|
|
18
|
-
dagster_airbyte-0.26.
|
|
19
|
-
dagster_airbyte-0.26.
|
|
20
|
-
dagster_airbyte-0.26.
|
|
21
|
-
dagster_airbyte-0.26.
|
|
22
|
-
dagster_airbyte-0.26.
|
|
23
|
-
dagster_airbyte-0.26.
|
|
18
|
+
dagster_airbyte-0.26.11rc0.dist-info/LICENSE,sha256=TMatHW4_G9ldRdodEAp-l2Xa2WvsdeOh60E3v1R2jis,11349
|
|
19
|
+
dagster_airbyte-0.26.11rc0.dist-info/METADATA,sha256=T96IiJ8u_IKB02rzxFjRzckiccbCmtHbKag-MTDHQyY,928
|
|
20
|
+
dagster_airbyte-0.26.11rc0.dist-info/WHEEL,sha256=yQN5g4mg4AybRjkgi-9yy4iQEFibGQmlz78Pik5Or-A,92
|
|
21
|
+
dagster_airbyte-0.26.11rc0.dist-info/entry_points.txt,sha256=XrbLOz3LpgPV5fdwMmgdP6Rp1AfSG07KeWIddLqh7Lw,61
|
|
22
|
+
dagster_airbyte-0.26.11rc0.dist-info/top_level.txt,sha256=HLwIRQCzqItn88_KbPP8DNTKKQEBUVKk6NCn4PrCtqY,16
|
|
23
|
+
dagster_airbyte-0.26.11rc0.dist-info/RECORD,,
|
|
File without changes
|
|
File without changes
|
|
File without changes
|
|
File without changes
|