datapipelab 0.1.7__py3-none-any.whl → 0.1.9__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -3,8 +3,6 @@ from datapipelab.logger import logger
3
3
 
4
4
  class BigQueryAPIProcessorNode(TNode):
5
5
  def __init__(self, spark, tnode_config):
6
- from google.cloud import bigquery
7
- from google.oauth2 import service_account
8
6
  super().__init__(spark=spark)
9
7
  self.sql_query = tnode_config['options']['query']
10
8
  self.node_name = tnode_config['name']
@@ -12,7 +10,9 @@ class BigQueryAPIProcessorNode(TNode):
12
10
  self.return_as_spark_df = tnode_config['options']['return_as_spark_df']
13
11
  self.project_name = tnode_config['options']['project_name']
14
12
 
15
- def __sql_query(self, sql_query):
13
+ def __sql_biqquery(self, sql_query):
14
+ from google.cloud import bigquery
15
+ from google.oauth2 import service_account
16
16
  credentials = service_account.Credentials.from_service_account_file(self.credentials_path)
17
17
  client = bigquery.Client(credentials=credentials, project=self.project_name)
18
18
 
@@ -28,6 +28,6 @@ class BigQueryAPIProcessorNode(TNode):
28
28
  logger.info(rows)
29
29
 
30
30
  def _process(self):
31
- self.__sql_query(self.sql_query)
31
+ self.__sql_biqquery(self.sql_query)
32
32
  self._createOrReplaceTempView()
33
33
  return self.node
@@ -1,30 +1,22 @@
1
1
  from datapipelab.app.node.tnode import TNode
2
2
  from datapipelab.logger import logger
3
3
 
4
+
4
5
  class BigQuerySparkProcessorNode(TNode):
5
6
  def __init__(self, spark, tnode_config):
6
7
  super().__init__(spark=spark)
7
8
  self.sql_query = tnode_config['options']['query']
8
9
  self.node_name = tnode_config['name']
9
- self.credentials_path = tnode_config['options']['materialization_dataset'] # materializationDataset
10
- self.return_as_spark_df = tnode_config['options']['parent_project'] # parentProject
11
-
12
- def __sql_query(self, sql_query):
13
- credentials = service_account.Credentials.from_service_account_file(self.credentials_path)
14
- client = bigquery.Client(credentials=credentials, project=self.project_name)
15
-
16
- # run the job
17
- query_job = client.query(sql_query)
10
+ self.materialization_dataset = tnode_config['options']['materialization_dataset'] # materializationDataset
11
+ self.parent_project = tnode_config['options']['parent_project'] # parentProject
18
12
 
19
- results = query_job.result()
20
- rows = [dict(row) for row in results]
21
- if self.return_as_spark_df:
22
- self.node = self.spark.createDataFrame(rows)
23
- else:
24
- self.node = None
25
- logger.info(rows)
13
+ def __sql_query(self):
14
+ self.node = self.spark.read.format("bigquery").option("materializationDataset",
15
+ self.materialization_dataset).option("query",
16
+ self.sql_query).option(
17
+ "parentProject", self.parent_project).load()
26
18
 
27
19
  def _process(self):
28
- self.__sql_query(self.sql_query)
20
+ self.__sql_query()
29
21
  self._createOrReplaceTempView()
30
22
  return self.node
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: datapipelab
3
- Version: 0.1.7
3
+ Version: 0.1.9
4
4
  Summary: A data pipeline library with connectors, sources, processors, and sinks.
5
5
  Requires-Dist: json5
6
6
  Requires-Dist: loguru
@@ -10,8 +10,8 @@ datapipelab/app/node/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hS
10
10
  datapipelab/app/node/custom_node.py,sha256=VvjwkECTobRhO_fYKUrJCd117B5MoR9P6UKYZfRLhV4,1017
11
11
  datapipelab/app/node/tnode.py,sha256=npHG4fFZty5JZ3F_okO9xml-BRhu4DkrZuNE6oaLbvw,446
12
12
  datapipelab/app/node/processor/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
13
- datapipelab/app/node/processor/bigquery_api_node.py,sha256=2kF6hgYOUi7te-aXXr-cTZfkvitBDEPJAHT8KG-i8fE,1293
14
- datapipelab/app/node/processor/bigquery_spark_node.py,sha256=t8JJqMWTZwBuJUKV7-l72ZLdLVoHHSFJHFNovAY-2nc,1179
13
+ datapipelab/app/node/processor/bigquery_api_node.py,sha256=uBq9Ruk791Sa4qUK7-dnGPklXH68xqWy0yvQsASwHP8,1299
14
+ datapipelab/app/node/processor/bigquery_spark_node.py,sha256=S9kIYW0RE5b0RjniKFFBTzA3Tx4_plFdkFQXzhl1xTY,1039
15
15
  datapipelab/app/node/processor/custom_node.py,sha256=1nqbJEhNiMP1rmN9ufpUuKO1IkuI2BEM5auW4JceGMA,933
16
16
  datapipelab/app/node/processor/shell_node.py,sha256=s3dKgfEqbpUIEiwORERgvp7FNDE5JkFHBo7EnJYBPnA,669
17
17
  datapipelab/app/node/processor/spark_node.py,sha256=jzqdffIHUCgOfMFcoqjXdl8wFag-3gafxfNCdssKnwc,483
@@ -25,7 +25,7 @@ datapipelab/app/node/source/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMp
25
25
  datapipelab/app/node/source/delta_node.py,sha256=gg7SfuKBAAfjk6OX2jNrot9XX61HoBe3us3D8O-dscE,529
26
26
  datapipelab/app/node/source/hive_node.py,sha256=h_AMCnnmon7uLRIGsaHAPWEReD3VaWZXnz9r0TpLGNM,478
27
27
  datapipelab/app/node/source/spark_node.py,sha256=S_x2atRFPDnXmhCUtcmaLc4BDFd2H4uQq6wnEJb7Uug,480
28
- datapipelab-0.1.7.dist-info/METADATA,sha256=oiRy9y4GTRDU-Yiz3EQJc831TRgaTiMRXOaNimFllTQ,220
29
- datapipelab-0.1.7.dist-info/WHEEL,sha256=Nw36Djuh_5VDukK0H78QzOX-_FQEo6V37m3nkm96gtU,91
30
- datapipelab-0.1.7.dist-info/top_level.txt,sha256=HgeBjHvXorKzvNqU5BNPutoI771HtiqVit9_-0Zyrb4,12
31
- datapipelab-0.1.7.dist-info/RECORD,,
28
+ datapipelab-0.1.9.dist-info/METADATA,sha256=GPQ-3fWDVD8G5z4rv-fjaRZ6L61U_a_pmuagIJw66Us,220
29
+ datapipelab-0.1.9.dist-info/WHEEL,sha256=zaaOINJESkSfm_4HQVc5ssNzHCPXhJm0kEUakpsEHaU,91
30
+ datapipelab-0.1.9.dist-info/top_level.txt,sha256=HgeBjHvXorKzvNqU5BNPutoI771HtiqVit9_-0Zyrb4,12
31
+ datapipelab-0.1.9.dist-info/RECORD,,
@@ -1,5 +1,5 @@
1
1
  Wheel-Version: 1.0
2
- Generator: setuptools (80.7.1)
2
+ Generator: setuptools (80.8.0)
3
3
  Root-Is-Purelib: true
4
4
  Tag: py3-none-any
5
5