google-cloud-data_catalog-v1 0.15.0 → 0.16.0
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/lib/google/cloud/data_catalog/v1/data_catalog/client.rb +18 -47
- data/lib/google/cloud/data_catalog/v1/data_catalog/rest/client.rb +18 -47
- data/lib/google/cloud/data_catalog/v1/version.rb +1 -1
- data/lib/google/cloud/datacatalog/v1/bigquery_pb.rb +24 -25
- data/lib/google/cloud/datacatalog/v1/common_pb.rb +25 -19
- data/lib/google/cloud/datacatalog/v1/data_source_pb.rb +24 -18
- data/lib/google/cloud/datacatalog/v1/datacatalog_pb.rb +41 -339
- data/lib/google/cloud/datacatalog/v1/dataplex_spec_pb.rb +25 -21
- data/lib/google/cloud/datacatalog/v1/dump_content_pb.rb +27 -13
- data/lib/google/cloud/datacatalog/v1/gcs_fileset_spec_pb.rb +25 -10
- data/lib/google/cloud/datacatalog/v1/physical_schema_pb.rb +24 -26
- data/lib/google/cloud/datacatalog/v1/policytagmanager_pb.rb +26 -72
- data/lib/google/cloud/datacatalog/v1/policytagmanagerserialization_pb.rb +25 -43
- data/lib/google/cloud/datacatalog/v1/schema_pb.rb +24 -36
- data/lib/google/cloud/datacatalog/v1/search_pb.rb +25 -21
- data/lib/google/cloud/datacatalog/v1/table_spec_pb.rb +24 -26
- data/lib/google/cloud/datacatalog/v1/tags_pb.rb +25 -59
- data/lib/google/cloud/datacatalog/v1/timestamps_pb.rb +25 -6
- data/lib/google/cloud/datacatalog/v1/usage_pb.rb +25 -16
- data/proto_docs/google/api/client.rb +58 -1
- data/proto_docs/google/cloud/datacatalog/v1/common.rb +6 -0
- data/proto_docs/google/cloud/datacatalog/v1/datacatalog.rb +84 -20
- data/proto_docs/google/cloud/datacatalog/v1/dump_content.rb +2 -2
- data/proto_docs/google/cloud/datacatalog/v1/policytagmanager.rb +1 -1
- data/proto_docs/google/protobuf/any.rb +7 -4
- data/proto_docs/google/protobuf/timestamp.rb +1 -3
- metadata +4 -4
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: 287a106028e030737475116a68767810728bbe6300dcd308072e53eef227495e
|
4
|
+
data.tar.gz: 3ff5fb58112bfb569cdef3307029a14e7f7d0f1de030cf03ccde349352d039a3
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: 95d4feec2aeb82509fb40dd46b2e1fcc5057f566fdfbd0a0c1038fb5bc7ac11abdb2f8c3cdd1c2c32edd0d10fc29ec9dbdf61efb21c60a5dada00f999fbcb379
|
7
|
+
data.tar.gz: e8c76153b2b5814d3f1243eff17e20462ab6166706d077499514315ee7dc07be0311ac9ddb30b928a41546db9313112bf96bf052bba70e95b0af213df887a843
|
@@ -67,50 +67,8 @@ module Google
|
|
67
67
|
default_config = Client::Configuration.new parent_config
|
68
68
|
|
69
69
|
default_config.timeout = 60.0
|
70
|
-
|
71
|
-
|
72
|
-
default_config.rpcs.search_catalog.retry_policy = {
|
73
|
-
initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
|
74
|
-
}
|
75
|
-
|
76
|
-
default_config.rpcs.get_entry_group.timeout = 60.0
|
77
|
-
default_config.rpcs.get_entry_group.retry_policy = {
|
78
|
-
initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
|
79
|
-
}
|
80
|
-
|
81
|
-
default_config.rpcs.list_entry_groups.timeout = 60.0
|
82
|
-
default_config.rpcs.list_entry_groups.retry_policy = {
|
83
|
-
initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
|
84
|
-
}
|
85
|
-
|
86
|
-
default_config.rpcs.get_entry.timeout = 60.0
|
87
|
-
default_config.rpcs.get_entry.retry_policy = {
|
88
|
-
initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
|
89
|
-
}
|
90
|
-
|
91
|
-
default_config.rpcs.lookup_entry.timeout = 60.0
|
92
|
-
default_config.rpcs.lookup_entry.retry_policy = {
|
93
|
-
initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
|
94
|
-
}
|
95
|
-
|
96
|
-
default_config.rpcs.list_entries.timeout = 60.0
|
97
|
-
default_config.rpcs.list_entries.retry_policy = {
|
98
|
-
initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
|
99
|
-
}
|
100
|
-
|
101
|
-
default_config.rpcs.list_tags.timeout = 60.0
|
102
|
-
default_config.rpcs.list_tags.retry_policy = {
|
103
|
-
initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
|
104
|
-
}
|
105
|
-
|
106
|
-
default_config.rpcs.get_iam_policy.timeout = 60.0
|
107
|
-
default_config.rpcs.get_iam_policy.retry_policy = {
|
108
|
-
initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
|
109
|
-
}
|
110
|
-
|
111
|
-
default_config.rpcs.import_entries.timeout = 60.0
|
112
|
-
default_config.rpcs.import_entries.retry_policy = {
|
113
|
-
initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
|
70
|
+
default_config.retry_policy = {
|
71
|
+
initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14, 8, 13]
|
114
72
|
}
|
115
73
|
|
116
74
|
default_config
|
@@ -1275,7 +1233,7 @@ module Google
|
|
1275
1233
|
# @param options [::Gapic::CallOptions, ::Hash]
|
1276
1234
|
# Overrides the default settings for this call, e.g, timeout, retries, etc. Optional.
|
1277
1235
|
#
|
1278
|
-
# @overload lookup_entry(linked_resource: nil, sql_resource: nil, fully_qualified_name: nil)
|
1236
|
+
# @overload lookup_entry(linked_resource: nil, sql_resource: nil, fully_qualified_name: nil, project: nil, location: nil)
|
1279
1237
|
# Pass arguments to `lookup_entry` via keyword arguments. Note that at
|
1280
1238
|
# least one keyword argument is required. To specify no parameters, or to keep all
|
1281
1239
|
# the default parameter values, pass an empty Hash as a request object (see above).
|
@@ -1304,7 +1262,9 @@ module Google
|
|
1304
1262
|
# [Lexical structure in Standard SQL]
|
1305
1263
|
# (https://cloud.google.com/bigquery/docs/reference/standard-sql/lexical).
|
1306
1264
|
# @param fully_qualified_name [::String]
|
1307
|
-
# Fully
|
1265
|
+
# [Fully Qualified Name
|
1266
|
+
# (FQN)](https://cloud.google.com//data-catalog/docs/fully-qualified-names)
|
1267
|
+
# of the resource.
|
1308
1268
|
#
|
1309
1269
|
# FQNs take two forms:
|
1310
1270
|
#
|
@@ -1319,6 +1279,14 @@ module Google
|
|
1319
1279
|
# Example for a DPMS table:
|
1320
1280
|
#
|
1321
1281
|
# `dataproc_metastore:{PROJECT_ID}.{LOCATION_ID}.{INSTANCE_ID}.{DATABASE_ID}.{TABLE_ID}`
|
1282
|
+
# @param project [::String]
|
1283
|
+
# Project where the lookup should be performed. Required to lookup
|
1284
|
+
# entry that is not a part of `DPMS` or `DATAPLEX` `integrated_system`
|
1285
|
+
# using its `fully_qualified_name`. Ignored in other cases.
|
1286
|
+
# @param location [::String]
|
1287
|
+
# Location where the lookup should be performed. Required to lookup
|
1288
|
+
# entry that is not a part of `DPMS` or `DATAPLEX` `integrated_system`
|
1289
|
+
# using its `fully_qualified_name`. Ignored in other cases.
|
1322
1290
|
#
|
1323
1291
|
# @yield [response, operation] Access the result along with the RPC operation
|
1324
1292
|
# @yieldparam response [::Google::Cloud::DataCatalog::V1::Entry]
|
@@ -3560,7 +3528,7 @@ module Google
|
|
3560
3528
|
# @param options [::Gapic::CallOptions, ::Hash]
|
3561
3529
|
# Overrides the default settings for this call, e.g, timeout, retries, etc. Optional.
|
3562
3530
|
#
|
3563
|
-
# @overload import_entries(parent: nil, gcs_bucket_path: nil)
|
3531
|
+
# @overload import_entries(parent: nil, gcs_bucket_path: nil, job_id: nil)
|
3564
3532
|
# Pass arguments to `import_entries` via keyword arguments. Note that at
|
3565
3533
|
# least one keyword argument is required. To specify no parameters, or to keep all
|
3566
3534
|
# the default parameter values, pass an empty Hash as a request object (see above).
|
@@ -3569,6 +3537,9 @@ module Google
|
|
3569
3537
|
# Required. Target entry group for ingested entries.
|
3570
3538
|
# @param gcs_bucket_path [::String]
|
3571
3539
|
# Path to a Cloud Storage bucket that contains a dump ready for ingestion.
|
3540
|
+
# @param job_id [::String]
|
3541
|
+
# Optional. (Optional) Dataplex task job id, if specified will be used as
|
3542
|
+
# part of ImportEntries LRO ID
|
3572
3543
|
#
|
3573
3544
|
# @yield [response, operation] Access the result along with the RPC operation
|
3574
3545
|
# @yieldparam response [::Gapic::Operation]
|
@@ -69,50 +69,8 @@ module Google
|
|
69
69
|
default_config = Client::Configuration.new parent_config
|
70
70
|
|
71
71
|
default_config.timeout = 60.0
|
72
|
-
|
73
|
-
|
74
|
-
default_config.rpcs.search_catalog.retry_policy = {
|
75
|
-
initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
|
76
|
-
}
|
77
|
-
|
78
|
-
default_config.rpcs.get_entry_group.timeout = 60.0
|
79
|
-
default_config.rpcs.get_entry_group.retry_policy = {
|
80
|
-
initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
|
81
|
-
}
|
82
|
-
|
83
|
-
default_config.rpcs.list_entry_groups.timeout = 60.0
|
84
|
-
default_config.rpcs.list_entry_groups.retry_policy = {
|
85
|
-
initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
|
86
|
-
}
|
87
|
-
|
88
|
-
default_config.rpcs.get_entry.timeout = 60.0
|
89
|
-
default_config.rpcs.get_entry.retry_policy = {
|
90
|
-
initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
|
91
|
-
}
|
92
|
-
|
93
|
-
default_config.rpcs.lookup_entry.timeout = 60.0
|
94
|
-
default_config.rpcs.lookup_entry.retry_policy = {
|
95
|
-
initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
|
96
|
-
}
|
97
|
-
|
98
|
-
default_config.rpcs.list_entries.timeout = 60.0
|
99
|
-
default_config.rpcs.list_entries.retry_policy = {
|
100
|
-
initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
|
101
|
-
}
|
102
|
-
|
103
|
-
default_config.rpcs.list_tags.timeout = 60.0
|
104
|
-
default_config.rpcs.list_tags.retry_policy = {
|
105
|
-
initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
|
106
|
-
}
|
107
|
-
|
108
|
-
default_config.rpcs.get_iam_policy.timeout = 60.0
|
109
|
-
default_config.rpcs.get_iam_policy.retry_policy = {
|
110
|
-
initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
|
111
|
-
}
|
112
|
-
|
113
|
-
default_config.rpcs.import_entries.timeout = 60.0
|
114
|
-
default_config.rpcs.import_entries.retry_policy = {
|
115
|
-
initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
|
72
|
+
default_config.retry_policy = {
|
73
|
+
initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14, 8, 13]
|
116
74
|
}
|
117
75
|
|
118
76
|
default_config
|
@@ -1035,7 +993,7 @@ module Google
|
|
1035
993
|
# @param options [::Gapic::CallOptions, ::Hash]
|
1036
994
|
# Overrides the default settings for this call, e.g, timeout, retries etc. Optional.
|
1037
995
|
#
|
1038
|
-
# @overload lookup_entry(linked_resource: nil, sql_resource: nil, fully_qualified_name: nil)
|
996
|
+
# @overload lookup_entry(linked_resource: nil, sql_resource: nil, fully_qualified_name: nil, project: nil, location: nil)
|
1039
997
|
# Pass arguments to `lookup_entry` via keyword arguments. Note that at
|
1040
998
|
# least one keyword argument is required. To specify no parameters, or to keep all
|
1041
999
|
# the default parameter values, pass an empty Hash as a request object (see above).
|
@@ -1064,7 +1022,9 @@ module Google
|
|
1064
1022
|
# [Lexical structure in Standard SQL]
|
1065
1023
|
# (https://cloud.google.com/bigquery/docs/reference/standard-sql/lexical).
|
1066
1024
|
# @param fully_qualified_name [::String]
|
1067
|
-
# Fully
|
1025
|
+
# [Fully Qualified Name
|
1026
|
+
# (FQN)](https://cloud.google.com//data-catalog/docs/fully-qualified-names)
|
1027
|
+
# of the resource.
|
1068
1028
|
#
|
1069
1029
|
# FQNs take two forms:
|
1070
1030
|
#
|
@@ -1079,6 +1039,14 @@ module Google
|
|
1079
1039
|
# Example for a DPMS table:
|
1080
1040
|
#
|
1081
1041
|
# `dataproc_metastore:{PROJECT_ID}.{LOCATION_ID}.{INSTANCE_ID}.{DATABASE_ID}.{TABLE_ID}`
|
1042
|
+
# @param project [::String]
|
1043
|
+
# Project where the lookup should be performed. Required to lookup
|
1044
|
+
# entry that is not a part of `DPMS` or `DATAPLEX` `integrated_system`
|
1045
|
+
# using its `fully_qualified_name`. Ignored in other cases.
|
1046
|
+
# @param location [::String]
|
1047
|
+
# Location where the lookup should be performed. Required to lookup
|
1048
|
+
# entry that is not a part of `DPMS` or `DATAPLEX` `integrated_system`
|
1049
|
+
# using its `fully_qualified_name`. Ignored in other cases.
|
1082
1050
|
# @yield [result, operation] Access the result along with the TransportOperation object
|
1083
1051
|
# @yieldparam result [::Google::Cloud::DataCatalog::V1::Entry]
|
1084
1052
|
# @yieldparam operation [::Gapic::Rest::TransportOperation]
|
@@ -2784,7 +2752,7 @@ module Google
|
|
2784
2752
|
# @param options [::Gapic::CallOptions, ::Hash]
|
2785
2753
|
# Overrides the default settings for this call, e.g, timeout, retries etc. Optional.
|
2786
2754
|
#
|
2787
|
-
# @overload import_entries(parent: nil, gcs_bucket_path: nil)
|
2755
|
+
# @overload import_entries(parent: nil, gcs_bucket_path: nil, job_id: nil)
|
2788
2756
|
# Pass arguments to `import_entries` via keyword arguments. Note that at
|
2789
2757
|
# least one keyword argument is required. To specify no parameters, or to keep all
|
2790
2758
|
# the default parameter values, pass an empty Hash as a request object (see above).
|
@@ -2793,6 +2761,9 @@ module Google
|
|
2793
2761
|
# Required. Target entry group for ingested entries.
|
2794
2762
|
# @param gcs_bucket_path [::String]
|
2795
2763
|
# Path to a Cloud Storage bucket that contains a dump ready for ingestion.
|
2764
|
+
# @param job_id [::String]
|
2765
|
+
# Optional. (Optional) Dataplex task job id, if specified will be used as
|
2766
|
+
# part of ImportEntries LRO ID
|
2796
2767
|
# @yield [result, operation] Access the result along with the TransportOperation object
|
2797
2768
|
# @yieldparam result [::Gapic::Operation]
|
2798
2769
|
# @yieldparam operation [::Gapic::Rest::TransportOperation]
|
@@ -1,3 +1,4 @@
|
|
1
|
+
# frozen_string_literal: true
|
1
2
|
# Generated by the protocol buffer compiler. DO NOT EDIT!
|
2
3
|
# source: google/cloud/datacatalog/v1/bigquery.proto
|
3
4
|
|
@@ -5,33 +6,31 @@ require 'google/protobuf'
|
|
5
6
|
|
6
7
|
require 'google/api/field_behavior_pb'
|
7
8
|
|
8
|
-
|
9
|
-
|
10
|
-
|
11
|
-
|
12
|
-
|
13
|
-
|
14
|
-
|
15
|
-
|
16
|
-
|
17
|
-
|
18
|
-
|
19
|
-
|
20
|
-
|
21
|
-
|
22
|
-
|
23
|
-
|
24
|
-
|
25
|
-
|
26
|
-
|
27
|
-
|
28
|
-
|
29
|
-
value :MYSQL, 2
|
30
|
-
end
|
31
|
-
add_message "google.cloud.datacatalog.v1.BigQueryRoutineSpec" do
|
32
|
-
repeated :imported_libraries, :string, 1
|
9
|
+
|
10
|
+
descriptor_data = "\n*google/cloud/datacatalog/v1/bigquery.proto\x12\x1bgoogle.cloud.datacatalog.v1\x1a\x1fgoogle/api/field_behavior.proto\"\xb4\x02\n\x16\x42igQueryConnectionSpec\x12[\n\x0f\x63onnection_type\x18\x01 \x01(\x0e\x32\x42.google.cloud.datacatalog.v1.BigQueryConnectionSpec.ConnectionType\x12P\n\tcloud_sql\x18\x02 \x01(\x0b\x32;.google.cloud.datacatalog.v1.CloudSqlBigQueryConnectionSpecH\x00\x12\x16\n\x0ehas_credential\x18\x03 \x01(\x08\"@\n\x0e\x43onnectionType\x12\x1f\n\x1b\x43ONNECTION_TYPE_UNSPECIFIED\x10\x00\x12\r\n\tCLOUD_SQL\x10\x01\x42\x11\n\x0f\x63onnection_spec\"\xe7\x01\n\x1e\x43loudSqlBigQueryConnectionSpec\x12\x13\n\x0binstance_id\x18\x01 \x01(\t\x12\x10\n\x08\x64\x61tabase\x18\x02 \x01(\t\x12V\n\x04type\x18\x03 \x01(\x0e\x32H.google.cloud.datacatalog.v1.CloudSqlBigQueryConnectionSpec.DatabaseType\"F\n\x0c\x44\x61tabaseType\x12\x1d\n\x19\x44\x41TABASE_TYPE_UNSPECIFIED\x10\x00\x12\x0c\n\x08POSTGRES\x10\x01\x12\t\n\x05MYSQL\x10\x02\"1\n\x13\x42igQueryRoutineSpec\x12\x1a\n\x12imported_libraries\x18\x01 \x03(\tB\xd5\x01\n\x1f\x63om.google.cloud.datacatalog.v1B\rBigQueryProtoP\x01ZAcloud.google.com/go/datacatalog/apiv1/datacatalogpb;datacatalogpb\xf8\x01\x01\xaa\x02\x1bGoogle.Cloud.DataCatalog.V1\xca\x02\x1bGoogle\\Cloud\\DataCatalog\\V1\xea\x02\x1eGoogle::Cloud::DataCatalog::V1b\x06proto3"
|
11
|
+
|
12
|
+
pool = Google::Protobuf::DescriptorPool.generated_pool
|
13
|
+
|
14
|
+
begin
|
15
|
+
pool.add_serialized_file(descriptor_data)
|
16
|
+
rescue TypeError => e
|
17
|
+
# Compatibility code: will be removed in the next major version.
|
18
|
+
require 'google/protobuf/descriptor_pb'
|
19
|
+
parsed = Google::Protobuf::FileDescriptorProto.decode(descriptor_data)
|
20
|
+
parsed.clear_dependency
|
21
|
+
serialized = parsed.class.encode(parsed)
|
22
|
+
file = pool.add_serialized_file(serialized)
|
23
|
+
warn "Warning: Protobuf detected an import path issue while loading generated file #{__FILE__}"
|
24
|
+
imports = [
|
25
|
+
]
|
26
|
+
imports.each do |type_name, expected_filename|
|
27
|
+
import_file = pool.lookup(type_name).file_descriptor
|
28
|
+
if import_file.name != expected_filename
|
29
|
+
warn "- #{file.name} imports #{expected_filename}, but that import was loaded as #{import_file.name}"
|
33
30
|
end
|
34
31
|
end
|
32
|
+
warn "Each proto file must use a consistent fully-qualified name."
|
33
|
+
warn "This will become an error in the next major version."
|
35
34
|
end
|
36
35
|
|
37
36
|
module Google
|
@@ -1,3 +1,4 @@
|
|
1
|
+
# frozen_string_literal: true
|
1
2
|
# Generated by the protocol buffer compiler. DO NOT EDIT!
|
2
3
|
# source: google/cloud/datacatalog/v1/common.proto
|
3
4
|
|
@@ -5,27 +6,32 @@ require 'google/protobuf'
|
|
5
6
|
|
6
7
|
require 'google/protobuf/timestamp_pb'
|
7
8
|
|
8
|
-
|
9
|
-
|
10
|
-
|
11
|
-
|
12
|
-
|
13
|
-
|
14
|
-
|
15
|
-
|
16
|
-
|
17
|
-
|
18
|
-
|
19
|
-
|
20
|
-
|
21
|
-
|
22
|
-
|
23
|
-
|
24
|
-
|
25
|
-
|
26
|
-
|
9
|
+
|
10
|
+
descriptor_data = "\n(google/cloud/datacatalog/v1/common.proto\x12\x1bgoogle.cloud.datacatalog.v1\x1a\x1fgoogle/protobuf/timestamp.proto\"Q\n\x0fPersonalDetails\x12\x0f\n\x07starred\x18\x01 \x01(\x08\x12-\n\tstar_time\x18\x02 \x01(\x0b\x32\x1a.google.protobuf.Timestamp*\xbd\x01\n\x10IntegratedSystem\x12!\n\x1dINTEGRATED_SYSTEM_UNSPECIFIED\x10\x00\x12\x0c\n\x08\x42IGQUERY\x10\x01\x12\x10\n\x0c\x43LOUD_PUBSUB\x10\x02\x12\x16\n\x12\x44\x41TAPROC_METASTORE\x10\x03\x12\x0c\n\x08\x44\x41TAPLEX\x10\x04\x12\x11\n\rCLOUD_SPANNER\x10\x06\x12\x12\n\x0e\x43LOUD_BIGTABLE\x10\x07\x12\r\n\tCLOUD_SQL\x10\x08\x12\n\n\x06LOOKER\x10\t*j\n\x0eManagingSystem\x12\x1f\n\x1bMANAGING_SYSTEM_UNSPECIFIED\x10\x00\x12\x1c\n\x18MANAGING_SYSTEM_DATAPLEX\x10\x01\x12\x19\n\x15MANAGING_SYSTEM_OTHER\x10\x02\x42\xc6\x01\n\x1f\x63om.google.cloud.datacatalog.v1P\x01ZAcloud.google.com/go/datacatalog/apiv1/datacatalogpb;datacatalogpb\xf8\x01\x01\xaa\x02\x1bGoogle.Cloud.DataCatalog.V1\xca\x02\x1bGoogle\\Cloud\\DataCatalog\\V1\xea\x02\x1eGoogle::Cloud::DataCatalog::V1b\x06proto3"
|
11
|
+
|
12
|
+
pool = Google::Protobuf::DescriptorPool.generated_pool
|
13
|
+
|
14
|
+
begin
|
15
|
+
pool.add_serialized_file(descriptor_data)
|
16
|
+
rescue TypeError => e
|
17
|
+
# Compatibility code: will be removed in the next major version.
|
18
|
+
require 'google/protobuf/descriptor_pb'
|
19
|
+
parsed = Google::Protobuf::FileDescriptorProto.decode(descriptor_data)
|
20
|
+
parsed.clear_dependency
|
21
|
+
serialized = parsed.class.encode(parsed)
|
22
|
+
file = pool.add_serialized_file(serialized)
|
23
|
+
warn "Warning: Protobuf detected an import path issue while loading generated file #{__FILE__}"
|
24
|
+
imports = [
|
25
|
+
["google.protobuf.Timestamp", "google/protobuf/timestamp.proto"],
|
26
|
+
]
|
27
|
+
imports.each do |type_name, expected_filename|
|
28
|
+
import_file = pool.lookup(type_name).file_descriptor
|
29
|
+
if import_file.name != expected_filename
|
30
|
+
warn "- #{file.name} imports #{expected_filename}, but that import was loaded as #{import_file.name}"
|
27
31
|
end
|
28
32
|
end
|
33
|
+
warn "Each proto file must use a consistent fully-qualified name."
|
34
|
+
warn "This will become an error in the next major version."
|
29
35
|
end
|
30
36
|
|
31
37
|
module Google
|
@@ -1,3 +1,4 @@
|
|
1
|
+
# frozen_string_literal: true
|
1
2
|
# Generated by the protocol buffer compiler. DO NOT EDIT!
|
2
3
|
# source: google/cloud/datacatalog/v1/data_source.proto
|
3
4
|
|
@@ -5,26 +6,31 @@ require 'google/protobuf'
|
|
5
6
|
|
6
7
|
require 'google/api/field_behavior_pb'
|
7
8
|
|
8
|
-
|
9
|
-
|
10
|
-
|
11
|
-
|
12
|
-
|
13
|
-
|
14
|
-
|
15
|
-
|
16
|
-
|
17
|
-
|
18
|
-
|
19
|
-
|
20
|
-
|
21
|
-
|
22
|
-
|
23
|
-
|
24
|
-
|
25
|
-
|
9
|
+
|
10
|
+
descriptor_data = "\n-google/cloud/datacatalog/v1/data_source.proto\x12\x1bgoogle.cloud.datacatalog.v1\x1a\x1fgoogle/api/field_behavior.proto\"\x9c\x02\n\nDataSource\x12@\n\x07service\x18\x01 \x01(\x0e\x32/.google.cloud.datacatalog.v1.DataSource.Service\x12\x10\n\x08resource\x18\x02 \x01(\t\x12\x19\n\x0csource_entry\x18\x03 \x01(\tB\x03\xe0\x41\x03\x12L\n\x12storage_properties\x18\x04 \x01(\x0b\x32..google.cloud.datacatalog.v1.StoragePropertiesH\x00\"C\n\x07Service\x12\x17\n\x13SERVICE_UNSPECIFIED\x10\x00\x12\x11\n\rCLOUD_STORAGE\x10\x01\x12\x0c\n\x08\x42IGQUERY\x10\x02\x42\x0c\n\nproperties\"<\n\x11StorageProperties\x12\x14\n\x0c\x66ile_pattern\x18\x01 \x03(\t\x12\x11\n\tfile_type\x18\x02 \x01(\tB\xd7\x01\n\x1f\x63om.google.cloud.datacatalog.v1B\x0f\x44\x61taSourceProtoP\x01ZAcloud.google.com/go/datacatalog/apiv1/datacatalogpb;datacatalogpb\xf8\x01\x01\xaa\x02\x1bGoogle.Cloud.DataCatalog.V1\xca\x02\x1bGoogle\\Cloud\\DataCatalog\\V1\xea\x02\x1eGoogle::Cloud::DataCatalog::V1b\x06proto3"
|
11
|
+
|
12
|
+
pool = Google::Protobuf::DescriptorPool.generated_pool
|
13
|
+
|
14
|
+
begin
|
15
|
+
pool.add_serialized_file(descriptor_data)
|
16
|
+
rescue TypeError => e
|
17
|
+
# Compatibility code: will be removed in the next major version.
|
18
|
+
require 'google/protobuf/descriptor_pb'
|
19
|
+
parsed = Google::Protobuf::FileDescriptorProto.decode(descriptor_data)
|
20
|
+
parsed.clear_dependency
|
21
|
+
serialized = parsed.class.encode(parsed)
|
22
|
+
file = pool.add_serialized_file(serialized)
|
23
|
+
warn "Warning: Protobuf detected an import path issue while loading generated file #{__FILE__}"
|
24
|
+
imports = [
|
25
|
+
]
|
26
|
+
imports.each do |type_name, expected_filename|
|
27
|
+
import_file = pool.lookup(type_name).file_descriptor
|
28
|
+
if import_file.name != expected_filename
|
29
|
+
warn "- #{file.name} imports #{expected_filename}, but that import was loaded as #{import_file.name}"
|
26
30
|
end
|
27
31
|
end
|
32
|
+
warn "Each proto file must use a consistent fully-qualified name."
|
33
|
+
warn "This will become an error in the next major version."
|
28
34
|
end
|
29
35
|
|
30
36
|
module Google
|