google-cloud-data_catalog-v1 0.15.0 → 0.16.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (28) hide show
  1. checksums.yaml +4 -4
  2. data/lib/google/cloud/data_catalog/v1/data_catalog/client.rb +18 -47
  3. data/lib/google/cloud/data_catalog/v1/data_catalog/rest/client.rb +18 -47
  4. data/lib/google/cloud/data_catalog/v1/version.rb +1 -1
  5. data/lib/google/cloud/datacatalog/v1/bigquery_pb.rb +24 -25
  6. data/lib/google/cloud/datacatalog/v1/common_pb.rb +25 -19
  7. data/lib/google/cloud/datacatalog/v1/data_source_pb.rb +24 -18
  8. data/lib/google/cloud/datacatalog/v1/datacatalog_pb.rb +41 -339
  9. data/lib/google/cloud/datacatalog/v1/dataplex_spec_pb.rb +25 -21
  10. data/lib/google/cloud/datacatalog/v1/dump_content_pb.rb +27 -13
  11. data/lib/google/cloud/datacatalog/v1/gcs_fileset_spec_pb.rb +25 -10
  12. data/lib/google/cloud/datacatalog/v1/physical_schema_pb.rb +24 -26
  13. data/lib/google/cloud/datacatalog/v1/policytagmanager_pb.rb +26 -72
  14. data/lib/google/cloud/datacatalog/v1/policytagmanagerserialization_pb.rb +25 -43
  15. data/lib/google/cloud/datacatalog/v1/schema_pb.rb +24 -36
  16. data/lib/google/cloud/datacatalog/v1/search_pb.rb +25 -21
  17. data/lib/google/cloud/datacatalog/v1/table_spec_pb.rb +24 -26
  18. data/lib/google/cloud/datacatalog/v1/tags_pb.rb +25 -59
  19. data/lib/google/cloud/datacatalog/v1/timestamps_pb.rb +25 -6
  20. data/lib/google/cloud/datacatalog/v1/usage_pb.rb +25 -16
  21. data/proto_docs/google/api/client.rb +58 -1
  22. data/proto_docs/google/cloud/datacatalog/v1/common.rb +6 -0
  23. data/proto_docs/google/cloud/datacatalog/v1/datacatalog.rb +84 -20
  24. data/proto_docs/google/cloud/datacatalog/v1/dump_content.rb +2 -2
  25. data/proto_docs/google/cloud/datacatalog/v1/policytagmanager.rb +1 -1
  26. data/proto_docs/google/protobuf/any.rb +7 -4
  27. data/proto_docs/google/protobuf/timestamp.rb +1 -3
  28. metadata +4 -4
checksums.yaml CHANGED
@@ -1,7 +1,7 @@
1
1
  ---
2
2
  SHA256:
3
- metadata.gz: 2d083c21d6aefccc6e2dc26d4f1adbc2776d3b8795bdc128e76bde149b408f6b
4
- data.tar.gz: 0a345f881b23995d3fddc3e8ce6cf090982ca65ffd7453fe37209b55f2810e0e
3
+ metadata.gz: 287a106028e030737475116a68767810728bbe6300dcd308072e53eef227495e
4
+ data.tar.gz: 3ff5fb58112bfb569cdef3307029a14e7f7d0f1de030cf03ccde349352d039a3
5
5
  SHA512:
6
- metadata.gz: e89e88507e4861c6f3c19b018a251bc0ed451a2fd4b6276609298a17682535f524c39eb6167ed63c8240e0c99fa0eaef031423a4cd0ea34927c0a88a460cc898
7
- data.tar.gz: bf2397e6fe505768b3c4b0f2f0ef12a5f951ee911e2e0c0cc50ccd3ca61eb330e1f68bc9725da5969c8c28815fa7f35b214fcb2b5701c316e4473b837e9e8696
6
+ metadata.gz: 95d4feec2aeb82509fb40dd46b2e1fcc5057f566fdfbd0a0c1038fb5bc7ac11abdb2f8c3cdd1c2c32edd0d10fc29ec9dbdf61efb21c60a5dada00f999fbcb379
7
+ data.tar.gz: e8c76153b2b5814d3f1243eff17e20462ab6166706d077499514315ee7dc07be0311ac9ddb30b928a41546db9313112bf96bf052bba70e95b0af213df887a843
@@ -67,50 +67,8 @@ module Google
67
67
  default_config = Client::Configuration.new parent_config
68
68
 
69
69
  default_config.timeout = 60.0
70
-
71
- default_config.rpcs.search_catalog.timeout = 60.0
72
- default_config.rpcs.search_catalog.retry_policy = {
73
- initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
74
- }
75
-
76
- default_config.rpcs.get_entry_group.timeout = 60.0
77
- default_config.rpcs.get_entry_group.retry_policy = {
78
- initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
79
- }
80
-
81
- default_config.rpcs.list_entry_groups.timeout = 60.0
82
- default_config.rpcs.list_entry_groups.retry_policy = {
83
- initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
84
- }
85
-
86
- default_config.rpcs.get_entry.timeout = 60.0
87
- default_config.rpcs.get_entry.retry_policy = {
88
- initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
89
- }
90
-
91
- default_config.rpcs.lookup_entry.timeout = 60.0
92
- default_config.rpcs.lookup_entry.retry_policy = {
93
- initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
94
- }
95
-
96
- default_config.rpcs.list_entries.timeout = 60.0
97
- default_config.rpcs.list_entries.retry_policy = {
98
- initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
99
- }
100
-
101
- default_config.rpcs.list_tags.timeout = 60.0
102
- default_config.rpcs.list_tags.retry_policy = {
103
- initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
104
- }
105
-
106
- default_config.rpcs.get_iam_policy.timeout = 60.0
107
- default_config.rpcs.get_iam_policy.retry_policy = {
108
- initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
109
- }
110
-
111
- default_config.rpcs.import_entries.timeout = 60.0
112
- default_config.rpcs.import_entries.retry_policy = {
113
- initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
70
+ default_config.retry_policy = {
71
+ initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14, 8, 13]
114
72
  }
115
73
 
116
74
  default_config
@@ -1275,7 +1233,7 @@ module Google
1275
1233
  # @param options [::Gapic::CallOptions, ::Hash]
1276
1234
  # Overrides the default settings for this call, e.g, timeout, retries, etc. Optional.
1277
1235
  #
1278
- # @overload lookup_entry(linked_resource: nil, sql_resource: nil, fully_qualified_name: nil)
1236
+ # @overload lookup_entry(linked_resource: nil, sql_resource: nil, fully_qualified_name: nil, project: nil, location: nil)
1279
1237
  # Pass arguments to `lookup_entry` via keyword arguments. Note that at
1280
1238
  # least one keyword argument is required. To specify no parameters, or to keep all
1281
1239
  # the default parameter values, pass an empty Hash as a request object (see above).
@@ -1304,7 +1262,9 @@ module Google
1304
1262
  # [Lexical structure in Standard SQL]
1305
1263
  # (https://cloud.google.com/bigquery/docs/reference/standard-sql/lexical).
1306
1264
  # @param fully_qualified_name [::String]
1307
- # Fully qualified name (FQN) of the resource.
1265
+ # [Fully Qualified Name
1266
+ # (FQN)](https://cloud.google.com//data-catalog/docs/fully-qualified-names)
1267
+ # of the resource.
1308
1268
  #
1309
1269
  # FQNs take two forms:
1310
1270
  #
@@ -1319,6 +1279,14 @@ module Google
1319
1279
  # Example for a DPMS table:
1320
1280
  #
1321
1281
  # `dataproc_metastore:{PROJECT_ID}.{LOCATION_ID}.{INSTANCE_ID}.{DATABASE_ID}.{TABLE_ID}`
1282
+ # @param project [::String]
1283
+ # Project where the lookup should be performed. Required to lookup
1284
+ # entry that is not a part of `DPMS` or `DATAPLEX` `integrated_system`
1285
+ # using its `fully_qualified_name`. Ignored in other cases.
1286
+ # @param location [::String]
1287
+ # Location where the lookup should be performed. Required to lookup
1288
+ # entry that is not a part of `DPMS` or `DATAPLEX` `integrated_system`
1289
+ # using its `fully_qualified_name`. Ignored in other cases.
1322
1290
  #
1323
1291
  # @yield [response, operation] Access the result along with the RPC operation
1324
1292
  # @yieldparam response [::Google::Cloud::DataCatalog::V1::Entry]
@@ -3560,7 +3528,7 @@ module Google
3560
3528
  # @param options [::Gapic::CallOptions, ::Hash]
3561
3529
  # Overrides the default settings for this call, e.g, timeout, retries, etc. Optional.
3562
3530
  #
3563
- # @overload import_entries(parent: nil, gcs_bucket_path: nil)
3531
+ # @overload import_entries(parent: nil, gcs_bucket_path: nil, job_id: nil)
3564
3532
  # Pass arguments to `import_entries` via keyword arguments. Note that at
3565
3533
  # least one keyword argument is required. To specify no parameters, or to keep all
3566
3534
  # the default parameter values, pass an empty Hash as a request object (see above).
@@ -3569,6 +3537,9 @@ module Google
3569
3537
  # Required. Target entry group for ingested entries.
3570
3538
  # @param gcs_bucket_path [::String]
3571
3539
  # Path to a Cloud Storage bucket that contains a dump ready for ingestion.
3540
+ # @param job_id [::String]
3541
+ # Optional. (Optional) Dataplex task job id, if specified will be used as
3542
+ # part of ImportEntries LRO ID
3572
3543
  #
3573
3544
  # @yield [response, operation] Access the result along with the RPC operation
3574
3545
  # @yieldparam response [::Gapic::Operation]
@@ -69,50 +69,8 @@ module Google
69
69
  default_config = Client::Configuration.new parent_config
70
70
 
71
71
  default_config.timeout = 60.0
72
-
73
- default_config.rpcs.search_catalog.timeout = 60.0
74
- default_config.rpcs.search_catalog.retry_policy = {
75
- initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
76
- }
77
-
78
- default_config.rpcs.get_entry_group.timeout = 60.0
79
- default_config.rpcs.get_entry_group.retry_policy = {
80
- initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
81
- }
82
-
83
- default_config.rpcs.list_entry_groups.timeout = 60.0
84
- default_config.rpcs.list_entry_groups.retry_policy = {
85
- initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
86
- }
87
-
88
- default_config.rpcs.get_entry.timeout = 60.0
89
- default_config.rpcs.get_entry.retry_policy = {
90
- initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
91
- }
92
-
93
- default_config.rpcs.lookup_entry.timeout = 60.0
94
- default_config.rpcs.lookup_entry.retry_policy = {
95
- initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
96
- }
97
-
98
- default_config.rpcs.list_entries.timeout = 60.0
99
- default_config.rpcs.list_entries.retry_policy = {
100
- initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
101
- }
102
-
103
- default_config.rpcs.list_tags.timeout = 60.0
104
- default_config.rpcs.list_tags.retry_policy = {
105
- initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
106
- }
107
-
108
- default_config.rpcs.get_iam_policy.timeout = 60.0
109
- default_config.rpcs.get_iam_policy.retry_policy = {
110
- initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
111
- }
112
-
113
- default_config.rpcs.import_entries.timeout = 60.0
114
- default_config.rpcs.import_entries.retry_policy = {
115
- initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14]
72
+ default_config.retry_policy = {
73
+ initial_delay: 0.1, max_delay: 60.0, multiplier: 1.3, retry_codes: [14, 8, 13]
116
74
  }
117
75
 
118
76
  default_config
@@ -1035,7 +993,7 @@ module Google
1035
993
  # @param options [::Gapic::CallOptions, ::Hash]
1036
994
  # Overrides the default settings for this call, e.g, timeout, retries etc. Optional.
1037
995
  #
1038
- # @overload lookup_entry(linked_resource: nil, sql_resource: nil, fully_qualified_name: nil)
996
+ # @overload lookup_entry(linked_resource: nil, sql_resource: nil, fully_qualified_name: nil, project: nil, location: nil)
1039
997
  # Pass arguments to `lookup_entry` via keyword arguments. Note that at
1040
998
  # least one keyword argument is required. To specify no parameters, or to keep all
1041
999
  # the default parameter values, pass an empty Hash as a request object (see above).
@@ -1064,7 +1022,9 @@ module Google
1064
1022
  # [Lexical structure in Standard SQL]
1065
1023
  # (https://cloud.google.com/bigquery/docs/reference/standard-sql/lexical).
1066
1024
  # @param fully_qualified_name [::String]
1067
- # Fully qualified name (FQN) of the resource.
1025
+ # [Fully Qualified Name
1026
+ # (FQN)](https://cloud.google.com//data-catalog/docs/fully-qualified-names)
1027
+ # of the resource.
1068
1028
  #
1069
1029
  # FQNs take two forms:
1070
1030
  #
@@ -1079,6 +1039,14 @@ module Google
1079
1039
  # Example for a DPMS table:
1080
1040
  #
1081
1041
  # `dataproc_metastore:{PROJECT_ID}.{LOCATION_ID}.{INSTANCE_ID}.{DATABASE_ID}.{TABLE_ID}`
1042
+ # @param project [::String]
1043
+ # Project where the lookup should be performed. Required to lookup
1044
+ # entry that is not a part of `DPMS` or `DATAPLEX` `integrated_system`
1045
+ # using its `fully_qualified_name`. Ignored in other cases.
1046
+ # @param location [::String]
1047
+ # Location where the lookup should be performed. Required to lookup
1048
+ # entry that is not a part of `DPMS` or `DATAPLEX` `integrated_system`
1049
+ # using its `fully_qualified_name`. Ignored in other cases.
1082
1050
  # @yield [result, operation] Access the result along with the TransportOperation object
1083
1051
  # @yieldparam result [::Google::Cloud::DataCatalog::V1::Entry]
1084
1052
  # @yieldparam operation [::Gapic::Rest::TransportOperation]
@@ -2784,7 +2752,7 @@ module Google
2784
2752
  # @param options [::Gapic::CallOptions, ::Hash]
2785
2753
  # Overrides the default settings for this call, e.g, timeout, retries etc. Optional.
2786
2754
  #
2787
- # @overload import_entries(parent: nil, gcs_bucket_path: nil)
2755
+ # @overload import_entries(parent: nil, gcs_bucket_path: nil, job_id: nil)
2788
2756
  # Pass arguments to `import_entries` via keyword arguments. Note that at
2789
2757
  # least one keyword argument is required. To specify no parameters, or to keep all
2790
2758
  # the default parameter values, pass an empty Hash as a request object (see above).
@@ -2793,6 +2761,9 @@ module Google
2793
2761
  # Required. Target entry group for ingested entries.
2794
2762
  # @param gcs_bucket_path [::String]
2795
2763
  # Path to a Cloud Storage bucket that contains a dump ready for ingestion.
2764
+ # @param job_id [::String]
2765
+ # Optional. (Optional) Dataplex task job id, if specified will be used as
2766
+ # part of ImportEntries LRO ID
2796
2767
  # @yield [result, operation] Access the result along with the TransportOperation object
2797
2768
  # @yieldparam result [::Gapic::Operation]
2798
2769
  # @yieldparam operation [::Gapic::Rest::TransportOperation]
@@ -21,7 +21,7 @@ module Google
21
21
  module Cloud
22
22
  module DataCatalog
23
23
  module V1
24
- VERSION = "0.15.0"
24
+ VERSION = "0.16.0"
25
25
  end
26
26
  end
27
27
  end
@@ -1,3 +1,4 @@
1
+ # frozen_string_literal: true
1
2
  # Generated by the protocol buffer compiler. DO NOT EDIT!
2
3
  # source: google/cloud/datacatalog/v1/bigquery.proto
3
4
 
@@ -5,33 +6,31 @@ require 'google/protobuf'
5
6
 
6
7
  require 'google/api/field_behavior_pb'
7
8
 
8
- Google::Protobuf::DescriptorPool.generated_pool.build do
9
- add_file("google/cloud/datacatalog/v1/bigquery.proto", :syntax => :proto3) do
10
- add_message "google.cloud.datacatalog.v1.BigQueryConnectionSpec" do
11
- optional :connection_type, :enum, 1, "google.cloud.datacatalog.v1.BigQueryConnectionSpec.ConnectionType"
12
- optional :has_credential, :bool, 3
13
- oneof :connection_spec do
14
- optional :cloud_sql, :message, 2, "google.cloud.datacatalog.v1.CloudSqlBigQueryConnectionSpec"
15
- end
16
- end
17
- add_enum "google.cloud.datacatalog.v1.BigQueryConnectionSpec.ConnectionType" do
18
- value :CONNECTION_TYPE_UNSPECIFIED, 0
19
- value :CLOUD_SQL, 1
20
- end
21
- add_message "google.cloud.datacatalog.v1.CloudSqlBigQueryConnectionSpec" do
22
- optional :instance_id, :string, 1
23
- optional :database, :string, 2
24
- optional :type, :enum, 3, "google.cloud.datacatalog.v1.CloudSqlBigQueryConnectionSpec.DatabaseType"
25
- end
26
- add_enum "google.cloud.datacatalog.v1.CloudSqlBigQueryConnectionSpec.DatabaseType" do
27
- value :DATABASE_TYPE_UNSPECIFIED, 0
28
- value :POSTGRES, 1
29
- value :MYSQL, 2
30
- end
31
- add_message "google.cloud.datacatalog.v1.BigQueryRoutineSpec" do
32
- repeated :imported_libraries, :string, 1
9
+
10
+ descriptor_data = "\n*google/cloud/datacatalog/v1/bigquery.proto\x12\x1bgoogle.cloud.datacatalog.v1\x1a\x1fgoogle/api/field_behavior.proto\"\xb4\x02\n\x16\x42igQueryConnectionSpec\x12[\n\x0f\x63onnection_type\x18\x01 \x01(\x0e\x32\x42.google.cloud.datacatalog.v1.BigQueryConnectionSpec.ConnectionType\x12P\n\tcloud_sql\x18\x02 \x01(\x0b\x32;.google.cloud.datacatalog.v1.CloudSqlBigQueryConnectionSpecH\x00\x12\x16\n\x0ehas_credential\x18\x03 \x01(\x08\"@\n\x0e\x43onnectionType\x12\x1f\n\x1b\x43ONNECTION_TYPE_UNSPECIFIED\x10\x00\x12\r\n\tCLOUD_SQL\x10\x01\x42\x11\n\x0f\x63onnection_spec\"\xe7\x01\n\x1e\x43loudSqlBigQueryConnectionSpec\x12\x13\n\x0binstance_id\x18\x01 \x01(\t\x12\x10\n\x08\x64\x61tabase\x18\x02 \x01(\t\x12V\n\x04type\x18\x03 \x01(\x0e\x32H.google.cloud.datacatalog.v1.CloudSqlBigQueryConnectionSpec.DatabaseType\"F\n\x0c\x44\x61tabaseType\x12\x1d\n\x19\x44\x41TABASE_TYPE_UNSPECIFIED\x10\x00\x12\x0c\n\x08POSTGRES\x10\x01\x12\t\n\x05MYSQL\x10\x02\"1\n\x13\x42igQueryRoutineSpec\x12\x1a\n\x12imported_libraries\x18\x01 \x03(\tB\xd5\x01\n\x1f\x63om.google.cloud.datacatalog.v1B\rBigQueryProtoP\x01ZAcloud.google.com/go/datacatalog/apiv1/datacatalogpb;datacatalogpb\xf8\x01\x01\xaa\x02\x1bGoogle.Cloud.DataCatalog.V1\xca\x02\x1bGoogle\\Cloud\\DataCatalog\\V1\xea\x02\x1eGoogle::Cloud::DataCatalog::V1b\x06proto3"
11
+
12
+ pool = Google::Protobuf::DescriptorPool.generated_pool
13
+
14
+ begin
15
+ pool.add_serialized_file(descriptor_data)
16
+ rescue TypeError => e
17
+ # Compatibility code: will be removed in the next major version.
18
+ require 'google/protobuf/descriptor_pb'
19
+ parsed = Google::Protobuf::FileDescriptorProto.decode(descriptor_data)
20
+ parsed.clear_dependency
21
+ serialized = parsed.class.encode(parsed)
22
+ file = pool.add_serialized_file(serialized)
23
+ warn "Warning: Protobuf detected an import path issue while loading generated file #{__FILE__}"
24
+ imports = [
25
+ ]
26
+ imports.each do |type_name, expected_filename|
27
+ import_file = pool.lookup(type_name).file_descriptor
28
+ if import_file.name != expected_filename
29
+ warn "- #{file.name} imports #{expected_filename}, but that import was loaded as #{import_file.name}"
33
30
  end
34
31
  end
32
+ warn "Each proto file must use a consistent fully-qualified name."
33
+ warn "This will become an error in the next major version."
35
34
  end
36
35
 
37
36
  module Google
@@ -1,3 +1,4 @@
1
+ # frozen_string_literal: true
1
2
  # Generated by the protocol buffer compiler. DO NOT EDIT!
2
3
  # source: google/cloud/datacatalog/v1/common.proto
3
4
 
@@ -5,27 +6,32 @@ require 'google/protobuf'
5
6
 
6
7
  require 'google/protobuf/timestamp_pb'
7
8
 
8
- Google::Protobuf::DescriptorPool.generated_pool.build do
9
- add_file("google/cloud/datacatalog/v1/common.proto", :syntax => :proto3) do
10
- add_message "google.cloud.datacatalog.v1.PersonalDetails" do
11
- optional :starred, :bool, 1
12
- optional :star_time, :message, 2, "google.protobuf.Timestamp"
13
- end
14
- add_enum "google.cloud.datacatalog.v1.IntegratedSystem" do
15
- value :INTEGRATED_SYSTEM_UNSPECIFIED, 0
16
- value :BIGQUERY, 1
17
- value :CLOUD_PUBSUB, 2
18
- value :DATAPROC_METASTORE, 3
19
- value :DATAPLEX, 4
20
- value :CLOUD_SQL, 8
21
- value :LOOKER, 9
22
- end
23
- add_enum "google.cloud.datacatalog.v1.ManagingSystem" do
24
- value :MANAGING_SYSTEM_UNSPECIFIED, 0
25
- value :MANAGING_SYSTEM_DATAPLEX, 1
26
- value :MANAGING_SYSTEM_OTHER, 2
9
+
10
+ descriptor_data = "\n(google/cloud/datacatalog/v1/common.proto\x12\x1bgoogle.cloud.datacatalog.v1\x1a\x1fgoogle/protobuf/timestamp.proto\"Q\n\x0fPersonalDetails\x12\x0f\n\x07starred\x18\x01 \x01(\x08\x12-\n\tstar_time\x18\x02 \x01(\x0b\x32\x1a.google.protobuf.Timestamp*\xbd\x01\n\x10IntegratedSystem\x12!\n\x1dINTEGRATED_SYSTEM_UNSPECIFIED\x10\x00\x12\x0c\n\x08\x42IGQUERY\x10\x01\x12\x10\n\x0c\x43LOUD_PUBSUB\x10\x02\x12\x16\n\x12\x44\x41TAPROC_METASTORE\x10\x03\x12\x0c\n\x08\x44\x41TAPLEX\x10\x04\x12\x11\n\rCLOUD_SPANNER\x10\x06\x12\x12\n\x0e\x43LOUD_BIGTABLE\x10\x07\x12\r\n\tCLOUD_SQL\x10\x08\x12\n\n\x06LOOKER\x10\t*j\n\x0eManagingSystem\x12\x1f\n\x1bMANAGING_SYSTEM_UNSPECIFIED\x10\x00\x12\x1c\n\x18MANAGING_SYSTEM_DATAPLEX\x10\x01\x12\x19\n\x15MANAGING_SYSTEM_OTHER\x10\x02\x42\xc6\x01\n\x1f\x63om.google.cloud.datacatalog.v1P\x01ZAcloud.google.com/go/datacatalog/apiv1/datacatalogpb;datacatalogpb\xf8\x01\x01\xaa\x02\x1bGoogle.Cloud.DataCatalog.V1\xca\x02\x1bGoogle\\Cloud\\DataCatalog\\V1\xea\x02\x1eGoogle::Cloud::DataCatalog::V1b\x06proto3"
11
+
12
+ pool = Google::Protobuf::DescriptorPool.generated_pool
13
+
14
+ begin
15
+ pool.add_serialized_file(descriptor_data)
16
+ rescue TypeError => e
17
+ # Compatibility code: will be removed in the next major version.
18
+ require 'google/protobuf/descriptor_pb'
19
+ parsed = Google::Protobuf::FileDescriptorProto.decode(descriptor_data)
20
+ parsed.clear_dependency
21
+ serialized = parsed.class.encode(parsed)
22
+ file = pool.add_serialized_file(serialized)
23
+ warn "Warning: Protobuf detected an import path issue while loading generated file #{__FILE__}"
24
+ imports = [
25
+ ["google.protobuf.Timestamp", "google/protobuf/timestamp.proto"],
26
+ ]
27
+ imports.each do |type_name, expected_filename|
28
+ import_file = pool.lookup(type_name).file_descriptor
29
+ if import_file.name != expected_filename
30
+ warn "- #{file.name} imports #{expected_filename}, but that import was loaded as #{import_file.name}"
27
31
  end
28
32
  end
33
+ warn "Each proto file must use a consistent fully-qualified name."
34
+ warn "This will become an error in the next major version."
29
35
  end
30
36
 
31
37
  module Google
@@ -1,3 +1,4 @@
1
+ # frozen_string_literal: true
1
2
  # Generated by the protocol buffer compiler. DO NOT EDIT!
2
3
  # source: google/cloud/datacatalog/v1/data_source.proto
3
4
 
@@ -5,26 +6,31 @@ require 'google/protobuf'
5
6
 
6
7
  require 'google/api/field_behavior_pb'
7
8
 
8
- Google::Protobuf::DescriptorPool.generated_pool.build do
9
- add_file("google/cloud/datacatalog/v1/data_source.proto", :syntax => :proto3) do
10
- add_message "google.cloud.datacatalog.v1.DataSource" do
11
- optional :service, :enum, 1, "google.cloud.datacatalog.v1.DataSource.Service"
12
- optional :resource, :string, 2
13
- optional :source_entry, :string, 3
14
- oneof :properties do
15
- optional :storage_properties, :message, 4, "google.cloud.datacatalog.v1.StorageProperties"
16
- end
17
- end
18
- add_enum "google.cloud.datacatalog.v1.DataSource.Service" do
19
- value :SERVICE_UNSPECIFIED, 0
20
- value :CLOUD_STORAGE, 1
21
- value :BIGQUERY, 2
22
- end
23
- add_message "google.cloud.datacatalog.v1.StorageProperties" do
24
- repeated :file_pattern, :string, 1
25
- optional :file_type, :string, 2
9
+
10
+ descriptor_data = "\n-google/cloud/datacatalog/v1/data_source.proto\x12\x1bgoogle.cloud.datacatalog.v1\x1a\x1fgoogle/api/field_behavior.proto\"\x9c\x02\n\nDataSource\x12@\n\x07service\x18\x01 \x01(\x0e\x32/.google.cloud.datacatalog.v1.DataSource.Service\x12\x10\n\x08resource\x18\x02 \x01(\t\x12\x19\n\x0csource_entry\x18\x03 \x01(\tB\x03\xe0\x41\x03\x12L\n\x12storage_properties\x18\x04 \x01(\x0b\x32..google.cloud.datacatalog.v1.StoragePropertiesH\x00\"C\n\x07Service\x12\x17\n\x13SERVICE_UNSPECIFIED\x10\x00\x12\x11\n\rCLOUD_STORAGE\x10\x01\x12\x0c\n\x08\x42IGQUERY\x10\x02\x42\x0c\n\nproperties\"<\n\x11StorageProperties\x12\x14\n\x0c\x66ile_pattern\x18\x01 \x03(\t\x12\x11\n\tfile_type\x18\x02 \x01(\tB\xd7\x01\n\x1f\x63om.google.cloud.datacatalog.v1B\x0f\x44\x61taSourceProtoP\x01ZAcloud.google.com/go/datacatalog/apiv1/datacatalogpb;datacatalogpb\xf8\x01\x01\xaa\x02\x1bGoogle.Cloud.DataCatalog.V1\xca\x02\x1bGoogle\\Cloud\\DataCatalog\\V1\xea\x02\x1eGoogle::Cloud::DataCatalog::V1b\x06proto3"
11
+
12
+ pool = Google::Protobuf::DescriptorPool.generated_pool
13
+
14
+ begin
15
+ pool.add_serialized_file(descriptor_data)
16
+ rescue TypeError => e
17
+ # Compatibility code: will be removed in the next major version.
18
+ require 'google/protobuf/descriptor_pb'
19
+ parsed = Google::Protobuf::FileDescriptorProto.decode(descriptor_data)
20
+ parsed.clear_dependency
21
+ serialized = parsed.class.encode(parsed)
22
+ file = pool.add_serialized_file(serialized)
23
+ warn "Warning: Protobuf detected an import path issue while loading generated file #{__FILE__}"
24
+ imports = [
25
+ ]
26
+ imports.each do |type_name, expected_filename|
27
+ import_file = pool.lookup(type_name).file_descriptor
28
+ if import_file.name != expected_filename
29
+ warn "- #{file.name} imports #{expected_filename}, but that import was loaded as #{import_file.name}"
26
30
  end
27
31
  end
32
+ warn "Each proto file must use a consistent fully-qualified name."
33
+ warn "This will become an error in the next major version."
28
34
  end
29
35
 
30
36
  module Google