aws-sdk-gluedatabrew 1.19.0 → 1.22.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- checksums.yaml +4 -4
- data/CHANGELOG.md +15 -0
- data/VERSION +1 -1
- data/lib/aws-sdk-gluedatabrew/client.rb +15 -5
- data/lib/aws-sdk-gluedatabrew/client_api.rb +2 -0
- data/lib/aws-sdk-gluedatabrew/types.rb +16 -6
- data/lib/aws-sdk-gluedatabrew.rb +1 -1
- metadata +4 -4
checksums.yaml
CHANGED
@@ -1,7 +1,7 @@
|
|
1
1
|
---
|
2
2
|
SHA256:
|
3
|
-
metadata.gz:
|
4
|
-
data.tar.gz:
|
3
|
+
metadata.gz: 80083adbd88cd99edbc125b4741389c96889f1aebea17667fecbe9eb28f2f248
|
4
|
+
data.tar.gz: f966bf41df2aedd1f9a62242fb5760f3c2fb5a91efc0500f5f698fa2b41c8188
|
5
5
|
SHA512:
|
6
|
-
metadata.gz:
|
7
|
-
data.tar.gz:
|
6
|
+
metadata.gz: 828fdc95208d3cb220bee5ddfe95ac9f99377dbf5f2647c9c5660c0b3db1bff27b3c09b379c0197ba1b161175c624aaa3e87b0db2ef12824ace1ef743bc0203c
|
7
|
+
data.tar.gz: 4376e1c2dc336a3627f8d875bf424b343d3df801cea7d0ef085f2634c2d6370c1dbbbbf4dfd731feaa107fc4f576a41261e908cb3368cb04b8754775a4cf3283
|
data/CHANGELOG.md
CHANGED
@@ -1,6 +1,21 @@
|
|
1
1
|
Unreleased Changes
|
2
2
|
------------------
|
3
3
|
|
4
|
+
1.22.0 (2022-03-31)
|
5
|
+
------------------
|
6
|
+
|
7
|
+
* Feature - This AWS Glue Databrew release adds feature to support ORC as an input format.
|
8
|
+
|
9
|
+
1.21.0 (2022-02-24)
|
10
|
+
------------------
|
11
|
+
|
12
|
+
* Feature - This AWS Glue Databrew release adds feature to merge job outputs into a max number of files for S3 File output type.
|
13
|
+
|
14
|
+
1.20.0 (2022-02-03)
|
15
|
+
------------------
|
16
|
+
|
17
|
+
* Feature - Code Generated Changes, see `./build_tools` or `aws-sdk-core`'s CHANGELOG.md for details.
|
18
|
+
|
4
19
|
1.19.0 (2022-01-10)
|
5
20
|
------------------
|
6
21
|
|
data/VERSION
CHANGED
@@ -1 +1 @@
|
|
1
|
-
1.
|
1
|
+
1.22.0
|
@@ -27,7 +27,9 @@ require 'aws-sdk-core/plugins/client_metrics_plugin.rb'
|
|
27
27
|
require 'aws-sdk-core/plugins/client_metrics_send_plugin.rb'
|
28
28
|
require 'aws-sdk-core/plugins/transfer_encoding.rb'
|
29
29
|
require 'aws-sdk-core/plugins/http_checksum.rb'
|
30
|
+
require 'aws-sdk-core/plugins/checksum_algorithm.rb'
|
30
31
|
require 'aws-sdk-core/plugins/defaults_mode.rb'
|
32
|
+
require 'aws-sdk-core/plugins/recursion_detection.rb'
|
31
33
|
require 'aws-sdk-core/plugins/signature_v4.rb'
|
32
34
|
require 'aws-sdk-core/plugins/protocols/rest_json.rb'
|
33
35
|
|
@@ -74,7 +76,9 @@ module Aws::GlueDataBrew
|
|
74
76
|
add_plugin(Aws::Plugins::ClientMetricsSendPlugin)
|
75
77
|
add_plugin(Aws::Plugins::TransferEncoding)
|
76
78
|
add_plugin(Aws::Plugins::HttpChecksum)
|
79
|
+
add_plugin(Aws::Plugins::ChecksumAlgorithm)
|
77
80
|
add_plugin(Aws::Plugins::DefaultsMode)
|
81
|
+
add_plugin(Aws::Plugins::RecursionDetection)
|
78
82
|
add_plugin(Aws::Plugins::SignatureV4)
|
79
83
|
add_plugin(Aws::Plugins::Protocols::RestJson)
|
80
84
|
|
@@ -446,7 +450,7 @@ module Aws::GlueDataBrew
|
|
446
450
|
#
|
447
451
|
# resp = client.create_dataset({
|
448
452
|
# name: "DatasetName", # required
|
449
|
-
# format: "CSV", # accepts CSV, JSON, PARQUET, EXCEL
|
453
|
+
# format: "CSV", # accepts CSV, JSON, PARQUET, EXCEL, ORC
|
450
454
|
# format_options: {
|
451
455
|
# json: {
|
452
456
|
# multi_line: false,
|
@@ -907,6 +911,7 @@ module Aws::GlueDataBrew
|
|
907
911
|
# delimiter: "Delimiter",
|
908
912
|
# },
|
909
913
|
# },
|
914
|
+
# max_output_files: 1,
|
910
915
|
# },
|
911
916
|
# ],
|
912
917
|
# data_catalog_outputs: [
|
@@ -1298,7 +1303,7 @@ module Aws::GlueDataBrew
|
|
1298
1303
|
# resp.created_by #=> String
|
1299
1304
|
# resp.create_date #=> Time
|
1300
1305
|
# resp.name #=> String
|
1301
|
-
# resp.format #=> String, one of "CSV", "JSON", "PARQUET", "EXCEL"
|
1306
|
+
# resp.format #=> String, one of "CSV", "JSON", "PARQUET", "EXCEL", "ORC"
|
1302
1307
|
# resp.format_options.json.multi_line #=> Boolean
|
1303
1308
|
# resp.format_options.excel.sheet_names #=> Array
|
1304
1309
|
# resp.format_options.excel.sheet_names[0] #=> String
|
@@ -1417,6 +1422,7 @@ module Aws::GlueDataBrew
|
|
1417
1422
|
# resp.outputs[0].location.bucket_owner #=> String
|
1418
1423
|
# resp.outputs[0].overwrite #=> Boolean
|
1419
1424
|
# resp.outputs[0].format_options.csv.delimiter #=> String
|
1425
|
+
# resp.outputs[0].max_output_files #=> Integer
|
1420
1426
|
# resp.data_catalog_outputs #=> Array
|
1421
1427
|
# resp.data_catalog_outputs[0].catalog_id #=> String
|
1422
1428
|
# resp.data_catalog_outputs[0].database_name #=> String
|
@@ -1569,6 +1575,7 @@ module Aws::GlueDataBrew
|
|
1569
1575
|
# resp.outputs[0].location.bucket_owner #=> String
|
1570
1576
|
# resp.outputs[0].overwrite #=> Boolean
|
1571
1577
|
# resp.outputs[0].format_options.csv.delimiter #=> String
|
1578
|
+
# resp.outputs[0].max_output_files #=> Integer
|
1572
1579
|
# resp.data_catalog_outputs #=> Array
|
1573
1580
|
# resp.data_catalog_outputs[0].catalog_id #=> String
|
1574
1581
|
# resp.data_catalog_outputs[0].database_name #=> String
|
@@ -1860,7 +1867,7 @@ module Aws::GlueDataBrew
|
|
1860
1867
|
# resp.datasets[0].created_by #=> String
|
1861
1868
|
# resp.datasets[0].create_date #=> Time
|
1862
1869
|
# resp.datasets[0].name #=> String
|
1863
|
-
# resp.datasets[0].format #=> String, one of "CSV", "JSON", "PARQUET", "EXCEL"
|
1870
|
+
# resp.datasets[0].format #=> String, one of "CSV", "JSON", "PARQUET", "EXCEL", "ORC"
|
1864
1871
|
# resp.datasets[0].format_options.json.multi_line #=> Boolean
|
1865
1872
|
# resp.datasets[0].format_options.excel.sheet_names #=> Array
|
1866
1873
|
# resp.datasets[0].format_options.excel.sheet_names[0] #=> String
|
@@ -1968,6 +1975,7 @@ module Aws::GlueDataBrew
|
|
1968
1975
|
# resp.job_runs[0].outputs[0].location.bucket_owner #=> String
|
1969
1976
|
# resp.job_runs[0].outputs[0].overwrite #=> Boolean
|
1970
1977
|
# resp.job_runs[0].outputs[0].format_options.csv.delimiter #=> String
|
1978
|
+
# resp.job_runs[0].outputs[0].max_output_files #=> Integer
|
1971
1979
|
# resp.job_runs[0].data_catalog_outputs #=> Array
|
1972
1980
|
# resp.job_runs[0].data_catalog_outputs[0].catalog_id #=> String
|
1973
1981
|
# resp.job_runs[0].data_catalog_outputs[0].database_name #=> String
|
@@ -2068,6 +2076,7 @@ module Aws::GlueDataBrew
|
|
2068
2076
|
# resp.jobs[0].outputs[0].location.bucket_owner #=> String
|
2069
2077
|
# resp.jobs[0].outputs[0].overwrite #=> Boolean
|
2070
2078
|
# resp.jobs[0].outputs[0].format_options.csv.delimiter #=> String
|
2079
|
+
# resp.jobs[0].outputs[0].max_output_files #=> Integer
|
2071
2080
|
# resp.jobs[0].data_catalog_outputs #=> Array
|
2072
2081
|
# resp.jobs[0].data_catalog_outputs[0].catalog_id #=> String
|
2073
2082
|
# resp.jobs[0].data_catalog_outputs[0].database_name #=> String
|
@@ -2728,7 +2737,7 @@ module Aws::GlueDataBrew
|
|
2728
2737
|
#
|
2729
2738
|
# resp = client.update_dataset({
|
2730
2739
|
# name: "DatasetName", # required
|
2731
|
-
# format: "CSV", # accepts CSV, JSON, PARQUET, EXCEL
|
2740
|
+
# format: "CSV", # accepts CSV, JSON, PARQUET, EXCEL, ORC
|
2732
2741
|
# format_options: {
|
2733
2742
|
# json: {
|
2734
2743
|
# multi_line: false,
|
@@ -3139,6 +3148,7 @@ module Aws::GlueDataBrew
|
|
3139
3148
|
# delimiter: "Delimiter",
|
3140
3149
|
# },
|
3141
3150
|
# },
|
3151
|
+
# max_output_files: 1,
|
3142
3152
|
# },
|
3143
3153
|
# ],
|
3144
3154
|
# data_catalog_outputs: [
|
@@ -3307,7 +3317,7 @@ module Aws::GlueDataBrew
|
|
3307
3317
|
params: params,
|
3308
3318
|
config: config)
|
3309
3319
|
context[:gem_name] = 'aws-sdk-gluedatabrew'
|
3310
|
-
context[:gem_version] = '1.
|
3320
|
+
context[:gem_version] = '1.22.0'
|
3311
3321
|
Seahorse::Client::Request.new(handlers, context)
|
3312
3322
|
end
|
3313
3323
|
|
@@ -161,6 +161,7 @@ module Aws::GlueDataBrew
|
|
161
161
|
LogSubscription = Shapes::StringShape.new(name: 'LogSubscription')
|
162
162
|
MaxCapacity = Shapes::IntegerShape.new(name: 'MaxCapacity')
|
163
163
|
MaxFiles = Shapes::IntegerShape.new(name: 'MaxFiles')
|
164
|
+
MaxOutputFiles = Shapes::IntegerShape.new(name: 'MaxOutputFiles')
|
164
165
|
MaxResults100 = Shapes::IntegerShape.new(name: 'MaxResults100')
|
165
166
|
MaxRetries = Shapes::IntegerShape.new(name: 'MaxRetries')
|
166
167
|
Message = Shapes::StringShape.new(name: 'Message')
|
@@ -842,6 +843,7 @@ module Aws::GlueDataBrew
|
|
842
843
|
Output.add_member(:location, Shapes::ShapeRef.new(shape: S3Location, required: true, location_name: "Location"))
|
843
844
|
Output.add_member(:overwrite, Shapes::ShapeRef.new(shape: OverwriteOutput, location_name: "Overwrite"))
|
844
845
|
Output.add_member(:format_options, Shapes::ShapeRef.new(shape: OutputFormatOptions, location_name: "FormatOptions"))
|
846
|
+
Output.add_member(:max_output_files, Shapes::ShapeRef.new(shape: MaxOutputFiles, location_name: "MaxOutputFiles"))
|
845
847
|
Output.struct_class = Types::Output
|
846
848
|
|
847
849
|
OutputFormatOptions.add_member(:csv, Shapes::ShapeRef.new(shape: CsvOutputOptions, location_name: "Csv"))
|
@@ -233,7 +233,7 @@ module Aws::GlueDataBrew
|
|
233
233
|
#
|
234
234
|
# {
|
235
235
|
# name: "DatasetName", # required
|
236
|
-
# format: "CSV", # accepts CSV, JSON, PARQUET, EXCEL
|
236
|
+
# format: "CSV", # accepts CSV, JSON, PARQUET, EXCEL, ORC
|
237
237
|
# format_options: {
|
238
238
|
# json: {
|
239
239
|
# multi_line: false,
|
@@ -656,6 +656,7 @@ module Aws::GlueDataBrew
|
|
656
656
|
# delimiter: "Delimiter",
|
657
657
|
# },
|
658
658
|
# },
|
659
|
+
# max_output_files: 1,
|
659
660
|
# },
|
660
661
|
# ],
|
661
662
|
# data_catalog_outputs: [
|
@@ -1401,7 +1402,7 @@ module Aws::GlueDataBrew
|
|
1401
1402
|
include Aws::Structure
|
1402
1403
|
end
|
1403
1404
|
|
1404
|
-
# Represents a dataset
|
1405
|
+
# Represents a dataset parameter that defines type and conditions for a
|
1405
1406
|
# parameter in the Amazon S3 path of the dataset.
|
1406
1407
|
#
|
1407
1408
|
# @note When making an API call, you may pass DatasetParameter
|
@@ -2606,7 +2607,7 @@ module Aws::GlueDataBrew
|
|
2606
2607
|
# @!attribute [rw] order
|
2607
2608
|
# A criteria to use for Amazon S3 files sorting before their
|
2608
2609
|
# selection. By default uses DESCENDING order, i.e. most recent files
|
2609
|
-
# are selected first.
|
2610
|
+
# are selected first. Another possible value is ASCENDING.
|
2610
2611
|
# @return [String]
|
2611
2612
|
#
|
2612
2613
|
# @see http://docs.aws.amazon.com/goto/WebAPI/databrew-2017-07-25/FilesLimit AWS API Documentation
|
@@ -3593,6 +3594,7 @@ module Aws::GlueDataBrew
|
|
3593
3594
|
# delimiter: "Delimiter",
|
3594
3595
|
# },
|
3595
3596
|
# },
|
3597
|
+
# max_output_files: 1,
|
3596
3598
|
# }
|
3597
3599
|
#
|
3598
3600
|
# @!attribute [rw] compression_format
|
@@ -3623,6 +3625,12 @@ module Aws::GlueDataBrew
|
|
3623
3625
|
# files.
|
3624
3626
|
# @return [Types::OutputFormatOptions]
|
3625
3627
|
#
|
3628
|
+
# @!attribute [rw] max_output_files
|
3629
|
+
# Maximum number of files to be generated by the job and written to
|
3630
|
+
# the output folder. For output partitioned by column(s), the
|
3631
|
+
# MaxOutputFiles value is the maximum number of files per partition.
|
3632
|
+
# @return [Integer]
|
3633
|
+
#
|
3626
3634
|
# @see http://docs.aws.amazon.com/goto/WebAPI/databrew-2017-07-25/Output AWS API Documentation
|
3627
3635
|
#
|
3628
3636
|
class Output < Struct.new(
|
@@ -3631,7 +3639,8 @@ module Aws::GlueDataBrew
|
|
3631
3639
|
:partition_columns,
|
3632
3640
|
:location,
|
3633
3641
|
:overwrite,
|
3634
|
-
:format_options
|
3642
|
+
:format_options,
|
3643
|
+
:max_output_files)
|
3635
3644
|
SENSITIVE = []
|
3636
3645
|
include Aws::Structure
|
3637
3646
|
end
|
@@ -4219,7 +4228,7 @@ module Aws::GlueDataBrew
|
|
4219
4228
|
# value or a column name. These values are defined in the
|
4220
4229
|
# SubstitutionMap. If a CheckExpression starts with a column
|
4221
4230
|
# reference, then ColumnSelectors in the rule should be null. If
|
4222
|
-
# ColumnSelectors has been defined, then there should be no
|
4231
|
+
# ColumnSelectors has been defined, then there should be no column
|
4223
4232
|
# reference in the left side of a condition, for example, `is_between
|
4224
4233
|
# :val1 and :val2`.
|
4225
4234
|
#
|
@@ -4878,7 +4887,7 @@ module Aws::GlueDataBrew
|
|
4878
4887
|
#
|
4879
4888
|
# {
|
4880
4889
|
# name: "DatasetName", # required
|
4881
|
-
# format: "CSV", # accepts CSV, JSON, PARQUET, EXCEL
|
4890
|
+
# format: "CSV", # accepts CSV, JSON, PARQUET, EXCEL, ORC
|
4882
4891
|
# format_options: {
|
4883
4892
|
# json: {
|
4884
4893
|
# multi_line: false,
|
@@ -5259,6 +5268,7 @@ module Aws::GlueDataBrew
|
|
5259
5268
|
# delimiter: "Delimiter",
|
5260
5269
|
# },
|
5261
5270
|
# },
|
5271
|
+
# max_output_files: 1,
|
5262
5272
|
# },
|
5263
5273
|
# ],
|
5264
5274
|
# data_catalog_outputs: [
|
data/lib/aws-sdk-gluedatabrew.rb
CHANGED
metadata
CHANGED
@@ -1,14 +1,14 @@
|
|
1
1
|
--- !ruby/object:Gem::Specification
|
2
2
|
name: aws-sdk-gluedatabrew
|
3
3
|
version: !ruby/object:Gem::Version
|
4
|
-
version: 1.
|
4
|
+
version: 1.22.0
|
5
5
|
platform: ruby
|
6
6
|
authors:
|
7
7
|
- Amazon Web Services
|
8
8
|
autorequire:
|
9
9
|
bindir: bin
|
10
10
|
cert_chain: []
|
11
|
-
date: 2022-
|
11
|
+
date: 2022-03-31 00:00:00.000000000 Z
|
12
12
|
dependencies:
|
13
13
|
- !ruby/object:Gem::Dependency
|
14
14
|
name: aws-sdk-core
|
@@ -19,7 +19,7 @@ dependencies:
|
|
19
19
|
version: '3'
|
20
20
|
- - ">="
|
21
21
|
- !ruby/object:Gem::Version
|
22
|
-
version: 3.
|
22
|
+
version: 3.127.0
|
23
23
|
type: :runtime
|
24
24
|
prerelease: false
|
25
25
|
version_requirements: !ruby/object:Gem::Requirement
|
@@ -29,7 +29,7 @@ dependencies:
|
|
29
29
|
version: '3'
|
30
30
|
- - ">="
|
31
31
|
- !ruby/object:Gem::Version
|
32
|
-
version: 3.
|
32
|
+
version: 3.127.0
|
33
33
|
- !ruby/object:Gem::Dependency
|
34
34
|
name: aws-sigv4
|
35
35
|
requirement: !ruby/object:Gem::Requirement
|