aws-sdk-datasync 1.37.0 → 1.41.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -275,6 +275,67 @@ module Aws::DataSync
275
275
  include Aws::Structure
276
276
  end
277
277
 
278
+ # @note When making an API call, you may pass CreateLocationFsxLustreRequest
279
+ # data as a hash:
280
+ #
281
+ # {
282
+ # fsx_filesystem_arn: "FsxFilesystemArn", # required
283
+ # security_group_arns: ["Ec2SecurityGroupArn"], # required
284
+ # subdirectory: "FsxLustreSubdirectory",
285
+ # tags: [
286
+ # {
287
+ # key: "TagKey", # required
288
+ # value: "TagValue",
289
+ # },
290
+ # ],
291
+ # }
292
+ #
293
+ # @!attribute [rw] fsx_filesystem_arn
294
+ # The Amazon Resource Name (ARN) for the FSx for Lustre file system.
295
+ # @return [String]
296
+ #
297
+ # @!attribute [rw] security_group_arns
298
+ # The Amazon Resource Names (ARNs) of the security groups that are
299
+ # used to configure the FSx for Lustre file system.
300
+ # @return [Array<String>]
301
+ #
302
+ # @!attribute [rw] subdirectory
303
+ # A subdirectory in the location's path. This subdirectory in the FSx
304
+ # for Lustre file system is used to read data from the FSx for Lustre
305
+ # source location or write data to the FSx for Lustre destination.
306
+ # @return [String]
307
+ #
308
+ # @!attribute [rw] tags
309
+ # The key-value pair that represents a tag that you want to add to the
310
+ # resource. The value can be an empty string. This value helps you
311
+ # manage, filter, and search for your resources. We recommend that you
312
+ # create a name tag for your location.
313
+ # @return [Array<Types::TagListEntry>]
314
+ #
315
+ # @see http://docs.aws.amazon.com/goto/WebAPI/datasync-2018-11-09/CreateLocationFsxLustreRequest AWS API Documentation
316
+ #
317
+ class CreateLocationFsxLustreRequest < Struct.new(
318
+ :fsx_filesystem_arn,
319
+ :security_group_arns,
320
+ :subdirectory,
321
+ :tags)
322
+ SENSITIVE = []
323
+ include Aws::Structure
324
+ end
325
+
326
+ # @!attribute [rw] location_arn
327
+ # The Amazon Resource Name (ARN) of the FSx for Lustre file system
328
+ # location that's created.
329
+ # @return [String]
330
+ #
331
+ # @see http://docs.aws.amazon.com/goto/WebAPI/datasync-2018-11-09/CreateLocationFsxLustreResponse AWS API Documentation
332
+ #
333
+ class CreateLocationFsxLustreResponse < Struct.new(
334
+ :location_arn)
335
+ SENSITIVE = []
336
+ include Aws::Structure
337
+ end
338
+
278
339
  # @note When making an API call, you may pass CreateLocationFsxWindowsRequest
279
340
  # data as a hash:
280
341
  #
@@ -294,7 +355,7 @@ module Aws::DataSync
294
355
  # }
295
356
  #
296
357
  # @!attribute [rw] subdirectory
297
- # A subdirectory in the locations path. This subdirectory in the
358
+ # A subdirectory in the location's path. This subdirectory in the
298
359
  # Amazon FSx for Windows File Server file system is used to read data
299
360
  # from the Amazon FSx for Windows File Server source location or write
300
361
  # data to the FSx for Windows File Server destination.
@@ -306,8 +367,8 @@ module Aws::DataSync
306
367
  # @return [String]
307
368
  #
308
369
  # @!attribute [rw] security_group_arns
309
- # The Amazon Resource Names (ARNs) of the security groups that are to
310
- # use to configure the FSx for Windows File Server file system.
370
+ # The Amazon Resource Names (ARNs) of the security groups that are
371
+ # used to configure the FSx for Windows File Server file system.
311
372
  # @return [Array<String>]
312
373
  #
313
374
  # @!attribute [rw] tags
@@ -363,6 +424,169 @@ module Aws::DataSync
363
424
  include Aws::Structure
364
425
  end
365
426
 
427
+ # @note When making an API call, you may pass CreateLocationHdfsRequest
428
+ # data as a hash:
429
+ #
430
+ # {
431
+ # subdirectory: "HdfsSubdirectory",
432
+ # name_nodes: [ # required
433
+ # {
434
+ # hostname: "HdfsServerHostname", # required
435
+ # port: 1, # required
436
+ # },
437
+ # ],
438
+ # block_size: 1,
439
+ # replication_factor: 1,
440
+ # kms_key_provider_uri: "KmsKeyProviderUri",
441
+ # qop_configuration: {
442
+ # rpc_protection: "DISABLED", # accepts DISABLED, AUTHENTICATION, INTEGRITY, PRIVACY
443
+ # data_transfer_protection: "DISABLED", # accepts DISABLED, AUTHENTICATION, INTEGRITY, PRIVACY
444
+ # },
445
+ # authentication_type: "SIMPLE", # required, accepts SIMPLE, KERBEROS
446
+ # simple_user: "HdfsUser",
447
+ # kerberos_principal: "KerberosPrincipal",
448
+ # kerberos_keytab: "data",
449
+ # kerberos_krb_5_conf: "data",
450
+ # agent_arns: ["AgentArn"], # required
451
+ # tags: [
452
+ # {
453
+ # key: "TagKey", # required
454
+ # value: "TagValue",
455
+ # },
456
+ # ],
457
+ # }
458
+ #
459
+ # @!attribute [rw] subdirectory
460
+ # A subdirectory in the HDFS cluster. This subdirectory is used to
461
+ # read data from or write data to the HDFS cluster. If the
462
+ # subdirectory isn't specified, it will default to `/`.
463
+ # @return [String]
464
+ #
465
+ # @!attribute [rw] name_nodes
466
+ # The NameNode that manages the HDFS namespace. The NameNode performs
467
+ # operations such as opening, closing, and renaming files and
468
+ # directories. The NameNode contains the information to map blocks of
469
+ # data to the DataNodes. You can use only one NameNode.
470
+ # @return [Array<Types::HdfsNameNode>]
471
+ #
472
+ # @!attribute [rw] block_size
473
+ # The size of data blocks to write into the HDFS cluster. The block
474
+ # size must be a multiple of 512 bytes. The default block size is 128
475
+ # mebibytes (MiB).
476
+ # @return [Integer]
477
+ #
478
+ # @!attribute [rw] replication_factor
479
+ # The number of DataNodes to replicate the data to when writing to the
480
+ # HDFS cluster. By default, data is replicated to three DataNodes.
481
+ # @return [Integer]
482
+ #
483
+ # @!attribute [rw] kms_key_provider_uri
484
+ # The URI of the HDFS cluster's Key Management Server (KMS).
485
+ # @return [String]
486
+ #
487
+ # @!attribute [rw] qop_configuration
488
+ # The Quality of Protection (QOP) configuration specifies the Remote
489
+ # Procedure Call (RPC) and data transfer protection settings
490
+ # configured on the Hadoop Distributed File System (HDFS) cluster. If
491
+ # `QopConfiguration` isn't specified, `RpcProtection` and
492
+ # `DataTransferProtection` default to `PRIVACY`. If you set
493
+ # `RpcProtection` or `DataTransferProtection`, the other parameter
494
+ # assumes the same value.
495
+ # @return [Types::QopConfiguration]
496
+ #
497
+ # @!attribute [rw] authentication_type
498
+ # The type of authentication used to determine the identity of the
499
+ # user.
500
+ # @return [String]
501
+ #
502
+ # @!attribute [rw] simple_user
503
+ # The user name used to identify the client on the host operating
504
+ # system.
505
+ #
506
+ # <note markdown="1"> If `SIMPLE` is specified for `AuthenticationType`, this parameter is
507
+ # required.
508
+ #
509
+ # </note>
510
+ # @return [String]
511
+ #
512
+ # @!attribute [rw] kerberos_principal
513
+ # The Kerberos principal with access to the files and folders on the
514
+ # HDFS cluster.
515
+ #
516
+ # <note markdown="1"> If `KERBEROS` is specified for `AuthenticationType`, this parameter
517
+ # is required.
518
+ #
519
+ # </note>
520
+ # @return [String]
521
+ #
522
+ # @!attribute [rw] kerberos_keytab
523
+ # The Kerberos key table (keytab) that contains mappings between the
524
+ # defined Kerberos principal and the encrypted keys. You can load the
525
+ # keytab from a file by providing the file's address. If you're
526
+ # using the CLI, it performs base64 encoding for you. Otherwise,
527
+ # provide the base64-encoded text.
528
+ #
529
+ # <note markdown="1"> If `KERBEROS` is specified for `AuthenticationType`, this parameter
530
+ # is required.
531
+ #
532
+ # </note>
533
+ # @return [String]
534
+ #
535
+ # @!attribute [rw] kerberos_krb_5_conf
536
+ # The `krb5.conf` file that contains the Kerberos configuration
537
+ # information. You can load the `krb5.conf` file by providing the
538
+ # file's address. If you're using the CLI, it performs the base64
539
+ # encoding for you. Otherwise, provide the base64-encoded text.
540
+ #
541
+ # <note markdown="1"> If `KERBEROS` is specified for `AuthenticationType`, this parameter
542
+ # is required.
543
+ #
544
+ # </note>
545
+ # @return [String]
546
+ #
547
+ # @!attribute [rw] agent_arns
548
+ # The Amazon Resource Names (ARNs) of the agents that are used to
549
+ # connect to the HDFS cluster.
550
+ # @return [Array<String>]
551
+ #
552
+ # @!attribute [rw] tags
553
+ # The key-value pair that represents the tag that you want to add to
554
+ # the location. The value can be an empty string. We recommend using
555
+ # tags to name your resources.
556
+ # @return [Array<Types::TagListEntry>]
557
+ #
558
+ # @see http://docs.aws.amazon.com/goto/WebAPI/datasync-2018-11-09/CreateLocationHdfsRequest AWS API Documentation
559
+ #
560
+ class CreateLocationHdfsRequest < Struct.new(
561
+ :subdirectory,
562
+ :name_nodes,
563
+ :block_size,
564
+ :replication_factor,
565
+ :kms_key_provider_uri,
566
+ :qop_configuration,
567
+ :authentication_type,
568
+ :simple_user,
569
+ :kerberos_principal,
570
+ :kerberos_keytab,
571
+ :kerberos_krb_5_conf,
572
+ :agent_arns,
573
+ :tags)
574
+ SENSITIVE = []
575
+ include Aws::Structure
576
+ end
577
+
578
+ # @!attribute [rw] location_arn
579
+ # The ARN of the source HDFS cluster location that's created.
580
+ # @return [String]
581
+ #
582
+ # @see http://docs.aws.amazon.com/goto/WebAPI/datasync-2018-11-09/CreateLocationHdfsResponse AWS API Documentation
583
+ #
584
+ class CreateLocationHdfsResponse < Struct.new(
585
+ :location_arn)
586
+ SENSITIVE = []
587
+ include Aws::Structure
588
+ end
589
+
366
590
  # CreateLocationNfsRequest
367
591
  #
368
592
  # @note When making an API call, you may pass CreateLocationNfsRequest
@@ -930,9 +1154,9 @@ module Aws::DataSync
930
1154
  #
931
1155
  # @!attribute [rw] includes
932
1156
  # A list of filter rules that determines which files to include when
933
- # running a task. The pattern should contain a single filter string
934
- # that consists of the patterns to include. The patterns are delimited
935
- # by "\|" (that is, a pipe). For example: `"/folder1|/folder2`"
1157
+ # running a task. The pattern contains a single filter string that
1158
+ # consists of the patterns to include. The patterns are delimited by
1159
+ # "\|" (that is, a pipe), for example, `"/folder1|/folder2"`.
936
1160
  # @return [Array<Types::FilterRule>]
937
1161
  #
938
1162
  # @see http://docs.aws.amazon.com/goto/WebAPI/datasync-2018-11-09/CreateTaskRequest AWS API Documentation
@@ -1171,6 +1395,55 @@ module Aws::DataSync
1171
1395
  include Aws::Structure
1172
1396
  end
1173
1397
 
1398
+ # @note When making an API call, you may pass DescribeLocationFsxLustreRequest
1399
+ # data as a hash:
1400
+ #
1401
+ # {
1402
+ # location_arn: "LocationArn", # required
1403
+ # }
1404
+ #
1405
+ # @!attribute [rw] location_arn
1406
+ # The Amazon Resource Name (ARN) of the FSx for Lustre location to
1407
+ # describe.
1408
+ # @return [String]
1409
+ #
1410
+ # @see http://docs.aws.amazon.com/goto/WebAPI/datasync-2018-11-09/DescribeLocationFsxLustreRequest AWS API Documentation
1411
+ #
1412
+ class DescribeLocationFsxLustreRequest < Struct.new(
1413
+ :location_arn)
1414
+ SENSITIVE = []
1415
+ include Aws::Structure
1416
+ end
1417
+
1418
+ # @!attribute [rw] location_arn
1419
+ # The Amazon Resource Name (ARN) of the FSx for Lustre location that
1420
+ # was described.
1421
+ # @return [String]
1422
+ #
1423
+ # @!attribute [rw] location_uri
1424
+ # The URI of the FSx for Lustre location that was described.
1425
+ # @return [String]
1426
+ #
1427
+ # @!attribute [rw] security_group_arns
1428
+ # The Amazon Resource Names (ARNs) of the security groups that are
1429
+ # configured for the FSx for Lustre file system.
1430
+ # @return [Array<String>]
1431
+ #
1432
+ # @!attribute [rw] creation_time
1433
+ # The time that the FSx for Lustre location was created.
1434
+ # @return [Time]
1435
+ #
1436
+ # @see http://docs.aws.amazon.com/goto/WebAPI/datasync-2018-11-09/DescribeLocationFsxLustreResponse AWS API Documentation
1437
+ #
1438
+ class DescribeLocationFsxLustreResponse < Struct.new(
1439
+ :location_arn,
1440
+ :location_uri,
1441
+ :security_group_arns,
1442
+ :creation_time)
1443
+ SENSITIVE = []
1444
+ include Aws::Structure
1445
+ end
1446
+
1174
1447
  # @note When making an API call, you may pass DescribeLocationFsxWindowsRequest
1175
1448
  # data as a hash:
1176
1449
  #
@@ -1233,6 +1506,101 @@ module Aws::DataSync
1233
1506
  include Aws::Structure
1234
1507
  end
1235
1508
 
1509
+ # @note When making an API call, you may pass DescribeLocationHdfsRequest
1510
+ # data as a hash:
1511
+ #
1512
+ # {
1513
+ # location_arn: "LocationArn", # required
1514
+ # }
1515
+ #
1516
+ # @!attribute [rw] location_arn
1517
+ # The Amazon Resource Name (ARN) of the HDFS cluster location to
1518
+ # describe.
1519
+ # @return [String]
1520
+ #
1521
+ # @see http://docs.aws.amazon.com/goto/WebAPI/datasync-2018-11-09/DescribeLocationHdfsRequest AWS API Documentation
1522
+ #
1523
+ class DescribeLocationHdfsRequest < Struct.new(
1524
+ :location_arn)
1525
+ SENSITIVE = []
1526
+ include Aws::Structure
1527
+ end
1528
+
1529
+ # @!attribute [rw] location_arn
1530
+ # The ARN of the HDFS cluster location.
1531
+ # @return [String]
1532
+ #
1533
+ # @!attribute [rw] location_uri
1534
+ # The URI of the HDFS cluster location.
1535
+ # @return [String]
1536
+ #
1537
+ # @!attribute [rw] name_nodes
1538
+ # The NameNode that manage the HDFS namespace.
1539
+ # @return [Array<Types::HdfsNameNode>]
1540
+ #
1541
+ # @!attribute [rw] block_size
1542
+ # The size of the data blocks to write into the HDFS cluster.
1543
+ # @return [Integer]
1544
+ #
1545
+ # @!attribute [rw] replication_factor
1546
+ # The number of DataNodes to replicate the data to when writing to the
1547
+ # HDFS cluster.
1548
+ # @return [Integer]
1549
+ #
1550
+ # @!attribute [rw] kms_key_provider_uri
1551
+ # The URI of the HDFS cluster's Key Management Server (KMS).
1552
+ # @return [String]
1553
+ #
1554
+ # @!attribute [rw] qop_configuration
1555
+ # The Quality of Protection (QOP) configuration specifies the Remote
1556
+ # Procedure Call (RPC) and data transfer protection settings
1557
+ # configured on the Hadoop Distributed File System (HDFS) cluster.
1558
+ # @return [Types::QopConfiguration]
1559
+ #
1560
+ # @!attribute [rw] authentication_type
1561
+ # The type of authentication used to determine the identity of the
1562
+ # user.
1563
+ # @return [String]
1564
+ #
1565
+ # @!attribute [rw] simple_user
1566
+ # The user name used to identify the client on the host operating
1567
+ # system. This parameter is used if the `AuthenticationType` is
1568
+ # defined as `SIMPLE`.
1569
+ # @return [String]
1570
+ #
1571
+ # @!attribute [rw] kerberos_principal
1572
+ # The Kerberos principal with access to the files and folders on the
1573
+ # HDFS cluster. This parameter is used if the `AuthenticationType` is
1574
+ # defined as `KERBEROS`.
1575
+ # @return [String]
1576
+ #
1577
+ # @!attribute [rw] agent_arns
1578
+ # The ARNs of the agents that are used to connect to the HDFS cluster.
1579
+ # @return [Array<String>]
1580
+ #
1581
+ # @!attribute [rw] creation_time
1582
+ # The time that the HDFS location was created.
1583
+ # @return [Time]
1584
+ #
1585
+ # @see http://docs.aws.amazon.com/goto/WebAPI/datasync-2018-11-09/DescribeLocationHdfsResponse AWS API Documentation
1586
+ #
1587
+ class DescribeLocationHdfsResponse < Struct.new(
1588
+ :location_arn,
1589
+ :location_uri,
1590
+ :name_nodes,
1591
+ :block_size,
1592
+ :replication_factor,
1593
+ :kms_key_provider_uri,
1594
+ :qop_configuration,
1595
+ :authentication_type,
1596
+ :simple_user,
1597
+ :kerberos_principal,
1598
+ :agent_arns,
1599
+ :creation_time)
1600
+ SENSITIVE = []
1601
+ include Aws::Structure
1602
+ end
1603
+
1236
1604
  # DescribeLocationNfsRequest
1237
1605
  #
1238
1606
  # @note When making an API call, you may pass DescribeLocationNfsRequest
@@ -1719,13 +2087,13 @@ module Aws::DataSync
1719
2087
  # @return [String]
1720
2088
  #
1721
2089
  # @!attribute [rw] source_network_interface_arns
1722
- # The Amazon Resource Name (ARN) of the source ENIs (Elastic Network
1723
- # Interface) that was created for your subnet.
2090
+ # The Amazon Resource Names (ARNs) of the source elastic network
2091
+ # interfaces (ENIs) that were created for your subnet.
1724
2092
  # @return [Array<String>]
1725
2093
  #
1726
2094
  # @!attribute [rw] destination_network_interface_arns
1727
- # The Amazon Resource Name (ARN) of the destination ENIs (Elastic
1728
- # Network Interface) that was created for your subnet.
2095
+ # The Amazon Resource Names (ARNs) of the destination elastic network
2096
+ # interfaces (ENIs) that were created for your subnet.
1729
2097
  # @return [Array<String>]
1730
2098
  #
1731
2099
  # @!attribute [rw] options
@@ -1748,7 +2116,7 @@ module Aws::DataSync
1748
2116
  # A list of filter rules that determines which files to exclude from a
1749
2117
  # task. The list should contain a single filter string that consists
1750
2118
  # of the patterns to exclude. The patterns are delimited by "\|"
1751
- # (that is, a pipe), for example: `"/folder1|/folder2"`
2119
+ # (that is, a pipe), for example, `"/folder1|/folder2"`.
1752
2120
  # @return [Array<Types::FilterRule>]
1753
2121
  #
1754
2122
  # @!attribute [rw] schedule
@@ -1773,9 +2141,9 @@ module Aws::DataSync
1773
2141
  #
1774
2142
  # @!attribute [rw] includes
1775
2143
  # A list of filter rules that determines which files to include when
1776
- # running a task. The pattern should contain a single filter string
1777
- # that consists of the patterns to include. The patterns are delimited
1778
- # by "\|" (that is, a pipe). For example: `"/folder1|/folder2`"
2144
+ # running a task. The pattern contains a single filter string that
2145
+ # consists of the patterns to include. The patterns are delimited by
2146
+ # "\|" (that is, a pipe), for example, `"/folder1|/folder2`".
1779
2147
  # @return [Array<Types::FilterRule>]
1780
2148
  #
1781
2149
  # @see http://docs.aws.amazon.com/goto/WebAPI/datasync-2018-11-09/DescribeTaskResponse AWS API Documentation
@@ -1865,6 +2233,40 @@ module Aws::DataSync
1865
2233
  include Aws::Structure
1866
2234
  end
1867
2235
 
2236
+ # The NameNode of the Hadoop Distributed File System (HDFS). The
2237
+ # NameNode manages the file system's namespace. The NameNode performs
2238
+ # operations such as opening, closing, and renaming files and
2239
+ # directories. The NameNode contains the information to map blocks of
2240
+ # data to the DataNodes.
2241
+ #
2242
+ # @note When making an API call, you may pass HdfsNameNode
2243
+ # data as a hash:
2244
+ #
2245
+ # {
2246
+ # hostname: "HdfsServerHostname", # required
2247
+ # port: 1, # required
2248
+ # }
2249
+ #
2250
+ # @!attribute [rw] hostname
2251
+ # The hostname of the NameNode in the HDFS cluster. This value is the
2252
+ # IP address or Domain Name Service (DNS) name of the NameNode. An
2253
+ # agent that's installed on-premises uses this hostname to
2254
+ # communicate with the NameNode in the network.
2255
+ # @return [String]
2256
+ #
2257
+ # @!attribute [rw] port
2258
+ # The port that the NameNode uses to listen to client requests.
2259
+ # @return [Integer]
2260
+ #
2261
+ # @see http://docs.aws.amazon.com/goto/WebAPI/datasync-2018-11-09/HdfsNameNode AWS API Documentation
2262
+ #
2263
+ class HdfsNameNode < Struct.new(
2264
+ :hostname,
2265
+ :port)
2266
+ SENSITIVE = []
2267
+ include Aws::Structure
2268
+ end
2269
+
1868
2270
  # This exception is thrown when an error occurs in the DataSync service.
1869
2271
  #
1870
2272
  # @!attribute [rw] message
@@ -2243,7 +2645,7 @@ module Aws::DataSync
2243
2645
  # @return [String]
2244
2646
  #
2245
2647
  # @!attribute [rw] location_uri
2246
- # Represents a list of URLs of a location. `LocationUri` returns an
2648
+ # Represents a list of URIs of a location. `LocationUri` returns an
2247
2649
  # array that contains a list of locations when the [ListLocations][1]
2248
2650
  # operation is called.
2249
2651
  #
@@ -2304,7 +2706,7 @@ module Aws::DataSync
2304
2706
  #
2305
2707
  # * <b> <a href="https://tools.ietf.org/html/rfc3530">NFSv4.0</a> </b>
2306
2708
  # - stateful, firewall-friendly protocol version that supports
2307
- # delegations and pseudo filesystems.
2709
+ # delegations and pseudo file systems.
2308
2710
  #
2309
2711
  # * <b> <a href="https://tools.ietf.org/html/rfc5661">NFSv4.1</a> </b>
2310
2712
  # - stateful protocol version that supports sessions, directory
@@ -2725,6 +3127,39 @@ module Aws::DataSync
2725
3127
  include Aws::Structure
2726
3128
  end
2727
3129
 
3130
+ # The Quality of Protection (QOP) configuration specifies the Remote
3131
+ # Procedure Call (RPC) and data transfer privacy settings configured on
3132
+ # the Hadoop Distributed File System (HDFS) cluster.
3133
+ #
3134
+ # @note When making an API call, you may pass QopConfiguration
3135
+ # data as a hash:
3136
+ #
3137
+ # {
3138
+ # rpc_protection: "DISABLED", # accepts DISABLED, AUTHENTICATION, INTEGRITY, PRIVACY
3139
+ # data_transfer_protection: "DISABLED", # accepts DISABLED, AUTHENTICATION, INTEGRITY, PRIVACY
3140
+ # }
3141
+ #
3142
+ # @!attribute [rw] rpc_protection
3143
+ # The RPC protection setting configured on the HDFS cluster. This
3144
+ # setting corresponds to your `hadoop.rpc.protection` setting in your
3145
+ # `core-site.xml` file on your Hadoop cluster.
3146
+ # @return [String]
3147
+ #
3148
+ # @!attribute [rw] data_transfer_protection
3149
+ # The data transfer protection setting configured on the HDFS cluster.
3150
+ # This setting corresponds to your `dfs.data.transfer.protection`
3151
+ # setting in the `hdfs-site.xml` file on your Hadoop cluster.
3152
+ # @return [String]
3153
+ #
3154
+ # @see http://docs.aws.amazon.com/goto/WebAPI/datasync-2018-11-09/QopConfiguration AWS API Documentation
3155
+ #
3156
+ class QopConfiguration < Struct.new(
3157
+ :rpc_protection,
3158
+ :data_transfer_protection)
3159
+ SENSITIVE = []
3160
+ include Aws::Structure
3161
+ end
3162
+
2728
3163
  # The Amazon Resource Name (ARN) of the Identity and Access Management
2729
3164
  # (IAM) role that is used to access an Amazon S3 bucket.
2730
3165
  #
@@ -2843,14 +3278,14 @@ module Aws::DataSync
2843
3278
  # A list of filter rules that determines which files to include when
2844
3279
  # running a task. The pattern should contain a single filter string
2845
3280
  # that consists of the patterns to include. The patterns are delimited
2846
- # by "\|" (that is, a pipe). For example: `"/folder1|/folder2"`
3281
+ # by "\|" (that is, a pipe), for example, `"/folder1|/folder2"`.
2847
3282
  # @return [Array<Types::FilterRule>]
2848
3283
  #
2849
3284
  # @!attribute [rw] excludes
2850
3285
  # A list of filter rules that determines which files to exclude from a
2851
- # task. The list should contain a single filter string that consists
2852
- # of the patterns to exclude. The patterns are delimited by "\|"
2853
- # (that is, a pipe), for example, `"/folder1|/folder2"`.
3286
+ # task. The list contains a single filter string that consists of the
3287
+ # patterns to exclude. The patterns are delimited by "\|" (that is,
3288
+ # a pipe), for example, `"/folder1|/folder2"`.
2854
3289
  # @return [Array<Types::FilterRule>]
2855
3290
  #
2856
3291
  # @see http://docs.aws.amazon.com/goto/WebAPI/datasync-2018-11-09/StartTaskExecutionRequest AWS API Documentation
@@ -2879,9 +3314,9 @@ module Aws::DataSync
2879
3314
  include Aws::Structure
2880
3315
  end
2881
3316
 
2882
- # Represents a single entry in a list of AWS resource tags.
2883
- # `TagListEntry` returns an array that contains a list of tasks when the
2884
- # [ListTagsForResource][1] operation is called.
3317
+ # Represents a single entry in a list of Amazon Web Services resource
3318
+ # tags. `TagListEntry` returns an array that contains a list of tasks
3319
+ # when the [ListTagsForResource][1] operation is called.
2885
3320
  #
2886
3321
  #
2887
3322
  #
@@ -3204,6 +3639,126 @@ module Aws::DataSync
3204
3639
  #
3205
3640
  class UpdateAgentResponse < Aws::EmptyStructure; end
3206
3641
 
3642
+ # @note When making an API call, you may pass UpdateLocationHdfsRequest
3643
+ # data as a hash:
3644
+ #
3645
+ # {
3646
+ # location_arn: "LocationArn", # required
3647
+ # subdirectory: "HdfsSubdirectory",
3648
+ # name_nodes: [
3649
+ # {
3650
+ # hostname: "HdfsServerHostname", # required
3651
+ # port: 1, # required
3652
+ # },
3653
+ # ],
3654
+ # block_size: 1,
3655
+ # replication_factor: 1,
3656
+ # kms_key_provider_uri: "KmsKeyProviderUri",
3657
+ # qop_configuration: {
3658
+ # rpc_protection: "DISABLED", # accepts DISABLED, AUTHENTICATION, INTEGRITY, PRIVACY
3659
+ # data_transfer_protection: "DISABLED", # accepts DISABLED, AUTHENTICATION, INTEGRITY, PRIVACY
3660
+ # },
3661
+ # authentication_type: "SIMPLE", # accepts SIMPLE, KERBEROS
3662
+ # simple_user: "HdfsUser",
3663
+ # kerberos_principal: "KerberosPrincipal",
3664
+ # kerberos_keytab: "data",
3665
+ # kerberos_krb_5_conf: "data",
3666
+ # agent_arns: ["AgentArn"],
3667
+ # }
3668
+ #
3669
+ # @!attribute [rw] location_arn
3670
+ # The Amazon Resource Name (ARN) of the source HDFS cluster location.
3671
+ # @return [String]
3672
+ #
3673
+ # @!attribute [rw] subdirectory
3674
+ # A subdirectory in the HDFS cluster. This subdirectory is used to
3675
+ # read data from or write data to the HDFS cluster.
3676
+ # @return [String]
3677
+ #
3678
+ # @!attribute [rw] name_nodes
3679
+ # The NameNode that manages the HDFS namespace. The NameNode performs
3680
+ # operations such as opening, closing, and renaming files and
3681
+ # directories. The NameNode contains the information to map blocks of
3682
+ # data to the DataNodes. You can use only one NameNode.
3683
+ # @return [Array<Types::HdfsNameNode>]
3684
+ #
3685
+ # @!attribute [rw] block_size
3686
+ # The size of the data blocks to write into the HDFS cluster.
3687
+ # @return [Integer]
3688
+ #
3689
+ # @!attribute [rw] replication_factor
3690
+ # The number of DataNodes to replicate the data to when writing to the
3691
+ # HDFS cluster.
3692
+ # @return [Integer]
3693
+ #
3694
+ # @!attribute [rw] kms_key_provider_uri
3695
+ # The URI of the HDFS cluster's Key Management Server (KMS).
3696
+ # @return [String]
3697
+ #
3698
+ # @!attribute [rw] qop_configuration
3699
+ # The Quality of Protection (QOP) configuration specifies the Remote
3700
+ # Procedure Call (RPC) and data transfer privacy settings configured
3701
+ # on the Hadoop Distributed File System (HDFS) cluster.
3702
+ # @return [Types::QopConfiguration]
3703
+ #
3704
+ # @!attribute [rw] authentication_type
3705
+ # The type of authentication used to determine the identity of the
3706
+ # user.
3707
+ # @return [String]
3708
+ #
3709
+ # @!attribute [rw] simple_user
3710
+ # The user name used to identify the client on the host operating
3711
+ # system.
3712
+ # @return [String]
3713
+ #
3714
+ # @!attribute [rw] kerberos_principal
3715
+ # The Kerberos principal with access to the files and folders on the
3716
+ # HDFS cluster.
3717
+ # @return [String]
3718
+ #
3719
+ # @!attribute [rw] kerberos_keytab
3720
+ # The Kerberos key table (keytab) that contains mappings between the
3721
+ # defined Kerberos principal and the encrypted keys. You can load the
3722
+ # keytab from a file by providing the file's address. If you use the
3723
+ # AWS CLI, it performs base64 encoding for you. Otherwise, provide the
3724
+ # base64-encoded text.
3725
+ # @return [String]
3726
+ #
3727
+ # @!attribute [rw] kerberos_krb_5_conf
3728
+ # The `krb5.conf` file that contains the Kerberos configuration
3729
+ # information. You can load the `krb5.conf` file by providing the
3730
+ # file's address. If you're using the AWS CLI, it performs the
3731
+ # base64 encoding for you. Otherwise, provide the base64-encoded text.
3732
+ # @return [String]
3733
+ #
3734
+ # @!attribute [rw] agent_arns
3735
+ # The ARNs of the agents that are used to connect to the HDFS cluster.
3736
+ # @return [Array<String>]
3737
+ #
3738
+ # @see http://docs.aws.amazon.com/goto/WebAPI/datasync-2018-11-09/UpdateLocationHdfsRequest AWS API Documentation
3739
+ #
3740
+ class UpdateLocationHdfsRequest < Struct.new(
3741
+ :location_arn,
3742
+ :subdirectory,
3743
+ :name_nodes,
3744
+ :block_size,
3745
+ :replication_factor,
3746
+ :kms_key_provider_uri,
3747
+ :qop_configuration,
3748
+ :authentication_type,
3749
+ :simple_user,
3750
+ :kerberos_principal,
3751
+ :kerberos_keytab,
3752
+ :kerberos_krb_5_conf,
3753
+ :agent_arns)
3754
+ SENSITIVE = []
3755
+ include Aws::Structure
3756
+ end
3757
+
3758
+ # @see http://docs.aws.amazon.com/goto/WebAPI/datasync-2018-11-09/UpdateLocationHdfsResponse AWS API Documentation
3759
+ #
3760
+ class UpdateLocationHdfsResponse < Aws::EmptyStructure; end
3761
+
3207
3762
  # @note When making an API call, you may pass UpdateLocationNfsRequest
3208
3763
  # data as a hash:
3209
3764
  #
@@ -3571,7 +4126,7 @@ module Aws::DataSync
3571
4126
  # A list of filter rules that determines which files to exclude from a
3572
4127
  # task. The list should contain a single filter string that consists
3573
4128
  # of the patterns to exclude. The patterns are delimited by "\|"
3574
- # (that is, a pipe), for example: `"/folder1|/folder2"`
4129
+ # (that is, a pipe), for example, `"/folder1|/folder2"`.
3575
4130
  # @return [Array<Types::FilterRule>]
3576
4131
  #
3577
4132
  # @!attribute [rw] schedule
@@ -3592,15 +4147,15 @@ module Aws::DataSync
3592
4147
  # @return [String]
3593
4148
  #
3594
4149
  # @!attribute [rw] cloud_watch_log_group_arn
3595
- # The Amazon Resource Name (ARN) of the resource name of the
3596
- # CloudWatch LogGroup.
4150
+ # The Amazon Resource Name (ARN) of the resource name of the Amazon
4151
+ # CloudWatch log group.
3597
4152
  # @return [String]
3598
4153
  #
3599
4154
  # @!attribute [rw] includes
3600
4155
  # A list of filter rules that determines which files to include when
3601
- # running a task. The pattern should contain a single filter string
3602
- # that consists of the patterns to include. The patterns are delimited
3603
- # by "\|" (that is, a pipe). For example: `"/folder1|/folder2`"
4156
+ # running a task. The pattern contains a single filter string that
4157
+ # consists of the patterns to include. The patterns are delimited by
4158
+ # "\|" (that is, a pipe), for example, `"/folder1|/folder2"`.
3604
4159
  # @return [Array<Types::FilterRule>]
3605
4160
  #
3606
4161
  # @see http://docs.aws.amazon.com/goto/WebAPI/datasync-2018-11-09/UpdateTaskRequest AWS API Documentation