cdk-lambda-subminute 2.0.278 → 2.0.280

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (37) hide show
  1. package/.gitattributes +0 -1
  2. package/.jsii +3 -3
  3. package/lib/cdk-lambda-subminute.js +3 -3
  4. package/node_modules/aws-sdk/CHANGELOG.md +17 -1
  5. package/node_modules/aws-sdk/README.md +1 -1
  6. package/node_modules/aws-sdk/apis/appstream-2016-12-01.min.json +28 -5
  7. package/node_modules/aws-sdk/apis/ec2-2016-11-15.min.json +115 -56
  8. package/node_modules/aws-sdk/apis/ec2-2016-11-15.paginators.json +6 -0
  9. package/node_modules/aws-sdk/apis/groundstation-2019-05-23.min.json +26 -25
  10. package/node_modules/aws-sdk/apis/network-firewall-2020-11-12.min.json +30 -16
  11. package/node_modules/aws-sdk/apis/opensearch-2021-01-01.min.json +184 -167
  12. package/node_modules/aws-sdk/apis/redshift-2012-12-01.min.json +168 -58
  13. package/node_modules/aws-sdk/apis/redshift-2012-12-01.paginators.json +6 -0
  14. package/node_modules/aws-sdk/apis/sagemaker-2017-07-24.min.json +877 -862
  15. package/node_modules/aws-sdk/apis/ssm-sap-2018-05-10.min.json +62 -27
  16. package/node_modules/aws-sdk/apis/transfer-2018-11-05.min.json +53 -53
  17. package/node_modules/aws-sdk/apis/transfer-2018-11-05.paginators.json +12 -24
  18. package/node_modules/aws-sdk/apis/transfer-2018-11-05.waiters2.json +33 -41
  19. package/node_modules/aws-sdk/clients/appstream.d.ts +46 -6
  20. package/node_modules/aws-sdk/clients/connectcases.d.ts +5 -5
  21. package/node_modules/aws-sdk/clients/ec2.d.ts +68 -0
  22. package/node_modules/aws-sdk/clients/groundstation.d.ts +5 -0
  23. package/node_modules/aws-sdk/clients/iam.d.ts +4 -4
  24. package/node_modules/aws-sdk/clients/networkfirewall.d.ts +31 -11
  25. package/node_modules/aws-sdk/clients/opensearch.d.ts +47 -22
  26. package/node_modules/aws-sdk/clients/redshift.d.ts +143 -0
  27. package/node_modules/aws-sdk/clients/sagemaker.d.ts +36 -14
  28. package/node_modules/aws-sdk/clients/sns.d.ts +4 -4
  29. package/node_modules/aws-sdk/clients/ssmsap.d.ts +75 -4
  30. package/node_modules/aws-sdk/clients/transfer.d.ts +3 -3
  31. package/node_modules/aws-sdk/dist/aws-sdk-core-react-native.js +1 -1
  32. package/node_modules/aws-sdk/dist/aws-sdk-react-native.js +14 -14
  33. package/node_modules/aws-sdk/dist/aws-sdk.js +298 -117
  34. package/node_modules/aws-sdk/dist/aws-sdk.min.js +68 -68
  35. package/node_modules/aws-sdk/lib/core.js +1 -1
  36. package/node_modules/aws-sdk/package.json +1 -1
  37. package/package.json +3 -4
@@ -260,11 +260,11 @@ declare class OpenSearch extends Service {
260
260
  */
261
261
  getCompatibleVersions(callback?: (err: AWSError, data: OpenSearch.Types.GetCompatibleVersionsResponse) => void): Request<OpenSearch.Types.GetCompatibleVersionsResponse, AWSError>;
262
262
  /**
263
- * Get the status of the maintenance action.
263
+ * The status of the maintenance action.
264
264
  */
265
265
  getDomainMaintenanceStatus(params: OpenSearch.Types.GetDomainMaintenanceStatusRequest, callback?: (err: AWSError, data: OpenSearch.Types.GetDomainMaintenanceStatusResponse) => void): Request<OpenSearch.Types.GetDomainMaintenanceStatusResponse, AWSError>;
266
266
  /**
267
- * Get the status of the maintenance action.
267
+ * The status of the maintenance action.
268
268
  */
269
269
  getDomainMaintenanceStatus(callback?: (err: AWSError, data: OpenSearch.Types.GetDomainMaintenanceStatusResponse) => void): Request<OpenSearch.Types.GetDomainMaintenanceStatusResponse, AWSError>;
270
270
  /**
@@ -292,11 +292,11 @@ declare class OpenSearch extends Service {
292
292
  */
293
293
  getUpgradeStatus(callback?: (err: AWSError, data: OpenSearch.Types.GetUpgradeStatusResponse) => void): Request<OpenSearch.Types.GetUpgradeStatusResponse, AWSError>;
294
294
  /**
295
- * Get the list of the maintenance action.
295
+ * A list of maintenance actions for the domain.
296
296
  */
297
297
  listDomainMaintenances(params: OpenSearch.Types.ListDomainMaintenancesRequest, callback?: (err: AWSError, data: OpenSearch.Types.ListDomainMaintenancesResponse) => void): Request<OpenSearch.Types.ListDomainMaintenancesResponse, AWSError>;
298
298
  /**
299
- * Get the list of the maintenance action.
299
+ * A list of maintenance actions for the domain.
300
300
  */
301
301
  listDomainMaintenances(callback?: (err: AWSError, data: OpenSearch.Types.ListDomainMaintenancesResponse) => void): Request<OpenSearch.Types.ListDomainMaintenancesResponse, AWSError>;
302
302
  /**
@@ -412,11 +412,11 @@ declare class OpenSearch extends Service {
412
412
  */
413
413
  revokeVpcEndpointAccess(callback?: (err: AWSError, data: OpenSearch.Types.RevokeVpcEndpointAccessResponse) => void): Request<OpenSearch.Types.RevokeVpcEndpointAccessResponse, AWSError>;
414
414
  /**
415
- * Starts the node maintenance (Node restart, Node reboot, Opensearch/Elasticsearch process restart, Dashboard/kibana restart) on the data node.
415
+ * Starts the node maintenance process on the data node. These processes can include a node reboot, an Opensearch or Elasticsearch process restart, or a Dashboard or Kibana restart.
416
416
  */
417
417
  startDomainMaintenance(params: OpenSearch.Types.StartDomainMaintenanceRequest, callback?: (err: AWSError, data: OpenSearch.Types.StartDomainMaintenanceResponse) => void): Request<OpenSearch.Types.StartDomainMaintenanceResponse, AWSError>;
418
418
  /**
419
- * Starts the node maintenance (Node restart, Node reboot, Opensearch/Elasticsearch process restart, Dashboard/kibana restart) on the data node.
419
+ * Starts the node maintenance process on the data node. These processes can include a node reboot, an Opensearch or Elasticsearch process restart, or a Dashboard or Kibana restart.
420
420
  */
421
421
  startDomainMaintenance(callback?: (err: AWSError, data: OpenSearch.Types.StartDomainMaintenanceResponse) => void): Request<OpenSearch.Types.StartDomainMaintenanceResponse, AWSError>;
422
422
  /**
@@ -1004,6 +1004,10 @@ declare namespace OpenSearch {
1004
1004
  * Identity and Access Management (IAM) policy document specifying the access policies for the new domain.
1005
1005
  */
1006
1006
  AccessPolicies?: PolicyDocument;
1007
+ /**
1008
+ * The type of IP addresses supported by the endpoint for the domain.
1009
+ */
1010
+ IPAddressType?: IPAddressType;
1007
1011
  /**
1008
1012
  * DEPRECATED. Container for the parameters required to configure automated snapshots of domain indexes.
1009
1013
  */
@@ -1611,6 +1615,10 @@ declare namespace OpenSearch {
1611
1615
  * Specifies the access policies for the domain.
1612
1616
  */
1613
1617
  AccessPolicies?: AccessPoliciesStatus;
1618
+ /**
1619
+ * The type of IP addresses supported by the endpoint for the domain.
1620
+ */
1621
+ IPAddressType?: IPAddressTypeStatus;
1614
1622
  /**
1615
1623
  * DEPRECATED. Container for parameters required to configure automated snapshots of domain indexes.
1616
1624
  */
@@ -1717,7 +1725,7 @@ declare namespace OpenSearch {
1717
1725
  }
1718
1726
  export interface DomainMaintenanceDetails {
1719
1727
  /**
1720
- * Id of the requested action.
1728
+ * The ID of the requested action.
1721
1729
  */
1722
1730
  MaintenanceId?: RequestId;
1723
1731
  /**
@@ -1729,7 +1737,7 @@ declare namespace OpenSearch {
1729
1737
  */
1730
1738
  Action?: MaintenanceType;
1731
1739
  /**
1732
- * Id of the data node.
1740
+ * The ID of the data node.
1733
1741
  */
1734
1742
  NodeId?: NodeId;
1735
1743
  /**
@@ -1737,15 +1745,15 @@ declare namespace OpenSearch {
1737
1745
  */
1738
1746
  Status?: MaintenanceStatus;
1739
1747
  /**
1740
- * The status message of the action.
1748
+ * The status message for the action.
1741
1749
  */
1742
1750
  StatusMessage?: MaintenanceStatusMessage;
1743
1751
  /**
1744
- * Contains time at which action created.
1752
+ * The time at which the action was created.
1745
1753
  */
1746
1754
  CreatedAt?: UpdateTimestamp;
1747
1755
  /**
1748
- * Contains time at which action updated.
1756
+ * The time at which the action was updated.
1749
1757
  */
1750
1758
  UpdatedAt?: UpdateTimestamp;
1751
1759
  }
@@ -1854,6 +1862,7 @@ declare namespace OpenSearch {
1854
1862
  * Domain-specific endpoint used to submit index, search, and data upload requests to the domain.
1855
1863
  */
1856
1864
  Endpoint?: ServiceUrl;
1865
+ EndpointV2?: ServiceUrl;
1857
1866
  /**
1858
1867
  * The key-value pair that exists if the OpenSearch Service domain uses VPC endpoints.. Example key, value: 'vpc','vpc-endpoint-h2dsd34efgyghrtguk5gt6j2foh4.us-east-1.es.amazonaws.com'.
1859
1868
  */
@@ -1882,6 +1891,10 @@ declare namespace OpenSearch {
1882
1891
  * Identity and Access Management (IAM) policy document specifying the access policies for the domain.
1883
1892
  */
1884
1893
  AccessPolicies?: PolicyDocument;
1894
+ /**
1895
+ * The type of IP addresses supported by the endpoint for the domain.
1896
+ */
1897
+ IPAddressType?: IPAddressType;
1885
1898
  /**
1886
1899
  * DEPRECATED. Container for parameters required to configure automated snapshots of domain indexes.
1887
1900
  */
@@ -2091,33 +2104,33 @@ declare namespace OpenSearch {
2091
2104
  */
2092
2105
  DomainName: DomainName;
2093
2106
  /**
2094
- * The request id of the maintenance action.
2107
+ * The request ID of the maintenance action.
2095
2108
  */
2096
2109
  MaintenanceId: RequestId;
2097
2110
  }
2098
2111
  export interface GetDomainMaintenanceStatusResponse {
2099
2112
  /**
2100
- * Contains status of the maintenance action.
2113
+ * The status of the maintenance action.
2101
2114
  */
2102
2115
  Status?: MaintenanceStatus;
2103
2116
  /**
2104
- * Contains status message of the maintenance action.
2117
+ * The status message of the maintenance action.
2105
2118
  */
2106
2119
  StatusMessage?: MaintenanceStatusMessage;
2107
2120
  /**
2108
- * Contains node id of maintenance action.
2121
+ * The node ID of the maintenance action.
2109
2122
  */
2110
2123
  NodeId?: NodeId;
2111
2124
  /**
2112
- * Contains action name.
2125
+ * The action name.
2113
2126
  */
2114
2127
  Action?: MaintenanceType;
2115
2128
  /**
2116
- * Contains time at which action created.
2129
+ * The time at which the action was created.
2117
2130
  */
2118
2131
  CreatedAt?: UpdateTimestamp;
2119
2132
  /**
2120
- * Contains time at which action updated.
2133
+ * The time at which the action was updated.
2121
2134
  */
2122
2135
  UpdatedAt?: UpdateTimestamp;
2123
2136
  }
@@ -2193,6 +2206,14 @@ declare namespace OpenSearch {
2193
2206
  */
2194
2207
  UpgradeName?: UpgradeName;
2195
2208
  }
2209
+ export type IPAddressType = "ipv4"|"dualstack"|string;
2210
+ export interface IPAddressTypeStatus {
2211
+ /**
2212
+ * The IP address options for the domain.
2213
+ */
2214
+ Options: IPAddressType;
2215
+ Status: OptionStatus;
2216
+ }
2196
2217
  export type IdentityPoolId = string;
2197
2218
  export interface InboundConnection {
2198
2219
  /**
@@ -2325,13 +2346,13 @@ declare namespace OpenSearch {
2325
2346
  */
2326
2347
  MaxResults?: MaxResults;
2327
2348
  /**
2328
- * If your initial ListDomainMaintenances operation returns a nextToken, you can include the returned nextToken in subsequent ListDomainMaintenances operations, which returns results in the next page.
2349
+ * If your initial ListDomainMaintenances operation returns a nextToken, include the returned nextToken in subsequent ListDomainMaintenances operations, which returns results in the next page.
2329
2350
  */
2330
2351
  NextToken?: NextToken;
2331
2352
  }
2332
2353
  export interface ListDomainMaintenancesResponse {
2333
2354
  /**
2334
- * List of the submitted maintenance actions.
2355
+ * A list of the submitted maintenance actions.
2335
2356
  */
2336
2357
  DomainMaintenances?: DomainMaintenanceList;
2337
2358
  /**
@@ -3199,13 +3220,13 @@ declare namespace OpenSearch {
3199
3220
  */
3200
3221
  Action: MaintenanceType;
3201
3222
  /**
3202
- * Id of the data node.
3223
+ * The ID of the data node.
3203
3224
  */
3204
3225
  NodeId?: NodeId;
3205
3226
  }
3206
3227
  export interface StartDomainMaintenanceResponse {
3207
3228
  /**
3208
- * Contains request id of requested action.
3229
+ * The request ID of requested action.
3209
3230
  */
3210
3231
  MaintenanceId?: RequestId;
3211
3232
  }
@@ -3313,6 +3334,10 @@ declare namespace OpenSearch {
3313
3334
  * Identity and Access Management (IAM) access policy as a JSON-formatted string.
3314
3335
  */
3315
3336
  AccessPolicies?: PolicyDocument;
3337
+ /**
3338
+ * The type of IP addresses supported by the endpoint for the domain.
3339
+ */
3340
+ IPAddressType?: IPAddressType;
3316
3341
  /**
3317
3342
  * Options to publish OpenSearch logs to Amazon CloudWatch Logs.
3318
3343
  */
@@ -332,6 +332,14 @@ declare class Redshift extends Service {
332
332
  * Deletes a partner integration from a cluster. Data can still flow to the cluster until the integration is deleted at the partner's website.
333
333
  */
334
334
  deletePartner(callback?: (err: AWSError, data: Redshift.Types.PartnerIntegrationOutputMessage) => void): Request<Redshift.Types.PartnerIntegrationOutputMessage, AWSError>;
335
+ /**
336
+ * Deletes the resource policy for a specified resource.
337
+ */
338
+ deleteResourcePolicy(params: Redshift.Types.DeleteResourcePolicyMessage, callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
339
+ /**
340
+ * Deletes the resource policy for a specified resource.
341
+ */
342
+ deleteResourcePolicy(callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
335
343
  /**
336
344
  * Deletes a scheduled action.
337
345
  */
@@ -556,6 +564,14 @@ declare class Redshift extends Service {
556
564
  * Returns information about the specified Amazon Redshift HSM configuration. If no configuration ID is specified, returns information about all the HSM configurations owned by your Amazon Web Services account. If you specify both tag keys and tag values in the same request, Amazon Redshift returns all HSM connections that match any combination of the specified keys and values. For example, if you have owner and environment for tag keys, and admin and test for tag values, all HSM connections that have any combination of those values are returned. If both tag keys and values are omitted from the request, HSM connections are returned regardless of whether they have tag keys or values associated with them.
557
565
  */
558
566
  describeHsmConfigurations(callback?: (err: AWSError, data: Redshift.Types.HsmConfigurationMessage) => void): Request<Redshift.Types.HsmConfigurationMessage, AWSError>;
567
+ /**
568
+ * Returns a list of inbound integrations.
569
+ */
570
+ describeInboundIntegrations(params: Redshift.Types.DescribeInboundIntegrationsMessage, callback?: (err: AWSError, data: Redshift.Types.InboundIntegrationsMessage) => void): Request<Redshift.Types.InboundIntegrationsMessage, AWSError>;
571
+ /**
572
+ * Returns a list of inbound integrations.
573
+ */
574
+ describeInboundIntegrations(callback?: (err: AWSError, data: Redshift.Types.InboundIntegrationsMessage) => void): Request<Redshift.Types.InboundIntegrationsMessage, AWSError>;
559
575
  /**
560
576
  * Describes whether information, such as queries and connection attempts, is being logged for the specified Amazon Redshift cluster.
561
577
  */
@@ -744,6 +760,14 @@ declare class Redshift extends Service {
744
760
  * Returns an array of DC2 ReservedNodeOfferings that matches the payment type, term, and usage price of the given DC1 reserved node.
745
761
  */
746
762
  getReservedNodeExchangeOfferings(callback?: (err: AWSError, data: Redshift.Types.GetReservedNodeExchangeOfferingsOutputMessage) => void): Request<Redshift.Types.GetReservedNodeExchangeOfferingsOutputMessage, AWSError>;
763
+ /**
764
+ * Get the resource policy for a specified resource.
765
+ */
766
+ getResourcePolicy(params: Redshift.Types.GetResourcePolicyMessage, callback?: (err: AWSError, data: Redshift.Types.GetResourcePolicyResult) => void): Request<Redshift.Types.GetResourcePolicyResult, AWSError>;
767
+ /**
768
+ * Get the resource policy for a specified resource.
769
+ */
770
+ getResourcePolicy(callback?: (err: AWSError, data: Redshift.Types.GetResourcePolicyResult) => void): Request<Redshift.Types.GetResourcePolicyResult, AWSError>;
747
771
  /**
748
772
  * This operation is retired. Calling this operation does not change AQUA configuration. Amazon Redshift automatically determines whether to use AQUA (Advanced Query Accelerator).
749
773
  */
@@ -896,6 +920,14 @@ declare class Redshift extends Service {
896
920
  * Allows you to purchase reserved nodes. Amazon Redshift offers a predefined set of reserved node offerings. You can purchase one or more of the offerings. You can call the DescribeReservedNodeOfferings API to obtain the available reserved node offerings. You can call this API by providing a specific reserved node offering and the number of nodes you want to reserve. For more information about reserved node offerings, go to Purchasing Reserved Nodes in the Amazon Redshift Cluster Management Guide.
897
921
  */
898
922
  purchaseReservedNodeOffering(callback?: (err: AWSError, data: Redshift.Types.PurchaseReservedNodeOfferingResult) => void): Request<Redshift.Types.PurchaseReservedNodeOfferingResult, AWSError>;
923
+ /**
924
+ * Updates the resource policy for a specified resource.
925
+ */
926
+ putResourcePolicy(params: Redshift.Types.PutResourcePolicyMessage, callback?: (err: AWSError, data: Redshift.Types.PutResourcePolicyResult) => void): Request<Redshift.Types.PutResourcePolicyResult, AWSError>;
927
+ /**
928
+ * Updates the resource policy for a specified resource.
929
+ */
930
+ putResourcePolicy(callback?: (err: AWSError, data: Redshift.Types.PutResourcePolicyResult) => void): Request<Redshift.Types.PutResourcePolicyResult, AWSError>;
899
931
  /**
900
932
  * Reboots a cluster. This action is taken as soon as possible. It results in a momentary outage to the cluster, during which the cluster status is set to rebooting. A cluster event is created when the reboot is completed. Any pending cluster modifications (see ModifyCluster) are applied at this reboot. For more information about managing clusters, go to Amazon Redshift Clusters in the Amazon Redshift Cluster Management Guide.
901
933
  */
@@ -2573,6 +2605,12 @@ declare namespace Redshift {
2573
2605
  */
2574
2606
  HsmConfigurationIdentifier: String;
2575
2607
  }
2608
+ export interface DeleteResourcePolicyMessage {
2609
+ /**
2610
+ * The Amazon Resource Name (ARN) of the resource of which its resource policy is deleted.
2611
+ */
2612
+ ResourceArn: String;
2613
+ }
2576
2614
  export interface DeleteScheduledActionMessage {
2577
2615
  /**
2578
2616
  * The name of the scheduled action to delete.
@@ -3096,6 +3134,24 @@ declare namespace Redshift {
3096
3134
  */
3097
3135
  TagValues?: TagValueList;
3098
3136
  }
3137
+ export interface DescribeInboundIntegrationsMessage {
3138
+ /**
3139
+ * The Amazon Resource Name (ARN) of the inbound integration.
3140
+ */
3141
+ IntegrationArn?: String;
3142
+ /**
3143
+ * The Amazon Resource Name (ARN) of the target of an inbound integration.
3144
+ */
3145
+ TargetArn?: String;
3146
+ /**
3147
+ * The maximum number of response records to return in each call. If the number of remaining response records exceeds the specified MaxRecords value, a value is returned in a marker field of the response. You can retrieve the next set of records by retrying the command with the returned marker value. Default: 100 Constraints: minimum 20, maximum 100.
3148
+ */
3149
+ MaxRecords?: IntegerOptional;
3150
+ /**
3151
+ * An optional parameter that specifies the starting point to return a set of response records. When the results of a DescribeInboundIntegrations request exceed the value specified in MaxRecords, Amazon Web Services returns a value in the Marker field of the response. You can retrieve the next set of response records by providing the returned marker value in the Marker parameter and retrying the request.
3152
+ */
3153
+ Marker?: String;
3154
+ }
3099
3155
  export interface DescribeLoggingStatusMessage {
3100
3156
  /**
3101
3157
  * The identifier of the cluster from which to get the logging status. Example: examplecluster
@@ -3870,6 +3926,18 @@ declare namespace Redshift {
3870
3926
  */
3871
3927
  ReservedNodeOfferings?: ReservedNodeOfferingList;
3872
3928
  }
3929
+ export interface GetResourcePolicyMessage {
3930
+ /**
3931
+ * The Amazon Resource Name (ARN) of the resource of which its resource policy is fetched.
3932
+ */
3933
+ ResourceArn: String;
3934
+ }
3935
+ export interface GetResourcePolicyResult {
3936
+ /**
3937
+ * The content of the resource policy.
3938
+ */
3939
+ ResourcePolicy?: ResourcePolicy;
3940
+ }
3873
3941
  export interface HsmClientCertificate {
3874
3942
  /**
3875
3943
  * The identifier of the HSM client certificate.
@@ -3961,8 +4029,56 @@ declare namespace Redshift {
3961
4029
  export type ImportTablesCompleted = String[];
3962
4030
  export type ImportTablesInProgress = String[];
3963
4031
  export type ImportTablesNotStarted = String[];
4032
+ export interface InboundIntegration {
4033
+ /**
4034
+ * The Amazon Resource Name (ARN) of an inbound integration.
4035
+ */
4036
+ IntegrationArn?: String;
4037
+ /**
4038
+ * The Amazon Resource Name (ARN) of the source of an inbound integration.
4039
+ */
4040
+ SourceArn?: String;
4041
+ /**
4042
+ * The Amazon Resource Name (ARN) of the target of an inbound integration.
4043
+ */
4044
+ TargetArn?: String;
4045
+ /**
4046
+ * The status of an inbound integration.
4047
+ */
4048
+ Status?: ZeroETLIntegrationStatus;
4049
+ /**
4050
+ * The outstanding errors of an inbound integration. Each item is an "IntegrationError". This is null if there is no error.
4051
+ */
4052
+ Errors?: IntegrationErrorList;
4053
+ /**
4054
+ * The creation time of an inbound integration.
4055
+ */
4056
+ CreateTime?: TStamp;
4057
+ }
4058
+ export type InboundIntegrationList = InboundIntegration[];
4059
+ export interface InboundIntegrationsMessage {
4060
+ /**
4061
+ * A value that indicates the starting point for the next set of response records in a subsequent request. If a value is returned in a response, you can retrieve the next set of records by providing this returned marker value in the Marker parameter and retrying the command. If the Marker field is empty, all response records have been retrieved for the request.
4062
+ */
4063
+ Marker?: String;
4064
+ /**
4065
+ * A list of InboundIntegration instances.
4066
+ */
4067
+ InboundIntegrations?: InboundIntegrationList;
4068
+ }
3964
4069
  export type Integer = number;
3965
4070
  export type IntegerOptional = number;
4071
+ export interface IntegrationError {
4072
+ /**
4073
+ * The error code of an inbound integration error.
4074
+ */
4075
+ ErrorCode: String;
4076
+ /**
4077
+ * The error message of an inbound integration error.
4078
+ */
4079
+ ErrorMessage?: String;
4080
+ }
4081
+ export type IntegrationErrorList = IntegrationError[];
3966
4082
  export type LogDestinationType = "s3"|"cloudwatch"|string;
3967
4083
  export type LogTypeList = String[];
3968
4084
  export interface LoggingStatus {
@@ -4701,6 +4817,22 @@ declare namespace Redshift {
4701
4817
  export interface PurchaseReservedNodeOfferingResult {
4702
4818
  ReservedNode?: ReservedNode;
4703
4819
  }
4820
+ export interface PutResourcePolicyMessage {
4821
+ /**
4822
+ * The Amazon Resource Name (ARN) of the resource of which its resource policy is updated.
4823
+ */
4824
+ ResourceArn: String;
4825
+ /**
4826
+ * The content of the resource policy being updated.
4827
+ */
4828
+ Policy: String;
4829
+ }
4830
+ export interface PutResourcePolicyResult {
4831
+ /**
4832
+ * The content of the updated resource policy.
4833
+ */
4834
+ ResourcePolicy?: ResourcePolicy;
4835
+ }
4704
4836
  export interface RebootClusterMessage {
4705
4837
  /**
4706
4838
  * The cluster identifier.
@@ -5015,6 +5147,16 @@ declare namespace Redshift {
5015
5147
  */
5016
5148
  DataTransferProgressPercent?: DoubleOptional;
5017
5149
  }
5150
+ export interface ResourcePolicy {
5151
+ /**
5152
+ * The resources that a policy is attached to.
5153
+ */
5154
+ ResourceArn?: String;
5155
+ /**
5156
+ * The content of a resource policy.
5157
+ */
5158
+ Policy?: String;
5159
+ }
5018
5160
  export type RestorableNodeTypeList = String[];
5019
5161
  export interface RestoreFromClusterSnapshotMessage {
5020
5162
  /**
@@ -5919,6 +6061,7 @@ declare namespace Redshift {
5919
6061
  Status?: String;
5920
6062
  }
5921
6063
  export type VpcSecurityGroupMembershipList = VpcSecurityGroupMembership[];
6064
+ export type ZeroETLIntegrationStatus = "creating"|"active"|"modifying"|"failed"|"deleting"|"syncing"|"needs_attention"|string;
5922
6065
  /**
5923
6066
  * A string in YYYY-MM-DD format that represents the latest possible API version that can be used in this service. Specify 'latest' to use the latest possible version.
5924
6067
  */
@@ -85,19 +85,19 @@ declare class SageMaker extends Service {
85
85
  */
86
86
  createArtifact(callback?: (err: AWSError, data: SageMaker.Types.CreateArtifactResponse) => void): Request<SageMaker.Types.CreateArtifactResponse, AWSError>;
87
87
  /**
88
- * Creates an Autopilot job also referred to as Autopilot experiment or AutoML job. We recommend using the new versions CreateAutoMLJobV2 and DescribeAutoMLJobV2, which offer backward compatibility. CreateAutoMLJobV2 can manage tabular problem types identical to those of its previous version CreateAutoMLJob, as well as time-series forecasting, and non-tabular problem types such as image or text classification. Find guidelines about how to migrate a CreateAutoMLJob to CreateAutoMLJobV2 in Migrate a CreateAutoMLJob to CreateAutoMLJobV2. You can find the best-performing model after you run an AutoML job by calling DescribeAutoMLJobV2 (recommended) or DescribeAutoMLJob.
88
+ * Creates an Autopilot job also referred to as Autopilot experiment or AutoML job. We recommend using the new versions CreateAutoMLJobV2 and DescribeAutoMLJobV2, which offer backward compatibility. CreateAutoMLJobV2 can manage tabular problem types identical to those of its previous version CreateAutoMLJob, as well as time-series forecasting, non-tabular problem types such as image or text classification, and text generation (LLMs fine-tuning). Find guidelines about how to migrate a CreateAutoMLJob to CreateAutoMLJobV2 in Migrate a CreateAutoMLJob to CreateAutoMLJobV2. You can find the best-performing model after you run an AutoML job by calling DescribeAutoMLJobV2 (recommended) or DescribeAutoMLJob.
89
89
  */
90
90
  createAutoMLJob(params: SageMaker.Types.CreateAutoMLJobRequest, callback?: (err: AWSError, data: SageMaker.Types.CreateAutoMLJobResponse) => void): Request<SageMaker.Types.CreateAutoMLJobResponse, AWSError>;
91
91
  /**
92
- * Creates an Autopilot job also referred to as Autopilot experiment or AutoML job. We recommend using the new versions CreateAutoMLJobV2 and DescribeAutoMLJobV2, which offer backward compatibility. CreateAutoMLJobV2 can manage tabular problem types identical to those of its previous version CreateAutoMLJob, as well as time-series forecasting, and non-tabular problem types such as image or text classification. Find guidelines about how to migrate a CreateAutoMLJob to CreateAutoMLJobV2 in Migrate a CreateAutoMLJob to CreateAutoMLJobV2. You can find the best-performing model after you run an AutoML job by calling DescribeAutoMLJobV2 (recommended) or DescribeAutoMLJob.
92
+ * Creates an Autopilot job also referred to as Autopilot experiment or AutoML job. We recommend using the new versions CreateAutoMLJobV2 and DescribeAutoMLJobV2, which offer backward compatibility. CreateAutoMLJobV2 can manage tabular problem types identical to those of its previous version CreateAutoMLJob, as well as time-series forecasting, non-tabular problem types such as image or text classification, and text generation (LLMs fine-tuning). Find guidelines about how to migrate a CreateAutoMLJob to CreateAutoMLJobV2 in Migrate a CreateAutoMLJob to CreateAutoMLJobV2. You can find the best-performing model after you run an AutoML job by calling DescribeAutoMLJobV2 (recommended) or DescribeAutoMLJob.
93
93
  */
94
94
  createAutoMLJob(callback?: (err: AWSError, data: SageMaker.Types.CreateAutoMLJobResponse) => void): Request<SageMaker.Types.CreateAutoMLJobResponse, AWSError>;
95
95
  /**
96
- * Creates an Autopilot job also referred to as Autopilot experiment or AutoML job V2. CreateAutoMLJobV2 and DescribeAutoMLJobV2 are new versions of CreateAutoMLJob and DescribeAutoMLJob which offer backward compatibility. CreateAutoMLJobV2 can manage tabular problem types identical to those of its previous version CreateAutoMLJob, as well as time-series forecasting, and non-tabular problem types such as image or text classification. Find guidelines about how to migrate a CreateAutoMLJob to CreateAutoMLJobV2 in Migrate a CreateAutoMLJob to CreateAutoMLJobV2. For the list of available problem types supported by CreateAutoMLJobV2, see AutoMLProblemTypeConfig. You can find the best-performing model after you run an AutoML job V2 by calling DescribeAutoMLJobV2.
96
+ * Creates an Autopilot job also referred to as Autopilot experiment or AutoML job V2. CreateAutoMLJobV2 and DescribeAutoMLJobV2 are new versions of CreateAutoMLJob and DescribeAutoMLJob which offer backward compatibility. CreateAutoMLJobV2 can manage tabular problem types identical to those of its previous version CreateAutoMLJob, as well as time-series forecasting, non-tabular problem types such as image or text classification, and text generation (LLMs fine-tuning). Find guidelines about how to migrate a CreateAutoMLJob to CreateAutoMLJobV2 in Migrate a CreateAutoMLJob to CreateAutoMLJobV2. For the list of available problem types supported by CreateAutoMLJobV2, see AutoMLProblemTypeConfig. You can find the best-performing model after you run an AutoML job V2 by calling DescribeAutoMLJobV2.
97
97
  */
98
98
  createAutoMLJobV2(params: SageMaker.Types.CreateAutoMLJobV2Request, callback?: (err: AWSError, data: SageMaker.Types.CreateAutoMLJobV2Response) => void): Request<SageMaker.Types.CreateAutoMLJobV2Response, AWSError>;
99
99
  /**
100
- * Creates an Autopilot job also referred to as Autopilot experiment or AutoML job V2. CreateAutoMLJobV2 and DescribeAutoMLJobV2 are new versions of CreateAutoMLJob and DescribeAutoMLJob which offer backward compatibility. CreateAutoMLJobV2 can manage tabular problem types identical to those of its previous version CreateAutoMLJob, as well as time-series forecasting, and non-tabular problem types such as image or text classification. Find guidelines about how to migrate a CreateAutoMLJob to CreateAutoMLJobV2 in Migrate a CreateAutoMLJob to CreateAutoMLJobV2. For the list of available problem types supported by CreateAutoMLJobV2, see AutoMLProblemTypeConfig. You can find the best-performing model after you run an AutoML job V2 by calling DescribeAutoMLJobV2.
100
+ * Creates an Autopilot job also referred to as Autopilot experiment or AutoML job V2. CreateAutoMLJobV2 and DescribeAutoMLJobV2 are new versions of CreateAutoMLJob and DescribeAutoMLJob which offer backward compatibility. CreateAutoMLJobV2 can manage tabular problem types identical to those of its previous version CreateAutoMLJob, as well as time-series forecasting, non-tabular problem types such as image or text classification, and text generation (LLMs fine-tuning). Find guidelines about how to migrate a CreateAutoMLJob to CreateAutoMLJobV2 in Migrate a CreateAutoMLJob to CreateAutoMLJobV2. For the list of available problem types supported by CreateAutoMLJobV2, see AutoMLProblemTypeConfig. You can find the best-performing model after you run an AutoML job V2 by calling DescribeAutoMLJobV2.
101
101
  */
102
102
  createAutoMLJobV2(callback?: (err: AWSError, data: SageMaker.Types.CreateAutoMLJobV2Response) => void): Request<SageMaker.Types.CreateAutoMLJobV2Response, AWSError>;
103
103
  /**
@@ -3244,7 +3244,7 @@ declare namespace SageMaker {
3244
3244
  */
3245
3245
  ChannelType?: AutoMLChannelType;
3246
3246
  /**
3247
- * The content type of the data from the input source. The following are the allowed content types for different problems: For tabular problem types: text/csv;header=present or x-application/vnd.amazon+parquet. The default value is text/csv;header=present. For image classification: image/png, image/jpeg, or image/*. The default value is image/*. For text classification: text/csv;header=present or x-application/vnd.amazon+parquet. The default value is text/csv;header=present. For time-series forecasting: text/csv;header=present or x-application/vnd.amazon+parquet. The default value is text/csv;header=present.
3247
+ * The content type of the data from the input source. The following are the allowed content types for different problems: For tabular problem types: text/csv;header=present or x-application/vnd.amazon+parquet. The default value is text/csv;header=present. For image classification: image/png, image/jpeg, or image/*. The default value is image/*. For text classification: text/csv;header=present or x-application/vnd.amazon+parquet. The default value is text/csv;header=present. For time-series forecasting: text/csv;header=present or x-application/vnd.amazon+parquet. The default value is text/csv;header=present. For text generation (LLMs fine-tuning): text/csv;header=present or x-application/vnd.amazon+parquet. The default value is text/csv;header=present.
3248
3248
  */
3249
3249
  ContentType?: ContentType;
3250
3250
  /**
@@ -3258,7 +3258,7 @@ declare namespace SageMaker {
3258
3258
  }
3259
3259
  export interface AutoMLJobCompletionCriteria {
3260
3260
  /**
3261
- * The maximum number of times a training job is allowed to run. For text and image classification, as well as time-series forecasting problem types, the supported value is 1. For tabular problem types, the maximum value is 750.
3261
+ * The maximum number of times a training job is allowed to run. For text and image classification, time-series forecasting, as well as text generation (LLMs fine-tuning) problem types, the supported value is 1. For tabular problem types, the maximum value is 750.
3262
3262
  */
3263
3263
  MaxCandidates?: MaxCandidates;
3264
3264
  /**
@@ -3296,7 +3296,7 @@ declare namespace SageMaker {
3296
3296
  export type AutoMLJobName = string;
3297
3297
  export interface AutoMLJobObjective {
3298
3298
  /**
3299
- * The name of the objective metric used to measure the predictive quality of a machine learning system. During training, the model's parameters are updated iteratively to optimize its performance based on the feedback provided by the objective metric when evaluating the model on the validation dataset. For the list of all available metrics supported by Autopilot, see Autopilot metrics. If you do not specify a metric explicitly, the default behavior is to automatically use: For tabular problem types: Regression: MSE. Binary classification: F1. Multiclass classification: Accuracy. For image or text classification problem types: Accuracy For time-series forecasting problem types: AverageWeightedQuantileLoss
3299
+ * The name of the objective metric used to measure the predictive quality of a machine learning system. During training, the model's parameters are updated iteratively to optimize its performance based on the feedback provided by the objective metric when evaluating the model on the validation dataset. The list of available metrics supported by Autopilot and the default metric applied when you do not specify a metric name explicitly depend on the problem type. For tabular problem types: List of available metrics: Regression: InferenceLatency, MAE, MSE, R2, RMSE Binary classification: Accuracy, AUC, BalancedAccuracy, F1, InferenceLatency, LogLoss, Precision, Recall Multiclass classification: Accuracy, BalancedAccuracy, F1macro, InferenceLatency, LogLoss, PrecisionMacro, RecallMacro For a description of each metric, see Autopilot metrics for classification and regression. Default objective metrics: Regression: MSE. Binary classification: F1. Multiclass classification: Accuracy. For image or text classification problem types: List of available metrics: Accuracy For a description of each metric, see Autopilot metrics for text and image classification. Default objective metrics: Accuracy For time-series forecasting problem types: List of available metrics: RMSE, wQL, Average wQL, MASE, MAPE, WAPE For a description of each metric, see Autopilot metrics for time-series forecasting. Default objective metrics: AverageWeightedQuantileLoss For text generation problem types (LLMs fine-tuning): Fine-tuning language models in Autopilot does not require setting the AutoMLJobObjective field. Autopilot fine-tunes LLMs without requiring multiple candidates to be trained and evaluated. Instead, using your dataset, Autopilot directly fine-tunes your target model to enhance a default objective metric, the cross-entropy loss. After fine-tuning a language model, you can evaluate the quality of its generated text using different metrics. For a list of the available metrics, see Metrics for fine-tuning LLMs in Autopilot.
3300
3300
  */
3301
3301
  MetricName: AutoMLMetricEnum;
3302
3302
  }
@@ -3350,7 +3350,7 @@ declare namespace SageMaker {
3350
3350
  }
3351
3351
  export type AutoMLMaxResults = number;
3352
3352
  export type AutoMLMetricEnum = "Accuracy"|"MSE"|"F1"|"F1macro"|"AUC"|"RMSE"|"MAE"|"R2"|"BalancedAccuracy"|"Precision"|"PrecisionMacro"|"Recall"|"RecallMacro"|"MAPE"|"MASE"|"WAPE"|"AverageWeightedQuantileLoss"|string;
3353
- export type AutoMLMetricExtendedEnum = "Accuracy"|"MSE"|"F1"|"F1macro"|"AUC"|"RMSE"|"MAE"|"R2"|"BalancedAccuracy"|"Precision"|"PrecisionMacro"|"Recall"|"RecallMacro"|"LogLoss"|"InferenceLatency"|"MAPE"|"MASE"|"WAPE"|"AverageWeightedQuantileLoss"|string;
3353
+ export type AutoMLMetricExtendedEnum = "Accuracy"|"MSE"|"F1"|"F1macro"|"AUC"|"RMSE"|"MAE"|"R2"|"BalancedAccuracy"|"Precision"|"PrecisionMacro"|"Recall"|"RecallMacro"|"LogLoss"|"InferenceLatency"|"MAPE"|"MASE"|"WAPE"|"AverageWeightedQuantileLoss"|"Rouge1"|"Rouge2"|"RougeL"|"RougeLSum"|"Perplexity"|"ValidationLoss"|"TrainingLoss"|string;
3354
3354
  export type AutoMLMode = "AUTO"|"ENSEMBLING"|"HYPERPARAMETER_TUNING"|string;
3355
3355
  export type AutoMLNameContains = string;
3356
3356
  export interface AutoMLOutputDataConfig {
@@ -3380,20 +3380,28 @@ declare namespace SageMaker {
3380
3380
  */
3381
3381
  TextClassificationJobConfig?: TextClassificationJobConfig;
3382
3382
  /**
3383
- * Settings used to configure an AutoML job V2 for a tabular problem type (regression, classification).
3383
+ * Settings used to configure an AutoML job V2 for the tabular problem type (regression, classification).
3384
3384
  */
3385
3385
  TabularJobConfig?: TabularJobConfig;
3386
3386
  /**
3387
- * Settings used to configure an AutoML job V2 for a time-series forecasting problem type.
3387
+ * Settings used to configure an AutoML job V2 for the time-series forecasting problem type.
3388
3388
  */
3389
3389
  TimeSeriesForecastingJobConfig?: TimeSeriesForecastingJobConfig;
3390
+ /**
3391
+ * Settings used to configure an AutoML job V2 for the text generation (LLMs fine-tuning) problem type. The text generation models that support fine-tuning in Autopilot are currently accessible exclusively in regions supported by Canvas. Refer to the documentation of Canvas for the full list of its supported Regions.
3392
+ */
3393
+ TextGenerationJobConfig?: TextGenerationJobConfig;
3390
3394
  }
3391
- export type AutoMLProblemTypeConfigName = "ImageClassification"|"TextClassification"|"Tabular"|"TimeSeriesForecasting"|string;
3395
+ export type AutoMLProblemTypeConfigName = "ImageClassification"|"TextClassification"|"Tabular"|"TimeSeriesForecasting"|"TextGeneration"|string;
3392
3396
  export interface AutoMLProblemTypeResolvedAttributes {
3393
3397
  /**
3394
- * Defines the resolved attributes for the TABULAR problem type.
3398
+ * The resolved attributes for the tabular problem type.
3395
3399
  */
3396
3400
  TabularResolvedAttributes?: TabularResolvedAttributes;
3401
+ /**
3402
+ * The resolved attributes for the text generation problem type.
3403
+ */
3404
+ TextGenerationResolvedAttributes?: TextGenerationResolvedAttributes;
3397
3405
  }
3398
3406
  export type AutoMLProcessingUnit = "CPU"|"GPU"|string;
3399
3407
  export interface AutoMLResolvedAttributes {
@@ -3457,6 +3465,7 @@ declare namespace SageMaker {
3457
3465
  export type AutotuneMode = "Enabled"|string;
3458
3466
  export type AwsManagedHumanLoopRequestSource = "AWS/Rekognition/DetectModerationLabels/Image/V3"|"AWS/Textract/AnalyzeDocument/Forms/V1"|string;
3459
3467
  export type BacktestResultsLocation = string;
3468
+ export type BaseModelName = string;
3460
3469
  export interface BatchDataCaptureConfig {
3461
3470
  /**
3462
3471
  * The Amazon S3 location being used to capture the data.
@@ -4492,7 +4501,7 @@ declare namespace SageMaker {
4492
4501
  */
4493
4502
  AutoMLJobName: AutoMLJobName;
4494
4503
  /**
4495
- * An array of channel objects describing the input data and their location. Each channel is a named input source. Similar to the InputDataConfig attribute in the CreateAutoMLJob input parameters. The supported formats depend on the problem type: For tabular problem types: S3Prefix, ManifestFile. For image classification: S3Prefix, ManifestFile, AugmentedManifestFile. For text classification: S3Prefix. For time-series forecasting: S3Prefix.
4504
+ * An array of channel objects describing the input data and their location. Each channel is a named input source. Similar to the InputDataConfig attribute in the CreateAutoMLJob input parameters. The supported formats depend on the problem type: For tabular problem types: S3Prefix, ManifestFile. For image classification: S3Prefix, ManifestFile, AugmentedManifestFile. For text classification: S3Prefix. For time-series forecasting: S3Prefix. For text generation (LLMs fine-tuning): S3Prefix.
4496
4505
  */
4497
4506
  AutoMLJobInputDataConfig: AutoMLJobInputDataConfig;
4498
4507
  /**
@@ -4516,7 +4525,7 @@ declare namespace SageMaker {
4516
4525
  */
4517
4526
  SecurityConfig?: AutoMLSecurityConfig;
4518
4527
  /**
4519
- * Specifies a metric to minimize or maximize as the objective of a job. If not specified, the default objective metric depends on the problem type. For the list of default values per problem type, see AutoMLJobObjective. For tabular problem types, you must either provide both the AutoMLJobObjective and indicate the type of supervised learning problem in AutoMLProblemTypeConfig (TabularJobConfig.ProblemType), or none at all.
4528
+ * Specifies a metric to minimize or maximize as the objective of a job. If not specified, the default objective metric depends on the problem type. For the list of default values per problem type, see AutoMLJobObjective. For tabular problem types: You must either provide both the AutoMLJobObjective and indicate the type of supervised learning problem in AutoMLProblemTypeConfig (TabularJobConfig.ProblemType), or none at all. For text generation problem types (LLMs fine-tuning): Fine-tuning language models in Autopilot does not require setting the AutoMLJobObjective field. Autopilot fine-tunes LLMs without requiring multiple candidates to be trained and evaluated. Instead, using your dataset, Autopilot directly fine-tunes your target model to enhance a default objective metric, the cross-entropy loss. After fine-tuning a language model, you can evaluate the quality of its generated text using different metrics. For a list of the available metrics, see Metrics for fine-tuning LLMs in Autopilot.
4520
4529
  */
4521
4530
  AutoMLJobObjective?: AutoMLJobObjective;
4522
4531
  /**
@@ -20846,6 +20855,19 @@ declare namespace SageMaker {
20846
20855
  */
20847
20856
  TargetLabelColumn: TargetLabelColumn;
20848
20857
  }
20858
+ export interface TextGenerationJobConfig {
20859
+ CompletionCriteria?: AutoMLJobCompletionCriteria;
20860
+ /**
20861
+ * The name of the base model to fine-tune. Autopilot supports fine-tuning a variety of large language models. For information on the list of supported models, see Text generation models supporting fine-tuning in Autopilot. If no BaseModelName is provided, the default model used is Falcon-7B-Instruct.
20862
+ */
20863
+ BaseModelName?: BaseModelName;
20864
+ }
20865
+ export interface TextGenerationResolvedAttributes {
20866
+ /**
20867
+ * The name of the base model to fine-tune.
20868
+ */
20869
+ BaseModelName?: BaseModelName;
20870
+ }
20849
20871
  export type ThingName = string;
20850
20872
  export interface TimeSeriesConfig {
20851
20873
  /**
@@ -308,11 +308,11 @@ declare class SNS extends Service {
308
308
  */
309
309
  setTopicAttributes(callback?: (err: AWSError, data: {}) => void): Request<{}, AWSError>;
310
310
  /**
311
- * Subscribes an endpoint to an Amazon SNS topic. If the endpoint type is HTTP/S or email, or if the endpoint and the topic are not in the same Amazon Web Services account, the endpoint owner must run the ConfirmSubscription action to confirm the subscription. You call the ConfirmSubscription action with the token from the subscription response. Confirmation tokens are valid for three days. This action is throttled at 100 transactions per second (TPS).
311
+ * Subscribes an endpoint to an Amazon SNS topic. If the endpoint type is HTTP/S or email, or if the endpoint and the topic are not in the same Amazon Web Services account, the endpoint owner must run the ConfirmSubscription action to confirm the subscription. You call the ConfirmSubscription action with the token from the subscription response. Confirmation tokens are valid for two days. This action is throttled at 100 transactions per second (TPS).
312
312
  */
313
313
  subscribe(params: SNS.Types.SubscribeInput, callback?: (err: AWSError, data: SNS.Types.SubscribeResponse) => void): Request<SNS.Types.SubscribeResponse, AWSError>;
314
314
  /**
315
- * Subscribes an endpoint to an Amazon SNS topic. If the endpoint type is HTTP/S or email, or if the endpoint and the topic are not in the same Amazon Web Services account, the endpoint owner must run the ConfirmSubscription action to confirm the subscription. You call the ConfirmSubscription action with the token from the subscription response. Confirmation tokens are valid for three days. This action is throttled at 100 transactions per second (TPS).
315
+ * Subscribes an endpoint to an Amazon SNS topic. If the endpoint type is HTTP/S or email, or if the endpoint and the topic are not in the same Amazon Web Services account, the endpoint owner must run the ConfirmSubscription action to confirm the subscription. You call the ConfirmSubscription action with the token from the subscription response. Confirmation tokens are valid for two days. This action is throttled at 100 transactions per second (TPS).
316
316
  */
317
317
  subscribe(callback?: (err: AWSError, data: SNS.Types.SubscribeResponse) => void): Request<SNS.Types.SubscribeResponse, AWSError>;
318
318
  /**
@@ -483,7 +483,7 @@ declare namespace SNS {
483
483
  */
484
484
  Name: topicName;
485
485
  /**
486
- * A map of attributes with their corresponding values. The following lists the names, descriptions, and values of the special request parameters that the CreateTopic action uses: DeliveryPolicy – The policy that defines how Amazon SNS retries failed deliveries to HTTP/S endpoints. DisplayName – The display name to use for a topic with SMS subscriptions. FifoTopic – Set to true to create a FIFO topic. Policy – The policy that defines who can access your topic. By default, only the topic owner can publish or subscribe to the topic. SignatureVersion – The signature version corresponds to the hashing algorithm used while creating the signature of the notifications, subscription confirmations, or unsubscribe confirmation messages sent by Amazon SNS. By default, SignatureVersion is set to 1. TracingConfig – Tracing mode of an Amazon SNS topic. By default TracingConfig is set to PassThrough, and the topic passes through the tracing header it receives from an Amazon SNS publisher to its subscriptions. If set to Active, Amazon SNS will vend X-Ray segment data to topic owner account if the sampled flag in the tracing header is true. This is only supported on standard topics. The following attribute applies only to server-side encryption: KmsMasterKeyId – The ID of an Amazon Web Services managed customer master key (CMK) for Amazon SNS or a custom CMK. For more information, see Key Terms. For more examples, see KeyId in the Key Management Service API Reference. The following attributes apply only to FIFO topics: FifoTopicWhen this is set to true, a FIFO topic is created. ContentBasedDeduplication – Enables content-based deduplication for FIFO topics. By default, ContentBasedDeduplication is set to false. If you create a FIFO topic and this attribute is false, you must specify a value for the MessageDeduplicationId parameter for the Publish action. When you set ContentBasedDeduplication to true, Amazon SNS uses a SHA-256 hash to generate the MessageDeduplicationId using the body of the message (but not the attributes of the message). (Optional) To override the generated value, you can specify a value for the MessageDeduplicationId parameter for the Publish action.
486
+ * A map of attributes with their corresponding values. The following lists the names, descriptions, and values of the special request parameters that the CreateTopic action uses: DeliveryPolicy – The policy that defines how Amazon SNS retries failed deliveries to HTTP/S endpoints. DisplayName – The display name to use for a topic with SMS subscriptions. FifoTopic – Set to true to create a FIFO topic. Policy – The policy that defines who can access your topic. By default, only the topic owner can publish or subscribe to the topic. SignatureVersion – The signature version corresponds to the hashing algorithm used while creating the signature of the notifications, subscription confirmations, or unsubscribe confirmation messages sent by Amazon SNS. By default, SignatureVersion is set to 1. TracingConfig – Tracing mode of an Amazon SNS topic. By default TracingConfig is set to PassThrough, and the topic passes through the tracing header it receives from an Amazon SNS publisher to its subscriptions. If set to Active, Amazon SNS will vend X-Ray segment data to topic owner account if the sampled flag in the tracing header is true. This is only supported on standard topics. The following attribute applies only to server-side encryption: KmsMasterKeyId – The ID of an Amazon Web Services managed customer master key (CMK) for Amazon SNS or a custom CMK. For more information, see Key Terms. For more examples, see KeyId in the Key Management Service API Reference. The following attributes apply only to FIFO topics: ArchivePolicyAdds or updates an inline policy document to archive messages stored in the specified Amazon SNS topic. BeginningArchiveTime – The earliest starting point at which a message in the topic’s archive can be replayed from. This point in time is based on the configured message retention period set by the topic’s message archiving policy. ContentBasedDeduplication – Enables content-based deduplication for FIFO topics. By default, ContentBasedDeduplication is set to false. If you create a FIFO topic and this attribute is false, you must specify a value for the MessageDeduplicationId parameter for the Publish action. When you set ContentBasedDeduplication to true, Amazon SNS uses a SHA-256 hash to generate the MessageDeduplicationId using the body of the message (but not the attributes of the message). (Optional) To override the generated value, you can specify a value for the MessageDeduplicationId parameter for the Publish action.
487
487
  */
488
488
  Attributes?: TopicAttributesMap;
489
489
  /**
@@ -1066,7 +1066,7 @@ declare namespace SNS {
1066
1066
  */
1067
1067
  Endpoint?: endpoint;
1068
1068
  /**
1069
- * A map of attributes with their corresponding values. The following lists the names, descriptions, and values of the special request parameters that the Subscribe action uses: DeliveryPolicy – The policy that defines how Amazon SNS retries failed deliveries to HTTP/S endpoints. FilterPolicy – The simple JSON object that lets your subscriber receive only a subset of messages, rather than receiving every message published to the topic. FilterPolicyScope – This attribute lets you choose the filtering scope by using one of the following string value types: MessageAttributes (default) – The filter is applied on the message attributes. MessageBody – The filter is applied on the message body. RawMessageDelivery – When set to true, enables raw message delivery to Amazon SQS or HTTP/S endpoints. This eliminates the need for the endpoints to process JSON formatting, which is otherwise created for Amazon SNS metadata. RedrivePolicy – When specified, sends undeliverable messages to the specified Amazon SQS dead-letter queue. Messages that can't be delivered due to client errors (for example, when the subscribed endpoint is unreachable) or server errors (for example, when the service that powers the subscribed endpoint becomes unavailable) are held in the dead-letter queue for further analysis or reprocessing. The following attribute applies only to Amazon Kinesis Data Firehose delivery stream subscriptions: SubscriptionRoleArn – The ARN of the IAM role that has the following: Permission to write to the Kinesis Data Firehose delivery stream Amazon SNS listed as a trusted entity Specifying a valid ARN for this attribute is required for Kinesis Data Firehose delivery stream subscriptions. For more information, see Fanout to Kinesis Data Firehose delivery streams in the Amazon SNS Developer Guide.
1069
+ * A map of attributes with their corresponding values. The following lists the names, descriptions, and values of the special request parameters that the Subscribe action uses: DeliveryPolicy – The policy that defines how Amazon SNS retries failed deliveries to HTTP/S endpoints. FilterPolicy – The simple JSON object that lets your subscriber receive only a subset of messages, rather than receiving every message published to the topic. FilterPolicyScope – This attribute lets you choose the filtering scope by using one of the following string value types: MessageAttributes (default) – The filter is applied on the message attributes. MessageBody – The filter is applied on the message body. RawMessageDelivery – When set to true, enables raw message delivery to Amazon SQS or HTTP/S endpoints. This eliminates the need for the endpoints to process JSON formatting, which is otherwise created for Amazon SNS metadata. RedrivePolicy – When specified, sends undeliverable messages to the specified Amazon SQS dead-letter queue. Messages that can't be delivered due to client errors (for example, when the subscribed endpoint is unreachable) or server errors (for example, when the service that powers the subscribed endpoint becomes unavailable) are held in the dead-letter queue for further analysis or reprocessing. The following attribute applies only to Amazon Kinesis Data Firehose delivery stream subscriptions: SubscriptionRoleArn – The ARN of the IAM role that has the following: Permission to write to the Kinesis Data Firehose delivery stream Amazon SNS listed as a trusted entity Specifying a valid ARN for this attribute is required for Kinesis Data Firehose delivery stream subscriptions. For more information, see Fanout to Kinesis Data Firehose delivery streams in the Amazon SNS Developer Guide. The following attributes apply only to FIFO topics: ReplayPolicy – Adds or updates an inline policy document for a subscription to replay messages stored in the specified Amazon SNS topic. ReplayStatus – Retrieves the status of the subscription message replay, which can be one of the following: Completed – The replay has successfully redelivered all messages, and is now delivering newly published messages. If an ending point was specified in the ReplayPolicy then the subscription will no longer receive newly published messages. In progress – The replay is currently replaying the selected messages. Failed – The replay was unable to complete. Pending – The default state while the replay initiates.
1070
1070
  */
1071
1071
  Attributes?: SubscriptionAttributesMap;
1072
1072
  /**