@aws-sdk/client-firehose 3.36.0 → 3.39.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/CHANGELOG.md +42 -0
- package/dist-cjs/Firehose.js +0 -6
- package/dist-cjs/FirehoseClient.js +0 -11
- package/dist-cjs/commands/CreateDeliveryStreamCommand.js +0 -92
- package/dist-cjs/commands/DeleteDeliveryStreamCommand.js +0 -33
- package/dist-cjs/commands/DescribeDeliveryStreamCommand.js +0 -30
- package/dist-cjs/commands/ListDeliveryStreamsCommand.js +0 -31
- package/dist-cjs/commands/ListTagsForDeliveryStreamCommand.js +0 -25
- package/dist-cjs/commands/PutRecordBatchCommand.js +0 -74
- package/dist-cjs/commands/PutRecordCommand.js +0 -55
- package/dist-cjs/commands/StartDeliveryStreamEncryptionCommand.js +0 -59
- package/dist-cjs/commands/StopDeliveryStreamEncryptionCommand.js +0 -43
- package/dist-cjs/commands/TagDeliveryStreamCommand.js +0 -32
- package/dist-cjs/commands/UntagDeliveryStreamCommand.js +0 -27
- package/dist-cjs/commands/UpdateDestinationCommand.js +0 -47
- package/dist-cjs/commands/index.js +15 -0
- package/dist-cjs/endpoints.js +6 -1
- package/dist-cjs/index.js +3 -14
- package/dist-cjs/models/models_0.js +0 -291
- package/dist-cjs/protocols/Aws_json1_1.js +0 -5
- package/dist-cjs/runtimeConfig.browser.js +1 -5
- package/dist-cjs/runtimeConfig.js +1 -5
- package/dist-cjs/runtimeConfig.native.js +0 -3
- package/dist-cjs/runtimeConfig.shared.js +0 -3
- package/dist-es/commands/index.js +12 -0
- package/dist-es/endpoints.js +6 -1
- package/dist-es/index.js +3 -14
- package/dist-types/commands/index.d.ts +12 -0
- package/dist-types/index.d.ts +3 -14
- package/dist-types/ts3.4/Firehose.d.ts +13 -302
- package/dist-types/ts3.4/FirehoseClient.d.ts +24 -92
- package/dist-types/ts3.4/commands/CreateDeliveryStreamCommand.d.ts +2 -88
- package/dist-types/ts3.4/commands/DeleteDeliveryStreamCommand.d.ts +2 -29
- package/dist-types/ts3.4/commands/DescribeDeliveryStreamCommand.d.ts +2 -26
- package/dist-types/ts3.4/commands/ListDeliveryStreamsCommand.d.ts +2 -27
- package/dist-types/ts3.4/commands/ListTagsForDeliveryStreamCommand.d.ts +2 -21
- package/dist-types/ts3.4/commands/PutRecordBatchCommand.d.ts +2 -70
- package/dist-types/ts3.4/commands/PutRecordCommand.d.ts +2 -51
- package/dist-types/ts3.4/commands/StartDeliveryStreamEncryptionCommand.d.ts +2 -55
- package/dist-types/ts3.4/commands/StopDeliveryStreamEncryptionCommand.d.ts +2 -39
- package/dist-types/ts3.4/commands/TagDeliveryStreamCommand.d.ts +2 -28
- package/dist-types/ts3.4/commands/UntagDeliveryStreamCommand.d.ts +2 -23
- package/dist-types/ts3.4/commands/UpdateDestinationCommand.d.ts +2 -43
- package/dist-types/ts3.4/commands/index.d.ts +12 -0
- package/dist-types/ts3.4/index.d.ts +3 -14
- package/dist-types/ts3.4/models/models_0.d.ts +540 -2345
- package/dist-types/ts3.4/runtimeConfig.browser.d.ts +1 -3
- package/dist-types/ts3.4/runtimeConfig.d.ts +1 -3
- package/dist-types/ts3.4/runtimeConfig.native.d.ts +1 -3
- package/dist-types/ts3.4/runtimeConfig.shared.d.ts +1 -3
- package/package.json +32 -32
|
@@ -4,34 +4,20 @@ export interface AmazonopensearchserviceBufferingHints {
|
|
|
4
4
|
SizeInMBs?: number;
|
|
5
5
|
}
|
|
6
6
|
export declare namespace AmazonopensearchserviceBufferingHints {
|
|
7
|
-
|
|
8
|
-
* @internal
|
|
9
|
-
*/
|
|
7
|
+
|
|
10
8
|
const filterSensitiveLog: (obj: AmazonopensearchserviceBufferingHints) => any;
|
|
11
9
|
}
|
|
12
|
-
|
|
13
|
-
* <p>Describes the Amazon CloudWatch logging options for your delivery stream.</p>
|
|
14
|
-
*/
|
|
10
|
+
|
|
15
11
|
export interface CloudWatchLoggingOptions {
|
|
16
|
-
|
|
17
|
-
* <p>Enables or disables CloudWatch logging.</p>
|
|
18
|
-
*/
|
|
12
|
+
|
|
19
13
|
Enabled?: boolean;
|
|
20
|
-
|
|
21
|
-
* <p>The CloudWatch group name for logging. This value is required if CloudWatch logging
|
|
22
|
-
* is enabled.</p>
|
|
23
|
-
*/
|
|
14
|
+
|
|
24
15
|
LogGroupName?: string;
|
|
25
|
-
|
|
26
|
-
* <p>The CloudWatch log stream name for logging. This value is required if CloudWatch
|
|
27
|
-
* logging is enabled.</p>
|
|
28
|
-
*/
|
|
16
|
+
|
|
29
17
|
LogStreamName?: string;
|
|
30
18
|
}
|
|
31
19
|
export declare namespace CloudWatchLoggingOptions {
|
|
32
|
-
|
|
33
|
-
* @internal
|
|
34
|
-
*/
|
|
20
|
+
|
|
35
21
|
const filterSensitiveLog: (obj: CloudWatchLoggingOptions) => any;
|
|
36
22
|
}
|
|
37
23
|
export declare enum AmazonopensearchserviceIndexRotationPeriod {
|
|
@@ -52,107 +38,60 @@ export declare enum ProcessorParameterName {
|
|
|
52
38
|
ROLE_ARN = "RoleArn",
|
|
53
39
|
SUB_RECORD_TYPE = "SubRecordType"
|
|
54
40
|
}
|
|
55
|
-
|
|
56
|
-
* <p>Describes the processor parameter.</p>
|
|
57
|
-
*/
|
|
41
|
+
|
|
58
42
|
export interface ProcessorParameter {
|
|
59
|
-
|
|
60
|
-
* <p>The name of the parameter.</p>
|
|
61
|
-
*/
|
|
43
|
+
|
|
62
44
|
ParameterName: ProcessorParameterName | string | undefined;
|
|
63
|
-
|
|
64
|
-
* <p>The parameter value.</p>
|
|
65
|
-
*/
|
|
45
|
+
|
|
66
46
|
ParameterValue: string | undefined;
|
|
67
47
|
}
|
|
68
48
|
export declare namespace ProcessorParameter {
|
|
69
|
-
|
|
70
|
-
* @internal
|
|
71
|
-
*/
|
|
49
|
+
|
|
72
50
|
const filterSensitiveLog: (obj: ProcessorParameter) => any;
|
|
73
51
|
}
|
|
74
52
|
export declare type ProcessorType = "AppendDelimiterToRecord" | "Lambda" | "MetadataExtraction" | "RecordDeAggregation";
|
|
75
|
-
|
|
76
|
-
* <p>Describes a data processor.</p>
|
|
77
|
-
*/
|
|
53
|
+
|
|
78
54
|
export interface Processor {
|
|
79
|
-
|
|
80
|
-
* <p>The type of processor.</p>
|
|
81
|
-
*/
|
|
55
|
+
|
|
82
56
|
Type: ProcessorType | string | undefined;
|
|
83
|
-
|
|
84
|
-
* <p>The processor parameters.</p>
|
|
85
|
-
*/
|
|
57
|
+
|
|
86
58
|
Parameters?: ProcessorParameter[];
|
|
87
59
|
}
|
|
88
60
|
export declare namespace Processor {
|
|
89
|
-
|
|
90
|
-
* @internal
|
|
91
|
-
*/
|
|
61
|
+
|
|
92
62
|
const filterSensitiveLog: (obj: Processor) => any;
|
|
93
63
|
}
|
|
94
|
-
|
|
95
|
-
* <p>Describes a data processing configuration.</p>
|
|
96
|
-
*/
|
|
64
|
+
|
|
97
65
|
export interface ProcessingConfiguration {
|
|
98
|
-
|
|
99
|
-
* <p>Enables or disables data processing.</p>
|
|
100
|
-
*/
|
|
66
|
+
|
|
101
67
|
Enabled?: boolean;
|
|
102
|
-
|
|
103
|
-
* <p>The data processors.</p>
|
|
104
|
-
*/
|
|
68
|
+
|
|
105
69
|
Processors?: Processor[];
|
|
106
70
|
}
|
|
107
71
|
export declare namespace ProcessingConfiguration {
|
|
108
|
-
|
|
109
|
-
* @internal
|
|
110
|
-
*/
|
|
72
|
+
|
|
111
73
|
const filterSensitiveLog: (obj: ProcessingConfiguration) => any;
|
|
112
74
|
}
|
|
113
75
|
export interface AmazonopensearchserviceRetryOptions {
|
|
114
76
|
DurationInSeconds?: number;
|
|
115
77
|
}
|
|
116
78
|
export declare namespace AmazonopensearchserviceRetryOptions {
|
|
117
|
-
|
|
118
|
-
* @internal
|
|
119
|
-
*/
|
|
79
|
+
|
|
120
80
|
const filterSensitiveLog: (obj: AmazonopensearchserviceRetryOptions) => any;
|
|
121
81
|
}
|
|
122
82
|
export declare enum AmazonopensearchserviceS3BackupMode {
|
|
123
83
|
AllDocuments = "AllDocuments",
|
|
124
84
|
FailedDocumentsOnly = "FailedDocumentsOnly"
|
|
125
85
|
}
|
|
126
|
-
|
|
127
|
-
* <p>Describes hints for the buffering to perform before delivering data to the
|
|
128
|
-
* destination. These options are treated as hints, and therefore Kinesis Data Firehose might
|
|
129
|
-
* choose to use different values when it is optimal. The <code>SizeInMBs</code> and
|
|
130
|
-
* <code>IntervalInSeconds</code> parameters are optional. However, if specify a value for
|
|
131
|
-
* one of them, you must also provide a value for the other.</p>
|
|
132
|
-
*/
|
|
86
|
+
|
|
133
87
|
export interface BufferingHints {
|
|
134
|
-
|
|
135
|
-
* <p>Buffer incoming data to the specified size, in MiBs, before delivering it to the
|
|
136
|
-
* destination. The default value is 5. This parameter is optional but if you specify a value
|
|
137
|
-
* for it, you must also specify a value for <code>IntervalInSeconds</code>, and vice
|
|
138
|
-
* versa.</p>
|
|
139
|
-
* <p>We recommend setting this parameter to a value greater than the amount of data you
|
|
140
|
-
* typically ingest into the delivery stream in 10 seconds. For example, if you typically
|
|
141
|
-
* ingest data at 1 MiB/sec, the value should be 10 MiB or higher.</p>
|
|
142
|
-
*/
|
|
88
|
+
|
|
143
89
|
SizeInMBs?: number;
|
|
144
|
-
|
|
145
|
-
* <p>Buffer incoming data for the specified period of time, in seconds, before delivering
|
|
146
|
-
* it to the destination. The default value is 300. This parameter is optional but if you
|
|
147
|
-
* specify a value for it, you must also specify a value for <code>SizeInMBs</code>, and vice
|
|
148
|
-
* versa.</p>
|
|
149
|
-
*/
|
|
90
|
+
|
|
150
91
|
IntervalInSeconds?: number;
|
|
151
92
|
}
|
|
152
93
|
export declare namespace BufferingHints {
|
|
153
|
-
|
|
154
|
-
* @internal
|
|
155
|
-
*/
|
|
94
|
+
|
|
156
95
|
const filterSensitiveLog: (obj: BufferingHints) => any;
|
|
157
96
|
}
|
|
158
97
|
export declare enum CompressionFormat {
|
|
@@ -162,190 +101,61 @@ export declare enum CompressionFormat {
|
|
|
162
101
|
UNCOMPRESSED = "UNCOMPRESSED",
|
|
163
102
|
ZIP = "ZIP"
|
|
164
103
|
}
|
|
165
|
-
|
|
166
|
-
* <p>Describes an encryption key for a destination in Amazon S3.</p>
|
|
167
|
-
*/
|
|
104
|
+
|
|
168
105
|
export interface KMSEncryptionConfig {
|
|
169
|
-
|
|
170
|
-
* <p>The Amazon Resource Name (ARN) of the encryption key. Must belong to the same AWS
|
|
171
|
-
* Region as the destination Amazon S3 bucket. For more information, see <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon
|
|
172
|
-
* Resource Names (ARNs) and AWS Service Namespaces</a>.</p>
|
|
173
|
-
*/
|
|
106
|
+
|
|
174
107
|
AWSKMSKeyARN: string | undefined;
|
|
175
108
|
}
|
|
176
109
|
export declare namespace KMSEncryptionConfig {
|
|
177
|
-
|
|
178
|
-
* @internal
|
|
179
|
-
*/
|
|
110
|
+
|
|
180
111
|
const filterSensitiveLog: (obj: KMSEncryptionConfig) => any;
|
|
181
112
|
}
|
|
182
113
|
export declare type NoEncryptionConfig = "NoEncryption";
|
|
183
|
-
|
|
184
|
-
* <p>Describes the encryption for a destination in Amazon S3.</p>
|
|
185
|
-
*/
|
|
114
|
+
|
|
186
115
|
export interface EncryptionConfiguration {
|
|
187
|
-
|
|
188
|
-
* <p>Specifically override existing encryption information to ensure that no encryption is
|
|
189
|
-
* used.</p>
|
|
190
|
-
*/
|
|
116
|
+
|
|
191
117
|
NoEncryptionConfig?: NoEncryptionConfig | string;
|
|
192
|
-
|
|
193
|
-
* <p>The encryption key.</p>
|
|
194
|
-
*/
|
|
118
|
+
|
|
195
119
|
KMSEncryptionConfig?: KMSEncryptionConfig;
|
|
196
120
|
}
|
|
197
121
|
export declare namespace EncryptionConfiguration {
|
|
198
|
-
|
|
199
|
-
* @internal
|
|
200
|
-
*/
|
|
122
|
+
|
|
201
123
|
const filterSensitiveLog: (obj: EncryptionConfiguration) => any;
|
|
202
124
|
}
|
|
203
|
-
|
|
204
|
-
* <p>Describes the configuration of a destination in Amazon S3.</p>
|
|
205
|
-
*/
|
|
125
|
+
|
|
206
126
|
export interface S3DestinationConfiguration {
|
|
207
|
-
|
|
208
|
-
* <p>The Amazon Resource Name (ARN) of the AWS credentials. For more information, see
|
|
209
|
-
* <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon
|
|
210
|
-
* Resource Names (ARNs) and AWS Service Namespaces</a>.</p>
|
|
211
|
-
*/
|
|
127
|
+
|
|
212
128
|
RoleARN: string | undefined;
|
|
213
|
-
|
|
214
|
-
* <p>The ARN of the S3 bucket. For more information, see <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon Resource Names (ARNs) and
|
|
215
|
-
* AWS Service Namespaces</a>.</p>
|
|
216
|
-
*/
|
|
129
|
+
|
|
217
130
|
BucketARN: string | undefined;
|
|
218
|
-
|
|
219
|
-
* <p>The "YYYY/MM/DD/HH" time format prefix is automatically used for delivered Amazon S3
|
|
220
|
-
* files. You can also specify a custom prefix, as described in <a href="https://docs.aws.amazon.com/firehose/latest/dev/s3-prefixes.html">Custom Prefixes for Amazon S3
|
|
221
|
-
* Objects</a>.</p>
|
|
222
|
-
*/
|
|
131
|
+
|
|
223
132
|
Prefix?: string;
|
|
224
|
-
|
|
225
|
-
* <p>A prefix that Kinesis Data Firehose evaluates and adds to failed records before writing
|
|
226
|
-
* them to S3. This prefix appears immediately following the bucket name. For information
|
|
227
|
-
* about how to specify this prefix, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/s3-prefixes.html">Custom Prefixes for Amazon S3
|
|
228
|
-
* Objects</a>.</p>
|
|
229
|
-
*/
|
|
133
|
+
|
|
230
134
|
ErrorOutputPrefix?: string;
|
|
231
|
-
|
|
232
|
-
* <p>The buffering option. If no value is specified, <code>BufferingHints</code> object
|
|
233
|
-
* default values are used.</p>
|
|
234
|
-
*/
|
|
135
|
+
|
|
235
136
|
BufferingHints?: BufferingHints;
|
|
236
|
-
|
|
237
|
-
* <p>The compression format. If no value is specified, the default is
|
|
238
|
-
* <code>UNCOMPRESSED</code>.</p>
|
|
239
|
-
* <p>The compression formats <code>SNAPPY</code> or <code>ZIP</code> cannot be specified
|
|
240
|
-
* for Amazon Redshift destinations because they are not supported by the Amazon Redshift
|
|
241
|
-
* <code>COPY</code> operation that reads from the S3 bucket.</p>
|
|
242
|
-
*/
|
|
137
|
+
|
|
243
138
|
CompressionFormat?: CompressionFormat | string;
|
|
244
|
-
|
|
245
|
-
* <p>The encryption configuration. If no value is specified, the default is no
|
|
246
|
-
* encryption.</p>
|
|
247
|
-
*/
|
|
139
|
+
|
|
248
140
|
EncryptionConfiguration?: EncryptionConfiguration;
|
|
249
|
-
|
|
250
|
-
* <p>The CloudWatch logging options for your delivery stream.</p>
|
|
251
|
-
*/
|
|
141
|
+
|
|
252
142
|
CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
|
|
253
143
|
}
|
|
254
144
|
export declare namespace S3DestinationConfiguration {
|
|
255
|
-
|
|
256
|
-
* @internal
|
|
257
|
-
*/
|
|
145
|
+
|
|
258
146
|
const filterSensitiveLog: (obj: S3DestinationConfiguration) => any;
|
|
259
147
|
}
|
|
260
|
-
|
|
261
|
-
* <p>The details of the VPC of the Amazon ES destination.</p>
|
|
262
|
-
*/
|
|
148
|
+
|
|
263
149
|
export interface VpcConfiguration {
|
|
264
|
-
|
|
265
|
-
* <p>The IDs of the subnets that you want Kinesis Data Firehose to use to create ENIs in the
|
|
266
|
-
* VPC of the Amazon ES destination. Make sure that the routing tables and inbound and
|
|
267
|
-
* outbound rules allow traffic to flow from the subnets whose IDs are specified here to the
|
|
268
|
-
* subnets that have the destination Amazon ES endpoints. Kinesis Data Firehose creates at
|
|
269
|
-
* least one ENI in each of the subnets that are specified here. Do not delete or modify these
|
|
270
|
-
* ENIs.</p>
|
|
271
|
-
* <p>The number of ENIs that Kinesis Data Firehose creates in the subnets specified here
|
|
272
|
-
* scales up and down automatically based on throughput. To enable Kinesis Data Firehose to
|
|
273
|
-
* scale up the number of ENIs to match throughput, ensure that you have sufficient quota. To
|
|
274
|
-
* help you calculate the quota you need, assume that Kinesis Data Firehose can create up to
|
|
275
|
-
* three ENIs for this delivery stream for each of the subnets specified here. For more
|
|
276
|
-
* information about ENI quota, see <a href="https://docs.aws.amazon.com/vpc/latest/userguide/amazon-vpc-limits.html#vpc-limits-enis">Network Interfaces
|
|
277
|
-
* </a> in the Amazon VPC Quotas topic.</p>
|
|
278
|
-
*/
|
|
150
|
+
|
|
279
151
|
SubnetIds: string[] | undefined;
|
|
280
|
-
|
|
281
|
-
* <p>The ARN of the IAM role that you want the delivery stream to use to create endpoints in
|
|
282
|
-
* the destination VPC. You can use your existing Kinesis Data Firehose delivery role or you
|
|
283
|
-
* can specify a new role. In either case, make sure that the role trusts the Kinesis Data
|
|
284
|
-
* Firehose service principal and that it grants the following permissions:</p>
|
|
285
|
-
* <ul>
|
|
286
|
-
* <li>
|
|
287
|
-
* <p>
|
|
288
|
-
* <code>ec2:DescribeVpcs</code>
|
|
289
|
-
* </p>
|
|
290
|
-
* </li>
|
|
291
|
-
* <li>
|
|
292
|
-
* <p>
|
|
293
|
-
* <code>ec2:DescribeVpcAttribute</code>
|
|
294
|
-
* </p>
|
|
295
|
-
* </li>
|
|
296
|
-
* <li>
|
|
297
|
-
* <p>
|
|
298
|
-
* <code>ec2:DescribeSubnets</code>
|
|
299
|
-
* </p>
|
|
300
|
-
* </li>
|
|
301
|
-
* <li>
|
|
302
|
-
* <p>
|
|
303
|
-
* <code>ec2:DescribeSecurityGroups</code>
|
|
304
|
-
* </p>
|
|
305
|
-
* </li>
|
|
306
|
-
* <li>
|
|
307
|
-
* <p>
|
|
308
|
-
* <code>ec2:DescribeNetworkInterfaces</code>
|
|
309
|
-
* </p>
|
|
310
|
-
* </li>
|
|
311
|
-
* <li>
|
|
312
|
-
* <p>
|
|
313
|
-
* <code>ec2:CreateNetworkInterface</code>
|
|
314
|
-
* </p>
|
|
315
|
-
* </li>
|
|
316
|
-
* <li>
|
|
317
|
-
* <p>
|
|
318
|
-
* <code>ec2:CreateNetworkInterfacePermission</code>
|
|
319
|
-
* </p>
|
|
320
|
-
* </li>
|
|
321
|
-
* <li>
|
|
322
|
-
* <p>
|
|
323
|
-
* <code>ec2:DeleteNetworkInterface</code>
|
|
324
|
-
* </p>
|
|
325
|
-
* </li>
|
|
326
|
-
* </ul>
|
|
327
|
-
* <p>If you revoke these permissions after you create the delivery stream, Kinesis Data
|
|
328
|
-
* Firehose can't scale out by creating more ENIs when necessary. You might therefore see a
|
|
329
|
-
* degradation in performance.</p>
|
|
330
|
-
*/
|
|
152
|
+
|
|
331
153
|
RoleARN: string | undefined;
|
|
332
|
-
|
|
333
|
-
* <p>The IDs of the security groups that you want Kinesis Data Firehose to use when it
|
|
334
|
-
* creates ENIs in the VPC of the Amazon ES destination. You can use the same security group
|
|
335
|
-
* that the Amazon ES domain uses or different ones. If you specify different security groups
|
|
336
|
-
* here, ensure that they allow outbound HTTPS traffic to the Amazon ES domain's security
|
|
337
|
-
* group. Also ensure that the Amazon ES domain's security group allows HTTPS traffic from the
|
|
338
|
-
* security groups specified here. If you use the same security group for both your delivery
|
|
339
|
-
* stream and the Amazon ES domain, make sure the security group inbound rule allows HTTPS
|
|
340
|
-
* traffic. For more information about security group rules, see <a href="https://docs.aws.amazon.com/vpc/latest/userguide/VPC_SecurityGroups.html#SecurityGroupRules">Security group
|
|
341
|
-
* rules</a> in the Amazon VPC documentation.</p>
|
|
342
|
-
*/
|
|
154
|
+
|
|
343
155
|
SecurityGroupIds: string[] | undefined;
|
|
344
156
|
}
|
|
345
157
|
export declare namespace VpcConfiguration {
|
|
346
|
-
|
|
347
|
-
* @internal
|
|
348
|
-
*/
|
|
158
|
+
|
|
349
159
|
const filterSensitiveLog: (obj: VpcConfiguration) => any;
|
|
350
160
|
}
|
|
351
161
|
export interface AmazonopensearchserviceDestinationConfiguration {
|
|
@@ -358,175 +168,55 @@ export interface AmazonopensearchserviceDestinationConfiguration {
|
|
|
358
168
|
BufferingHints?: AmazonopensearchserviceBufferingHints;
|
|
359
169
|
RetryOptions?: AmazonopensearchserviceRetryOptions;
|
|
360
170
|
S3BackupMode?: AmazonopensearchserviceS3BackupMode | string;
|
|
361
|
-
|
|
362
|
-
* <p>Describes the configuration of a destination in Amazon S3.</p>
|
|
363
|
-
*/
|
|
171
|
+
|
|
364
172
|
S3Configuration: S3DestinationConfiguration | undefined;
|
|
365
|
-
|
|
366
|
-
* <p>Describes a data processing configuration.</p>
|
|
367
|
-
*/
|
|
173
|
+
|
|
368
174
|
ProcessingConfiguration?: ProcessingConfiguration;
|
|
369
|
-
|
|
370
|
-
* <p>Describes the Amazon CloudWatch logging options for your delivery stream.</p>
|
|
371
|
-
*/
|
|
175
|
+
|
|
372
176
|
CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
|
|
373
|
-
|
|
374
|
-
* <p>The details of the VPC of the Amazon ES destination.</p>
|
|
375
|
-
*/
|
|
177
|
+
|
|
376
178
|
VpcConfiguration?: VpcConfiguration;
|
|
377
179
|
}
|
|
378
180
|
export declare namespace AmazonopensearchserviceDestinationConfiguration {
|
|
379
|
-
|
|
380
|
-
* @internal
|
|
381
|
-
*/
|
|
181
|
+
|
|
382
182
|
const filterSensitiveLog: (obj: AmazonopensearchserviceDestinationConfiguration) => any;
|
|
383
183
|
}
|
|
384
|
-
|
|
385
|
-
* <p>Describes a destination in Amazon S3.</p>
|
|
386
|
-
*/
|
|
184
|
+
|
|
387
185
|
export interface S3DestinationDescription {
|
|
388
|
-
|
|
389
|
-
* <p>The Amazon Resource Name (ARN) of the AWS credentials. For more information, see
|
|
390
|
-
* <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon
|
|
391
|
-
* Resource Names (ARNs) and AWS Service Namespaces</a>.</p>
|
|
392
|
-
*/
|
|
186
|
+
|
|
393
187
|
RoleARN: string | undefined;
|
|
394
|
-
|
|
395
|
-
* <p>The ARN of the S3 bucket. For more information, see <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon Resource Names (ARNs) and
|
|
396
|
-
* AWS Service Namespaces</a>.</p>
|
|
397
|
-
*/
|
|
188
|
+
|
|
398
189
|
BucketARN: string | undefined;
|
|
399
|
-
|
|
400
|
-
* <p>The "YYYY/MM/DD/HH" time format prefix is automatically used for delivered Amazon S3
|
|
401
|
-
* files. You can also specify a custom prefix, as described in <a href="https://docs.aws.amazon.com/firehose/latest/dev/s3-prefixes.html">Custom Prefixes for Amazon S3
|
|
402
|
-
* Objects</a>.</p>
|
|
403
|
-
*/
|
|
190
|
+
|
|
404
191
|
Prefix?: string;
|
|
405
|
-
|
|
406
|
-
* <p>A prefix that Kinesis Data Firehose evaluates and adds to failed records before writing
|
|
407
|
-
* them to S3. This prefix appears immediately following the bucket name. For information
|
|
408
|
-
* about how to specify this prefix, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/s3-prefixes.html">Custom Prefixes for Amazon S3
|
|
409
|
-
* Objects</a>.</p>
|
|
410
|
-
*/
|
|
192
|
+
|
|
411
193
|
ErrorOutputPrefix?: string;
|
|
412
|
-
|
|
413
|
-
* <p>The buffering option. If no value is specified, <code>BufferingHints</code> object
|
|
414
|
-
* default values are used.</p>
|
|
415
|
-
*/
|
|
194
|
+
|
|
416
195
|
BufferingHints: BufferingHints | undefined;
|
|
417
|
-
|
|
418
|
-
* <p>The compression format. If no value is specified, the default is
|
|
419
|
-
* <code>UNCOMPRESSED</code>.</p>
|
|
420
|
-
*/
|
|
196
|
+
|
|
421
197
|
CompressionFormat: CompressionFormat | string | undefined;
|
|
422
|
-
|
|
423
|
-
* <p>The encryption configuration. If no value is specified, the default is no
|
|
424
|
-
* encryption.</p>
|
|
425
|
-
*/
|
|
198
|
+
|
|
426
199
|
EncryptionConfiguration: EncryptionConfiguration | undefined;
|
|
427
|
-
|
|
428
|
-
* <p>The Amazon CloudWatch logging options for your delivery stream.</p>
|
|
429
|
-
*/
|
|
200
|
+
|
|
430
201
|
CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
|
|
431
202
|
}
|
|
432
203
|
export declare namespace S3DestinationDescription {
|
|
433
|
-
|
|
434
|
-
* @internal
|
|
435
|
-
*/
|
|
204
|
+
|
|
436
205
|
const filterSensitiveLog: (obj: S3DestinationDescription) => any;
|
|
437
206
|
}
|
|
438
|
-
|
|
439
|
-
* <p>The details of the VPC of the Amazon ES destination.</p>
|
|
440
|
-
*/
|
|
207
|
+
|
|
441
208
|
export interface VpcConfigurationDescription {
|
|
442
|
-
|
|
443
|
-
* <p>The IDs of the subnets that Kinesis Data Firehose uses to create ENIs in the VPC of the
|
|
444
|
-
* Amazon ES destination. Make sure that the routing tables and inbound and outbound rules
|
|
445
|
-
* allow traffic to flow from the subnets whose IDs are specified here to the subnets that
|
|
446
|
-
* have the destination Amazon ES endpoints. Kinesis Data Firehose creates at least one ENI in
|
|
447
|
-
* each of the subnets that are specified here. Do not delete or modify these ENIs.</p>
|
|
448
|
-
* <p>The number of ENIs that Kinesis Data Firehose creates in the subnets specified here
|
|
449
|
-
* scales up and down automatically based on throughput. To enable Kinesis Data Firehose to
|
|
450
|
-
* scale up the number of ENIs to match throughput, ensure that you have sufficient quota. To
|
|
451
|
-
* help you calculate the quota you need, assume that Kinesis Data Firehose can create up to
|
|
452
|
-
* three ENIs for this delivery stream for each of the subnets specified here. For more
|
|
453
|
-
* information about ENI quota, see <a href="https://docs.aws.amazon.com/vpc/latest/userguide/amazon-vpc-limits.html#vpc-limits-enis">Network Interfaces
|
|
454
|
-
* </a> in the Amazon VPC Quotas topic.</p>
|
|
455
|
-
*/
|
|
209
|
+
|
|
456
210
|
SubnetIds: string[] | undefined;
|
|
457
|
-
|
|
458
|
-
* <p>The ARN of the IAM role that the delivery stream uses to create endpoints in the
|
|
459
|
-
* destination VPC. You can use your existing Kinesis Data Firehose delivery role or you can
|
|
460
|
-
* specify a new role. In either case, make sure that the role trusts the Kinesis Data
|
|
461
|
-
* Firehose service principal and that it grants the following permissions:</p>
|
|
462
|
-
* <ul>
|
|
463
|
-
* <li>
|
|
464
|
-
* <p>
|
|
465
|
-
* <code>ec2:DescribeVpcs</code>
|
|
466
|
-
* </p>
|
|
467
|
-
* </li>
|
|
468
|
-
* <li>
|
|
469
|
-
* <p>
|
|
470
|
-
* <code>ec2:DescribeVpcAttribute</code>
|
|
471
|
-
* </p>
|
|
472
|
-
* </li>
|
|
473
|
-
* <li>
|
|
474
|
-
* <p>
|
|
475
|
-
* <code>ec2:DescribeSubnets</code>
|
|
476
|
-
* </p>
|
|
477
|
-
* </li>
|
|
478
|
-
* <li>
|
|
479
|
-
* <p>
|
|
480
|
-
* <code>ec2:DescribeSecurityGroups</code>
|
|
481
|
-
* </p>
|
|
482
|
-
* </li>
|
|
483
|
-
* <li>
|
|
484
|
-
* <p>
|
|
485
|
-
* <code>ec2:DescribeNetworkInterfaces</code>
|
|
486
|
-
* </p>
|
|
487
|
-
* </li>
|
|
488
|
-
* <li>
|
|
489
|
-
* <p>
|
|
490
|
-
* <code>ec2:CreateNetworkInterface</code>
|
|
491
|
-
* </p>
|
|
492
|
-
* </li>
|
|
493
|
-
* <li>
|
|
494
|
-
* <p>
|
|
495
|
-
* <code>ec2:CreateNetworkInterfacePermission</code>
|
|
496
|
-
* </p>
|
|
497
|
-
* </li>
|
|
498
|
-
* <li>
|
|
499
|
-
* <p>
|
|
500
|
-
* <code>ec2:DeleteNetworkInterface</code>
|
|
501
|
-
* </p>
|
|
502
|
-
* </li>
|
|
503
|
-
* </ul>
|
|
504
|
-
* <p>If you revoke these permissions after you create the delivery stream, Kinesis Data
|
|
505
|
-
* Firehose can't scale out by creating more ENIs when necessary. You might therefore see a
|
|
506
|
-
* degradation in performance.</p>
|
|
507
|
-
*/
|
|
211
|
+
|
|
508
212
|
RoleARN: string | undefined;
|
|
509
|
-
|
|
510
|
-
* <p>The IDs of the security groups that Kinesis Data Firehose uses when it creates ENIs in
|
|
511
|
-
* the VPC of the Amazon ES destination. You can use the same security group that the Amazon
|
|
512
|
-
* ES domain uses or different ones. If you specify different security groups, ensure that
|
|
513
|
-
* they allow outbound HTTPS traffic to the Amazon ES domain's security group. Also ensure
|
|
514
|
-
* that the Amazon ES domain's security group allows HTTPS traffic from the security groups
|
|
515
|
-
* specified here. If you use the same security group for both your delivery stream and the
|
|
516
|
-
* Amazon ES domain, make sure the security group inbound rule allows HTTPS traffic. For more
|
|
517
|
-
* information about security group rules, see <a href="https://docs.aws.amazon.com/vpc/latest/userguide/VPC_SecurityGroups.html#SecurityGroupRules">Security group
|
|
518
|
-
* rules</a> in the Amazon VPC documentation.</p>
|
|
519
|
-
*/
|
|
213
|
+
|
|
520
214
|
SecurityGroupIds: string[] | undefined;
|
|
521
|
-
|
|
522
|
-
* <p>The ID of the Amazon ES destination's VPC.</p>
|
|
523
|
-
*/
|
|
215
|
+
|
|
524
216
|
VpcId: string | undefined;
|
|
525
217
|
}
|
|
526
218
|
export declare namespace VpcConfigurationDescription {
|
|
527
|
-
|
|
528
|
-
* @internal
|
|
529
|
-
*/
|
|
219
|
+
|
|
530
220
|
const filterSensitiveLog: (obj: VpcConfigurationDescription) => any;
|
|
531
221
|
}
|
|
532
222
|
export interface AmazonopensearchserviceDestinationDescription {
|
|
@@ -539,84 +229,40 @@ export interface AmazonopensearchserviceDestinationDescription {
|
|
|
539
229
|
BufferingHints?: AmazonopensearchserviceBufferingHints;
|
|
540
230
|
RetryOptions?: AmazonopensearchserviceRetryOptions;
|
|
541
231
|
S3BackupMode?: AmazonopensearchserviceS3BackupMode | string;
|
|
542
|
-
|
|
543
|
-
* <p>Describes a destination in Amazon S3.</p>
|
|
544
|
-
*/
|
|
232
|
+
|
|
545
233
|
S3DestinationDescription?: S3DestinationDescription;
|
|
546
|
-
|
|
547
|
-
* <p>Describes a data processing configuration.</p>
|
|
548
|
-
*/
|
|
234
|
+
|
|
549
235
|
ProcessingConfiguration?: ProcessingConfiguration;
|
|
550
|
-
|
|
551
|
-
* <p>Describes the Amazon CloudWatch logging options for your delivery stream.</p>
|
|
552
|
-
*/
|
|
236
|
+
|
|
553
237
|
CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
|
|
554
|
-
|
|
555
|
-
* <p>The details of the VPC of the Amazon ES destination.</p>
|
|
556
|
-
*/
|
|
238
|
+
|
|
557
239
|
VpcConfigurationDescription?: VpcConfigurationDescription;
|
|
558
240
|
}
|
|
559
241
|
export declare namespace AmazonopensearchserviceDestinationDescription {
|
|
560
|
-
|
|
561
|
-
* @internal
|
|
562
|
-
*/
|
|
242
|
+
|
|
563
243
|
const filterSensitiveLog: (obj: AmazonopensearchserviceDestinationDescription) => any;
|
|
564
244
|
}
|
|
565
|
-
|
|
566
|
-
* <p>Describes an update for a destination in Amazon S3.</p>
|
|
567
|
-
*/
|
|
245
|
+
|
|
568
246
|
export interface S3DestinationUpdate {
|
|
569
|
-
|
|
570
|
-
* <p>The Amazon Resource Name (ARN) of the AWS credentials. For more information, see
|
|
571
|
-
* <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon
|
|
572
|
-
* Resource Names (ARNs) and AWS Service Namespaces</a>.</p>
|
|
573
|
-
*/
|
|
247
|
+
|
|
574
248
|
RoleARN?: string;
|
|
575
|
-
|
|
576
|
-
* <p>The ARN of the S3 bucket. For more information, see <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon Resource Names (ARNs) and
|
|
577
|
-
* AWS Service Namespaces</a>.</p>
|
|
578
|
-
*/
|
|
249
|
+
|
|
579
250
|
BucketARN?: string;
|
|
580
|
-
|
|
581
|
-
* <p>The "YYYY/MM/DD/HH" time format prefix is automatically used for delivered Amazon S3
|
|
582
|
-
* files. You can also specify a custom prefix, as described in <a href="https://docs.aws.amazon.com/firehose/latest/dev/s3-prefixes.html">Custom Prefixes for Amazon S3
|
|
583
|
-
* Objects</a>.</p>
|
|
584
|
-
*/
|
|
251
|
+
|
|
585
252
|
Prefix?: string;
|
|
586
|
-
|
|
587
|
-
* <p>A prefix that Kinesis Data Firehose evaluates and adds to failed records before writing
|
|
588
|
-
* them to S3. This prefix appears immediately following the bucket name. For information
|
|
589
|
-
* about how to specify this prefix, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/s3-prefixes.html">Custom Prefixes for Amazon S3
|
|
590
|
-
* Objects</a>.</p>
|
|
591
|
-
*/
|
|
253
|
+
|
|
592
254
|
ErrorOutputPrefix?: string;
|
|
593
|
-
|
|
594
|
-
* <p>The buffering option. If no value is specified, <code>BufferingHints</code> object
|
|
595
|
-
* default values are used.</p>
|
|
596
|
-
*/
|
|
255
|
+
|
|
597
256
|
BufferingHints?: BufferingHints;
|
|
598
|
-
|
|
599
|
-
* <p>The compression format. If no value is specified, the default is
|
|
600
|
-
* <code>UNCOMPRESSED</code>.</p>
|
|
601
|
-
* <p>The compression formats <code>SNAPPY</code> or <code>ZIP</code> cannot be specified
|
|
602
|
-
* for Amazon Redshift destinations because they are not supported by the Amazon Redshift
|
|
603
|
-
* <code>COPY</code> operation that reads from the S3 bucket.</p>
|
|
604
|
-
*/
|
|
257
|
+
|
|
605
258
|
CompressionFormat?: CompressionFormat | string;
|
|
606
|
-
|
|
607
|
-
* <p>The encryption configuration. If no value is specified, the default is no
|
|
608
|
-
* encryption.</p>
|
|
609
|
-
*/
|
|
259
|
+
|
|
610
260
|
EncryptionConfiguration?: EncryptionConfiguration;
|
|
611
|
-
|
|
612
|
-
* <p>The CloudWatch logging options for your delivery stream.</p>
|
|
613
|
-
*/
|
|
261
|
+
|
|
614
262
|
CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
|
|
615
263
|
}
|
|
616
264
|
export declare namespace S3DestinationUpdate {
|
|
617
|
-
|
|
618
|
-
* @internal
|
|
619
|
-
*/
|
|
265
|
+
|
|
620
266
|
const filterSensitiveLog: (obj: S3DestinationUpdate) => any;
|
|
621
267
|
}
|
|
622
268
|
export interface AmazonopensearchserviceDestinationUpdate {
|
|
@@ -628,376 +274,158 @@ export interface AmazonopensearchserviceDestinationUpdate {
|
|
|
628
274
|
IndexRotationPeriod?: AmazonopensearchserviceIndexRotationPeriod | string;
|
|
629
275
|
BufferingHints?: AmazonopensearchserviceBufferingHints;
|
|
630
276
|
RetryOptions?: AmazonopensearchserviceRetryOptions;
|
|
631
|
-
|
|
632
|
-
* <p>Describes an update for a destination in Amazon S3.</p>
|
|
633
|
-
*/
|
|
277
|
+
|
|
634
278
|
S3Update?: S3DestinationUpdate;
|
|
635
|
-
|
|
636
|
-
* <p>Describes a data processing configuration.</p>
|
|
637
|
-
*/
|
|
279
|
+
|
|
638
280
|
ProcessingConfiguration?: ProcessingConfiguration;
|
|
639
|
-
|
|
640
|
-
* <p>Describes the Amazon CloudWatch logging options for your delivery stream.</p>
|
|
641
|
-
*/
|
|
281
|
+
|
|
642
282
|
CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
|
|
643
283
|
}
|
|
644
284
|
export declare namespace AmazonopensearchserviceDestinationUpdate {
|
|
645
|
-
|
|
646
|
-
* @internal
|
|
647
|
-
*/
|
|
285
|
+
|
|
648
286
|
const filterSensitiveLog: (obj: AmazonopensearchserviceDestinationUpdate) => any;
|
|
649
287
|
}
|
|
650
|
-
|
|
651
|
-
* <p>Another modification has already happened. Fetch <code>VersionId</code> again and use
|
|
652
|
-
* it to update the destination.</p>
|
|
653
|
-
*/
|
|
288
|
+
|
|
654
289
|
export interface ConcurrentModificationException extends __SmithyException, $MetadataBearer {
|
|
655
290
|
name: "ConcurrentModificationException";
|
|
656
291
|
$fault: "client";
|
|
657
|
-
|
|
658
|
-
* <p>A message that provides information about the error.</p>
|
|
659
|
-
*/
|
|
292
|
+
|
|
660
293
|
message?: string;
|
|
661
294
|
}
|
|
662
295
|
export declare namespace ConcurrentModificationException {
|
|
663
|
-
|
|
664
|
-
* @internal
|
|
665
|
-
*/
|
|
296
|
+
|
|
666
297
|
const filterSensitiveLog: (obj: ConcurrentModificationException) => any;
|
|
667
298
|
}
|
|
668
299
|
export declare enum ContentEncoding {
|
|
669
300
|
GZIP = "GZIP",
|
|
670
301
|
NONE = "NONE"
|
|
671
302
|
}
|
|
672
|
-
|
|
673
|
-
* <p>Describes a <code>COPY</code> command for Amazon Redshift.</p>
|
|
674
|
-
*/
|
|
303
|
+
|
|
675
304
|
export interface CopyCommand {
|
|
676
|
-
|
|
677
|
-
* <p>The name of the target table. The table must already exist in the database.</p>
|
|
678
|
-
*/
|
|
305
|
+
|
|
679
306
|
DataTableName: string | undefined;
|
|
680
|
-
|
|
681
|
-
* <p>A comma-separated list of column names.</p>
|
|
682
|
-
*/
|
|
307
|
+
|
|
683
308
|
DataTableColumns?: string;
|
|
684
|
-
|
|
685
|
-
* <p>Optional parameters to use with the Amazon Redshift <code>COPY</code> command. For
|
|
686
|
-
* more information, see the "Optional Parameters" section of <a href="https://docs.aws.amazon.com/redshift/latest/dg/r_COPY.html">Amazon Redshift COPY command</a>. Some possible
|
|
687
|
-
* examples that would apply to Kinesis Data Firehose are as follows:</p>
|
|
688
|
-
* <p>
|
|
689
|
-
* <code>delimiter '\t' lzop;</code> - fields are delimited with "\t" (TAB character) and
|
|
690
|
-
* compressed using lzop.</p>
|
|
691
|
-
* <p>
|
|
692
|
-
* <code>delimiter '|'</code> - fields are delimited with "|" (this is the default
|
|
693
|
-
* delimiter).</p>
|
|
694
|
-
* <p>
|
|
695
|
-
* <code>delimiter '|' escape</code> - the delimiter should be escaped.</p>
|
|
696
|
-
* <p>
|
|
697
|
-
* <code>fixedwidth 'venueid:3,venuename:25,venuecity:12,venuestate:2,venueseats:6'</code> -
|
|
698
|
-
* fields are fixed width in the source, with each width specified after every column in the
|
|
699
|
-
* table.</p>
|
|
700
|
-
* <p>
|
|
701
|
-
* <code>JSON 's3://mybucket/jsonpaths.txt'</code> - data is in JSON format, and the path
|
|
702
|
-
* specified is the format of the data.</p>
|
|
703
|
-
* <p>For more examples, see <a href="https://docs.aws.amazon.com/redshift/latest/dg/r_COPY_command_examples.html">Amazon Redshift COPY command
|
|
704
|
-
* examples</a>.</p>
|
|
705
|
-
*/
|
|
309
|
+
|
|
706
310
|
CopyOptions?: string;
|
|
707
311
|
}
|
|
708
312
|
export declare namespace CopyCommand {
|
|
709
|
-
|
|
710
|
-
* @internal
|
|
711
|
-
*/
|
|
313
|
+
|
|
712
314
|
const filterSensitiveLog: (obj: CopyCommand) => any;
|
|
713
315
|
}
|
|
714
316
|
export declare enum KeyType {
|
|
715
317
|
AWS_OWNED_CMK = "AWS_OWNED_CMK",
|
|
716
318
|
CUSTOMER_MANAGED_CMK = "CUSTOMER_MANAGED_CMK"
|
|
717
319
|
}
|
|
718
|
-
|
|
719
|
-
* <p>Specifies the type and Amazon Resource Name (ARN) of the CMK to use for Server-Side
|
|
720
|
-
* Encryption (SSE). </p>
|
|
721
|
-
*/
|
|
320
|
+
|
|
722
321
|
export interface DeliveryStreamEncryptionConfigurationInput {
|
|
723
|
-
|
|
724
|
-
* <p>If you set <code>KeyType</code> to <code>CUSTOMER_MANAGED_CMK</code>, you must specify
|
|
725
|
-
* the Amazon Resource Name (ARN) of the CMK. If you set <code>KeyType</code> to
|
|
726
|
-
* <code>AWS_OWNED_CMK</code>, Kinesis Data Firehose uses a service-account CMK.</p>
|
|
727
|
-
*/
|
|
322
|
+
|
|
728
323
|
KeyARN?: string;
|
|
729
|
-
|
|
730
|
-
* <p>Indicates the type of customer master key (CMK) to use for encryption. The default
|
|
731
|
-
* setting is <code>AWS_OWNED_CMK</code>. For more information about CMKs, see <a href="https://docs.aws.amazon.com/kms/latest/developerguide/concepts.html#master_keys">Customer
|
|
732
|
-
* Master Keys (CMKs)</a>. When you invoke <a>CreateDeliveryStream</a> or
|
|
733
|
-
* <a>StartDeliveryStreamEncryption</a> with <code>KeyType</code> set to
|
|
734
|
-
* CUSTOMER_MANAGED_CMK, Kinesis Data Firehose invokes the Amazon KMS operation <a href="https://docs.aws.amazon.com/kms/latest/APIReference/API_CreateGrant.html">CreateGrant</a> to create a grant that allows the Kinesis Data Firehose service to
|
|
735
|
-
* use the customer managed CMK to perform encryption and decryption. Kinesis Data Firehose
|
|
736
|
-
* manages that grant. </p>
|
|
737
|
-
* <p>When you invoke <a>StartDeliveryStreamEncryption</a> to change the CMK for a
|
|
738
|
-
* delivery stream that is encrypted with a customer managed CMK, Kinesis Data Firehose
|
|
739
|
-
* schedules the grant it had on the old CMK for retirement.</p>
|
|
740
|
-
* <p>You can use a CMK of type CUSTOMER_MANAGED_CMK to encrypt up to 500 delivery streams. If
|
|
741
|
-
* a <a>CreateDeliveryStream</a> or <a>StartDeliveryStreamEncryption</a>
|
|
742
|
-
* operation exceeds this limit, Kinesis Data Firehose throws a
|
|
743
|
-
* <code>LimitExceededException</code>. </p>
|
|
744
|
-
* <important>
|
|
745
|
-
* <p>To encrypt your delivery stream, use symmetric CMKs. Kinesis Data Firehose doesn't
|
|
746
|
-
* support asymmetric CMKs. For information about symmetric and asymmetric CMKs, see <a href="https://docs.aws.amazon.com/kms/latest/developerguide/symm-asymm-concepts.html">About
|
|
747
|
-
* Symmetric and Asymmetric CMKs</a> in the AWS Key Management Service developer
|
|
748
|
-
* guide.</p>
|
|
749
|
-
* </important>
|
|
750
|
-
*/
|
|
324
|
+
|
|
751
325
|
KeyType: KeyType | string | undefined;
|
|
752
326
|
}
|
|
753
327
|
export declare namespace DeliveryStreamEncryptionConfigurationInput {
|
|
754
|
-
|
|
755
|
-
* @internal
|
|
756
|
-
*/
|
|
328
|
+
|
|
757
329
|
const filterSensitiveLog: (obj: DeliveryStreamEncryptionConfigurationInput) => any;
|
|
758
330
|
}
|
|
759
331
|
export declare type DeliveryStreamType = "DirectPut" | "KinesisStreamAsSource";
|
|
760
|
-
|
|
761
|
-
* <p>Describes the buffering to perform before delivering data to the Amazon ES
|
|
762
|
-
* destination.</p>
|
|
763
|
-
*/
|
|
332
|
+
|
|
764
333
|
export interface ElasticsearchBufferingHints {
|
|
765
|
-
|
|
766
|
-
* <p>Buffer incoming data for the specified period of time, in seconds, before delivering
|
|
767
|
-
* it to the destination. The default value is 300 (5 minutes).</p>
|
|
768
|
-
*/
|
|
334
|
+
|
|
769
335
|
IntervalInSeconds?: number;
|
|
770
|
-
|
|
771
|
-
* <p>Buffer incoming data to the specified size, in MBs, before delivering it to the
|
|
772
|
-
* destination. The default value is 5.</p>
|
|
773
|
-
* <p>We recommend setting this parameter to a value greater than the amount of data you
|
|
774
|
-
* typically ingest into the delivery stream in 10 seconds. For example, if you typically
|
|
775
|
-
* ingest data at 1 MB/sec, the value should be 10 MB or higher.</p>
|
|
776
|
-
*/
|
|
336
|
+
|
|
777
337
|
SizeInMBs?: number;
|
|
778
338
|
}
|
|
779
339
|
export declare namespace ElasticsearchBufferingHints {
|
|
780
|
-
|
|
781
|
-
* @internal
|
|
782
|
-
*/
|
|
340
|
+
|
|
783
341
|
const filterSensitiveLog: (obj: ElasticsearchBufferingHints) => any;
|
|
784
342
|
}
|
|
785
343
|
export declare type ElasticsearchIndexRotationPeriod = "NoRotation" | "OneDay" | "OneHour" | "OneMonth" | "OneWeek";
|
|
786
|
-
|
|
787
|
-
* <p>Configures retry behavior in case Kinesis Data Firehose is unable to deliver
|
|
788
|
-
* documents to Amazon ES.</p>
|
|
789
|
-
*/
|
|
344
|
+
|
|
790
345
|
export interface ElasticsearchRetryOptions {
|
|
791
|
-
|
|
792
|
-
* <p>After an initial failure to deliver to Amazon ES, the total amount of time during
|
|
793
|
-
* which Kinesis Data Firehose retries delivery (including the first attempt). After this time
|
|
794
|
-
* has elapsed, the failed documents are written to Amazon S3. Default value is 300 seconds (5
|
|
795
|
-
* minutes). A value of 0 (zero) results in no retries.</p>
|
|
796
|
-
*/
|
|
346
|
+
|
|
797
347
|
DurationInSeconds?: number;
|
|
798
348
|
}
|
|
799
349
|
export declare namespace ElasticsearchRetryOptions {
|
|
800
|
-
|
|
801
|
-
* @internal
|
|
802
|
-
*/
|
|
350
|
+
|
|
803
351
|
const filterSensitiveLog: (obj: ElasticsearchRetryOptions) => any;
|
|
804
352
|
}
|
|
805
353
|
export declare type ElasticsearchS3BackupMode = "AllDocuments" | "FailedDocumentsOnly";
|
|
806
|
-
|
|
807
|
-
* <p>Describes the configuration of a destination in Amazon ES.</p>
|
|
808
|
-
*/
|
|
354
|
+
|
|
809
355
|
export interface ElasticsearchDestinationConfiguration {
|
|
810
|
-
|
|
811
|
-
* <p>The Amazon Resource Name (ARN) of the IAM role to be assumed by Kinesis Data Firehose
|
|
812
|
-
* for calling the Amazon ES Configuration API and for indexing documents. For more
|
|
813
|
-
* information, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/controlling-access.html#using-iam-s3">Grant Kinesis Data
|
|
814
|
-
* Firehose Access to an Amazon S3 Destination</a> and <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon Resource Names (ARNs) and
|
|
815
|
-
* AWS Service Namespaces</a>.</p>
|
|
816
|
-
*/
|
|
356
|
+
|
|
817
357
|
RoleARN: string | undefined;
|
|
818
|
-
|
|
819
|
-
* <p>The ARN of the Amazon ES domain. The IAM role must have permissions
|
|
820
|
-
* for <code>DescribeElasticsearchDomain</code>, <code>DescribeElasticsearchDomains</code>,
|
|
821
|
-
* and <code>DescribeElasticsearchDomainConfig</code> after assuming the role specified in
|
|
822
|
-
* <b>RoleARN</b>. For more information, see <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon
|
|
823
|
-
* Resource Names (ARNs) and AWS Service Namespaces</a>.</p>
|
|
824
|
-
*
|
|
825
|
-
* <p>Specify either <code>ClusterEndpoint</code> or <code>DomainARN</code>.</p>
|
|
826
|
-
*/
|
|
358
|
+
|
|
827
359
|
DomainARN?: string;
|
|
828
|
-
|
|
829
|
-
* <p>The endpoint to use when communicating with the cluster. Specify either this
|
|
830
|
-
* <code>ClusterEndpoint</code> or the <code>DomainARN</code> field.</p>
|
|
831
|
-
*/
|
|
360
|
+
|
|
832
361
|
ClusterEndpoint?: string;
|
|
833
|
-
|
|
834
|
-
* <p>The Elasticsearch index name.</p>
|
|
835
|
-
*/
|
|
362
|
+
|
|
836
363
|
IndexName: string | undefined;
|
|
837
|
-
|
|
838
|
-
* <p>The Elasticsearch type name. For Elasticsearch 6.x, there can be only one type per
|
|
839
|
-
* index. If you try to specify a new type for an existing index that already has another
|
|
840
|
-
* type, Kinesis Data Firehose returns an error during run time.</p>
|
|
841
|
-
*
|
|
842
|
-
* <p>For Elasticsearch 7.x, don't specify a <code>TypeName</code>.</p>
|
|
843
|
-
*/
|
|
364
|
+
|
|
844
365
|
TypeName?: string;
|
|
845
|
-
|
|
846
|
-
* <p>The Elasticsearch index rotation period. Index rotation appends a timestamp to the
|
|
847
|
-
* <code>IndexName</code> to facilitate the expiration of old data. For more information,
|
|
848
|
-
* see <a href="https://docs.aws.amazon.com/firehose/latest/dev/basic-deliver.html#es-index-rotation">Index Rotation for the
|
|
849
|
-
* Amazon ES Destination</a>. The default value is <code>OneDay</code>.</p>
|
|
850
|
-
*/
|
|
366
|
+
|
|
851
367
|
IndexRotationPeriod?: ElasticsearchIndexRotationPeriod | string;
|
|
852
|
-
|
|
853
|
-
* <p>The buffering options. If no value is specified, the default values for
|
|
854
|
-
* <code>ElasticsearchBufferingHints</code> are used.</p>
|
|
855
|
-
*/
|
|
368
|
+
|
|
856
369
|
BufferingHints?: ElasticsearchBufferingHints;
|
|
857
|
-
|
|
858
|
-
* <p>The retry behavior in case Kinesis Data Firehose is unable to deliver documents to
|
|
859
|
-
* Amazon ES. The default value is 300 (5 minutes).</p>
|
|
860
|
-
*/
|
|
370
|
+
|
|
861
371
|
RetryOptions?: ElasticsearchRetryOptions;
|
|
862
|
-
|
|
863
|
-
* <p>Defines how documents should be delivered to Amazon S3. When it is set to
|
|
864
|
-
* <code>FailedDocumentsOnly</code>, Kinesis Data Firehose writes any documents that could
|
|
865
|
-
* not be indexed to the configured Amazon S3 destination, with
|
|
866
|
-
* <code>elasticsearch-failed/</code> appended to the key prefix. When set to
|
|
867
|
-
* <code>AllDocuments</code>, Kinesis Data Firehose delivers all incoming records to Amazon
|
|
868
|
-
* S3, and also writes failed documents with <code>elasticsearch-failed/</code> appended to
|
|
869
|
-
* the prefix. For more information, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/basic-deliver.html#es-s3-backup">Amazon S3 Backup for the
|
|
870
|
-
* Amazon ES Destination</a>. Default value is
|
|
871
|
-
* <code>FailedDocumentsOnly</code>.</p>
|
|
872
|
-
* <p>You can't change this backup mode after you create the delivery stream. </p>
|
|
873
|
-
*/
|
|
372
|
+
|
|
874
373
|
S3BackupMode?: ElasticsearchS3BackupMode | string;
|
|
875
|
-
|
|
876
|
-
* <p>The configuration for the backup Amazon S3 location.</p>
|
|
877
|
-
*/
|
|
374
|
+
|
|
878
375
|
S3Configuration: S3DestinationConfiguration | undefined;
|
|
879
|
-
|
|
880
|
-
* <p>The data processing configuration.</p>
|
|
881
|
-
*/
|
|
376
|
+
|
|
882
377
|
ProcessingConfiguration?: ProcessingConfiguration;
|
|
883
|
-
|
|
884
|
-
* <p>The Amazon CloudWatch logging options for your delivery stream.</p>
|
|
885
|
-
*/
|
|
378
|
+
|
|
886
379
|
CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
|
|
887
|
-
|
|
888
|
-
* <p>The details of the VPC of the Amazon ES destination.</p>
|
|
889
|
-
*/
|
|
380
|
+
|
|
890
381
|
VpcConfiguration?: VpcConfiguration;
|
|
891
382
|
}
|
|
892
383
|
export declare namespace ElasticsearchDestinationConfiguration {
|
|
893
|
-
|
|
894
|
-
* @internal
|
|
895
|
-
*/
|
|
384
|
+
|
|
896
385
|
const filterSensitiveLog: (obj: ElasticsearchDestinationConfiguration) => any;
|
|
897
386
|
}
|
|
898
|
-
|
|
899
|
-
* <p>The native Hive / HCatalog JsonSerDe. Used by Kinesis Data Firehose for deserializing
|
|
900
|
-
* data, which means converting it from the JSON format in preparation for serializing it to
|
|
901
|
-
* the Parquet or ORC format. This is one of two deserializers you can choose, depending on
|
|
902
|
-
* which one offers the functionality you need. The other option is the OpenX SerDe.</p>
|
|
903
|
-
*/
|
|
387
|
+
|
|
904
388
|
export interface HiveJsonSerDe {
|
|
905
|
-
|
|
906
|
-
* <p>Indicates how you want Kinesis Data Firehose to parse the date and timestamps that
|
|
907
|
-
* may be present in your input data JSON. To specify these format strings, follow the pattern
|
|
908
|
-
* syntax of JodaTime's DateTimeFormat format strings. For more information, see <a href="https://www.joda.org/joda-time/apidocs/org/joda/time/format/DateTimeFormat.html">Class DateTimeFormat</a>. You can also use the special value <code>millis</code> to
|
|
909
|
-
* parse timestamps in epoch milliseconds. If you don't specify a format, Kinesis Data
|
|
910
|
-
* Firehose uses <code>java.sql.Timestamp::valueOf</code> by default.</p>
|
|
911
|
-
*/
|
|
389
|
+
|
|
912
390
|
TimestampFormats?: string[];
|
|
913
391
|
}
|
|
914
392
|
export declare namespace HiveJsonSerDe {
|
|
915
|
-
|
|
916
|
-
* @internal
|
|
917
|
-
*/
|
|
393
|
+
|
|
918
394
|
const filterSensitiveLog: (obj: HiveJsonSerDe) => any;
|
|
919
395
|
}
|
|
920
|
-
|
|
921
|
-
* <p>The OpenX SerDe. Used by Kinesis Data Firehose for deserializing data, which means
|
|
922
|
-
* converting it from the JSON format in preparation for serializing it to the Parquet or ORC
|
|
923
|
-
* format. This is one of two deserializers you can choose, depending on which one offers the
|
|
924
|
-
* functionality you need. The other option is the native Hive / HCatalog JsonSerDe.</p>
|
|
925
|
-
*/
|
|
396
|
+
|
|
926
397
|
export interface OpenXJsonSerDe {
|
|
927
|
-
|
|
928
|
-
* <p>When set to <code>true</code>, specifies that the names of the keys include dots and
|
|
929
|
-
* that you want Kinesis Data Firehose to replace them with underscores. This is useful
|
|
930
|
-
* because Apache Hive does not allow dots in column names. For example, if the JSON contains
|
|
931
|
-
* a key whose name is "a.b", you can define the column name to be "a_b" when using this
|
|
932
|
-
* option.</p>
|
|
933
|
-
* <p>The default is <code>false</code>.</p>
|
|
934
|
-
*/
|
|
398
|
+
|
|
935
399
|
ConvertDotsInJsonKeysToUnderscores?: boolean;
|
|
936
|
-
|
|
937
|
-
* <p>When set to <code>true</code>, which is the default, Kinesis Data Firehose converts
|
|
938
|
-
* JSON keys to lowercase before deserializing them.</p>
|
|
939
|
-
*/
|
|
400
|
+
|
|
940
401
|
CaseInsensitive?: boolean;
|
|
941
|
-
|
|
942
|
-
* <p>Maps column names to JSON keys that aren't identical to the column names. This is
|
|
943
|
-
* useful when the JSON contains keys that are Hive keywords. For example,
|
|
944
|
-
* <code>timestamp</code> is a Hive keyword. If you have a JSON key named
|
|
945
|
-
* <code>timestamp</code>, set this parameter to <code>{"ts": "timestamp"}</code> to map
|
|
946
|
-
* this key to a column named <code>ts</code>.</p>
|
|
947
|
-
*/
|
|
402
|
+
|
|
948
403
|
ColumnToJsonKeyMappings?: {
|
|
949
404
|
[key: string]: string;
|
|
950
405
|
};
|
|
951
406
|
}
|
|
952
407
|
export declare namespace OpenXJsonSerDe {
|
|
953
|
-
|
|
954
|
-
* @internal
|
|
955
|
-
*/
|
|
408
|
+
|
|
956
409
|
const filterSensitiveLog: (obj: OpenXJsonSerDe) => any;
|
|
957
410
|
}
|
|
958
|
-
|
|
959
|
-
* <p>The deserializer you want Kinesis Data Firehose to use for converting the input data
|
|
960
|
-
* from JSON. Kinesis Data Firehose then serializes the data to its final format using the
|
|
961
|
-
* <a>Serializer</a>. Kinesis Data Firehose supports two types of deserializers:
|
|
962
|
-
* the <a href="https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-JSON">Apache Hive JSON SerDe</a> and the <a href="https://github.com/rcongiu/Hive-JSON-Serde">OpenX JSON SerDe</a>.</p>
|
|
963
|
-
*/
|
|
411
|
+
|
|
964
412
|
export interface Deserializer {
|
|
965
|
-
|
|
966
|
-
* <p>The OpenX SerDe. Used by Kinesis Data Firehose for deserializing data, which means
|
|
967
|
-
* converting it from the JSON format in preparation for serializing it to the Parquet or ORC
|
|
968
|
-
* format. This is one of two deserializers you can choose, depending on which one offers the
|
|
969
|
-
* functionality you need. The other option is the native Hive / HCatalog JsonSerDe.</p>
|
|
970
|
-
*/
|
|
413
|
+
|
|
971
414
|
OpenXJsonSerDe?: OpenXJsonSerDe;
|
|
972
|
-
|
|
973
|
-
* <p>The native Hive / HCatalog JsonSerDe. Used by Kinesis Data Firehose for deserializing
|
|
974
|
-
* data, which means converting it from the JSON format in preparation for serializing it to
|
|
975
|
-
* the Parquet or ORC format. This is one of two deserializers you can choose, depending on
|
|
976
|
-
* which one offers the functionality you need. The other option is the OpenX SerDe.</p>
|
|
977
|
-
*/
|
|
415
|
+
|
|
978
416
|
HiveJsonSerDe?: HiveJsonSerDe;
|
|
979
417
|
}
|
|
980
418
|
export declare namespace Deserializer {
|
|
981
|
-
|
|
982
|
-
* @internal
|
|
983
|
-
*/
|
|
419
|
+
|
|
984
420
|
const filterSensitiveLog: (obj: Deserializer) => any;
|
|
985
421
|
}
|
|
986
|
-
|
|
987
|
-
* <p>Specifies the deserializer you want to use to convert the format of the input data.
|
|
988
|
-
* This parameter is required if <code>Enabled</code> is set to true.</p>
|
|
989
|
-
*/
|
|
422
|
+
|
|
990
423
|
export interface InputFormatConfiguration {
|
|
991
|
-
|
|
992
|
-
* <p>Specifies which deserializer to use. You can choose either the Apache Hive JSON SerDe
|
|
993
|
-
* or the OpenX JSON SerDe. If both are non-null, the server rejects the request.</p>
|
|
994
|
-
*/
|
|
424
|
+
|
|
995
425
|
Deserializer?: Deserializer;
|
|
996
426
|
}
|
|
997
427
|
export declare namespace InputFormatConfiguration {
|
|
998
|
-
|
|
999
|
-
* @internal
|
|
1000
|
-
*/
|
|
428
|
+
|
|
1001
429
|
const filterSensitiveLog: (obj: InputFormatConfiguration) => any;
|
|
1002
430
|
}
|
|
1003
431
|
export declare enum OrcCompression {
|
|
@@ -1009,76 +437,31 @@ export declare enum OrcFormatVersion {
|
|
|
1009
437
|
V0_11 = "V0_11",
|
|
1010
438
|
V0_12 = "V0_12"
|
|
1011
439
|
}
|
|
1012
|
-
|
|
1013
|
-
* <p>A serializer to use for converting data to the ORC format before storing it in Amazon
|
|
1014
|
-
* S3. For more information, see <a href="https://orc.apache.org/docs/">Apache
|
|
1015
|
-
* ORC</a>.</p>
|
|
1016
|
-
*/
|
|
440
|
+
|
|
1017
441
|
export interface OrcSerDe {
|
|
1018
|
-
|
|
1019
|
-
* <p>The number of bytes in each stripe. The default is 64 MiB and the minimum is 8
|
|
1020
|
-
* MiB.</p>
|
|
1021
|
-
*/
|
|
442
|
+
|
|
1022
443
|
StripeSizeBytes?: number;
|
|
1023
|
-
|
|
1024
|
-
* <p>The Hadoop Distributed File System (HDFS) block size. This is useful if you intend to
|
|
1025
|
-
* copy the data from Amazon S3 to HDFS before querying. The default is 256 MiB and the
|
|
1026
|
-
* minimum is 64 MiB. Kinesis Data Firehose uses this value for padding calculations.</p>
|
|
1027
|
-
*/
|
|
444
|
+
|
|
1028
445
|
BlockSizeBytes?: number;
|
|
1029
|
-
|
|
1030
|
-
* <p>The number of rows between index entries. The default is 10,000 and the minimum is
|
|
1031
|
-
* 1,000.</p>
|
|
1032
|
-
*/
|
|
446
|
+
|
|
1033
447
|
RowIndexStride?: number;
|
|
1034
|
-
|
|
1035
|
-
* <p>Set this to <code>true</code> to indicate that you want stripes to be padded to the HDFS
|
|
1036
|
-
* block boundaries. This is useful if you intend to copy the data from Amazon S3 to HDFS
|
|
1037
|
-
* before querying. The default is <code>false</code>.</p>
|
|
1038
|
-
*/
|
|
448
|
+
|
|
1039
449
|
EnablePadding?: boolean;
|
|
1040
|
-
|
|
1041
|
-
* <p>A number between 0 and 1 that defines the tolerance for block padding as a decimal
|
|
1042
|
-
* fraction of stripe size. The default value is 0.05, which means 5 percent of stripe
|
|
1043
|
-
* size.</p>
|
|
1044
|
-
* <p>For the default values of 64 MiB ORC stripes and 256 MiB HDFS blocks, the default block
|
|
1045
|
-
* padding tolerance of 5 percent reserves a maximum of 3.2 MiB for padding within the 256 MiB
|
|
1046
|
-
* block. In such a case, if the available size within the block is more than 3.2 MiB, a new,
|
|
1047
|
-
* smaller stripe is inserted to fit within that space. This ensures that no stripe crosses
|
|
1048
|
-
* block boundaries and causes remote reads within a node-local task.</p>
|
|
1049
|
-
* <p>Kinesis Data Firehose ignores this parameter when <a>OrcSerDe$EnablePadding</a> is <code>false</code>.</p>
|
|
1050
|
-
*/
|
|
450
|
+
|
|
1051
451
|
PaddingTolerance?: number;
|
|
1052
|
-
|
|
1053
|
-
* <p>The compression code to use over data blocks. The default is <code>SNAPPY</code>.</p>
|
|
1054
|
-
*/
|
|
452
|
+
|
|
1055
453
|
Compression?: OrcCompression | string;
|
|
1056
|
-
|
|
1057
|
-
* <p>The column names for which you want Kinesis Data Firehose to create bloom filters. The
|
|
1058
|
-
* default is <code>null</code>.</p>
|
|
1059
|
-
*/
|
|
454
|
+
|
|
1060
455
|
BloomFilterColumns?: string[];
|
|
1061
|
-
|
|
1062
|
-
* <p>The Bloom filter false positive probability (FPP). The lower the FPP, the bigger the
|
|
1063
|
-
* Bloom filter. The default value is 0.05, the minimum is 0, and the maximum is 1.</p>
|
|
1064
|
-
*/
|
|
456
|
+
|
|
1065
457
|
BloomFilterFalsePositiveProbability?: number;
|
|
1066
|
-
|
|
1067
|
-
* <p>Represents the fraction of the total number of non-null rows. To turn off dictionary
|
|
1068
|
-
* encoding, set this fraction to a number that is less than the number of distinct keys in a
|
|
1069
|
-
* dictionary. To always use dictionary encoding, set this threshold to 1.</p>
|
|
1070
|
-
*/
|
|
458
|
+
|
|
1071
459
|
DictionaryKeyThreshold?: number;
|
|
1072
|
-
|
|
1073
|
-
* <p>The version of the file to write. The possible values are <code>V0_11</code> and
|
|
1074
|
-
* <code>V0_12</code>. The default is <code>V0_12</code>.</p>
|
|
1075
|
-
*/
|
|
460
|
+
|
|
1076
461
|
FormatVersion?: OrcFormatVersion | string;
|
|
1077
462
|
}
|
|
1078
463
|
export declare namespace OrcSerDe {
|
|
1079
|
-
|
|
1080
|
-
* @internal
|
|
1081
|
-
*/
|
|
464
|
+
|
|
1082
465
|
const filterSensitiveLog: (obj: OrcSerDe) => any;
|
|
1083
466
|
}
|
|
1084
467
|
export declare enum ParquetCompression {
|
|
@@ -1090,818 +473,360 @@ export declare enum ParquetWriterVersion {
|
|
|
1090
473
|
V1 = "V1",
|
|
1091
474
|
V2 = "V2"
|
|
1092
475
|
}
|
|
1093
|
-
|
|
1094
|
-
* <p>A serializer to use for converting data to the Parquet format before storing it in
|
|
1095
|
-
* Amazon S3. For more information, see <a href="https://parquet.apache.org/documentation/latest/">Apache Parquet</a>.</p>
|
|
1096
|
-
*/
|
|
476
|
+
|
|
1097
477
|
export interface ParquetSerDe {
|
|
1098
|
-
|
|
1099
|
-
* <p>The Hadoop Distributed File System (HDFS) block size. This is useful if you intend to
|
|
1100
|
-
* copy the data from Amazon S3 to HDFS before querying. The default is 256 MiB and the
|
|
1101
|
-
* minimum is 64 MiB. Kinesis Data Firehose uses this value for padding calculations.</p>
|
|
1102
|
-
*/
|
|
478
|
+
|
|
1103
479
|
BlockSizeBytes?: number;
|
|
1104
|
-
|
|
1105
|
-
* <p>The Parquet page size. Column chunks are divided into pages. A page is conceptually an
|
|
1106
|
-
* indivisible unit (in terms of compression and encoding). The minimum value is 64 KiB and
|
|
1107
|
-
* the default is 1 MiB.</p>
|
|
1108
|
-
*/
|
|
480
|
+
|
|
1109
481
|
PageSizeBytes?: number;
|
|
1110
|
-
|
|
1111
|
-
* <p>The compression code to use over data blocks. The possible values are
|
|
1112
|
-
* <code>UNCOMPRESSED</code>, <code>SNAPPY</code>, and <code>GZIP</code>, with the default
|
|
1113
|
-
* being <code>SNAPPY</code>. Use <code>SNAPPY</code> for higher decompression speed. Use
|
|
1114
|
-
* <code>GZIP</code> if the compression ratio is more important than speed.</p>
|
|
1115
|
-
*/
|
|
482
|
+
|
|
1116
483
|
Compression?: ParquetCompression | string;
|
|
1117
|
-
|
|
1118
|
-
* <p>Indicates whether to enable dictionary compression.</p>
|
|
1119
|
-
*/
|
|
484
|
+
|
|
1120
485
|
EnableDictionaryCompression?: boolean;
|
|
1121
|
-
|
|
1122
|
-
* <p>The maximum amount of padding to apply. This is useful if you intend to copy the data
|
|
1123
|
-
* from Amazon S3 to HDFS before querying. The default is 0.</p>
|
|
1124
|
-
*/
|
|
486
|
+
|
|
1125
487
|
MaxPaddingBytes?: number;
|
|
1126
|
-
|
|
1127
|
-
* <p>Indicates the version of row format to output. The possible values are <code>V1</code>
|
|
1128
|
-
* and <code>V2</code>. The default is <code>V1</code>.</p>
|
|
1129
|
-
*/
|
|
488
|
+
|
|
1130
489
|
WriterVersion?: ParquetWriterVersion | string;
|
|
1131
490
|
}
|
|
1132
491
|
export declare namespace ParquetSerDe {
|
|
1133
|
-
|
|
1134
|
-
* @internal
|
|
1135
|
-
*/
|
|
492
|
+
|
|
1136
493
|
const filterSensitiveLog: (obj: ParquetSerDe) => any;
|
|
1137
494
|
}
|
|
1138
|
-
|
|
1139
|
-
* <p>The serializer that you want Kinesis Data Firehose to use to convert data to the target
|
|
1140
|
-
* format before writing it to Amazon S3. Kinesis Data Firehose supports two types of
|
|
1141
|
-
* serializers: the <a href="https://hive.apache.org/javadocs/r1.2.2/api/org/apache/hadoop/hive/ql/io/orc/OrcSerde.html">ORC SerDe</a> and the <a href="https://hive.apache.org/javadocs/r1.2.2/api/org/apache/hadoop/hive/ql/io/parquet/serde/ParquetHiveSerDe.html">Parquet SerDe</a>.</p>
|
|
1142
|
-
*/
|
|
495
|
+
|
|
1143
496
|
export interface Serializer {
|
|
1144
|
-
|
|
1145
|
-
* <p>A serializer to use for converting data to the Parquet format before storing it in
|
|
1146
|
-
* Amazon S3. For more information, see <a href="https://parquet.apache.org/documentation/latest/">Apache Parquet</a>.</p>
|
|
1147
|
-
*/
|
|
497
|
+
|
|
1148
498
|
ParquetSerDe?: ParquetSerDe;
|
|
1149
|
-
|
|
1150
|
-
* <p>A serializer to use for converting data to the ORC format before storing it in Amazon
|
|
1151
|
-
* S3. For more information, see <a href="https://orc.apache.org/docs/">Apache
|
|
1152
|
-
* ORC</a>.</p>
|
|
1153
|
-
*/
|
|
499
|
+
|
|
1154
500
|
OrcSerDe?: OrcSerDe;
|
|
1155
501
|
}
|
|
1156
502
|
export declare namespace Serializer {
|
|
1157
|
-
|
|
1158
|
-
* @internal
|
|
1159
|
-
*/
|
|
503
|
+
|
|
1160
504
|
const filterSensitiveLog: (obj: Serializer) => any;
|
|
1161
505
|
}
|
|
1162
|
-
|
|
1163
|
-
* <p>Specifies the serializer that you want Kinesis Data Firehose to use to convert the
|
|
1164
|
-
* format of your data before it writes it to Amazon S3. This parameter is required if
|
|
1165
|
-
* <code>Enabled</code> is set to true.</p>
|
|
1166
|
-
*/
|
|
506
|
+
|
|
1167
507
|
export interface OutputFormatConfiguration {
|
|
1168
|
-
|
|
1169
|
-
* <p>Specifies which serializer to use. You can choose either the ORC SerDe or the Parquet
|
|
1170
|
-
* SerDe. If both are non-null, the server rejects the request.</p>
|
|
1171
|
-
*/
|
|
508
|
+
|
|
1172
509
|
Serializer?: Serializer;
|
|
1173
510
|
}
|
|
1174
511
|
export declare namespace OutputFormatConfiguration {
|
|
1175
|
-
|
|
1176
|
-
* @internal
|
|
1177
|
-
*/
|
|
512
|
+
|
|
1178
513
|
const filterSensitiveLog: (obj: OutputFormatConfiguration) => any;
|
|
1179
514
|
}
|
|
1180
|
-
|
|
1181
|
-
* <p>Specifies the schema to which you want Kinesis Data Firehose to configure your data
|
|
1182
|
-
* before it writes it to Amazon S3. This parameter is required if <code>Enabled</code> is set
|
|
1183
|
-
* to true.</p>
|
|
1184
|
-
*/
|
|
515
|
+
|
|
1185
516
|
export interface SchemaConfiguration {
|
|
1186
|
-
|
|
1187
|
-
* <p>The role that Kinesis Data Firehose can use to access AWS Glue. This role must be in
|
|
1188
|
-
* the same account you use for Kinesis Data Firehose. Cross-account roles aren't
|
|
1189
|
-
* allowed.</p>
|
|
1190
|
-
* <important>
|
|
1191
|
-
* <p>If the <code>SchemaConfiguration</code> request parameter is used as part of invoking
|
|
1192
|
-
* the <code>CreateDeliveryStream</code> API, then the <code>RoleARN</code> property is
|
|
1193
|
-
* required and its value must be specified.</p>
|
|
1194
|
-
* </important>
|
|
1195
|
-
*/
|
|
517
|
+
|
|
1196
518
|
RoleARN?: string;
|
|
1197
|
-
|
|
1198
|
-
* <p>The ID of the AWS Glue Data Catalog. If you don't supply this, the AWS account ID is
|
|
1199
|
-
* used by default.</p>
|
|
1200
|
-
*/
|
|
519
|
+
|
|
1201
520
|
CatalogId?: string;
|
|
1202
|
-
|
|
1203
|
-
* <p>Specifies the name of the AWS Glue database that contains the schema for the output
|
|
1204
|
-
* data.</p>
|
|
1205
|
-
* <important>
|
|
1206
|
-
* <p>If the <code>SchemaConfiguration</code> request parameter is used as part of invoking
|
|
1207
|
-
* the <code>CreateDeliveryStream</code> API, then the <code>DatabaseName</code> property
|
|
1208
|
-
* is required and its value must be specified.</p>
|
|
1209
|
-
* </important>
|
|
1210
|
-
*/
|
|
521
|
+
|
|
1211
522
|
DatabaseName?: string;
|
|
1212
|
-
|
|
1213
|
-
* <p>Specifies the AWS Glue table that contains the column information that constitutes your
|
|
1214
|
-
* data schema.</p>
|
|
1215
|
-
* <important>
|
|
1216
|
-
* <p>If the <code>SchemaConfiguration</code> request parameter is used as part of invoking
|
|
1217
|
-
* the <code>CreateDeliveryStream</code> API, then the <code>TableName</code> property is
|
|
1218
|
-
* required and its value must be specified.</p>
|
|
1219
|
-
* </important>
|
|
1220
|
-
*/
|
|
523
|
+
|
|
1221
524
|
TableName?: string;
|
|
1222
|
-
|
|
1223
|
-
* <p>If you don't specify an AWS Region, the default is the current Region.</p>
|
|
1224
|
-
*/
|
|
525
|
+
|
|
1225
526
|
Region?: string;
|
|
1226
|
-
|
|
1227
|
-
* <p>Specifies the table version for the output data schema. If you don't specify this
|
|
1228
|
-
* version ID, or if you set it to <code>LATEST</code>, Kinesis Data Firehose uses the most
|
|
1229
|
-
* recent version. This means that any updates to the table are automatically picked
|
|
1230
|
-
* up.</p>
|
|
1231
|
-
*/
|
|
527
|
+
|
|
1232
528
|
VersionId?: string;
|
|
1233
529
|
}
|
|
1234
530
|
export declare namespace SchemaConfiguration {
|
|
1235
|
-
|
|
1236
|
-
* @internal
|
|
1237
|
-
*/
|
|
531
|
+
|
|
1238
532
|
const filterSensitiveLog: (obj: SchemaConfiguration) => any;
|
|
1239
533
|
}
|
|
1240
|
-
|
|
1241
|
-
* <p>Specifies that you want Kinesis Data Firehose to convert data from the JSON format to
|
|
1242
|
-
* the Parquet or ORC format before writing it to Amazon S3. Kinesis Data Firehose uses the
|
|
1243
|
-
* serializer and deserializer that you specify, in addition to the column information from
|
|
1244
|
-
* the AWS Glue table, to deserialize your input data from JSON and then serialize it to the
|
|
1245
|
-
* Parquet or ORC format. For more information, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/record-format-conversion.html">Kinesis Data Firehose Record
|
|
1246
|
-
* Format Conversion</a>.</p>
|
|
1247
|
-
*/
|
|
534
|
+
|
|
1248
535
|
export interface DataFormatConversionConfiguration {
|
|
1249
|
-
|
|
1250
|
-
* <p>Specifies the AWS Glue Data Catalog table that contains the column information. This
|
|
1251
|
-
* parameter is required if <code>Enabled</code> is set to true.</p>
|
|
1252
|
-
*/
|
|
536
|
+
|
|
1253
537
|
SchemaConfiguration?: SchemaConfiguration;
|
|
1254
|
-
|
|
1255
|
-
* <p>Specifies the deserializer that you want Kinesis Data Firehose to use to convert the
|
|
1256
|
-
* format of your data from JSON. This parameter is required if <code>Enabled</code> is set to
|
|
1257
|
-
* true.</p>
|
|
1258
|
-
*/
|
|
538
|
+
|
|
1259
539
|
InputFormatConfiguration?: InputFormatConfiguration;
|
|
1260
|
-
|
|
1261
|
-
* <p>Specifies the serializer that you want Kinesis Data Firehose to use to convert the
|
|
1262
|
-
* format of your data to the Parquet or ORC format. This parameter is required if
|
|
1263
|
-
* <code>Enabled</code> is set to true.</p>
|
|
1264
|
-
*/
|
|
540
|
+
|
|
1265
541
|
OutputFormatConfiguration?: OutputFormatConfiguration;
|
|
1266
|
-
|
|
1267
|
-
* <p>Defaults to <code>true</code>. Set it to <code>false</code> if you want to disable
|
|
1268
|
-
* format conversion while preserving the configuration details.</p>
|
|
1269
|
-
*/
|
|
542
|
+
|
|
1270
543
|
Enabled?: boolean;
|
|
1271
544
|
}
|
|
1272
545
|
export declare namespace DataFormatConversionConfiguration {
|
|
1273
|
-
|
|
1274
|
-
* @internal
|
|
1275
|
-
*/
|
|
546
|
+
|
|
1276
547
|
const filterSensitiveLog: (obj: DataFormatConversionConfiguration) => any;
|
|
1277
548
|
}
|
|
1278
|
-
|
|
1279
|
-
* <p> The retry behavior in case Kinesis Data Firehose is unable to deliver data to an Amazon
|
|
1280
|
-
* S3 prefix.</p>
|
|
1281
|
-
*/
|
|
549
|
+
|
|
1282
550
|
export interface RetryOptions {
|
|
1283
|
-
|
|
1284
|
-
* <p>The period of time during which Kinesis Data Firehose retries to deliver data to the
|
|
1285
|
-
* specified Amazon S3 prefix.</p>
|
|
1286
|
-
*/
|
|
551
|
+
|
|
1287
552
|
DurationInSeconds?: number;
|
|
1288
553
|
}
|
|
1289
554
|
export declare namespace RetryOptions {
|
|
1290
|
-
|
|
1291
|
-
* @internal
|
|
1292
|
-
*/
|
|
555
|
+
|
|
1293
556
|
const filterSensitiveLog: (obj: RetryOptions) => any;
|
|
1294
557
|
}
|
|
1295
|
-
|
|
1296
|
-
* <p>The configuration of the dynamic partitioning mechanism that creates smaller data sets
|
|
1297
|
-
* from the streaming data by partitioning it based on partition keys. Currently, dynamic
|
|
1298
|
-
* partitioning is only supported for Amazon S3 destinations. For more information, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/dynamic-partitioning.html">https://docs.aws.amazon.com/firehose/latest/dev/dynamic-partitioning.html</a>
|
|
1299
|
-
* </p>
|
|
1300
|
-
*/
|
|
558
|
+
|
|
1301
559
|
export interface DynamicPartitioningConfiguration {
|
|
1302
|
-
|
|
1303
|
-
* <p>The retry behavior in case Kinesis Data Firehose is unable to deliver data to an Amazon
|
|
1304
|
-
* S3 prefix.</p>
|
|
1305
|
-
*/
|
|
560
|
+
|
|
1306
561
|
RetryOptions?: RetryOptions;
|
|
1307
|
-
|
|
1308
|
-
* <p>Specifies that the dynamic partitioning is enabled for this Kinesis Data Firehose
|
|
1309
|
-
* delivery stream.</p>
|
|
1310
|
-
*/
|
|
562
|
+
|
|
1311
563
|
Enabled?: boolean;
|
|
1312
564
|
}
|
|
1313
565
|
export declare namespace DynamicPartitioningConfiguration {
|
|
1314
|
-
|
|
1315
|
-
* @internal
|
|
1316
|
-
*/
|
|
566
|
+
|
|
1317
567
|
const filterSensitiveLog: (obj: DynamicPartitioningConfiguration) => any;
|
|
1318
568
|
}
|
|
1319
569
|
export declare type S3BackupMode = "Disabled" | "Enabled";
|
|
1320
|
-
|
|
1321
|
-
* <p>Describes the configuration of a destination in Amazon S3.</p>
|
|
1322
|
-
*/
|
|
570
|
+
|
|
1323
571
|
export interface ExtendedS3DestinationConfiguration {
|
|
1324
|
-
|
|
1325
|
-
* <p>The Amazon Resource Name (ARN) of the AWS credentials. For more information, see
|
|
1326
|
-
* <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon
|
|
1327
|
-
* Resource Names (ARNs) and AWS Service Namespaces</a>.</p>
|
|
1328
|
-
*/
|
|
572
|
+
|
|
1329
573
|
RoleARN: string | undefined;
|
|
1330
|
-
|
|
1331
|
-
* <p>The ARN of the S3 bucket. For more information, see <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon Resource Names (ARNs) and
|
|
1332
|
-
* AWS Service Namespaces</a>.</p>
|
|
1333
|
-
*/
|
|
574
|
+
|
|
1334
575
|
BucketARN: string | undefined;
|
|
1335
|
-
|
|
1336
|
-
* <p>The "YYYY/MM/DD/HH" time format prefix is automatically used for delivered Amazon S3
|
|
1337
|
-
* files. You can also specify a custom prefix, as described in <a href="https://docs.aws.amazon.com/firehose/latest/dev/s3-prefixes.html">Custom Prefixes for Amazon S3
|
|
1338
|
-
* Objects</a>.</p>
|
|
1339
|
-
*/
|
|
576
|
+
|
|
1340
577
|
Prefix?: string;
|
|
1341
|
-
|
|
1342
|
-
* <p>A prefix that Kinesis Data Firehose evaluates and adds to failed records before writing
|
|
1343
|
-
* them to S3. This prefix appears immediately following the bucket name. For information
|
|
1344
|
-
* about how to specify this prefix, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/s3-prefixes.html">Custom Prefixes for Amazon S3
|
|
1345
|
-
* Objects</a>.</p>
|
|
1346
|
-
*/
|
|
578
|
+
|
|
1347
579
|
ErrorOutputPrefix?: string;
|
|
1348
|
-
|
|
1349
|
-
* <p>The buffering option.</p>
|
|
1350
|
-
*/
|
|
580
|
+
|
|
1351
581
|
BufferingHints?: BufferingHints;
|
|
1352
|
-
|
|
1353
|
-
* <p>The compression format. If no value is specified, the default is
|
|
1354
|
-
* UNCOMPRESSED.</p>
|
|
1355
|
-
*/
|
|
582
|
+
|
|
1356
583
|
CompressionFormat?: CompressionFormat | string;
|
|
1357
|
-
|
|
1358
|
-
* <p>The encryption configuration. If no value is specified, the default is no
|
|
1359
|
-
* encryption.</p>
|
|
1360
|
-
*/
|
|
584
|
+
|
|
1361
585
|
EncryptionConfiguration?: EncryptionConfiguration;
|
|
1362
|
-
|
|
1363
|
-
* <p>The Amazon CloudWatch logging options for your delivery stream.</p>
|
|
1364
|
-
*/
|
|
586
|
+
|
|
1365
587
|
CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
|
|
1366
|
-
|
|
1367
|
-
* <p>The data processing configuration.</p>
|
|
1368
|
-
*/
|
|
588
|
+
|
|
1369
589
|
ProcessingConfiguration?: ProcessingConfiguration;
|
|
1370
|
-
|
|
1371
|
-
* <p>The Amazon S3 backup mode. After you create a delivery stream, you can update it to
|
|
1372
|
-
* enable Amazon S3 backup if it is disabled. If backup is enabled, you can't update the
|
|
1373
|
-
* delivery stream to disable it. </p>
|
|
1374
|
-
*/
|
|
590
|
+
|
|
1375
591
|
S3BackupMode?: S3BackupMode | string;
|
|
1376
|
-
|
|
1377
|
-
* <p>The configuration for backup in Amazon S3.</p>
|
|
1378
|
-
*/
|
|
592
|
+
|
|
1379
593
|
S3BackupConfiguration?: S3DestinationConfiguration;
|
|
1380
|
-
|
|
1381
|
-
* <p>The serializer, deserializer, and schema for converting data from the JSON format to
|
|
1382
|
-
* the Parquet or ORC format before writing it to Amazon S3.</p>
|
|
1383
|
-
*/
|
|
594
|
+
|
|
1384
595
|
DataFormatConversionConfiguration?: DataFormatConversionConfiguration;
|
|
1385
|
-
|
|
1386
|
-
* <p>The configuration of the dynamic partitioning mechanism that creates smaller data sets
|
|
1387
|
-
* from the streaming data by partitioning it based on partition keys. Currently, dynamic
|
|
1388
|
-
* partitioning is only supported for Amazon S3 destinations. For more information, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/dynamic-partitioning.html">https://docs.aws.amazon.com/firehose/latest/dev/dynamic-partitioning.html</a>
|
|
1389
|
-
* </p>
|
|
1390
|
-
*/
|
|
596
|
+
|
|
1391
597
|
DynamicPartitioningConfiguration?: DynamicPartitioningConfiguration;
|
|
1392
598
|
}
|
|
1393
599
|
export declare namespace ExtendedS3DestinationConfiguration {
|
|
1394
|
-
|
|
1395
|
-
* @internal
|
|
1396
|
-
*/
|
|
600
|
+
|
|
1397
601
|
const filterSensitiveLog: (obj: ExtendedS3DestinationConfiguration) => any;
|
|
1398
602
|
}
|
|
1399
|
-
|
|
1400
|
-
* <p>Describes the buffering options that can be applied before data is delivered to the HTTP
|
|
1401
|
-
* endpoint destination. Kinesis Data Firehose treats these options as hints, and it might
|
|
1402
|
-
* choose to use more optimal values. The <code>SizeInMBs</code> and
|
|
1403
|
-
* <code>IntervalInSeconds</code> parameters are optional. However, if specify a value for
|
|
1404
|
-
* one of them, you must also provide a value for the other. </p>
|
|
1405
|
-
*/
|
|
603
|
+
|
|
1406
604
|
export interface HttpEndpointBufferingHints {
|
|
1407
|
-
|
|
1408
|
-
* <p>Buffer incoming data to the specified size, in MBs, before delivering it to the
|
|
1409
|
-
* destination. The default value is 5. </p>
|
|
1410
|
-
* <p>We recommend setting this parameter to a value greater than the amount of data you
|
|
1411
|
-
* typically ingest into the delivery stream in 10 seconds. For example, if you typically
|
|
1412
|
-
* ingest data at 1 MB/sec, the value should be 10 MB or higher. </p>
|
|
1413
|
-
*/
|
|
605
|
+
|
|
1414
606
|
SizeInMBs?: number;
|
|
1415
|
-
|
|
1416
|
-
* <p>Buffer incoming data for the specified period of time, in seconds, before delivering it
|
|
1417
|
-
* to the destination. The default value is 300 (5 minutes). </p>
|
|
1418
|
-
*/
|
|
607
|
+
|
|
1419
608
|
IntervalInSeconds?: number;
|
|
1420
609
|
}
|
|
1421
610
|
export declare namespace HttpEndpointBufferingHints {
|
|
1422
|
-
|
|
1423
|
-
* @internal
|
|
1424
|
-
*/
|
|
611
|
+
|
|
1425
612
|
const filterSensitiveLog: (obj: HttpEndpointBufferingHints) => any;
|
|
1426
613
|
}
|
|
1427
|
-
|
|
1428
|
-
* <p>Describes the configuration of the HTTP endpoint to which Kinesis Firehose delivers
|
|
1429
|
-
* data.</p>
|
|
1430
|
-
*/
|
|
614
|
+
|
|
1431
615
|
export interface HttpEndpointConfiguration {
|
|
1432
|
-
|
|
1433
|
-
* <p>The URL of the HTTP endpoint selected as the destination.</p>
|
|
1434
|
-
* <important>
|
|
1435
|
-
* <p>If you choose an HTTP endpoint as your destination, review and follow the
|
|
1436
|
-
* instructions in the <a href="https://docs.aws.amazon.com/firehose/latest/dev/httpdeliveryrequestresponse.html">Appendix - HTTP Endpoint
|
|
1437
|
-
* Delivery Request and Response Specifications</a>.</p>
|
|
1438
|
-
* </important>
|
|
1439
|
-
*/
|
|
616
|
+
|
|
1440
617
|
Url: string | undefined;
|
|
1441
|
-
|
|
1442
|
-
* <p>The name of the HTTP endpoint selected as the destination.</p>
|
|
1443
|
-
*/
|
|
618
|
+
|
|
1444
619
|
Name?: string;
|
|
1445
|
-
|
|
1446
|
-
* <p>The access key required for Kinesis Firehose to authenticate with the HTTP endpoint
|
|
1447
|
-
* selected as the destination.</p>
|
|
1448
|
-
*/
|
|
620
|
+
|
|
1449
621
|
AccessKey?: string;
|
|
1450
622
|
}
|
|
1451
623
|
export declare namespace HttpEndpointConfiguration {
|
|
1452
|
-
|
|
1453
|
-
* @internal
|
|
1454
|
-
*/
|
|
624
|
+
|
|
1455
625
|
const filterSensitiveLog: (obj: HttpEndpointConfiguration) => any;
|
|
1456
626
|
}
|
|
1457
|
-
|
|
1458
|
-
* <p>Describes the metadata that's delivered to the specified HTTP endpoint
|
|
1459
|
-
* destination.</p>
|
|
1460
|
-
*/
|
|
627
|
+
|
|
1461
628
|
export interface HttpEndpointCommonAttribute {
|
|
1462
|
-
|
|
1463
|
-
* <p>The name of the HTTP endpoint common attribute.</p>
|
|
1464
|
-
*/
|
|
629
|
+
|
|
1465
630
|
AttributeName: string | undefined;
|
|
1466
|
-
|
|
1467
|
-
* <p>The value of the HTTP endpoint common attribute.</p>
|
|
1468
|
-
*/
|
|
631
|
+
|
|
1469
632
|
AttributeValue: string | undefined;
|
|
1470
633
|
}
|
|
1471
634
|
export declare namespace HttpEndpointCommonAttribute {
|
|
1472
|
-
|
|
1473
|
-
* @internal
|
|
1474
|
-
*/
|
|
635
|
+
|
|
1475
636
|
const filterSensitiveLog: (obj: HttpEndpointCommonAttribute) => any;
|
|
1476
637
|
}
|
|
1477
|
-
|
|
1478
|
-
* <p>The configuration of the HTTP endpoint request.</p>
|
|
1479
|
-
*/
|
|
638
|
+
|
|
1480
639
|
export interface HttpEndpointRequestConfiguration {
|
|
1481
|
-
|
|
1482
|
-
* <p>Kinesis Data Firehose uses the content encoding to compress the body of a request before
|
|
1483
|
-
* sending the request to the destination. For more information, see <a href="https://developer.mozilla.org/en-US/docs/Web/HTTP/Headers/Content-Encoding">Content-Encoding</a> in MDN Web Docs, the official Mozilla documentation.</p>
|
|
1484
|
-
*/
|
|
640
|
+
|
|
1485
641
|
ContentEncoding?: ContentEncoding | string;
|
|
1486
|
-
|
|
1487
|
-
* <p>Describes the metadata sent to the HTTP endpoint destination.</p>
|
|
1488
|
-
*/
|
|
642
|
+
|
|
1489
643
|
CommonAttributes?: HttpEndpointCommonAttribute[];
|
|
1490
644
|
}
|
|
1491
645
|
export declare namespace HttpEndpointRequestConfiguration {
|
|
1492
|
-
|
|
1493
|
-
* @internal
|
|
1494
|
-
*/
|
|
646
|
+
|
|
1495
647
|
const filterSensitiveLog: (obj: HttpEndpointRequestConfiguration) => any;
|
|
1496
648
|
}
|
|
1497
|
-
|
|
1498
|
-
* <p>Describes the retry behavior in case Kinesis Data Firehose is unable to deliver data to
|
|
1499
|
-
* the specified HTTP endpoint destination, or if it doesn't receive a valid acknowledgment of
|
|
1500
|
-
* receipt from the specified HTTP endpoint destination.</p>
|
|
1501
|
-
*/
|
|
649
|
+
|
|
1502
650
|
export interface HttpEndpointRetryOptions {
|
|
1503
|
-
|
|
1504
|
-
* <p>The total amount of time that Kinesis Data Firehose spends on retries. This duration
|
|
1505
|
-
* starts after the initial attempt to send data to the custom destination via HTTPS endpoint
|
|
1506
|
-
* fails. It doesn't include the periods during which Kinesis Data Firehose waits for
|
|
1507
|
-
* acknowledgment from the specified destination after each attempt. </p>
|
|
1508
|
-
*/
|
|
651
|
+
|
|
1509
652
|
DurationInSeconds?: number;
|
|
1510
653
|
}
|
|
1511
654
|
export declare namespace HttpEndpointRetryOptions {
|
|
1512
|
-
|
|
1513
|
-
* @internal
|
|
1514
|
-
*/
|
|
655
|
+
|
|
1515
656
|
const filterSensitiveLog: (obj: HttpEndpointRetryOptions) => any;
|
|
1516
657
|
}
|
|
1517
658
|
export declare type HttpEndpointS3BackupMode = "AllData" | "FailedDataOnly";
|
|
1518
|
-
|
|
1519
|
-
* <p>Describes the configuration of the HTTP endpoint destination.</p>
|
|
1520
|
-
*/
|
|
659
|
+
|
|
1521
660
|
export interface HttpEndpointDestinationConfiguration {
|
|
1522
|
-
|
|
1523
|
-
* <p>The configuration of the HTTP endpoint selected as the destination.</p>
|
|
1524
|
-
*/
|
|
661
|
+
|
|
1525
662
|
EndpointConfiguration: HttpEndpointConfiguration | undefined;
|
|
1526
|
-
|
|
1527
|
-
* <p>The buffering options that can be used before data is delivered to the specified
|
|
1528
|
-
* destination. Kinesis Data Firehose treats these options as hints, and it might choose to
|
|
1529
|
-
* use more optimal values. The <code>SizeInMBs</code> and <code>IntervalInSeconds</code>
|
|
1530
|
-
* parameters are optional. However, if you specify a value for one of them, you must also
|
|
1531
|
-
* provide a value for the other. </p>
|
|
1532
|
-
*/
|
|
663
|
+
|
|
1533
664
|
BufferingHints?: HttpEndpointBufferingHints;
|
|
1534
|
-
|
|
1535
|
-
* <p>Describes the Amazon CloudWatch logging options for your delivery stream.</p>
|
|
1536
|
-
*/
|
|
665
|
+
|
|
1537
666
|
CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
|
|
1538
|
-
|
|
1539
|
-
* <p>The configuration of the requeste sent to the HTTP endpoint specified as the
|
|
1540
|
-
* destination.</p>
|
|
1541
|
-
*/
|
|
667
|
+
|
|
1542
668
|
RequestConfiguration?: HttpEndpointRequestConfiguration;
|
|
1543
|
-
|
|
1544
|
-
* <p>Describes a data processing configuration.</p>
|
|
1545
|
-
*/
|
|
669
|
+
|
|
1546
670
|
ProcessingConfiguration?: ProcessingConfiguration;
|
|
1547
|
-
|
|
1548
|
-
* <p>Kinesis Data Firehose uses this IAM role for all the permissions that the delivery
|
|
1549
|
-
* stream needs.</p>
|
|
1550
|
-
*/
|
|
671
|
+
|
|
1551
672
|
RoleARN?: string;
|
|
1552
|
-
|
|
1553
|
-
* <p>Describes the retry behavior in case Kinesis Data Firehose is unable to deliver data to
|
|
1554
|
-
* the specified HTTP endpoint destination, or if it doesn't receive a valid acknowledgment of
|
|
1555
|
-
* receipt from the specified HTTP endpoint destination.</p>
|
|
1556
|
-
*/
|
|
673
|
+
|
|
1557
674
|
RetryOptions?: HttpEndpointRetryOptions;
|
|
1558
|
-
|
|
1559
|
-
* <p>Describes the S3 bucket backup options for the data that Kinesis Data Firehose delivers
|
|
1560
|
-
* to the HTTP endpoint destination. You can back up all documents (<code>AllData</code>) or
|
|
1561
|
-
* only the documents that Kinesis Data Firehose could not deliver to the specified HTTP
|
|
1562
|
-
* endpoint destination (<code>FailedDataOnly</code>).</p>
|
|
1563
|
-
*/
|
|
675
|
+
|
|
1564
676
|
S3BackupMode?: HttpEndpointS3BackupMode | string;
|
|
1565
|
-
|
|
1566
|
-
* <p>Describes the configuration of a destination in Amazon S3.</p>
|
|
1567
|
-
*/
|
|
677
|
+
|
|
1568
678
|
S3Configuration: S3DestinationConfiguration | undefined;
|
|
1569
679
|
}
|
|
1570
680
|
export declare namespace HttpEndpointDestinationConfiguration {
|
|
1571
|
-
|
|
1572
|
-
* @internal
|
|
1573
|
-
*/
|
|
681
|
+
|
|
1574
682
|
const filterSensitiveLog: (obj: HttpEndpointDestinationConfiguration) => any;
|
|
1575
683
|
}
|
|
1576
|
-
|
|
1577
|
-
* <p>The stream and role Amazon Resource Names (ARNs) for a Kinesis data stream used as
|
|
1578
|
-
* the source for a delivery stream.</p>
|
|
1579
|
-
*/
|
|
684
|
+
|
|
1580
685
|
export interface KinesisStreamSourceConfiguration {
|
|
1581
|
-
|
|
1582
|
-
* <p>The ARN of the source Kinesis data stream. For more information, see <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html#arn-syntax-kinesis-streams">Amazon
|
|
1583
|
-
* Kinesis Data Streams ARN Format</a>.</p>
|
|
1584
|
-
*/
|
|
686
|
+
|
|
1585
687
|
KinesisStreamARN: string | undefined;
|
|
1586
|
-
|
|
1587
|
-
* <p>The ARN of the role that provides access to the source Kinesis data stream. For more
|
|
1588
|
-
* information, see <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html#arn-syntax-iam">AWS Identity and
|
|
1589
|
-
* Access Management (IAM) ARN Format</a>.</p>
|
|
1590
|
-
*/
|
|
688
|
+
|
|
1591
689
|
RoleARN: string | undefined;
|
|
1592
690
|
}
|
|
1593
691
|
export declare namespace KinesisStreamSourceConfiguration {
|
|
1594
|
-
|
|
1595
|
-
* @internal
|
|
1596
|
-
*/
|
|
692
|
+
|
|
1597
693
|
const filterSensitiveLog: (obj: KinesisStreamSourceConfiguration) => any;
|
|
1598
694
|
}
|
|
1599
|
-
|
|
1600
|
-
* <p>Configures retry behavior in case Kinesis Data Firehose is unable to deliver
|
|
1601
|
-
* documents to Amazon Redshift.</p>
|
|
1602
|
-
*/
|
|
695
|
+
|
|
1603
696
|
export interface RedshiftRetryOptions {
|
|
1604
|
-
|
|
1605
|
-
* <p>The length of time during which Kinesis Data Firehose retries delivery after a
|
|
1606
|
-
* failure, starting from the initial request and including the first attempt. The default
|
|
1607
|
-
* value is 3600 seconds (60 minutes). Kinesis Data Firehose does not retry if the value of
|
|
1608
|
-
* <code>DurationInSeconds</code> is 0 (zero) or if the first delivery attempt takes longer
|
|
1609
|
-
* than the current value.</p>
|
|
1610
|
-
*/
|
|
697
|
+
|
|
1611
698
|
DurationInSeconds?: number;
|
|
1612
699
|
}
|
|
1613
700
|
export declare namespace RedshiftRetryOptions {
|
|
1614
|
-
|
|
1615
|
-
* @internal
|
|
1616
|
-
*/
|
|
701
|
+
|
|
1617
702
|
const filterSensitiveLog: (obj: RedshiftRetryOptions) => any;
|
|
1618
703
|
}
|
|
1619
704
|
export declare type RedshiftS3BackupMode = "Disabled" | "Enabled";
|
|
1620
|
-
|
|
1621
|
-
* <p>Describes the configuration of a destination in Amazon Redshift.</p>
|
|
1622
|
-
*/
|
|
705
|
+
|
|
1623
706
|
export interface RedshiftDestinationConfiguration {
|
|
1624
|
-
|
|
1625
|
-
* <p>The Amazon Resource Name (ARN) of the AWS credentials. For more information, see
|
|
1626
|
-
* <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon
|
|
1627
|
-
* Resource Names (ARNs) and AWS Service Namespaces</a>.</p>
|
|
1628
|
-
*/
|
|
707
|
+
|
|
1629
708
|
RoleARN: string | undefined;
|
|
1630
|
-
|
|
1631
|
-
* <p>The database connection string.</p>
|
|
1632
|
-
*/
|
|
709
|
+
|
|
1633
710
|
ClusterJDBCURL: string | undefined;
|
|
1634
|
-
|
|
1635
|
-
* <p>The <code>COPY</code> command.</p>
|
|
1636
|
-
*/
|
|
711
|
+
|
|
1637
712
|
CopyCommand: CopyCommand | undefined;
|
|
1638
|
-
|
|
1639
|
-
* <p>The name of the user.</p>
|
|
1640
|
-
*/
|
|
713
|
+
|
|
1641
714
|
Username: string | undefined;
|
|
1642
|
-
|
|
1643
|
-
* <p>The user password.</p>
|
|
1644
|
-
*/
|
|
715
|
+
|
|
1645
716
|
Password: string | undefined;
|
|
1646
|
-
|
|
1647
|
-
* <p>The retry behavior in case Kinesis Data Firehose is unable to deliver documents to
|
|
1648
|
-
* Amazon Redshift. Default value is 3600 (60 minutes).</p>
|
|
1649
|
-
*/
|
|
717
|
+
|
|
1650
718
|
RetryOptions?: RedshiftRetryOptions;
|
|
1651
|
-
|
|
1652
|
-
* <p>The configuration for the intermediate Amazon S3 location from which Amazon Redshift
|
|
1653
|
-
* obtains data. Restrictions are described in the topic for <a>CreateDeliveryStream</a>.</p>
|
|
1654
|
-
* <p>The compression formats <code>SNAPPY</code> or <code>ZIP</code> cannot be specified
|
|
1655
|
-
* in <code>RedshiftDestinationConfiguration.S3Configuration</code> because the Amazon
|
|
1656
|
-
* Redshift <code>COPY</code> operation that reads from the S3 bucket doesn't support these
|
|
1657
|
-
* compression formats.</p>
|
|
1658
|
-
*/
|
|
719
|
+
|
|
1659
720
|
S3Configuration: S3DestinationConfiguration | undefined;
|
|
1660
|
-
|
|
1661
|
-
* <p>The data processing configuration.</p>
|
|
1662
|
-
*/
|
|
721
|
+
|
|
1663
722
|
ProcessingConfiguration?: ProcessingConfiguration;
|
|
1664
|
-
|
|
1665
|
-
* <p>The Amazon S3 backup mode. After you create a delivery stream, you can update it to
|
|
1666
|
-
* enable Amazon S3 backup if it is disabled. If backup is enabled, you can't update the
|
|
1667
|
-
* delivery stream to disable it. </p>
|
|
1668
|
-
*/
|
|
723
|
+
|
|
1669
724
|
S3BackupMode?: RedshiftS3BackupMode | string;
|
|
1670
|
-
|
|
1671
|
-
* <p>The configuration for backup in Amazon S3.</p>
|
|
1672
|
-
*/
|
|
725
|
+
|
|
1673
726
|
S3BackupConfiguration?: S3DestinationConfiguration;
|
|
1674
|
-
|
|
1675
|
-
* <p>The CloudWatch logging options for your delivery stream.</p>
|
|
1676
|
-
*/
|
|
727
|
+
|
|
1677
728
|
CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
|
|
1678
729
|
}
|
|
1679
730
|
export declare namespace RedshiftDestinationConfiguration {
|
|
1680
|
-
|
|
1681
|
-
* @internal
|
|
1682
|
-
*/
|
|
731
|
+
|
|
1683
732
|
const filterSensitiveLog: (obj: RedshiftDestinationConfiguration) => any;
|
|
1684
733
|
}
|
|
1685
734
|
export declare type HECEndpointType = "Event" | "Raw";
|
|
1686
|
-
|
|
1687
|
-
* <p>Configures retry behavior in case Kinesis Data Firehose is unable to deliver
|
|
1688
|
-
* documents to Splunk, or if it doesn't receive an acknowledgment from Splunk.</p>
|
|
1689
|
-
*/
|
|
735
|
+
|
|
1690
736
|
export interface SplunkRetryOptions {
|
|
1691
|
-
|
|
1692
|
-
* <p>The total amount of time that Kinesis Data Firehose spends on retries. This duration
|
|
1693
|
-
* starts after the initial attempt to send data to Splunk fails. It doesn't include the
|
|
1694
|
-
* periods during which Kinesis Data Firehose waits for acknowledgment from Splunk after each
|
|
1695
|
-
* attempt.</p>
|
|
1696
|
-
*/
|
|
737
|
+
|
|
1697
738
|
DurationInSeconds?: number;
|
|
1698
739
|
}
|
|
1699
740
|
export declare namespace SplunkRetryOptions {
|
|
1700
|
-
|
|
1701
|
-
* @internal
|
|
1702
|
-
*/
|
|
741
|
+
|
|
1703
742
|
const filterSensitiveLog: (obj: SplunkRetryOptions) => any;
|
|
1704
743
|
}
|
|
1705
744
|
export declare type SplunkS3BackupMode = "AllEvents" | "FailedEventsOnly";
|
|
1706
|
-
|
|
1707
|
-
* <p>Describes the configuration of a destination in Splunk.</p>
|
|
1708
|
-
*/
|
|
745
|
+
|
|
1709
746
|
export interface SplunkDestinationConfiguration {
|
|
1710
|
-
|
|
1711
|
-
* <p>The HTTP Event Collector (HEC) endpoint to which Kinesis Data Firehose sends your
|
|
1712
|
-
* data.</p>
|
|
1713
|
-
*/
|
|
747
|
+
|
|
1714
748
|
HECEndpoint: string | undefined;
|
|
1715
|
-
|
|
1716
|
-
* <p>This type can be either "Raw" or "Event."</p>
|
|
1717
|
-
*/
|
|
749
|
+
|
|
1718
750
|
HECEndpointType: HECEndpointType | string | undefined;
|
|
1719
|
-
|
|
1720
|
-
* <p>This is a GUID that you obtain from your Splunk cluster when you create a new HEC
|
|
1721
|
-
* endpoint.</p>
|
|
1722
|
-
*/
|
|
751
|
+
|
|
1723
752
|
HECToken: string | undefined;
|
|
1724
|
-
|
|
1725
|
-
* <p>The amount of time that Kinesis Data Firehose waits to receive an acknowledgment from
|
|
1726
|
-
* Splunk after it sends it data. At the end of the timeout period, Kinesis Data Firehose
|
|
1727
|
-
* either tries to send the data again or considers it an error, based on your retry
|
|
1728
|
-
* settings.</p>
|
|
1729
|
-
*/
|
|
753
|
+
|
|
1730
754
|
HECAcknowledgmentTimeoutInSeconds?: number;
|
|
1731
|
-
|
|
1732
|
-
* <p>The retry behavior in case Kinesis Data Firehose is unable to deliver data to Splunk,
|
|
1733
|
-
* or if it doesn't receive an acknowledgment of receipt from Splunk.</p>
|
|
1734
|
-
*/
|
|
755
|
+
|
|
1735
756
|
RetryOptions?: SplunkRetryOptions;
|
|
1736
|
-
|
|
1737
|
-
* <p>Defines how documents should be delivered to Amazon S3. When set to
|
|
1738
|
-
* <code>FailedEventsOnly</code>, Kinesis Data Firehose writes any data that could not be
|
|
1739
|
-
* indexed to the configured Amazon S3 destination. When set to <code>AllEvents</code>,
|
|
1740
|
-
* Kinesis Data Firehose delivers all incoming records to Amazon S3, and also writes failed
|
|
1741
|
-
* documents to Amazon S3. The default value is <code>FailedEventsOnly</code>.</p>
|
|
1742
|
-
* <p>You can update this backup mode from <code>FailedEventsOnly</code> to
|
|
1743
|
-
* <code>AllEvents</code>. You can't update it from <code>AllEvents</code> to
|
|
1744
|
-
* <code>FailedEventsOnly</code>.</p>
|
|
1745
|
-
*/
|
|
757
|
+
|
|
1746
758
|
S3BackupMode?: SplunkS3BackupMode | string;
|
|
1747
|
-
|
|
1748
|
-
* <p>The configuration for the backup Amazon S3 location.</p>
|
|
1749
|
-
*/
|
|
759
|
+
|
|
1750
760
|
S3Configuration: S3DestinationConfiguration | undefined;
|
|
1751
|
-
|
|
1752
|
-
* <p>The data processing configuration.</p>
|
|
1753
|
-
*/
|
|
761
|
+
|
|
1754
762
|
ProcessingConfiguration?: ProcessingConfiguration;
|
|
1755
|
-
|
|
1756
|
-
* <p>The Amazon CloudWatch logging options for your delivery stream.</p>
|
|
1757
|
-
*/
|
|
763
|
+
|
|
1758
764
|
CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
|
|
1759
765
|
}
|
|
1760
766
|
export declare namespace SplunkDestinationConfiguration {
|
|
1761
|
-
|
|
1762
|
-
* @internal
|
|
1763
|
-
*/
|
|
767
|
+
|
|
1764
768
|
const filterSensitiveLog: (obj: SplunkDestinationConfiguration) => any;
|
|
1765
769
|
}
|
|
1766
|
-
|
|
1767
|
-
* <p>Metadata that you can assign to a delivery stream, consisting of a key-value
|
|
1768
|
-
* pair.</p>
|
|
1769
|
-
*/
|
|
770
|
+
|
|
1770
771
|
export interface Tag {
|
|
1771
|
-
|
|
1772
|
-
* <p>A unique identifier for the tag. Maximum length: 128 characters. Valid characters:
|
|
1773
|
-
* Unicode letters, digits, white space, _ . / = + - % @</p>
|
|
1774
|
-
*/
|
|
772
|
+
|
|
1775
773
|
Key: string | undefined;
|
|
1776
|
-
|
|
1777
|
-
* <p>An optional string, which you can use to describe or define the tag. Maximum length:
|
|
1778
|
-
* 256 characters. Valid characters: Unicode letters, digits, white space, _ . / = + - %
|
|
1779
|
-
* @</p>
|
|
1780
|
-
*/
|
|
774
|
+
|
|
1781
775
|
Value?: string;
|
|
1782
776
|
}
|
|
1783
777
|
export declare namespace Tag {
|
|
1784
|
-
|
|
1785
|
-
* @internal
|
|
1786
|
-
*/
|
|
778
|
+
|
|
1787
779
|
const filterSensitiveLog: (obj: Tag) => any;
|
|
1788
780
|
}
|
|
1789
781
|
export interface CreateDeliveryStreamInput {
|
|
1790
|
-
|
|
1791
|
-
* <p>The name of the delivery stream. This name must be unique per AWS account in the same
|
|
1792
|
-
* AWS Region. If the delivery streams are in different accounts or different Regions, you can
|
|
1793
|
-
* have multiple delivery streams with the same name.</p>
|
|
1794
|
-
*/
|
|
782
|
+
|
|
1795
783
|
DeliveryStreamName: string | undefined;
|
|
1796
|
-
|
|
1797
|
-
* <p>The delivery stream type. This parameter can be one of the following
|
|
1798
|
-
* values:</p>
|
|
1799
|
-
* <ul>
|
|
1800
|
-
* <li>
|
|
1801
|
-
* <p>
|
|
1802
|
-
* <code>DirectPut</code>: Provider applications access the delivery stream
|
|
1803
|
-
* directly.</p>
|
|
1804
|
-
* </li>
|
|
1805
|
-
* <li>
|
|
1806
|
-
* <p>
|
|
1807
|
-
* <code>KinesisStreamAsSource</code>: The delivery stream uses a Kinesis data
|
|
1808
|
-
* stream as a source.</p>
|
|
1809
|
-
* </li>
|
|
1810
|
-
* </ul>
|
|
1811
|
-
*/
|
|
784
|
+
|
|
1812
785
|
DeliveryStreamType?: DeliveryStreamType | string;
|
|
1813
|
-
|
|
1814
|
-
* <p>When a Kinesis data stream is used as the source for the delivery stream, a <a>KinesisStreamSourceConfiguration</a> containing the Kinesis data stream Amazon
|
|
1815
|
-
* Resource Name (ARN) and the role ARN for the source stream.</p>
|
|
1816
|
-
*/
|
|
786
|
+
|
|
1817
787
|
KinesisStreamSourceConfiguration?: KinesisStreamSourceConfiguration;
|
|
1818
|
-
|
|
1819
|
-
* <p>Used to specify the type and Amazon Resource Name (ARN) of the KMS key needed for
|
|
1820
|
-
* Server-Side Encryption (SSE).</p>
|
|
1821
|
-
*/
|
|
788
|
+
|
|
1822
789
|
DeliveryStreamEncryptionConfigurationInput?: DeliveryStreamEncryptionConfigurationInput;
|
|
1823
|
-
|
|
1824
|
-
* @deprecated
|
|
1825
|
-
*
|
|
1826
|
-
* <p>[Deprecated]
|
|
1827
|
-
* The destination in Amazon S3. You can specify only one destination.</p>
|
|
1828
|
-
*/
|
|
790
|
+
|
|
1829
791
|
S3DestinationConfiguration?: S3DestinationConfiguration;
|
|
1830
|
-
|
|
1831
|
-
* <p>The destination in Amazon S3. You can specify only one destination.</p>
|
|
1832
|
-
*/
|
|
792
|
+
|
|
1833
793
|
ExtendedS3DestinationConfiguration?: ExtendedS3DestinationConfiguration;
|
|
1834
|
-
|
|
1835
|
-
* <p>The destination in Amazon Redshift. You can specify only one destination.</p>
|
|
1836
|
-
*/
|
|
794
|
+
|
|
1837
795
|
RedshiftDestinationConfiguration?: RedshiftDestinationConfiguration;
|
|
1838
|
-
|
|
1839
|
-
* <p>The destination in Amazon ES. You can specify only one destination.</p>
|
|
1840
|
-
*/
|
|
796
|
+
|
|
1841
797
|
ElasticsearchDestinationConfiguration?: ElasticsearchDestinationConfiguration;
|
|
1842
798
|
AmazonopensearchserviceDestinationConfiguration?: AmazonopensearchserviceDestinationConfiguration;
|
|
1843
|
-
|
|
1844
|
-
* <p>The destination in Splunk. You can specify only one destination.</p>
|
|
1845
|
-
*/
|
|
799
|
+
|
|
1846
800
|
SplunkDestinationConfiguration?: SplunkDestinationConfiguration;
|
|
1847
|
-
|
|
1848
|
-
* <p>Enables configuring Kinesis Firehose to deliver data to any HTTP endpoint destination.
|
|
1849
|
-
* You can specify only one destination.</p>
|
|
1850
|
-
*/
|
|
801
|
+
|
|
1851
802
|
HttpEndpointDestinationConfiguration?: HttpEndpointDestinationConfiguration;
|
|
1852
|
-
|
|
1853
|
-
* <p>A set of tags to assign to the delivery stream. A tag is a key-value pair that you can
|
|
1854
|
-
* define and assign to AWS resources. Tags are metadata. For example, you can add friendly
|
|
1855
|
-
* names and descriptions or other types of information that can help you distinguish the
|
|
1856
|
-
* delivery stream. For more information about tags, see <a href="https://docs.aws.amazon.com/awsaccountbilling/latest/aboutv2/cost-alloc-tags.html">Using Cost Allocation
|
|
1857
|
-
* Tags</a> in the AWS Billing and Cost Management User Guide.</p>
|
|
1858
|
-
*
|
|
1859
|
-
* <p>You can specify up to 50 tags when creating a delivery stream.</p>
|
|
1860
|
-
*/
|
|
803
|
+
|
|
1861
804
|
Tags?: Tag[];
|
|
1862
805
|
}
|
|
1863
806
|
export declare namespace CreateDeliveryStreamInput {
|
|
1864
|
-
|
|
1865
|
-
* @internal
|
|
1866
|
-
*/
|
|
807
|
+
|
|
1867
808
|
const filterSensitiveLog: (obj: CreateDeliveryStreamInput) => any;
|
|
1868
809
|
}
|
|
1869
810
|
export interface CreateDeliveryStreamOutput {
|
|
1870
|
-
|
|
1871
|
-
* <p>The ARN of the delivery stream.</p>
|
|
1872
|
-
*/
|
|
811
|
+
|
|
1873
812
|
DeliveryStreamARN?: string;
|
|
1874
813
|
}
|
|
1875
814
|
export declare namespace CreateDeliveryStreamOutput {
|
|
1876
|
-
|
|
1877
|
-
* @internal
|
|
1878
|
-
*/
|
|
815
|
+
|
|
1879
816
|
const filterSensitiveLog: (obj: CreateDeliveryStreamOutput) => any;
|
|
1880
817
|
}
|
|
1881
|
-
|
|
1882
|
-
* <p>The specified input parameter has a value that is not valid.</p>
|
|
1883
|
-
*/
|
|
818
|
+
|
|
1884
819
|
export interface InvalidArgumentException extends __SmithyException, $MetadataBearer {
|
|
1885
820
|
name: "InvalidArgumentException";
|
|
1886
821
|
$fault: "client";
|
|
1887
|
-
|
|
1888
|
-
* <p>A message that provides information about the error.</p>
|
|
1889
|
-
*/
|
|
822
|
+
|
|
1890
823
|
message?: string;
|
|
1891
824
|
}
|
|
1892
825
|
export declare namespace InvalidArgumentException {
|
|
1893
|
-
|
|
1894
|
-
* @internal
|
|
1895
|
-
*/
|
|
826
|
+
|
|
1896
827
|
const filterSensitiveLog: (obj: InvalidArgumentException) => any;
|
|
1897
828
|
}
|
|
1898
|
-
|
|
1899
|
-
* <p>Kinesis Data Firehose throws this exception when an attempt to put records or to start
|
|
1900
|
-
* or stop delivery stream encryption fails. This happens when the KMS service throws one of
|
|
1901
|
-
* the following exception types: <code>AccessDeniedException</code>,
|
|
1902
|
-
* <code>InvalidStateException</code>, <code>DisabledException</code>, or
|
|
1903
|
-
* <code>NotFoundException</code>.</p>
|
|
1904
|
-
*/
|
|
829
|
+
|
|
1905
830
|
export interface InvalidKMSResourceException extends __SmithyException, $MetadataBearer {
|
|
1906
831
|
name: "InvalidKMSResourceException";
|
|
1907
832
|
$fault: "client";
|
|
@@ -1909,91 +834,56 @@ export interface InvalidKMSResourceException extends __SmithyException, $Metadat
|
|
|
1909
834
|
message?: string;
|
|
1910
835
|
}
|
|
1911
836
|
export declare namespace InvalidKMSResourceException {
|
|
1912
|
-
|
|
1913
|
-
* @internal
|
|
1914
|
-
*/
|
|
837
|
+
|
|
1915
838
|
const filterSensitiveLog: (obj: InvalidKMSResourceException) => any;
|
|
1916
839
|
}
|
|
1917
|
-
|
|
1918
|
-
* <p>You have already reached the limit for a requested resource.</p>
|
|
1919
|
-
*/
|
|
840
|
+
|
|
1920
841
|
export interface LimitExceededException extends __SmithyException, $MetadataBearer {
|
|
1921
842
|
name: "LimitExceededException";
|
|
1922
843
|
$fault: "client";
|
|
1923
|
-
|
|
1924
|
-
* <p>A message that provides information about the error.</p>
|
|
1925
|
-
*/
|
|
844
|
+
|
|
1926
845
|
message?: string;
|
|
1927
846
|
}
|
|
1928
847
|
export declare namespace LimitExceededException {
|
|
1929
|
-
|
|
1930
|
-
* @internal
|
|
1931
|
-
*/
|
|
848
|
+
|
|
1932
849
|
const filterSensitiveLog: (obj: LimitExceededException) => any;
|
|
1933
850
|
}
|
|
1934
|
-
|
|
1935
|
-
* <p>The resource is already in use and not available for this operation.</p>
|
|
1936
|
-
*/
|
|
851
|
+
|
|
1937
852
|
export interface ResourceInUseException extends __SmithyException, $MetadataBearer {
|
|
1938
853
|
name: "ResourceInUseException";
|
|
1939
854
|
$fault: "client";
|
|
1940
|
-
|
|
1941
|
-
* <p>A message that provides information about the error.</p>
|
|
1942
|
-
*/
|
|
855
|
+
|
|
1943
856
|
message?: string;
|
|
1944
857
|
}
|
|
1945
858
|
export declare namespace ResourceInUseException {
|
|
1946
|
-
|
|
1947
|
-
* @internal
|
|
1948
|
-
*/
|
|
859
|
+
|
|
1949
860
|
const filterSensitiveLog: (obj: ResourceInUseException) => any;
|
|
1950
861
|
}
|
|
1951
862
|
export interface DeleteDeliveryStreamInput {
|
|
1952
|
-
|
|
1953
|
-
* <p>The name of the delivery stream.</p>
|
|
1954
|
-
*/
|
|
863
|
+
|
|
1955
864
|
DeliveryStreamName: string | undefined;
|
|
1956
|
-
|
|
1957
|
-
* <p>Set this to true if you want to delete the delivery stream even if Kinesis Data Firehose
|
|
1958
|
-
* is unable to retire the grant for the CMK. Kinesis Data Firehose might be unable to retire
|
|
1959
|
-
* the grant due to a customer error, such as when the CMK or the grant are in an invalid
|
|
1960
|
-
* state. If you force deletion, you can then use the <a href="https://docs.aws.amazon.com/kms/latest/APIReference/API_RevokeGrant.html">RevokeGrant</a> operation to
|
|
1961
|
-
* revoke the grant you gave to Kinesis Data Firehose. If a failure to retire the grant
|
|
1962
|
-
* happens due to an AWS KMS issue, Kinesis Data Firehose keeps retrying the delete
|
|
1963
|
-
* operation.</p>
|
|
1964
|
-
* <p>The default value is false.</p>
|
|
1965
|
-
*/
|
|
865
|
+
|
|
1966
866
|
AllowForceDelete?: boolean;
|
|
1967
867
|
}
|
|
1968
868
|
export declare namespace DeleteDeliveryStreamInput {
|
|
1969
|
-
|
|
1970
|
-
* @internal
|
|
1971
|
-
*/
|
|
869
|
+
|
|
1972
870
|
const filterSensitiveLog: (obj: DeleteDeliveryStreamInput) => any;
|
|
1973
871
|
}
|
|
1974
872
|
export interface DeleteDeliveryStreamOutput {
|
|
1975
873
|
}
|
|
1976
874
|
export declare namespace DeleteDeliveryStreamOutput {
|
|
1977
|
-
|
|
1978
|
-
* @internal
|
|
1979
|
-
*/
|
|
875
|
+
|
|
1980
876
|
const filterSensitiveLog: (obj: DeleteDeliveryStreamOutput) => any;
|
|
1981
877
|
}
|
|
1982
|
-
|
|
1983
|
-
* <p>The specified resource could not be found.</p>
|
|
1984
|
-
*/
|
|
878
|
+
|
|
1985
879
|
export interface ResourceNotFoundException extends __SmithyException, $MetadataBearer {
|
|
1986
880
|
name: "ResourceNotFoundException";
|
|
1987
881
|
$fault: "client";
|
|
1988
|
-
|
|
1989
|
-
* <p>A message that provides information about the error.</p>
|
|
1990
|
-
*/
|
|
882
|
+
|
|
1991
883
|
message?: string;
|
|
1992
884
|
}
|
|
1993
885
|
export declare namespace ResourceNotFoundException {
|
|
1994
|
-
|
|
1995
|
-
* @internal
|
|
1996
|
-
*/
|
|
886
|
+
|
|
1997
887
|
const filterSensitiveLog: (obj: ResourceNotFoundException) => any;
|
|
1998
888
|
}
|
|
1999
889
|
export declare enum DeliveryStreamFailureType {
|
|
@@ -2013,25 +903,15 @@ export declare enum DeliveryStreamFailureType {
|
|
|
2013
903
|
SUBNET_NOT_FOUND = "SUBNET_NOT_FOUND",
|
|
2014
904
|
UNKNOWN_ERROR = "UNKNOWN_ERROR"
|
|
2015
905
|
}
|
|
2016
|
-
|
|
2017
|
-
* <p>Provides details in case one of the following operations fails due to an error related
|
|
2018
|
-
* to KMS: <a>CreateDeliveryStream</a>, <a>DeleteDeliveryStream</a>,
|
|
2019
|
-
* <a>StartDeliveryStreamEncryption</a>, <a>StopDeliveryStreamEncryption</a>.</p>
|
|
2020
|
-
*/
|
|
906
|
+
|
|
2021
907
|
export interface FailureDescription {
|
|
2022
|
-
|
|
2023
|
-
* <p>The type of error that caused the failure.</p>
|
|
2024
|
-
*/
|
|
908
|
+
|
|
2025
909
|
Type: DeliveryStreamFailureType | string | undefined;
|
|
2026
|
-
|
|
2027
|
-
* <p>A message providing details about the error that caused the failure.</p>
|
|
2028
|
-
*/
|
|
910
|
+
|
|
2029
911
|
Details: string | undefined;
|
|
2030
912
|
}
|
|
2031
913
|
export declare namespace FailureDescription {
|
|
2032
|
-
|
|
2033
|
-
* @internal
|
|
2034
|
-
*/
|
|
914
|
+
|
|
2035
915
|
const filterSensitiveLog: (obj: FailureDescription) => any;
|
|
2036
916
|
}
|
|
2037
917
|
export declare enum DeliveryStreamEncryptionStatus {
|
|
@@ -2042,43 +922,19 @@ export declare enum DeliveryStreamEncryptionStatus {
|
|
|
2042
922
|
ENABLING = "ENABLING",
|
|
2043
923
|
ENABLING_FAILED = "ENABLING_FAILED"
|
|
2044
924
|
}
|
|
2045
|
-
|
|
2046
|
-
* <p>Contains information about the server-side encryption (SSE) status for the delivery
|
|
2047
|
-
* stream, the type customer master key (CMK) in use, if any, and the ARN of the CMK. You can
|
|
2048
|
-
* get <code>DeliveryStreamEncryptionConfiguration</code> by invoking the <a>DescribeDeliveryStream</a> operation. </p>
|
|
2049
|
-
*/
|
|
925
|
+
|
|
2050
926
|
export interface DeliveryStreamEncryptionConfiguration {
|
|
2051
|
-
|
|
2052
|
-
* <p>If <code>KeyType</code> is <code>CUSTOMER_MANAGED_CMK</code>, this field contains the
|
|
2053
|
-
* ARN of the customer managed CMK. If <code>KeyType</code> is <code>AWS_OWNED_CMK</code>,
|
|
2054
|
-
* <code>DeliveryStreamEncryptionConfiguration</code> doesn't contain a value for
|
|
2055
|
-
* <code>KeyARN</code>.</p>
|
|
2056
|
-
*/
|
|
927
|
+
|
|
2057
928
|
KeyARN?: string;
|
|
2058
|
-
|
|
2059
|
-
* <p>Indicates the type of customer master key (CMK) that is used for encryption. The default
|
|
2060
|
-
* setting is <code>AWS_OWNED_CMK</code>. For more information about CMKs, see <a href="https://docs.aws.amazon.com/kms/latest/developerguide/concepts.html#master_keys">Customer
|
|
2061
|
-
* Master Keys (CMKs)</a>.</p>
|
|
2062
|
-
*/
|
|
929
|
+
|
|
2063
930
|
KeyType?: KeyType | string;
|
|
2064
|
-
|
|
2065
|
-
* <p>This is the server-side encryption (SSE) status for the delivery stream. For a full
|
|
2066
|
-
* description of the different values of this status, see <a>StartDeliveryStreamEncryption</a> and <a>StopDeliveryStreamEncryption</a>. If this status is <code>ENABLING_FAILED</code>
|
|
2067
|
-
* or <code>DISABLING_FAILED</code>, it is the status of the most recent attempt to enable or
|
|
2068
|
-
* disable SSE, respectively.</p>
|
|
2069
|
-
*/
|
|
931
|
+
|
|
2070
932
|
Status?: DeliveryStreamEncryptionStatus | string;
|
|
2071
|
-
|
|
2072
|
-
* <p>Provides details in case one of the following operations fails due to an error related
|
|
2073
|
-
* to KMS: <a>CreateDeliveryStream</a>, <a>DeleteDeliveryStream</a>,
|
|
2074
|
-
* <a>StartDeliveryStreamEncryption</a>, <a>StopDeliveryStreamEncryption</a>.</p>
|
|
2075
|
-
*/
|
|
933
|
+
|
|
2076
934
|
FailureDescription?: FailureDescription;
|
|
2077
935
|
}
|
|
2078
936
|
export declare namespace DeliveryStreamEncryptionConfiguration {
|
|
2079
|
-
|
|
2080
|
-
* @internal
|
|
2081
|
-
*/
|
|
937
|
+
|
|
2082
938
|
const filterSensitiveLog: (obj: DeliveryStreamEncryptionConfiguration) => any;
|
|
2083
939
|
}
|
|
2084
940
|
export declare enum DeliveryStreamStatus {
|
|
@@ -2088,1266 +944,605 @@ export declare enum DeliveryStreamStatus {
|
|
|
2088
944
|
DELETING = "DELETING",
|
|
2089
945
|
DELETING_FAILED = "DELETING_FAILED"
|
|
2090
946
|
}
|
|
2091
|
-
|
|
2092
|
-
* <p>The destination description in Amazon ES.</p>
|
|
2093
|
-
*/
|
|
947
|
+
|
|
2094
948
|
export interface ElasticsearchDestinationDescription {
|
|
2095
|
-
|
|
2096
|
-
* <p>The Amazon Resource Name (ARN) of the AWS credentials. For more information, see
|
|
2097
|
-
* <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon
|
|
2098
|
-
* Resource Names (ARNs) and AWS Service Namespaces</a>.</p>
|
|
2099
|
-
*/
|
|
949
|
+
|
|
2100
950
|
RoleARN?: string;
|
|
2101
|
-
|
|
2102
|
-
* <p>The ARN of the Amazon ES domain. For more information, see <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon
|
|
2103
|
-
* Resource Names (ARNs) and AWS Service Namespaces</a>.</p>
|
|
2104
|
-
*
|
|
2105
|
-
* <p>Kinesis Data Firehose uses either <code>ClusterEndpoint</code> or <code>DomainARN</code>
|
|
2106
|
-
* to send data to Amazon ES.</p>
|
|
2107
|
-
*/
|
|
951
|
+
|
|
2108
952
|
DomainARN?: string;
|
|
2109
|
-
|
|
2110
|
-
* <p>The endpoint to use when communicating with the cluster. Kinesis Data Firehose uses
|
|
2111
|
-
* either this <code>ClusterEndpoint</code> or the <code>DomainARN</code> field to send data
|
|
2112
|
-
* to Amazon ES.</p>
|
|
2113
|
-
*/
|
|
953
|
+
|
|
2114
954
|
ClusterEndpoint?: string;
|
|
2115
|
-
|
|
2116
|
-
* <p>The Elasticsearch index name.</p>
|
|
2117
|
-
*/
|
|
955
|
+
|
|
2118
956
|
IndexName?: string;
|
|
2119
|
-
|
|
2120
|
-
* <p>The Elasticsearch type name. This applies to Elasticsearch 6.x and lower versions.
|
|
2121
|
-
* For Elasticsearch 7.x, there's no value for <code>TypeName</code>.</p>
|
|
2122
|
-
*/
|
|
957
|
+
|
|
2123
958
|
TypeName?: string;
|
|
2124
|
-
|
|
2125
|
-
* <p>The Elasticsearch index rotation period</p>
|
|
2126
|
-
*/
|
|
959
|
+
|
|
2127
960
|
IndexRotationPeriod?: ElasticsearchIndexRotationPeriod | string;
|
|
2128
|
-
|
|
2129
|
-
* <p>The buffering options.</p>
|
|
2130
|
-
*/
|
|
961
|
+
|
|
2131
962
|
BufferingHints?: ElasticsearchBufferingHints;
|
|
2132
|
-
|
|
2133
|
-
* <p>The Amazon ES retry options.</p>
|
|
2134
|
-
*/
|
|
963
|
+
|
|
2135
964
|
RetryOptions?: ElasticsearchRetryOptions;
|
|
2136
|
-
|
|
2137
|
-
* <p>The Amazon S3 backup mode.</p>
|
|
2138
|
-
*/
|
|
965
|
+
|
|
2139
966
|
S3BackupMode?: ElasticsearchS3BackupMode | string;
|
|
2140
|
-
|
|
2141
|
-
* <p>The Amazon S3 destination.</p>
|
|
2142
|
-
*/
|
|
967
|
+
|
|
2143
968
|
S3DestinationDescription?: S3DestinationDescription;
|
|
2144
|
-
|
|
2145
|
-
* <p>The data processing configuration.</p>
|
|
2146
|
-
*/
|
|
969
|
+
|
|
2147
970
|
ProcessingConfiguration?: ProcessingConfiguration;
|
|
2148
|
-
|
|
2149
|
-
* <p>The Amazon CloudWatch logging options.</p>
|
|
2150
|
-
*/
|
|
971
|
+
|
|
2151
972
|
CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
|
|
2152
|
-
|
|
2153
|
-
* <p>The details of the VPC of the Amazon ES destination.</p>
|
|
2154
|
-
*/
|
|
973
|
+
|
|
2155
974
|
VpcConfigurationDescription?: VpcConfigurationDescription;
|
|
2156
975
|
}
|
|
2157
976
|
export declare namespace ElasticsearchDestinationDescription {
|
|
2158
|
-
|
|
2159
|
-
* @internal
|
|
2160
|
-
*/
|
|
977
|
+
|
|
2161
978
|
const filterSensitiveLog: (obj: ElasticsearchDestinationDescription) => any;
|
|
2162
979
|
}
|
|
2163
|
-
|
|
2164
|
-
* <p>Describes a destination in Amazon S3.</p>
|
|
2165
|
-
*/
|
|
980
|
+
|
|
2166
981
|
export interface ExtendedS3DestinationDescription {
|
|
2167
|
-
|
|
2168
|
-
* <p>The Amazon Resource Name (ARN) of the AWS credentials. For more information, see
|
|
2169
|
-
* <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon
|
|
2170
|
-
* Resource Names (ARNs) and AWS Service Namespaces</a>.</p>
|
|
2171
|
-
*/
|
|
982
|
+
|
|
2172
983
|
RoleARN: string | undefined;
|
|
2173
|
-
|
|
2174
|
-
* <p>The ARN of the S3 bucket. For more information, see <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon Resource Names (ARNs) and
|
|
2175
|
-
* AWS Service Namespaces</a>.</p>
|
|
2176
|
-
*/
|
|
984
|
+
|
|
2177
985
|
BucketARN: string | undefined;
|
|
2178
|
-
|
|
2179
|
-
* <p>The "YYYY/MM/DD/HH" time format prefix is automatically used for delivered Amazon S3
|
|
2180
|
-
* files. You can also specify a custom prefix, as described in <a href="https://docs.aws.amazon.com/firehose/latest/dev/s3-prefixes.html">Custom Prefixes for Amazon S3
|
|
2181
|
-
* Objects</a>.</p>
|
|
2182
|
-
*/
|
|
986
|
+
|
|
2183
987
|
Prefix?: string;
|
|
2184
|
-
|
|
2185
|
-
* <p>A prefix that Kinesis Data Firehose evaluates and adds to failed records before writing
|
|
2186
|
-
* them to S3. This prefix appears immediately following the bucket name. For information
|
|
2187
|
-
* about how to specify this prefix, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/s3-prefixes.html">Custom Prefixes for Amazon S3
|
|
2188
|
-
* Objects</a>.</p>
|
|
2189
|
-
*/
|
|
988
|
+
|
|
2190
989
|
ErrorOutputPrefix?: string;
|
|
2191
|
-
|
|
2192
|
-
* <p>The buffering option.</p>
|
|
2193
|
-
*/
|
|
990
|
+
|
|
2194
991
|
BufferingHints: BufferingHints | undefined;
|
|
2195
|
-
|
|
2196
|
-
* <p>The compression format. If no value is specified, the default is
|
|
2197
|
-
* <code>UNCOMPRESSED</code>.</p>
|
|
2198
|
-
*/
|
|
992
|
+
|
|
2199
993
|
CompressionFormat: CompressionFormat | string | undefined;
|
|
2200
|
-
|
|
2201
|
-
* <p>The encryption configuration. If no value is specified, the default is no
|
|
2202
|
-
* encryption.</p>
|
|
2203
|
-
*/
|
|
994
|
+
|
|
2204
995
|
EncryptionConfiguration: EncryptionConfiguration | undefined;
|
|
2205
|
-
|
|
2206
|
-
* <p>The Amazon CloudWatch logging options for your delivery stream.</p>
|
|
2207
|
-
*/
|
|
996
|
+
|
|
2208
997
|
CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
|
|
2209
|
-
|
|
2210
|
-
* <p>The data processing configuration.</p>
|
|
2211
|
-
*/
|
|
998
|
+
|
|
2212
999
|
ProcessingConfiguration?: ProcessingConfiguration;
|
|
2213
|
-
|
|
2214
|
-
* <p>The Amazon S3 backup mode.</p>
|
|
2215
|
-
*/
|
|
1000
|
+
|
|
2216
1001
|
S3BackupMode?: S3BackupMode | string;
|
|
2217
|
-
|
|
2218
|
-
* <p>The configuration for backup in Amazon S3.</p>
|
|
2219
|
-
*/
|
|
1002
|
+
|
|
2220
1003
|
S3BackupDescription?: S3DestinationDescription;
|
|
2221
|
-
|
|
2222
|
-
* <p>The serializer, deserializer, and schema for converting data from the JSON format to
|
|
2223
|
-
* the Parquet or ORC format before writing it to Amazon S3.</p>
|
|
2224
|
-
*/
|
|
1004
|
+
|
|
2225
1005
|
DataFormatConversionConfiguration?: DataFormatConversionConfiguration;
|
|
2226
|
-
|
|
2227
|
-
* <p>The configuration of the dynamic partitioning mechanism that creates smaller data sets
|
|
2228
|
-
* from the streaming data by partitioning it based on partition keys. Currently, dynamic
|
|
2229
|
-
* partitioning is only supported for Amazon S3 destinations. For more information, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/dynamic-partitioning.html">https://docs.aws.amazon.com/firehose/latest/dev/dynamic-partitioning.html</a>
|
|
2230
|
-
* </p>
|
|
2231
|
-
*/
|
|
1006
|
+
|
|
2232
1007
|
DynamicPartitioningConfiguration?: DynamicPartitioningConfiguration;
|
|
2233
1008
|
}
|
|
2234
1009
|
export declare namespace ExtendedS3DestinationDescription {
|
|
2235
|
-
|
|
2236
|
-
* @internal
|
|
2237
|
-
*/
|
|
1010
|
+
|
|
2238
1011
|
const filterSensitiveLog: (obj: ExtendedS3DestinationDescription) => any;
|
|
2239
1012
|
}
|
|
2240
|
-
|
|
2241
|
-
* <p>Describes the HTTP endpoint selected as the destination. </p>
|
|
2242
|
-
*/
|
|
1013
|
+
|
|
2243
1014
|
export interface HttpEndpointDescription {
|
|
2244
|
-
|
|
2245
|
-
* <p>The URL of the HTTP endpoint selected as the destination.</p>
|
|
2246
|
-
*/
|
|
1015
|
+
|
|
2247
1016
|
Url?: string;
|
|
2248
|
-
|
|
2249
|
-
* <p>The name of the HTTP endpoint selected as the destination.</p>
|
|
2250
|
-
*/
|
|
1017
|
+
|
|
2251
1018
|
Name?: string;
|
|
2252
1019
|
}
|
|
2253
1020
|
export declare namespace HttpEndpointDescription {
|
|
2254
|
-
|
|
2255
|
-
* @internal
|
|
2256
|
-
*/
|
|
1021
|
+
|
|
2257
1022
|
const filterSensitiveLog: (obj: HttpEndpointDescription) => any;
|
|
2258
1023
|
}
|
|
2259
|
-
|
|
2260
|
-
* <p>Describes the HTTP endpoint destination.</p>
|
|
2261
|
-
*/
|
|
1024
|
+
|
|
2262
1025
|
export interface HttpEndpointDestinationDescription {
|
|
2263
|
-
|
|
2264
|
-
* <p>The configuration of the specified HTTP endpoint destination.</p>
|
|
2265
|
-
*/
|
|
1026
|
+
|
|
2266
1027
|
EndpointConfiguration?: HttpEndpointDescription;
|
|
2267
|
-
|
|
2268
|
-
* <p>Describes buffering options that can be applied to the data before it is delivered to
|
|
2269
|
-
* the HTTPS endpoint destination. Kinesis Data Firehose teats these options as hints, and it
|
|
2270
|
-
* might choose to use more optimal values. The <code>SizeInMBs</code> and
|
|
2271
|
-
* <code>IntervalInSeconds</code> parameters are optional. However, if specify a value for
|
|
2272
|
-
* one of them, you must also provide a value for the other. </p>
|
|
2273
|
-
*/
|
|
1028
|
+
|
|
2274
1029
|
BufferingHints?: HttpEndpointBufferingHints;
|
|
2275
|
-
|
|
2276
|
-
* <p>Describes the Amazon CloudWatch logging options for your delivery stream.</p>
|
|
2277
|
-
*/
|
|
1030
|
+
|
|
2278
1031
|
CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
|
|
2279
|
-
|
|
2280
|
-
* <p>The configuration of request sent to the HTTP endpoint specified as the
|
|
2281
|
-
* destination.</p>
|
|
2282
|
-
*/
|
|
1032
|
+
|
|
2283
1033
|
RequestConfiguration?: HttpEndpointRequestConfiguration;
|
|
2284
|
-
|
|
2285
|
-
* <p>Describes a data processing configuration.</p>
|
|
2286
|
-
*/
|
|
1034
|
+
|
|
2287
1035
|
ProcessingConfiguration?: ProcessingConfiguration;
|
|
2288
|
-
|
|
2289
|
-
* <p>Kinesis Data Firehose uses this IAM role for all the permissions that the delivery
|
|
2290
|
-
* stream needs.</p>
|
|
2291
|
-
*/
|
|
1036
|
+
|
|
2292
1037
|
RoleARN?: string;
|
|
2293
|
-
|
|
2294
|
-
* <p>Describes the retry behavior in case Kinesis Data Firehose is unable to deliver data to
|
|
2295
|
-
* the specified HTTP endpoint destination, or if it doesn't receive a valid acknowledgment of
|
|
2296
|
-
* receipt from the specified HTTP endpoint destination.</p>
|
|
2297
|
-
*/
|
|
1038
|
+
|
|
2298
1039
|
RetryOptions?: HttpEndpointRetryOptions;
|
|
2299
|
-
|
|
2300
|
-
* <p>Describes the S3 bucket backup options for the data that Kinesis Firehose delivers to
|
|
2301
|
-
* the HTTP endpoint destination. You can back up all documents (<code>AllData</code>) or only
|
|
2302
|
-
* the documents that Kinesis Data Firehose could not deliver to the specified HTTP endpoint
|
|
2303
|
-
* destination (<code>FailedDataOnly</code>).</p>
|
|
2304
|
-
*/
|
|
1040
|
+
|
|
2305
1041
|
S3BackupMode?: HttpEndpointS3BackupMode | string;
|
|
2306
|
-
|
|
2307
|
-
* <p>Describes a destination in Amazon S3.</p>
|
|
2308
|
-
*/
|
|
1042
|
+
|
|
2309
1043
|
S3DestinationDescription?: S3DestinationDescription;
|
|
2310
1044
|
}
|
|
2311
1045
|
export declare namespace HttpEndpointDestinationDescription {
|
|
2312
|
-
|
|
2313
|
-
* @internal
|
|
2314
|
-
*/
|
|
1046
|
+
|
|
2315
1047
|
const filterSensitiveLog: (obj: HttpEndpointDestinationDescription) => any;
|
|
2316
1048
|
}
|
|
2317
|
-
|
|
2318
|
-
* <p>Describes a destination in Amazon Redshift.</p>
|
|
2319
|
-
*/
|
|
1049
|
+
|
|
2320
1050
|
export interface RedshiftDestinationDescription {
|
|
2321
|
-
|
|
2322
|
-
* <p>The Amazon Resource Name (ARN) of the AWS credentials. For more information, see
|
|
2323
|
-
* <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon
|
|
2324
|
-
* Resource Names (ARNs) and AWS Service Namespaces</a>.</p>
|
|
2325
|
-
*/
|
|
1051
|
+
|
|
2326
1052
|
RoleARN: string | undefined;
|
|
2327
|
-
|
|
2328
|
-
* <p>The database connection string.</p>
|
|
2329
|
-
*/
|
|
1053
|
+
|
|
2330
1054
|
ClusterJDBCURL: string | undefined;
|
|
2331
|
-
|
|
2332
|
-
* <p>The <code>COPY</code> command.</p>
|
|
2333
|
-
*/
|
|
1055
|
+
|
|
2334
1056
|
CopyCommand: CopyCommand | undefined;
|
|
2335
|
-
|
|
2336
|
-
* <p>The name of the user.</p>
|
|
2337
|
-
*/
|
|
1057
|
+
|
|
2338
1058
|
Username: string | undefined;
|
|
2339
|
-
|
|
2340
|
-
* <p>The retry behavior in case Kinesis Data Firehose is unable to deliver documents to
|
|
2341
|
-
* Amazon Redshift. Default value is 3600 (60 minutes).</p>
|
|
2342
|
-
*/
|
|
1059
|
+
|
|
2343
1060
|
RetryOptions?: RedshiftRetryOptions;
|
|
2344
|
-
|
|
2345
|
-
* <p>The Amazon S3 destination.</p>
|
|
2346
|
-
*/
|
|
1061
|
+
|
|
2347
1062
|
S3DestinationDescription: S3DestinationDescription | undefined;
|
|
2348
|
-
|
|
2349
|
-
* <p>The data processing configuration.</p>
|
|
2350
|
-
*/
|
|
1063
|
+
|
|
2351
1064
|
ProcessingConfiguration?: ProcessingConfiguration;
|
|
2352
|
-
|
|
2353
|
-
* <p>The Amazon S3 backup mode.</p>
|
|
2354
|
-
*/
|
|
1065
|
+
|
|
2355
1066
|
S3BackupMode?: RedshiftS3BackupMode | string;
|
|
2356
|
-
|
|
2357
|
-
* <p>The configuration for backup in Amazon S3.</p>
|
|
2358
|
-
*/
|
|
1067
|
+
|
|
2359
1068
|
S3BackupDescription?: S3DestinationDescription;
|
|
2360
|
-
|
|
2361
|
-
* <p>The Amazon CloudWatch logging options for your delivery stream.</p>
|
|
2362
|
-
*/
|
|
1069
|
+
|
|
2363
1070
|
CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
|
|
2364
1071
|
}
|
|
2365
1072
|
export declare namespace RedshiftDestinationDescription {
|
|
2366
|
-
|
|
2367
|
-
* @internal
|
|
2368
|
-
*/
|
|
1073
|
+
|
|
2369
1074
|
const filterSensitiveLog: (obj: RedshiftDestinationDescription) => any;
|
|
2370
1075
|
}
|
|
2371
|
-
|
|
2372
|
-
* <p>Describes a destination in Splunk.</p>
|
|
2373
|
-
*/
|
|
1076
|
+
|
|
2374
1077
|
export interface SplunkDestinationDescription {
|
|
2375
|
-
|
|
2376
|
-
* <p>The HTTP Event Collector (HEC) endpoint to which Kinesis Data Firehose sends your
|
|
2377
|
-
* data.</p>
|
|
2378
|
-
*/
|
|
1078
|
+
|
|
2379
1079
|
HECEndpoint?: string;
|
|
2380
|
-
|
|
2381
|
-
* <p>This type can be either "Raw" or "Event."</p>
|
|
2382
|
-
*/
|
|
1080
|
+
|
|
2383
1081
|
HECEndpointType?: HECEndpointType | string;
|
|
2384
|
-
|
|
2385
|
-
* <p>A GUID you obtain from your Splunk cluster when you create a new HEC
|
|
2386
|
-
* endpoint.</p>
|
|
2387
|
-
*/
|
|
1082
|
+
|
|
2388
1083
|
HECToken?: string;
|
|
2389
|
-
|
|
2390
|
-
* <p>The amount of time that Kinesis Data Firehose waits to receive an acknowledgment from
|
|
2391
|
-
* Splunk after it sends it data. At the end of the timeout period, Kinesis Data Firehose
|
|
2392
|
-
* either tries to send the data again or considers it an error, based on your retry
|
|
2393
|
-
* settings.</p>
|
|
2394
|
-
*/
|
|
1084
|
+
|
|
2395
1085
|
HECAcknowledgmentTimeoutInSeconds?: number;
|
|
2396
|
-
|
|
2397
|
-
* <p>The retry behavior in case Kinesis Data Firehose is unable to deliver data to Splunk
|
|
2398
|
-
* or if it doesn't receive an acknowledgment of receipt from Splunk.</p>
|
|
2399
|
-
*/
|
|
1086
|
+
|
|
2400
1087
|
RetryOptions?: SplunkRetryOptions;
|
|
2401
|
-
|
|
2402
|
-
* <p>Defines how documents should be delivered to Amazon S3. When set to
|
|
2403
|
-
* <code>FailedDocumentsOnly</code>, Kinesis Data Firehose writes any data that could not
|
|
2404
|
-
* be indexed to the configured Amazon S3 destination. When set to <code>AllDocuments</code>,
|
|
2405
|
-
* Kinesis Data Firehose delivers all incoming records to Amazon S3, and also writes failed
|
|
2406
|
-
* documents to Amazon S3. Default value is <code>FailedDocumentsOnly</code>. </p>
|
|
2407
|
-
*/
|
|
1088
|
+
|
|
2408
1089
|
S3BackupMode?: SplunkS3BackupMode | string;
|
|
2409
|
-
|
|
2410
|
-
* <p>The Amazon S3 destination.></p>
|
|
2411
|
-
*/
|
|
1090
|
+
|
|
2412
1091
|
S3DestinationDescription?: S3DestinationDescription;
|
|
2413
|
-
|
|
2414
|
-
* <p>The data processing configuration.</p>
|
|
2415
|
-
*/
|
|
1092
|
+
|
|
2416
1093
|
ProcessingConfiguration?: ProcessingConfiguration;
|
|
2417
|
-
|
|
2418
|
-
* <p>The Amazon CloudWatch logging options for your delivery stream.</p>
|
|
2419
|
-
*/
|
|
1094
|
+
|
|
2420
1095
|
CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
|
|
2421
1096
|
}
|
|
2422
1097
|
export declare namespace SplunkDestinationDescription {
|
|
2423
|
-
|
|
2424
|
-
* @internal
|
|
2425
|
-
*/
|
|
1098
|
+
|
|
2426
1099
|
const filterSensitiveLog: (obj: SplunkDestinationDescription) => any;
|
|
2427
1100
|
}
|
|
2428
|
-
|
|
2429
|
-
* <p>Describes the destination for a delivery stream.</p>
|
|
2430
|
-
*/
|
|
1101
|
+
|
|
2431
1102
|
export interface DestinationDescription {
|
|
2432
|
-
|
|
2433
|
-
* <p>The ID of the destination.</p>
|
|
2434
|
-
*/
|
|
1103
|
+
|
|
2435
1104
|
DestinationId: string | undefined;
|
|
2436
|
-
|
|
2437
|
-
* <p>[Deprecated] The destination in Amazon S3.</p>
|
|
2438
|
-
*/
|
|
1105
|
+
|
|
2439
1106
|
S3DestinationDescription?: S3DestinationDescription;
|
|
2440
|
-
|
|
2441
|
-
* <p>The destination in Amazon S3.</p>
|
|
2442
|
-
*/
|
|
1107
|
+
|
|
2443
1108
|
ExtendedS3DestinationDescription?: ExtendedS3DestinationDescription;
|
|
2444
|
-
|
|
2445
|
-
* <p>The destination in Amazon Redshift.</p>
|
|
2446
|
-
*/
|
|
1109
|
+
|
|
2447
1110
|
RedshiftDestinationDescription?: RedshiftDestinationDescription;
|
|
2448
|
-
|
|
2449
|
-
* <p>The destination in Amazon ES.</p>
|
|
2450
|
-
*/
|
|
1111
|
+
|
|
2451
1112
|
ElasticsearchDestinationDescription?: ElasticsearchDestinationDescription;
|
|
2452
1113
|
AmazonopensearchserviceDestinationDescription?: AmazonopensearchserviceDestinationDescription;
|
|
2453
|
-
|
|
2454
|
-
* <p>The destination in Splunk.</p>
|
|
2455
|
-
*/
|
|
1114
|
+
|
|
2456
1115
|
SplunkDestinationDescription?: SplunkDestinationDescription;
|
|
2457
|
-
|
|
2458
|
-
* <p>Describes the specified HTTP endpoint destination.</p>
|
|
2459
|
-
*/
|
|
1116
|
+
|
|
2460
1117
|
HttpEndpointDestinationDescription?: HttpEndpointDestinationDescription;
|
|
2461
1118
|
}
|
|
2462
1119
|
export declare namespace DestinationDescription {
|
|
2463
|
-
|
|
2464
|
-
* @internal
|
|
2465
|
-
*/
|
|
1120
|
+
|
|
2466
1121
|
const filterSensitiveLog: (obj: DestinationDescription) => any;
|
|
2467
1122
|
}
|
|
2468
|
-
|
|
2469
|
-
* <p>Details about a Kinesis data stream used as the source for a Kinesis Data Firehose
|
|
2470
|
-
* delivery stream.</p>
|
|
2471
|
-
*/
|
|
1123
|
+
|
|
2472
1124
|
export interface KinesisStreamSourceDescription {
|
|
2473
|
-
|
|
2474
|
-
* <p>The Amazon Resource Name (ARN) of the source Kinesis data stream. For more
|
|
2475
|
-
* information, see <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html#arn-syntax-kinesis-streams">Amazon
|
|
2476
|
-
* Kinesis Data Streams ARN Format</a>.</p>
|
|
2477
|
-
*/
|
|
1125
|
+
|
|
2478
1126
|
KinesisStreamARN?: string;
|
|
2479
|
-
|
|
2480
|
-
* <p>The ARN of the role used by the source Kinesis data stream. For more information, see
|
|
2481
|
-
* <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html#arn-syntax-iam">AWS Identity and
|
|
2482
|
-
* Access Management (IAM) ARN Format</a>.</p>
|
|
2483
|
-
*/
|
|
1127
|
+
|
|
2484
1128
|
RoleARN?: string;
|
|
2485
|
-
|
|
2486
|
-
* <p>Kinesis Data Firehose starts retrieving records from the Kinesis data stream starting
|
|
2487
|
-
* with this timestamp.</p>
|
|
2488
|
-
*/
|
|
1129
|
+
|
|
2489
1130
|
DeliveryStartTimestamp?: Date;
|
|
2490
1131
|
}
|
|
2491
1132
|
export declare namespace KinesisStreamSourceDescription {
|
|
2492
|
-
|
|
2493
|
-
* @internal
|
|
2494
|
-
*/
|
|
1133
|
+
|
|
2495
1134
|
const filterSensitiveLog: (obj: KinesisStreamSourceDescription) => any;
|
|
2496
1135
|
}
|
|
2497
|
-
|
|
2498
|
-
* <p>Details about a Kinesis data stream used as the source for a Kinesis Data Firehose
|
|
2499
|
-
* delivery stream.</p>
|
|
2500
|
-
*/
|
|
1136
|
+
|
|
2501
1137
|
export interface SourceDescription {
|
|
2502
|
-
|
|
2503
|
-
* <p>The <a>KinesisStreamSourceDescription</a> value for the source Kinesis
|
|
2504
|
-
* data stream.</p>
|
|
2505
|
-
*/
|
|
1138
|
+
|
|
2506
1139
|
KinesisStreamSourceDescription?: KinesisStreamSourceDescription;
|
|
2507
1140
|
}
|
|
2508
1141
|
export declare namespace SourceDescription {
|
|
2509
|
-
|
|
2510
|
-
* @internal
|
|
2511
|
-
*/
|
|
1142
|
+
|
|
2512
1143
|
const filterSensitiveLog: (obj: SourceDescription) => any;
|
|
2513
1144
|
}
|
|
2514
|
-
|
|
2515
|
-
* <p>Contains information about a delivery stream.</p>
|
|
2516
|
-
*/
|
|
1145
|
+
|
|
2517
1146
|
export interface DeliveryStreamDescription {
|
|
2518
|
-
|
|
2519
|
-
* <p>The name of the delivery stream.</p>
|
|
2520
|
-
*/
|
|
1147
|
+
|
|
2521
1148
|
DeliveryStreamName: string | undefined;
|
|
2522
|
-
|
|
2523
|
-
* <p>The Amazon Resource Name (ARN) of the delivery stream. For more information, see
|
|
2524
|
-
* <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon
|
|
2525
|
-
* Resource Names (ARNs) and AWS Service Namespaces</a>.</p>
|
|
2526
|
-
*/
|
|
1149
|
+
|
|
2527
1150
|
DeliveryStreamARN: string | undefined;
|
|
2528
|
-
|
|
2529
|
-
* <p>The status of the delivery stream. If the status of a delivery stream is
|
|
2530
|
-
* <code>CREATING_FAILED</code>, this status doesn't change, and you can't invoke
|
|
2531
|
-
* <code>CreateDeliveryStream</code> again on it. However, you can invoke the <a>DeleteDeliveryStream</a> operation to delete it.</p>
|
|
2532
|
-
*/
|
|
1151
|
+
|
|
2533
1152
|
DeliveryStreamStatus: DeliveryStreamStatus | string | undefined;
|
|
2534
|
-
|
|
2535
|
-
* <p>Provides details in case one of the following operations fails due to an error related
|
|
2536
|
-
* to KMS: <a>CreateDeliveryStream</a>, <a>DeleteDeliveryStream</a>,
|
|
2537
|
-
* <a>StartDeliveryStreamEncryption</a>, <a>StopDeliveryStreamEncryption</a>.</p>
|
|
2538
|
-
*/
|
|
1153
|
+
|
|
2539
1154
|
FailureDescription?: FailureDescription;
|
|
2540
|
-
|
|
2541
|
-
* <p>Indicates the server-side encryption (SSE) status for the delivery stream.</p>
|
|
2542
|
-
*/
|
|
1155
|
+
|
|
2543
1156
|
DeliveryStreamEncryptionConfiguration?: DeliveryStreamEncryptionConfiguration;
|
|
2544
|
-
|
|
2545
|
-
* <p>The delivery stream type. This can be one of the following values:</p>
|
|
2546
|
-
* <ul>
|
|
2547
|
-
* <li>
|
|
2548
|
-
* <p>
|
|
2549
|
-
* <code>DirectPut</code>: Provider applications access the delivery stream
|
|
2550
|
-
* directly.</p>
|
|
2551
|
-
* </li>
|
|
2552
|
-
* <li>
|
|
2553
|
-
* <p>
|
|
2554
|
-
* <code>KinesisStreamAsSource</code>: The delivery stream uses a Kinesis data
|
|
2555
|
-
* stream as a source.</p>
|
|
2556
|
-
* </li>
|
|
2557
|
-
* </ul>
|
|
2558
|
-
*/
|
|
1157
|
+
|
|
2559
1158
|
DeliveryStreamType: DeliveryStreamType | string | undefined;
|
|
2560
|
-
|
|
2561
|
-
* <p>Each time the destination is updated for a delivery stream, the version ID is
|
|
2562
|
-
* changed, and the current version ID is required when updating the destination. This is so
|
|
2563
|
-
* that the service knows it is applying the changes to the correct version of the delivery
|
|
2564
|
-
* stream.</p>
|
|
2565
|
-
*/
|
|
1159
|
+
|
|
2566
1160
|
VersionId: string | undefined;
|
|
2567
|
-
|
|
2568
|
-
* <p>The date and time that the delivery stream was created.</p>
|
|
2569
|
-
*/
|
|
1161
|
+
|
|
2570
1162
|
CreateTimestamp?: Date;
|
|
2571
|
-
|
|
2572
|
-
* <p>The date and time that the delivery stream was last updated.</p>
|
|
2573
|
-
*/
|
|
1163
|
+
|
|
2574
1164
|
LastUpdateTimestamp?: Date;
|
|
2575
|
-
|
|
2576
|
-
* <p>If the <code>DeliveryStreamType</code> parameter is
|
|
2577
|
-
* <code>KinesisStreamAsSource</code>, a <a>SourceDescription</a> object
|
|
2578
|
-
* describing the source Kinesis data stream.</p>
|
|
2579
|
-
*/
|
|
1165
|
+
|
|
2580
1166
|
Source?: SourceDescription;
|
|
2581
|
-
|
|
2582
|
-
* <p>The destinations.</p>
|
|
2583
|
-
*/
|
|
1167
|
+
|
|
2584
1168
|
Destinations: DestinationDescription[] | undefined;
|
|
2585
|
-
|
|
2586
|
-
* <p>Indicates whether there are more destinations available to list.</p>
|
|
2587
|
-
*/
|
|
1169
|
+
|
|
2588
1170
|
HasMoreDestinations: boolean | undefined;
|
|
2589
1171
|
}
|
|
2590
1172
|
export declare namespace DeliveryStreamDescription {
|
|
2591
|
-
|
|
2592
|
-
* @internal
|
|
2593
|
-
*/
|
|
1173
|
+
|
|
2594
1174
|
const filterSensitiveLog: (obj: DeliveryStreamDescription) => any;
|
|
2595
1175
|
}
|
|
2596
1176
|
export interface DescribeDeliveryStreamInput {
|
|
2597
|
-
|
|
2598
|
-
* <p>The name of the delivery stream.</p>
|
|
2599
|
-
*/
|
|
1177
|
+
|
|
2600
1178
|
DeliveryStreamName: string | undefined;
|
|
2601
|
-
|
|
2602
|
-
* <p>The limit on the number of destinations to return. You can have one destination per
|
|
2603
|
-
* delivery stream.</p>
|
|
2604
|
-
*/
|
|
1179
|
+
|
|
2605
1180
|
Limit?: number;
|
|
2606
|
-
|
|
2607
|
-
* <p>The ID of the destination to start returning the destination information. Kinesis
|
|
2608
|
-
* Data Firehose supports one destination per delivery stream.</p>
|
|
2609
|
-
*/
|
|
1181
|
+
|
|
2610
1182
|
ExclusiveStartDestinationId?: string;
|
|
2611
1183
|
}
|
|
2612
1184
|
export declare namespace DescribeDeliveryStreamInput {
|
|
2613
|
-
|
|
2614
|
-
* @internal
|
|
2615
|
-
*/
|
|
1185
|
+
|
|
2616
1186
|
const filterSensitiveLog: (obj: DescribeDeliveryStreamInput) => any;
|
|
2617
1187
|
}
|
|
2618
1188
|
export interface DescribeDeliveryStreamOutput {
|
|
2619
|
-
|
|
2620
|
-
* <p>Information about the delivery stream.</p>
|
|
2621
|
-
*/
|
|
1189
|
+
|
|
2622
1190
|
DeliveryStreamDescription: DeliveryStreamDescription | undefined;
|
|
2623
1191
|
}
|
|
2624
1192
|
export declare namespace DescribeDeliveryStreamOutput {
|
|
2625
|
-
|
|
2626
|
-
* @internal
|
|
2627
|
-
*/
|
|
1193
|
+
|
|
2628
1194
|
const filterSensitiveLog: (obj: DescribeDeliveryStreamOutput) => any;
|
|
2629
1195
|
}
|
|
2630
|
-
|
|
2631
|
-
* <p>Describes an update for a destination in Amazon ES.</p>
|
|
2632
|
-
*/
|
|
1196
|
+
|
|
2633
1197
|
export interface ElasticsearchDestinationUpdate {
|
|
2634
|
-
|
|
2635
|
-
* <p>The Amazon Resource Name (ARN) of the IAM role to be assumed by Kinesis Data Firehose
|
|
2636
|
-
* for calling the Amazon ES Configuration API and for indexing documents. For more
|
|
2637
|
-
* information, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/controlling-access.html#using-iam-s3">Grant Kinesis Data
|
|
2638
|
-
* Firehose Access to an Amazon S3 Destination</a> and <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon Resource Names (ARNs) and
|
|
2639
|
-
* AWS Service Namespaces</a>.</p>
|
|
2640
|
-
*/
|
|
1198
|
+
|
|
2641
1199
|
RoleARN?: string;
|
|
2642
|
-
|
|
2643
|
-
* <p>The ARN of the Amazon ES domain. The IAM role must have permissions
|
|
2644
|
-
* for <code>DescribeElasticsearchDomain</code>, <code>DescribeElasticsearchDomains</code>,
|
|
2645
|
-
* and <code>DescribeElasticsearchDomainConfig</code> after assuming the IAM role specified in
|
|
2646
|
-
* <code>RoleARN</code>. For more information, see <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon Resource Names (ARNs) and
|
|
2647
|
-
* AWS Service Namespaces</a>.</p>
|
|
2648
|
-
*
|
|
2649
|
-
* <p>Specify either <code>ClusterEndpoint</code> or <code>DomainARN</code>.</p>
|
|
2650
|
-
*/
|
|
1200
|
+
|
|
2651
1201
|
DomainARN?: string;
|
|
2652
|
-
|
|
2653
|
-
* <p>The endpoint to use when communicating with the cluster. Specify either this
|
|
2654
|
-
* <code>ClusterEndpoint</code> or the <code>DomainARN</code> field.</p>
|
|
2655
|
-
*/
|
|
1202
|
+
|
|
2656
1203
|
ClusterEndpoint?: string;
|
|
2657
|
-
|
|
2658
|
-
* <p>The Elasticsearch index name.</p>
|
|
2659
|
-
*/
|
|
1204
|
+
|
|
2660
1205
|
IndexName?: string;
|
|
2661
|
-
|
|
2662
|
-
* <p>The Elasticsearch type name. For Elasticsearch 6.x, there can be only one type per
|
|
2663
|
-
* index. If you try to specify a new type for an existing index that already has another
|
|
2664
|
-
* type, Kinesis Data Firehose returns an error during runtime.</p>
|
|
2665
|
-
*
|
|
2666
|
-
* <p>If you upgrade Elasticsearch from 6.x to 7.x and don’t update your delivery stream,
|
|
2667
|
-
* Kinesis Data Firehose still delivers data to Elasticsearch with the old index name and type
|
|
2668
|
-
* name. If you want to update your delivery stream with a new index name, provide an empty
|
|
2669
|
-
* string for <code>TypeName</code>. </p>
|
|
2670
|
-
*/
|
|
1206
|
+
|
|
2671
1207
|
TypeName?: string;
|
|
2672
|
-
|
|
2673
|
-
* <p>The Elasticsearch index rotation period. Index rotation appends a timestamp to
|
|
2674
|
-
* <code>IndexName</code> to facilitate the expiration of old data. For more information,
|
|
2675
|
-
* see <a href="https://docs.aws.amazon.com/firehose/latest/dev/basic-deliver.html#es-index-rotation">Index Rotation for the
|
|
2676
|
-
* Amazon ES Destination</a>. Default value is <code>OneDay</code>.</p>
|
|
2677
|
-
*/
|
|
1208
|
+
|
|
2678
1209
|
IndexRotationPeriod?: ElasticsearchIndexRotationPeriod | string;
|
|
2679
|
-
|
|
2680
|
-
* <p>The buffering options. If no value is specified,
|
|
2681
|
-
* <code>ElasticsearchBufferingHints</code> object default values are used. </p>
|
|
2682
|
-
*/
|
|
1210
|
+
|
|
2683
1211
|
BufferingHints?: ElasticsearchBufferingHints;
|
|
2684
|
-
|
|
2685
|
-
* <p>The retry behavior in case Kinesis Data Firehose is unable to deliver documents to
|
|
2686
|
-
* Amazon ES. The default value is 300 (5 minutes).</p>
|
|
2687
|
-
*/
|
|
1212
|
+
|
|
2688
1213
|
RetryOptions?: ElasticsearchRetryOptions;
|
|
2689
|
-
|
|
2690
|
-
* <p>The Amazon S3 destination.</p>
|
|
2691
|
-
*/
|
|
1214
|
+
|
|
2692
1215
|
S3Update?: S3DestinationUpdate;
|
|
2693
|
-
|
|
2694
|
-
* <p>The data processing configuration.</p>
|
|
2695
|
-
*/
|
|
1216
|
+
|
|
2696
1217
|
ProcessingConfiguration?: ProcessingConfiguration;
|
|
2697
|
-
|
|
2698
|
-
* <p>The CloudWatch logging options for your delivery stream.</p>
|
|
2699
|
-
*/
|
|
1218
|
+
|
|
2700
1219
|
CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
|
|
2701
1220
|
}
|
|
2702
1221
|
export declare namespace ElasticsearchDestinationUpdate {
|
|
2703
|
-
|
|
2704
|
-
* @internal
|
|
2705
|
-
*/
|
|
1222
|
+
|
|
2706
1223
|
const filterSensitiveLog: (obj: ElasticsearchDestinationUpdate) => any;
|
|
2707
1224
|
}
|
|
2708
|
-
|
|
2709
|
-
* <p>Describes an update for a destination in Amazon S3.</p>
|
|
2710
|
-
*/
|
|
1225
|
+
|
|
2711
1226
|
export interface ExtendedS3DestinationUpdate {
|
|
2712
|
-
|
|
2713
|
-
* <p>The Amazon Resource Name (ARN) of the AWS credentials. For more information, see
|
|
2714
|
-
* <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon
|
|
2715
|
-
* Resource Names (ARNs) and AWS Service Namespaces</a>.</p>
|
|
2716
|
-
*/
|
|
1227
|
+
|
|
2717
1228
|
RoleARN?: string;
|
|
2718
|
-
|
|
2719
|
-
* <p>The ARN of the S3 bucket. For more information, see <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon Resource Names (ARNs) and
|
|
2720
|
-
* AWS Service Namespaces</a>.</p>
|
|
2721
|
-
*/
|
|
1229
|
+
|
|
2722
1230
|
BucketARN?: string;
|
|
2723
|
-
|
|
2724
|
-
* <p>The "YYYY/MM/DD/HH" time format prefix is automatically used for delivered Amazon S3
|
|
2725
|
-
* files. You can also specify a custom prefix, as described in <a href="https://docs.aws.amazon.com/firehose/latest/dev/s3-prefixes.html">Custom Prefixes for Amazon S3
|
|
2726
|
-
* Objects</a>.</p>
|
|
2727
|
-
*/
|
|
1231
|
+
|
|
2728
1232
|
Prefix?: string;
|
|
2729
|
-
|
|
2730
|
-
* <p>A prefix that Kinesis Data Firehose evaluates and adds to failed records before writing
|
|
2731
|
-
* them to S3. This prefix appears immediately following the bucket name. For information
|
|
2732
|
-
* about how to specify this prefix, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/s3-prefixes.html">Custom Prefixes for Amazon S3
|
|
2733
|
-
* Objects</a>.</p>
|
|
2734
|
-
*/
|
|
1233
|
+
|
|
2735
1234
|
ErrorOutputPrefix?: string;
|
|
2736
|
-
|
|
2737
|
-
* <p>The buffering option.</p>
|
|
2738
|
-
*/
|
|
1235
|
+
|
|
2739
1236
|
BufferingHints?: BufferingHints;
|
|
2740
|
-
|
|
2741
|
-
* <p>The compression format. If no value is specified, the default is
|
|
2742
|
-
* <code>UNCOMPRESSED</code>. </p>
|
|
2743
|
-
*/
|
|
1237
|
+
|
|
2744
1238
|
CompressionFormat?: CompressionFormat | string;
|
|
2745
|
-
|
|
2746
|
-
* <p>The encryption configuration. If no value is specified, the default is no
|
|
2747
|
-
* encryption.</p>
|
|
2748
|
-
*/
|
|
1239
|
+
|
|
2749
1240
|
EncryptionConfiguration?: EncryptionConfiguration;
|
|
2750
|
-
|
|
2751
|
-
* <p>The Amazon CloudWatch logging options for your delivery stream.</p>
|
|
2752
|
-
*/
|
|
1241
|
+
|
|
2753
1242
|
CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
|
|
2754
|
-
|
|
2755
|
-
* <p>The data processing configuration.</p>
|
|
2756
|
-
*/
|
|
1243
|
+
|
|
2757
1244
|
ProcessingConfiguration?: ProcessingConfiguration;
|
|
2758
|
-
|
|
2759
|
-
* <p>You can update a delivery stream to enable Amazon S3 backup if it is disabled. If
|
|
2760
|
-
* backup is enabled, you can't update the delivery stream to disable it. </p>
|
|
2761
|
-
*/
|
|
1245
|
+
|
|
2762
1246
|
S3BackupMode?: S3BackupMode | string;
|
|
2763
|
-
|
|
2764
|
-
* <p>The Amazon S3 destination for backup.</p>
|
|
2765
|
-
*/
|
|
1247
|
+
|
|
2766
1248
|
S3BackupUpdate?: S3DestinationUpdate;
|
|
2767
|
-
|
|
2768
|
-
* <p>The serializer, deserializer, and schema for converting data from the JSON format to
|
|
2769
|
-
* the Parquet or ORC format before writing it to Amazon S3.</p>
|
|
2770
|
-
*/
|
|
1249
|
+
|
|
2771
1250
|
DataFormatConversionConfiguration?: DataFormatConversionConfiguration;
|
|
2772
|
-
|
|
2773
|
-
* <p>The configuration of the dynamic partitioning mechanism that creates smaller data sets
|
|
2774
|
-
* from the streaming data by partitioning it based on partition keys. Currently, dynamic
|
|
2775
|
-
* partitioning is only supported for Amazon S3 destinations. For more information, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/dynamic-partitioning.html">https://docs.aws.amazon.com/firehose/latest/dev/dynamic-partitioning.html</a>
|
|
2776
|
-
* </p>
|
|
2777
|
-
*/
|
|
1251
|
+
|
|
2778
1252
|
DynamicPartitioningConfiguration?: DynamicPartitioningConfiguration;
|
|
2779
1253
|
}
|
|
2780
1254
|
export declare namespace ExtendedS3DestinationUpdate {
|
|
2781
|
-
|
|
2782
|
-
* @internal
|
|
2783
|
-
*/
|
|
1255
|
+
|
|
2784
1256
|
const filterSensitiveLog: (obj: ExtendedS3DestinationUpdate) => any;
|
|
2785
1257
|
}
|
|
2786
1258
|
export interface ListDeliveryStreamsInput {
|
|
2787
|
-
|
|
2788
|
-
* <p>The maximum number of delivery streams to list. The default value is 10.</p>
|
|
2789
|
-
*/
|
|
1259
|
+
|
|
2790
1260
|
Limit?: number;
|
|
2791
|
-
|
|
2792
|
-
* <p>The delivery stream type. This can be one of the following values:</p>
|
|
2793
|
-
* <ul>
|
|
2794
|
-
* <li>
|
|
2795
|
-
* <p>
|
|
2796
|
-
* <code>DirectPut</code>: Provider applications access the delivery stream
|
|
2797
|
-
* directly.</p>
|
|
2798
|
-
* </li>
|
|
2799
|
-
* <li>
|
|
2800
|
-
* <p>
|
|
2801
|
-
* <code>KinesisStreamAsSource</code>: The delivery stream uses a Kinesis data
|
|
2802
|
-
* stream as a source.</p>
|
|
2803
|
-
* </li>
|
|
2804
|
-
* </ul>
|
|
2805
|
-
* <p>This parameter is optional. If this parameter is omitted, delivery streams of all
|
|
2806
|
-
* types are returned.</p>
|
|
2807
|
-
*/
|
|
1261
|
+
|
|
2808
1262
|
DeliveryStreamType?: DeliveryStreamType | string;
|
|
2809
|
-
|
|
2810
|
-
* <p>The list of delivery streams returned by this call to
|
|
2811
|
-
* <code>ListDeliveryStreams</code> will start with the delivery stream whose name comes
|
|
2812
|
-
* alphabetically immediately after the name you specify in
|
|
2813
|
-
* <code>ExclusiveStartDeliveryStreamName</code>.</p>
|
|
2814
|
-
*/
|
|
1263
|
+
|
|
2815
1264
|
ExclusiveStartDeliveryStreamName?: string;
|
|
2816
1265
|
}
|
|
2817
1266
|
export declare namespace ListDeliveryStreamsInput {
|
|
2818
|
-
|
|
2819
|
-
* @internal
|
|
2820
|
-
*/
|
|
1267
|
+
|
|
2821
1268
|
const filterSensitiveLog: (obj: ListDeliveryStreamsInput) => any;
|
|
2822
1269
|
}
|
|
2823
1270
|
export interface ListDeliveryStreamsOutput {
|
|
2824
|
-
|
|
2825
|
-
* <p>The names of the delivery streams.</p>
|
|
2826
|
-
*/
|
|
1271
|
+
|
|
2827
1272
|
DeliveryStreamNames: string[] | undefined;
|
|
2828
|
-
|
|
2829
|
-
* <p>Indicates whether there are more delivery streams available to list.</p>
|
|
2830
|
-
*/
|
|
1273
|
+
|
|
2831
1274
|
HasMoreDeliveryStreams: boolean | undefined;
|
|
2832
1275
|
}
|
|
2833
1276
|
export declare namespace ListDeliveryStreamsOutput {
|
|
2834
|
-
|
|
2835
|
-
* @internal
|
|
2836
|
-
*/
|
|
1277
|
+
|
|
2837
1278
|
const filterSensitiveLog: (obj: ListDeliveryStreamsOutput) => any;
|
|
2838
1279
|
}
|
|
2839
1280
|
export interface ListTagsForDeliveryStreamInput {
|
|
2840
|
-
|
|
2841
|
-
* <p>The name of the delivery stream whose tags you want to list.</p>
|
|
2842
|
-
*/
|
|
1281
|
+
|
|
2843
1282
|
DeliveryStreamName: string | undefined;
|
|
2844
|
-
|
|
2845
|
-
* <p>The key to use as the starting point for the list of tags. If you set this parameter,
|
|
2846
|
-
* <code>ListTagsForDeliveryStream</code> gets all tags that occur after
|
|
2847
|
-
* <code>ExclusiveStartTagKey</code>.</p>
|
|
2848
|
-
*/
|
|
1283
|
+
|
|
2849
1284
|
ExclusiveStartTagKey?: string;
|
|
2850
|
-
|
|
2851
|
-
* <p>The number of tags to return. If this number is less than the total number of tags
|
|
2852
|
-
* associated with the delivery stream, <code>HasMoreTags</code> is set to <code>true</code>
|
|
2853
|
-
* in the response. To list additional tags, set <code>ExclusiveStartTagKey</code> to the last
|
|
2854
|
-
* key in the response. </p>
|
|
2855
|
-
*/
|
|
1285
|
+
|
|
2856
1286
|
Limit?: number;
|
|
2857
1287
|
}
|
|
2858
1288
|
export declare namespace ListTagsForDeliveryStreamInput {
|
|
2859
|
-
|
|
2860
|
-
* @internal
|
|
2861
|
-
*/
|
|
1289
|
+
|
|
2862
1290
|
const filterSensitiveLog: (obj: ListTagsForDeliveryStreamInput) => any;
|
|
2863
1291
|
}
|
|
2864
1292
|
export interface ListTagsForDeliveryStreamOutput {
|
|
2865
|
-
|
|
2866
|
-
* <p>A list of tags associated with <code>DeliveryStreamName</code>, starting with the
|
|
2867
|
-
* first tag after <code>ExclusiveStartTagKey</code> and up to the specified
|
|
2868
|
-
* <code>Limit</code>.</p>
|
|
2869
|
-
*/
|
|
1293
|
+
|
|
2870
1294
|
Tags: Tag[] | undefined;
|
|
2871
|
-
|
|
2872
|
-
* <p>If this is <code>true</code> in the response, more tags are available. To list the
|
|
2873
|
-
* remaining tags, set <code>ExclusiveStartTagKey</code> to the key of the last tag returned
|
|
2874
|
-
* and call <code>ListTagsForDeliveryStream</code> again.</p>
|
|
2875
|
-
*/
|
|
1295
|
+
|
|
2876
1296
|
HasMoreTags: boolean | undefined;
|
|
2877
1297
|
}
|
|
2878
1298
|
export declare namespace ListTagsForDeliveryStreamOutput {
|
|
2879
|
-
|
|
2880
|
-
* @internal
|
|
2881
|
-
*/
|
|
1299
|
+
|
|
2882
1300
|
const filterSensitiveLog: (obj: ListTagsForDeliveryStreamOutput) => any;
|
|
2883
1301
|
}
|
|
2884
|
-
|
|
2885
|
-
* <p>The unit of data in a delivery stream.</p>
|
|
2886
|
-
*/
|
|
1302
|
+
|
|
2887
1303
|
export interface _Record {
|
|
2888
|
-
|
|
2889
|
-
* <p>The data blob, which is base64-encoded when the blob is serialized. The maximum size
|
|
2890
|
-
* of the data blob, before base64-encoding, is 1,000 KiB.</p>
|
|
2891
|
-
*/
|
|
1304
|
+
|
|
2892
1305
|
Data: Uint8Array | undefined;
|
|
2893
1306
|
}
|
|
2894
1307
|
export declare namespace _Record {
|
|
2895
|
-
|
|
2896
|
-
* @internal
|
|
2897
|
-
*/
|
|
1308
|
+
|
|
2898
1309
|
const filterSensitiveLog: (obj: _Record) => any;
|
|
2899
1310
|
}
|
|
2900
1311
|
export interface PutRecordInput {
|
|
2901
|
-
|
|
2902
|
-
* <p>The name of the delivery stream.</p>
|
|
2903
|
-
*/
|
|
1312
|
+
|
|
2904
1313
|
DeliveryStreamName: string | undefined;
|
|
2905
|
-
|
|
2906
|
-
* <p>The record.</p>
|
|
2907
|
-
*/
|
|
1314
|
+
|
|
2908
1315
|
Record: _Record | undefined;
|
|
2909
1316
|
}
|
|
2910
1317
|
export declare namespace PutRecordInput {
|
|
2911
|
-
|
|
2912
|
-
* @internal
|
|
2913
|
-
*/
|
|
1318
|
+
|
|
2914
1319
|
const filterSensitiveLog: (obj: PutRecordInput) => any;
|
|
2915
1320
|
}
|
|
2916
1321
|
export interface PutRecordOutput {
|
|
2917
|
-
|
|
2918
|
-
* <p>The ID of the record.</p>
|
|
2919
|
-
*/
|
|
1322
|
+
|
|
2920
1323
|
RecordId: string | undefined;
|
|
2921
|
-
|
|
2922
|
-
* <p>Indicates whether server-side encryption (SSE) was enabled during this operation.</p>
|
|
2923
|
-
*/
|
|
1324
|
+
|
|
2924
1325
|
Encrypted?: boolean;
|
|
2925
1326
|
}
|
|
2926
1327
|
export declare namespace PutRecordOutput {
|
|
2927
|
-
|
|
2928
|
-
* @internal
|
|
2929
|
-
*/
|
|
1328
|
+
|
|
2930
1329
|
const filterSensitiveLog: (obj: PutRecordOutput) => any;
|
|
2931
1330
|
}
|
|
2932
|
-
|
|
2933
|
-
* <p>The service is unavailable. Back off and retry the operation. If you continue to see
|
|
2934
|
-
* the exception, throughput limits for the delivery stream may have been exceeded. For more
|
|
2935
|
-
* information about limits and how to request an increase, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/limits.html">Amazon Kinesis Data Firehose
|
|
2936
|
-
* Limits</a>.</p>
|
|
2937
|
-
*/
|
|
1331
|
+
|
|
2938
1332
|
export interface ServiceUnavailableException extends __SmithyException, $MetadataBearer {
|
|
2939
1333
|
name: "ServiceUnavailableException";
|
|
2940
1334
|
$fault: "server";
|
|
2941
|
-
|
|
2942
|
-
* <p>A message that provides information about the error.</p>
|
|
2943
|
-
*/
|
|
1335
|
+
|
|
2944
1336
|
message?: string;
|
|
2945
1337
|
}
|
|
2946
1338
|
export declare namespace ServiceUnavailableException {
|
|
2947
|
-
|
|
2948
|
-
* @internal
|
|
2949
|
-
*/
|
|
1339
|
+
|
|
2950
1340
|
const filterSensitiveLog: (obj: ServiceUnavailableException) => any;
|
|
2951
1341
|
}
|
|
2952
1342
|
export interface PutRecordBatchInput {
|
|
2953
|
-
|
|
2954
|
-
* <p>The name of the delivery stream.</p>
|
|
2955
|
-
*/
|
|
1343
|
+
|
|
2956
1344
|
DeliveryStreamName: string | undefined;
|
|
2957
|
-
|
|
2958
|
-
* <p>One or more records.</p>
|
|
2959
|
-
*/
|
|
1345
|
+
|
|
2960
1346
|
Records: _Record[] | undefined;
|
|
2961
1347
|
}
|
|
2962
1348
|
export declare namespace PutRecordBatchInput {
|
|
2963
|
-
|
|
2964
|
-
* @internal
|
|
2965
|
-
*/
|
|
1349
|
+
|
|
2966
1350
|
const filterSensitiveLog: (obj: PutRecordBatchInput) => any;
|
|
2967
1351
|
}
|
|
2968
|
-
|
|
2969
|
-
* <p>Contains the result for an individual record from a <a>PutRecordBatch</a>
|
|
2970
|
-
* request. If the record is successfully added to your delivery stream, it receives a record
|
|
2971
|
-
* ID. If the record fails to be added to your delivery stream, the result includes an error
|
|
2972
|
-
* code and an error message.</p>
|
|
2973
|
-
*/
|
|
1352
|
+
|
|
2974
1353
|
export interface PutRecordBatchResponseEntry {
|
|
2975
|
-
|
|
2976
|
-
* <p>The ID of the record.</p>
|
|
2977
|
-
*/
|
|
1354
|
+
|
|
2978
1355
|
RecordId?: string;
|
|
2979
|
-
|
|
2980
|
-
* <p>The error code for an individual record result.</p>
|
|
2981
|
-
*/
|
|
1356
|
+
|
|
2982
1357
|
ErrorCode?: string;
|
|
2983
|
-
|
|
2984
|
-
* <p>The error message for an individual record result.</p>
|
|
2985
|
-
*/
|
|
1358
|
+
|
|
2986
1359
|
ErrorMessage?: string;
|
|
2987
1360
|
}
|
|
2988
1361
|
export declare namespace PutRecordBatchResponseEntry {
|
|
2989
|
-
|
|
2990
|
-
* @internal
|
|
2991
|
-
*/
|
|
1362
|
+
|
|
2992
1363
|
const filterSensitiveLog: (obj: PutRecordBatchResponseEntry) => any;
|
|
2993
1364
|
}
|
|
2994
1365
|
export interface PutRecordBatchOutput {
|
|
2995
|
-
|
|
2996
|
-
* <p>The number of records that might have failed processing. This number might be greater
|
|
2997
|
-
* than 0 even if the <a>PutRecordBatch</a> call succeeds. Check
|
|
2998
|
-
* <code>FailedPutCount</code> to determine whether there are records that you need to
|
|
2999
|
-
* resend.</p>
|
|
3000
|
-
*/
|
|
1366
|
+
|
|
3001
1367
|
FailedPutCount: number | undefined;
|
|
3002
|
-
|
|
3003
|
-
* <p>Indicates whether server-side encryption (SSE) was enabled during this operation.</p>
|
|
3004
|
-
*/
|
|
1368
|
+
|
|
3005
1369
|
Encrypted?: boolean;
|
|
3006
|
-
|
|
3007
|
-
* <p>The results array. For each record, the index of the response element is the same as
|
|
3008
|
-
* the index used in the request array.</p>
|
|
3009
|
-
*/
|
|
1370
|
+
|
|
3010
1371
|
RequestResponses: PutRecordBatchResponseEntry[] | undefined;
|
|
3011
1372
|
}
|
|
3012
1373
|
export declare namespace PutRecordBatchOutput {
|
|
3013
|
-
|
|
3014
|
-
* @internal
|
|
3015
|
-
*/
|
|
1374
|
+
|
|
3016
1375
|
const filterSensitiveLog: (obj: PutRecordBatchOutput) => any;
|
|
3017
1376
|
}
|
|
3018
1377
|
export interface StartDeliveryStreamEncryptionInput {
|
|
3019
|
-
|
|
3020
|
-
* <p>The name of the delivery stream for which you want to enable server-side encryption
|
|
3021
|
-
* (SSE).</p>
|
|
3022
|
-
*/
|
|
1378
|
+
|
|
3023
1379
|
DeliveryStreamName: string | undefined;
|
|
3024
|
-
|
|
3025
|
-
* <p>Used to specify the type and Amazon Resource Name (ARN) of the KMS key needed for
|
|
3026
|
-
* Server-Side Encryption (SSE).</p>
|
|
3027
|
-
*/
|
|
1380
|
+
|
|
3028
1381
|
DeliveryStreamEncryptionConfigurationInput?: DeliveryStreamEncryptionConfigurationInput;
|
|
3029
1382
|
}
|
|
3030
1383
|
export declare namespace StartDeliveryStreamEncryptionInput {
|
|
3031
|
-
|
|
3032
|
-
* @internal
|
|
3033
|
-
*/
|
|
1384
|
+
|
|
3034
1385
|
const filterSensitiveLog: (obj: StartDeliveryStreamEncryptionInput) => any;
|
|
3035
1386
|
}
|
|
3036
1387
|
export interface StartDeliveryStreamEncryptionOutput {
|
|
3037
1388
|
}
|
|
3038
1389
|
export declare namespace StartDeliveryStreamEncryptionOutput {
|
|
3039
|
-
|
|
3040
|
-
* @internal
|
|
3041
|
-
*/
|
|
1390
|
+
|
|
3042
1391
|
const filterSensitiveLog: (obj: StartDeliveryStreamEncryptionOutput) => any;
|
|
3043
1392
|
}
|
|
3044
1393
|
export interface StopDeliveryStreamEncryptionInput {
|
|
3045
|
-
|
|
3046
|
-
* <p>The name of the delivery stream for which you want to disable server-side encryption
|
|
3047
|
-
* (SSE).</p>
|
|
3048
|
-
*/
|
|
1394
|
+
|
|
3049
1395
|
DeliveryStreamName: string | undefined;
|
|
3050
1396
|
}
|
|
3051
1397
|
export declare namespace StopDeliveryStreamEncryptionInput {
|
|
3052
|
-
|
|
3053
|
-
* @internal
|
|
3054
|
-
*/
|
|
1398
|
+
|
|
3055
1399
|
const filterSensitiveLog: (obj: StopDeliveryStreamEncryptionInput) => any;
|
|
3056
1400
|
}
|
|
3057
1401
|
export interface StopDeliveryStreamEncryptionOutput {
|
|
3058
1402
|
}
|
|
3059
1403
|
export declare namespace StopDeliveryStreamEncryptionOutput {
|
|
3060
|
-
|
|
3061
|
-
* @internal
|
|
3062
|
-
*/
|
|
1404
|
+
|
|
3063
1405
|
const filterSensitiveLog: (obj: StopDeliveryStreamEncryptionOutput) => any;
|
|
3064
1406
|
}
|
|
3065
1407
|
export interface TagDeliveryStreamInput {
|
|
3066
|
-
|
|
3067
|
-
* <p>The name of the delivery stream to which you want to add the tags.</p>
|
|
3068
|
-
*/
|
|
1408
|
+
|
|
3069
1409
|
DeliveryStreamName: string | undefined;
|
|
3070
|
-
|
|
3071
|
-
* <p>A set of key-value pairs to use to create the tags.</p>
|
|
3072
|
-
*/
|
|
1410
|
+
|
|
3073
1411
|
Tags: Tag[] | undefined;
|
|
3074
1412
|
}
|
|
3075
1413
|
export declare namespace TagDeliveryStreamInput {
|
|
3076
|
-
|
|
3077
|
-
* @internal
|
|
3078
|
-
*/
|
|
1414
|
+
|
|
3079
1415
|
const filterSensitiveLog: (obj: TagDeliveryStreamInput) => any;
|
|
3080
1416
|
}
|
|
3081
1417
|
export interface TagDeliveryStreamOutput {
|
|
3082
1418
|
}
|
|
3083
1419
|
export declare namespace TagDeliveryStreamOutput {
|
|
3084
|
-
|
|
3085
|
-
* @internal
|
|
3086
|
-
*/
|
|
1420
|
+
|
|
3087
1421
|
const filterSensitiveLog: (obj: TagDeliveryStreamOutput) => any;
|
|
3088
1422
|
}
|
|
3089
1423
|
export interface UntagDeliveryStreamInput {
|
|
3090
|
-
|
|
3091
|
-
* <p>The name of the delivery stream.</p>
|
|
3092
|
-
*/
|
|
1424
|
+
|
|
3093
1425
|
DeliveryStreamName: string | undefined;
|
|
3094
|
-
|
|
3095
|
-
* <p>A list of tag keys. Each corresponding tag is removed from the delivery
|
|
3096
|
-
* stream.</p>
|
|
3097
|
-
*/
|
|
1426
|
+
|
|
3098
1427
|
TagKeys: string[] | undefined;
|
|
3099
1428
|
}
|
|
3100
1429
|
export declare namespace UntagDeliveryStreamInput {
|
|
3101
|
-
|
|
3102
|
-
* @internal
|
|
3103
|
-
*/
|
|
1430
|
+
|
|
3104
1431
|
const filterSensitiveLog: (obj: UntagDeliveryStreamInput) => any;
|
|
3105
1432
|
}
|
|
3106
1433
|
export interface UntagDeliveryStreamOutput {
|
|
3107
1434
|
}
|
|
3108
1435
|
export declare namespace UntagDeliveryStreamOutput {
|
|
3109
|
-
|
|
3110
|
-
* @internal
|
|
3111
|
-
*/
|
|
1436
|
+
|
|
3112
1437
|
const filterSensitiveLog: (obj: UntagDeliveryStreamOutput) => any;
|
|
3113
1438
|
}
|
|
3114
|
-
|
|
3115
|
-
* <p>Updates the specified HTTP endpoint destination.</p>
|
|
3116
|
-
*/
|
|
1439
|
+
|
|
3117
1440
|
export interface HttpEndpointDestinationUpdate {
|
|
3118
|
-
|
|
3119
|
-
* <p>Describes the configuration of the HTTP endpoint destination.</p>
|
|
3120
|
-
*/
|
|
1441
|
+
|
|
3121
1442
|
EndpointConfiguration?: HttpEndpointConfiguration;
|
|
3122
|
-
|
|
3123
|
-
* <p>Describes buffering options that can be applied to the data before it is delivered to
|
|
3124
|
-
* the HTTPS endpoint destination. Kinesis Data Firehose teats these options as hints, and it
|
|
3125
|
-
* might choose to use more optimal values. The <code>SizeInMBs</code> and
|
|
3126
|
-
* <code>IntervalInSeconds</code> parameters are optional. However, if specify a value for
|
|
3127
|
-
* one of them, you must also provide a value for the other. </p>
|
|
3128
|
-
*/
|
|
1443
|
+
|
|
3129
1444
|
BufferingHints?: HttpEndpointBufferingHints;
|
|
3130
|
-
|
|
3131
|
-
* <p>Describes the Amazon CloudWatch logging options for your delivery stream.</p>
|
|
3132
|
-
*/
|
|
1445
|
+
|
|
3133
1446
|
CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
|
|
3134
|
-
|
|
3135
|
-
* <p>The configuration of the request sent to the HTTP endpoint specified as the
|
|
3136
|
-
* destination.</p>
|
|
3137
|
-
*/
|
|
1447
|
+
|
|
3138
1448
|
RequestConfiguration?: HttpEndpointRequestConfiguration;
|
|
3139
|
-
|
|
3140
|
-
* <p>Describes a data processing configuration.</p>
|
|
3141
|
-
*/
|
|
1449
|
+
|
|
3142
1450
|
ProcessingConfiguration?: ProcessingConfiguration;
|
|
3143
|
-
|
|
3144
|
-
* <p>Kinesis Data Firehose uses this IAM role for all the permissions that the delivery
|
|
3145
|
-
* stream needs.</p>
|
|
3146
|
-
*/
|
|
1451
|
+
|
|
3147
1452
|
RoleARN?: string;
|
|
3148
|
-
|
|
3149
|
-
* <p>Describes the retry behavior in case Kinesis Data Firehose is unable to deliver data to
|
|
3150
|
-
* the specified HTTP endpoint destination, or if it doesn't receive a valid acknowledgment of
|
|
3151
|
-
* receipt from the specified HTTP endpoint destination.</p>
|
|
3152
|
-
*/
|
|
1453
|
+
|
|
3153
1454
|
RetryOptions?: HttpEndpointRetryOptions;
|
|
3154
|
-
|
|
3155
|
-
* <p>Describes the S3 bucket backup options for the data that Kinesis Firehose delivers to
|
|
3156
|
-
* the HTTP endpoint destination. You can back up all documents (<code>AllData</code>) or only
|
|
3157
|
-
* the documents that Kinesis Data Firehose could not deliver to the specified HTTP endpoint
|
|
3158
|
-
* destination (<code>FailedDataOnly</code>).</p>
|
|
3159
|
-
*/
|
|
1455
|
+
|
|
3160
1456
|
S3BackupMode?: HttpEndpointS3BackupMode | string;
|
|
3161
|
-
|
|
3162
|
-
* <p>Describes an update for a destination in Amazon S3.</p>
|
|
3163
|
-
*/
|
|
1457
|
+
|
|
3164
1458
|
S3Update?: S3DestinationUpdate;
|
|
3165
1459
|
}
|
|
3166
1460
|
export declare namespace HttpEndpointDestinationUpdate {
|
|
3167
|
-
|
|
3168
|
-
* @internal
|
|
3169
|
-
*/
|
|
1461
|
+
|
|
3170
1462
|
const filterSensitiveLog: (obj: HttpEndpointDestinationUpdate) => any;
|
|
3171
1463
|
}
|
|
3172
|
-
|
|
3173
|
-
* <p>Describes an update for a destination in Amazon Redshift.</p>
|
|
3174
|
-
*/
|
|
1464
|
+
|
|
3175
1465
|
export interface RedshiftDestinationUpdate {
|
|
3176
|
-
|
|
3177
|
-
* <p>The Amazon Resource Name (ARN) of the AWS credentials. For more information, see
|
|
3178
|
-
* <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon
|
|
3179
|
-
* Resource Names (ARNs) and AWS Service Namespaces</a>.</p>
|
|
3180
|
-
*/
|
|
1466
|
+
|
|
3181
1467
|
RoleARN?: string;
|
|
3182
|
-
|
|
3183
|
-
* <p>The database connection string.</p>
|
|
3184
|
-
*/
|
|
1468
|
+
|
|
3185
1469
|
ClusterJDBCURL?: string;
|
|
3186
|
-
|
|
3187
|
-
* <p>The <code>COPY</code> command.</p>
|
|
3188
|
-
*/
|
|
1470
|
+
|
|
3189
1471
|
CopyCommand?: CopyCommand;
|
|
3190
|
-
|
|
3191
|
-
* <p>The name of the user.</p>
|
|
3192
|
-
*/
|
|
1472
|
+
|
|
3193
1473
|
Username?: string;
|
|
3194
|
-
|
|
3195
|
-
* <p>The user password.</p>
|
|
3196
|
-
*/
|
|
1474
|
+
|
|
3197
1475
|
Password?: string;
|
|
3198
|
-
|
|
3199
|
-
* <p>The retry behavior in case Kinesis Data Firehose is unable to deliver documents to
|
|
3200
|
-
* Amazon Redshift. Default value is 3600 (60 minutes).</p>
|
|
3201
|
-
*/
|
|
1476
|
+
|
|
3202
1477
|
RetryOptions?: RedshiftRetryOptions;
|
|
3203
|
-
|
|
3204
|
-
* <p>The Amazon S3 destination.</p>
|
|
3205
|
-
* <p>The compression formats <code>SNAPPY</code> or <code>ZIP</code> cannot be specified
|
|
3206
|
-
* in <code>RedshiftDestinationUpdate.S3Update</code> because the Amazon Redshift
|
|
3207
|
-
* <code>COPY</code> operation that reads from the S3 bucket doesn't support these
|
|
3208
|
-
* compression formats.</p>
|
|
3209
|
-
*/
|
|
1478
|
+
|
|
3210
1479
|
S3Update?: S3DestinationUpdate;
|
|
3211
|
-
|
|
3212
|
-
* <p>The data processing configuration.</p>
|
|
3213
|
-
*/
|
|
1480
|
+
|
|
3214
1481
|
ProcessingConfiguration?: ProcessingConfiguration;
|
|
3215
|
-
|
|
3216
|
-
* <p>You can update a delivery stream to enable Amazon S3 backup if it is disabled. If
|
|
3217
|
-
* backup is enabled, you can't update the delivery stream to disable it. </p>
|
|
3218
|
-
*/
|
|
1482
|
+
|
|
3219
1483
|
S3BackupMode?: RedshiftS3BackupMode | string;
|
|
3220
|
-
|
|
3221
|
-
* <p>The Amazon S3 destination for backup.</p>
|
|
3222
|
-
*/
|
|
1484
|
+
|
|
3223
1485
|
S3BackupUpdate?: S3DestinationUpdate;
|
|
3224
|
-
|
|
3225
|
-
* <p>The Amazon CloudWatch logging options for your delivery stream.</p>
|
|
3226
|
-
*/
|
|
1486
|
+
|
|
3227
1487
|
CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
|
|
3228
1488
|
}
|
|
3229
1489
|
export declare namespace RedshiftDestinationUpdate {
|
|
3230
|
-
|
|
3231
|
-
* @internal
|
|
3232
|
-
*/
|
|
1490
|
+
|
|
3233
1491
|
const filterSensitiveLog: (obj: RedshiftDestinationUpdate) => any;
|
|
3234
1492
|
}
|
|
3235
|
-
|
|
3236
|
-
* <p>Describes an update for a destination in Splunk.</p>
|
|
3237
|
-
*/
|
|
1493
|
+
|
|
3238
1494
|
export interface SplunkDestinationUpdate {
|
|
3239
|
-
|
|
3240
|
-
* <p>The HTTP Event Collector (HEC) endpoint to which Kinesis Data Firehose sends your
|
|
3241
|
-
* data.</p>
|
|
3242
|
-
*/
|
|
1495
|
+
|
|
3243
1496
|
HECEndpoint?: string;
|
|
3244
|
-
|
|
3245
|
-
* <p>This type can be either "Raw" or "Event."</p>
|
|
3246
|
-
*/
|
|
1497
|
+
|
|
3247
1498
|
HECEndpointType?: HECEndpointType | string;
|
|
3248
|
-
|
|
3249
|
-
* <p>A GUID that you obtain from your Splunk cluster when you create a new HEC
|
|
3250
|
-
* endpoint.</p>
|
|
3251
|
-
*/
|
|
1499
|
+
|
|
3252
1500
|
HECToken?: string;
|
|
3253
|
-
|
|
3254
|
-
* <p>The amount of time that Kinesis Data Firehose waits to receive an acknowledgment from
|
|
3255
|
-
* Splunk after it sends data. At the end of the timeout period, Kinesis Data Firehose either
|
|
3256
|
-
* tries to send the data again or considers it an error, based on your retry
|
|
3257
|
-
* settings.</p>
|
|
3258
|
-
*/
|
|
1501
|
+
|
|
3259
1502
|
HECAcknowledgmentTimeoutInSeconds?: number;
|
|
3260
|
-
|
|
3261
|
-
* <p>The retry behavior in case Kinesis Data Firehose is unable to deliver data to Splunk
|
|
3262
|
-
* or if it doesn't receive an acknowledgment of receipt from Splunk.</p>
|
|
3263
|
-
*/
|
|
1503
|
+
|
|
3264
1504
|
RetryOptions?: SplunkRetryOptions;
|
|
3265
|
-
|
|
3266
|
-
* <p>Specifies how you want Kinesis Data Firehose to back up documents to Amazon S3. When
|
|
3267
|
-
* set to <code>FailedDocumentsOnly</code>, Kinesis Data Firehose writes any data that could
|
|
3268
|
-
* not be indexed to the configured Amazon S3 destination. When set to <code>AllEvents</code>,
|
|
3269
|
-
* Kinesis Data Firehose delivers all incoming records to Amazon S3, and also writes failed
|
|
3270
|
-
* documents to Amazon S3. The default value is <code>FailedEventsOnly</code>.</p>
|
|
3271
|
-
* <p>You can update this backup mode from <code>FailedEventsOnly</code> to
|
|
3272
|
-
* <code>AllEvents</code>. You can't update it from <code>AllEvents</code> to
|
|
3273
|
-
* <code>FailedEventsOnly</code>.</p>
|
|
3274
|
-
*/
|
|
1505
|
+
|
|
3275
1506
|
S3BackupMode?: SplunkS3BackupMode | string;
|
|
3276
|
-
|
|
3277
|
-
* <p>Your update to the configuration of the backup Amazon S3 location.</p>
|
|
3278
|
-
*/
|
|
1507
|
+
|
|
3279
1508
|
S3Update?: S3DestinationUpdate;
|
|
3280
|
-
|
|
3281
|
-
* <p>The data processing configuration.</p>
|
|
3282
|
-
*/
|
|
1509
|
+
|
|
3283
1510
|
ProcessingConfiguration?: ProcessingConfiguration;
|
|
3284
|
-
|
|
3285
|
-
* <p>The Amazon CloudWatch logging options for your delivery stream.</p>
|
|
3286
|
-
*/
|
|
1511
|
+
|
|
3287
1512
|
CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
|
|
3288
1513
|
}
|
|
3289
1514
|
export declare namespace SplunkDestinationUpdate {
|
|
3290
|
-
|
|
3291
|
-
* @internal
|
|
3292
|
-
*/
|
|
1515
|
+
|
|
3293
1516
|
const filterSensitiveLog: (obj: SplunkDestinationUpdate) => any;
|
|
3294
1517
|
}
|
|
3295
1518
|
export interface UpdateDestinationInput {
|
|
3296
|
-
|
|
3297
|
-
* <p>The name of the delivery stream.</p>
|
|
3298
|
-
*/
|
|
1519
|
+
|
|
3299
1520
|
DeliveryStreamName: string | undefined;
|
|
3300
|
-
|
|
3301
|
-
* <p>Obtain this value from the <code>VersionId</code> result of <a>DeliveryStreamDescription</a>. This value is required, and helps the service
|
|
3302
|
-
* perform conditional operations. For example, if there is an interleaving update and this
|
|
3303
|
-
* value is null, then the update destination fails. After the update is successful, the
|
|
3304
|
-
* <code>VersionId</code> value is updated. The service then performs a merge of the old
|
|
3305
|
-
* configuration with the new configuration.</p>
|
|
3306
|
-
*/
|
|
1521
|
+
|
|
3307
1522
|
CurrentDeliveryStreamVersionId: string | undefined;
|
|
3308
|
-
|
|
3309
|
-
* <p>The ID of the destination.</p>
|
|
3310
|
-
*/
|
|
1523
|
+
|
|
3311
1524
|
DestinationId: string | undefined;
|
|
3312
|
-
|
|
3313
|
-
* @deprecated
|
|
3314
|
-
*
|
|
3315
|
-
* <p>[Deprecated] Describes an update for a destination in Amazon S3.</p>
|
|
3316
|
-
*/
|
|
1525
|
+
|
|
3317
1526
|
S3DestinationUpdate?: S3DestinationUpdate;
|
|
3318
|
-
|
|
3319
|
-
* <p>Describes an update for a destination in Amazon S3.</p>
|
|
3320
|
-
*/
|
|
1527
|
+
|
|
3321
1528
|
ExtendedS3DestinationUpdate?: ExtendedS3DestinationUpdate;
|
|
3322
|
-
|
|
3323
|
-
* <p>Describes an update for a destination in Amazon Redshift.</p>
|
|
3324
|
-
*/
|
|
1529
|
+
|
|
3325
1530
|
RedshiftDestinationUpdate?: RedshiftDestinationUpdate;
|
|
3326
|
-
|
|
3327
|
-
* <p>Describes an update for a destination in Amazon ES.</p>
|
|
3328
|
-
*/
|
|
1531
|
+
|
|
3329
1532
|
ElasticsearchDestinationUpdate?: ElasticsearchDestinationUpdate;
|
|
3330
1533
|
AmazonopensearchserviceDestinationUpdate?: AmazonopensearchserviceDestinationUpdate;
|
|
3331
|
-
|
|
3332
|
-
* <p>Describes an update for a destination in Splunk.</p>
|
|
3333
|
-
*/
|
|
1534
|
+
|
|
3334
1535
|
SplunkDestinationUpdate?: SplunkDestinationUpdate;
|
|
3335
|
-
|
|
3336
|
-
* <p>Describes an update to the specified HTTP endpoint destination.</p>
|
|
3337
|
-
*/
|
|
1536
|
+
|
|
3338
1537
|
HttpEndpointDestinationUpdate?: HttpEndpointDestinationUpdate;
|
|
3339
1538
|
}
|
|
3340
1539
|
export declare namespace UpdateDestinationInput {
|
|
3341
|
-
|
|
3342
|
-
* @internal
|
|
3343
|
-
*/
|
|
1540
|
+
|
|
3344
1541
|
const filterSensitiveLog: (obj: UpdateDestinationInput) => any;
|
|
3345
1542
|
}
|
|
3346
1543
|
export interface UpdateDestinationOutput {
|
|
3347
1544
|
}
|
|
3348
1545
|
export declare namespace UpdateDestinationOutput {
|
|
3349
|
-
|
|
3350
|
-
* @internal
|
|
3351
|
-
*/
|
|
1546
|
+
|
|
3352
1547
|
const filterSensitiveLog: (obj: UpdateDestinationOutput) => any;
|
|
3353
1548
|
}
|