@aws-sdk/client-firehose 3.36.0 → 3.39.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (51) hide show
  1. package/CHANGELOG.md +42 -0
  2. package/dist-cjs/Firehose.js +0 -6
  3. package/dist-cjs/FirehoseClient.js +0 -11
  4. package/dist-cjs/commands/CreateDeliveryStreamCommand.js +0 -92
  5. package/dist-cjs/commands/DeleteDeliveryStreamCommand.js +0 -33
  6. package/dist-cjs/commands/DescribeDeliveryStreamCommand.js +0 -30
  7. package/dist-cjs/commands/ListDeliveryStreamsCommand.js +0 -31
  8. package/dist-cjs/commands/ListTagsForDeliveryStreamCommand.js +0 -25
  9. package/dist-cjs/commands/PutRecordBatchCommand.js +0 -74
  10. package/dist-cjs/commands/PutRecordCommand.js +0 -55
  11. package/dist-cjs/commands/StartDeliveryStreamEncryptionCommand.js +0 -59
  12. package/dist-cjs/commands/StopDeliveryStreamEncryptionCommand.js +0 -43
  13. package/dist-cjs/commands/TagDeliveryStreamCommand.js +0 -32
  14. package/dist-cjs/commands/UntagDeliveryStreamCommand.js +0 -27
  15. package/dist-cjs/commands/UpdateDestinationCommand.js +0 -47
  16. package/dist-cjs/commands/index.js +15 -0
  17. package/dist-cjs/endpoints.js +6 -1
  18. package/dist-cjs/index.js +3 -14
  19. package/dist-cjs/models/models_0.js +0 -291
  20. package/dist-cjs/protocols/Aws_json1_1.js +0 -5
  21. package/dist-cjs/runtimeConfig.browser.js +1 -5
  22. package/dist-cjs/runtimeConfig.js +1 -5
  23. package/dist-cjs/runtimeConfig.native.js +0 -3
  24. package/dist-cjs/runtimeConfig.shared.js +0 -3
  25. package/dist-es/commands/index.js +12 -0
  26. package/dist-es/endpoints.js +6 -1
  27. package/dist-es/index.js +3 -14
  28. package/dist-types/commands/index.d.ts +12 -0
  29. package/dist-types/index.d.ts +3 -14
  30. package/dist-types/ts3.4/Firehose.d.ts +13 -302
  31. package/dist-types/ts3.4/FirehoseClient.d.ts +24 -92
  32. package/dist-types/ts3.4/commands/CreateDeliveryStreamCommand.d.ts +2 -88
  33. package/dist-types/ts3.4/commands/DeleteDeliveryStreamCommand.d.ts +2 -29
  34. package/dist-types/ts3.4/commands/DescribeDeliveryStreamCommand.d.ts +2 -26
  35. package/dist-types/ts3.4/commands/ListDeliveryStreamsCommand.d.ts +2 -27
  36. package/dist-types/ts3.4/commands/ListTagsForDeliveryStreamCommand.d.ts +2 -21
  37. package/dist-types/ts3.4/commands/PutRecordBatchCommand.d.ts +2 -70
  38. package/dist-types/ts3.4/commands/PutRecordCommand.d.ts +2 -51
  39. package/dist-types/ts3.4/commands/StartDeliveryStreamEncryptionCommand.d.ts +2 -55
  40. package/dist-types/ts3.4/commands/StopDeliveryStreamEncryptionCommand.d.ts +2 -39
  41. package/dist-types/ts3.4/commands/TagDeliveryStreamCommand.d.ts +2 -28
  42. package/dist-types/ts3.4/commands/UntagDeliveryStreamCommand.d.ts +2 -23
  43. package/dist-types/ts3.4/commands/UpdateDestinationCommand.d.ts +2 -43
  44. package/dist-types/ts3.4/commands/index.d.ts +12 -0
  45. package/dist-types/ts3.4/index.d.ts +3 -14
  46. package/dist-types/ts3.4/models/models_0.d.ts +540 -2345
  47. package/dist-types/ts3.4/runtimeConfig.browser.d.ts +1 -3
  48. package/dist-types/ts3.4/runtimeConfig.d.ts +1 -3
  49. package/dist-types/ts3.4/runtimeConfig.native.d.ts +1 -3
  50. package/dist-types/ts3.4/runtimeConfig.shared.d.ts +1 -3
  51. package/package.json +32 -32
@@ -4,34 +4,20 @@ export interface AmazonopensearchserviceBufferingHints {
4
4
  SizeInMBs?: number;
5
5
  }
6
6
  export declare namespace AmazonopensearchserviceBufferingHints {
7
- /**
8
- * @internal
9
- */
7
+
10
8
  const filterSensitiveLog: (obj: AmazonopensearchserviceBufferingHints) => any;
11
9
  }
12
- /**
13
- * <p>Describes the Amazon CloudWatch logging options for your delivery stream.</p>
14
- */
10
+
15
11
  export interface CloudWatchLoggingOptions {
16
- /**
17
- * <p>Enables or disables CloudWatch logging.</p>
18
- */
12
+
19
13
  Enabled?: boolean;
20
- /**
21
- * <p>The CloudWatch group name for logging. This value is required if CloudWatch logging
22
- * is enabled.</p>
23
- */
14
+
24
15
  LogGroupName?: string;
25
- /**
26
- * <p>The CloudWatch log stream name for logging. This value is required if CloudWatch
27
- * logging is enabled.</p>
28
- */
16
+
29
17
  LogStreamName?: string;
30
18
  }
31
19
  export declare namespace CloudWatchLoggingOptions {
32
- /**
33
- * @internal
34
- */
20
+
35
21
  const filterSensitiveLog: (obj: CloudWatchLoggingOptions) => any;
36
22
  }
37
23
  export declare enum AmazonopensearchserviceIndexRotationPeriod {
@@ -52,107 +38,60 @@ export declare enum ProcessorParameterName {
52
38
  ROLE_ARN = "RoleArn",
53
39
  SUB_RECORD_TYPE = "SubRecordType"
54
40
  }
55
- /**
56
- * <p>Describes the processor parameter.</p>
57
- */
41
+
58
42
  export interface ProcessorParameter {
59
- /**
60
- * <p>The name of the parameter.</p>
61
- */
43
+
62
44
  ParameterName: ProcessorParameterName | string | undefined;
63
- /**
64
- * <p>The parameter value.</p>
65
- */
45
+
66
46
  ParameterValue: string | undefined;
67
47
  }
68
48
  export declare namespace ProcessorParameter {
69
- /**
70
- * @internal
71
- */
49
+
72
50
  const filterSensitiveLog: (obj: ProcessorParameter) => any;
73
51
  }
74
52
  export declare type ProcessorType = "AppendDelimiterToRecord" | "Lambda" | "MetadataExtraction" | "RecordDeAggregation";
75
- /**
76
- * <p>Describes a data processor.</p>
77
- */
53
+
78
54
  export interface Processor {
79
- /**
80
- * <p>The type of processor.</p>
81
- */
55
+
82
56
  Type: ProcessorType | string | undefined;
83
- /**
84
- * <p>The processor parameters.</p>
85
- */
57
+
86
58
  Parameters?: ProcessorParameter[];
87
59
  }
88
60
  export declare namespace Processor {
89
- /**
90
- * @internal
91
- */
61
+
92
62
  const filterSensitiveLog: (obj: Processor) => any;
93
63
  }
94
- /**
95
- * <p>Describes a data processing configuration.</p>
96
- */
64
+
97
65
  export interface ProcessingConfiguration {
98
- /**
99
- * <p>Enables or disables data processing.</p>
100
- */
66
+
101
67
  Enabled?: boolean;
102
- /**
103
- * <p>The data processors.</p>
104
- */
68
+
105
69
  Processors?: Processor[];
106
70
  }
107
71
  export declare namespace ProcessingConfiguration {
108
- /**
109
- * @internal
110
- */
72
+
111
73
  const filterSensitiveLog: (obj: ProcessingConfiguration) => any;
112
74
  }
113
75
  export interface AmazonopensearchserviceRetryOptions {
114
76
  DurationInSeconds?: number;
115
77
  }
116
78
  export declare namespace AmazonopensearchserviceRetryOptions {
117
- /**
118
- * @internal
119
- */
79
+
120
80
  const filterSensitiveLog: (obj: AmazonopensearchserviceRetryOptions) => any;
121
81
  }
122
82
  export declare enum AmazonopensearchserviceS3BackupMode {
123
83
  AllDocuments = "AllDocuments",
124
84
  FailedDocumentsOnly = "FailedDocumentsOnly"
125
85
  }
126
- /**
127
- * <p>Describes hints for the buffering to perform before delivering data to the
128
- * destination. These options are treated as hints, and therefore Kinesis Data Firehose might
129
- * choose to use different values when it is optimal. The <code>SizeInMBs</code> and
130
- * <code>IntervalInSeconds</code> parameters are optional. However, if specify a value for
131
- * one of them, you must also provide a value for the other.</p>
132
- */
86
+
133
87
  export interface BufferingHints {
134
- /**
135
- * <p>Buffer incoming data to the specified size, in MiBs, before delivering it to the
136
- * destination. The default value is 5. This parameter is optional but if you specify a value
137
- * for it, you must also specify a value for <code>IntervalInSeconds</code>, and vice
138
- * versa.</p>
139
- * <p>We recommend setting this parameter to a value greater than the amount of data you
140
- * typically ingest into the delivery stream in 10 seconds. For example, if you typically
141
- * ingest data at 1 MiB/sec, the value should be 10 MiB or higher.</p>
142
- */
88
+
143
89
  SizeInMBs?: number;
144
- /**
145
- * <p>Buffer incoming data for the specified period of time, in seconds, before delivering
146
- * it to the destination. The default value is 300. This parameter is optional but if you
147
- * specify a value for it, you must also specify a value for <code>SizeInMBs</code>, and vice
148
- * versa.</p>
149
- */
90
+
150
91
  IntervalInSeconds?: number;
151
92
  }
152
93
  export declare namespace BufferingHints {
153
- /**
154
- * @internal
155
- */
94
+
156
95
  const filterSensitiveLog: (obj: BufferingHints) => any;
157
96
  }
158
97
  export declare enum CompressionFormat {
@@ -162,190 +101,61 @@ export declare enum CompressionFormat {
162
101
  UNCOMPRESSED = "UNCOMPRESSED",
163
102
  ZIP = "ZIP"
164
103
  }
165
- /**
166
- * <p>Describes an encryption key for a destination in Amazon S3.</p>
167
- */
104
+
168
105
  export interface KMSEncryptionConfig {
169
- /**
170
- * <p>The Amazon Resource Name (ARN) of the encryption key. Must belong to the same AWS
171
- * Region as the destination Amazon S3 bucket. For more information, see <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon
172
- * Resource Names (ARNs) and AWS Service Namespaces</a>.</p>
173
- */
106
+
174
107
  AWSKMSKeyARN: string | undefined;
175
108
  }
176
109
  export declare namespace KMSEncryptionConfig {
177
- /**
178
- * @internal
179
- */
110
+
180
111
  const filterSensitiveLog: (obj: KMSEncryptionConfig) => any;
181
112
  }
182
113
  export declare type NoEncryptionConfig = "NoEncryption";
183
- /**
184
- * <p>Describes the encryption for a destination in Amazon S3.</p>
185
- */
114
+
186
115
  export interface EncryptionConfiguration {
187
- /**
188
- * <p>Specifically override existing encryption information to ensure that no encryption is
189
- * used.</p>
190
- */
116
+
191
117
  NoEncryptionConfig?: NoEncryptionConfig | string;
192
- /**
193
- * <p>The encryption key.</p>
194
- */
118
+
195
119
  KMSEncryptionConfig?: KMSEncryptionConfig;
196
120
  }
197
121
  export declare namespace EncryptionConfiguration {
198
- /**
199
- * @internal
200
- */
122
+
201
123
  const filterSensitiveLog: (obj: EncryptionConfiguration) => any;
202
124
  }
203
- /**
204
- * <p>Describes the configuration of a destination in Amazon S3.</p>
205
- */
125
+
206
126
  export interface S3DestinationConfiguration {
207
- /**
208
- * <p>The Amazon Resource Name (ARN) of the AWS credentials. For more information, see
209
- * <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon
210
- * Resource Names (ARNs) and AWS Service Namespaces</a>.</p>
211
- */
127
+
212
128
  RoleARN: string | undefined;
213
- /**
214
- * <p>The ARN of the S3 bucket. For more information, see <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon Resource Names (ARNs) and
215
- * AWS Service Namespaces</a>.</p>
216
- */
129
+
217
130
  BucketARN: string | undefined;
218
- /**
219
- * <p>The "YYYY/MM/DD/HH" time format prefix is automatically used for delivered Amazon S3
220
- * files. You can also specify a custom prefix, as described in <a href="https://docs.aws.amazon.com/firehose/latest/dev/s3-prefixes.html">Custom Prefixes for Amazon S3
221
- * Objects</a>.</p>
222
- */
131
+
223
132
  Prefix?: string;
224
- /**
225
- * <p>A prefix that Kinesis Data Firehose evaluates and adds to failed records before writing
226
- * them to S3. This prefix appears immediately following the bucket name. For information
227
- * about how to specify this prefix, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/s3-prefixes.html">Custom Prefixes for Amazon S3
228
- * Objects</a>.</p>
229
- */
133
+
230
134
  ErrorOutputPrefix?: string;
231
- /**
232
- * <p>The buffering option. If no value is specified, <code>BufferingHints</code> object
233
- * default values are used.</p>
234
- */
135
+
235
136
  BufferingHints?: BufferingHints;
236
- /**
237
- * <p>The compression format. If no value is specified, the default is
238
- * <code>UNCOMPRESSED</code>.</p>
239
- * <p>The compression formats <code>SNAPPY</code> or <code>ZIP</code> cannot be specified
240
- * for Amazon Redshift destinations because they are not supported by the Amazon Redshift
241
- * <code>COPY</code> operation that reads from the S3 bucket.</p>
242
- */
137
+
243
138
  CompressionFormat?: CompressionFormat | string;
244
- /**
245
- * <p>The encryption configuration. If no value is specified, the default is no
246
- * encryption.</p>
247
- */
139
+
248
140
  EncryptionConfiguration?: EncryptionConfiguration;
249
- /**
250
- * <p>The CloudWatch logging options for your delivery stream.</p>
251
- */
141
+
252
142
  CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
253
143
  }
254
144
  export declare namespace S3DestinationConfiguration {
255
- /**
256
- * @internal
257
- */
145
+
258
146
  const filterSensitiveLog: (obj: S3DestinationConfiguration) => any;
259
147
  }
260
- /**
261
- * <p>The details of the VPC of the Amazon ES destination.</p>
262
- */
148
+
263
149
  export interface VpcConfiguration {
264
- /**
265
- * <p>The IDs of the subnets that you want Kinesis Data Firehose to use to create ENIs in the
266
- * VPC of the Amazon ES destination. Make sure that the routing tables and inbound and
267
- * outbound rules allow traffic to flow from the subnets whose IDs are specified here to the
268
- * subnets that have the destination Amazon ES endpoints. Kinesis Data Firehose creates at
269
- * least one ENI in each of the subnets that are specified here. Do not delete or modify these
270
- * ENIs.</p>
271
- * <p>The number of ENIs that Kinesis Data Firehose creates in the subnets specified here
272
- * scales up and down automatically based on throughput. To enable Kinesis Data Firehose to
273
- * scale up the number of ENIs to match throughput, ensure that you have sufficient quota. To
274
- * help you calculate the quota you need, assume that Kinesis Data Firehose can create up to
275
- * three ENIs for this delivery stream for each of the subnets specified here. For more
276
- * information about ENI quota, see <a href="https://docs.aws.amazon.com/vpc/latest/userguide/amazon-vpc-limits.html#vpc-limits-enis">Network Interfaces
277
- * </a> in the Amazon VPC Quotas topic.</p>
278
- */
150
+
279
151
  SubnetIds: string[] | undefined;
280
- /**
281
- * <p>The ARN of the IAM role that you want the delivery stream to use to create endpoints in
282
- * the destination VPC. You can use your existing Kinesis Data Firehose delivery role or you
283
- * can specify a new role. In either case, make sure that the role trusts the Kinesis Data
284
- * Firehose service principal and that it grants the following permissions:</p>
285
- * <ul>
286
- * <li>
287
- * <p>
288
- * <code>ec2:DescribeVpcs</code>
289
- * </p>
290
- * </li>
291
- * <li>
292
- * <p>
293
- * <code>ec2:DescribeVpcAttribute</code>
294
- * </p>
295
- * </li>
296
- * <li>
297
- * <p>
298
- * <code>ec2:DescribeSubnets</code>
299
- * </p>
300
- * </li>
301
- * <li>
302
- * <p>
303
- * <code>ec2:DescribeSecurityGroups</code>
304
- * </p>
305
- * </li>
306
- * <li>
307
- * <p>
308
- * <code>ec2:DescribeNetworkInterfaces</code>
309
- * </p>
310
- * </li>
311
- * <li>
312
- * <p>
313
- * <code>ec2:CreateNetworkInterface</code>
314
- * </p>
315
- * </li>
316
- * <li>
317
- * <p>
318
- * <code>ec2:CreateNetworkInterfacePermission</code>
319
- * </p>
320
- * </li>
321
- * <li>
322
- * <p>
323
- * <code>ec2:DeleteNetworkInterface</code>
324
- * </p>
325
- * </li>
326
- * </ul>
327
- * <p>If you revoke these permissions after you create the delivery stream, Kinesis Data
328
- * Firehose can't scale out by creating more ENIs when necessary. You might therefore see a
329
- * degradation in performance.</p>
330
- */
152
+
331
153
  RoleARN: string | undefined;
332
- /**
333
- * <p>The IDs of the security groups that you want Kinesis Data Firehose to use when it
334
- * creates ENIs in the VPC of the Amazon ES destination. You can use the same security group
335
- * that the Amazon ES domain uses or different ones. If you specify different security groups
336
- * here, ensure that they allow outbound HTTPS traffic to the Amazon ES domain's security
337
- * group. Also ensure that the Amazon ES domain's security group allows HTTPS traffic from the
338
- * security groups specified here. If you use the same security group for both your delivery
339
- * stream and the Amazon ES domain, make sure the security group inbound rule allows HTTPS
340
- * traffic. For more information about security group rules, see <a href="https://docs.aws.amazon.com/vpc/latest/userguide/VPC_SecurityGroups.html#SecurityGroupRules">Security group
341
- * rules</a> in the Amazon VPC documentation.</p>
342
- */
154
+
343
155
  SecurityGroupIds: string[] | undefined;
344
156
  }
345
157
  export declare namespace VpcConfiguration {
346
- /**
347
- * @internal
348
- */
158
+
349
159
  const filterSensitiveLog: (obj: VpcConfiguration) => any;
350
160
  }
351
161
  export interface AmazonopensearchserviceDestinationConfiguration {
@@ -358,175 +168,55 @@ export interface AmazonopensearchserviceDestinationConfiguration {
358
168
  BufferingHints?: AmazonopensearchserviceBufferingHints;
359
169
  RetryOptions?: AmazonopensearchserviceRetryOptions;
360
170
  S3BackupMode?: AmazonopensearchserviceS3BackupMode | string;
361
- /**
362
- * <p>Describes the configuration of a destination in Amazon S3.</p>
363
- */
171
+
364
172
  S3Configuration: S3DestinationConfiguration | undefined;
365
- /**
366
- * <p>Describes a data processing configuration.</p>
367
- */
173
+
368
174
  ProcessingConfiguration?: ProcessingConfiguration;
369
- /**
370
- * <p>Describes the Amazon CloudWatch logging options for your delivery stream.</p>
371
- */
175
+
372
176
  CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
373
- /**
374
- * <p>The details of the VPC of the Amazon ES destination.</p>
375
- */
177
+
376
178
  VpcConfiguration?: VpcConfiguration;
377
179
  }
378
180
  export declare namespace AmazonopensearchserviceDestinationConfiguration {
379
- /**
380
- * @internal
381
- */
181
+
382
182
  const filterSensitiveLog: (obj: AmazonopensearchserviceDestinationConfiguration) => any;
383
183
  }
384
- /**
385
- * <p>Describes a destination in Amazon S3.</p>
386
- */
184
+
387
185
  export interface S3DestinationDescription {
388
- /**
389
- * <p>The Amazon Resource Name (ARN) of the AWS credentials. For more information, see
390
- * <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon
391
- * Resource Names (ARNs) and AWS Service Namespaces</a>.</p>
392
- */
186
+
393
187
  RoleARN: string | undefined;
394
- /**
395
- * <p>The ARN of the S3 bucket. For more information, see <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon Resource Names (ARNs) and
396
- * AWS Service Namespaces</a>.</p>
397
- */
188
+
398
189
  BucketARN: string | undefined;
399
- /**
400
- * <p>The "YYYY/MM/DD/HH" time format prefix is automatically used for delivered Amazon S3
401
- * files. You can also specify a custom prefix, as described in <a href="https://docs.aws.amazon.com/firehose/latest/dev/s3-prefixes.html">Custom Prefixes for Amazon S3
402
- * Objects</a>.</p>
403
- */
190
+
404
191
  Prefix?: string;
405
- /**
406
- * <p>A prefix that Kinesis Data Firehose evaluates and adds to failed records before writing
407
- * them to S3. This prefix appears immediately following the bucket name. For information
408
- * about how to specify this prefix, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/s3-prefixes.html">Custom Prefixes for Amazon S3
409
- * Objects</a>.</p>
410
- */
192
+
411
193
  ErrorOutputPrefix?: string;
412
- /**
413
- * <p>The buffering option. If no value is specified, <code>BufferingHints</code> object
414
- * default values are used.</p>
415
- */
194
+
416
195
  BufferingHints: BufferingHints | undefined;
417
- /**
418
- * <p>The compression format. If no value is specified, the default is
419
- * <code>UNCOMPRESSED</code>.</p>
420
- */
196
+
421
197
  CompressionFormat: CompressionFormat | string | undefined;
422
- /**
423
- * <p>The encryption configuration. If no value is specified, the default is no
424
- * encryption.</p>
425
- */
198
+
426
199
  EncryptionConfiguration: EncryptionConfiguration | undefined;
427
- /**
428
- * <p>The Amazon CloudWatch logging options for your delivery stream.</p>
429
- */
200
+
430
201
  CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
431
202
  }
432
203
  export declare namespace S3DestinationDescription {
433
- /**
434
- * @internal
435
- */
204
+
436
205
  const filterSensitiveLog: (obj: S3DestinationDescription) => any;
437
206
  }
438
- /**
439
- * <p>The details of the VPC of the Amazon ES destination.</p>
440
- */
207
+
441
208
  export interface VpcConfigurationDescription {
442
- /**
443
- * <p>The IDs of the subnets that Kinesis Data Firehose uses to create ENIs in the VPC of the
444
- * Amazon ES destination. Make sure that the routing tables and inbound and outbound rules
445
- * allow traffic to flow from the subnets whose IDs are specified here to the subnets that
446
- * have the destination Amazon ES endpoints. Kinesis Data Firehose creates at least one ENI in
447
- * each of the subnets that are specified here. Do not delete or modify these ENIs.</p>
448
- * <p>The number of ENIs that Kinesis Data Firehose creates in the subnets specified here
449
- * scales up and down automatically based on throughput. To enable Kinesis Data Firehose to
450
- * scale up the number of ENIs to match throughput, ensure that you have sufficient quota. To
451
- * help you calculate the quota you need, assume that Kinesis Data Firehose can create up to
452
- * three ENIs for this delivery stream for each of the subnets specified here. For more
453
- * information about ENI quota, see <a href="https://docs.aws.amazon.com/vpc/latest/userguide/amazon-vpc-limits.html#vpc-limits-enis">Network Interfaces
454
- * </a> in the Amazon VPC Quotas topic.</p>
455
- */
209
+
456
210
  SubnetIds: string[] | undefined;
457
- /**
458
- * <p>The ARN of the IAM role that the delivery stream uses to create endpoints in the
459
- * destination VPC. You can use your existing Kinesis Data Firehose delivery role or you can
460
- * specify a new role. In either case, make sure that the role trusts the Kinesis Data
461
- * Firehose service principal and that it grants the following permissions:</p>
462
- * <ul>
463
- * <li>
464
- * <p>
465
- * <code>ec2:DescribeVpcs</code>
466
- * </p>
467
- * </li>
468
- * <li>
469
- * <p>
470
- * <code>ec2:DescribeVpcAttribute</code>
471
- * </p>
472
- * </li>
473
- * <li>
474
- * <p>
475
- * <code>ec2:DescribeSubnets</code>
476
- * </p>
477
- * </li>
478
- * <li>
479
- * <p>
480
- * <code>ec2:DescribeSecurityGroups</code>
481
- * </p>
482
- * </li>
483
- * <li>
484
- * <p>
485
- * <code>ec2:DescribeNetworkInterfaces</code>
486
- * </p>
487
- * </li>
488
- * <li>
489
- * <p>
490
- * <code>ec2:CreateNetworkInterface</code>
491
- * </p>
492
- * </li>
493
- * <li>
494
- * <p>
495
- * <code>ec2:CreateNetworkInterfacePermission</code>
496
- * </p>
497
- * </li>
498
- * <li>
499
- * <p>
500
- * <code>ec2:DeleteNetworkInterface</code>
501
- * </p>
502
- * </li>
503
- * </ul>
504
- * <p>If you revoke these permissions after you create the delivery stream, Kinesis Data
505
- * Firehose can't scale out by creating more ENIs when necessary. You might therefore see a
506
- * degradation in performance.</p>
507
- */
211
+
508
212
  RoleARN: string | undefined;
509
- /**
510
- * <p>The IDs of the security groups that Kinesis Data Firehose uses when it creates ENIs in
511
- * the VPC of the Amazon ES destination. You can use the same security group that the Amazon
512
- * ES domain uses or different ones. If you specify different security groups, ensure that
513
- * they allow outbound HTTPS traffic to the Amazon ES domain's security group. Also ensure
514
- * that the Amazon ES domain's security group allows HTTPS traffic from the security groups
515
- * specified here. If you use the same security group for both your delivery stream and the
516
- * Amazon ES domain, make sure the security group inbound rule allows HTTPS traffic. For more
517
- * information about security group rules, see <a href="https://docs.aws.amazon.com/vpc/latest/userguide/VPC_SecurityGroups.html#SecurityGroupRules">Security group
518
- * rules</a> in the Amazon VPC documentation.</p>
519
- */
213
+
520
214
  SecurityGroupIds: string[] | undefined;
521
- /**
522
- * <p>The ID of the Amazon ES destination's VPC.</p>
523
- */
215
+
524
216
  VpcId: string | undefined;
525
217
  }
526
218
  export declare namespace VpcConfigurationDescription {
527
- /**
528
- * @internal
529
- */
219
+
530
220
  const filterSensitiveLog: (obj: VpcConfigurationDescription) => any;
531
221
  }
532
222
  export interface AmazonopensearchserviceDestinationDescription {
@@ -539,84 +229,40 @@ export interface AmazonopensearchserviceDestinationDescription {
539
229
  BufferingHints?: AmazonopensearchserviceBufferingHints;
540
230
  RetryOptions?: AmazonopensearchserviceRetryOptions;
541
231
  S3BackupMode?: AmazonopensearchserviceS3BackupMode | string;
542
- /**
543
- * <p>Describes a destination in Amazon S3.</p>
544
- */
232
+
545
233
  S3DestinationDescription?: S3DestinationDescription;
546
- /**
547
- * <p>Describes a data processing configuration.</p>
548
- */
234
+
549
235
  ProcessingConfiguration?: ProcessingConfiguration;
550
- /**
551
- * <p>Describes the Amazon CloudWatch logging options for your delivery stream.</p>
552
- */
236
+
553
237
  CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
554
- /**
555
- * <p>The details of the VPC of the Amazon ES destination.</p>
556
- */
238
+
557
239
  VpcConfigurationDescription?: VpcConfigurationDescription;
558
240
  }
559
241
  export declare namespace AmazonopensearchserviceDestinationDescription {
560
- /**
561
- * @internal
562
- */
242
+
563
243
  const filterSensitiveLog: (obj: AmazonopensearchserviceDestinationDescription) => any;
564
244
  }
565
- /**
566
- * <p>Describes an update for a destination in Amazon S3.</p>
567
- */
245
+
568
246
  export interface S3DestinationUpdate {
569
- /**
570
- * <p>The Amazon Resource Name (ARN) of the AWS credentials. For more information, see
571
- * <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon
572
- * Resource Names (ARNs) and AWS Service Namespaces</a>.</p>
573
- */
247
+
574
248
  RoleARN?: string;
575
- /**
576
- * <p>The ARN of the S3 bucket. For more information, see <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon Resource Names (ARNs) and
577
- * AWS Service Namespaces</a>.</p>
578
- */
249
+
579
250
  BucketARN?: string;
580
- /**
581
- * <p>The "YYYY/MM/DD/HH" time format prefix is automatically used for delivered Amazon S3
582
- * files. You can also specify a custom prefix, as described in <a href="https://docs.aws.amazon.com/firehose/latest/dev/s3-prefixes.html">Custom Prefixes for Amazon S3
583
- * Objects</a>.</p>
584
- */
251
+
585
252
  Prefix?: string;
586
- /**
587
- * <p>A prefix that Kinesis Data Firehose evaluates and adds to failed records before writing
588
- * them to S3. This prefix appears immediately following the bucket name. For information
589
- * about how to specify this prefix, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/s3-prefixes.html">Custom Prefixes for Amazon S3
590
- * Objects</a>.</p>
591
- */
253
+
592
254
  ErrorOutputPrefix?: string;
593
- /**
594
- * <p>The buffering option. If no value is specified, <code>BufferingHints</code> object
595
- * default values are used.</p>
596
- */
255
+
597
256
  BufferingHints?: BufferingHints;
598
- /**
599
- * <p>The compression format. If no value is specified, the default is
600
- * <code>UNCOMPRESSED</code>.</p>
601
- * <p>The compression formats <code>SNAPPY</code> or <code>ZIP</code> cannot be specified
602
- * for Amazon Redshift destinations because they are not supported by the Amazon Redshift
603
- * <code>COPY</code> operation that reads from the S3 bucket.</p>
604
- */
257
+
605
258
  CompressionFormat?: CompressionFormat | string;
606
- /**
607
- * <p>The encryption configuration. If no value is specified, the default is no
608
- * encryption.</p>
609
- */
259
+
610
260
  EncryptionConfiguration?: EncryptionConfiguration;
611
- /**
612
- * <p>The CloudWatch logging options for your delivery stream.</p>
613
- */
261
+
614
262
  CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
615
263
  }
616
264
  export declare namespace S3DestinationUpdate {
617
- /**
618
- * @internal
619
- */
265
+
620
266
  const filterSensitiveLog: (obj: S3DestinationUpdate) => any;
621
267
  }
622
268
  export interface AmazonopensearchserviceDestinationUpdate {
@@ -628,376 +274,158 @@ export interface AmazonopensearchserviceDestinationUpdate {
628
274
  IndexRotationPeriod?: AmazonopensearchserviceIndexRotationPeriod | string;
629
275
  BufferingHints?: AmazonopensearchserviceBufferingHints;
630
276
  RetryOptions?: AmazonopensearchserviceRetryOptions;
631
- /**
632
- * <p>Describes an update for a destination in Amazon S3.</p>
633
- */
277
+
634
278
  S3Update?: S3DestinationUpdate;
635
- /**
636
- * <p>Describes a data processing configuration.</p>
637
- */
279
+
638
280
  ProcessingConfiguration?: ProcessingConfiguration;
639
- /**
640
- * <p>Describes the Amazon CloudWatch logging options for your delivery stream.</p>
641
- */
281
+
642
282
  CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
643
283
  }
644
284
  export declare namespace AmazonopensearchserviceDestinationUpdate {
645
- /**
646
- * @internal
647
- */
285
+
648
286
  const filterSensitiveLog: (obj: AmazonopensearchserviceDestinationUpdate) => any;
649
287
  }
650
- /**
651
- * <p>Another modification has already happened. Fetch <code>VersionId</code> again and use
652
- * it to update the destination.</p>
653
- */
288
+
654
289
  export interface ConcurrentModificationException extends __SmithyException, $MetadataBearer {
655
290
  name: "ConcurrentModificationException";
656
291
  $fault: "client";
657
- /**
658
- * <p>A message that provides information about the error.</p>
659
- */
292
+
660
293
  message?: string;
661
294
  }
662
295
  export declare namespace ConcurrentModificationException {
663
- /**
664
- * @internal
665
- */
296
+
666
297
  const filterSensitiveLog: (obj: ConcurrentModificationException) => any;
667
298
  }
668
299
  export declare enum ContentEncoding {
669
300
  GZIP = "GZIP",
670
301
  NONE = "NONE"
671
302
  }
672
- /**
673
- * <p>Describes a <code>COPY</code> command for Amazon Redshift.</p>
674
- */
303
+
675
304
  export interface CopyCommand {
676
- /**
677
- * <p>The name of the target table. The table must already exist in the database.</p>
678
- */
305
+
679
306
  DataTableName: string | undefined;
680
- /**
681
- * <p>A comma-separated list of column names.</p>
682
- */
307
+
683
308
  DataTableColumns?: string;
684
- /**
685
- * <p>Optional parameters to use with the Amazon Redshift <code>COPY</code> command. For
686
- * more information, see the "Optional Parameters" section of <a href="https://docs.aws.amazon.com/redshift/latest/dg/r_COPY.html">Amazon Redshift COPY command</a>. Some possible
687
- * examples that would apply to Kinesis Data Firehose are as follows:</p>
688
- * <p>
689
- * <code>delimiter '\t' lzop;</code> - fields are delimited with "\t" (TAB character) and
690
- * compressed using lzop.</p>
691
- * <p>
692
- * <code>delimiter '|'</code> - fields are delimited with "|" (this is the default
693
- * delimiter).</p>
694
- * <p>
695
- * <code>delimiter '|' escape</code> - the delimiter should be escaped.</p>
696
- * <p>
697
- * <code>fixedwidth 'venueid:3,venuename:25,venuecity:12,venuestate:2,venueseats:6'</code> -
698
- * fields are fixed width in the source, with each width specified after every column in the
699
- * table.</p>
700
- * <p>
701
- * <code>JSON 's3://mybucket/jsonpaths.txt'</code> - data is in JSON format, and the path
702
- * specified is the format of the data.</p>
703
- * <p>For more examples, see <a href="https://docs.aws.amazon.com/redshift/latest/dg/r_COPY_command_examples.html">Amazon Redshift COPY command
704
- * examples</a>.</p>
705
- */
309
+
706
310
  CopyOptions?: string;
707
311
  }
708
312
  export declare namespace CopyCommand {
709
- /**
710
- * @internal
711
- */
313
+
712
314
  const filterSensitiveLog: (obj: CopyCommand) => any;
713
315
  }
714
316
  export declare enum KeyType {
715
317
  AWS_OWNED_CMK = "AWS_OWNED_CMK",
716
318
  CUSTOMER_MANAGED_CMK = "CUSTOMER_MANAGED_CMK"
717
319
  }
718
- /**
719
- * <p>Specifies the type and Amazon Resource Name (ARN) of the CMK to use for Server-Side
720
- * Encryption (SSE). </p>
721
- */
320
+
722
321
  export interface DeliveryStreamEncryptionConfigurationInput {
723
- /**
724
- * <p>If you set <code>KeyType</code> to <code>CUSTOMER_MANAGED_CMK</code>, you must specify
725
- * the Amazon Resource Name (ARN) of the CMK. If you set <code>KeyType</code> to
726
- * <code>AWS_OWNED_CMK</code>, Kinesis Data Firehose uses a service-account CMK.</p>
727
- */
322
+
728
323
  KeyARN?: string;
729
- /**
730
- * <p>Indicates the type of customer master key (CMK) to use for encryption. The default
731
- * setting is <code>AWS_OWNED_CMK</code>. For more information about CMKs, see <a href="https://docs.aws.amazon.com/kms/latest/developerguide/concepts.html#master_keys">Customer
732
- * Master Keys (CMKs)</a>. When you invoke <a>CreateDeliveryStream</a> or
733
- * <a>StartDeliveryStreamEncryption</a> with <code>KeyType</code> set to
734
- * CUSTOMER_MANAGED_CMK, Kinesis Data Firehose invokes the Amazon KMS operation <a href="https://docs.aws.amazon.com/kms/latest/APIReference/API_CreateGrant.html">CreateGrant</a> to create a grant that allows the Kinesis Data Firehose service to
735
- * use the customer managed CMK to perform encryption and decryption. Kinesis Data Firehose
736
- * manages that grant. </p>
737
- * <p>When you invoke <a>StartDeliveryStreamEncryption</a> to change the CMK for a
738
- * delivery stream that is encrypted with a customer managed CMK, Kinesis Data Firehose
739
- * schedules the grant it had on the old CMK for retirement.</p>
740
- * <p>You can use a CMK of type CUSTOMER_MANAGED_CMK to encrypt up to 500 delivery streams. If
741
- * a <a>CreateDeliveryStream</a> or <a>StartDeliveryStreamEncryption</a>
742
- * operation exceeds this limit, Kinesis Data Firehose throws a
743
- * <code>LimitExceededException</code>. </p>
744
- * <important>
745
- * <p>To encrypt your delivery stream, use symmetric CMKs. Kinesis Data Firehose doesn't
746
- * support asymmetric CMKs. For information about symmetric and asymmetric CMKs, see <a href="https://docs.aws.amazon.com/kms/latest/developerguide/symm-asymm-concepts.html">About
747
- * Symmetric and Asymmetric CMKs</a> in the AWS Key Management Service developer
748
- * guide.</p>
749
- * </important>
750
- */
324
+
751
325
  KeyType: KeyType | string | undefined;
752
326
  }
753
327
  export declare namespace DeliveryStreamEncryptionConfigurationInput {
754
- /**
755
- * @internal
756
- */
328
+
757
329
  const filterSensitiveLog: (obj: DeliveryStreamEncryptionConfigurationInput) => any;
758
330
  }
759
331
  export declare type DeliveryStreamType = "DirectPut" | "KinesisStreamAsSource";
760
- /**
761
- * <p>Describes the buffering to perform before delivering data to the Amazon ES
762
- * destination.</p>
763
- */
332
+
764
333
  export interface ElasticsearchBufferingHints {
765
- /**
766
- * <p>Buffer incoming data for the specified period of time, in seconds, before delivering
767
- * it to the destination. The default value is 300 (5 minutes).</p>
768
- */
334
+
769
335
  IntervalInSeconds?: number;
770
- /**
771
- * <p>Buffer incoming data to the specified size, in MBs, before delivering it to the
772
- * destination. The default value is 5.</p>
773
- * <p>We recommend setting this parameter to a value greater than the amount of data you
774
- * typically ingest into the delivery stream in 10 seconds. For example, if you typically
775
- * ingest data at 1 MB/sec, the value should be 10 MB or higher.</p>
776
- */
336
+
777
337
  SizeInMBs?: number;
778
338
  }
779
339
  export declare namespace ElasticsearchBufferingHints {
780
- /**
781
- * @internal
782
- */
340
+
783
341
  const filterSensitiveLog: (obj: ElasticsearchBufferingHints) => any;
784
342
  }
785
343
  export declare type ElasticsearchIndexRotationPeriod = "NoRotation" | "OneDay" | "OneHour" | "OneMonth" | "OneWeek";
786
- /**
787
- * <p>Configures retry behavior in case Kinesis Data Firehose is unable to deliver
788
- * documents to Amazon ES.</p>
789
- */
344
+
790
345
  export interface ElasticsearchRetryOptions {
791
- /**
792
- * <p>After an initial failure to deliver to Amazon ES, the total amount of time during
793
- * which Kinesis Data Firehose retries delivery (including the first attempt). After this time
794
- * has elapsed, the failed documents are written to Amazon S3. Default value is 300 seconds (5
795
- * minutes). A value of 0 (zero) results in no retries.</p>
796
- */
346
+
797
347
  DurationInSeconds?: number;
798
348
  }
799
349
  export declare namespace ElasticsearchRetryOptions {
800
- /**
801
- * @internal
802
- */
350
+
803
351
  const filterSensitiveLog: (obj: ElasticsearchRetryOptions) => any;
804
352
  }
805
353
  export declare type ElasticsearchS3BackupMode = "AllDocuments" | "FailedDocumentsOnly";
806
- /**
807
- * <p>Describes the configuration of a destination in Amazon ES.</p>
808
- */
354
+
809
355
  export interface ElasticsearchDestinationConfiguration {
810
- /**
811
- * <p>The Amazon Resource Name (ARN) of the IAM role to be assumed by Kinesis Data Firehose
812
- * for calling the Amazon ES Configuration API and for indexing documents. For more
813
- * information, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/controlling-access.html#using-iam-s3">Grant Kinesis Data
814
- * Firehose Access to an Amazon S3 Destination</a> and <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon Resource Names (ARNs) and
815
- * AWS Service Namespaces</a>.</p>
816
- */
356
+
817
357
  RoleARN: string | undefined;
818
- /**
819
- * <p>The ARN of the Amazon ES domain. The IAM role must have permissions
820
- * for <code>DescribeElasticsearchDomain</code>, <code>DescribeElasticsearchDomains</code>,
821
- * and <code>DescribeElasticsearchDomainConfig</code> after assuming the role specified in
822
- * <b>RoleARN</b>. For more information, see <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon
823
- * Resource Names (ARNs) and AWS Service Namespaces</a>.</p>
824
- *
825
- * <p>Specify either <code>ClusterEndpoint</code> or <code>DomainARN</code>.</p>
826
- */
358
+
827
359
  DomainARN?: string;
828
- /**
829
- * <p>The endpoint to use when communicating with the cluster. Specify either this
830
- * <code>ClusterEndpoint</code> or the <code>DomainARN</code> field.</p>
831
- */
360
+
832
361
  ClusterEndpoint?: string;
833
- /**
834
- * <p>The Elasticsearch index name.</p>
835
- */
362
+
836
363
  IndexName: string | undefined;
837
- /**
838
- * <p>The Elasticsearch type name. For Elasticsearch 6.x, there can be only one type per
839
- * index. If you try to specify a new type for an existing index that already has another
840
- * type, Kinesis Data Firehose returns an error during run time.</p>
841
- *
842
- * <p>For Elasticsearch 7.x, don't specify a <code>TypeName</code>.</p>
843
- */
364
+
844
365
  TypeName?: string;
845
- /**
846
- * <p>The Elasticsearch index rotation period. Index rotation appends a timestamp to the
847
- * <code>IndexName</code> to facilitate the expiration of old data. For more information,
848
- * see <a href="https://docs.aws.amazon.com/firehose/latest/dev/basic-deliver.html#es-index-rotation">Index Rotation for the
849
- * Amazon ES Destination</a>. The default value is <code>OneDay</code>.</p>
850
- */
366
+
851
367
  IndexRotationPeriod?: ElasticsearchIndexRotationPeriod | string;
852
- /**
853
- * <p>The buffering options. If no value is specified, the default values for
854
- * <code>ElasticsearchBufferingHints</code> are used.</p>
855
- */
368
+
856
369
  BufferingHints?: ElasticsearchBufferingHints;
857
- /**
858
- * <p>The retry behavior in case Kinesis Data Firehose is unable to deliver documents to
859
- * Amazon ES. The default value is 300 (5 minutes).</p>
860
- */
370
+
861
371
  RetryOptions?: ElasticsearchRetryOptions;
862
- /**
863
- * <p>Defines how documents should be delivered to Amazon S3. When it is set to
864
- * <code>FailedDocumentsOnly</code>, Kinesis Data Firehose writes any documents that could
865
- * not be indexed to the configured Amazon S3 destination, with
866
- * <code>elasticsearch-failed/</code> appended to the key prefix. When set to
867
- * <code>AllDocuments</code>, Kinesis Data Firehose delivers all incoming records to Amazon
868
- * S3, and also writes failed documents with <code>elasticsearch-failed/</code> appended to
869
- * the prefix. For more information, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/basic-deliver.html#es-s3-backup">Amazon S3 Backup for the
870
- * Amazon ES Destination</a>. Default value is
871
- * <code>FailedDocumentsOnly</code>.</p>
872
- * <p>You can't change this backup mode after you create the delivery stream. </p>
873
- */
372
+
874
373
  S3BackupMode?: ElasticsearchS3BackupMode | string;
875
- /**
876
- * <p>The configuration for the backup Amazon S3 location.</p>
877
- */
374
+
878
375
  S3Configuration: S3DestinationConfiguration | undefined;
879
- /**
880
- * <p>The data processing configuration.</p>
881
- */
376
+
882
377
  ProcessingConfiguration?: ProcessingConfiguration;
883
- /**
884
- * <p>The Amazon CloudWatch logging options for your delivery stream.</p>
885
- */
378
+
886
379
  CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
887
- /**
888
- * <p>The details of the VPC of the Amazon ES destination.</p>
889
- */
380
+
890
381
  VpcConfiguration?: VpcConfiguration;
891
382
  }
892
383
  export declare namespace ElasticsearchDestinationConfiguration {
893
- /**
894
- * @internal
895
- */
384
+
896
385
  const filterSensitiveLog: (obj: ElasticsearchDestinationConfiguration) => any;
897
386
  }
898
- /**
899
- * <p>The native Hive / HCatalog JsonSerDe. Used by Kinesis Data Firehose for deserializing
900
- * data, which means converting it from the JSON format in preparation for serializing it to
901
- * the Parquet or ORC format. This is one of two deserializers you can choose, depending on
902
- * which one offers the functionality you need. The other option is the OpenX SerDe.</p>
903
- */
387
+
904
388
  export interface HiveJsonSerDe {
905
- /**
906
- * <p>Indicates how you want Kinesis Data Firehose to parse the date and timestamps that
907
- * may be present in your input data JSON. To specify these format strings, follow the pattern
908
- * syntax of JodaTime's DateTimeFormat format strings. For more information, see <a href="https://www.joda.org/joda-time/apidocs/org/joda/time/format/DateTimeFormat.html">Class DateTimeFormat</a>. You can also use the special value <code>millis</code> to
909
- * parse timestamps in epoch milliseconds. If you don't specify a format, Kinesis Data
910
- * Firehose uses <code>java.sql.Timestamp::valueOf</code> by default.</p>
911
- */
389
+
912
390
  TimestampFormats?: string[];
913
391
  }
914
392
  export declare namespace HiveJsonSerDe {
915
- /**
916
- * @internal
917
- */
393
+
918
394
  const filterSensitiveLog: (obj: HiveJsonSerDe) => any;
919
395
  }
920
- /**
921
- * <p>The OpenX SerDe. Used by Kinesis Data Firehose for deserializing data, which means
922
- * converting it from the JSON format in preparation for serializing it to the Parquet or ORC
923
- * format. This is one of two deserializers you can choose, depending on which one offers the
924
- * functionality you need. The other option is the native Hive / HCatalog JsonSerDe.</p>
925
- */
396
+
926
397
  export interface OpenXJsonSerDe {
927
- /**
928
- * <p>When set to <code>true</code>, specifies that the names of the keys include dots and
929
- * that you want Kinesis Data Firehose to replace them with underscores. This is useful
930
- * because Apache Hive does not allow dots in column names. For example, if the JSON contains
931
- * a key whose name is "a.b", you can define the column name to be "a_b" when using this
932
- * option.</p>
933
- * <p>The default is <code>false</code>.</p>
934
- */
398
+
935
399
  ConvertDotsInJsonKeysToUnderscores?: boolean;
936
- /**
937
- * <p>When set to <code>true</code>, which is the default, Kinesis Data Firehose converts
938
- * JSON keys to lowercase before deserializing them.</p>
939
- */
400
+
940
401
  CaseInsensitive?: boolean;
941
- /**
942
- * <p>Maps column names to JSON keys that aren't identical to the column names. This is
943
- * useful when the JSON contains keys that are Hive keywords. For example,
944
- * <code>timestamp</code> is a Hive keyword. If you have a JSON key named
945
- * <code>timestamp</code>, set this parameter to <code>{"ts": "timestamp"}</code> to map
946
- * this key to a column named <code>ts</code>.</p>
947
- */
402
+
948
403
  ColumnToJsonKeyMappings?: {
949
404
  [key: string]: string;
950
405
  };
951
406
  }
952
407
  export declare namespace OpenXJsonSerDe {
953
- /**
954
- * @internal
955
- */
408
+
956
409
  const filterSensitiveLog: (obj: OpenXJsonSerDe) => any;
957
410
  }
958
- /**
959
- * <p>The deserializer you want Kinesis Data Firehose to use for converting the input data
960
- * from JSON. Kinesis Data Firehose then serializes the data to its final format using the
961
- * <a>Serializer</a>. Kinesis Data Firehose supports two types of deserializers:
962
- * the <a href="https://cwiki.apache.org/confluence/display/Hive/LanguageManual+DDL#LanguageManualDDL-JSON">Apache Hive JSON SerDe</a> and the <a href="https://github.com/rcongiu/Hive-JSON-Serde">OpenX JSON SerDe</a>.</p>
963
- */
411
+
964
412
  export interface Deserializer {
965
- /**
966
- * <p>The OpenX SerDe. Used by Kinesis Data Firehose for deserializing data, which means
967
- * converting it from the JSON format in preparation for serializing it to the Parquet or ORC
968
- * format. This is one of two deserializers you can choose, depending on which one offers the
969
- * functionality you need. The other option is the native Hive / HCatalog JsonSerDe.</p>
970
- */
413
+
971
414
  OpenXJsonSerDe?: OpenXJsonSerDe;
972
- /**
973
- * <p>The native Hive / HCatalog JsonSerDe. Used by Kinesis Data Firehose for deserializing
974
- * data, which means converting it from the JSON format in preparation for serializing it to
975
- * the Parquet or ORC format. This is one of two deserializers you can choose, depending on
976
- * which one offers the functionality you need. The other option is the OpenX SerDe.</p>
977
- */
415
+
978
416
  HiveJsonSerDe?: HiveJsonSerDe;
979
417
  }
980
418
  export declare namespace Deserializer {
981
- /**
982
- * @internal
983
- */
419
+
984
420
  const filterSensitiveLog: (obj: Deserializer) => any;
985
421
  }
986
- /**
987
- * <p>Specifies the deserializer you want to use to convert the format of the input data.
988
- * This parameter is required if <code>Enabled</code> is set to true.</p>
989
- */
422
+
990
423
  export interface InputFormatConfiguration {
991
- /**
992
- * <p>Specifies which deserializer to use. You can choose either the Apache Hive JSON SerDe
993
- * or the OpenX JSON SerDe. If both are non-null, the server rejects the request.</p>
994
- */
424
+
995
425
  Deserializer?: Deserializer;
996
426
  }
997
427
  export declare namespace InputFormatConfiguration {
998
- /**
999
- * @internal
1000
- */
428
+
1001
429
  const filterSensitiveLog: (obj: InputFormatConfiguration) => any;
1002
430
  }
1003
431
  export declare enum OrcCompression {
@@ -1009,76 +437,31 @@ export declare enum OrcFormatVersion {
1009
437
  V0_11 = "V0_11",
1010
438
  V0_12 = "V0_12"
1011
439
  }
1012
- /**
1013
- * <p>A serializer to use for converting data to the ORC format before storing it in Amazon
1014
- * S3. For more information, see <a href="https://orc.apache.org/docs/">Apache
1015
- * ORC</a>.</p>
1016
- */
440
+
1017
441
  export interface OrcSerDe {
1018
- /**
1019
- * <p>The number of bytes in each stripe. The default is 64 MiB and the minimum is 8
1020
- * MiB.</p>
1021
- */
442
+
1022
443
  StripeSizeBytes?: number;
1023
- /**
1024
- * <p>The Hadoop Distributed File System (HDFS) block size. This is useful if you intend to
1025
- * copy the data from Amazon S3 to HDFS before querying. The default is 256 MiB and the
1026
- * minimum is 64 MiB. Kinesis Data Firehose uses this value for padding calculations.</p>
1027
- */
444
+
1028
445
  BlockSizeBytes?: number;
1029
- /**
1030
- * <p>The number of rows between index entries. The default is 10,000 and the minimum is
1031
- * 1,000.</p>
1032
- */
446
+
1033
447
  RowIndexStride?: number;
1034
- /**
1035
- * <p>Set this to <code>true</code> to indicate that you want stripes to be padded to the HDFS
1036
- * block boundaries. This is useful if you intend to copy the data from Amazon S3 to HDFS
1037
- * before querying. The default is <code>false</code>.</p>
1038
- */
448
+
1039
449
  EnablePadding?: boolean;
1040
- /**
1041
- * <p>A number between 0 and 1 that defines the tolerance for block padding as a decimal
1042
- * fraction of stripe size. The default value is 0.05, which means 5 percent of stripe
1043
- * size.</p>
1044
- * <p>For the default values of 64 MiB ORC stripes and 256 MiB HDFS blocks, the default block
1045
- * padding tolerance of 5 percent reserves a maximum of 3.2 MiB for padding within the 256 MiB
1046
- * block. In such a case, if the available size within the block is more than 3.2 MiB, a new,
1047
- * smaller stripe is inserted to fit within that space. This ensures that no stripe crosses
1048
- * block boundaries and causes remote reads within a node-local task.</p>
1049
- * <p>Kinesis Data Firehose ignores this parameter when <a>OrcSerDe$EnablePadding</a> is <code>false</code>.</p>
1050
- */
450
+
1051
451
  PaddingTolerance?: number;
1052
- /**
1053
- * <p>The compression code to use over data blocks. The default is <code>SNAPPY</code>.</p>
1054
- */
452
+
1055
453
  Compression?: OrcCompression | string;
1056
- /**
1057
- * <p>The column names for which you want Kinesis Data Firehose to create bloom filters. The
1058
- * default is <code>null</code>.</p>
1059
- */
454
+
1060
455
  BloomFilterColumns?: string[];
1061
- /**
1062
- * <p>The Bloom filter false positive probability (FPP). The lower the FPP, the bigger the
1063
- * Bloom filter. The default value is 0.05, the minimum is 0, and the maximum is 1.</p>
1064
- */
456
+
1065
457
  BloomFilterFalsePositiveProbability?: number;
1066
- /**
1067
- * <p>Represents the fraction of the total number of non-null rows. To turn off dictionary
1068
- * encoding, set this fraction to a number that is less than the number of distinct keys in a
1069
- * dictionary. To always use dictionary encoding, set this threshold to 1.</p>
1070
- */
458
+
1071
459
  DictionaryKeyThreshold?: number;
1072
- /**
1073
- * <p>The version of the file to write. The possible values are <code>V0_11</code> and
1074
- * <code>V0_12</code>. The default is <code>V0_12</code>.</p>
1075
- */
460
+
1076
461
  FormatVersion?: OrcFormatVersion | string;
1077
462
  }
1078
463
  export declare namespace OrcSerDe {
1079
- /**
1080
- * @internal
1081
- */
464
+
1082
465
  const filterSensitiveLog: (obj: OrcSerDe) => any;
1083
466
  }
1084
467
  export declare enum ParquetCompression {
@@ -1090,818 +473,360 @@ export declare enum ParquetWriterVersion {
1090
473
  V1 = "V1",
1091
474
  V2 = "V2"
1092
475
  }
1093
- /**
1094
- * <p>A serializer to use for converting data to the Parquet format before storing it in
1095
- * Amazon S3. For more information, see <a href="https://parquet.apache.org/documentation/latest/">Apache Parquet</a>.</p>
1096
- */
476
+
1097
477
  export interface ParquetSerDe {
1098
- /**
1099
- * <p>The Hadoop Distributed File System (HDFS) block size. This is useful if you intend to
1100
- * copy the data from Amazon S3 to HDFS before querying. The default is 256 MiB and the
1101
- * minimum is 64 MiB. Kinesis Data Firehose uses this value for padding calculations.</p>
1102
- */
478
+
1103
479
  BlockSizeBytes?: number;
1104
- /**
1105
- * <p>The Parquet page size. Column chunks are divided into pages. A page is conceptually an
1106
- * indivisible unit (in terms of compression and encoding). The minimum value is 64 KiB and
1107
- * the default is 1 MiB.</p>
1108
- */
480
+
1109
481
  PageSizeBytes?: number;
1110
- /**
1111
- * <p>The compression code to use over data blocks. The possible values are
1112
- * <code>UNCOMPRESSED</code>, <code>SNAPPY</code>, and <code>GZIP</code>, with the default
1113
- * being <code>SNAPPY</code>. Use <code>SNAPPY</code> for higher decompression speed. Use
1114
- * <code>GZIP</code> if the compression ratio is more important than speed.</p>
1115
- */
482
+
1116
483
  Compression?: ParquetCompression | string;
1117
- /**
1118
- * <p>Indicates whether to enable dictionary compression.</p>
1119
- */
484
+
1120
485
  EnableDictionaryCompression?: boolean;
1121
- /**
1122
- * <p>The maximum amount of padding to apply. This is useful if you intend to copy the data
1123
- * from Amazon S3 to HDFS before querying. The default is 0.</p>
1124
- */
486
+
1125
487
  MaxPaddingBytes?: number;
1126
- /**
1127
- * <p>Indicates the version of row format to output. The possible values are <code>V1</code>
1128
- * and <code>V2</code>. The default is <code>V1</code>.</p>
1129
- */
488
+
1130
489
  WriterVersion?: ParquetWriterVersion | string;
1131
490
  }
1132
491
  export declare namespace ParquetSerDe {
1133
- /**
1134
- * @internal
1135
- */
492
+
1136
493
  const filterSensitiveLog: (obj: ParquetSerDe) => any;
1137
494
  }
1138
- /**
1139
- * <p>The serializer that you want Kinesis Data Firehose to use to convert data to the target
1140
- * format before writing it to Amazon S3. Kinesis Data Firehose supports two types of
1141
- * serializers: the <a href="https://hive.apache.org/javadocs/r1.2.2/api/org/apache/hadoop/hive/ql/io/orc/OrcSerde.html">ORC SerDe</a> and the <a href="https://hive.apache.org/javadocs/r1.2.2/api/org/apache/hadoop/hive/ql/io/parquet/serde/ParquetHiveSerDe.html">Parquet SerDe</a>.</p>
1142
- */
495
+
1143
496
  export interface Serializer {
1144
- /**
1145
- * <p>A serializer to use for converting data to the Parquet format before storing it in
1146
- * Amazon S3. For more information, see <a href="https://parquet.apache.org/documentation/latest/">Apache Parquet</a>.</p>
1147
- */
497
+
1148
498
  ParquetSerDe?: ParquetSerDe;
1149
- /**
1150
- * <p>A serializer to use for converting data to the ORC format before storing it in Amazon
1151
- * S3. For more information, see <a href="https://orc.apache.org/docs/">Apache
1152
- * ORC</a>.</p>
1153
- */
499
+
1154
500
  OrcSerDe?: OrcSerDe;
1155
501
  }
1156
502
  export declare namespace Serializer {
1157
- /**
1158
- * @internal
1159
- */
503
+
1160
504
  const filterSensitiveLog: (obj: Serializer) => any;
1161
505
  }
1162
- /**
1163
- * <p>Specifies the serializer that you want Kinesis Data Firehose to use to convert the
1164
- * format of your data before it writes it to Amazon S3. This parameter is required if
1165
- * <code>Enabled</code> is set to true.</p>
1166
- */
506
+
1167
507
  export interface OutputFormatConfiguration {
1168
- /**
1169
- * <p>Specifies which serializer to use. You can choose either the ORC SerDe or the Parquet
1170
- * SerDe. If both are non-null, the server rejects the request.</p>
1171
- */
508
+
1172
509
  Serializer?: Serializer;
1173
510
  }
1174
511
  export declare namespace OutputFormatConfiguration {
1175
- /**
1176
- * @internal
1177
- */
512
+
1178
513
  const filterSensitiveLog: (obj: OutputFormatConfiguration) => any;
1179
514
  }
1180
- /**
1181
- * <p>Specifies the schema to which you want Kinesis Data Firehose to configure your data
1182
- * before it writes it to Amazon S3. This parameter is required if <code>Enabled</code> is set
1183
- * to true.</p>
1184
- */
515
+
1185
516
  export interface SchemaConfiguration {
1186
- /**
1187
- * <p>The role that Kinesis Data Firehose can use to access AWS Glue. This role must be in
1188
- * the same account you use for Kinesis Data Firehose. Cross-account roles aren't
1189
- * allowed.</p>
1190
- * <important>
1191
- * <p>If the <code>SchemaConfiguration</code> request parameter is used as part of invoking
1192
- * the <code>CreateDeliveryStream</code> API, then the <code>RoleARN</code> property is
1193
- * required and its value must be specified.</p>
1194
- * </important>
1195
- */
517
+
1196
518
  RoleARN?: string;
1197
- /**
1198
- * <p>The ID of the AWS Glue Data Catalog. If you don't supply this, the AWS account ID is
1199
- * used by default.</p>
1200
- */
519
+
1201
520
  CatalogId?: string;
1202
- /**
1203
- * <p>Specifies the name of the AWS Glue database that contains the schema for the output
1204
- * data.</p>
1205
- * <important>
1206
- * <p>If the <code>SchemaConfiguration</code> request parameter is used as part of invoking
1207
- * the <code>CreateDeliveryStream</code> API, then the <code>DatabaseName</code> property
1208
- * is required and its value must be specified.</p>
1209
- * </important>
1210
- */
521
+
1211
522
  DatabaseName?: string;
1212
- /**
1213
- * <p>Specifies the AWS Glue table that contains the column information that constitutes your
1214
- * data schema.</p>
1215
- * <important>
1216
- * <p>If the <code>SchemaConfiguration</code> request parameter is used as part of invoking
1217
- * the <code>CreateDeliveryStream</code> API, then the <code>TableName</code> property is
1218
- * required and its value must be specified.</p>
1219
- * </important>
1220
- */
523
+
1221
524
  TableName?: string;
1222
- /**
1223
- * <p>If you don't specify an AWS Region, the default is the current Region.</p>
1224
- */
525
+
1225
526
  Region?: string;
1226
- /**
1227
- * <p>Specifies the table version for the output data schema. If you don't specify this
1228
- * version ID, or if you set it to <code>LATEST</code>, Kinesis Data Firehose uses the most
1229
- * recent version. This means that any updates to the table are automatically picked
1230
- * up.</p>
1231
- */
527
+
1232
528
  VersionId?: string;
1233
529
  }
1234
530
  export declare namespace SchemaConfiguration {
1235
- /**
1236
- * @internal
1237
- */
531
+
1238
532
  const filterSensitiveLog: (obj: SchemaConfiguration) => any;
1239
533
  }
1240
- /**
1241
- * <p>Specifies that you want Kinesis Data Firehose to convert data from the JSON format to
1242
- * the Parquet or ORC format before writing it to Amazon S3. Kinesis Data Firehose uses the
1243
- * serializer and deserializer that you specify, in addition to the column information from
1244
- * the AWS Glue table, to deserialize your input data from JSON and then serialize it to the
1245
- * Parquet or ORC format. For more information, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/record-format-conversion.html">Kinesis Data Firehose Record
1246
- * Format Conversion</a>.</p>
1247
- */
534
+
1248
535
  export interface DataFormatConversionConfiguration {
1249
- /**
1250
- * <p>Specifies the AWS Glue Data Catalog table that contains the column information. This
1251
- * parameter is required if <code>Enabled</code> is set to true.</p>
1252
- */
536
+
1253
537
  SchemaConfiguration?: SchemaConfiguration;
1254
- /**
1255
- * <p>Specifies the deserializer that you want Kinesis Data Firehose to use to convert the
1256
- * format of your data from JSON. This parameter is required if <code>Enabled</code> is set to
1257
- * true.</p>
1258
- */
538
+
1259
539
  InputFormatConfiguration?: InputFormatConfiguration;
1260
- /**
1261
- * <p>Specifies the serializer that you want Kinesis Data Firehose to use to convert the
1262
- * format of your data to the Parquet or ORC format. This parameter is required if
1263
- * <code>Enabled</code> is set to true.</p>
1264
- */
540
+
1265
541
  OutputFormatConfiguration?: OutputFormatConfiguration;
1266
- /**
1267
- * <p>Defaults to <code>true</code>. Set it to <code>false</code> if you want to disable
1268
- * format conversion while preserving the configuration details.</p>
1269
- */
542
+
1270
543
  Enabled?: boolean;
1271
544
  }
1272
545
  export declare namespace DataFormatConversionConfiguration {
1273
- /**
1274
- * @internal
1275
- */
546
+
1276
547
  const filterSensitiveLog: (obj: DataFormatConversionConfiguration) => any;
1277
548
  }
1278
- /**
1279
- * <p> The retry behavior in case Kinesis Data Firehose is unable to deliver data to an Amazon
1280
- * S3 prefix.</p>
1281
- */
549
+
1282
550
  export interface RetryOptions {
1283
- /**
1284
- * <p>The period of time during which Kinesis Data Firehose retries to deliver data to the
1285
- * specified Amazon S3 prefix.</p>
1286
- */
551
+
1287
552
  DurationInSeconds?: number;
1288
553
  }
1289
554
  export declare namespace RetryOptions {
1290
- /**
1291
- * @internal
1292
- */
555
+
1293
556
  const filterSensitiveLog: (obj: RetryOptions) => any;
1294
557
  }
1295
- /**
1296
- * <p>The configuration of the dynamic partitioning mechanism that creates smaller data sets
1297
- * from the streaming data by partitioning it based on partition keys. Currently, dynamic
1298
- * partitioning is only supported for Amazon S3 destinations. For more information, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/dynamic-partitioning.html">https://docs.aws.amazon.com/firehose/latest/dev/dynamic-partitioning.html</a>
1299
- * </p>
1300
- */
558
+
1301
559
  export interface DynamicPartitioningConfiguration {
1302
- /**
1303
- * <p>The retry behavior in case Kinesis Data Firehose is unable to deliver data to an Amazon
1304
- * S3 prefix.</p>
1305
- */
560
+
1306
561
  RetryOptions?: RetryOptions;
1307
- /**
1308
- * <p>Specifies that the dynamic partitioning is enabled for this Kinesis Data Firehose
1309
- * delivery stream.</p>
1310
- */
562
+
1311
563
  Enabled?: boolean;
1312
564
  }
1313
565
  export declare namespace DynamicPartitioningConfiguration {
1314
- /**
1315
- * @internal
1316
- */
566
+
1317
567
  const filterSensitiveLog: (obj: DynamicPartitioningConfiguration) => any;
1318
568
  }
1319
569
  export declare type S3BackupMode = "Disabled" | "Enabled";
1320
- /**
1321
- * <p>Describes the configuration of a destination in Amazon S3.</p>
1322
- */
570
+
1323
571
  export interface ExtendedS3DestinationConfiguration {
1324
- /**
1325
- * <p>The Amazon Resource Name (ARN) of the AWS credentials. For more information, see
1326
- * <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon
1327
- * Resource Names (ARNs) and AWS Service Namespaces</a>.</p>
1328
- */
572
+
1329
573
  RoleARN: string | undefined;
1330
- /**
1331
- * <p>The ARN of the S3 bucket. For more information, see <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon Resource Names (ARNs) and
1332
- * AWS Service Namespaces</a>.</p>
1333
- */
574
+
1334
575
  BucketARN: string | undefined;
1335
- /**
1336
- * <p>The "YYYY/MM/DD/HH" time format prefix is automatically used for delivered Amazon S3
1337
- * files. You can also specify a custom prefix, as described in <a href="https://docs.aws.amazon.com/firehose/latest/dev/s3-prefixes.html">Custom Prefixes for Amazon S3
1338
- * Objects</a>.</p>
1339
- */
576
+
1340
577
  Prefix?: string;
1341
- /**
1342
- * <p>A prefix that Kinesis Data Firehose evaluates and adds to failed records before writing
1343
- * them to S3. This prefix appears immediately following the bucket name. For information
1344
- * about how to specify this prefix, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/s3-prefixes.html">Custom Prefixes for Amazon S3
1345
- * Objects</a>.</p>
1346
- */
578
+
1347
579
  ErrorOutputPrefix?: string;
1348
- /**
1349
- * <p>The buffering option.</p>
1350
- */
580
+
1351
581
  BufferingHints?: BufferingHints;
1352
- /**
1353
- * <p>The compression format. If no value is specified, the default is
1354
- * UNCOMPRESSED.</p>
1355
- */
582
+
1356
583
  CompressionFormat?: CompressionFormat | string;
1357
- /**
1358
- * <p>The encryption configuration. If no value is specified, the default is no
1359
- * encryption.</p>
1360
- */
584
+
1361
585
  EncryptionConfiguration?: EncryptionConfiguration;
1362
- /**
1363
- * <p>The Amazon CloudWatch logging options for your delivery stream.</p>
1364
- */
586
+
1365
587
  CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
1366
- /**
1367
- * <p>The data processing configuration.</p>
1368
- */
588
+
1369
589
  ProcessingConfiguration?: ProcessingConfiguration;
1370
- /**
1371
- * <p>The Amazon S3 backup mode. After you create a delivery stream, you can update it to
1372
- * enable Amazon S3 backup if it is disabled. If backup is enabled, you can't update the
1373
- * delivery stream to disable it. </p>
1374
- */
590
+
1375
591
  S3BackupMode?: S3BackupMode | string;
1376
- /**
1377
- * <p>The configuration for backup in Amazon S3.</p>
1378
- */
592
+
1379
593
  S3BackupConfiguration?: S3DestinationConfiguration;
1380
- /**
1381
- * <p>The serializer, deserializer, and schema for converting data from the JSON format to
1382
- * the Parquet or ORC format before writing it to Amazon S3.</p>
1383
- */
594
+
1384
595
  DataFormatConversionConfiguration?: DataFormatConversionConfiguration;
1385
- /**
1386
- * <p>The configuration of the dynamic partitioning mechanism that creates smaller data sets
1387
- * from the streaming data by partitioning it based on partition keys. Currently, dynamic
1388
- * partitioning is only supported for Amazon S3 destinations. For more information, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/dynamic-partitioning.html">https://docs.aws.amazon.com/firehose/latest/dev/dynamic-partitioning.html</a>
1389
- * </p>
1390
- */
596
+
1391
597
  DynamicPartitioningConfiguration?: DynamicPartitioningConfiguration;
1392
598
  }
1393
599
  export declare namespace ExtendedS3DestinationConfiguration {
1394
- /**
1395
- * @internal
1396
- */
600
+
1397
601
  const filterSensitiveLog: (obj: ExtendedS3DestinationConfiguration) => any;
1398
602
  }
1399
- /**
1400
- * <p>Describes the buffering options that can be applied before data is delivered to the HTTP
1401
- * endpoint destination. Kinesis Data Firehose treats these options as hints, and it might
1402
- * choose to use more optimal values. The <code>SizeInMBs</code> and
1403
- * <code>IntervalInSeconds</code> parameters are optional. However, if specify a value for
1404
- * one of them, you must also provide a value for the other. </p>
1405
- */
603
+
1406
604
  export interface HttpEndpointBufferingHints {
1407
- /**
1408
- * <p>Buffer incoming data to the specified size, in MBs, before delivering it to the
1409
- * destination. The default value is 5. </p>
1410
- * <p>We recommend setting this parameter to a value greater than the amount of data you
1411
- * typically ingest into the delivery stream in 10 seconds. For example, if you typically
1412
- * ingest data at 1 MB/sec, the value should be 10 MB or higher. </p>
1413
- */
605
+
1414
606
  SizeInMBs?: number;
1415
- /**
1416
- * <p>Buffer incoming data for the specified period of time, in seconds, before delivering it
1417
- * to the destination. The default value is 300 (5 minutes). </p>
1418
- */
607
+
1419
608
  IntervalInSeconds?: number;
1420
609
  }
1421
610
  export declare namespace HttpEndpointBufferingHints {
1422
- /**
1423
- * @internal
1424
- */
611
+
1425
612
  const filterSensitiveLog: (obj: HttpEndpointBufferingHints) => any;
1426
613
  }
1427
- /**
1428
- * <p>Describes the configuration of the HTTP endpoint to which Kinesis Firehose delivers
1429
- * data.</p>
1430
- */
614
+
1431
615
  export interface HttpEndpointConfiguration {
1432
- /**
1433
- * <p>The URL of the HTTP endpoint selected as the destination.</p>
1434
- * <important>
1435
- * <p>If you choose an HTTP endpoint as your destination, review and follow the
1436
- * instructions in the <a href="https://docs.aws.amazon.com/firehose/latest/dev/httpdeliveryrequestresponse.html">Appendix - HTTP Endpoint
1437
- * Delivery Request and Response Specifications</a>.</p>
1438
- * </important>
1439
- */
616
+
1440
617
  Url: string | undefined;
1441
- /**
1442
- * <p>The name of the HTTP endpoint selected as the destination.</p>
1443
- */
618
+
1444
619
  Name?: string;
1445
- /**
1446
- * <p>The access key required for Kinesis Firehose to authenticate with the HTTP endpoint
1447
- * selected as the destination.</p>
1448
- */
620
+
1449
621
  AccessKey?: string;
1450
622
  }
1451
623
  export declare namespace HttpEndpointConfiguration {
1452
- /**
1453
- * @internal
1454
- */
624
+
1455
625
  const filterSensitiveLog: (obj: HttpEndpointConfiguration) => any;
1456
626
  }
1457
- /**
1458
- * <p>Describes the metadata that's delivered to the specified HTTP endpoint
1459
- * destination.</p>
1460
- */
627
+
1461
628
  export interface HttpEndpointCommonAttribute {
1462
- /**
1463
- * <p>The name of the HTTP endpoint common attribute.</p>
1464
- */
629
+
1465
630
  AttributeName: string | undefined;
1466
- /**
1467
- * <p>The value of the HTTP endpoint common attribute.</p>
1468
- */
631
+
1469
632
  AttributeValue: string | undefined;
1470
633
  }
1471
634
  export declare namespace HttpEndpointCommonAttribute {
1472
- /**
1473
- * @internal
1474
- */
635
+
1475
636
  const filterSensitiveLog: (obj: HttpEndpointCommonAttribute) => any;
1476
637
  }
1477
- /**
1478
- * <p>The configuration of the HTTP endpoint request.</p>
1479
- */
638
+
1480
639
  export interface HttpEndpointRequestConfiguration {
1481
- /**
1482
- * <p>Kinesis Data Firehose uses the content encoding to compress the body of a request before
1483
- * sending the request to the destination. For more information, see <a href="https://developer.mozilla.org/en-US/docs/Web/HTTP/Headers/Content-Encoding">Content-Encoding</a> in MDN Web Docs, the official Mozilla documentation.</p>
1484
- */
640
+
1485
641
  ContentEncoding?: ContentEncoding | string;
1486
- /**
1487
- * <p>Describes the metadata sent to the HTTP endpoint destination.</p>
1488
- */
642
+
1489
643
  CommonAttributes?: HttpEndpointCommonAttribute[];
1490
644
  }
1491
645
  export declare namespace HttpEndpointRequestConfiguration {
1492
- /**
1493
- * @internal
1494
- */
646
+
1495
647
  const filterSensitiveLog: (obj: HttpEndpointRequestConfiguration) => any;
1496
648
  }
1497
- /**
1498
- * <p>Describes the retry behavior in case Kinesis Data Firehose is unable to deliver data to
1499
- * the specified HTTP endpoint destination, or if it doesn't receive a valid acknowledgment of
1500
- * receipt from the specified HTTP endpoint destination.</p>
1501
- */
649
+
1502
650
  export interface HttpEndpointRetryOptions {
1503
- /**
1504
- * <p>The total amount of time that Kinesis Data Firehose spends on retries. This duration
1505
- * starts after the initial attempt to send data to the custom destination via HTTPS endpoint
1506
- * fails. It doesn't include the periods during which Kinesis Data Firehose waits for
1507
- * acknowledgment from the specified destination after each attempt. </p>
1508
- */
651
+
1509
652
  DurationInSeconds?: number;
1510
653
  }
1511
654
  export declare namespace HttpEndpointRetryOptions {
1512
- /**
1513
- * @internal
1514
- */
655
+
1515
656
  const filterSensitiveLog: (obj: HttpEndpointRetryOptions) => any;
1516
657
  }
1517
658
  export declare type HttpEndpointS3BackupMode = "AllData" | "FailedDataOnly";
1518
- /**
1519
- * <p>Describes the configuration of the HTTP endpoint destination.</p>
1520
- */
659
+
1521
660
  export interface HttpEndpointDestinationConfiguration {
1522
- /**
1523
- * <p>The configuration of the HTTP endpoint selected as the destination.</p>
1524
- */
661
+
1525
662
  EndpointConfiguration: HttpEndpointConfiguration | undefined;
1526
- /**
1527
- * <p>The buffering options that can be used before data is delivered to the specified
1528
- * destination. Kinesis Data Firehose treats these options as hints, and it might choose to
1529
- * use more optimal values. The <code>SizeInMBs</code> and <code>IntervalInSeconds</code>
1530
- * parameters are optional. However, if you specify a value for one of them, you must also
1531
- * provide a value for the other. </p>
1532
- */
663
+
1533
664
  BufferingHints?: HttpEndpointBufferingHints;
1534
- /**
1535
- * <p>Describes the Amazon CloudWatch logging options for your delivery stream.</p>
1536
- */
665
+
1537
666
  CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
1538
- /**
1539
- * <p>The configuration of the requeste sent to the HTTP endpoint specified as the
1540
- * destination.</p>
1541
- */
667
+
1542
668
  RequestConfiguration?: HttpEndpointRequestConfiguration;
1543
- /**
1544
- * <p>Describes a data processing configuration.</p>
1545
- */
669
+
1546
670
  ProcessingConfiguration?: ProcessingConfiguration;
1547
- /**
1548
- * <p>Kinesis Data Firehose uses this IAM role for all the permissions that the delivery
1549
- * stream needs.</p>
1550
- */
671
+
1551
672
  RoleARN?: string;
1552
- /**
1553
- * <p>Describes the retry behavior in case Kinesis Data Firehose is unable to deliver data to
1554
- * the specified HTTP endpoint destination, or if it doesn't receive a valid acknowledgment of
1555
- * receipt from the specified HTTP endpoint destination.</p>
1556
- */
673
+
1557
674
  RetryOptions?: HttpEndpointRetryOptions;
1558
- /**
1559
- * <p>Describes the S3 bucket backup options for the data that Kinesis Data Firehose delivers
1560
- * to the HTTP endpoint destination. You can back up all documents (<code>AllData</code>) or
1561
- * only the documents that Kinesis Data Firehose could not deliver to the specified HTTP
1562
- * endpoint destination (<code>FailedDataOnly</code>).</p>
1563
- */
675
+
1564
676
  S3BackupMode?: HttpEndpointS3BackupMode | string;
1565
- /**
1566
- * <p>Describes the configuration of a destination in Amazon S3.</p>
1567
- */
677
+
1568
678
  S3Configuration: S3DestinationConfiguration | undefined;
1569
679
  }
1570
680
  export declare namespace HttpEndpointDestinationConfiguration {
1571
- /**
1572
- * @internal
1573
- */
681
+
1574
682
  const filterSensitiveLog: (obj: HttpEndpointDestinationConfiguration) => any;
1575
683
  }
1576
- /**
1577
- * <p>The stream and role Amazon Resource Names (ARNs) for a Kinesis data stream used as
1578
- * the source for a delivery stream.</p>
1579
- */
684
+
1580
685
  export interface KinesisStreamSourceConfiguration {
1581
- /**
1582
- * <p>The ARN of the source Kinesis data stream. For more information, see <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html#arn-syntax-kinesis-streams">Amazon
1583
- * Kinesis Data Streams ARN Format</a>.</p>
1584
- */
686
+
1585
687
  KinesisStreamARN: string | undefined;
1586
- /**
1587
- * <p>The ARN of the role that provides access to the source Kinesis data stream. For more
1588
- * information, see <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html#arn-syntax-iam">AWS Identity and
1589
- * Access Management (IAM) ARN Format</a>.</p>
1590
- */
688
+
1591
689
  RoleARN: string | undefined;
1592
690
  }
1593
691
  export declare namespace KinesisStreamSourceConfiguration {
1594
- /**
1595
- * @internal
1596
- */
692
+
1597
693
  const filterSensitiveLog: (obj: KinesisStreamSourceConfiguration) => any;
1598
694
  }
1599
- /**
1600
- * <p>Configures retry behavior in case Kinesis Data Firehose is unable to deliver
1601
- * documents to Amazon Redshift.</p>
1602
- */
695
+
1603
696
  export interface RedshiftRetryOptions {
1604
- /**
1605
- * <p>The length of time during which Kinesis Data Firehose retries delivery after a
1606
- * failure, starting from the initial request and including the first attempt. The default
1607
- * value is 3600 seconds (60 minutes). Kinesis Data Firehose does not retry if the value of
1608
- * <code>DurationInSeconds</code> is 0 (zero) or if the first delivery attempt takes longer
1609
- * than the current value.</p>
1610
- */
697
+
1611
698
  DurationInSeconds?: number;
1612
699
  }
1613
700
  export declare namespace RedshiftRetryOptions {
1614
- /**
1615
- * @internal
1616
- */
701
+
1617
702
  const filterSensitiveLog: (obj: RedshiftRetryOptions) => any;
1618
703
  }
1619
704
  export declare type RedshiftS3BackupMode = "Disabled" | "Enabled";
1620
- /**
1621
- * <p>Describes the configuration of a destination in Amazon Redshift.</p>
1622
- */
705
+
1623
706
  export interface RedshiftDestinationConfiguration {
1624
- /**
1625
- * <p>The Amazon Resource Name (ARN) of the AWS credentials. For more information, see
1626
- * <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon
1627
- * Resource Names (ARNs) and AWS Service Namespaces</a>.</p>
1628
- */
707
+
1629
708
  RoleARN: string | undefined;
1630
- /**
1631
- * <p>The database connection string.</p>
1632
- */
709
+
1633
710
  ClusterJDBCURL: string | undefined;
1634
- /**
1635
- * <p>The <code>COPY</code> command.</p>
1636
- */
711
+
1637
712
  CopyCommand: CopyCommand | undefined;
1638
- /**
1639
- * <p>The name of the user.</p>
1640
- */
713
+
1641
714
  Username: string | undefined;
1642
- /**
1643
- * <p>The user password.</p>
1644
- */
715
+
1645
716
  Password: string | undefined;
1646
- /**
1647
- * <p>The retry behavior in case Kinesis Data Firehose is unable to deliver documents to
1648
- * Amazon Redshift. Default value is 3600 (60 minutes).</p>
1649
- */
717
+
1650
718
  RetryOptions?: RedshiftRetryOptions;
1651
- /**
1652
- * <p>The configuration for the intermediate Amazon S3 location from which Amazon Redshift
1653
- * obtains data. Restrictions are described in the topic for <a>CreateDeliveryStream</a>.</p>
1654
- * <p>The compression formats <code>SNAPPY</code> or <code>ZIP</code> cannot be specified
1655
- * in <code>RedshiftDestinationConfiguration.S3Configuration</code> because the Amazon
1656
- * Redshift <code>COPY</code> operation that reads from the S3 bucket doesn't support these
1657
- * compression formats.</p>
1658
- */
719
+
1659
720
  S3Configuration: S3DestinationConfiguration | undefined;
1660
- /**
1661
- * <p>The data processing configuration.</p>
1662
- */
721
+
1663
722
  ProcessingConfiguration?: ProcessingConfiguration;
1664
- /**
1665
- * <p>The Amazon S3 backup mode. After you create a delivery stream, you can update it to
1666
- * enable Amazon S3 backup if it is disabled. If backup is enabled, you can't update the
1667
- * delivery stream to disable it. </p>
1668
- */
723
+
1669
724
  S3BackupMode?: RedshiftS3BackupMode | string;
1670
- /**
1671
- * <p>The configuration for backup in Amazon S3.</p>
1672
- */
725
+
1673
726
  S3BackupConfiguration?: S3DestinationConfiguration;
1674
- /**
1675
- * <p>The CloudWatch logging options for your delivery stream.</p>
1676
- */
727
+
1677
728
  CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
1678
729
  }
1679
730
  export declare namespace RedshiftDestinationConfiguration {
1680
- /**
1681
- * @internal
1682
- */
731
+
1683
732
  const filterSensitiveLog: (obj: RedshiftDestinationConfiguration) => any;
1684
733
  }
1685
734
  export declare type HECEndpointType = "Event" | "Raw";
1686
- /**
1687
- * <p>Configures retry behavior in case Kinesis Data Firehose is unable to deliver
1688
- * documents to Splunk, or if it doesn't receive an acknowledgment from Splunk.</p>
1689
- */
735
+
1690
736
  export interface SplunkRetryOptions {
1691
- /**
1692
- * <p>The total amount of time that Kinesis Data Firehose spends on retries. This duration
1693
- * starts after the initial attempt to send data to Splunk fails. It doesn't include the
1694
- * periods during which Kinesis Data Firehose waits for acknowledgment from Splunk after each
1695
- * attempt.</p>
1696
- */
737
+
1697
738
  DurationInSeconds?: number;
1698
739
  }
1699
740
  export declare namespace SplunkRetryOptions {
1700
- /**
1701
- * @internal
1702
- */
741
+
1703
742
  const filterSensitiveLog: (obj: SplunkRetryOptions) => any;
1704
743
  }
1705
744
  export declare type SplunkS3BackupMode = "AllEvents" | "FailedEventsOnly";
1706
- /**
1707
- * <p>Describes the configuration of a destination in Splunk.</p>
1708
- */
745
+
1709
746
  export interface SplunkDestinationConfiguration {
1710
- /**
1711
- * <p>The HTTP Event Collector (HEC) endpoint to which Kinesis Data Firehose sends your
1712
- * data.</p>
1713
- */
747
+
1714
748
  HECEndpoint: string | undefined;
1715
- /**
1716
- * <p>This type can be either "Raw" or "Event."</p>
1717
- */
749
+
1718
750
  HECEndpointType: HECEndpointType | string | undefined;
1719
- /**
1720
- * <p>This is a GUID that you obtain from your Splunk cluster when you create a new HEC
1721
- * endpoint.</p>
1722
- */
751
+
1723
752
  HECToken: string | undefined;
1724
- /**
1725
- * <p>The amount of time that Kinesis Data Firehose waits to receive an acknowledgment from
1726
- * Splunk after it sends it data. At the end of the timeout period, Kinesis Data Firehose
1727
- * either tries to send the data again or considers it an error, based on your retry
1728
- * settings.</p>
1729
- */
753
+
1730
754
  HECAcknowledgmentTimeoutInSeconds?: number;
1731
- /**
1732
- * <p>The retry behavior in case Kinesis Data Firehose is unable to deliver data to Splunk,
1733
- * or if it doesn't receive an acknowledgment of receipt from Splunk.</p>
1734
- */
755
+
1735
756
  RetryOptions?: SplunkRetryOptions;
1736
- /**
1737
- * <p>Defines how documents should be delivered to Amazon S3. When set to
1738
- * <code>FailedEventsOnly</code>, Kinesis Data Firehose writes any data that could not be
1739
- * indexed to the configured Amazon S3 destination. When set to <code>AllEvents</code>,
1740
- * Kinesis Data Firehose delivers all incoming records to Amazon S3, and also writes failed
1741
- * documents to Amazon S3. The default value is <code>FailedEventsOnly</code>.</p>
1742
- * <p>You can update this backup mode from <code>FailedEventsOnly</code> to
1743
- * <code>AllEvents</code>. You can't update it from <code>AllEvents</code> to
1744
- * <code>FailedEventsOnly</code>.</p>
1745
- */
757
+
1746
758
  S3BackupMode?: SplunkS3BackupMode | string;
1747
- /**
1748
- * <p>The configuration for the backup Amazon S3 location.</p>
1749
- */
759
+
1750
760
  S3Configuration: S3DestinationConfiguration | undefined;
1751
- /**
1752
- * <p>The data processing configuration.</p>
1753
- */
761
+
1754
762
  ProcessingConfiguration?: ProcessingConfiguration;
1755
- /**
1756
- * <p>The Amazon CloudWatch logging options for your delivery stream.</p>
1757
- */
763
+
1758
764
  CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
1759
765
  }
1760
766
  export declare namespace SplunkDestinationConfiguration {
1761
- /**
1762
- * @internal
1763
- */
767
+
1764
768
  const filterSensitiveLog: (obj: SplunkDestinationConfiguration) => any;
1765
769
  }
1766
- /**
1767
- * <p>Metadata that you can assign to a delivery stream, consisting of a key-value
1768
- * pair.</p>
1769
- */
770
+
1770
771
  export interface Tag {
1771
- /**
1772
- * <p>A unique identifier for the tag. Maximum length: 128 characters. Valid characters:
1773
- * Unicode letters, digits, white space, _ . / = + - % @</p>
1774
- */
772
+
1775
773
  Key: string | undefined;
1776
- /**
1777
- * <p>An optional string, which you can use to describe or define the tag. Maximum length:
1778
- * 256 characters. Valid characters: Unicode letters, digits, white space, _ . / = + - %
1779
- * @</p>
1780
- */
774
+
1781
775
  Value?: string;
1782
776
  }
1783
777
  export declare namespace Tag {
1784
- /**
1785
- * @internal
1786
- */
778
+
1787
779
  const filterSensitiveLog: (obj: Tag) => any;
1788
780
  }
1789
781
  export interface CreateDeliveryStreamInput {
1790
- /**
1791
- * <p>The name of the delivery stream. This name must be unique per AWS account in the same
1792
- * AWS Region. If the delivery streams are in different accounts or different Regions, you can
1793
- * have multiple delivery streams with the same name.</p>
1794
- */
782
+
1795
783
  DeliveryStreamName: string | undefined;
1796
- /**
1797
- * <p>The delivery stream type. This parameter can be one of the following
1798
- * values:</p>
1799
- * <ul>
1800
- * <li>
1801
- * <p>
1802
- * <code>DirectPut</code>: Provider applications access the delivery stream
1803
- * directly.</p>
1804
- * </li>
1805
- * <li>
1806
- * <p>
1807
- * <code>KinesisStreamAsSource</code>: The delivery stream uses a Kinesis data
1808
- * stream as a source.</p>
1809
- * </li>
1810
- * </ul>
1811
- */
784
+
1812
785
  DeliveryStreamType?: DeliveryStreamType | string;
1813
- /**
1814
- * <p>When a Kinesis data stream is used as the source for the delivery stream, a <a>KinesisStreamSourceConfiguration</a> containing the Kinesis data stream Amazon
1815
- * Resource Name (ARN) and the role ARN for the source stream.</p>
1816
- */
786
+
1817
787
  KinesisStreamSourceConfiguration?: KinesisStreamSourceConfiguration;
1818
- /**
1819
- * <p>Used to specify the type and Amazon Resource Name (ARN) of the KMS key needed for
1820
- * Server-Side Encryption (SSE).</p>
1821
- */
788
+
1822
789
  DeliveryStreamEncryptionConfigurationInput?: DeliveryStreamEncryptionConfigurationInput;
1823
- /**
1824
- * @deprecated
1825
- *
1826
- * <p>[Deprecated]
1827
- * The destination in Amazon S3. You can specify only one destination.</p>
1828
- */
790
+
1829
791
  S3DestinationConfiguration?: S3DestinationConfiguration;
1830
- /**
1831
- * <p>The destination in Amazon S3. You can specify only one destination.</p>
1832
- */
792
+
1833
793
  ExtendedS3DestinationConfiguration?: ExtendedS3DestinationConfiguration;
1834
- /**
1835
- * <p>The destination in Amazon Redshift. You can specify only one destination.</p>
1836
- */
794
+
1837
795
  RedshiftDestinationConfiguration?: RedshiftDestinationConfiguration;
1838
- /**
1839
- * <p>The destination in Amazon ES. You can specify only one destination.</p>
1840
- */
796
+
1841
797
  ElasticsearchDestinationConfiguration?: ElasticsearchDestinationConfiguration;
1842
798
  AmazonopensearchserviceDestinationConfiguration?: AmazonopensearchserviceDestinationConfiguration;
1843
- /**
1844
- * <p>The destination in Splunk. You can specify only one destination.</p>
1845
- */
799
+
1846
800
  SplunkDestinationConfiguration?: SplunkDestinationConfiguration;
1847
- /**
1848
- * <p>Enables configuring Kinesis Firehose to deliver data to any HTTP endpoint destination.
1849
- * You can specify only one destination.</p>
1850
- */
801
+
1851
802
  HttpEndpointDestinationConfiguration?: HttpEndpointDestinationConfiguration;
1852
- /**
1853
- * <p>A set of tags to assign to the delivery stream. A tag is a key-value pair that you can
1854
- * define and assign to AWS resources. Tags are metadata. For example, you can add friendly
1855
- * names and descriptions or other types of information that can help you distinguish the
1856
- * delivery stream. For more information about tags, see <a href="https://docs.aws.amazon.com/awsaccountbilling/latest/aboutv2/cost-alloc-tags.html">Using Cost Allocation
1857
- * Tags</a> in the AWS Billing and Cost Management User Guide.</p>
1858
- *
1859
- * <p>You can specify up to 50 tags when creating a delivery stream.</p>
1860
- */
803
+
1861
804
  Tags?: Tag[];
1862
805
  }
1863
806
  export declare namespace CreateDeliveryStreamInput {
1864
- /**
1865
- * @internal
1866
- */
807
+
1867
808
  const filterSensitiveLog: (obj: CreateDeliveryStreamInput) => any;
1868
809
  }
1869
810
  export interface CreateDeliveryStreamOutput {
1870
- /**
1871
- * <p>The ARN of the delivery stream.</p>
1872
- */
811
+
1873
812
  DeliveryStreamARN?: string;
1874
813
  }
1875
814
  export declare namespace CreateDeliveryStreamOutput {
1876
- /**
1877
- * @internal
1878
- */
815
+
1879
816
  const filterSensitiveLog: (obj: CreateDeliveryStreamOutput) => any;
1880
817
  }
1881
- /**
1882
- * <p>The specified input parameter has a value that is not valid.</p>
1883
- */
818
+
1884
819
  export interface InvalidArgumentException extends __SmithyException, $MetadataBearer {
1885
820
  name: "InvalidArgumentException";
1886
821
  $fault: "client";
1887
- /**
1888
- * <p>A message that provides information about the error.</p>
1889
- */
822
+
1890
823
  message?: string;
1891
824
  }
1892
825
  export declare namespace InvalidArgumentException {
1893
- /**
1894
- * @internal
1895
- */
826
+
1896
827
  const filterSensitiveLog: (obj: InvalidArgumentException) => any;
1897
828
  }
1898
- /**
1899
- * <p>Kinesis Data Firehose throws this exception when an attempt to put records or to start
1900
- * or stop delivery stream encryption fails. This happens when the KMS service throws one of
1901
- * the following exception types: <code>AccessDeniedException</code>,
1902
- * <code>InvalidStateException</code>, <code>DisabledException</code>, or
1903
- * <code>NotFoundException</code>.</p>
1904
- */
829
+
1905
830
  export interface InvalidKMSResourceException extends __SmithyException, $MetadataBearer {
1906
831
  name: "InvalidKMSResourceException";
1907
832
  $fault: "client";
@@ -1909,91 +834,56 @@ export interface InvalidKMSResourceException extends __SmithyException, $Metadat
1909
834
  message?: string;
1910
835
  }
1911
836
  export declare namespace InvalidKMSResourceException {
1912
- /**
1913
- * @internal
1914
- */
837
+
1915
838
  const filterSensitiveLog: (obj: InvalidKMSResourceException) => any;
1916
839
  }
1917
- /**
1918
- * <p>You have already reached the limit for a requested resource.</p>
1919
- */
840
+
1920
841
  export interface LimitExceededException extends __SmithyException, $MetadataBearer {
1921
842
  name: "LimitExceededException";
1922
843
  $fault: "client";
1923
- /**
1924
- * <p>A message that provides information about the error.</p>
1925
- */
844
+
1926
845
  message?: string;
1927
846
  }
1928
847
  export declare namespace LimitExceededException {
1929
- /**
1930
- * @internal
1931
- */
848
+
1932
849
  const filterSensitiveLog: (obj: LimitExceededException) => any;
1933
850
  }
1934
- /**
1935
- * <p>The resource is already in use and not available for this operation.</p>
1936
- */
851
+
1937
852
  export interface ResourceInUseException extends __SmithyException, $MetadataBearer {
1938
853
  name: "ResourceInUseException";
1939
854
  $fault: "client";
1940
- /**
1941
- * <p>A message that provides information about the error.</p>
1942
- */
855
+
1943
856
  message?: string;
1944
857
  }
1945
858
  export declare namespace ResourceInUseException {
1946
- /**
1947
- * @internal
1948
- */
859
+
1949
860
  const filterSensitiveLog: (obj: ResourceInUseException) => any;
1950
861
  }
1951
862
  export interface DeleteDeliveryStreamInput {
1952
- /**
1953
- * <p>The name of the delivery stream.</p>
1954
- */
863
+
1955
864
  DeliveryStreamName: string | undefined;
1956
- /**
1957
- * <p>Set this to true if you want to delete the delivery stream even if Kinesis Data Firehose
1958
- * is unable to retire the grant for the CMK. Kinesis Data Firehose might be unable to retire
1959
- * the grant due to a customer error, such as when the CMK or the grant are in an invalid
1960
- * state. If you force deletion, you can then use the <a href="https://docs.aws.amazon.com/kms/latest/APIReference/API_RevokeGrant.html">RevokeGrant</a> operation to
1961
- * revoke the grant you gave to Kinesis Data Firehose. If a failure to retire the grant
1962
- * happens due to an AWS KMS issue, Kinesis Data Firehose keeps retrying the delete
1963
- * operation.</p>
1964
- * <p>The default value is false.</p>
1965
- */
865
+
1966
866
  AllowForceDelete?: boolean;
1967
867
  }
1968
868
  export declare namespace DeleteDeliveryStreamInput {
1969
- /**
1970
- * @internal
1971
- */
869
+
1972
870
  const filterSensitiveLog: (obj: DeleteDeliveryStreamInput) => any;
1973
871
  }
1974
872
  export interface DeleteDeliveryStreamOutput {
1975
873
  }
1976
874
  export declare namespace DeleteDeliveryStreamOutput {
1977
- /**
1978
- * @internal
1979
- */
875
+
1980
876
  const filterSensitiveLog: (obj: DeleteDeliveryStreamOutput) => any;
1981
877
  }
1982
- /**
1983
- * <p>The specified resource could not be found.</p>
1984
- */
878
+
1985
879
  export interface ResourceNotFoundException extends __SmithyException, $MetadataBearer {
1986
880
  name: "ResourceNotFoundException";
1987
881
  $fault: "client";
1988
- /**
1989
- * <p>A message that provides information about the error.</p>
1990
- */
882
+
1991
883
  message?: string;
1992
884
  }
1993
885
  export declare namespace ResourceNotFoundException {
1994
- /**
1995
- * @internal
1996
- */
886
+
1997
887
  const filterSensitiveLog: (obj: ResourceNotFoundException) => any;
1998
888
  }
1999
889
  export declare enum DeliveryStreamFailureType {
@@ -2013,25 +903,15 @@ export declare enum DeliveryStreamFailureType {
2013
903
  SUBNET_NOT_FOUND = "SUBNET_NOT_FOUND",
2014
904
  UNKNOWN_ERROR = "UNKNOWN_ERROR"
2015
905
  }
2016
- /**
2017
- * <p>Provides details in case one of the following operations fails due to an error related
2018
- * to KMS: <a>CreateDeliveryStream</a>, <a>DeleteDeliveryStream</a>,
2019
- * <a>StartDeliveryStreamEncryption</a>, <a>StopDeliveryStreamEncryption</a>.</p>
2020
- */
906
+
2021
907
  export interface FailureDescription {
2022
- /**
2023
- * <p>The type of error that caused the failure.</p>
2024
- */
908
+
2025
909
  Type: DeliveryStreamFailureType | string | undefined;
2026
- /**
2027
- * <p>A message providing details about the error that caused the failure.</p>
2028
- */
910
+
2029
911
  Details: string | undefined;
2030
912
  }
2031
913
  export declare namespace FailureDescription {
2032
- /**
2033
- * @internal
2034
- */
914
+
2035
915
  const filterSensitiveLog: (obj: FailureDescription) => any;
2036
916
  }
2037
917
  export declare enum DeliveryStreamEncryptionStatus {
@@ -2042,43 +922,19 @@ export declare enum DeliveryStreamEncryptionStatus {
2042
922
  ENABLING = "ENABLING",
2043
923
  ENABLING_FAILED = "ENABLING_FAILED"
2044
924
  }
2045
- /**
2046
- * <p>Contains information about the server-side encryption (SSE) status for the delivery
2047
- * stream, the type customer master key (CMK) in use, if any, and the ARN of the CMK. You can
2048
- * get <code>DeliveryStreamEncryptionConfiguration</code> by invoking the <a>DescribeDeliveryStream</a> operation. </p>
2049
- */
925
+
2050
926
  export interface DeliveryStreamEncryptionConfiguration {
2051
- /**
2052
- * <p>If <code>KeyType</code> is <code>CUSTOMER_MANAGED_CMK</code>, this field contains the
2053
- * ARN of the customer managed CMK. If <code>KeyType</code> is <code>AWS_OWNED_CMK</code>,
2054
- * <code>DeliveryStreamEncryptionConfiguration</code> doesn't contain a value for
2055
- * <code>KeyARN</code>.</p>
2056
- */
927
+
2057
928
  KeyARN?: string;
2058
- /**
2059
- * <p>Indicates the type of customer master key (CMK) that is used for encryption. The default
2060
- * setting is <code>AWS_OWNED_CMK</code>. For more information about CMKs, see <a href="https://docs.aws.amazon.com/kms/latest/developerguide/concepts.html#master_keys">Customer
2061
- * Master Keys (CMKs)</a>.</p>
2062
- */
929
+
2063
930
  KeyType?: KeyType | string;
2064
- /**
2065
- * <p>This is the server-side encryption (SSE) status for the delivery stream. For a full
2066
- * description of the different values of this status, see <a>StartDeliveryStreamEncryption</a> and <a>StopDeliveryStreamEncryption</a>. If this status is <code>ENABLING_FAILED</code>
2067
- * or <code>DISABLING_FAILED</code>, it is the status of the most recent attempt to enable or
2068
- * disable SSE, respectively.</p>
2069
- */
931
+
2070
932
  Status?: DeliveryStreamEncryptionStatus | string;
2071
- /**
2072
- * <p>Provides details in case one of the following operations fails due to an error related
2073
- * to KMS: <a>CreateDeliveryStream</a>, <a>DeleteDeliveryStream</a>,
2074
- * <a>StartDeliveryStreamEncryption</a>, <a>StopDeliveryStreamEncryption</a>.</p>
2075
- */
933
+
2076
934
  FailureDescription?: FailureDescription;
2077
935
  }
2078
936
  export declare namespace DeliveryStreamEncryptionConfiguration {
2079
- /**
2080
- * @internal
2081
- */
937
+
2082
938
  const filterSensitiveLog: (obj: DeliveryStreamEncryptionConfiguration) => any;
2083
939
  }
2084
940
  export declare enum DeliveryStreamStatus {
@@ -2088,1266 +944,605 @@ export declare enum DeliveryStreamStatus {
2088
944
  DELETING = "DELETING",
2089
945
  DELETING_FAILED = "DELETING_FAILED"
2090
946
  }
2091
- /**
2092
- * <p>The destination description in Amazon ES.</p>
2093
- */
947
+
2094
948
  export interface ElasticsearchDestinationDescription {
2095
- /**
2096
- * <p>The Amazon Resource Name (ARN) of the AWS credentials. For more information, see
2097
- * <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon
2098
- * Resource Names (ARNs) and AWS Service Namespaces</a>.</p>
2099
- */
949
+
2100
950
  RoleARN?: string;
2101
- /**
2102
- * <p>The ARN of the Amazon ES domain. For more information, see <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon
2103
- * Resource Names (ARNs) and AWS Service Namespaces</a>.</p>
2104
- *
2105
- * <p>Kinesis Data Firehose uses either <code>ClusterEndpoint</code> or <code>DomainARN</code>
2106
- * to send data to Amazon ES.</p>
2107
- */
951
+
2108
952
  DomainARN?: string;
2109
- /**
2110
- * <p>The endpoint to use when communicating with the cluster. Kinesis Data Firehose uses
2111
- * either this <code>ClusterEndpoint</code> or the <code>DomainARN</code> field to send data
2112
- * to Amazon ES.</p>
2113
- */
953
+
2114
954
  ClusterEndpoint?: string;
2115
- /**
2116
- * <p>The Elasticsearch index name.</p>
2117
- */
955
+
2118
956
  IndexName?: string;
2119
- /**
2120
- * <p>The Elasticsearch type name. This applies to Elasticsearch 6.x and lower versions.
2121
- * For Elasticsearch 7.x, there's no value for <code>TypeName</code>.</p>
2122
- */
957
+
2123
958
  TypeName?: string;
2124
- /**
2125
- * <p>The Elasticsearch index rotation period</p>
2126
- */
959
+
2127
960
  IndexRotationPeriod?: ElasticsearchIndexRotationPeriod | string;
2128
- /**
2129
- * <p>The buffering options.</p>
2130
- */
961
+
2131
962
  BufferingHints?: ElasticsearchBufferingHints;
2132
- /**
2133
- * <p>The Amazon ES retry options.</p>
2134
- */
963
+
2135
964
  RetryOptions?: ElasticsearchRetryOptions;
2136
- /**
2137
- * <p>The Amazon S3 backup mode.</p>
2138
- */
965
+
2139
966
  S3BackupMode?: ElasticsearchS3BackupMode | string;
2140
- /**
2141
- * <p>The Amazon S3 destination.</p>
2142
- */
967
+
2143
968
  S3DestinationDescription?: S3DestinationDescription;
2144
- /**
2145
- * <p>The data processing configuration.</p>
2146
- */
969
+
2147
970
  ProcessingConfiguration?: ProcessingConfiguration;
2148
- /**
2149
- * <p>The Amazon CloudWatch logging options.</p>
2150
- */
971
+
2151
972
  CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
2152
- /**
2153
- * <p>The details of the VPC of the Amazon ES destination.</p>
2154
- */
973
+
2155
974
  VpcConfigurationDescription?: VpcConfigurationDescription;
2156
975
  }
2157
976
  export declare namespace ElasticsearchDestinationDescription {
2158
- /**
2159
- * @internal
2160
- */
977
+
2161
978
  const filterSensitiveLog: (obj: ElasticsearchDestinationDescription) => any;
2162
979
  }
2163
- /**
2164
- * <p>Describes a destination in Amazon S3.</p>
2165
- */
980
+
2166
981
  export interface ExtendedS3DestinationDescription {
2167
- /**
2168
- * <p>The Amazon Resource Name (ARN) of the AWS credentials. For more information, see
2169
- * <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon
2170
- * Resource Names (ARNs) and AWS Service Namespaces</a>.</p>
2171
- */
982
+
2172
983
  RoleARN: string | undefined;
2173
- /**
2174
- * <p>The ARN of the S3 bucket. For more information, see <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon Resource Names (ARNs) and
2175
- * AWS Service Namespaces</a>.</p>
2176
- */
984
+
2177
985
  BucketARN: string | undefined;
2178
- /**
2179
- * <p>The "YYYY/MM/DD/HH" time format prefix is automatically used for delivered Amazon S3
2180
- * files. You can also specify a custom prefix, as described in <a href="https://docs.aws.amazon.com/firehose/latest/dev/s3-prefixes.html">Custom Prefixes for Amazon S3
2181
- * Objects</a>.</p>
2182
- */
986
+
2183
987
  Prefix?: string;
2184
- /**
2185
- * <p>A prefix that Kinesis Data Firehose evaluates and adds to failed records before writing
2186
- * them to S3. This prefix appears immediately following the bucket name. For information
2187
- * about how to specify this prefix, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/s3-prefixes.html">Custom Prefixes for Amazon S3
2188
- * Objects</a>.</p>
2189
- */
988
+
2190
989
  ErrorOutputPrefix?: string;
2191
- /**
2192
- * <p>The buffering option.</p>
2193
- */
990
+
2194
991
  BufferingHints: BufferingHints | undefined;
2195
- /**
2196
- * <p>The compression format. If no value is specified, the default is
2197
- * <code>UNCOMPRESSED</code>.</p>
2198
- */
992
+
2199
993
  CompressionFormat: CompressionFormat | string | undefined;
2200
- /**
2201
- * <p>The encryption configuration. If no value is specified, the default is no
2202
- * encryption.</p>
2203
- */
994
+
2204
995
  EncryptionConfiguration: EncryptionConfiguration | undefined;
2205
- /**
2206
- * <p>The Amazon CloudWatch logging options for your delivery stream.</p>
2207
- */
996
+
2208
997
  CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
2209
- /**
2210
- * <p>The data processing configuration.</p>
2211
- */
998
+
2212
999
  ProcessingConfiguration?: ProcessingConfiguration;
2213
- /**
2214
- * <p>The Amazon S3 backup mode.</p>
2215
- */
1000
+
2216
1001
  S3BackupMode?: S3BackupMode | string;
2217
- /**
2218
- * <p>The configuration for backup in Amazon S3.</p>
2219
- */
1002
+
2220
1003
  S3BackupDescription?: S3DestinationDescription;
2221
- /**
2222
- * <p>The serializer, deserializer, and schema for converting data from the JSON format to
2223
- * the Parquet or ORC format before writing it to Amazon S3.</p>
2224
- */
1004
+
2225
1005
  DataFormatConversionConfiguration?: DataFormatConversionConfiguration;
2226
- /**
2227
- * <p>The configuration of the dynamic partitioning mechanism that creates smaller data sets
2228
- * from the streaming data by partitioning it based on partition keys. Currently, dynamic
2229
- * partitioning is only supported for Amazon S3 destinations. For more information, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/dynamic-partitioning.html">https://docs.aws.amazon.com/firehose/latest/dev/dynamic-partitioning.html</a>
2230
- * </p>
2231
- */
1006
+
2232
1007
  DynamicPartitioningConfiguration?: DynamicPartitioningConfiguration;
2233
1008
  }
2234
1009
  export declare namespace ExtendedS3DestinationDescription {
2235
- /**
2236
- * @internal
2237
- */
1010
+
2238
1011
  const filterSensitiveLog: (obj: ExtendedS3DestinationDescription) => any;
2239
1012
  }
2240
- /**
2241
- * <p>Describes the HTTP endpoint selected as the destination. </p>
2242
- */
1013
+
2243
1014
  export interface HttpEndpointDescription {
2244
- /**
2245
- * <p>The URL of the HTTP endpoint selected as the destination.</p>
2246
- */
1015
+
2247
1016
  Url?: string;
2248
- /**
2249
- * <p>The name of the HTTP endpoint selected as the destination.</p>
2250
- */
1017
+
2251
1018
  Name?: string;
2252
1019
  }
2253
1020
  export declare namespace HttpEndpointDescription {
2254
- /**
2255
- * @internal
2256
- */
1021
+
2257
1022
  const filterSensitiveLog: (obj: HttpEndpointDescription) => any;
2258
1023
  }
2259
- /**
2260
- * <p>Describes the HTTP endpoint destination.</p>
2261
- */
1024
+
2262
1025
  export interface HttpEndpointDestinationDescription {
2263
- /**
2264
- * <p>The configuration of the specified HTTP endpoint destination.</p>
2265
- */
1026
+
2266
1027
  EndpointConfiguration?: HttpEndpointDescription;
2267
- /**
2268
- * <p>Describes buffering options that can be applied to the data before it is delivered to
2269
- * the HTTPS endpoint destination. Kinesis Data Firehose teats these options as hints, and it
2270
- * might choose to use more optimal values. The <code>SizeInMBs</code> and
2271
- * <code>IntervalInSeconds</code> parameters are optional. However, if specify a value for
2272
- * one of them, you must also provide a value for the other. </p>
2273
- */
1028
+
2274
1029
  BufferingHints?: HttpEndpointBufferingHints;
2275
- /**
2276
- * <p>Describes the Amazon CloudWatch logging options for your delivery stream.</p>
2277
- */
1030
+
2278
1031
  CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
2279
- /**
2280
- * <p>The configuration of request sent to the HTTP endpoint specified as the
2281
- * destination.</p>
2282
- */
1032
+
2283
1033
  RequestConfiguration?: HttpEndpointRequestConfiguration;
2284
- /**
2285
- * <p>Describes a data processing configuration.</p>
2286
- */
1034
+
2287
1035
  ProcessingConfiguration?: ProcessingConfiguration;
2288
- /**
2289
- * <p>Kinesis Data Firehose uses this IAM role for all the permissions that the delivery
2290
- * stream needs.</p>
2291
- */
1036
+
2292
1037
  RoleARN?: string;
2293
- /**
2294
- * <p>Describes the retry behavior in case Kinesis Data Firehose is unable to deliver data to
2295
- * the specified HTTP endpoint destination, or if it doesn't receive a valid acknowledgment of
2296
- * receipt from the specified HTTP endpoint destination.</p>
2297
- */
1038
+
2298
1039
  RetryOptions?: HttpEndpointRetryOptions;
2299
- /**
2300
- * <p>Describes the S3 bucket backup options for the data that Kinesis Firehose delivers to
2301
- * the HTTP endpoint destination. You can back up all documents (<code>AllData</code>) or only
2302
- * the documents that Kinesis Data Firehose could not deliver to the specified HTTP endpoint
2303
- * destination (<code>FailedDataOnly</code>).</p>
2304
- */
1040
+
2305
1041
  S3BackupMode?: HttpEndpointS3BackupMode | string;
2306
- /**
2307
- * <p>Describes a destination in Amazon S3.</p>
2308
- */
1042
+
2309
1043
  S3DestinationDescription?: S3DestinationDescription;
2310
1044
  }
2311
1045
  export declare namespace HttpEndpointDestinationDescription {
2312
- /**
2313
- * @internal
2314
- */
1046
+
2315
1047
  const filterSensitiveLog: (obj: HttpEndpointDestinationDescription) => any;
2316
1048
  }
2317
- /**
2318
- * <p>Describes a destination in Amazon Redshift.</p>
2319
- */
1049
+
2320
1050
  export interface RedshiftDestinationDescription {
2321
- /**
2322
- * <p>The Amazon Resource Name (ARN) of the AWS credentials. For more information, see
2323
- * <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon
2324
- * Resource Names (ARNs) and AWS Service Namespaces</a>.</p>
2325
- */
1051
+
2326
1052
  RoleARN: string | undefined;
2327
- /**
2328
- * <p>The database connection string.</p>
2329
- */
1053
+
2330
1054
  ClusterJDBCURL: string | undefined;
2331
- /**
2332
- * <p>The <code>COPY</code> command.</p>
2333
- */
1055
+
2334
1056
  CopyCommand: CopyCommand | undefined;
2335
- /**
2336
- * <p>The name of the user.</p>
2337
- */
1057
+
2338
1058
  Username: string | undefined;
2339
- /**
2340
- * <p>The retry behavior in case Kinesis Data Firehose is unable to deliver documents to
2341
- * Amazon Redshift. Default value is 3600 (60 minutes).</p>
2342
- */
1059
+
2343
1060
  RetryOptions?: RedshiftRetryOptions;
2344
- /**
2345
- * <p>The Amazon S3 destination.</p>
2346
- */
1061
+
2347
1062
  S3DestinationDescription: S3DestinationDescription | undefined;
2348
- /**
2349
- * <p>The data processing configuration.</p>
2350
- */
1063
+
2351
1064
  ProcessingConfiguration?: ProcessingConfiguration;
2352
- /**
2353
- * <p>The Amazon S3 backup mode.</p>
2354
- */
1065
+
2355
1066
  S3BackupMode?: RedshiftS3BackupMode | string;
2356
- /**
2357
- * <p>The configuration for backup in Amazon S3.</p>
2358
- */
1067
+
2359
1068
  S3BackupDescription?: S3DestinationDescription;
2360
- /**
2361
- * <p>The Amazon CloudWatch logging options for your delivery stream.</p>
2362
- */
1069
+
2363
1070
  CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
2364
1071
  }
2365
1072
  export declare namespace RedshiftDestinationDescription {
2366
- /**
2367
- * @internal
2368
- */
1073
+
2369
1074
  const filterSensitiveLog: (obj: RedshiftDestinationDescription) => any;
2370
1075
  }
2371
- /**
2372
- * <p>Describes a destination in Splunk.</p>
2373
- */
1076
+
2374
1077
  export interface SplunkDestinationDescription {
2375
- /**
2376
- * <p>The HTTP Event Collector (HEC) endpoint to which Kinesis Data Firehose sends your
2377
- * data.</p>
2378
- */
1078
+
2379
1079
  HECEndpoint?: string;
2380
- /**
2381
- * <p>This type can be either "Raw" or "Event."</p>
2382
- */
1080
+
2383
1081
  HECEndpointType?: HECEndpointType | string;
2384
- /**
2385
- * <p>A GUID you obtain from your Splunk cluster when you create a new HEC
2386
- * endpoint.</p>
2387
- */
1082
+
2388
1083
  HECToken?: string;
2389
- /**
2390
- * <p>The amount of time that Kinesis Data Firehose waits to receive an acknowledgment from
2391
- * Splunk after it sends it data. At the end of the timeout period, Kinesis Data Firehose
2392
- * either tries to send the data again or considers it an error, based on your retry
2393
- * settings.</p>
2394
- */
1084
+
2395
1085
  HECAcknowledgmentTimeoutInSeconds?: number;
2396
- /**
2397
- * <p>The retry behavior in case Kinesis Data Firehose is unable to deliver data to Splunk
2398
- * or if it doesn't receive an acknowledgment of receipt from Splunk.</p>
2399
- */
1086
+
2400
1087
  RetryOptions?: SplunkRetryOptions;
2401
- /**
2402
- * <p>Defines how documents should be delivered to Amazon S3. When set to
2403
- * <code>FailedDocumentsOnly</code>, Kinesis Data Firehose writes any data that could not
2404
- * be indexed to the configured Amazon S3 destination. When set to <code>AllDocuments</code>,
2405
- * Kinesis Data Firehose delivers all incoming records to Amazon S3, and also writes failed
2406
- * documents to Amazon S3. Default value is <code>FailedDocumentsOnly</code>. </p>
2407
- */
1088
+
2408
1089
  S3BackupMode?: SplunkS3BackupMode | string;
2409
- /**
2410
- * <p>The Amazon S3 destination.></p>
2411
- */
1090
+
2412
1091
  S3DestinationDescription?: S3DestinationDescription;
2413
- /**
2414
- * <p>The data processing configuration.</p>
2415
- */
1092
+
2416
1093
  ProcessingConfiguration?: ProcessingConfiguration;
2417
- /**
2418
- * <p>The Amazon CloudWatch logging options for your delivery stream.</p>
2419
- */
1094
+
2420
1095
  CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
2421
1096
  }
2422
1097
  export declare namespace SplunkDestinationDescription {
2423
- /**
2424
- * @internal
2425
- */
1098
+
2426
1099
  const filterSensitiveLog: (obj: SplunkDestinationDescription) => any;
2427
1100
  }
2428
- /**
2429
- * <p>Describes the destination for a delivery stream.</p>
2430
- */
1101
+
2431
1102
  export interface DestinationDescription {
2432
- /**
2433
- * <p>The ID of the destination.</p>
2434
- */
1103
+
2435
1104
  DestinationId: string | undefined;
2436
- /**
2437
- * <p>[Deprecated] The destination in Amazon S3.</p>
2438
- */
1105
+
2439
1106
  S3DestinationDescription?: S3DestinationDescription;
2440
- /**
2441
- * <p>The destination in Amazon S3.</p>
2442
- */
1107
+
2443
1108
  ExtendedS3DestinationDescription?: ExtendedS3DestinationDescription;
2444
- /**
2445
- * <p>The destination in Amazon Redshift.</p>
2446
- */
1109
+
2447
1110
  RedshiftDestinationDescription?: RedshiftDestinationDescription;
2448
- /**
2449
- * <p>The destination in Amazon ES.</p>
2450
- */
1111
+
2451
1112
  ElasticsearchDestinationDescription?: ElasticsearchDestinationDescription;
2452
1113
  AmazonopensearchserviceDestinationDescription?: AmazonopensearchserviceDestinationDescription;
2453
- /**
2454
- * <p>The destination in Splunk.</p>
2455
- */
1114
+
2456
1115
  SplunkDestinationDescription?: SplunkDestinationDescription;
2457
- /**
2458
- * <p>Describes the specified HTTP endpoint destination.</p>
2459
- */
1116
+
2460
1117
  HttpEndpointDestinationDescription?: HttpEndpointDestinationDescription;
2461
1118
  }
2462
1119
  export declare namespace DestinationDescription {
2463
- /**
2464
- * @internal
2465
- */
1120
+
2466
1121
  const filterSensitiveLog: (obj: DestinationDescription) => any;
2467
1122
  }
2468
- /**
2469
- * <p>Details about a Kinesis data stream used as the source for a Kinesis Data Firehose
2470
- * delivery stream.</p>
2471
- */
1123
+
2472
1124
  export interface KinesisStreamSourceDescription {
2473
- /**
2474
- * <p>The Amazon Resource Name (ARN) of the source Kinesis data stream. For more
2475
- * information, see <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html#arn-syntax-kinesis-streams">Amazon
2476
- * Kinesis Data Streams ARN Format</a>.</p>
2477
- */
1125
+
2478
1126
  KinesisStreamARN?: string;
2479
- /**
2480
- * <p>The ARN of the role used by the source Kinesis data stream. For more information, see
2481
- * <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html#arn-syntax-iam">AWS Identity and
2482
- * Access Management (IAM) ARN Format</a>.</p>
2483
- */
1127
+
2484
1128
  RoleARN?: string;
2485
- /**
2486
- * <p>Kinesis Data Firehose starts retrieving records from the Kinesis data stream starting
2487
- * with this timestamp.</p>
2488
- */
1129
+
2489
1130
  DeliveryStartTimestamp?: Date;
2490
1131
  }
2491
1132
  export declare namespace KinesisStreamSourceDescription {
2492
- /**
2493
- * @internal
2494
- */
1133
+
2495
1134
  const filterSensitiveLog: (obj: KinesisStreamSourceDescription) => any;
2496
1135
  }
2497
- /**
2498
- * <p>Details about a Kinesis data stream used as the source for a Kinesis Data Firehose
2499
- * delivery stream.</p>
2500
- */
1136
+
2501
1137
  export interface SourceDescription {
2502
- /**
2503
- * <p>The <a>KinesisStreamSourceDescription</a> value for the source Kinesis
2504
- * data stream.</p>
2505
- */
1138
+
2506
1139
  KinesisStreamSourceDescription?: KinesisStreamSourceDescription;
2507
1140
  }
2508
1141
  export declare namespace SourceDescription {
2509
- /**
2510
- * @internal
2511
- */
1142
+
2512
1143
  const filterSensitiveLog: (obj: SourceDescription) => any;
2513
1144
  }
2514
- /**
2515
- * <p>Contains information about a delivery stream.</p>
2516
- */
1145
+
2517
1146
  export interface DeliveryStreamDescription {
2518
- /**
2519
- * <p>The name of the delivery stream.</p>
2520
- */
1147
+
2521
1148
  DeliveryStreamName: string | undefined;
2522
- /**
2523
- * <p>The Amazon Resource Name (ARN) of the delivery stream. For more information, see
2524
- * <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon
2525
- * Resource Names (ARNs) and AWS Service Namespaces</a>.</p>
2526
- */
1149
+
2527
1150
  DeliveryStreamARN: string | undefined;
2528
- /**
2529
- * <p>The status of the delivery stream. If the status of a delivery stream is
2530
- * <code>CREATING_FAILED</code>, this status doesn't change, and you can't invoke
2531
- * <code>CreateDeliveryStream</code> again on it. However, you can invoke the <a>DeleteDeliveryStream</a> operation to delete it.</p>
2532
- */
1151
+
2533
1152
  DeliveryStreamStatus: DeliveryStreamStatus | string | undefined;
2534
- /**
2535
- * <p>Provides details in case one of the following operations fails due to an error related
2536
- * to KMS: <a>CreateDeliveryStream</a>, <a>DeleteDeliveryStream</a>,
2537
- * <a>StartDeliveryStreamEncryption</a>, <a>StopDeliveryStreamEncryption</a>.</p>
2538
- */
1153
+
2539
1154
  FailureDescription?: FailureDescription;
2540
- /**
2541
- * <p>Indicates the server-side encryption (SSE) status for the delivery stream.</p>
2542
- */
1155
+
2543
1156
  DeliveryStreamEncryptionConfiguration?: DeliveryStreamEncryptionConfiguration;
2544
- /**
2545
- * <p>The delivery stream type. This can be one of the following values:</p>
2546
- * <ul>
2547
- * <li>
2548
- * <p>
2549
- * <code>DirectPut</code>: Provider applications access the delivery stream
2550
- * directly.</p>
2551
- * </li>
2552
- * <li>
2553
- * <p>
2554
- * <code>KinesisStreamAsSource</code>: The delivery stream uses a Kinesis data
2555
- * stream as a source.</p>
2556
- * </li>
2557
- * </ul>
2558
- */
1157
+
2559
1158
  DeliveryStreamType: DeliveryStreamType | string | undefined;
2560
- /**
2561
- * <p>Each time the destination is updated for a delivery stream, the version ID is
2562
- * changed, and the current version ID is required when updating the destination. This is so
2563
- * that the service knows it is applying the changes to the correct version of the delivery
2564
- * stream.</p>
2565
- */
1159
+
2566
1160
  VersionId: string | undefined;
2567
- /**
2568
- * <p>The date and time that the delivery stream was created.</p>
2569
- */
1161
+
2570
1162
  CreateTimestamp?: Date;
2571
- /**
2572
- * <p>The date and time that the delivery stream was last updated.</p>
2573
- */
1163
+
2574
1164
  LastUpdateTimestamp?: Date;
2575
- /**
2576
- * <p>If the <code>DeliveryStreamType</code> parameter is
2577
- * <code>KinesisStreamAsSource</code>, a <a>SourceDescription</a> object
2578
- * describing the source Kinesis data stream.</p>
2579
- */
1165
+
2580
1166
  Source?: SourceDescription;
2581
- /**
2582
- * <p>The destinations.</p>
2583
- */
1167
+
2584
1168
  Destinations: DestinationDescription[] | undefined;
2585
- /**
2586
- * <p>Indicates whether there are more destinations available to list.</p>
2587
- */
1169
+
2588
1170
  HasMoreDestinations: boolean | undefined;
2589
1171
  }
2590
1172
  export declare namespace DeliveryStreamDescription {
2591
- /**
2592
- * @internal
2593
- */
1173
+
2594
1174
  const filterSensitiveLog: (obj: DeliveryStreamDescription) => any;
2595
1175
  }
2596
1176
  export interface DescribeDeliveryStreamInput {
2597
- /**
2598
- * <p>The name of the delivery stream.</p>
2599
- */
1177
+
2600
1178
  DeliveryStreamName: string | undefined;
2601
- /**
2602
- * <p>The limit on the number of destinations to return. You can have one destination per
2603
- * delivery stream.</p>
2604
- */
1179
+
2605
1180
  Limit?: number;
2606
- /**
2607
- * <p>The ID of the destination to start returning the destination information. Kinesis
2608
- * Data Firehose supports one destination per delivery stream.</p>
2609
- */
1181
+
2610
1182
  ExclusiveStartDestinationId?: string;
2611
1183
  }
2612
1184
  export declare namespace DescribeDeliveryStreamInput {
2613
- /**
2614
- * @internal
2615
- */
1185
+
2616
1186
  const filterSensitiveLog: (obj: DescribeDeliveryStreamInput) => any;
2617
1187
  }
2618
1188
  export interface DescribeDeliveryStreamOutput {
2619
- /**
2620
- * <p>Information about the delivery stream.</p>
2621
- */
1189
+
2622
1190
  DeliveryStreamDescription: DeliveryStreamDescription | undefined;
2623
1191
  }
2624
1192
  export declare namespace DescribeDeliveryStreamOutput {
2625
- /**
2626
- * @internal
2627
- */
1193
+
2628
1194
  const filterSensitiveLog: (obj: DescribeDeliveryStreamOutput) => any;
2629
1195
  }
2630
- /**
2631
- * <p>Describes an update for a destination in Amazon ES.</p>
2632
- */
1196
+
2633
1197
  export interface ElasticsearchDestinationUpdate {
2634
- /**
2635
- * <p>The Amazon Resource Name (ARN) of the IAM role to be assumed by Kinesis Data Firehose
2636
- * for calling the Amazon ES Configuration API and for indexing documents. For more
2637
- * information, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/controlling-access.html#using-iam-s3">Grant Kinesis Data
2638
- * Firehose Access to an Amazon S3 Destination</a> and <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon Resource Names (ARNs) and
2639
- * AWS Service Namespaces</a>.</p>
2640
- */
1198
+
2641
1199
  RoleARN?: string;
2642
- /**
2643
- * <p>The ARN of the Amazon ES domain. The IAM role must have permissions
2644
- * for <code>DescribeElasticsearchDomain</code>, <code>DescribeElasticsearchDomains</code>,
2645
- * and <code>DescribeElasticsearchDomainConfig</code> after assuming the IAM role specified in
2646
- * <code>RoleARN</code>. For more information, see <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon Resource Names (ARNs) and
2647
- * AWS Service Namespaces</a>.</p>
2648
- *
2649
- * <p>Specify either <code>ClusterEndpoint</code> or <code>DomainARN</code>.</p>
2650
- */
1200
+
2651
1201
  DomainARN?: string;
2652
- /**
2653
- * <p>The endpoint to use when communicating with the cluster. Specify either this
2654
- * <code>ClusterEndpoint</code> or the <code>DomainARN</code> field.</p>
2655
- */
1202
+
2656
1203
  ClusterEndpoint?: string;
2657
- /**
2658
- * <p>The Elasticsearch index name.</p>
2659
- */
1204
+
2660
1205
  IndexName?: string;
2661
- /**
2662
- * <p>The Elasticsearch type name. For Elasticsearch 6.x, there can be only one type per
2663
- * index. If you try to specify a new type for an existing index that already has another
2664
- * type, Kinesis Data Firehose returns an error during runtime.</p>
2665
- *
2666
- * <p>If you upgrade Elasticsearch from 6.x to 7.x and don’t update your delivery stream,
2667
- * Kinesis Data Firehose still delivers data to Elasticsearch with the old index name and type
2668
- * name. If you want to update your delivery stream with a new index name, provide an empty
2669
- * string for <code>TypeName</code>. </p>
2670
- */
1206
+
2671
1207
  TypeName?: string;
2672
- /**
2673
- * <p>The Elasticsearch index rotation period. Index rotation appends a timestamp to
2674
- * <code>IndexName</code> to facilitate the expiration of old data. For more information,
2675
- * see <a href="https://docs.aws.amazon.com/firehose/latest/dev/basic-deliver.html#es-index-rotation">Index Rotation for the
2676
- * Amazon ES Destination</a>. Default value is <code>OneDay</code>.</p>
2677
- */
1208
+
2678
1209
  IndexRotationPeriod?: ElasticsearchIndexRotationPeriod | string;
2679
- /**
2680
- * <p>The buffering options. If no value is specified,
2681
- * <code>ElasticsearchBufferingHints</code> object default values are used. </p>
2682
- */
1210
+
2683
1211
  BufferingHints?: ElasticsearchBufferingHints;
2684
- /**
2685
- * <p>The retry behavior in case Kinesis Data Firehose is unable to deliver documents to
2686
- * Amazon ES. The default value is 300 (5 minutes).</p>
2687
- */
1212
+
2688
1213
  RetryOptions?: ElasticsearchRetryOptions;
2689
- /**
2690
- * <p>The Amazon S3 destination.</p>
2691
- */
1214
+
2692
1215
  S3Update?: S3DestinationUpdate;
2693
- /**
2694
- * <p>The data processing configuration.</p>
2695
- */
1216
+
2696
1217
  ProcessingConfiguration?: ProcessingConfiguration;
2697
- /**
2698
- * <p>The CloudWatch logging options for your delivery stream.</p>
2699
- */
1218
+
2700
1219
  CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
2701
1220
  }
2702
1221
  export declare namespace ElasticsearchDestinationUpdate {
2703
- /**
2704
- * @internal
2705
- */
1222
+
2706
1223
  const filterSensitiveLog: (obj: ElasticsearchDestinationUpdate) => any;
2707
1224
  }
2708
- /**
2709
- * <p>Describes an update for a destination in Amazon S3.</p>
2710
- */
1225
+
2711
1226
  export interface ExtendedS3DestinationUpdate {
2712
- /**
2713
- * <p>The Amazon Resource Name (ARN) of the AWS credentials. For more information, see
2714
- * <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon
2715
- * Resource Names (ARNs) and AWS Service Namespaces</a>.</p>
2716
- */
1227
+
2717
1228
  RoleARN?: string;
2718
- /**
2719
- * <p>The ARN of the S3 bucket. For more information, see <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon Resource Names (ARNs) and
2720
- * AWS Service Namespaces</a>.</p>
2721
- */
1229
+
2722
1230
  BucketARN?: string;
2723
- /**
2724
- * <p>The "YYYY/MM/DD/HH" time format prefix is automatically used for delivered Amazon S3
2725
- * files. You can also specify a custom prefix, as described in <a href="https://docs.aws.amazon.com/firehose/latest/dev/s3-prefixes.html">Custom Prefixes for Amazon S3
2726
- * Objects</a>.</p>
2727
- */
1231
+
2728
1232
  Prefix?: string;
2729
- /**
2730
- * <p>A prefix that Kinesis Data Firehose evaluates and adds to failed records before writing
2731
- * them to S3. This prefix appears immediately following the bucket name. For information
2732
- * about how to specify this prefix, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/s3-prefixes.html">Custom Prefixes for Amazon S3
2733
- * Objects</a>.</p>
2734
- */
1233
+
2735
1234
  ErrorOutputPrefix?: string;
2736
- /**
2737
- * <p>The buffering option.</p>
2738
- */
1235
+
2739
1236
  BufferingHints?: BufferingHints;
2740
- /**
2741
- * <p>The compression format. If no value is specified, the default is
2742
- * <code>UNCOMPRESSED</code>. </p>
2743
- */
1237
+
2744
1238
  CompressionFormat?: CompressionFormat | string;
2745
- /**
2746
- * <p>The encryption configuration. If no value is specified, the default is no
2747
- * encryption.</p>
2748
- */
1239
+
2749
1240
  EncryptionConfiguration?: EncryptionConfiguration;
2750
- /**
2751
- * <p>The Amazon CloudWatch logging options for your delivery stream.</p>
2752
- */
1241
+
2753
1242
  CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
2754
- /**
2755
- * <p>The data processing configuration.</p>
2756
- */
1243
+
2757
1244
  ProcessingConfiguration?: ProcessingConfiguration;
2758
- /**
2759
- * <p>You can update a delivery stream to enable Amazon S3 backup if it is disabled. If
2760
- * backup is enabled, you can't update the delivery stream to disable it. </p>
2761
- */
1245
+
2762
1246
  S3BackupMode?: S3BackupMode | string;
2763
- /**
2764
- * <p>The Amazon S3 destination for backup.</p>
2765
- */
1247
+
2766
1248
  S3BackupUpdate?: S3DestinationUpdate;
2767
- /**
2768
- * <p>The serializer, deserializer, and schema for converting data from the JSON format to
2769
- * the Parquet or ORC format before writing it to Amazon S3.</p>
2770
- */
1249
+
2771
1250
  DataFormatConversionConfiguration?: DataFormatConversionConfiguration;
2772
- /**
2773
- * <p>The configuration of the dynamic partitioning mechanism that creates smaller data sets
2774
- * from the streaming data by partitioning it based on partition keys. Currently, dynamic
2775
- * partitioning is only supported for Amazon S3 destinations. For more information, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/dynamic-partitioning.html">https://docs.aws.amazon.com/firehose/latest/dev/dynamic-partitioning.html</a>
2776
- * </p>
2777
- */
1251
+
2778
1252
  DynamicPartitioningConfiguration?: DynamicPartitioningConfiguration;
2779
1253
  }
2780
1254
  export declare namespace ExtendedS3DestinationUpdate {
2781
- /**
2782
- * @internal
2783
- */
1255
+
2784
1256
  const filterSensitiveLog: (obj: ExtendedS3DestinationUpdate) => any;
2785
1257
  }
2786
1258
  export interface ListDeliveryStreamsInput {
2787
- /**
2788
- * <p>The maximum number of delivery streams to list. The default value is 10.</p>
2789
- */
1259
+
2790
1260
  Limit?: number;
2791
- /**
2792
- * <p>The delivery stream type. This can be one of the following values:</p>
2793
- * <ul>
2794
- * <li>
2795
- * <p>
2796
- * <code>DirectPut</code>: Provider applications access the delivery stream
2797
- * directly.</p>
2798
- * </li>
2799
- * <li>
2800
- * <p>
2801
- * <code>KinesisStreamAsSource</code>: The delivery stream uses a Kinesis data
2802
- * stream as a source.</p>
2803
- * </li>
2804
- * </ul>
2805
- * <p>This parameter is optional. If this parameter is omitted, delivery streams of all
2806
- * types are returned.</p>
2807
- */
1261
+
2808
1262
  DeliveryStreamType?: DeliveryStreamType | string;
2809
- /**
2810
- * <p>The list of delivery streams returned by this call to
2811
- * <code>ListDeliveryStreams</code> will start with the delivery stream whose name comes
2812
- * alphabetically immediately after the name you specify in
2813
- * <code>ExclusiveStartDeliveryStreamName</code>.</p>
2814
- */
1263
+
2815
1264
  ExclusiveStartDeliveryStreamName?: string;
2816
1265
  }
2817
1266
  export declare namespace ListDeliveryStreamsInput {
2818
- /**
2819
- * @internal
2820
- */
1267
+
2821
1268
  const filterSensitiveLog: (obj: ListDeliveryStreamsInput) => any;
2822
1269
  }
2823
1270
  export interface ListDeliveryStreamsOutput {
2824
- /**
2825
- * <p>The names of the delivery streams.</p>
2826
- */
1271
+
2827
1272
  DeliveryStreamNames: string[] | undefined;
2828
- /**
2829
- * <p>Indicates whether there are more delivery streams available to list.</p>
2830
- */
1273
+
2831
1274
  HasMoreDeliveryStreams: boolean | undefined;
2832
1275
  }
2833
1276
  export declare namespace ListDeliveryStreamsOutput {
2834
- /**
2835
- * @internal
2836
- */
1277
+
2837
1278
  const filterSensitiveLog: (obj: ListDeliveryStreamsOutput) => any;
2838
1279
  }
2839
1280
  export interface ListTagsForDeliveryStreamInput {
2840
- /**
2841
- * <p>The name of the delivery stream whose tags you want to list.</p>
2842
- */
1281
+
2843
1282
  DeliveryStreamName: string | undefined;
2844
- /**
2845
- * <p>The key to use as the starting point for the list of tags. If you set this parameter,
2846
- * <code>ListTagsForDeliveryStream</code> gets all tags that occur after
2847
- * <code>ExclusiveStartTagKey</code>.</p>
2848
- */
1283
+
2849
1284
  ExclusiveStartTagKey?: string;
2850
- /**
2851
- * <p>The number of tags to return. If this number is less than the total number of tags
2852
- * associated with the delivery stream, <code>HasMoreTags</code> is set to <code>true</code>
2853
- * in the response. To list additional tags, set <code>ExclusiveStartTagKey</code> to the last
2854
- * key in the response. </p>
2855
- */
1285
+
2856
1286
  Limit?: number;
2857
1287
  }
2858
1288
  export declare namespace ListTagsForDeliveryStreamInput {
2859
- /**
2860
- * @internal
2861
- */
1289
+
2862
1290
  const filterSensitiveLog: (obj: ListTagsForDeliveryStreamInput) => any;
2863
1291
  }
2864
1292
  export interface ListTagsForDeliveryStreamOutput {
2865
- /**
2866
- * <p>A list of tags associated with <code>DeliveryStreamName</code>, starting with the
2867
- * first tag after <code>ExclusiveStartTagKey</code> and up to the specified
2868
- * <code>Limit</code>.</p>
2869
- */
1293
+
2870
1294
  Tags: Tag[] | undefined;
2871
- /**
2872
- * <p>If this is <code>true</code> in the response, more tags are available. To list the
2873
- * remaining tags, set <code>ExclusiveStartTagKey</code> to the key of the last tag returned
2874
- * and call <code>ListTagsForDeliveryStream</code> again.</p>
2875
- */
1295
+
2876
1296
  HasMoreTags: boolean | undefined;
2877
1297
  }
2878
1298
  export declare namespace ListTagsForDeliveryStreamOutput {
2879
- /**
2880
- * @internal
2881
- */
1299
+
2882
1300
  const filterSensitiveLog: (obj: ListTagsForDeliveryStreamOutput) => any;
2883
1301
  }
2884
- /**
2885
- * <p>The unit of data in a delivery stream.</p>
2886
- */
1302
+
2887
1303
  export interface _Record {
2888
- /**
2889
- * <p>The data blob, which is base64-encoded when the blob is serialized. The maximum size
2890
- * of the data blob, before base64-encoding, is 1,000 KiB.</p>
2891
- */
1304
+
2892
1305
  Data: Uint8Array | undefined;
2893
1306
  }
2894
1307
  export declare namespace _Record {
2895
- /**
2896
- * @internal
2897
- */
1308
+
2898
1309
  const filterSensitiveLog: (obj: _Record) => any;
2899
1310
  }
2900
1311
  export interface PutRecordInput {
2901
- /**
2902
- * <p>The name of the delivery stream.</p>
2903
- */
1312
+
2904
1313
  DeliveryStreamName: string | undefined;
2905
- /**
2906
- * <p>The record.</p>
2907
- */
1314
+
2908
1315
  Record: _Record | undefined;
2909
1316
  }
2910
1317
  export declare namespace PutRecordInput {
2911
- /**
2912
- * @internal
2913
- */
1318
+
2914
1319
  const filterSensitiveLog: (obj: PutRecordInput) => any;
2915
1320
  }
2916
1321
  export interface PutRecordOutput {
2917
- /**
2918
- * <p>The ID of the record.</p>
2919
- */
1322
+
2920
1323
  RecordId: string | undefined;
2921
- /**
2922
- * <p>Indicates whether server-side encryption (SSE) was enabled during this operation.</p>
2923
- */
1324
+
2924
1325
  Encrypted?: boolean;
2925
1326
  }
2926
1327
  export declare namespace PutRecordOutput {
2927
- /**
2928
- * @internal
2929
- */
1328
+
2930
1329
  const filterSensitiveLog: (obj: PutRecordOutput) => any;
2931
1330
  }
2932
- /**
2933
- * <p>The service is unavailable. Back off and retry the operation. If you continue to see
2934
- * the exception, throughput limits for the delivery stream may have been exceeded. For more
2935
- * information about limits and how to request an increase, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/limits.html">Amazon Kinesis Data Firehose
2936
- * Limits</a>.</p>
2937
- */
1331
+
2938
1332
  export interface ServiceUnavailableException extends __SmithyException, $MetadataBearer {
2939
1333
  name: "ServiceUnavailableException";
2940
1334
  $fault: "server";
2941
- /**
2942
- * <p>A message that provides information about the error.</p>
2943
- */
1335
+
2944
1336
  message?: string;
2945
1337
  }
2946
1338
  export declare namespace ServiceUnavailableException {
2947
- /**
2948
- * @internal
2949
- */
1339
+
2950
1340
  const filterSensitiveLog: (obj: ServiceUnavailableException) => any;
2951
1341
  }
2952
1342
  export interface PutRecordBatchInput {
2953
- /**
2954
- * <p>The name of the delivery stream.</p>
2955
- */
1343
+
2956
1344
  DeliveryStreamName: string | undefined;
2957
- /**
2958
- * <p>One or more records.</p>
2959
- */
1345
+
2960
1346
  Records: _Record[] | undefined;
2961
1347
  }
2962
1348
  export declare namespace PutRecordBatchInput {
2963
- /**
2964
- * @internal
2965
- */
1349
+
2966
1350
  const filterSensitiveLog: (obj: PutRecordBatchInput) => any;
2967
1351
  }
2968
- /**
2969
- * <p>Contains the result for an individual record from a <a>PutRecordBatch</a>
2970
- * request. If the record is successfully added to your delivery stream, it receives a record
2971
- * ID. If the record fails to be added to your delivery stream, the result includes an error
2972
- * code and an error message.</p>
2973
- */
1352
+
2974
1353
  export interface PutRecordBatchResponseEntry {
2975
- /**
2976
- * <p>The ID of the record.</p>
2977
- */
1354
+
2978
1355
  RecordId?: string;
2979
- /**
2980
- * <p>The error code for an individual record result.</p>
2981
- */
1356
+
2982
1357
  ErrorCode?: string;
2983
- /**
2984
- * <p>The error message for an individual record result.</p>
2985
- */
1358
+
2986
1359
  ErrorMessage?: string;
2987
1360
  }
2988
1361
  export declare namespace PutRecordBatchResponseEntry {
2989
- /**
2990
- * @internal
2991
- */
1362
+
2992
1363
  const filterSensitiveLog: (obj: PutRecordBatchResponseEntry) => any;
2993
1364
  }
2994
1365
  export interface PutRecordBatchOutput {
2995
- /**
2996
- * <p>The number of records that might have failed processing. This number might be greater
2997
- * than 0 even if the <a>PutRecordBatch</a> call succeeds. Check
2998
- * <code>FailedPutCount</code> to determine whether there are records that you need to
2999
- * resend.</p>
3000
- */
1366
+
3001
1367
  FailedPutCount: number | undefined;
3002
- /**
3003
- * <p>Indicates whether server-side encryption (SSE) was enabled during this operation.</p>
3004
- */
1368
+
3005
1369
  Encrypted?: boolean;
3006
- /**
3007
- * <p>The results array. For each record, the index of the response element is the same as
3008
- * the index used in the request array.</p>
3009
- */
1370
+
3010
1371
  RequestResponses: PutRecordBatchResponseEntry[] | undefined;
3011
1372
  }
3012
1373
  export declare namespace PutRecordBatchOutput {
3013
- /**
3014
- * @internal
3015
- */
1374
+
3016
1375
  const filterSensitiveLog: (obj: PutRecordBatchOutput) => any;
3017
1376
  }
3018
1377
  export interface StartDeliveryStreamEncryptionInput {
3019
- /**
3020
- * <p>The name of the delivery stream for which you want to enable server-side encryption
3021
- * (SSE).</p>
3022
- */
1378
+
3023
1379
  DeliveryStreamName: string | undefined;
3024
- /**
3025
- * <p>Used to specify the type and Amazon Resource Name (ARN) of the KMS key needed for
3026
- * Server-Side Encryption (SSE).</p>
3027
- */
1380
+
3028
1381
  DeliveryStreamEncryptionConfigurationInput?: DeliveryStreamEncryptionConfigurationInput;
3029
1382
  }
3030
1383
  export declare namespace StartDeliveryStreamEncryptionInput {
3031
- /**
3032
- * @internal
3033
- */
1384
+
3034
1385
  const filterSensitiveLog: (obj: StartDeliveryStreamEncryptionInput) => any;
3035
1386
  }
3036
1387
  export interface StartDeliveryStreamEncryptionOutput {
3037
1388
  }
3038
1389
  export declare namespace StartDeliveryStreamEncryptionOutput {
3039
- /**
3040
- * @internal
3041
- */
1390
+
3042
1391
  const filterSensitiveLog: (obj: StartDeliveryStreamEncryptionOutput) => any;
3043
1392
  }
3044
1393
  export interface StopDeliveryStreamEncryptionInput {
3045
- /**
3046
- * <p>The name of the delivery stream for which you want to disable server-side encryption
3047
- * (SSE).</p>
3048
- */
1394
+
3049
1395
  DeliveryStreamName: string | undefined;
3050
1396
  }
3051
1397
  export declare namespace StopDeliveryStreamEncryptionInput {
3052
- /**
3053
- * @internal
3054
- */
1398
+
3055
1399
  const filterSensitiveLog: (obj: StopDeliveryStreamEncryptionInput) => any;
3056
1400
  }
3057
1401
  export interface StopDeliveryStreamEncryptionOutput {
3058
1402
  }
3059
1403
  export declare namespace StopDeliveryStreamEncryptionOutput {
3060
- /**
3061
- * @internal
3062
- */
1404
+
3063
1405
  const filterSensitiveLog: (obj: StopDeliveryStreamEncryptionOutput) => any;
3064
1406
  }
3065
1407
  export interface TagDeliveryStreamInput {
3066
- /**
3067
- * <p>The name of the delivery stream to which you want to add the tags.</p>
3068
- */
1408
+
3069
1409
  DeliveryStreamName: string | undefined;
3070
- /**
3071
- * <p>A set of key-value pairs to use to create the tags.</p>
3072
- */
1410
+
3073
1411
  Tags: Tag[] | undefined;
3074
1412
  }
3075
1413
  export declare namespace TagDeliveryStreamInput {
3076
- /**
3077
- * @internal
3078
- */
1414
+
3079
1415
  const filterSensitiveLog: (obj: TagDeliveryStreamInput) => any;
3080
1416
  }
3081
1417
  export interface TagDeliveryStreamOutput {
3082
1418
  }
3083
1419
  export declare namespace TagDeliveryStreamOutput {
3084
- /**
3085
- * @internal
3086
- */
1420
+
3087
1421
  const filterSensitiveLog: (obj: TagDeliveryStreamOutput) => any;
3088
1422
  }
3089
1423
  export interface UntagDeliveryStreamInput {
3090
- /**
3091
- * <p>The name of the delivery stream.</p>
3092
- */
1424
+
3093
1425
  DeliveryStreamName: string | undefined;
3094
- /**
3095
- * <p>A list of tag keys. Each corresponding tag is removed from the delivery
3096
- * stream.</p>
3097
- */
1426
+
3098
1427
  TagKeys: string[] | undefined;
3099
1428
  }
3100
1429
  export declare namespace UntagDeliveryStreamInput {
3101
- /**
3102
- * @internal
3103
- */
1430
+
3104
1431
  const filterSensitiveLog: (obj: UntagDeliveryStreamInput) => any;
3105
1432
  }
3106
1433
  export interface UntagDeliveryStreamOutput {
3107
1434
  }
3108
1435
  export declare namespace UntagDeliveryStreamOutput {
3109
- /**
3110
- * @internal
3111
- */
1436
+
3112
1437
  const filterSensitiveLog: (obj: UntagDeliveryStreamOutput) => any;
3113
1438
  }
3114
- /**
3115
- * <p>Updates the specified HTTP endpoint destination.</p>
3116
- */
1439
+
3117
1440
  export interface HttpEndpointDestinationUpdate {
3118
- /**
3119
- * <p>Describes the configuration of the HTTP endpoint destination.</p>
3120
- */
1441
+
3121
1442
  EndpointConfiguration?: HttpEndpointConfiguration;
3122
- /**
3123
- * <p>Describes buffering options that can be applied to the data before it is delivered to
3124
- * the HTTPS endpoint destination. Kinesis Data Firehose teats these options as hints, and it
3125
- * might choose to use more optimal values. The <code>SizeInMBs</code> and
3126
- * <code>IntervalInSeconds</code> parameters are optional. However, if specify a value for
3127
- * one of them, you must also provide a value for the other. </p>
3128
- */
1443
+
3129
1444
  BufferingHints?: HttpEndpointBufferingHints;
3130
- /**
3131
- * <p>Describes the Amazon CloudWatch logging options for your delivery stream.</p>
3132
- */
1445
+
3133
1446
  CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
3134
- /**
3135
- * <p>The configuration of the request sent to the HTTP endpoint specified as the
3136
- * destination.</p>
3137
- */
1447
+
3138
1448
  RequestConfiguration?: HttpEndpointRequestConfiguration;
3139
- /**
3140
- * <p>Describes a data processing configuration.</p>
3141
- */
1449
+
3142
1450
  ProcessingConfiguration?: ProcessingConfiguration;
3143
- /**
3144
- * <p>Kinesis Data Firehose uses this IAM role for all the permissions that the delivery
3145
- * stream needs.</p>
3146
- */
1451
+
3147
1452
  RoleARN?: string;
3148
- /**
3149
- * <p>Describes the retry behavior in case Kinesis Data Firehose is unable to deliver data to
3150
- * the specified HTTP endpoint destination, or if it doesn't receive a valid acknowledgment of
3151
- * receipt from the specified HTTP endpoint destination.</p>
3152
- */
1453
+
3153
1454
  RetryOptions?: HttpEndpointRetryOptions;
3154
- /**
3155
- * <p>Describes the S3 bucket backup options for the data that Kinesis Firehose delivers to
3156
- * the HTTP endpoint destination. You can back up all documents (<code>AllData</code>) or only
3157
- * the documents that Kinesis Data Firehose could not deliver to the specified HTTP endpoint
3158
- * destination (<code>FailedDataOnly</code>).</p>
3159
- */
1455
+
3160
1456
  S3BackupMode?: HttpEndpointS3BackupMode | string;
3161
- /**
3162
- * <p>Describes an update for a destination in Amazon S3.</p>
3163
- */
1457
+
3164
1458
  S3Update?: S3DestinationUpdate;
3165
1459
  }
3166
1460
  export declare namespace HttpEndpointDestinationUpdate {
3167
- /**
3168
- * @internal
3169
- */
1461
+
3170
1462
  const filterSensitiveLog: (obj: HttpEndpointDestinationUpdate) => any;
3171
1463
  }
3172
- /**
3173
- * <p>Describes an update for a destination in Amazon Redshift.</p>
3174
- */
1464
+
3175
1465
  export interface RedshiftDestinationUpdate {
3176
- /**
3177
- * <p>The Amazon Resource Name (ARN) of the AWS credentials. For more information, see
3178
- * <a href="https://docs.aws.amazon.com/general/latest/gr/aws-arns-and-namespaces.html">Amazon
3179
- * Resource Names (ARNs) and AWS Service Namespaces</a>.</p>
3180
- */
1466
+
3181
1467
  RoleARN?: string;
3182
- /**
3183
- * <p>The database connection string.</p>
3184
- */
1468
+
3185
1469
  ClusterJDBCURL?: string;
3186
- /**
3187
- * <p>The <code>COPY</code> command.</p>
3188
- */
1470
+
3189
1471
  CopyCommand?: CopyCommand;
3190
- /**
3191
- * <p>The name of the user.</p>
3192
- */
1472
+
3193
1473
  Username?: string;
3194
- /**
3195
- * <p>The user password.</p>
3196
- */
1474
+
3197
1475
  Password?: string;
3198
- /**
3199
- * <p>The retry behavior in case Kinesis Data Firehose is unable to deliver documents to
3200
- * Amazon Redshift. Default value is 3600 (60 minutes).</p>
3201
- */
1476
+
3202
1477
  RetryOptions?: RedshiftRetryOptions;
3203
- /**
3204
- * <p>The Amazon S3 destination.</p>
3205
- * <p>The compression formats <code>SNAPPY</code> or <code>ZIP</code> cannot be specified
3206
- * in <code>RedshiftDestinationUpdate.S3Update</code> because the Amazon Redshift
3207
- * <code>COPY</code> operation that reads from the S3 bucket doesn't support these
3208
- * compression formats.</p>
3209
- */
1478
+
3210
1479
  S3Update?: S3DestinationUpdate;
3211
- /**
3212
- * <p>The data processing configuration.</p>
3213
- */
1480
+
3214
1481
  ProcessingConfiguration?: ProcessingConfiguration;
3215
- /**
3216
- * <p>You can update a delivery stream to enable Amazon S3 backup if it is disabled. If
3217
- * backup is enabled, you can't update the delivery stream to disable it. </p>
3218
- */
1482
+
3219
1483
  S3BackupMode?: RedshiftS3BackupMode | string;
3220
- /**
3221
- * <p>The Amazon S3 destination for backup.</p>
3222
- */
1484
+
3223
1485
  S3BackupUpdate?: S3DestinationUpdate;
3224
- /**
3225
- * <p>The Amazon CloudWatch logging options for your delivery stream.</p>
3226
- */
1486
+
3227
1487
  CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
3228
1488
  }
3229
1489
  export declare namespace RedshiftDestinationUpdate {
3230
- /**
3231
- * @internal
3232
- */
1490
+
3233
1491
  const filterSensitiveLog: (obj: RedshiftDestinationUpdate) => any;
3234
1492
  }
3235
- /**
3236
- * <p>Describes an update for a destination in Splunk.</p>
3237
- */
1493
+
3238
1494
  export interface SplunkDestinationUpdate {
3239
- /**
3240
- * <p>The HTTP Event Collector (HEC) endpoint to which Kinesis Data Firehose sends your
3241
- * data.</p>
3242
- */
1495
+
3243
1496
  HECEndpoint?: string;
3244
- /**
3245
- * <p>This type can be either "Raw" or "Event."</p>
3246
- */
1497
+
3247
1498
  HECEndpointType?: HECEndpointType | string;
3248
- /**
3249
- * <p>A GUID that you obtain from your Splunk cluster when you create a new HEC
3250
- * endpoint.</p>
3251
- */
1499
+
3252
1500
  HECToken?: string;
3253
- /**
3254
- * <p>The amount of time that Kinesis Data Firehose waits to receive an acknowledgment from
3255
- * Splunk after it sends data. At the end of the timeout period, Kinesis Data Firehose either
3256
- * tries to send the data again or considers it an error, based on your retry
3257
- * settings.</p>
3258
- */
1501
+
3259
1502
  HECAcknowledgmentTimeoutInSeconds?: number;
3260
- /**
3261
- * <p>The retry behavior in case Kinesis Data Firehose is unable to deliver data to Splunk
3262
- * or if it doesn't receive an acknowledgment of receipt from Splunk.</p>
3263
- */
1503
+
3264
1504
  RetryOptions?: SplunkRetryOptions;
3265
- /**
3266
- * <p>Specifies how you want Kinesis Data Firehose to back up documents to Amazon S3. When
3267
- * set to <code>FailedDocumentsOnly</code>, Kinesis Data Firehose writes any data that could
3268
- * not be indexed to the configured Amazon S3 destination. When set to <code>AllEvents</code>,
3269
- * Kinesis Data Firehose delivers all incoming records to Amazon S3, and also writes failed
3270
- * documents to Amazon S3. The default value is <code>FailedEventsOnly</code>.</p>
3271
- * <p>You can update this backup mode from <code>FailedEventsOnly</code> to
3272
- * <code>AllEvents</code>. You can't update it from <code>AllEvents</code> to
3273
- * <code>FailedEventsOnly</code>.</p>
3274
- */
1505
+
3275
1506
  S3BackupMode?: SplunkS3BackupMode | string;
3276
- /**
3277
- * <p>Your update to the configuration of the backup Amazon S3 location.</p>
3278
- */
1507
+
3279
1508
  S3Update?: S3DestinationUpdate;
3280
- /**
3281
- * <p>The data processing configuration.</p>
3282
- */
1509
+
3283
1510
  ProcessingConfiguration?: ProcessingConfiguration;
3284
- /**
3285
- * <p>The Amazon CloudWatch logging options for your delivery stream.</p>
3286
- */
1511
+
3287
1512
  CloudWatchLoggingOptions?: CloudWatchLoggingOptions;
3288
1513
  }
3289
1514
  export declare namespace SplunkDestinationUpdate {
3290
- /**
3291
- * @internal
3292
- */
1515
+
3293
1516
  const filterSensitiveLog: (obj: SplunkDestinationUpdate) => any;
3294
1517
  }
3295
1518
  export interface UpdateDestinationInput {
3296
- /**
3297
- * <p>The name of the delivery stream.</p>
3298
- */
1519
+
3299
1520
  DeliveryStreamName: string | undefined;
3300
- /**
3301
- * <p>Obtain this value from the <code>VersionId</code> result of <a>DeliveryStreamDescription</a>. This value is required, and helps the service
3302
- * perform conditional operations. For example, if there is an interleaving update and this
3303
- * value is null, then the update destination fails. After the update is successful, the
3304
- * <code>VersionId</code> value is updated. The service then performs a merge of the old
3305
- * configuration with the new configuration.</p>
3306
- */
1521
+
3307
1522
  CurrentDeliveryStreamVersionId: string | undefined;
3308
- /**
3309
- * <p>The ID of the destination.</p>
3310
- */
1523
+
3311
1524
  DestinationId: string | undefined;
3312
- /**
3313
- * @deprecated
3314
- *
3315
- * <p>[Deprecated] Describes an update for a destination in Amazon S3.</p>
3316
- */
1525
+
3317
1526
  S3DestinationUpdate?: S3DestinationUpdate;
3318
- /**
3319
- * <p>Describes an update for a destination in Amazon S3.</p>
3320
- */
1527
+
3321
1528
  ExtendedS3DestinationUpdate?: ExtendedS3DestinationUpdate;
3322
- /**
3323
- * <p>Describes an update for a destination in Amazon Redshift.</p>
3324
- */
1529
+
3325
1530
  RedshiftDestinationUpdate?: RedshiftDestinationUpdate;
3326
- /**
3327
- * <p>Describes an update for a destination in Amazon ES.</p>
3328
- */
1531
+
3329
1532
  ElasticsearchDestinationUpdate?: ElasticsearchDestinationUpdate;
3330
1533
  AmazonopensearchserviceDestinationUpdate?: AmazonopensearchserviceDestinationUpdate;
3331
- /**
3332
- * <p>Describes an update for a destination in Splunk.</p>
3333
- */
1534
+
3334
1535
  SplunkDestinationUpdate?: SplunkDestinationUpdate;
3335
- /**
3336
- * <p>Describes an update to the specified HTTP endpoint destination.</p>
3337
- */
1536
+
3338
1537
  HttpEndpointDestinationUpdate?: HttpEndpointDestinationUpdate;
3339
1538
  }
3340
1539
  export declare namespace UpdateDestinationInput {
3341
- /**
3342
- * @internal
3343
- */
1540
+
3344
1541
  const filterSensitiveLog: (obj: UpdateDestinationInput) => any;
3345
1542
  }
3346
1543
  export interface UpdateDestinationOutput {
3347
1544
  }
3348
1545
  export declare namespace UpdateDestinationOutput {
3349
- /**
3350
- * @internal
3351
- */
1546
+
3352
1547
  const filterSensitiveLog: (obj: UpdateDestinationOutput) => any;
3353
1548
  }