@aws-sdk/client-firehose 3.36.0 → 3.39.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (51) hide show
  1. package/CHANGELOG.md +42 -0
  2. package/dist-cjs/Firehose.js +0 -6
  3. package/dist-cjs/FirehoseClient.js +0 -11
  4. package/dist-cjs/commands/CreateDeliveryStreamCommand.js +0 -92
  5. package/dist-cjs/commands/DeleteDeliveryStreamCommand.js +0 -33
  6. package/dist-cjs/commands/DescribeDeliveryStreamCommand.js +0 -30
  7. package/dist-cjs/commands/ListDeliveryStreamsCommand.js +0 -31
  8. package/dist-cjs/commands/ListTagsForDeliveryStreamCommand.js +0 -25
  9. package/dist-cjs/commands/PutRecordBatchCommand.js +0 -74
  10. package/dist-cjs/commands/PutRecordCommand.js +0 -55
  11. package/dist-cjs/commands/StartDeliveryStreamEncryptionCommand.js +0 -59
  12. package/dist-cjs/commands/StopDeliveryStreamEncryptionCommand.js +0 -43
  13. package/dist-cjs/commands/TagDeliveryStreamCommand.js +0 -32
  14. package/dist-cjs/commands/UntagDeliveryStreamCommand.js +0 -27
  15. package/dist-cjs/commands/UpdateDestinationCommand.js +0 -47
  16. package/dist-cjs/commands/index.js +15 -0
  17. package/dist-cjs/endpoints.js +6 -1
  18. package/dist-cjs/index.js +3 -14
  19. package/dist-cjs/models/models_0.js +0 -291
  20. package/dist-cjs/protocols/Aws_json1_1.js +0 -5
  21. package/dist-cjs/runtimeConfig.browser.js +1 -5
  22. package/dist-cjs/runtimeConfig.js +1 -5
  23. package/dist-cjs/runtimeConfig.native.js +0 -3
  24. package/dist-cjs/runtimeConfig.shared.js +0 -3
  25. package/dist-es/commands/index.js +12 -0
  26. package/dist-es/endpoints.js +6 -1
  27. package/dist-es/index.js +3 -14
  28. package/dist-types/commands/index.d.ts +12 -0
  29. package/dist-types/index.d.ts +3 -14
  30. package/dist-types/ts3.4/Firehose.d.ts +13 -302
  31. package/dist-types/ts3.4/FirehoseClient.d.ts +24 -92
  32. package/dist-types/ts3.4/commands/CreateDeliveryStreamCommand.d.ts +2 -88
  33. package/dist-types/ts3.4/commands/DeleteDeliveryStreamCommand.d.ts +2 -29
  34. package/dist-types/ts3.4/commands/DescribeDeliveryStreamCommand.d.ts +2 -26
  35. package/dist-types/ts3.4/commands/ListDeliveryStreamsCommand.d.ts +2 -27
  36. package/dist-types/ts3.4/commands/ListTagsForDeliveryStreamCommand.d.ts +2 -21
  37. package/dist-types/ts3.4/commands/PutRecordBatchCommand.d.ts +2 -70
  38. package/dist-types/ts3.4/commands/PutRecordCommand.d.ts +2 -51
  39. package/dist-types/ts3.4/commands/StartDeliveryStreamEncryptionCommand.d.ts +2 -55
  40. package/dist-types/ts3.4/commands/StopDeliveryStreamEncryptionCommand.d.ts +2 -39
  41. package/dist-types/ts3.4/commands/TagDeliveryStreamCommand.d.ts +2 -28
  42. package/dist-types/ts3.4/commands/UntagDeliveryStreamCommand.d.ts +2 -23
  43. package/dist-types/ts3.4/commands/UpdateDestinationCommand.d.ts +2 -43
  44. package/dist-types/ts3.4/commands/index.d.ts +12 -0
  45. package/dist-types/ts3.4/index.d.ts +3 -14
  46. package/dist-types/ts3.4/models/models_0.d.ts +540 -2345
  47. package/dist-types/ts3.4/runtimeConfig.browser.d.ts +1 -3
  48. package/dist-types/ts3.4/runtimeConfig.d.ts +1 -3
  49. package/dist-types/ts3.4/runtimeConfig.native.d.ts +1 -3
  50. package/dist-types/ts3.4/runtimeConfig.shared.d.ts +1 -3
  51. package/package.json +32 -32
@@ -21,127 +21,59 @@ import { UpdateDestinationCommandInput, UpdateDestinationCommandOutput } from ".
21
21
  export declare type ServiceInputTypes = CreateDeliveryStreamCommandInput | DeleteDeliveryStreamCommandInput | DescribeDeliveryStreamCommandInput | ListDeliveryStreamsCommandInput | ListTagsForDeliveryStreamCommandInput | PutRecordBatchCommandInput | PutRecordCommandInput | StartDeliveryStreamEncryptionCommandInput | StopDeliveryStreamEncryptionCommandInput | TagDeliveryStreamCommandInput | UntagDeliveryStreamCommandInput | UpdateDestinationCommandInput;
22
22
  export declare type ServiceOutputTypes = CreateDeliveryStreamCommandOutput | DeleteDeliveryStreamCommandOutput | DescribeDeliveryStreamCommandOutput | ListDeliveryStreamsCommandOutput | ListTagsForDeliveryStreamCommandOutput | PutRecordBatchCommandOutput | PutRecordCommandOutput | StartDeliveryStreamEncryptionCommandOutput | StopDeliveryStreamEncryptionCommandOutput | TagDeliveryStreamCommandOutput | UntagDeliveryStreamCommandOutput | UpdateDestinationCommandOutput;
23
23
  export interface ClientDefaults extends Partial<__SmithyResolvedConfiguration<__HttpHandlerOptions>> {
24
- /**
25
- * The HTTP handler to use. Fetch in browser and Https in Nodejs.
26
- */
24
+
27
25
  requestHandler?: __HttpHandler;
28
- /**
29
- * A constructor for a class implementing the {@link __Hash} interface
30
- * that computes the SHA-256 HMAC or checksum of a string or binary buffer.
31
- * @internal
32
- */
26
+
33
27
  sha256?: __HashConstructor;
34
- /**
35
- * The function that will be used to convert strings into HTTP endpoints.
36
- * @internal
37
- */
28
+
38
29
  urlParser?: __UrlParser;
39
- /**
40
- * A function that can calculate the length of a request body.
41
- * @internal
42
- */
30
+
43
31
  bodyLengthChecker?: (body: any) => number | undefined;
44
- /**
45
- * A function that converts a stream into an array of bytes.
46
- * @internal
47
- */
32
+
48
33
  streamCollector?: __StreamCollector;
49
- /**
50
- * The function that will be used to convert a base64-encoded string to a byte array.
51
- * @internal
52
- */
34
+
53
35
  base64Decoder?: __Decoder;
54
- /**
55
- * The function that will be used to convert binary data to a base64-encoded string.
56
- * @internal
57
- */
36
+
58
37
  base64Encoder?: __Encoder;
59
- /**
60
- * The function that will be used to convert a UTF8-encoded string to a byte array.
61
- * @internal
62
- */
38
+
63
39
  utf8Decoder?: __Decoder;
64
- /**
65
- * The function that will be used to convert binary data to a UTF-8 encoded string.
66
- * @internal
67
- */
40
+
68
41
  utf8Encoder?: __Encoder;
69
- /**
70
- * The runtime environment.
71
- * @internal
72
- */
42
+
73
43
  runtime?: string;
74
- /**
75
- * Disable dyanamically changing the endpoint of the client based on the hostPrefix
76
- * trait of an operation.
77
- */
44
+
78
45
  disableHostPrefix?: boolean;
79
- /**
80
- * Value for how many times a request will be made at most in case of retry.
81
- */
46
+
82
47
  maxAttempts?: number | __Provider<number>;
83
- /**
84
- * Specifies which retry algorithm to use.
85
- */
48
+
86
49
  retryMode?: string | __Provider<string>;
87
- /**
88
- * Optional logger for logging debug/info/warn/error.
89
- */
50
+
90
51
  logger?: __Logger;
91
- /**
92
- * Unique service identifier.
93
- * @internal
94
- */
52
+
95
53
  serviceId?: string;
96
- /**
97
- * The AWS region to which this client will send requests
98
- */
54
+
99
55
  region?: string | __Provider<string>;
100
- /**
101
- * Default credentials provider; Not available in browser runtime.
102
- * @internal
103
- */
56
+
104
57
  credentialDefaultProvider?: (input: any) => __Provider<__Credentials>;
105
- /**
106
- * Fetch related hostname, signing name or signing region with given region.
107
- * @internal
108
- */
58
+
109
59
  regionInfoProvider?: RegionInfoProvider;
110
- /**
111
- * The provider populating default tracking information to be sent with `user-agent`, `x-amz-user-agent` header
112
- * @internal
113
- */
60
+
114
61
  defaultUserAgentProvider?: Provider<__UserAgent>;
115
62
  }
116
63
  declare type FirehoseClientConfigType = Partial<__SmithyConfiguration<__HttpHandlerOptions>> & ClientDefaults & RegionInputConfig & EndpointsInputConfig & RetryInputConfig & HostHeaderInputConfig & AwsAuthInputConfig & UserAgentInputConfig;
117
- /**
118
- * The configuration interface of FirehoseClient class constructor that set the region, credentials and other options.
119
- */
64
+
120
65
  export interface FirehoseClientConfig extends FirehoseClientConfigType {
121
66
  }
122
67
  declare type FirehoseClientResolvedConfigType = __SmithyResolvedConfiguration<__HttpHandlerOptions> & Required<ClientDefaults> & RegionResolvedConfig & EndpointsResolvedConfig & RetryResolvedConfig & HostHeaderResolvedConfig & AwsAuthResolvedConfig & UserAgentResolvedConfig;
123
- /**
124
- * The resolved configuration interface of FirehoseClient class. This is resolved and normalized from the {@link FirehoseClientConfig | constructor configuration interface}.
125
- */
68
+
126
69
  export interface FirehoseClientResolvedConfig extends FirehoseClientResolvedConfigType {
127
70
  }
128
- /**
129
- * <fullname>Amazon Kinesis Data Firehose API Reference</fullname>
130
- * <p>Amazon Kinesis Data Firehose is a fully managed service that delivers real-time
131
- * streaming data to destinations such as Amazon Simple Storage Service (Amazon S3), Amazon
132
- * Elasticsearch Service (Amazon ES), Amazon Redshift, and Splunk.</p>
133
- */
71
+
134
72
  export declare class FirehoseClient extends __Client<__HttpHandlerOptions, ServiceInputTypes, ServiceOutputTypes, FirehoseClientResolvedConfig> {
135
- /**
136
- * The resolved configuration of FirehoseClient class. This is resolved and normalized from the {@link FirehoseClientConfig | constructor configuration interface}.
137
- */
73
+
138
74
  readonly config: FirehoseClientResolvedConfig;
139
75
  constructor(configuration: FirehoseClientConfig);
140
- /**
141
- * Destroy underlying resources, like sockets. It's usually not necessary to do this.
142
- * However in Node.js, it's best to explicitly shut down the client's agent when it is no longer needed.
143
- * Otherwise, sockets might stay open for quite a long time before the server terminates them.
144
- */
76
+
145
77
  destroy(): void;
146
78
  }
147
79
  export {};
@@ -6,97 +6,11 @@ export interface CreateDeliveryStreamCommandInput extends CreateDeliveryStreamIn
6
6
  }
7
7
  export interface CreateDeliveryStreamCommandOutput extends CreateDeliveryStreamOutput, __MetadataBearer {
8
8
  }
9
- /**
10
- * <p>Creates a Kinesis Data Firehose delivery stream.</p>
11
- *
12
- * <p>By default, you can create up to 50 delivery streams per AWS Region.</p>
13
- * <p>This is an asynchronous operation that immediately returns. The initial status of the
14
- * delivery stream is <code>CREATING</code>. After the delivery stream is created, its status
15
- * is <code>ACTIVE</code> and it now accepts data. If the delivery stream creation fails, the
16
- * status transitions to <code>CREATING_FAILED</code>. Attempts to send data to a delivery
17
- * stream that is not in the <code>ACTIVE</code> state cause an exception. To check the state
18
- * of a delivery stream, use <a>DescribeDeliveryStream</a>.</p>
19
- * <p>If the status of a delivery stream is <code>CREATING_FAILED</code>, this status
20
- * doesn't change, and you can't invoke <code>CreateDeliveryStream</code> again on it.
21
- * However, you can invoke the <a>DeleteDeliveryStream</a> operation to delete
22
- * it.</p>
23
- * <p>A Kinesis Data Firehose delivery stream can be configured to receive records directly
24
- * from providers using <a>PutRecord</a> or <a>PutRecordBatch</a>, or it
25
- * can be configured to use an existing Kinesis stream as its source. To specify a Kinesis
26
- * data stream as input, set the <code>DeliveryStreamType</code> parameter to
27
- * <code>KinesisStreamAsSource</code>, and provide the Kinesis stream Amazon Resource Name
28
- * (ARN) and role ARN in the <code>KinesisStreamSourceConfiguration</code>
29
- * parameter.</p>
30
- * <p>To create a delivery stream with server-side encryption (SSE) enabled, include <a>DeliveryStreamEncryptionConfigurationInput</a> in your request. This is
31
- * optional. You can also invoke <a>StartDeliveryStreamEncryption</a> to turn on
32
- * SSE for an existing delivery stream that doesn't have SSE enabled.</p>
33
- * <p>A delivery stream is configured with a single destination: Amazon S3, Amazon ES,
34
- * Amazon Redshift, or Splunk. You must specify only one of the following destination
35
- * configuration parameters: <code>ExtendedS3DestinationConfiguration</code>,
36
- * <code>S3DestinationConfiguration</code>,
37
- * <code>ElasticsearchDestinationConfiguration</code>,
38
- * <code>RedshiftDestinationConfiguration</code>, or
39
- * <code>SplunkDestinationConfiguration</code>.</p>
40
- * <p>When you specify <code>S3DestinationConfiguration</code>, you can also provide the
41
- * following optional values: BufferingHints, <code>EncryptionConfiguration</code>, and
42
- * <code>CompressionFormat</code>. By default, if no <code>BufferingHints</code> value is
43
- * provided, Kinesis Data Firehose buffers data up to 5 MB or for 5 minutes, whichever
44
- * condition is satisfied first. <code>BufferingHints</code> is a hint, so there are some
45
- * cases where the service cannot adhere to these conditions strictly. For example, record
46
- * boundaries might be such that the size is a little over or under the configured buffering
47
- * size. By default, no encryption is performed. We strongly recommend that you enable
48
- * encryption to ensure secure data storage in Amazon S3.</p>
49
- *
50
- * <p>A few notes about Amazon Redshift as a destination:</p>
51
- * <ul>
52
- * <li>
53
- * <p>An Amazon Redshift destination requires an S3 bucket as intermediate location.
54
- * Kinesis Data Firehose first delivers data to Amazon S3 and then uses
55
- * <code>COPY</code> syntax to load data into an Amazon Redshift table. This is
56
- * specified in the <code>RedshiftDestinationConfiguration.S3Configuration</code>
57
- * parameter.</p>
58
- *
59
- * </li>
60
- * <li>
61
- * <p>The compression formats <code>SNAPPY</code> or <code>ZIP</code> cannot be
62
- * specified in <code>RedshiftDestinationConfiguration.S3Configuration</code> because
63
- * the Amazon Redshift <code>COPY</code> operation that reads from the S3 bucket doesn't
64
- * support these compression formats.</p>
65
- * </li>
66
- * <li>
67
- * <p>We strongly recommend that you use the user name and password you provide
68
- * exclusively with Kinesis Data Firehose, and that the permissions for the account are
69
- * restricted for Amazon Redshift <code>INSERT</code> permissions.</p>
70
- *
71
- * </li>
72
- * </ul>
73
- * <p>Kinesis Data Firehose assumes the IAM role that is configured as part of the
74
- * destination. The role should allow the Kinesis Data Firehose principal to assume the role,
75
- * and the role should have permissions that allow the service to deliver the data. For more
76
- * information, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/controlling-access.html#using-iam-s3">Grant Kinesis Data
77
- * Firehose Access to an Amazon S3 Destination</a> in the <i>Amazon Kinesis Data
78
- * Firehose Developer Guide</i>.</p>
79
- * @example
80
- * Use a bare-bones client and the command you need to make an API call.
81
- * ```javascript
82
- * import { FirehoseClient, CreateDeliveryStreamCommand } from "@aws-sdk/client-firehose"; // ES Modules import
83
- * // const { FirehoseClient, CreateDeliveryStreamCommand } = require("@aws-sdk/client-firehose"); // CommonJS import
84
- * const client = new FirehoseClient(config);
85
- * const command = new CreateDeliveryStreamCommand(input);
86
- * const response = await client.send(command);
87
- * ```
88
- *
89
- * @see {@link CreateDeliveryStreamCommandInput} for command's `input` shape.
90
- * @see {@link CreateDeliveryStreamCommandOutput} for command's `response` shape.
91
- * @see {@link FirehoseClientResolvedConfig | config} for command's `input` shape.
92
- *
93
- */
9
+
94
10
  export declare class CreateDeliveryStreamCommand extends $Command<CreateDeliveryStreamCommandInput, CreateDeliveryStreamCommandOutput, FirehoseClientResolvedConfig> {
95
11
  readonly input: CreateDeliveryStreamCommandInput;
96
12
  constructor(input: CreateDeliveryStreamCommandInput);
97
- /**
98
- * @internal
99
- */
13
+
100
14
  resolveMiddleware(clientStack: MiddlewareStack<ServiceInputTypes, ServiceOutputTypes>, configuration: FirehoseClientResolvedConfig, options?: __HttpHandlerOptions): Handler<CreateDeliveryStreamCommandInput, CreateDeliveryStreamCommandOutput>;
101
15
  private serialize;
102
16
  private deserialize;
@@ -6,38 +6,11 @@ export interface DeleteDeliveryStreamCommandInput extends DeleteDeliveryStreamIn
6
6
  }
7
7
  export interface DeleteDeliveryStreamCommandOutput extends DeleteDeliveryStreamOutput, __MetadataBearer {
8
8
  }
9
- /**
10
- * <p>Deletes a delivery stream and its data.</p>
11
- * <p>To check the state of a delivery stream, use <a>DescribeDeliveryStream</a>. You can delete a delivery stream only if it is in one of the following states:
12
- * <code>ACTIVE</code>, <code>DELETING</code>, <code>CREATING_FAILED</code>, or
13
- * <code>DELETING_FAILED</code>. You can't delete a delivery stream that is in the
14
- * <code>CREATING</code> state. While the deletion request is in process, the delivery
15
- * stream is in the <code>DELETING</code> state.</p>
16
- * <p>While the delivery stream is in the <code>DELETING</code> state, the service might
17
- * continue to accept records, but it doesn't make any guarantees with respect to delivering
18
- * the data. Therefore, as a best practice, first stop any applications that are sending
19
- * records before you delete a delivery stream.</p>
20
- * @example
21
- * Use a bare-bones client and the command you need to make an API call.
22
- * ```javascript
23
- * import { FirehoseClient, DeleteDeliveryStreamCommand } from "@aws-sdk/client-firehose"; // ES Modules import
24
- * // const { FirehoseClient, DeleteDeliveryStreamCommand } = require("@aws-sdk/client-firehose"); // CommonJS import
25
- * const client = new FirehoseClient(config);
26
- * const command = new DeleteDeliveryStreamCommand(input);
27
- * const response = await client.send(command);
28
- * ```
29
- *
30
- * @see {@link DeleteDeliveryStreamCommandInput} for command's `input` shape.
31
- * @see {@link DeleteDeliveryStreamCommandOutput} for command's `response` shape.
32
- * @see {@link FirehoseClientResolvedConfig | config} for command's `input` shape.
33
- *
34
- */
9
+
35
10
  export declare class DeleteDeliveryStreamCommand extends $Command<DeleteDeliveryStreamCommandInput, DeleteDeliveryStreamCommandOutput, FirehoseClientResolvedConfig> {
36
11
  readonly input: DeleteDeliveryStreamCommandInput;
37
12
  constructor(input: DeleteDeliveryStreamCommandInput);
38
- /**
39
- * @internal
40
- */
13
+
41
14
  resolveMiddleware(clientStack: MiddlewareStack<ServiceInputTypes, ServiceOutputTypes>, configuration: FirehoseClientResolvedConfig, options?: __HttpHandlerOptions): Handler<DeleteDeliveryStreamCommandInput, DeleteDeliveryStreamCommandOutput>;
42
15
  private serialize;
43
16
  private deserialize;
@@ -6,35 +6,11 @@ export interface DescribeDeliveryStreamCommandInput extends DescribeDeliveryStre
6
6
  }
7
7
  export interface DescribeDeliveryStreamCommandOutput extends DescribeDeliveryStreamOutput, __MetadataBearer {
8
8
  }
9
- /**
10
- * <p>Describes the specified delivery stream and its status. For example, after your
11
- * delivery stream is created, call <code>DescribeDeliveryStream</code> to see whether the
12
- * delivery stream is <code>ACTIVE</code> and therefore ready for data to be sent to it. </p>
13
- * <p>If the status of a delivery stream is <code>CREATING_FAILED</code>, this status
14
- * doesn't change, and you can't invoke <a>CreateDeliveryStream</a> again on it.
15
- * However, you can invoke the <a>DeleteDeliveryStream</a> operation to delete it.
16
- * If the status is <code>DELETING_FAILED</code>, you can force deletion by invoking <a>DeleteDeliveryStream</a> again but with <a>DeleteDeliveryStreamInput$AllowForceDelete</a> set to true.</p>
17
- * @example
18
- * Use a bare-bones client and the command you need to make an API call.
19
- * ```javascript
20
- * import { FirehoseClient, DescribeDeliveryStreamCommand } from "@aws-sdk/client-firehose"; // ES Modules import
21
- * // const { FirehoseClient, DescribeDeliveryStreamCommand } = require("@aws-sdk/client-firehose"); // CommonJS import
22
- * const client = new FirehoseClient(config);
23
- * const command = new DescribeDeliveryStreamCommand(input);
24
- * const response = await client.send(command);
25
- * ```
26
- *
27
- * @see {@link DescribeDeliveryStreamCommandInput} for command's `input` shape.
28
- * @see {@link DescribeDeliveryStreamCommandOutput} for command's `response` shape.
29
- * @see {@link FirehoseClientResolvedConfig | config} for command's `input` shape.
30
- *
31
- */
9
+
32
10
  export declare class DescribeDeliveryStreamCommand extends $Command<DescribeDeliveryStreamCommandInput, DescribeDeliveryStreamCommandOutput, FirehoseClientResolvedConfig> {
33
11
  readonly input: DescribeDeliveryStreamCommandInput;
34
12
  constructor(input: DescribeDeliveryStreamCommandInput);
35
- /**
36
- * @internal
37
- */
13
+
38
14
  resolveMiddleware(clientStack: MiddlewareStack<ServiceInputTypes, ServiceOutputTypes>, configuration: FirehoseClientResolvedConfig, options?: __HttpHandlerOptions): Handler<DescribeDeliveryStreamCommandInput, DescribeDeliveryStreamCommandOutput>;
39
15
  private serialize;
40
16
  private deserialize;
@@ -6,36 +6,11 @@ export interface ListDeliveryStreamsCommandInput extends ListDeliveryStreamsInpu
6
6
  }
7
7
  export interface ListDeliveryStreamsCommandOutput extends ListDeliveryStreamsOutput, __MetadataBearer {
8
8
  }
9
- /**
10
- * <p>Lists your delivery streams in alphabetical order of their names.</p>
11
- * <p>The number of delivery streams might be too large to return using a single call to
12
- * <code>ListDeliveryStreams</code>. You can limit the number of delivery streams returned,
13
- * using the <code>Limit</code> parameter. To determine whether there are more delivery
14
- * streams to list, check the value of <code>HasMoreDeliveryStreams</code> in the output. If
15
- * there are more delivery streams to list, you can request them by calling this operation
16
- * again and setting the <code>ExclusiveStartDeliveryStreamName</code> parameter to the name
17
- * of the last delivery stream returned in the last call.</p>
18
- * @example
19
- * Use a bare-bones client and the command you need to make an API call.
20
- * ```javascript
21
- * import { FirehoseClient, ListDeliveryStreamsCommand } from "@aws-sdk/client-firehose"; // ES Modules import
22
- * // const { FirehoseClient, ListDeliveryStreamsCommand } = require("@aws-sdk/client-firehose"); // CommonJS import
23
- * const client = new FirehoseClient(config);
24
- * const command = new ListDeliveryStreamsCommand(input);
25
- * const response = await client.send(command);
26
- * ```
27
- *
28
- * @see {@link ListDeliveryStreamsCommandInput} for command's `input` shape.
29
- * @see {@link ListDeliveryStreamsCommandOutput} for command's `response` shape.
30
- * @see {@link FirehoseClientResolvedConfig | config} for command's `input` shape.
31
- *
32
- */
9
+
33
10
  export declare class ListDeliveryStreamsCommand extends $Command<ListDeliveryStreamsCommandInput, ListDeliveryStreamsCommandOutput, FirehoseClientResolvedConfig> {
34
11
  readonly input: ListDeliveryStreamsCommandInput;
35
12
  constructor(input: ListDeliveryStreamsCommandInput);
36
- /**
37
- * @internal
38
- */
13
+
39
14
  resolveMiddleware(clientStack: MiddlewareStack<ServiceInputTypes, ServiceOutputTypes>, configuration: FirehoseClientResolvedConfig, options?: __HttpHandlerOptions): Handler<ListDeliveryStreamsCommandInput, ListDeliveryStreamsCommandOutput>;
40
15
  private serialize;
41
16
  private deserialize;
@@ -6,30 +6,11 @@ export interface ListTagsForDeliveryStreamCommandInput extends ListTagsForDelive
6
6
  }
7
7
  export interface ListTagsForDeliveryStreamCommandOutput extends ListTagsForDeliveryStreamOutput, __MetadataBearer {
8
8
  }
9
- /**
10
- * <p>Lists the tags for the specified delivery stream. This operation has a limit of five
11
- * transactions per second per account. </p>
12
- * @example
13
- * Use a bare-bones client and the command you need to make an API call.
14
- * ```javascript
15
- * import { FirehoseClient, ListTagsForDeliveryStreamCommand } from "@aws-sdk/client-firehose"; // ES Modules import
16
- * // const { FirehoseClient, ListTagsForDeliveryStreamCommand } = require("@aws-sdk/client-firehose"); // CommonJS import
17
- * const client = new FirehoseClient(config);
18
- * const command = new ListTagsForDeliveryStreamCommand(input);
19
- * const response = await client.send(command);
20
- * ```
21
- *
22
- * @see {@link ListTagsForDeliveryStreamCommandInput} for command's `input` shape.
23
- * @see {@link ListTagsForDeliveryStreamCommandOutput} for command's `response` shape.
24
- * @see {@link FirehoseClientResolvedConfig | config} for command's `input` shape.
25
- *
26
- */
9
+
27
10
  export declare class ListTagsForDeliveryStreamCommand extends $Command<ListTagsForDeliveryStreamCommandInput, ListTagsForDeliveryStreamCommandOutput, FirehoseClientResolvedConfig> {
28
11
  readonly input: ListTagsForDeliveryStreamCommandInput;
29
12
  constructor(input: ListTagsForDeliveryStreamCommandInput);
30
- /**
31
- * @internal
32
- */
13
+
33
14
  resolveMiddleware(clientStack: MiddlewareStack<ServiceInputTypes, ServiceOutputTypes>, configuration: FirehoseClientResolvedConfig, options?: __HttpHandlerOptions): Handler<ListTagsForDeliveryStreamCommandInput, ListTagsForDeliveryStreamCommandOutput>;
34
15
  private serialize;
35
16
  private deserialize;
@@ -6,79 +6,11 @@ export interface PutRecordBatchCommandInput extends PutRecordBatchInput {
6
6
  }
7
7
  export interface PutRecordBatchCommandOutput extends PutRecordBatchOutput, __MetadataBearer {
8
8
  }
9
- /**
10
- * <p>Writes multiple data records into a delivery stream in a single call, which can
11
- * achieve higher throughput per producer than when writing single records. To write single
12
- * data records into a delivery stream, use <a>PutRecord</a>. Applications using
13
- * these operations are referred to as producers.</p>
14
- * <p>For information about service quota, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/limits.html">Amazon Kinesis Data Firehose
15
- * Quota</a>.</p>
16
- * <p>Each <a>PutRecordBatch</a> request supports up to 500 records. Each record
17
- * in the request can be as large as 1,000 KB (before base64 encoding), up to a limit of 4 MB
18
- * for the entire request. These limits cannot be changed.</p>
19
- * <p>You must specify the name of the delivery stream and the data record when using <a>PutRecord</a>. The data record consists of a data blob that can be up to 1,000
20
- * KB in size, and any kind of data. For example, it could be a segment from a log file,
21
- * geographic location data, website clickstream data, and so on.</p>
22
- * <p>Kinesis Data Firehose buffers records before delivering them to the destination. To
23
- * disambiguate the data blobs at the destination, a common solution is to use delimiters in
24
- * the data, such as a newline (<code>\n</code>) or some other character unique within the
25
- * data. This allows the consumer application to parse individual data items when reading the
26
- * data from the destination.</p>
27
- * <p>The <a>PutRecordBatch</a> response includes a count of failed records,
28
- * <code>FailedPutCount</code>, and an array of responses, <code>RequestResponses</code>.
29
- * Even if the <a>PutRecordBatch</a> call succeeds, the value of
30
- * <code>FailedPutCount</code> may be greater than 0, indicating that there are records for
31
- * which the operation didn't succeed. Each entry in the <code>RequestResponses</code> array
32
- * provides additional information about the processed record. It directly correlates with a
33
- * record in the request array using the same ordering, from the top to the bottom. The
34
- * response array always includes the same number of records as the request array.
35
- * <code>RequestResponses</code> includes both successfully and unsuccessfully processed
36
- * records. Kinesis Data Firehose tries to process all records in each <a>PutRecordBatch</a> request. A single record failure does not stop the processing
37
- * of subsequent records. </p>
38
- * <p>A successfully processed record includes a <code>RecordId</code> value, which is
39
- * unique for the record. An unsuccessfully processed record includes <code>ErrorCode</code>
40
- * and <code>ErrorMessage</code> values. <code>ErrorCode</code> reflects the type of error,
41
- * and is one of the following values: <code>ServiceUnavailableException</code> or
42
- * <code>InternalFailure</code>. <code>ErrorMessage</code> provides more detailed
43
- * information about the error.</p>
44
- * <p>If there is an internal server error or a timeout, the write might have completed or
45
- * it might have failed. If <code>FailedPutCount</code> is greater than 0, retry the request,
46
- * resending only those records that might have failed processing. This minimizes the possible
47
- * duplicate records and also reduces the total bytes sent (and corresponding charges). We
48
- * recommend that you handle any duplicates at the destination.</p>
49
- * <p>If <a>PutRecordBatch</a> throws <code>ServiceUnavailableException</code>,
50
- * back off and retry. If the exception persists, it is possible that the throughput limits
51
- * have been exceeded for the delivery stream.</p>
52
- *
53
- * <p>Data records sent to Kinesis Data Firehose are stored for 24 hours from the time they
54
- * are added to a delivery stream as it attempts to send the records to the destination. If
55
- * the destination is unreachable for more than 24 hours, the data is no longer
56
- * available.</p>
57
- * <important>
58
- * <p>Don't concatenate two or more base64 strings to form the data fields of your records.
59
- * Instead, concatenate the raw data, then perform base64 encoding.</p>
60
- * </important>
61
- * @example
62
- * Use a bare-bones client and the command you need to make an API call.
63
- * ```javascript
64
- * import { FirehoseClient, PutRecordBatchCommand } from "@aws-sdk/client-firehose"; // ES Modules import
65
- * // const { FirehoseClient, PutRecordBatchCommand } = require("@aws-sdk/client-firehose"); // CommonJS import
66
- * const client = new FirehoseClient(config);
67
- * const command = new PutRecordBatchCommand(input);
68
- * const response = await client.send(command);
69
- * ```
70
- *
71
- * @see {@link PutRecordBatchCommandInput} for command's `input` shape.
72
- * @see {@link PutRecordBatchCommandOutput} for command's `response` shape.
73
- * @see {@link FirehoseClientResolvedConfig | config} for command's `input` shape.
74
- *
75
- */
9
+
76
10
  export declare class PutRecordBatchCommand extends $Command<PutRecordBatchCommandInput, PutRecordBatchCommandOutput, FirehoseClientResolvedConfig> {
77
11
  readonly input: PutRecordBatchCommandInput;
78
12
  constructor(input: PutRecordBatchCommandInput);
79
- /**
80
- * @internal
81
- */
13
+
82
14
  resolveMiddleware(clientStack: MiddlewareStack<ServiceInputTypes, ServiceOutputTypes>, configuration: FirehoseClientResolvedConfig, options?: __HttpHandlerOptions): Handler<PutRecordBatchCommandInput, PutRecordBatchCommandOutput>;
83
15
  private serialize;
84
16
  private deserialize;
@@ -6,60 +6,11 @@ export interface PutRecordCommandInput extends PutRecordInput {
6
6
  }
7
7
  export interface PutRecordCommandOutput extends PutRecordOutput, __MetadataBearer {
8
8
  }
9
- /**
10
- * <p>Writes a single data record into an Amazon Kinesis Data Firehose delivery stream. To
11
- * write multiple data records into a delivery stream, use <a>PutRecordBatch</a>.
12
- * Applications using these operations are referred to as producers.</p>
13
- * <p>By default, each delivery stream can take in up to 2,000 transactions per second,
14
- * 5,000 records per second, or 5 MB per second. If you use <a>PutRecord</a> and
15
- * <a>PutRecordBatch</a>, the limits are an aggregate across these two
16
- * operations for each delivery stream. For more information about limits and how to request
17
- * an increase, see <a href="https://docs.aws.amazon.com/firehose/latest/dev/limits.html">Amazon
18
- * Kinesis Data Firehose Limits</a>. </p>
19
- * <p>You must specify the name of the delivery stream and the data record when using <a>PutRecord</a>. The data record consists of a data blob that can be up to 1,000
20
- * KiB in size, and any kind of data. For example, it can be a segment from a log file,
21
- * geographic location data, website clickstream data, and so on.</p>
22
- * <p>Kinesis Data Firehose buffers records before delivering them to the destination. To
23
- * disambiguate the data blobs at the destination, a common solution is to use delimiters in
24
- * the data, such as a newline (<code>\n</code>) or some other character unique within the
25
- * data. This allows the consumer application to parse individual data items when reading the
26
- * data from the destination.</p>
27
- * <p>The <code>PutRecord</code> operation returns a <code>RecordId</code>, which is a
28
- * unique string assigned to each record. Producer applications can use this ID for purposes
29
- * such as auditability and investigation.</p>
30
- * <p>If the <code>PutRecord</code> operation throws a
31
- * <code>ServiceUnavailableException</code>, back off and retry. If the exception persists,
32
- * it is possible that the throughput limits have been exceeded for the delivery stream. </p>
33
- * <p>Data records sent to Kinesis Data Firehose are stored for 24 hours from the time they
34
- * are added to a delivery stream as it tries to send the records to the destination. If the
35
- * destination is unreachable for more than 24 hours, the data is no longer
36
- * available.</p>
37
- *
38
- * <important>
39
- * <p>Don't concatenate two or more base64 strings to form the data fields of your records.
40
- * Instead, concatenate the raw data, then perform base64 encoding.</p>
41
- * </important>
42
- * @example
43
- * Use a bare-bones client and the command you need to make an API call.
44
- * ```javascript
45
- * import { FirehoseClient, PutRecordCommand } from "@aws-sdk/client-firehose"; // ES Modules import
46
- * // const { FirehoseClient, PutRecordCommand } = require("@aws-sdk/client-firehose"); // CommonJS import
47
- * const client = new FirehoseClient(config);
48
- * const command = new PutRecordCommand(input);
49
- * const response = await client.send(command);
50
- * ```
51
- *
52
- * @see {@link PutRecordCommandInput} for command's `input` shape.
53
- * @see {@link PutRecordCommandOutput} for command's `response` shape.
54
- * @see {@link FirehoseClientResolvedConfig | config} for command's `input` shape.
55
- *
56
- */
9
+
57
10
  export declare class PutRecordCommand extends $Command<PutRecordCommandInput, PutRecordCommandOutput, FirehoseClientResolvedConfig> {
58
11
  readonly input: PutRecordCommandInput;
59
12
  constructor(input: PutRecordCommandInput);
60
- /**
61
- * @internal
62
- */
13
+
63
14
  resolveMiddleware(clientStack: MiddlewareStack<ServiceInputTypes, ServiceOutputTypes>, configuration: FirehoseClientResolvedConfig, options?: __HttpHandlerOptions): Handler<PutRecordCommandInput, PutRecordCommandOutput>;
64
15
  private serialize;
65
16
  private deserialize;
@@ -6,64 +6,11 @@ export interface StartDeliveryStreamEncryptionCommandInput extends StartDelivery
6
6
  }
7
7
  export interface StartDeliveryStreamEncryptionCommandOutput extends StartDeliveryStreamEncryptionOutput, __MetadataBearer {
8
8
  }
9
- /**
10
- * <p>Enables server-side encryption (SSE) for the delivery stream. </p>
11
- * <p>This operation is asynchronous. It returns immediately. When you invoke it, Kinesis Data
12
- * Firehose first sets the encryption status of the stream to <code>ENABLING</code>, and then
13
- * to <code>ENABLED</code>. The encryption status of a delivery stream is the
14
- * <code>Status</code> property in <a>DeliveryStreamEncryptionConfiguration</a>.
15
- * If the operation fails, the encryption status changes to <code>ENABLING_FAILED</code>. You
16
- * can continue to read and write data to your delivery stream while the encryption status is
17
- * <code>ENABLING</code>, but the data is not encrypted. It can take up to 5 seconds after
18
- * the encryption status changes to <code>ENABLED</code> before all records written to the
19
- * delivery stream are encrypted. To find out whether a record or a batch of records was
20
- * encrypted, check the response elements <a>PutRecordOutput$Encrypted</a> and
21
- * <a>PutRecordBatchOutput$Encrypted</a>, respectively.</p>
22
- * <p>To check the encryption status of a delivery stream, use <a>DescribeDeliveryStream</a>.</p>
23
- * <p>Even if encryption is currently enabled for a delivery stream, you can still invoke this
24
- * operation on it to change the ARN of the CMK or both its type and ARN. If you invoke this
25
- * method to change the CMK, and the old CMK is of type <code>CUSTOMER_MANAGED_CMK</code>,
26
- * Kinesis Data Firehose schedules the grant it had on the old CMK for retirement. If the new
27
- * CMK is of type <code>CUSTOMER_MANAGED_CMK</code>, Kinesis Data Firehose creates a grant
28
- * that enables it to use the new CMK to encrypt and decrypt data and to manage the
29
- * grant.</p>
30
- * <p>If a delivery stream already has encryption enabled and then you invoke this operation
31
- * to change the ARN of the CMK or both its type and ARN and you get
32
- * <code>ENABLING_FAILED</code>, this only means that the attempt to change the CMK failed.
33
- * In this case, encryption remains enabled with the old CMK.</p>
34
- * <p>If the encryption status of your delivery stream is <code>ENABLING_FAILED</code>, you
35
- * can invoke this operation again with a valid CMK. The CMK must be enabled and the key
36
- * policy mustn't explicitly deny the permission for Kinesis Data Firehose to invoke KMS
37
- * encrypt and decrypt operations.</p>
38
- * <p>You can enable SSE for a delivery stream only if it's a delivery stream that uses
39
- * <code>DirectPut</code> as its source. </p>
40
- * <p>The <code>StartDeliveryStreamEncryption</code> and
41
- * <code>StopDeliveryStreamEncryption</code> operations have a combined limit of 25 calls
42
- * per delivery stream per 24 hours. For example, you reach the limit if you call
43
- * <code>StartDeliveryStreamEncryption</code> 13 times and
44
- * <code>StopDeliveryStreamEncryption</code> 12 times for the same delivery stream in a
45
- * 24-hour period.</p>
46
- * @example
47
- * Use a bare-bones client and the command you need to make an API call.
48
- * ```javascript
49
- * import { FirehoseClient, StartDeliveryStreamEncryptionCommand } from "@aws-sdk/client-firehose"; // ES Modules import
50
- * // const { FirehoseClient, StartDeliveryStreamEncryptionCommand } = require("@aws-sdk/client-firehose"); // CommonJS import
51
- * const client = new FirehoseClient(config);
52
- * const command = new StartDeliveryStreamEncryptionCommand(input);
53
- * const response = await client.send(command);
54
- * ```
55
- *
56
- * @see {@link StartDeliveryStreamEncryptionCommandInput} for command's `input` shape.
57
- * @see {@link StartDeliveryStreamEncryptionCommandOutput} for command's `response` shape.
58
- * @see {@link FirehoseClientResolvedConfig | config} for command's `input` shape.
59
- *
60
- */
9
+
61
10
  export declare class StartDeliveryStreamEncryptionCommand extends $Command<StartDeliveryStreamEncryptionCommandInput, StartDeliveryStreamEncryptionCommandOutput, FirehoseClientResolvedConfig> {
62
11
  readonly input: StartDeliveryStreamEncryptionCommandInput;
63
12
  constructor(input: StartDeliveryStreamEncryptionCommandInput);
64
- /**
65
- * @internal
66
- */
13
+
67
14
  resolveMiddleware(clientStack: MiddlewareStack<ServiceInputTypes, ServiceOutputTypes>, configuration: FirehoseClientResolvedConfig, options?: __HttpHandlerOptions): Handler<StartDeliveryStreamEncryptionCommandInput, StartDeliveryStreamEncryptionCommandOutput>;
68
15
  private serialize;
69
16
  private deserialize;