cdk-comprehend-s3olap 2.0.121 → 2.0.123
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.jsii +4 -4
- package/lib/cdk-comprehend-s3olap.js +2 -2
- package/lib/comprehend-lambdas.js +2 -2
- package/lib/iam-roles.js +4 -4
- package/node_modules/@esbuild/linux-x64/bin/esbuild +0 -0
- package/node_modules/@esbuild/linux-x64/package.json +1 -1
- package/node_modules/aws-sdk/CHANGELOG.md +25 -1
- package/node_modules/aws-sdk/README.md +1 -1
- package/node_modules/aws-sdk/apis/connect-2017-08-08.min.json +196 -187
- package/node_modules/aws-sdk/apis/ecr-public-2020-10-30.min.json +8 -8
- package/node_modules/aws-sdk/apis/emr-serverless-2021-07-13.min.json +3 -0
- package/node_modules/aws-sdk/apis/eventbridge-2015-10-07.min.json +18 -6
- package/node_modules/aws-sdk/apis/iot-data-2015-05-28.min.json +3 -0
- package/node_modules/aws-sdk/apis/marketplace-catalog-2018-09-17.min.json +82 -15
- package/node_modules/aws-sdk/apis/mediaconvert-2017-08-29.min.json +56 -32
- package/node_modules/aws-sdk/apis/rekognition-2016-06-27.min.json +224 -138
- package/node_modules/aws-sdk/apis/wafv2-2019-07-29.min.json +55 -28
- package/node_modules/aws-sdk/clients/connect.d.ts +15 -4
- package/node_modules/aws-sdk/clients/ecrpublic.d.ts +2 -2
- package/node_modules/aws-sdk/clients/ecs.d.ts +24 -24
- package/node_modules/aws-sdk/clients/emrserverless.d.ts +7 -3
- package/node_modules/aws-sdk/clients/eventbridge.d.ts +3 -1
- package/node_modules/aws-sdk/clients/iotdata.d.ts +5 -0
- package/node_modules/aws-sdk/clients/marketplacecatalog.d.ts +67 -7
- package/node_modules/aws-sdk/clients/mediaconvert.d.ts +20 -1
- package/node_modules/aws-sdk/clients/omics.d.ts +2 -2
- package/node_modules/aws-sdk/clients/rekognition.d.ts +114 -16
- package/node_modules/aws-sdk/clients/wafv2.d.ts +89 -62
- package/node_modules/aws-sdk/dist/aws-sdk-core-react-native.js +1 -1
- package/node_modules/aws-sdk/dist/aws-sdk-react-native.js +2537 -880
- package/node_modules/aws-sdk/dist/aws-sdk.js +508 -343
- package/node_modules/aws-sdk/dist/aws-sdk.min.js +79 -79
- package/node_modules/aws-sdk/dist/xml2js.js +2579 -922
- package/node_modules/aws-sdk/lib/core.js +1 -1
- package/node_modules/aws-sdk/package.json +2 -2
- package/node_modules/esbuild/bin/esbuild +1 -1
- package/node_modules/esbuild/lib/main.js +8 -8
- package/node_modules/esbuild/package.json +23 -23
- package/node_modules/xml2js/README.md +108 -7
- package/node_modules/xml2js/lib/parser.js +35 -7
- package/node_modules/xml2js/lib/xml2js.js +2 -0
- package/node_modules/xml2js/node_modules/xmlbuilder/CHANGELOG.md +47 -0
- package/node_modules/xml2js/node_modules/xmlbuilder/LICENSE +21 -21
- package/node_modules/xml2js/node_modules/xmlbuilder/README.md +86 -85
- package/node_modules/xml2js/node_modules/xmlbuilder/appveyor.yml +20 -0
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/Derivation.js +10 -0
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/DocumentPosition.js +12 -0
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/NodeType.js +23 -0
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/OperationType.js +11 -0
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/Utility.js +11 -1
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/WriterState.js +10 -0
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLAttribute.js +86 -9
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLCData.js +10 -6
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLCharacterData.js +79 -0
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLComment.js +10 -6
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLDOMConfiguration.js +64 -0
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLDOMErrorHandler.js +16 -0
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLDOMImplementation.js +32 -0
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLDOMStringList.js +28 -0
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLDTDAttList.js +16 -11
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLDTDElement.js +7 -4
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLDTDEntity.js +49 -8
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLDTDNotation.js +20 -5
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLDeclaration.js +5 -2
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLDocType.js +84 -5
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLDocument.js +199 -5
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLDocumentCB.js +165 -39
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLDocumentFragment.js +24 -0
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLDummy.js +31 -0
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLElement.js +207 -20
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLNamedNodeMap.js +58 -0
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLNode.js +403 -50
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLNodeFilter.js +48 -0
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLNodeList.js +28 -0
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLProcessingInstruction.js +19 -5
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLRaw.js +6 -3
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLStreamWriter.js +94 -197
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLStringWriter.js +6 -305
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLStringifier.js +109 -32
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLText.js +43 -6
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLTypeInfo.js +21 -0
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLUserDataHandler.js +16 -0
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/XMLWriterBase.js +397 -59
- package/node_modules/xml2js/node_modules/xmlbuilder/lib/index.js +15 -3
- package/node_modules/xml2js/node_modules/xmlbuilder/package.json +5 -3
- package/node_modules/xml2js/node_modules/xmlbuilder/typings/index.d.ts +153 -0
- package/node_modules/xml2js/package.json +11 -5
- package/package.json +4 -4
- package/node_modules/xml2js/node_modules/xmlbuilder/.npmignore +0 -5
@@ -2523,7 +2523,7 @@ declare namespace EventBridge {
|
|
2523
2523
|
/**
|
2524
2524
|
* The SQL statement text to run.
|
2525
2525
|
*/
|
2526
|
-
Sql
|
2526
|
+
Sql?: Sql;
|
2527
2527
|
/**
|
2528
2528
|
* The name of the SQL statement. You can name the SQL statement when you create it to identify the query.
|
2529
2529
|
*/
|
@@ -2532,6 +2532,7 @@ declare namespace EventBridge {
|
|
2532
2532
|
* Indicates whether to send an event back to EventBridge after the SQL statement runs.
|
2533
2533
|
*/
|
2534
2534
|
WithEvent?: Boolean;
|
2535
|
+
Sqls?: Sqls;
|
2535
2536
|
}
|
2536
2537
|
export type RedshiftSecretManagerArn = string;
|
2537
2538
|
export type ReferenceId = string;
|
@@ -2766,6 +2767,7 @@ declare namespace EventBridge {
|
|
2766
2767
|
}
|
2767
2768
|
export type SecretsManagerSecretArn = string;
|
2768
2769
|
export type Sql = string;
|
2770
|
+
export type Sqls = Sql[];
|
2769
2771
|
export interface SqsParameters {
|
2770
2772
|
/**
|
2771
2773
|
* The FIFO message group ID to use as the target.
|
@@ -110,6 +110,10 @@ declare namespace IotData {
|
|
110
110
|
* The Epoch date and time, in milliseconds, when the retained message was stored by IoT.
|
111
111
|
*/
|
112
112
|
lastModifiedTime?: Timestamp;
|
113
|
+
/**
|
114
|
+
* A base64-encoded JSON string that includes an array of JSON objects, or null if the retained message doesn't include any user properties. The following example userProperties parameter is a JSON string that represents two user properties. Note that it will be base64-encoded: [{"deviceName": "alpha"}, {"deviceCnt": "45"}]
|
115
|
+
*/
|
116
|
+
userProperties?: UserPropertiesBlob;
|
113
117
|
}
|
114
118
|
export interface GetThingShadowRequest {
|
115
119
|
/**
|
@@ -273,6 +277,7 @@ declare namespace IotData {
|
|
273
277
|
payload?: JsonDocument;
|
274
278
|
}
|
275
279
|
export type UserProperties = string;
|
280
|
+
export type UserPropertiesBlob = Buffer|Uint8Array|Blob|string;
|
276
281
|
/**
|
277
282
|
* A string in YYYY-MM-DD format that represents the latest possible API version that can be used in this service. Specify 'latest' to use the latest possible version.
|
278
283
|
*/
|
@@ -19,6 +19,14 @@ declare class MarketplaceCatalog extends Service {
|
|
19
19
|
* Used to cancel an open change request. Must be sent before the status of the request changes to APPLYING, the final stage of completing your change request. You can describe a change during the 60-day request history retention period for API calls.
|
20
20
|
*/
|
21
21
|
cancelChangeSet(callback?: (err: AWSError, data: MarketplaceCatalog.Types.CancelChangeSetResponse) => void): Request<MarketplaceCatalog.Types.CancelChangeSetResponse, AWSError>;
|
22
|
+
/**
|
23
|
+
* Deletes a resource-based policy on an Entity that is identified by its resource ARN.
|
24
|
+
*/
|
25
|
+
deleteResourcePolicy(params: MarketplaceCatalog.Types.DeleteResourcePolicyRequest, callback?: (err: AWSError, data: MarketplaceCatalog.Types.DeleteResourcePolicyResponse) => void): Request<MarketplaceCatalog.Types.DeleteResourcePolicyResponse, AWSError>;
|
26
|
+
/**
|
27
|
+
* Deletes a resource-based policy on an Entity that is identified by its resource ARN.
|
28
|
+
*/
|
29
|
+
deleteResourcePolicy(callback?: (err: AWSError, data: MarketplaceCatalog.Types.DeleteResourcePolicyResponse) => void): Request<MarketplaceCatalog.Types.DeleteResourcePolicyResponse, AWSError>;
|
22
30
|
/**
|
23
31
|
* Provides information about a given change set.
|
24
32
|
*/
|
@@ -35,6 +43,14 @@ declare class MarketplaceCatalog extends Service {
|
|
35
43
|
* Returns the metadata and content of the entity.
|
36
44
|
*/
|
37
45
|
describeEntity(callback?: (err: AWSError, data: MarketplaceCatalog.Types.DescribeEntityResponse) => void): Request<MarketplaceCatalog.Types.DescribeEntityResponse, AWSError>;
|
46
|
+
/**
|
47
|
+
* Gets a resource-based policy of an Entity that is identified by its resource ARN.
|
48
|
+
*/
|
49
|
+
getResourcePolicy(params: MarketplaceCatalog.Types.GetResourcePolicyRequest, callback?: (err: AWSError, data: MarketplaceCatalog.Types.GetResourcePolicyResponse) => void): Request<MarketplaceCatalog.Types.GetResourcePolicyResponse, AWSError>;
|
50
|
+
/**
|
51
|
+
* Gets a resource-based policy of an Entity that is identified by its resource ARN.
|
52
|
+
*/
|
53
|
+
getResourcePolicy(callback?: (err: AWSError, data: MarketplaceCatalog.Types.GetResourcePolicyResponse) => void): Request<MarketplaceCatalog.Types.GetResourcePolicyResponse, AWSError>;
|
38
54
|
/**
|
39
55
|
* Returns the list of change sets owned by the account being used to make the call. You can filter this list by providing any combination of entityId, ChangeSetName, and status. If you provide more than one filter, the API operation applies a logical AND between the filters. You can describe a change during the 60-day request history retention period for API calls.
|
40
56
|
*/
|
@@ -60,11 +76,19 @@ declare class MarketplaceCatalog extends Service {
|
|
60
76
|
*/
|
61
77
|
listTagsForResource(callback?: (err: AWSError, data: MarketplaceCatalog.Types.ListTagsForResourceResponse) => void): Request<MarketplaceCatalog.Types.ListTagsForResourceResponse, AWSError>;
|
62
78
|
/**
|
63
|
-
*
|
79
|
+
* Attaches a resource-based policy to an Entity. Examples of an entity include: AmiProduct and ContainerProduct.
|
80
|
+
*/
|
81
|
+
putResourcePolicy(params: MarketplaceCatalog.Types.PutResourcePolicyRequest, callback?: (err: AWSError, data: MarketplaceCatalog.Types.PutResourcePolicyResponse) => void): Request<MarketplaceCatalog.Types.PutResourcePolicyResponse, AWSError>;
|
82
|
+
/**
|
83
|
+
* Attaches a resource-based policy to an Entity. Examples of an entity include: AmiProduct and ContainerProduct.
|
84
|
+
*/
|
85
|
+
putResourcePolicy(callback?: (err: AWSError, data: MarketplaceCatalog.Types.PutResourcePolicyResponse) => void): Request<MarketplaceCatalog.Types.PutResourcePolicyResponse, AWSError>;
|
86
|
+
/**
|
87
|
+
* Allows you to request changes for your entities. Within a single ChangeSet, you can't start the same change type against the same entity multiple times. Additionally, when a ChangeSet is running, all the entities targeted by the different changes are locked until the change set has completed (either succeeded, cancelled, or failed). If you try to start a change set containing a change against an entity that is already locked, you will receive a ResourceInUseException error. For example, you can't start the ChangeSet described in the example later in this topic because it contains two changes to run the same change type (AddRevisions) against the same entity (entity-id@1). For more information about working with change sets, see Working with change sets. For information on change types for single-AMI products, see Working with single-AMI products. Als, for more information on change types available for container-based products, see Working with container products.
|
64
88
|
*/
|
65
89
|
startChangeSet(params: MarketplaceCatalog.Types.StartChangeSetRequest, callback?: (err: AWSError, data: MarketplaceCatalog.Types.StartChangeSetResponse) => void): Request<MarketplaceCatalog.Types.StartChangeSetResponse, AWSError>;
|
66
90
|
/**
|
67
|
-
* Allows you to request changes for your entities. Within a single ChangeSet, you can't start the same change type against the same entity multiple times. Additionally, when a ChangeSet is running, all the entities targeted by the different changes are locked until the change set has completed (either succeeded, cancelled, or failed). If you try to start a change set containing a change against an entity that is already locked, you will receive a ResourceInUseException error. For example, you can't start the ChangeSet described in the example later in this topic because it contains two changes to run the same change type (AddRevisions) against the same entity (entity-id@1). For more information about working with change sets, see Working with change sets.
|
91
|
+
* Allows you to request changes for your entities. Within a single ChangeSet, you can't start the same change type against the same entity multiple times. Additionally, when a ChangeSet is running, all the entities targeted by the different changes are locked until the change set has completed (either succeeded, cancelled, or failed). If you try to start a change set containing a change against an entity that is already locked, you will receive a ResourceInUseException error. For example, you can't start the ChangeSet described in the example later in this topic because it contains two changes to run the same change type (AddRevisions) against the same entity (entity-id@1). For more information about working with change sets, see Working with change sets. For information on change types for single-AMI products, see Working with single-AMI products. Als, for more information on change types available for container-based products, see Working with container products.
|
68
92
|
*/
|
69
93
|
startChangeSet(callback?: (err: AWSError, data: MarketplaceCatalog.Types.StartChangeSetResponse) => void): Request<MarketplaceCatalog.Types.StartChangeSetResponse, AWSError>;
|
70
94
|
/**
|
@@ -109,7 +133,7 @@ declare namespace MarketplaceCatalog {
|
|
109
133
|
export type Catalog = string;
|
110
134
|
export interface Change {
|
111
135
|
/**
|
112
|
-
* Change types are single string values that describe your intention for the change. Each change type is unique for each EntityType provided in the change's scope.
|
136
|
+
* Change types are single string values that describe your intention for the change. Each change type is unique for each EntityType provided in the change's scope. For more information on change types available for single-AMI products, see Working with single-AMI products. Also, for more information on change types available for container-based products, see Working with container products.
|
113
137
|
*/
|
114
138
|
ChangeType: ChangeType;
|
115
139
|
/**
|
@@ -121,7 +145,7 @@ declare namespace MarketplaceCatalog {
|
|
121
145
|
*/
|
122
146
|
EntityTags?: TagList;
|
123
147
|
/**
|
124
|
-
* This object contains details specific to the change type of the requested change.
|
148
|
+
* This object contains details specific to the change type of the requested change. For more information on change types available for single-AMI products, see Working with single-AMI products. Also, for more information on change types available for container-based products, see Working with container products.
|
125
149
|
*/
|
126
150
|
Details: Json;
|
127
151
|
/**
|
@@ -193,6 +217,14 @@ declare namespace MarketplaceCatalog {
|
|
193
217
|
export type ChangeType = string;
|
194
218
|
export type ClientRequestToken = string;
|
195
219
|
export type DateTimeISO8601 = string;
|
220
|
+
export interface DeleteResourcePolicyRequest {
|
221
|
+
/**
|
222
|
+
* The Amazon Resource Name (ARN) of the Entity resource that is associated with the resource policy.
|
223
|
+
*/
|
224
|
+
ResourceArn: ResourceARN;
|
225
|
+
}
|
226
|
+
export interface DeleteResourcePolicyResponse {
|
227
|
+
}
|
196
228
|
export interface DescribeChangeSetRequest {
|
197
229
|
/**
|
198
230
|
* Required. The catalog related to the request. Fixed value: AWSMarketplace
|
@@ -339,8 +371,21 @@ declare namespace MarketplaceCatalog {
|
|
339
371
|
export type FilterList = Filter[];
|
340
372
|
export type FilterName = string;
|
341
373
|
export type FilterValueContent = string;
|
374
|
+
export interface GetResourcePolicyRequest {
|
375
|
+
/**
|
376
|
+
* The Amazon Resource Name (ARN) of the Entity resource that is associated with the resource policy.
|
377
|
+
*/
|
378
|
+
ResourceArn: ResourceARN;
|
379
|
+
}
|
380
|
+
export interface GetResourcePolicyResponse {
|
381
|
+
/**
|
382
|
+
* The policy document to set; formatted in JSON.
|
383
|
+
*/
|
384
|
+
Policy?: ResourcePolicyJson;
|
385
|
+
}
|
342
386
|
export type Identifier = string;
|
343
387
|
export type Json = string;
|
388
|
+
export type ListChangeSetsMaxResultInteger = number;
|
344
389
|
export interface ListChangeSetsRequest {
|
345
390
|
/**
|
346
391
|
* The catalog related to the request. Fixed value: AWSMarketplace
|
@@ -357,7 +402,7 @@ declare namespace MarketplaceCatalog {
|
|
357
402
|
/**
|
358
403
|
* The maximum number of results returned by a single call. This value must be provided in the next call to retrieve the next set of results. By default, this value is 20.
|
359
404
|
*/
|
360
|
-
MaxResults?:
|
405
|
+
MaxResults?: ListChangeSetsMaxResultInteger;
|
361
406
|
/**
|
362
407
|
* The token value retrieved from a previous call to access the next page of results.
|
363
408
|
*/
|
@@ -373,6 +418,7 @@ declare namespace MarketplaceCatalog {
|
|
373
418
|
*/
|
374
419
|
NextToken?: NextToken;
|
375
420
|
}
|
421
|
+
export type ListEntitiesMaxResultInteger = number;
|
376
422
|
export interface ListEntitiesRequest {
|
377
423
|
/**
|
378
424
|
* The catalog related to the request. Fixed value: AWSMarketplace
|
@@ -397,7 +443,8 @@ declare namespace MarketplaceCatalog {
|
|
397
443
|
/**
|
398
444
|
* Specifies the upper limit of the elements on a single page. If a value isn't provided, the default value is 20.
|
399
445
|
*/
|
400
|
-
MaxResults?:
|
446
|
+
MaxResults?: ListEntitiesMaxResultInteger;
|
447
|
+
OwnershipType?: OwnershipType;
|
401
448
|
}
|
402
449
|
export interface ListEntitiesResponse {
|
403
450
|
/**
|
@@ -425,12 +472,25 @@ declare namespace MarketplaceCatalog {
|
|
425
472
|
*/
|
426
473
|
Tags?: TagList;
|
427
474
|
}
|
428
|
-
export type MaxResultInteger = number;
|
429
475
|
export type NextToken = string;
|
476
|
+
export type OwnershipType = "SELF"|"SHARED"|string;
|
477
|
+
export interface PutResourcePolicyRequest {
|
478
|
+
/**
|
479
|
+
* The Amazon Resource Name (ARN) of the Entity resource you want to associate with a resource policy.
|
480
|
+
*/
|
481
|
+
ResourceArn: ResourceARN;
|
482
|
+
/**
|
483
|
+
* The policy document to set; formatted in JSON.
|
484
|
+
*/
|
485
|
+
Policy: ResourcePolicyJson;
|
486
|
+
}
|
487
|
+
export interface PutResourcePolicyResponse {
|
488
|
+
}
|
430
489
|
export type RequestedChangeList = Change[];
|
431
490
|
export type ResourceARN = string;
|
432
491
|
export type ResourceId = string;
|
433
492
|
export type ResourceIdList = ResourceId[];
|
493
|
+
export type ResourcePolicyJson = string;
|
434
494
|
export interface Sort {
|
435
495
|
/**
|
436
496
|
* For ListEntities, supported attributes include LastModifiedDate (default), Visibility, EntityId, and Name. For ListChangeSets, supported attributes include StartTime and EndTime.
|
@@ -3144,7 +3144,7 @@ Within your job settings, all of your DVB-Sub settings must be identical.
|
|
3144
3144
|
*/
|
3145
3145
|
Queue?: __string;
|
3146
3146
|
/**
|
3147
|
-
* Required for setting up a job to use queue hopping. Minimum wait time in minutes until the job can hop to the destination queue. Valid range is 1 to
|
3147
|
+
* Required for setting up a job to use queue hopping. Minimum wait time in minutes until the job can hop to the destination queue. Valid range is 1 to 4320 minutes, inclusive.
|
3148
3148
|
*/
|
3149
3149
|
WaitMinutes?: __integer;
|
3150
3150
|
}
|
@@ -3450,6 +3450,10 @@ Within your job settings, all of your DVB-Sub settings must be identical.
|
|
3450
3450
|
* The tag type that AWS Billing and Cost Management will use to sort your AWS Elemental MediaConvert costs on any billing report that you set up.
|
3451
3451
|
*/
|
3452
3452
|
BillingTagsSource?: BillingTagsSource;
|
3453
|
+
/**
|
3454
|
+
* Prevent duplicate jobs from being created and ensure idempotency for your requests. A client request token can be any string that includes up to 64 ASCII characters. If you reuse a client request token within one minute of a successful request, the API returns the job details of the original request instead. For more information see https://docs.aws.amazon.com/mediaconvert/latest/apireference/idempotency.html.
|
3455
|
+
*/
|
3456
|
+
ClientRequestToken?: __string;
|
3453
3457
|
/**
|
3454
3458
|
* The time, in Unix epoch format in seconds, when the job got created.
|
3455
3459
|
*/
|
@@ -3534,6 +3538,10 @@ Within your job settings, all of your DVB-Sub settings must be identical.
|
|
3534
3538
|
* User-defined metadata that you want to associate with an MediaConvert job. You specify metadata in key/value pairs.
|
3535
3539
|
*/
|
3536
3540
|
UserMetadata?: __mapOf__string;
|
3541
|
+
/**
|
3542
|
+
* Contains any warning messages for the job. Use to help identify potential issues with your input, output, or job. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/warning_codes.html
|
3543
|
+
*/
|
3544
|
+
Warnings?: __listOfWarningGroup;
|
3537
3545
|
}
|
3538
3546
|
export interface JobMessages {
|
3539
3547
|
/**
|
@@ -5871,6 +5879,16 @@ When you specify Version 1, you must also set ID3 metadata (timedMetadata) to Pa
|
|
5871
5879
|
*/
|
5872
5880
|
RateControlMode?: Vp9RateControlMode;
|
5873
5881
|
}
|
5882
|
+
export interface WarningGroup {
|
5883
|
+
/**
|
5884
|
+
* Warning code that identifies a specific warning in the job. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/warning_codes.html
|
5885
|
+
*/
|
5886
|
+
Code: __integer;
|
5887
|
+
/**
|
5888
|
+
* The number of times this warning occurred in the job.
|
5889
|
+
*/
|
5890
|
+
Count: __integer;
|
5891
|
+
}
|
5874
5892
|
export type WatermarkingStrength = "LIGHTEST"|"LIGHTER"|"DEFAULT"|"STRONGER"|"STRONGEST"|string;
|
5875
5893
|
export type WavFormat = "RIFF"|"RF64"|string;
|
5876
5894
|
export interface WavSettings {
|
@@ -6231,6 +6249,7 @@ When you specify Version 1, you must also set ID3 metadata (timedMetadata) to Pa
|
|
6231
6249
|
export type __listOfQueue = Queue[];
|
6232
6250
|
export type __listOfQueueTransition = QueueTransition[];
|
6233
6251
|
export type __listOfTeletextPageType = TeletextPageType[];
|
6252
|
+
export type __listOfWarningGroup = WarningGroup[];
|
6234
6253
|
export type __listOf__doubleMinNegative60Max6 = __doubleMinNegative60Max6[];
|
6235
6254
|
export type __listOf__integerMin1Max2147483647 = __integerMin1Max2147483647[];
|
6236
6255
|
export type __listOf__integerMin32Max8182 = __integerMin32Max8182[];
|
@@ -4157,10 +4157,10 @@ declare namespace Omics {
|
|
4157
4157
|
export type WorkflowParameterName = string;
|
4158
4158
|
export type WorkflowParameterTemplate = {[key: string]: WorkflowParameter};
|
4159
4159
|
export type WorkflowRequestId = string;
|
4160
|
-
export type WorkflowStatus = "CREATING"|"ACTIVE"|"UPDATING"|"DELETED"|"FAILED"|
|
4160
|
+
export type WorkflowStatus = "CREATING"|"ACTIVE"|"UPDATING"|"DELETED"|"FAILED"|string;
|
4161
4161
|
export type WorkflowStatusMessage = string;
|
4162
4162
|
export type WorkflowTimestamp = Date;
|
4163
|
-
export type WorkflowType = "PRIVATE"|
|
4163
|
+
export type WorkflowType = "PRIVATE"|string;
|
4164
4164
|
/**
|
4165
4165
|
* A string in YYYY-MM-DD format that represents the latest possible API version that can be used in this service. Specify 'latest' to use the latest possible version.
|
4166
4166
|
*/
|
@@ -21,11 +21,11 @@ declare class Rekognition extends Service {
|
|
21
21
|
*/
|
22
22
|
compareFaces(callback?: (err: AWSError, data: Rekognition.Types.CompareFacesResponse) => void): Request<Rekognition.Types.CompareFacesResponse, AWSError>;
|
23
23
|
/**
|
24
|
-
* Copies a version of an Amazon Rekognition Custom Labels model from a source project to a destination project. The source and destination projects can be in different AWS accounts but must be in the same AWS Region. You can't copy a model to another AWS service. To copy a model version to a different AWS account, you need to create a resource-based policy known as a project policy. You attach the project policy to the source project by calling PutProjectPolicy. The project policy gives permission to copy the model version from a trusting AWS account to a trusted account. For more information creating and attaching a project policy, see Attaching a project policy (SDK) in the Amazon Rekognition Custom Labels Developer Guide. If you are copying a model version to a project in the same AWS account, you don't need to create a project policy. To copy a model, the destination project, source project, and source model version must already exist. Copying a model version takes a while to complete. To get the current status, call DescribeProjectVersions and check the value of Status in the ProjectVersionDescription object. The copy operation has finished when the value of Status is COPYING_COMPLETED.
|
24
|
+
* Copies a version of an Amazon Rekognition Custom Labels model from a source project to a destination project. The source and destination projects can be in different AWS accounts but must be in the same AWS Region. You can't copy a model to another AWS service. To copy a model version to a different AWS account, you need to create a resource-based policy known as a project policy. You attach the project policy to the source project by calling PutProjectPolicy. The project policy gives permission to copy the model version from a trusting AWS account to a trusted account. For more information creating and attaching a project policy, see Attaching a project policy (SDK) in the Amazon Rekognition Custom Labels Developer Guide. If you are copying a model version to a project in the same AWS account, you don't need to create a project policy. To copy a model, the destination project, source project, and source model version must already exist. Copying a model version takes a while to complete. To get the current status, call DescribeProjectVersions and check the value of Status in the ProjectVersionDescription object. The copy operation has finished when the value of Status is COPYING_COMPLETED. This operation requires permissions to perform the rekognition:CopyProjectVersion action.
|
25
25
|
*/
|
26
26
|
copyProjectVersion(params: Rekognition.Types.CopyProjectVersionRequest, callback?: (err: AWSError, data: Rekognition.Types.CopyProjectVersionResponse) => void): Request<Rekognition.Types.CopyProjectVersionResponse, AWSError>;
|
27
27
|
/**
|
28
|
-
* Copies a version of an Amazon Rekognition Custom Labels model from a source project to a destination project. The source and destination projects can be in different AWS accounts but must be in the same AWS Region. You can't copy a model to another AWS service. To copy a model version to a different AWS account, you need to create a resource-based policy known as a project policy. You attach the project policy to the source project by calling PutProjectPolicy. The project policy gives permission to copy the model version from a trusting AWS account to a trusted account. For more information creating and attaching a project policy, see Attaching a project policy (SDK) in the Amazon Rekognition Custom Labels Developer Guide. If you are copying a model version to a project in the same AWS account, you don't need to create a project policy. To copy a model, the destination project, source project, and source model version must already exist. Copying a model version takes a while to complete. To get the current status, call DescribeProjectVersions and check the value of Status in the ProjectVersionDescription object. The copy operation has finished when the value of Status is COPYING_COMPLETED.
|
28
|
+
* Copies a version of an Amazon Rekognition Custom Labels model from a source project to a destination project. The source and destination projects can be in different AWS accounts but must be in the same AWS Region. You can't copy a model to another AWS service. To copy a model version to a different AWS account, you need to create a resource-based policy known as a project policy. You attach the project policy to the source project by calling PutProjectPolicy. The project policy gives permission to copy the model version from a trusting AWS account to a trusted account. For more information creating and attaching a project policy, see Attaching a project policy (SDK) in the Amazon Rekognition Custom Labels Developer Guide. If you are copying a model version to a project in the same AWS account, you don't need to create a project policy. To copy a model, the destination project, source project, and source model version must already exist. Copying a model version takes a while to complete. To get the current status, call DescribeProjectVersions and check the value of Status in the ProjectVersionDescription object. The copy operation has finished when the value of Status is COPYING_COMPLETED. This operation requires permissions to perform the rekognition:CopyProjectVersion action.
|
29
29
|
*/
|
30
30
|
copyProjectVersion(callback?: (err: AWSError, data: Rekognition.Types.CopyProjectVersionResponse) => void): Request<Rekognition.Types.CopyProjectVersionResponse, AWSError>;
|
31
31
|
/**
|
@@ -44,6 +44,14 @@ declare class Rekognition extends Service {
|
|
44
44
|
* Creates a new Amazon Rekognition Custom Labels dataset. You can create a dataset by using an Amazon Sagemaker format manifest file or by copying an existing Amazon Rekognition Custom Labels dataset. To create a training dataset for a project, specify train for the value of DatasetType. To create the test dataset for a project, specify test for the value of DatasetType. The response from CreateDataset is the Amazon Resource Name (ARN) for the dataset. Creating a dataset takes a while to complete. Use DescribeDataset to check the current status. The dataset created successfully if the value of Status is CREATE_COMPLETE. To check if any non-terminal errors occurred, call ListDatasetEntries and check for the presence of errors lists in the JSON Lines. Dataset creation fails if a terminal error occurs (Status = CREATE_FAILED). Currently, you can't access the terminal error information. For more information, see Creating dataset in the Amazon Rekognition Custom Labels Developer Guide. This operation requires permissions to perform the rekognition:CreateDataset action. If you want to copy an existing dataset, you also require permission to perform the rekognition:ListDatasetEntries action.
|
45
45
|
*/
|
46
46
|
createDataset(callback?: (err: AWSError, data: Rekognition.Types.CreateDatasetResponse) => void): Request<Rekognition.Types.CreateDatasetResponse, AWSError>;
|
47
|
+
/**
|
48
|
+
* This API operation initiates a Face Liveness session. It returns a SessionId, which you can use to start streaming Face Liveness video and get the results for a Face Liveness session. You can use the OutputConfig option in the Settings parameter to provide an Amazon S3 bucket location. The Amazon S3 bucket stores reference images and audit images. You can use AuditImagesLimit to limit of audit images returned. This number is between 0 and 4. By default, it is set to 0. The limit is best effort and based on the duration of the selfie-video.
|
49
|
+
*/
|
50
|
+
createFaceLivenessSession(params: Rekognition.Types.CreateFaceLivenessSessionRequest, callback?: (err: AWSError, data: Rekognition.Types.CreateFaceLivenessSessionResponse) => void): Request<Rekognition.Types.CreateFaceLivenessSessionResponse, AWSError>;
|
51
|
+
/**
|
52
|
+
* This API operation initiates a Face Liveness session. It returns a SessionId, which you can use to start streaming Face Liveness video and get the results for a Face Liveness session. You can use the OutputConfig option in the Settings parameter to provide an Amazon S3 bucket location. The Amazon S3 bucket stores reference images and audit images. You can use AuditImagesLimit to limit of audit images returned. This number is between 0 and 4. By default, it is set to 0. The limit is best effort and based on the duration of the selfie-video.
|
53
|
+
*/
|
54
|
+
createFaceLivenessSession(callback?: (err: AWSError, data: Rekognition.Types.CreateFaceLivenessSessionResponse) => void): Request<Rekognition.Types.CreateFaceLivenessSessionResponse, AWSError>;
|
47
55
|
/**
|
48
56
|
* Creates a new Amazon Rekognition Custom Labels project. A project is a group of resources (datasets, model versions) that you use to create and manage Amazon Rekognition Custom Labels models. This operation requires permissions to perform the rekognition:CreateProject action.
|
49
57
|
*/
|
@@ -61,11 +69,11 @@ declare class Rekognition extends Service {
|
|
61
69
|
*/
|
62
70
|
createProjectVersion(callback?: (err: AWSError, data: Rekognition.Types.CreateProjectVersionResponse) => void): Request<Rekognition.Types.CreateProjectVersionResponse, AWSError>;
|
63
71
|
/**
|
64
|
-
* Creates an Amazon Rekognition stream processor that you can use to detect and recognize faces or to detect labels in a streaming video. Amazon Rekognition Video is a consumer of live video from Amazon Kinesis Video Streams. There are two different settings for stream processors in Amazon Rekognition: detecting faces and detecting labels. If you are creating a stream processor for detecting faces, you provide as input a Kinesis video stream (Input) and a Kinesis data stream (Output) stream. You
|
72
|
+
* Creates an Amazon Rekognition stream processor that you can use to detect and recognize faces or to detect labels in a streaming video. Amazon Rekognition Video is a consumer of live video from Amazon Kinesis Video Streams. There are two different settings for stream processors in Amazon Rekognition: detecting faces and detecting labels. If you are creating a stream processor for detecting faces, you provide as input a Kinesis video stream (Input) and a Kinesis data stream (Output) stream for receiving the output. You must use the FaceSearch option in Settings, specifying the collection that contains the faces you want to recognize. After you have finished analyzing a streaming video, use StopStreamProcessor to stop processing. If you are creating a stream processor to detect labels, you provide as input a Kinesis video stream (Input), Amazon S3 bucket information (Output), and an Amazon SNS topic ARN (NotificationChannel). You can also provide a KMS key ID to encrypt the data sent to your Amazon S3 bucket. You specify what you want to detect by using the ConnectedHome option in settings, and selecting one of the following: PERSON, PET, PACKAGE, ALL You can also specify where in the frame you want Amazon Rekognition to monitor with RegionsOfInterest. When you run the StartStreamProcessor operation on a label detection stream processor, you input start and stop information to determine the length of the processing time. Use Name to assign an identifier for the stream processor. You use Name to manage the stream processor. For example, you can start processing the source video by calling StartStreamProcessor with the Name field. This operation requires permissions to perform the rekognition:CreateStreamProcessor action. If you want to tag your stream processor, you also require permission to perform the rekognition:TagResource operation.
|
65
73
|
*/
|
66
74
|
createStreamProcessor(params: Rekognition.Types.CreateStreamProcessorRequest, callback?: (err: AWSError, data: Rekognition.Types.CreateStreamProcessorResponse) => void): Request<Rekognition.Types.CreateStreamProcessorResponse, AWSError>;
|
67
75
|
/**
|
68
|
-
* Creates an Amazon Rekognition stream processor that you can use to detect and recognize faces or to detect labels in a streaming video. Amazon Rekognition Video is a consumer of live video from Amazon Kinesis Video Streams. There are two different settings for stream processors in Amazon Rekognition: detecting faces and detecting labels. If you are creating a stream processor for detecting faces, you provide as input a Kinesis video stream (Input) and a Kinesis data stream (Output) stream. You
|
76
|
+
* Creates an Amazon Rekognition stream processor that you can use to detect and recognize faces or to detect labels in a streaming video. Amazon Rekognition Video is a consumer of live video from Amazon Kinesis Video Streams. There are two different settings for stream processors in Amazon Rekognition: detecting faces and detecting labels. If you are creating a stream processor for detecting faces, you provide as input a Kinesis video stream (Input) and a Kinesis data stream (Output) stream for receiving the output. You must use the FaceSearch option in Settings, specifying the collection that contains the faces you want to recognize. After you have finished analyzing a streaming video, use StopStreamProcessor to stop processing. If you are creating a stream processor to detect labels, you provide as input a Kinesis video stream (Input), Amazon S3 bucket information (Output), and an Amazon SNS topic ARN (NotificationChannel). You can also provide a KMS key ID to encrypt the data sent to your Amazon S3 bucket. You specify what you want to detect by using the ConnectedHome option in settings, and selecting one of the following: PERSON, PET, PACKAGE, ALL You can also specify where in the frame you want Amazon Rekognition to monitor with RegionsOfInterest. When you run the StartStreamProcessor operation on a label detection stream processor, you input start and stop information to determine the length of the processing time. Use Name to assign an identifier for the stream processor. You use Name to manage the stream processor. For example, you can start processing the source video by calling StartStreamProcessor with the Name field. This operation requires permissions to perform the rekognition:CreateStreamProcessor action. If you want to tag your stream processor, you also require permission to perform the rekognition:TagResource operation.
|
69
77
|
*/
|
70
78
|
createStreamProcessor(callback?: (err: AWSError, data: Rekognition.Types.CreateStreamProcessorResponse) => void): Request<Rekognition.Types.CreateStreamProcessorResponse, AWSError>;
|
71
79
|
/**
|
@@ -101,11 +109,11 @@ declare class Rekognition extends Service {
|
|
101
109
|
*/
|
102
110
|
deleteProject(callback?: (err: AWSError, data: Rekognition.Types.DeleteProjectResponse) => void): Request<Rekognition.Types.DeleteProjectResponse, AWSError>;
|
103
111
|
/**
|
104
|
-
* Deletes an existing project policy. To get a list of project policies attached to a project, call ListProjectPolicies. To attach a project policy to a project, call PutProjectPolicy.
|
112
|
+
* Deletes an existing project policy. To get a list of project policies attached to a project, call ListProjectPolicies. To attach a project policy to a project, call PutProjectPolicy. This operation requires permissions to perform the rekognition:DeleteProjectPolicy action.
|
105
113
|
*/
|
106
114
|
deleteProjectPolicy(params: Rekognition.Types.DeleteProjectPolicyRequest, callback?: (err: AWSError, data: Rekognition.Types.DeleteProjectPolicyResponse) => void): Request<Rekognition.Types.DeleteProjectPolicyResponse, AWSError>;
|
107
115
|
/**
|
108
|
-
* Deletes an existing project policy. To get a list of project policies attached to a project, call ListProjectPolicies. To attach a project policy to a project, call PutProjectPolicy.
|
116
|
+
* Deletes an existing project policy. To get a list of project policies attached to a project, call ListProjectPolicies. To attach a project policy to a project, call PutProjectPolicy. This operation requires permissions to perform the rekognition:DeleteProjectPolicy action.
|
109
117
|
*/
|
110
118
|
deleteProjectPolicy(callback?: (err: AWSError, data: Rekognition.Types.DeleteProjectPolicyResponse) => void): Request<Rekognition.Types.DeleteProjectPolicyResponse, AWSError>;
|
111
119
|
/**
|
@@ -181,11 +189,11 @@ declare class Rekognition extends Service {
|
|
181
189
|
*/
|
182
190
|
detectFaces(callback?: (err: AWSError, data: Rekognition.Types.DetectFacesResponse) => void): Request<Rekognition.Types.DetectFacesResponse, AWSError>;
|
183
191
|
/**
|
184
|
-
* Detects instances of real-world entities within an image (JPEG or PNG) provided as input. This includes objects like flower, tree, and table; events like wedding, graduation, and birthday party; and concepts like landscape, evening, and nature. For an example, see Analyzing images stored in an Amazon S3 bucket in the Amazon Rekognition Developer Guide. You pass the input image as base64-encoded image bytes or as a reference to an image in an Amazon S3 bucket. If you use the AWS CLI to call Amazon Rekognition operations, passing image bytes is not supported. The image must be either a PNG or JPEG formatted file. Optional Parameters You can specify one or both of the GENERAL_LABELS and IMAGE_PROPERTIES feature types when calling the DetectLabels API. Including GENERAL_LABELS will ensure the response includes the labels detected in the input image, while including IMAGE_PROPERTIES will ensure the response includes information about the image quality and color. When using GENERAL_LABELS and/or IMAGE_PROPERTIES you can provide filtering criteria to the Settings parameter. You can filter with sets of individual labels or with label categories. You can specify inclusive filters, exclusive filters, or a combination of inclusive and exclusive filters. For more information on filtering see Detecting Labels in an Image. You can specify MinConfidence to control the confidence threshold for the labels returned. The default is 55%. You can also add the MaxLabels parameter to limit the number of labels returned. The default and upper limit is 1000 labels. Response Elements For each object, scene, and concept the API returns one or more labels. The API returns the following types of information
|
192
|
+
* Detects instances of real-world entities within an image (JPEG or PNG) provided as input. This includes objects like flower, tree, and table; events like wedding, graduation, and birthday party; and concepts like landscape, evening, and nature. For an example, see Analyzing images stored in an Amazon S3 bucket in the Amazon Rekognition Developer Guide. You pass the input image as base64-encoded image bytes or as a reference to an image in an Amazon S3 bucket. If you use the AWS CLI to call Amazon Rekognition operations, passing image bytes is not supported. The image must be either a PNG or JPEG formatted file. Optional Parameters You can specify one or both of the GENERAL_LABELS and IMAGE_PROPERTIES feature types when calling the DetectLabels API. Including GENERAL_LABELS will ensure the response includes the labels detected in the input image, while including IMAGE_PROPERTIES will ensure the response includes information about the image quality and color. When using GENERAL_LABELS and/or IMAGE_PROPERTIES you can provide filtering criteria to the Settings parameter. You can filter with sets of individual labels or with label categories. You can specify inclusive filters, exclusive filters, or a combination of inclusive and exclusive filters. For more information on filtering see Detecting Labels in an Image. You can specify MinConfidence to control the confidence threshold for the labels returned. The default is 55%. You can also add the MaxLabels parameter to limit the number of labels returned. The default and upper limit is 1000 labels. Response Elements For each object, scene, and concept the API returns one or more labels. The API returns the following types of information about labels: Name - The name of the detected label. Confidence - The level of confidence in the label assigned to a detected object. Parents - The ancestor labels for a detected label. DetectLabels returns a hierarchical taxonomy of detected labels. For example, a detected car might be assigned the label car. The label car has two parent labels: Vehicle (its parent) and Transportation (its grandparent). The response includes the all ancestors for a label, where every ancestor is a unique label. In the previous example, Car, Vehicle, and Transportation are returned as unique labels in the response. Aliases - Possible Aliases for the label. Categories - The label categories that the detected label belongs to. BoundingBox — Bounding boxes are described for all instances of detected common object labels, returned in an array of Instance objects. An Instance object contains a BoundingBox object, describing the location of the label on the input image. It also includes the confidence for the accuracy of the detected bounding box. The API returns the following information regarding the image, as part of the ImageProperties structure: Quality - Information about the Sharpness, Brightness, and Contrast of the input image, scored between 0 to 100. Image quality is returned for the entire image, as well as the background and the foreground. Dominant Color - An array of the dominant colors in the image. Foreground - Information about the sharpness, brightness, and dominant colors of the input image’s foreground. Background - Information about the sharpness, brightness, and dominant colors of the input image’s background. The list of returned labels will include at least one label for every detected object, along with information about that label. In the following example, suppose the input image has a lighthouse, the sea, and a rock. The response includes all three labels, one for each object, as well as the confidence in the label: {Name: lighthouse, Confidence: 98.4629} {Name: rock,Confidence: 79.2097} {Name: sea,Confidence: 75.061} The list of labels can include multiple labels for the same object. For example, if the input image shows a flower (for example, a tulip), the operation might return the following three labels. {Name: flower,Confidence: 99.0562} {Name: plant,Confidence: 99.0562} {Name: tulip,Confidence: 99.0562} In this example, the detection algorithm more precisely identifies the flower as a tulip. If the object detected is a person, the operation doesn't provide the same facial details that the DetectFaces operation provides. This is a stateless API operation that doesn't return any data. This operation requires permissions to perform the rekognition:DetectLabels action.
|
185
193
|
*/
|
186
194
|
detectLabels(params: Rekognition.Types.DetectLabelsRequest, callback?: (err: AWSError, data: Rekognition.Types.DetectLabelsResponse) => void): Request<Rekognition.Types.DetectLabelsResponse, AWSError>;
|
187
195
|
/**
|
188
|
-
* Detects instances of real-world entities within an image (JPEG or PNG) provided as input. This includes objects like flower, tree, and table; events like wedding, graduation, and birthday party; and concepts like landscape, evening, and nature. For an example, see Analyzing images stored in an Amazon S3 bucket in the Amazon Rekognition Developer Guide. You pass the input image as base64-encoded image bytes or as a reference to an image in an Amazon S3 bucket. If you use the AWS CLI to call Amazon Rekognition operations, passing image bytes is not supported. The image must be either a PNG or JPEG formatted file. Optional Parameters You can specify one or both of the GENERAL_LABELS and IMAGE_PROPERTIES feature types when calling the DetectLabels API. Including GENERAL_LABELS will ensure the response includes the labels detected in the input image, while including IMAGE_PROPERTIES will ensure the response includes information about the image quality and color. When using GENERAL_LABELS and/or IMAGE_PROPERTIES you can provide filtering criteria to the Settings parameter. You can filter with sets of individual labels or with label categories. You can specify inclusive filters, exclusive filters, or a combination of inclusive and exclusive filters. For more information on filtering see Detecting Labels in an Image. You can specify MinConfidence to control the confidence threshold for the labels returned. The default is 55%. You can also add the MaxLabels parameter to limit the number of labels returned. The default and upper limit is 1000 labels. Response Elements For each object, scene, and concept the API returns one or more labels. The API returns the following types of information
|
196
|
+
* Detects instances of real-world entities within an image (JPEG or PNG) provided as input. This includes objects like flower, tree, and table; events like wedding, graduation, and birthday party; and concepts like landscape, evening, and nature. For an example, see Analyzing images stored in an Amazon S3 bucket in the Amazon Rekognition Developer Guide. You pass the input image as base64-encoded image bytes or as a reference to an image in an Amazon S3 bucket. If you use the AWS CLI to call Amazon Rekognition operations, passing image bytes is not supported. The image must be either a PNG or JPEG formatted file. Optional Parameters You can specify one or both of the GENERAL_LABELS and IMAGE_PROPERTIES feature types when calling the DetectLabels API. Including GENERAL_LABELS will ensure the response includes the labels detected in the input image, while including IMAGE_PROPERTIES will ensure the response includes information about the image quality and color. When using GENERAL_LABELS and/or IMAGE_PROPERTIES you can provide filtering criteria to the Settings parameter. You can filter with sets of individual labels or with label categories. You can specify inclusive filters, exclusive filters, or a combination of inclusive and exclusive filters. For more information on filtering see Detecting Labels in an Image. You can specify MinConfidence to control the confidence threshold for the labels returned. The default is 55%. You can also add the MaxLabels parameter to limit the number of labels returned. The default and upper limit is 1000 labels. Response Elements For each object, scene, and concept the API returns one or more labels. The API returns the following types of information about labels: Name - The name of the detected label. Confidence - The level of confidence in the label assigned to a detected object. Parents - The ancestor labels for a detected label. DetectLabels returns a hierarchical taxonomy of detected labels. For example, a detected car might be assigned the label car. The label car has two parent labels: Vehicle (its parent) and Transportation (its grandparent). The response includes the all ancestors for a label, where every ancestor is a unique label. In the previous example, Car, Vehicle, and Transportation are returned as unique labels in the response. Aliases - Possible Aliases for the label. Categories - The label categories that the detected label belongs to. BoundingBox — Bounding boxes are described for all instances of detected common object labels, returned in an array of Instance objects. An Instance object contains a BoundingBox object, describing the location of the label on the input image. It also includes the confidence for the accuracy of the detected bounding box. The API returns the following information regarding the image, as part of the ImageProperties structure: Quality - Information about the Sharpness, Brightness, and Contrast of the input image, scored between 0 to 100. Image quality is returned for the entire image, as well as the background and the foreground. Dominant Color - An array of the dominant colors in the image. Foreground - Information about the sharpness, brightness, and dominant colors of the input image’s foreground. Background - Information about the sharpness, brightness, and dominant colors of the input image’s background. The list of returned labels will include at least one label for every detected object, along with information about that label. In the following example, suppose the input image has a lighthouse, the sea, and a rock. The response includes all three labels, one for each object, as well as the confidence in the label: {Name: lighthouse, Confidence: 98.4629} {Name: rock,Confidence: 79.2097} {Name: sea,Confidence: 75.061} The list of labels can include multiple labels for the same object. For example, if the input image shows a flower (for example, a tulip), the operation might return the following three labels. {Name: flower,Confidence: 99.0562} {Name: plant,Confidence: 99.0562} {Name: tulip,Confidence: 99.0562} In this example, the detection algorithm more precisely identifies the flower as a tulip. If the object detected is a person, the operation doesn't provide the same facial details that the DetectFaces operation provides. This is a stateless API operation that doesn't return any data. This operation requires permissions to perform the rekognition:DetectLabels action.
|
189
197
|
*/
|
190
198
|
detectLabels(callback?: (err: AWSError, data: Rekognition.Types.DetectLabelsResponse) => void): Request<Rekognition.Types.DetectLabelsResponse, AWSError>;
|
191
199
|
/**
|
@@ -252,6 +260,14 @@ declare class Rekognition extends Service {
|
|
252
260
|
* Gets face detection results for a Amazon Rekognition Video analysis started by StartFaceDetection. Face detection with Amazon Rekognition Video is an asynchronous operation. You start face detection by calling StartFaceDetection which returns a job identifier (JobId). When the face detection operation finishes, Amazon Rekognition Video publishes a completion status to the Amazon Simple Notification Service topic registered in the initial call to StartFaceDetection. To get the results of the face detection operation, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call GetFaceDetection and pass the job identifier (JobId) from the initial call to StartFaceDetection. GetFaceDetection returns an array of detected faces (Faces) sorted by the time the faces were detected. Use MaxResults parameter to limit the number of labels returned. If there are more results than specified in MaxResults, the value of NextToken in the operation response contains a pagination token for getting the next set of results. To get the next page of results, call GetFaceDetection and populate the NextToken request parameter with the token value returned from the previous call to GetFaceDetection.
|
253
261
|
*/
|
254
262
|
getFaceDetection(callback?: (err: AWSError, data: Rekognition.Types.GetFaceDetectionResponse) => void): Request<Rekognition.Types.GetFaceDetectionResponse, AWSError>;
|
263
|
+
/**
|
264
|
+
* Retrieves the results of a specific Face Liveness session. It requires the sessionId as input, which was created using CreateFaceLivenessSession. Returns the corresponding Face Liveness confidence score, a reference image that includes a face bounding box, and audit images that also contain face bounding boxes. The Face Liveness confidence score ranges from 0 to 100. The reference image can optionally be returned.
|
265
|
+
*/
|
266
|
+
getFaceLivenessSessionResults(params: Rekognition.Types.GetFaceLivenessSessionResultsRequest, callback?: (err: AWSError, data: Rekognition.Types.GetFaceLivenessSessionResultsResponse) => void): Request<Rekognition.Types.GetFaceLivenessSessionResultsResponse, AWSError>;
|
267
|
+
/**
|
268
|
+
* Retrieves the results of a specific Face Liveness session. It requires the sessionId as input, which was created using CreateFaceLivenessSession. Returns the corresponding Face Liveness confidence score, a reference image that includes a face bounding box, and audit images that also contain face bounding boxes. The Face Liveness confidence score ranges from 0 to 100. The reference image can optionally be returned.
|
269
|
+
*/
|
270
|
+
getFaceLivenessSessionResults(callback?: (err: AWSError, data: Rekognition.Types.GetFaceLivenessSessionResultsResponse) => void): Request<Rekognition.Types.GetFaceLivenessSessionResultsResponse, AWSError>;
|
255
271
|
/**
|
256
272
|
* Gets the face search results for Amazon Rekognition Video face search started by StartFaceSearch. The search returns faces in a collection that match the faces of persons detected in a video. It also includes the time(s) that faces are matched in the video. Face search in a video is an asynchronous operation. You start face search by calling to StartFaceSearch which returns a job identifier (JobId). When the search operation finishes, Amazon Rekognition Video publishes a completion status to the Amazon Simple Notification Service topic registered in the initial call to StartFaceSearch. To get the search results, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call GetFaceSearch and pass the job identifier (JobId) from the initial call to StartFaceSearch. For more information, see Searching Faces in a Collection in the Amazon Rekognition Developer Guide. The search results are retured in an array, Persons, of PersonMatch objects. EachPersonMatch element contains details about the matching faces in the input collection, person information (facial attributes, bounding boxes, and person identifer) for the matched person, and the time the person was matched in the video. GetFaceSearch only returns the default facial attributes (BoundingBox, Confidence, Landmarks, Pose, and Quality). The other facial attributes listed in the Face object of the following response syntax are not returned. For more information, see FaceDetail in the Amazon Rekognition Developer Guide. By default, the Persons array is sorted by the time, in milliseconds from the start of the video, persons are matched. You can also sort by persons by specifying INDEX for the SORTBY input parameter.
|
257
273
|
*/
|
@@ -333,11 +349,11 @@ declare class Rekognition extends Service {
|
|
333
349
|
*/
|
334
350
|
listFaces(callback?: (err: AWSError, data: Rekognition.Types.ListFacesResponse) => void): Request<Rekognition.Types.ListFacesResponse, AWSError>;
|
335
351
|
/**
|
336
|
-
* Gets a list of the project policies attached to a project. To attach a project policy to a project, call PutProjectPolicy. To remove a project policy from a project, call DeleteProjectPolicy.
|
352
|
+
* Gets a list of the project policies attached to a project. To attach a project policy to a project, call PutProjectPolicy. To remove a project policy from a project, call DeleteProjectPolicy. This operation requires permissions to perform the rekognition:ListProjectPolicies action.
|
337
353
|
*/
|
338
354
|
listProjectPolicies(params: Rekognition.Types.ListProjectPoliciesRequest, callback?: (err: AWSError, data: Rekognition.Types.ListProjectPoliciesResponse) => void): Request<Rekognition.Types.ListProjectPoliciesResponse, AWSError>;
|
339
355
|
/**
|
340
|
-
* Gets a list of the project policies attached to a project. To attach a project policy to a project, call PutProjectPolicy. To remove a project policy from a project, call DeleteProjectPolicy.
|
356
|
+
* Gets a list of the project policies attached to a project. To attach a project policy to a project, call PutProjectPolicy. To remove a project policy from a project, call DeleteProjectPolicy. This operation requires permissions to perform the rekognition:ListProjectPolicies action.
|
341
357
|
*/
|
342
358
|
listProjectPolicies(callback?: (err: AWSError, data: Rekognition.Types.ListProjectPoliciesResponse) => void): Request<Rekognition.Types.ListProjectPoliciesResponse, AWSError>;
|
343
359
|
/**
|
@@ -357,11 +373,11 @@ declare class Rekognition extends Service {
|
|
357
373
|
*/
|
358
374
|
listTagsForResource(callback?: (err: AWSError, data: Rekognition.Types.ListTagsForResourceResponse) => void): Request<Rekognition.Types.ListTagsForResourceResponse, AWSError>;
|
359
375
|
/**
|
360
|
-
* Attaches a project policy to a Amazon Rekognition Custom Labels project in a trusting AWS account. A project policy specifies that a trusted AWS account can copy a model version from a trusting AWS account to a project in the trusted AWS account. To copy a model version you use the CopyProjectVersion operation. For more information about the format of a project policy document, see Attaching a project policy (SDK) in the Amazon Rekognition Custom Labels Developer Guide. The response from PutProjectPolicy is a revision ID for the project policy. You can attach multiple project policies to a project. You can also update an existing project policy by specifying the policy revision ID of the existing policy. To remove a project policy from a project, call DeleteProjectPolicy. To get a list of project policies attached to a project, call ListProjectPolicies. You copy a model version by calling CopyProjectVersion.
|
376
|
+
* Attaches a project policy to a Amazon Rekognition Custom Labels project in a trusting AWS account. A project policy specifies that a trusted AWS account can copy a model version from a trusting AWS account to a project in the trusted AWS account. To copy a model version you use the CopyProjectVersion operation. For more information about the format of a project policy document, see Attaching a project policy (SDK) in the Amazon Rekognition Custom Labels Developer Guide. The response from PutProjectPolicy is a revision ID for the project policy. You can attach multiple project policies to a project. You can also update an existing project policy by specifying the policy revision ID of the existing policy. To remove a project policy from a project, call DeleteProjectPolicy. To get a list of project policies attached to a project, call ListProjectPolicies. You copy a model version by calling CopyProjectVersion. This operation requires permissions to perform the rekognition:PutProjectPolicy action.
|
361
377
|
*/
|
362
378
|
putProjectPolicy(params: Rekognition.Types.PutProjectPolicyRequest, callback?: (err: AWSError, data: Rekognition.Types.PutProjectPolicyResponse) => void): Request<Rekognition.Types.PutProjectPolicyResponse, AWSError>;
|
363
379
|
/**
|
364
|
-
* Attaches a project policy to a Amazon Rekognition Custom Labels project in a trusting AWS account. A project policy specifies that a trusted AWS account can copy a model version from a trusting AWS account to a project in the trusted AWS account. To copy a model version you use the CopyProjectVersion operation. For more information about the format of a project policy document, see Attaching a project policy (SDK) in the Amazon Rekognition Custom Labels Developer Guide. The response from PutProjectPolicy is a revision ID for the project policy. You can attach multiple project policies to a project. You can also update an existing project policy by specifying the policy revision ID of the existing policy. To remove a project policy from a project, call DeleteProjectPolicy. To get a list of project policies attached to a project, call ListProjectPolicies. You copy a model version by calling CopyProjectVersion.
|
380
|
+
* Attaches a project policy to a Amazon Rekognition Custom Labels project in a trusting AWS account. A project policy specifies that a trusted AWS account can copy a model version from a trusting AWS account to a project in the trusted AWS account. To copy a model version you use the CopyProjectVersion operation. For more information about the format of a project policy document, see Attaching a project policy (SDK) in the Amazon Rekognition Custom Labels Developer Guide. The response from PutProjectPolicy is a revision ID for the project policy. You can attach multiple project policies to a project. You can also update an existing project policy by specifying the policy revision ID of the existing policy. To remove a project policy from a project, call DeleteProjectPolicy. To get a list of project policies attached to a project, call ListProjectPolicies. You copy a model version by calling CopyProjectVersion. This operation requires permissions to perform the rekognition:PutProjectPolicy action.
|
365
381
|
*/
|
366
382
|
putProjectPolicy(callback?: (err: AWSError, data: Rekognition.Types.PutProjectPolicyResponse) => void): Request<Rekognition.Types.PutProjectPolicyResponse, AWSError>;
|
367
383
|
/**
|
@@ -469,11 +485,11 @@ declare class Rekognition extends Service {
|
|
469
485
|
*/
|
470
486
|
startTextDetection(callback?: (err: AWSError, data: Rekognition.Types.StartTextDetectionResponse) => void): Request<Rekognition.Types.StartTextDetectionResponse, AWSError>;
|
471
487
|
/**
|
472
|
-
* Stops a running model. The operation might take a while to complete. To check the current status, call DescribeProjectVersions.
|
488
|
+
* Stops a running model. The operation might take a while to complete. To check the current status, call DescribeProjectVersions. This operation requires permissions to perform the rekognition:StopProjectVersion action.
|
473
489
|
*/
|
474
490
|
stopProjectVersion(params: Rekognition.Types.StopProjectVersionRequest, callback?: (err: AWSError, data: Rekognition.Types.StopProjectVersionResponse) => void): Request<Rekognition.Types.StopProjectVersionResponse, AWSError>;
|
475
491
|
/**
|
476
|
-
* Stops a running model. The operation might take a while to complete. To check the current status, call DescribeProjectVersions.
|
492
|
+
* Stops a running model. The operation might take a while to complete. To check the current status, call DescribeProjectVersions. This operation requires permissions to perform the rekognition:StopProjectVersion action.
|
477
493
|
*/
|
478
494
|
stopProjectVersion(callback?: (err: AWSError, data: Rekognition.Types.StopProjectVersionResponse) => void): Request<Rekognition.Types.StopProjectVersionResponse, AWSError>;
|
479
495
|
/**
|
@@ -569,6 +585,16 @@ declare namespace Rekognition {
|
|
569
585
|
NumberOfChannels?: ULong;
|
570
586
|
}
|
571
587
|
export type AudioMetadataList = AudioMetadata[];
|
588
|
+
export interface AuditImage {
|
589
|
+
/**
|
590
|
+
* The Base64-encoded bytes representing an image selected from the Face Liveness video and returned for audit purposes.
|
591
|
+
*/
|
592
|
+
Bytes?: LivenessImageBlob;
|
593
|
+
S3Object?: S3Object;
|
594
|
+
BoundingBox?: BoundingBox;
|
595
|
+
}
|
596
|
+
export type AuditImages = AuditImage[];
|
597
|
+
export type AuditImagesLimit = number;
|
572
598
|
export interface Beard {
|
573
599
|
/**
|
574
600
|
* Boolean value that indicates whether the face has beard or not.
|
@@ -900,6 +926,36 @@ declare namespace Rekognition {
|
|
900
926
|
*/
|
901
927
|
DatasetArn?: DatasetArn;
|
902
928
|
}
|
929
|
+
export interface CreateFaceLivenessSessionRequest {
|
930
|
+
/**
|
931
|
+
* The identifier for your AWS Key Management Service key (AWS KMS key). Used to encrypt audit images and reference images.
|
932
|
+
*/
|
933
|
+
KmsKeyId?: KmsKeyId;
|
934
|
+
/**
|
935
|
+
* A session settings object. It contains settings for the operation to be performed. For Face Liveness, it accepts OutputConfig and AuditImagesLimit.
|
936
|
+
*/
|
937
|
+
Settings?: CreateFaceLivenessSessionRequestSettings;
|
938
|
+
/**
|
939
|
+
* Idempotent token is used to recognize the Face Liveness request. If the same token is used with multiple CreateFaceLivenessSession requests, the same session is returned. This token is employed to avoid unintentionally creating the same session multiple times.
|
940
|
+
*/
|
941
|
+
ClientRequestToken?: ClientRequestToken;
|
942
|
+
}
|
943
|
+
export interface CreateFaceLivenessSessionRequestSettings {
|
944
|
+
/**
|
945
|
+
* Can specify the location of an Amazon S3 bucket, where reference and audit images will be stored. Note that the Amazon S3 bucket must be located in the caller's AWS account and in the same region as the Face Liveness end-point. Additionally, the Amazon S3 object keys are auto-generated by the Face Liveness system.
|
946
|
+
*/
|
947
|
+
OutputConfig?: LivenessOutputConfig;
|
948
|
+
/**
|
949
|
+
* Number of audit images to be returned back. Takes an integer between 0-4. Any integer less than 0 will return 0, any integer above 4 will return 4 images in the response. By default, it is set to 0. The limit is best effort and is based on the actual duration of the selfie-video.
|
950
|
+
*/
|
951
|
+
AuditImagesLimit?: AuditImagesLimit;
|
952
|
+
}
|
953
|
+
export interface CreateFaceLivenessSessionResponse {
|
954
|
+
/**
|
955
|
+
* A unique 128-bit UUID identifying a Face Liveness session.
|
956
|
+
*/
|
957
|
+
SessionId: LivenessSessionId;
|
958
|
+
}
|
903
959
|
export interface CreateProjectRequest {
|
904
960
|
/**
|
905
961
|
* The name of the project to create.
|
@@ -1476,7 +1532,7 @@ declare namespace Rekognition {
|
|
1476
1532
|
*/
|
1477
1533
|
Features?: DetectLabelsFeatureList;
|
1478
1534
|
/**
|
1479
|
-
* A list of the filters to be applied to returned detected labels and image properties. Specified filters can be inclusive, exclusive, or a combination of both. Filters can be used for individual labels or label categories. The exact label names or label categories must be supplied. For a full list of labels and label categories, see
|
1535
|
+
* A list of the filters to be applied to returned detected labels and image properties. Specified filters can be inclusive, exclusive, or a combination of both. Filters can be used for individual labels or label categories. The exact label names or label categories must be supplied. For a full list of labels and label categories, see Detecting labels.
|
1480
1536
|
*/
|
1481
1537
|
Settings?: DetectLabelsSettings;
|
1482
1538
|
}
|
@@ -2031,6 +2087,34 @@ declare namespace Rekognition {
|
|
2031
2087
|
*/
|
2032
2088
|
Faces?: FaceDetections;
|
2033
2089
|
}
|
2090
|
+
export interface GetFaceLivenessSessionResultsRequest {
|
2091
|
+
/**
|
2092
|
+
* A unique 128-bit UUID. This is used to uniquely identify the session and also acts as an idempotency token for all operations associated with the session.
|
2093
|
+
*/
|
2094
|
+
SessionId: LivenessSessionId;
|
2095
|
+
}
|
2096
|
+
export interface GetFaceLivenessSessionResultsResponse {
|
2097
|
+
/**
|
2098
|
+
* The sessionId for which this request was called.
|
2099
|
+
*/
|
2100
|
+
SessionId: LivenessSessionId;
|
2101
|
+
/**
|
2102
|
+
* Represents a status corresponding to the state of the session. Possible statuses are: CREATED, IN_PROGRESS, SUCCEEDED, FAILED, EXPIRED.
|
2103
|
+
*/
|
2104
|
+
Status: LivenessSessionStatus;
|
2105
|
+
/**
|
2106
|
+
* Probabalistic confidence score for if the person in the given video was live, represented as a float value between 0 to 100.
|
2107
|
+
*/
|
2108
|
+
Confidence?: Percent;
|
2109
|
+
/**
|
2110
|
+
* A high-quality image from the Face Liveness video that can be used for face comparison or search. It includes a bounding box of the face and the Base64-encoded bytes that return an image. If the CreateFaceLivenessSession request included an OutputConfig argument, the image will be uploaded to an S3Object specified in the output configuration. In case the reference image is not returned, it's recommended to retry the Liveness check.
|
2111
|
+
*/
|
2112
|
+
ReferenceImage?: AuditImage;
|
2113
|
+
/**
|
2114
|
+
* A set of images from the Face Liveness video that can be used for audit purposes. It includes a bounding box of the face and the Base64-encoded bytes that return an image. If the CreateFaceLivenessSession request included an OutputConfig argument, the image will be uploaded to an S3Object specified in the output configuration.
|
2115
|
+
*/
|
2116
|
+
AuditImages?: AuditImages;
|
2117
|
+
}
|
2034
2118
|
export interface GetFaceSearchRequest {
|
2035
2119
|
/**
|
2036
2120
|
* The job identifer for the search request. You get the job identifier from an initial call to StartFaceSearch.
|
@@ -2286,7 +2370,7 @@ declare namespace Rekognition {
|
|
2286
2370
|
export type HumanLoopName = string;
|
2287
2371
|
export interface Image {
|
2288
2372
|
/**
|
2289
|
-
* Blob of image bytes up to 5 MBs.
|
2373
|
+
* Blob of image bytes up to 5 MBs. Note that the maximum image size you can pass to DetectCustomLabels is 4MB.
|
2290
2374
|
*/
|
2291
2375
|
Bytes?: ImageBlob;
|
2292
2376
|
/**
|
@@ -2665,6 +2749,20 @@ declare namespace Rekognition {
|
|
2665
2749
|
*/
|
2666
2750
|
Tags?: TagMap;
|
2667
2751
|
}
|
2752
|
+
export type LivenessImageBlob = Buffer|Uint8Array|Blob|string;
|
2753
|
+
export interface LivenessOutputConfig {
|
2754
|
+
/**
|
2755
|
+
* The path to an AWS Amazon S3 bucket used to store Face Liveness session results.
|
2756
|
+
*/
|
2757
|
+
S3Bucket: S3Bucket;
|
2758
|
+
/**
|
2759
|
+
* The prefix appended to the output files for the Face Liveness session results.
|
2760
|
+
*/
|
2761
|
+
S3KeyPrefix?: LivenessS3KeyPrefix;
|
2762
|
+
}
|
2763
|
+
export type LivenessS3KeyPrefix = string;
|
2764
|
+
export type LivenessSessionId = string;
|
2765
|
+
export type LivenessSessionStatus = "CREATED"|"IN_PROGRESS"|"SUCCEEDED"|"FAILED"|string;
|
2668
2766
|
export type MaxDurationInSecondsULong = number;
|
2669
2767
|
export type MaxFaces = number;
|
2670
2768
|
export type MaxFacesToIndex = number;
|