cdk-lambda-subminute 2.0.276 → 2.0.278
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/.jsii +3 -3
- package/lib/cdk-lambda-subminute.js +3 -3
- package/node_modules/aws-sdk/CHANGELOG.md +16 -1
- package/node_modules/aws-sdk/README.md +1 -1
- package/node_modules/aws-sdk/apis/codepipeline-2015-07-09.min.json +146 -55
- package/node_modules/aws-sdk/apis/marketplacecommerceanalytics-2015-07-01.min.json +9 -3
- package/node_modules/aws-sdk/apis/migrationhub-config-2019-06-30.min.json +15 -0
- package/node_modules/aws-sdk/apis/migrationhubstrategy-2020-02-19.min.json +43 -8
- package/node_modules/aws-sdk/apis/migrationhubstrategy-2020-02-19.paginators.json +6 -0
- package/node_modules/aws-sdk/apis/networkmanager-2019-07-05.min.json +28 -26
- package/node_modules/aws-sdk/apis/opensearchserverless-2021-11-01.min.json +296 -45
- package/node_modules/aws-sdk/apis/opensearchserverless-2021-11-01.paginators.json +5 -0
- package/node_modules/aws-sdk/apis/redshift-serverless-2021-04-21.min.json +3 -1
- package/node_modules/aws-sdk/apis/rekognition-2016-06-27.examples.json +141 -0
- package/node_modules/aws-sdk/apis/rekognition-2016-06-27.min.json +222 -12
- package/node_modules/aws-sdk/apis/rekognition-2016-06-27.paginators.json +5 -0
- package/node_modules/aws-sdk/clients/codepipeline.d.ts +108 -1
- package/node_modules/aws-sdk/clients/ec2.d.ts +3 -3
- package/node_modules/aws-sdk/clients/eks.d.ts +3 -3
- package/node_modules/aws-sdk/clients/marketplacecommerceanalytics.d.ts +10 -10
- package/node_modules/aws-sdk/clients/migrationhubconfig.d.ts +16 -0
- package/node_modules/aws-sdk/clients/migrationhubstrategy.d.ts +57 -1
- package/node_modules/aws-sdk/clients/networkmanager.d.ts +15 -3
- package/node_modules/aws-sdk/clients/opensearchserverless.d.ts +349 -3
- package/node_modules/aws-sdk/clients/redshiftserverless.d.ts +9 -1
- package/node_modules/aws-sdk/clients/rekognition.d.ts +234 -0
- package/node_modules/aws-sdk/dist/aws-sdk-core-react-native.js +6 -6
- package/node_modules/aws-sdk/dist/aws-sdk-react-native.js +22 -22
- package/node_modules/aws-sdk/dist/aws-sdk.js +385 -73
- package/node_modules/aws-sdk/dist/aws-sdk.min.js +86 -86
- package/node_modules/aws-sdk/lib/core.js +1 -1
- package/node_modules/aws-sdk/package.json +1 -1
- package/package.json +6 -6
@@ -316,6 +316,14 @@ declare class Rekognition extends Service {
|
|
316
316
|
* Gets the label detection results of a Amazon Rekognition Video analysis started by StartLabelDetection. The label detection operation is started by a call to StartLabelDetection which returns a job identifier (JobId). When the label detection operation finishes, Amazon Rekognition publishes a completion status to the Amazon Simple Notification Service topic registered in the initial call to StartlabelDetection. To get the results of the label detection operation, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call GetLabelDetection and pass the job identifier (JobId) from the initial call to StartLabelDetection. GetLabelDetection returns an array of detected labels (Labels) sorted by the time the labels were detected. You can also sort by the label name by specifying NAME for the SortBy input parameter. If there is no NAME specified, the default sort is by timestamp. You can select how results are aggregated by using the AggregateBy input parameter. The default aggregation method is TIMESTAMPS. You can also aggregate by SEGMENTS, which aggregates all instances of labels detected in a given segment. The returned Labels array may include the following attributes: Name - The name of the detected label. Confidence - The level of confidence in the label assigned to a detected object. Parents - The ancestor labels for a detected label. GetLabelDetection returns a hierarchical taxonomy of detected labels. For example, a detected car might be assigned the label car. The label car has two parent labels: Vehicle (its parent) and Transportation (its grandparent). The response includes the all ancestors for a label, where every ancestor is a unique label. In the previous example, Car, Vehicle, and Transportation are returned as unique labels in the response. Aliases - Possible Aliases for the label. Categories - The label categories that the detected label belongs to. BoundingBox — Bounding boxes are described for all instances of detected common object labels, returned in an array of Instance objects. An Instance object contains a BoundingBox object, describing the location of the label on the input image. It also includes the confidence for the accuracy of the detected bounding box. Timestamp - Time, in milliseconds from the start of the video, that the label was detected. For aggregation by SEGMENTS, the StartTimestampMillis, EndTimestampMillis, and DurationMillis structures are what define a segment. Although the “Timestamp” structure is still returned with each label, its value is set to be the same as StartTimestampMillis. Timestamp and Bounding box information are returned for detected Instances, only if aggregation is done by TIMESTAMPS. If aggregating by SEGMENTS, information about detected instances isn’t returned. The version of the label model used for the detection is also returned. Note DominantColors isn't returned for Instances, although it is shown as part of the response in the sample seen below. Use MaxResults parameter to limit the number of labels returned. If there are more results than specified in MaxResults, the value of NextToken in the operation response contains a pagination token for getting the next set of results. To get the next page of results, call GetlabelDetection and populate the NextToken request parameter with the token value returned from the previous call to GetLabelDetection.
|
317
317
|
*/
|
318
318
|
getLabelDetection(callback?: (err: AWSError, data: Rekognition.Types.GetLabelDetectionResponse) => void): Request<Rekognition.Types.GetLabelDetectionResponse, AWSError>;
|
319
|
+
/**
|
320
|
+
* Retrieves the results for a given media analysis job. Takes a JobId returned by StartMediaAnalysisJob.
|
321
|
+
*/
|
322
|
+
getMediaAnalysisJob(params: Rekognition.Types.GetMediaAnalysisJobRequest, callback?: (err: AWSError, data: Rekognition.Types.GetMediaAnalysisJobResponse) => void): Request<Rekognition.Types.GetMediaAnalysisJobResponse, AWSError>;
|
323
|
+
/**
|
324
|
+
* Retrieves the results for a given media analysis job. Takes a JobId returned by StartMediaAnalysisJob.
|
325
|
+
*/
|
326
|
+
getMediaAnalysisJob(callback?: (err: AWSError, data: Rekognition.Types.GetMediaAnalysisJobResponse) => void): Request<Rekognition.Types.GetMediaAnalysisJobResponse, AWSError>;
|
319
327
|
/**
|
320
328
|
* Gets the path tracking results of a Amazon Rekognition Video analysis started by StartPersonTracking. The person path tracking operation is started by a call to StartPersonTracking which returns a job identifier (JobId). When the operation finishes, Amazon Rekognition Video publishes a completion status to the Amazon Simple Notification Service topic registered in the initial call to StartPersonTracking. To get the results of the person path tracking operation, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call GetPersonTracking and pass the job identifier (JobId) from the initial call to StartPersonTracking. GetPersonTracking returns an array, Persons, of tracked persons and the time(s) their paths were tracked in the video. GetPersonTracking only returns the default facial attributes (BoundingBox, Confidence, Landmarks, Pose, and Quality). The other facial attributes listed in the Face object of the following response syntax are not returned. For more information, see FaceDetail in the Amazon Rekognition Developer Guide. By default, the array is sorted by the time(s) a person's path is tracked in the video. You can sort by tracked persons by specifying INDEX for the SortBy input parameter. Use the MaxResults parameter to limit the number of items returned. If there are more results than specified in MaxResults, the value of NextToken in the operation response contains a pagination token for getting the next set of results. To get the next page of results, call GetPersonTracking and populate the NextToken request parameter with the token value returned from the previous call to GetPersonTracking.
|
321
329
|
*/
|
@@ -380,6 +388,14 @@ declare class Rekognition extends Service {
|
|
380
388
|
* Returns metadata for faces in the specified collection. This metadata includes information such as the bounding box coordinates, the confidence (that the bounding box contains a face), and face ID. For an example, see Listing Faces in a Collection in the Amazon Rekognition Developer Guide. This operation requires permissions to perform the rekognition:ListFaces action.
|
381
389
|
*/
|
382
390
|
listFaces(callback?: (err: AWSError, data: Rekognition.Types.ListFacesResponse) => void): Request<Rekognition.Types.ListFacesResponse, AWSError>;
|
391
|
+
/**
|
392
|
+
* Returns a list of media analysis jobs. Results are sorted by CreationTimestamp in descending order.
|
393
|
+
*/
|
394
|
+
listMediaAnalysisJobs(params: Rekognition.Types.ListMediaAnalysisJobsRequest, callback?: (err: AWSError, data: Rekognition.Types.ListMediaAnalysisJobsResponse) => void): Request<Rekognition.Types.ListMediaAnalysisJobsResponse, AWSError>;
|
395
|
+
/**
|
396
|
+
* Returns a list of media analysis jobs. Results are sorted by CreationTimestamp in descending order.
|
397
|
+
*/
|
398
|
+
listMediaAnalysisJobs(callback?: (err: AWSError, data: Rekognition.Types.ListMediaAnalysisJobsResponse) => void): Request<Rekognition.Types.ListMediaAnalysisJobsResponse, AWSError>;
|
383
399
|
/**
|
384
400
|
* This operation applies only to Amazon Rekognition Custom Labels. Gets a list of the project policies attached to a project. To attach a project policy to a project, call PutProjectPolicy. To remove a project policy from a project, call DeleteProjectPolicy. This operation requires permissions to perform the rekognition:ListProjectPolicies action.
|
385
401
|
*/
|
@@ -500,6 +516,14 @@ declare class Rekognition extends Service {
|
|
500
516
|
* Starts asynchronous detection of labels in a stored video. Amazon Rekognition Video can detect labels in a video. Labels are instances of real-world entities. This includes objects like flower, tree, and table; events like wedding, graduation, and birthday party; concepts like landscape, evening, and nature; and activities like a person getting out of a car or a person skiing. The video must be stored in an Amazon S3 bucket. Use Video to specify the bucket name and the filename of the video. StartLabelDetection returns a job identifier (JobId) which you use to get the results of the operation. When label detection is finished, Amazon Rekognition Video publishes a completion status to the Amazon Simple Notification Service topic that you specify in NotificationChannel. To get the results of the label detection operation, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call GetLabelDetection and pass the job identifier (JobId) from the initial call to StartLabelDetection. Optional Parameters StartLabelDetection has the GENERAL_LABELS Feature applied by default. This feature allows you to provide filtering criteria to the Settings parameter. You can filter with sets of individual labels or with label categories. You can specify inclusive filters, exclusive filters, or a combination of inclusive and exclusive filters. For more information on filtering, see Detecting labels in a video. You can specify MinConfidence to control the confidence threshold for the labels returned. The default is 50.
|
501
517
|
*/
|
502
518
|
startLabelDetection(callback?: (err: AWSError, data: Rekognition.Types.StartLabelDetectionResponse) => void): Request<Rekognition.Types.StartLabelDetectionResponse, AWSError>;
|
519
|
+
/**
|
520
|
+
* Initiates a new media analysis job. Accepts a manifest file in an Amazon S3 bucket. The output is a manifest file and a summary of the manifest stored in the Amazon S3 bucket.
|
521
|
+
*/
|
522
|
+
startMediaAnalysisJob(params: Rekognition.Types.StartMediaAnalysisJobRequest, callback?: (err: AWSError, data: Rekognition.Types.StartMediaAnalysisJobResponse) => void): Request<Rekognition.Types.StartMediaAnalysisJobResponse, AWSError>;
|
523
|
+
/**
|
524
|
+
* Initiates a new media analysis job. Accepts a manifest file in an Amazon S3 bucket. The output is a manifest file and a summary of the manifest stored in the Amazon S3 bucket.
|
525
|
+
*/
|
526
|
+
startMediaAnalysisJob(callback?: (err: AWSError, data: Rekognition.Types.StartMediaAnalysisJobResponse) => void): Request<Rekognition.Types.StartMediaAnalysisJobResponse, AWSError>;
|
503
527
|
/**
|
504
528
|
* Starts the asynchronous tracking of a person's path in a stored video. Amazon Rekognition Video can track the path of people in a video stored in an Amazon S3 bucket. Use Video to specify the bucket name and the filename of the video. StartPersonTracking returns a job identifier (JobId) which you use to get the results of the operation. When label detection is finished, Amazon Rekognition publishes a completion status to the Amazon Simple Notification Service topic that you specify in NotificationChannel. To get the results of the person detection operation, first check that the status value published to the Amazon SNS topic is SUCCEEDED. If so, call GetPersonTracking and pass the job identifier (JobId) from the initial call to StartPersonTracking.
|
505
529
|
*/
|
@@ -2549,6 +2573,62 @@ declare namespace Rekognition {
|
|
2549
2573
|
*/
|
2550
2574
|
GetRequestMetadata?: GetLabelDetectionRequestMetadata;
|
2551
2575
|
}
|
2576
|
+
export interface GetMediaAnalysisJobRequest {
|
2577
|
+
/**
|
2578
|
+
* Unique identifier for the media analysis job for which you want to retrieve results.
|
2579
|
+
*/
|
2580
|
+
JobId: MediaAnalysisJobId;
|
2581
|
+
}
|
2582
|
+
export interface GetMediaAnalysisJobResponse {
|
2583
|
+
/**
|
2584
|
+
* The identifier for the media analysis job.
|
2585
|
+
*/
|
2586
|
+
JobId: MediaAnalysisJobId;
|
2587
|
+
/**
|
2588
|
+
* The name of the media analysis job.
|
2589
|
+
*/
|
2590
|
+
JobName?: MediaAnalysisJobName;
|
2591
|
+
/**
|
2592
|
+
* Operation configurations that were provided during job creation.
|
2593
|
+
*/
|
2594
|
+
OperationsConfig: MediaAnalysisOperationsConfig;
|
2595
|
+
/**
|
2596
|
+
* The current status of the media analysis job.
|
2597
|
+
*/
|
2598
|
+
Status: MediaAnalysisJobStatus;
|
2599
|
+
/**
|
2600
|
+
* Details about the error that resulted in failure of the job.
|
2601
|
+
*/
|
2602
|
+
FailureDetails?: MediaAnalysisJobFailureDetails;
|
2603
|
+
/**
|
2604
|
+
* The Unix date and time when the job was started.
|
2605
|
+
*/
|
2606
|
+
CreationTimestamp: DateTime;
|
2607
|
+
/**
|
2608
|
+
* The Unix date and time when the job finished.
|
2609
|
+
*/
|
2610
|
+
CompletionTimestamp?: DateTime;
|
2611
|
+
/**
|
2612
|
+
* Reference to the input manifest that was provided in the job creation request.
|
2613
|
+
*/
|
2614
|
+
Input: MediaAnalysisInput;
|
2615
|
+
/**
|
2616
|
+
* Output configuration that was provided in the creation request.
|
2617
|
+
*/
|
2618
|
+
OutputConfig: MediaAnalysisOutputConfig;
|
2619
|
+
/**
|
2620
|
+
* KMS Key that was provided in the creation request.
|
2621
|
+
*/
|
2622
|
+
KmsKeyId?: KmsKeyId;
|
2623
|
+
/**
|
2624
|
+
* Output manifest that contains prediction results.
|
2625
|
+
*/
|
2626
|
+
Results?: MediaAnalysisResults;
|
2627
|
+
/**
|
2628
|
+
* The summary manifest provides statistics on input manifest and errors identified in the input manifest.
|
2629
|
+
*/
|
2630
|
+
ManifestSummary?: MediaAnalysisManifestSummary;
|
2631
|
+
}
|
2552
2632
|
export interface GetPersonTrackingRequest {
|
2553
2633
|
/**
|
2554
2634
|
* The identifier for a job that tracks persons in a video. You get the JobId from a call to StartPersonTracking.
|
@@ -3073,6 +3153,27 @@ declare namespace Rekognition {
|
|
3073
3153
|
*/
|
3074
3154
|
FaceModelVersion?: String;
|
3075
3155
|
}
|
3156
|
+
export type ListMediaAnalysisJobsPageSize = number;
|
3157
|
+
export interface ListMediaAnalysisJobsRequest {
|
3158
|
+
/**
|
3159
|
+
* Pagination token, if the previous response was incomplete.
|
3160
|
+
*/
|
3161
|
+
NextToken?: ExtendedPaginationToken;
|
3162
|
+
/**
|
3163
|
+
* The maximum number of results to return per paginated call. The largest value user can specify is 100. If user specifies a value greater than 100, an InvalidParameterException error occurs. The default value is 100.
|
3164
|
+
*/
|
3165
|
+
MaxResults?: ListMediaAnalysisJobsPageSize;
|
3166
|
+
}
|
3167
|
+
export interface ListMediaAnalysisJobsResponse {
|
3168
|
+
/**
|
3169
|
+
* Pagination token, if the previous response was incomplete.
|
3170
|
+
*/
|
3171
|
+
NextToken?: ExtendedPaginationToken;
|
3172
|
+
/**
|
3173
|
+
* Contains a list of all media analysis jobs.
|
3174
|
+
*/
|
3175
|
+
MediaAnalysisJobs: MediaAnalysisJobDescriptions;
|
3176
|
+
}
|
3076
3177
|
export type ListProjectPoliciesPageSize = number;
|
3077
3178
|
export interface ListProjectPoliciesRequest {
|
3078
3179
|
/**
|
@@ -3184,6 +3285,107 @@ declare namespace Rekognition {
|
|
3184
3285
|
export type MaxPixelThreshold = number;
|
3185
3286
|
export type MaxResults = number;
|
3186
3287
|
export type MaxUserResults = number;
|
3288
|
+
export interface MediaAnalysisDetectModerationLabelsConfig {
|
3289
|
+
/**
|
3290
|
+
* Specifies the minimum confidence level for the moderation labels to return. Amazon Rekognition doesn't return any labels with a confidence level lower than this specified value.
|
3291
|
+
*/
|
3292
|
+
MinConfidence?: Percent;
|
3293
|
+
/**
|
3294
|
+
* Specifies the custom moderation model to be used during the label detection job. If not provided the pre-trained model is used.
|
3295
|
+
*/
|
3296
|
+
ProjectVersion?: ProjectVersionId;
|
3297
|
+
}
|
3298
|
+
export interface MediaAnalysisInput {
|
3299
|
+
S3Object: S3Object;
|
3300
|
+
}
|
3301
|
+
export interface MediaAnalysisJobDescription {
|
3302
|
+
/**
|
3303
|
+
* The identifier for a media analysis job.
|
3304
|
+
*/
|
3305
|
+
JobId: MediaAnalysisJobId;
|
3306
|
+
/**
|
3307
|
+
* The name of a media analysis job.
|
3308
|
+
*/
|
3309
|
+
JobName?: MediaAnalysisJobName;
|
3310
|
+
/**
|
3311
|
+
* Operation configurations that were provided during job creation.
|
3312
|
+
*/
|
3313
|
+
OperationsConfig: MediaAnalysisOperationsConfig;
|
3314
|
+
/**
|
3315
|
+
* The status of the media analysis job being retrieved.
|
3316
|
+
*/
|
3317
|
+
Status: MediaAnalysisJobStatus;
|
3318
|
+
/**
|
3319
|
+
* Details about the error that resulted in failure of the job.
|
3320
|
+
*/
|
3321
|
+
FailureDetails?: MediaAnalysisJobFailureDetails;
|
3322
|
+
/**
|
3323
|
+
* The Unix date and time when the job was started.
|
3324
|
+
*/
|
3325
|
+
CreationTimestamp: DateTime;
|
3326
|
+
/**
|
3327
|
+
* The Unix date and time when the job finished.
|
3328
|
+
*/
|
3329
|
+
CompletionTimestamp?: DateTime;
|
3330
|
+
/**
|
3331
|
+
* Reference to the input manifest that was provided in the job creation request.
|
3332
|
+
*/
|
3333
|
+
Input: MediaAnalysisInput;
|
3334
|
+
/**
|
3335
|
+
* Output configuration that was provided in the creation request.
|
3336
|
+
*/
|
3337
|
+
OutputConfig: MediaAnalysisOutputConfig;
|
3338
|
+
/**
|
3339
|
+
* KMS Key that was provided in the creation request.
|
3340
|
+
*/
|
3341
|
+
KmsKeyId?: KmsKeyId;
|
3342
|
+
/**
|
3343
|
+
* Output manifest that contains prediction results.
|
3344
|
+
*/
|
3345
|
+
Results?: MediaAnalysisResults;
|
3346
|
+
/**
|
3347
|
+
* Provides statistics on input manifest and errors identified in the input manifest.
|
3348
|
+
*/
|
3349
|
+
ManifestSummary?: MediaAnalysisManifestSummary;
|
3350
|
+
}
|
3351
|
+
export type MediaAnalysisJobDescriptions = MediaAnalysisJobDescription[];
|
3352
|
+
export type MediaAnalysisJobFailureCode = "INTERNAL_ERROR"|"INVALID_S3_OBJECT"|"INVALID_MANIFEST"|"INVALID_OUTPUT_CONFIG"|"INVALID_KMS_KEY"|"ACCESS_DENIED"|"RESOURCE_NOT_FOUND"|"RESOURCE_NOT_READY"|"THROTTLED"|string;
|
3353
|
+
export interface MediaAnalysisJobFailureDetails {
|
3354
|
+
/**
|
3355
|
+
* Error code for the failed job.
|
3356
|
+
*/
|
3357
|
+
Code?: MediaAnalysisJobFailureCode;
|
3358
|
+
/**
|
3359
|
+
* Human readable error message.
|
3360
|
+
*/
|
3361
|
+
Message?: String;
|
3362
|
+
}
|
3363
|
+
export type MediaAnalysisJobId = string;
|
3364
|
+
export type MediaAnalysisJobName = string;
|
3365
|
+
export type MediaAnalysisJobStatus = "CREATED"|"QUEUED"|"IN_PROGRESS"|"SUCCEEDED"|"FAILED"|string;
|
3366
|
+
export interface MediaAnalysisManifestSummary {
|
3367
|
+
S3Object?: S3Object;
|
3368
|
+
}
|
3369
|
+
export interface MediaAnalysisOperationsConfig {
|
3370
|
+
/**
|
3371
|
+
* Contains configuration options for a DetectModerationLabels job.
|
3372
|
+
*/
|
3373
|
+
DetectModerationLabels?: MediaAnalysisDetectModerationLabelsConfig;
|
3374
|
+
}
|
3375
|
+
export interface MediaAnalysisOutputConfig {
|
3376
|
+
/**
|
3377
|
+
* Specifies the Amazon S3 bucket to contain the output of the media analysis job.
|
3378
|
+
*/
|
3379
|
+
S3Bucket: S3Bucket;
|
3380
|
+
/**
|
3381
|
+
* Specifies the Amazon S3 key prefix that comes after the name of the bucket you have designated for storage.
|
3382
|
+
*/
|
3383
|
+
S3KeyPrefix?: MediaAnalysisS3KeyPrefix;
|
3384
|
+
}
|
3385
|
+
export interface MediaAnalysisResults {
|
3386
|
+
S3Object?: S3Object;
|
3387
|
+
}
|
3388
|
+
export type MediaAnalysisS3KeyPrefix = string;
|
3187
3389
|
export type MinCoveragePercentage = number;
|
3188
3390
|
export interface ModerationLabel {
|
3189
3391
|
/**
|
@@ -4012,6 +4214,38 @@ declare namespace Rekognition {
|
|
4012
4214
|
*/
|
4013
4215
|
JobId?: JobId;
|
4014
4216
|
}
|
4217
|
+
export interface StartMediaAnalysisJobRequest {
|
4218
|
+
/**
|
4219
|
+
* Idempotency token used to prevent the accidental creation of duplicate versions. If you use the same token with multiple StartMediaAnalysisJobRequest requests, the same response is returned. Use ClientRequestToken to prevent the same request from being processed more than once.
|
4220
|
+
*/
|
4221
|
+
ClientRequestToken?: ClientRequestToken;
|
4222
|
+
/**
|
4223
|
+
* The name of the job. Does not have to be unique.
|
4224
|
+
*/
|
4225
|
+
JobName?: MediaAnalysisJobName;
|
4226
|
+
/**
|
4227
|
+
* Configuration options for the media analysis job to be created.
|
4228
|
+
*/
|
4229
|
+
OperationsConfig: MediaAnalysisOperationsConfig;
|
4230
|
+
/**
|
4231
|
+
* Input data to be analyzed by the job.
|
4232
|
+
*/
|
4233
|
+
Input: MediaAnalysisInput;
|
4234
|
+
/**
|
4235
|
+
* The Amazon S3 bucket location to store the results.
|
4236
|
+
*/
|
4237
|
+
OutputConfig: MediaAnalysisOutputConfig;
|
4238
|
+
/**
|
4239
|
+
* The identifier of customer managed AWS KMS key (name or ARN). The key is used to encrypt images copied into the service. The key is also used to encrypt results and manifest files written to the output Amazon S3 bucket.
|
4240
|
+
*/
|
4241
|
+
KmsKeyId?: KmsKeyId;
|
4242
|
+
}
|
4243
|
+
export interface StartMediaAnalysisJobResponse {
|
4244
|
+
/**
|
4245
|
+
* Identifier for the created job.
|
4246
|
+
*/
|
4247
|
+
JobId: MediaAnalysisJobId;
|
4248
|
+
}
|
4015
4249
|
export interface StartPersonTrackingRequest {
|
4016
4250
|
/**
|
4017
4251
|
* The video in which you want to detect people. The video must be stored in an Amazon S3 bucket.
|
@@ -83,7 +83,7 @@ return /******/ (function(modules) { // webpackBootstrap
|
|
83
83
|
/**
|
84
84
|
* @constant
|
85
85
|
*/
|
86
|
-
VERSION: '2.
|
86
|
+
VERSION: '2.1480.0',
|
87
87
|
|
88
88
|
/**
|
89
89
|
* @api private
|
@@ -9980,16 +9980,16 @@ return /******/ (function(modules) { // webpackBootstrap
|
|
9980
9980
|
|
9981
9981
|
/***/ }),
|
9982
9982
|
/* 63 */
|
9983
|
-
/***/ (function(module, exports) {
|
9983
|
+
/***/ (function(module, exports, __webpack_require__) {
|
9984
9984
|
|
9985
9985
|
'use strict';
|
9986
9986
|
|
9987
|
-
var hasOwnProperty = {}.hasOwnProperty;
|
9988
9987
|
var call = Function.prototype.call;
|
9988
|
+
var $hasOwn = Object.prototype.hasOwnProperty;
|
9989
|
+
var bind = __webpack_require__(61);
|
9989
9990
|
|
9990
|
-
|
9991
|
-
|
9992
|
-
};
|
9991
|
+
/** @type {(o: {}, p: PropertyKey) => p is keyof o} */
|
9992
|
+
module.exports = bind.call(call, $hasOwn);
|
9993
9993
|
|
9994
9994
|
|
9995
9995
|
/***/ }),
|