@aws-sdk/client-rekognition 3.686.0 → 3.691.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -10,12 +10,12 @@ export interface MatchedUser {
10
10
  * <p>A provided ID for the UserID. Unique within the collection.</p>
11
11
  * @public
12
12
  */
13
- UserId?: string;
13
+ UserId?: string | undefined;
14
14
  /**
15
15
  * <p>The status of the user matched to a provided FaceID.</p>
16
16
  * @public
17
17
  */
18
- UserStatus?: UserStatus;
18
+ UserStatus?: UserStatus | undefined;
19
19
  }
20
20
  /**
21
21
  * <p>The Amazon Simple Notification Service topic to which Amazon Rekognition publishes the completion status of a video analysis operation. For more information, see
@@ -55,7 +55,7 @@ export interface PutProjectPolicyRequest {
55
55
  * policy.</p>
56
56
  * @public
57
57
  */
58
- PolicyRevisionId?: string;
58
+ PolicyRevisionId?: string | undefined;
59
59
  /**
60
60
  * <p>A resource policy to add to the model. The policy is a JSON structure that contains
61
61
  * one or more statements that define the policy.
@@ -74,7 +74,7 @@ export interface PutProjectPolicyResponse {
74
74
  * <p>The ID of the project policy.</p>
75
75
  * @public
76
76
  */
77
- PolicyRevisionId?: string;
77
+ PolicyRevisionId?: string | undefined;
78
78
  }
79
79
  /**
80
80
  * @public
@@ -103,12 +103,12 @@ export interface RecognizeCelebritiesResponse {
103
103
  * <code>Urls</code>.</p>
104
104
  * @public
105
105
  */
106
- CelebrityFaces?: Celebrity[];
106
+ CelebrityFaces?: Celebrity[] | undefined;
107
107
  /**
108
108
  * <p>Details about each unrecognized face in the image.</p>
109
109
  * @public
110
110
  */
111
- UnrecognizedFaces?: ComparedFace[];
111
+ UnrecognizedFaces?: ComparedFace[] | undefined;
112
112
  /**
113
113
  * <note>
114
114
  * <p>Support for estimating image orientation using the the OrientationCorrection field
@@ -129,7 +129,7 @@ export interface RecognizeCelebritiesResponse {
129
129
  * </note>
130
130
  * @public
131
131
  */
132
- OrientationCorrection?: OrientationCorrection;
132
+ OrientationCorrection?: OrientationCorrection | undefined;
133
133
  }
134
134
  /**
135
135
  * @public
@@ -150,14 +150,14 @@ export interface SearchFacesRequest {
150
150
  * with the highest confidence in the match.</p>
151
151
  * @public
152
152
  */
153
- MaxFaces?: number;
153
+ MaxFaces?: number | undefined;
154
154
  /**
155
155
  * <p>Optional value specifying the minimum confidence in the face match to return. For
156
156
  * example, don't return any matches where confidence in matches is less than 70%. The default
157
157
  * value is 80%. </p>
158
158
  * @public
159
159
  */
160
- FaceMatchThreshold?: number;
160
+ FaceMatchThreshold?: number | undefined;
161
161
  }
162
162
  /**
163
163
  * @public
@@ -167,19 +167,19 @@ export interface SearchFacesResponse {
167
167
  * <p>ID of the face that was searched for matches in a collection.</p>
168
168
  * @public
169
169
  */
170
- SearchedFaceId?: string;
170
+ SearchedFaceId?: string | undefined;
171
171
  /**
172
172
  * <p>An array of faces that matched the input face, along with the confidence in the
173
173
  * match.</p>
174
174
  * @public
175
175
  */
176
- FaceMatches?: FaceMatch[];
176
+ FaceMatches?: FaceMatch[] | undefined;
177
177
  /**
178
178
  * <p>Version number of the face detection model associated with the input collection
179
179
  * (<code>CollectionId</code>).</p>
180
180
  * @public
181
181
  */
182
- FaceModelVersion?: string;
182
+ FaceModelVersion?: string | undefined;
183
183
  }
184
184
  /**
185
185
  * @public
@@ -204,14 +204,14 @@ export interface SearchFacesByImageRequest {
204
204
  * with the highest confidence in the match.</p>
205
205
  * @public
206
206
  */
207
- MaxFaces?: number;
207
+ MaxFaces?: number | undefined;
208
208
  /**
209
209
  * <p>(Optional) Specifies the minimum confidence in the face match to return. For example,
210
210
  * don't return any matches where confidence in matches is less than 70%. The default value is
211
211
  * 80%.</p>
212
212
  * @public
213
213
  */
214
- FaceMatchThreshold?: number;
214
+ FaceMatchThreshold?: number | undefined;
215
215
  /**
216
216
  * <p>A filter that specifies a quality bar for how much filtering is done to identify faces.
217
217
  * Filtered faces aren't searched for in the collection. If you specify <code>AUTO</code>,
@@ -226,7 +226,7 @@ export interface SearchFacesByImageRequest {
226
226
  * of the face model or higher.</p>
227
227
  * @public
228
228
  */
229
- QualityFilter?: QualityFilter;
229
+ QualityFilter?: QualityFilter | undefined;
230
230
  }
231
231
  /**
232
232
  * @public
@@ -237,25 +237,25 @@ export interface SearchFacesByImageResponse {
237
237
  * search.</p>
238
238
  * @public
239
239
  */
240
- SearchedFaceBoundingBox?: BoundingBox;
240
+ SearchedFaceBoundingBox?: BoundingBox | undefined;
241
241
  /**
242
242
  * <p>The level of confidence that the <code>searchedFaceBoundingBox</code>, contains a
243
243
  * face.</p>
244
244
  * @public
245
245
  */
246
- SearchedFaceConfidence?: number;
246
+ SearchedFaceConfidence?: number | undefined;
247
247
  /**
248
248
  * <p>An array of faces that match the input face, along with the confidence in the
249
249
  * match.</p>
250
250
  * @public
251
251
  */
252
- FaceMatches?: FaceMatch[];
252
+ FaceMatches?: FaceMatch[] | undefined;
253
253
  /**
254
254
  * <p>Version number of the face detection model associated with the input collection
255
255
  * (<code>CollectionId</code>).</p>
256
256
  * @public
257
257
  */
258
- FaceModelVersion?: string;
258
+ FaceModelVersion?: string | undefined;
259
259
  }
260
260
  /**
261
261
  * @public
@@ -271,23 +271,23 @@ export interface SearchUsersRequest {
271
271
  * <p>ID for the existing User.</p>
272
272
  * @public
273
273
  */
274
- UserId?: string;
274
+ UserId?: string | undefined;
275
275
  /**
276
276
  * <p>ID for the existing face.</p>
277
277
  * @public
278
278
  */
279
- FaceId?: string;
279
+ FaceId?: string | undefined;
280
280
  /**
281
281
  * <p>Optional value that specifies the minimum confidence in the matched UserID to return.
282
282
  * Default value of 80.</p>
283
283
  * @public
284
284
  */
285
- UserMatchThreshold?: number;
285
+ UserMatchThreshold?: number | undefined;
286
286
  /**
287
287
  * <p>Maximum number of identities to return.</p>
288
288
  * @public
289
289
  */
290
- MaxUsers?: number;
290
+ MaxUsers?: number | undefined;
291
291
  }
292
292
  /**
293
293
  * <p>Provides face metadata such as FaceId, BoundingBox, Confidence of the input face used for
@@ -299,7 +299,7 @@ export interface SearchedFace {
299
299
  * <p> Unique identifier assigned to the face.</p>
300
300
  * @public
301
301
  */
302
- FaceId?: string;
302
+ FaceId?: string | undefined;
303
303
  }
304
304
  /**
305
305
  * <p>Contains metadata about a User searched for within a collection.</p>
@@ -310,7 +310,7 @@ export interface SearchedUser {
310
310
  * <p> A provided ID for the UserID. Unique within the collection. </p>
311
311
  * @public
312
312
  */
313
- UserId?: string;
313
+ UserId?: string | undefined;
314
314
  }
315
315
  /**
316
316
  * <p>Provides UserID metadata along with the confidence in the match of this UserID with the
@@ -322,12 +322,12 @@ export interface UserMatch {
322
322
  * <p> Describes the UserID metadata.</p>
323
323
  * @public
324
324
  */
325
- Similarity?: number;
325
+ Similarity?: number | undefined;
326
326
  /**
327
327
  * <p> Confidence in the match of this UserID with the input face. </p>
328
328
  * @public
329
329
  */
330
- User?: MatchedUser;
330
+ User?: MatchedUser | undefined;
331
331
  }
332
332
  /**
333
333
  * @public
@@ -338,22 +338,22 @@ export interface SearchUsersResponse {
338
338
  * match. Array will be empty if there are no matches.</p>
339
339
  * @public
340
340
  */
341
- UserMatches?: UserMatch[];
341
+ UserMatches?: UserMatch[] | undefined;
342
342
  /**
343
343
  * <p>Version number of the face detection model associated with the input CollectionId.</p>
344
344
  * @public
345
345
  */
346
- FaceModelVersion?: string;
346
+ FaceModelVersion?: string | undefined;
347
347
  /**
348
348
  * <p>Contains the ID of a face that was used to search for matches in a collection.</p>
349
349
  * @public
350
350
  */
351
- SearchedFace?: SearchedFace;
351
+ SearchedFace?: SearchedFace | undefined;
352
352
  /**
353
353
  * <p>Contains the ID of the UserID that was used to search for matches in a collection.</p>
354
354
  * @public
355
355
  */
356
- SearchedUser?: SearchedUser;
356
+ SearchedUser?: SearchedUser | undefined;
357
357
  }
358
358
  /**
359
359
  * @public
@@ -392,18 +392,18 @@ export interface SearchUsersByImageRequest {
392
392
  * 80.</p>
393
393
  * @public
394
394
  */
395
- UserMatchThreshold?: number;
395
+ UserMatchThreshold?: number | undefined;
396
396
  /**
397
397
  * <p>Maximum number of UserIDs to return.</p>
398
398
  * @public
399
399
  */
400
- MaxUsers?: number;
400
+ MaxUsers?: number | undefined;
401
401
  /**
402
402
  * <p>A filter that specifies a quality bar for how much filtering is done to identify faces.
403
403
  * Filtered faces aren't searched for in the collection. The default value is NONE.</p>
404
404
  * @public
405
405
  */
406
- QualityFilter?: QualityFilter;
406
+ QualityFilter?: QualityFilter | undefined;
407
407
  }
408
408
  /**
409
409
  * <p>Contains data regarding the input face used for a search.</p>
@@ -439,7 +439,7 @@ export interface SearchedFaceDetails {
439
439
  * <code>IndexFaces</code>, use the <code>DetectAttributes</code> input parameter.</p>
440
440
  * @public
441
441
  */
442
- FaceDetail?: FaceDetail;
442
+ FaceDetail?: FaceDetail | undefined;
443
443
  }
444
444
  /**
445
445
  * @public
@@ -494,12 +494,12 @@ export interface UnsearchedFace {
494
494
  * <code>IndexFaces</code>, use the <code>DetectAttributes</code> input parameter.</p>
495
495
  * @public
496
496
  */
497
- FaceDetails?: FaceDetail;
497
+ FaceDetails?: FaceDetail | undefined;
498
498
  /**
499
499
  * <p> Reasons why a face wasn't used for Search. </p>
500
500
  * @public
501
501
  */
502
- Reasons?: UnsearchedFaceReason[];
502
+ Reasons?: UnsearchedFaceReason[] | undefined;
503
503
  }
504
504
  /**
505
505
  * @public
@@ -511,27 +511,27 @@ export interface SearchUsersByImageResponse {
511
511
  * SearchUsersByImageResponse action is successful.</p>
512
512
  * @public
513
513
  */
514
- UserMatches?: UserMatch[];
514
+ UserMatches?: UserMatch[] | undefined;
515
515
  /**
516
516
  * <p>Version number of the face detection model associated with the input collection
517
517
  * CollectionId.</p>
518
518
  * @public
519
519
  */
520
- FaceModelVersion?: string;
520
+ FaceModelVersion?: string | undefined;
521
521
  /**
522
522
  * <p>A list of FaceDetail objects containing the BoundingBox for the largest face in image, as
523
523
  * well as the confidence in the bounding box, that was searched for matches. If no valid face is
524
524
  * detected in the image the response will contain no SearchedFace object.</p>
525
525
  * @public
526
526
  */
527
- SearchedFace?: SearchedFaceDetails;
527
+ SearchedFace?: SearchedFaceDetails | undefined;
528
528
  /**
529
529
  * <p>List of UnsearchedFace objects. Contains the face details infered from the specified image
530
530
  * but not used for search. Contains reasons that describe why a face wasn't used for Search.
531
531
  * </p>
532
532
  * @public
533
533
  */
534
- UnsearchedFaces?: UnsearchedFace[];
534
+ UnsearchedFaces?: UnsearchedFace[] | undefined;
535
535
  }
536
536
  /**
537
537
  * @public
@@ -549,19 +549,19 @@ export interface StartCelebrityRecognitionRequest {
549
549
  * <code>ClientRequestToken</code> to prevent the same job from being accidently started more than once. </p>
550
550
  * @public
551
551
  */
552
- ClientRequestToken?: string;
552
+ ClientRequestToken?: string | undefined;
553
553
  /**
554
554
  * <p>The Amazon SNS topic ARN that you want Amazon Rekognition Video to publish the completion status of the
555
555
  * celebrity recognition analysis to. The Amazon SNS topic must have a topic name that begins with <i>AmazonRekognition</i> if you are using the AmazonRekognitionServiceRole permissions policy.</p>
556
556
  * @public
557
557
  */
558
- NotificationChannel?: NotificationChannel;
558
+ NotificationChannel?: NotificationChannel | undefined;
559
559
  /**
560
560
  * <p>An identifier you specify that's returned in the completion notification that's published to your Amazon Simple Notification Service topic.
561
561
  * For example, you can use <code>JobTag</code> to group related jobs and identify them in the completion notification.</p>
562
562
  * @public
563
563
  */
564
- JobTag?: string;
564
+ JobTag?: string | undefined;
565
565
  }
566
566
  /**
567
567
  * @public
@@ -572,7 +572,7 @@ export interface StartCelebrityRecognitionResponse {
572
572
  * a subsequent call to <code>GetCelebrityRecognition</code>.</p>
573
573
  * @public
574
574
  */
575
- JobId?: string;
575
+ JobId?: string | undefined;
576
576
  }
577
577
  /**
578
578
  * <p>The file size or duration of the supplied media is too large. The maximum file size is 10GB.
@@ -582,13 +582,13 @@ export interface StartCelebrityRecognitionResponse {
582
582
  export declare class VideoTooLargeException extends __BaseException {
583
583
  readonly name: "VideoTooLargeException";
584
584
  readonly $fault: "client";
585
- Message?: string;
586
- Code?: string;
585
+ Message?: string | undefined;
586
+ Code?: string | undefined;
587
587
  /**
588
588
  * <p>A universally unique identifier (UUID) for the request.</p>
589
589
  * @public
590
590
  */
591
- Logref?: string;
591
+ Logref?: string | undefined;
592
592
  /**
593
593
  * @internal
594
594
  */
@@ -612,26 +612,26 @@ export interface StartContentModerationRequest {
612
612
  * returns labels with confidence values greater than or equal to 50 percent.</p>
613
613
  * @public
614
614
  */
615
- MinConfidence?: number;
615
+ MinConfidence?: number | undefined;
616
616
  /**
617
617
  * <p>Idempotent token used to identify the start request. If you use the same token with multiple
618
618
  * <code>StartContentModeration</code> requests, the same <code>JobId</code> is returned. Use
619
619
  * <code>ClientRequestToken</code> to prevent the same job from being accidently started more than once. </p>
620
620
  * @public
621
621
  */
622
- ClientRequestToken?: string;
622
+ ClientRequestToken?: string | undefined;
623
623
  /**
624
624
  * <p>The Amazon SNS topic ARN that you want Amazon Rekognition Video to publish the completion status of the
625
625
  * content analysis to. The Amazon SNS topic must have a topic name that begins with <i>AmazonRekognition</i> if you are using the AmazonRekognitionServiceRole permissions policy to access the topic.</p>
626
626
  * @public
627
627
  */
628
- NotificationChannel?: NotificationChannel;
628
+ NotificationChannel?: NotificationChannel | undefined;
629
629
  /**
630
630
  * <p>An identifier you specify that's returned in the completion notification that's published to your Amazon Simple Notification Service topic.
631
631
  * For example, you can use <code>JobTag</code> to group related jobs and identify them in the completion notification.</p>
632
632
  * @public
633
633
  */
634
- JobTag?: string;
634
+ JobTag?: string | undefined;
635
635
  }
636
636
  /**
637
637
  * @public
@@ -642,7 +642,7 @@ export interface StartContentModerationResponse {
642
642
  * a subsequent call to <code>GetContentModeration</code>.</p>
643
643
  * @public
644
644
  */
645
- JobId?: string;
645
+ JobId?: string | undefined;
646
646
  }
647
647
  /**
648
648
  * @public
@@ -660,13 +660,13 @@ export interface StartFaceDetectionRequest {
660
660
  * <code>ClientRequestToken</code> to prevent the same job from being accidently started more than once. </p>
661
661
  * @public
662
662
  */
663
- ClientRequestToken?: string;
663
+ ClientRequestToken?: string | undefined;
664
664
  /**
665
665
  * <p>The ARN of the Amazon SNS topic to which you want Amazon Rekognition Video to publish the completion status of the
666
666
  * face detection operation. The Amazon SNS topic must have a topic name that begins with <i>AmazonRekognition</i> if you are using the AmazonRekognitionServiceRole permissions policy.</p>
667
667
  * @public
668
668
  */
669
- NotificationChannel?: NotificationChannel;
669
+ NotificationChannel?: NotificationChannel | undefined;
670
670
  /**
671
671
  * <p>The face attributes you want returned.</p>
672
672
  * <p>
@@ -675,13 +675,13 @@ export interface StartFaceDetectionRequest {
675
675
  * <code>ALL</code> - All facial attributes are returned.</p>
676
676
  * @public
677
677
  */
678
- FaceAttributes?: FaceAttributes;
678
+ FaceAttributes?: FaceAttributes | undefined;
679
679
  /**
680
680
  * <p>An identifier you specify that's returned in the completion notification that's published to your Amazon Simple Notification Service topic.
681
681
  * For example, you can use <code>JobTag</code> to group related jobs and identify them in the completion notification.</p>
682
682
  * @public
683
683
  */
684
- JobTag?: string;
684
+ JobTag?: string | undefined;
685
685
  }
686
686
  /**
687
687
  * @public
@@ -692,7 +692,7 @@ export interface StartFaceDetectionResponse {
692
692
  * a subsequent call to <code>GetFaceDetection</code>.</p>
693
693
  * @public
694
694
  */
695
- JobId?: string;
695
+ JobId?: string | undefined;
696
696
  }
697
697
  /**
698
698
  * @public
@@ -709,13 +709,13 @@ export interface StartFaceSearchRequest {
709
709
  * <code>ClientRequestToken</code> to prevent the same job from being accidently started more than once. </p>
710
710
  * @public
711
711
  */
712
- ClientRequestToken?: string;
712
+ ClientRequestToken?: string | undefined;
713
713
  /**
714
714
  * <p>The minimum confidence in the person match to return. For example, don't return any matches where confidence in matches is less than 70%.
715
715
  * The default value is 80%.</p>
716
716
  * @public
717
717
  */
718
- FaceMatchThreshold?: number;
718
+ FaceMatchThreshold?: number | undefined;
719
719
  /**
720
720
  * <p>ID of the collection that contains the faces you want to search for.</p>
721
721
  * @public
@@ -725,13 +725,13 @@ export interface StartFaceSearchRequest {
725
725
  * <p>The ARN of the Amazon SNS topic to which you want Amazon Rekognition Video to publish the completion status of the search. The Amazon SNS topic must have a topic name that begins with <i>AmazonRekognition</i> if you are using the AmazonRekognitionServiceRole permissions policy to access the topic.</p>
726
726
  * @public
727
727
  */
728
- NotificationChannel?: NotificationChannel;
728
+ NotificationChannel?: NotificationChannel | undefined;
729
729
  /**
730
730
  * <p>An identifier you specify that's returned in the completion notification that's published to your Amazon Simple Notification Service topic.
731
731
  * For example, you can use <code>JobTag</code> to group related jobs and identify them in the completion notification.</p>
732
732
  * @public
733
733
  */
734
- JobTag?: string;
734
+ JobTag?: string | undefined;
735
735
  }
736
736
  /**
737
737
  * @public
@@ -741,7 +741,7 @@ export interface StartFaceSearchResponse {
741
741
  * <p>The identifier for the search job. Use <code>JobId</code> to identify the job in a subsequent call to <code>GetFaceSearch</code>. </p>
742
742
  * @public
743
743
  */
744
- JobId?: string;
744
+ JobId?: string | undefined;
745
745
  }
746
746
  /**
747
747
  * @public
@@ -759,7 +759,7 @@ export interface StartLabelDetectionRequest {
759
759
  * <code>ClientRequestToken</code> to prevent the same job from being accidently started more than once. </p>
760
760
  * @public
761
761
  */
762
- ClientRequestToken?: string;
762
+ ClientRequestToken?: string | undefined;
763
763
  /**
764
764
  * <p>Specifies the minimum confidence that Amazon Rekognition Video must have in order to return a detected label. Confidence
765
765
  * represents how certain Amazon Rekognition is that a label is correctly identified.0 is the lowest confidence.
@@ -770,30 +770,30 @@ export interface StartLabelDetectionRequest {
770
770
  * percent.</p>
771
771
  * @public
772
772
  */
773
- MinConfidence?: number;
773
+ MinConfidence?: number | undefined;
774
774
  /**
775
775
  * <p>The Amazon SNS topic ARN you want Amazon Rekognition Video to publish the completion status of the label detection
776
776
  * operation to. The Amazon SNS topic must have a topic name that begins with <i>AmazonRekognition</i> if you are using the AmazonRekognitionServiceRole permissions policy.</p>
777
777
  * @public
778
778
  */
779
- NotificationChannel?: NotificationChannel;
779
+ NotificationChannel?: NotificationChannel | undefined;
780
780
  /**
781
781
  * <p>An identifier you specify that's returned in the completion notification that's published to your Amazon Simple Notification Service topic.
782
782
  * For example, you can use <code>JobTag</code> to group related jobs and identify them in the completion notification.</p>
783
783
  * @public
784
784
  */
785
- JobTag?: string;
785
+ JobTag?: string | undefined;
786
786
  /**
787
787
  * <p>The features to return after video analysis. You can specify that GENERAL_LABELS are returned.</p>
788
788
  * @public
789
789
  */
790
- Features?: LabelDetectionFeatureName[];
790
+ Features?: LabelDetectionFeatureName[] | undefined;
791
791
  /**
792
792
  * <p>The settings for a StartLabelDetection request.Contains the specified parameters for the label detection request of an asynchronous label analysis operation.
793
793
  * Settings can include filters for GENERAL_LABELS.</p>
794
794
  * @public
795
795
  */
796
- Settings?: LabelDetectionSettings;
796
+ Settings?: LabelDetectionSettings | undefined;
797
797
  }
798
798
  /**
799
799
  * @public
@@ -804,7 +804,7 @@ export interface StartLabelDetectionResponse {
804
804
  * a subsequent call to <code>GetLabelDetection</code>. </p>
805
805
  * @public
806
806
  */
807
- JobId?: string;
807
+ JobId?: string | undefined;
808
808
  }
809
809
  /**
810
810
  * @public
@@ -817,12 +817,12 @@ export interface StartMediaAnalysisJobRequest {
817
817
  * processed more than once.</p>
818
818
  * @public
819
819
  */
820
- ClientRequestToken?: string;
820
+ ClientRequestToken?: string | undefined;
821
821
  /**
822
822
  * <p>The name of the job. Does not have to be unique.</p>
823
823
  * @public
824
824
  */
825
- JobName?: string;
825
+ JobName?: string | undefined;
826
826
  /**
827
827
  * <p>Configuration options for the media analysis job to be created.</p>
828
828
  * @public
@@ -844,7 +844,7 @@ export interface StartMediaAnalysisJobRequest {
844
844
  * to encrypt results and manifest files written to the output Amazon S3 bucket.</p>
845
845
  * @public
846
846
  */
847
- KmsKeyId?: string;
847
+ KmsKeyId?: string | undefined;
848
848
  }
849
849
  /**
850
850
  * @public
@@ -872,19 +872,19 @@ export interface StartPersonTrackingRequest {
872
872
  * <code>ClientRequestToken</code> to prevent the same job from being accidently started more than once. </p>
873
873
  * @public
874
874
  */
875
- ClientRequestToken?: string;
875
+ ClientRequestToken?: string | undefined;
876
876
  /**
877
877
  * <p>The Amazon SNS topic ARN you want Amazon Rekognition Video to publish the completion status of the people detection
878
878
  * operation to. The Amazon SNS topic must have a topic name that begins with <i>AmazonRekognition</i> if you are using the AmazonRekognitionServiceRole permissions policy.</p>
879
879
  * @public
880
880
  */
881
- NotificationChannel?: NotificationChannel;
881
+ NotificationChannel?: NotificationChannel | undefined;
882
882
  /**
883
883
  * <p>An identifier you specify that's returned in the completion notification that's published to your Amazon Simple Notification Service topic.
884
884
  * For example, you can use <code>JobTag</code> to group related jobs and identify them in the completion notification.</p>
885
885
  * @public
886
886
  */
887
- JobTag?: string;
887
+ JobTag?: string | undefined;
888
888
  }
889
889
  /**
890
890
  * @public
@@ -895,7 +895,7 @@ export interface StartPersonTrackingResponse {
895
895
  * a subsequent call to <code>GetPersonTracking</code>.</p>
896
896
  * @public
897
897
  */
898
- JobId?: string;
898
+ JobId?: string | undefined;
899
899
  }
900
900
  /**
901
901
  * @public
@@ -920,7 +920,7 @@ export interface StartProjectVersionRequest {
920
920
  * specify a value, Amazon Rekognition Custom Labels doesn't auto-scale the model.</p>
921
921
  * @public
922
922
  */
923
- MaxInferenceUnits?: number;
923
+ MaxInferenceUnits?: number | undefined;
924
924
  }
925
925
  /**
926
926
  * @public
@@ -930,7 +930,7 @@ export interface StartProjectVersionResponse {
930
930
  * <p>The current running status of the model. </p>
931
931
  * @public
932
932
  */
933
- Status?: ProjectVersionStatus;
933
+ Status?: ProjectVersionStatus | undefined;
934
934
  }
935
935
  /**
936
936
  * <p>Filters for the shot detection segments returned by <code>GetSegmentDetection</code>.
@@ -947,7 +947,7 @@ export interface StartShotDetectionFilter {
947
947
  * segments with confidence values greater than or equal to 50 percent.</p>
948
948
  * @public
949
949
  */
950
- MinSegmentConfidence?: number;
950
+ MinSegmentConfidence?: number | undefined;
951
951
  }
952
952
  /**
953
953
  * <p>Filters for the technical segments returned by <a>GetSegmentDetection</a>. For more information,
@@ -964,7 +964,7 @@ export interface StartTechnicalCueDetectionFilter {
964
964
  * segments with confidence values greater than or equal to 50 percent.</p>
965
965
  * @public
966
966
  */
967
- MinSegmentConfidence?: number;
967
+ MinSegmentConfidence?: number | undefined;
968
968
  /**
969
969
  * <p>
970
970
  * A filter that allows you to control the black frame detection by specifying the black levels and pixel coverage of black pixels in a frame.
@@ -972,7 +972,7 @@ export interface StartTechnicalCueDetectionFilter {
972
972
  * </p>
973
973
  * @public
974
974
  */
975
- BlackFrame?: BlackFrame;
975
+ BlackFrame?: BlackFrame | undefined;
976
976
  }
977
977
  /**
978
978
  * <p>Filters applied to the technical cue or shot detection segments.
@@ -985,12 +985,12 @@ export interface StartSegmentDetectionFilters {
985
985
  * <p>Filters that are specific to technical cues.</p>
986
986
  * @public
987
987
  */
988
- TechnicalCueFilter?: StartTechnicalCueDetectionFilter;
988
+ TechnicalCueFilter?: StartTechnicalCueDetectionFilter | undefined;
989
989
  /**
990
990
  * <p>Filters that are specific to shot detections.</p>
991
991
  * @public
992
992
  */
993
- ShotFilter?: StartShotDetectionFilter;
993
+ ShotFilter?: StartShotDetectionFilter | undefined;
994
994
  }
995
995
  /**
996
996
  * @public
@@ -1008,24 +1008,24 @@ export interface StartSegmentDetectionRequest {
1008
1008
  * <code>ClientRequestToken</code> to prevent the same job from being accidently started more than once. </p>
1009
1009
  * @public
1010
1010
  */
1011
- ClientRequestToken?: string;
1011
+ ClientRequestToken?: string | undefined;
1012
1012
  /**
1013
1013
  * <p>The ARN of the Amazon SNS topic to which you want Amazon Rekognition Video to publish the completion status of the
1014
1014
  * segment detection operation. Note that the Amazon SNS topic must have a topic name that begins with <i>AmazonRekognition</i> if you are using the AmazonRekognitionServiceRole permissions policy to access the topic.</p>
1015
1015
  * @public
1016
1016
  */
1017
- NotificationChannel?: NotificationChannel;
1017
+ NotificationChannel?: NotificationChannel | undefined;
1018
1018
  /**
1019
1019
  * <p>An identifier you specify that's returned in the completion notification that's published to your Amazon Simple Notification Service topic.
1020
1020
  * For example, you can use <code>JobTag</code> to group related jobs and identify them in the completion notification.</p>
1021
1021
  * @public
1022
1022
  */
1023
- JobTag?: string;
1023
+ JobTag?: string | undefined;
1024
1024
  /**
1025
1025
  * <p>Filters for technical cue or shot detection.</p>
1026
1026
  * @public
1027
1027
  */
1028
- Filters?: StartSegmentDetectionFilters;
1028
+ Filters?: StartSegmentDetectionFilters | undefined;
1029
1029
  /**
1030
1030
  * <p>An array of segment types to detect in the video. Valid values are TECHNICAL_CUE and SHOT.</p>
1031
1031
  * @public
@@ -1041,7 +1041,7 @@ export interface StartSegmentDetectionResponse {
1041
1041
  * </p>
1042
1042
  * @public
1043
1043
  */
1044
- JobId?: string;
1044
+ JobId?: string | undefined;
1045
1045
  }
1046
1046
  /**
1047
1047
  * <p>This is a required parameter for label detection stream processors and should not be used
@@ -1055,7 +1055,7 @@ export interface StreamProcessingStartSelector {
1055
1055
  * </p>
1056
1056
  * @public
1057
1057
  */
1058
- KVSStreamStartSelector?: KinesisVideoStreamStartSelector;
1058
+ KVSStreamStartSelector?: KinesisVideoStreamStartSelector | undefined;
1059
1059
  }
1060
1060
  /**
1061
1061
  * <p>
@@ -1071,7 +1071,7 @@ export interface StreamProcessingStopSelector {
1071
1071
  * </p>
1072
1072
  * @public
1073
1073
  */
1074
- MaxDurationInSeconds?: number;
1074
+ MaxDurationInSeconds?: number | undefined;
1075
1075
  }
1076
1076
  /**
1077
1077
  * @public
@@ -1091,7 +1091,7 @@ export interface StartStreamProcessorRequest {
1091
1091
  * <p>This is a required parameter for label detection stream processors and should not be used to start a face search stream processor.</p>
1092
1092
  * @public
1093
1093
  */
1094
- StartSelector?: StreamProcessingStartSelector;
1094
+ StartSelector?: StreamProcessingStartSelector | undefined;
1095
1095
  /**
1096
1096
  * <p>
1097
1097
  * Specifies when to stop processing the stream. You can specify a
@@ -1100,7 +1100,7 @@ export interface StartStreamProcessorRequest {
1100
1100
  * <p>This is a required parameter for label detection stream processors and should not be used to start a face search stream processor.</p>
1101
1101
  * @public
1102
1102
  */
1103
- StopSelector?: StreamProcessingStopSelector;
1103
+ StopSelector?: StreamProcessingStopSelector | undefined;
1104
1104
  }
1105
1105
  /**
1106
1106
  * @public
@@ -1112,7 +1112,7 @@ export interface StartStreamProcessorResponse {
1112
1112
  * </p>
1113
1113
  * @public
1114
1114
  */
1115
- SessionId?: string;
1115
+ SessionId?: string | undefined;
1116
1116
  }
1117
1117
  /**
1118
1118
  * <p>Set of optional parameters that let you set the criteria text must meet to be included in your response.
@@ -1125,13 +1125,13 @@ export interface StartTextDetectionFilters {
1125
1125
  * <p>Filters focusing on qualities of the text, such as confidence or size.</p>
1126
1126
  * @public
1127
1127
  */
1128
- WordFilter?: DetectionFilter;
1128
+ WordFilter?: DetectionFilter | undefined;
1129
1129
  /**
1130
1130
  * <p>Filter focusing on a certain area of the frame. Uses a <code>BoundingBox</code> object to set the region
1131
1131
  * of the screen.</p>
1132
1132
  * @public
1133
1133
  */
1134
- RegionsOfInterest?: RegionOfInterest[];
1134
+ RegionsOfInterest?: RegionOfInterest[] | undefined;
1135
1135
  }
1136
1136
  /**
1137
1137
  * @public
@@ -1149,25 +1149,25 @@ export interface StartTextDetectionRequest {
1149
1149
  * from being accidentaly started more than once.</p>
1150
1150
  * @public
1151
1151
  */
1152
- ClientRequestToken?: string;
1152
+ ClientRequestToken?: string | undefined;
1153
1153
  /**
1154
1154
  * <p>The Amazon Simple Notification Service topic to which Amazon Rekognition publishes the completion status of a video analysis operation. For more information, see
1155
1155
  * <a href="https://docs.aws.amazon.com/rekognition/latest/dg/api-video.html">Calling Amazon Rekognition Video operations</a>. Note that the Amazon SNS topic must have a topic name that begins with <i>AmazonRekognition</i> if you are using the AmazonRekognitionServiceRole permissions policy to access the topic.
1156
1156
  * For more information, see <a href="https://docs.aws.amazon.com/rekognition/latest/dg/api-video-roles.html#api-video-roles-all-topics">Giving access to multiple Amazon SNS topics</a>.</p>
1157
1157
  * @public
1158
1158
  */
1159
- NotificationChannel?: NotificationChannel;
1159
+ NotificationChannel?: NotificationChannel | undefined;
1160
1160
  /**
1161
1161
  * <p>An identifier returned in the completion status published by your Amazon Simple Notification Service topic. For example, you can use <code>JobTag</code> to group related jobs
1162
1162
  * and identify them in the completion notification.</p>
1163
1163
  * @public
1164
1164
  */
1165
- JobTag?: string;
1165
+ JobTag?: string | undefined;
1166
1166
  /**
1167
1167
  * <p>Optional parameters that let you set criteria the text must meet to be included in your response.</p>
1168
1168
  * @public
1169
1169
  */
1170
- Filters?: StartTextDetectionFilters;
1170
+ Filters?: StartTextDetectionFilters | undefined;
1171
1171
  }
1172
1172
  /**
1173
1173
  * @public
@@ -1177,7 +1177,7 @@ export interface StartTextDetectionResponse {
1177
1177
  * <p>Identifier for the text detection job. Use <code>JobId</code> to identify the job in a subsequent call to <code>GetTextDetection</code>.</p>
1178
1178
  * @public
1179
1179
  */
1180
- JobId?: string;
1180
+ JobId?: string | undefined;
1181
1181
  }
1182
1182
  /**
1183
1183
  * @public
@@ -1198,7 +1198,7 @@ export interface StopProjectVersionResponse {
1198
1198
  * <p>The current status of the stop operation. </p>
1199
1199
  * @public
1200
1200
  */
1201
- Status?: ProjectVersionStatus;
1201
+ Status?: ProjectVersionStatus | undefined;
1202
1202
  }
1203
1203
  /**
1204
1204
  * @public
@@ -1306,7 +1306,7 @@ export interface StreamProcessorSettingsForUpdate {
1306
1306
  * </p>
1307
1307
  * @public
1308
1308
  */
1309
- ConnectedHomeForUpdate?: ConnectedHomeSettingsForUpdate;
1309
+ ConnectedHomeForUpdate?: ConnectedHomeSettingsForUpdate | undefined;
1310
1310
  }
1311
1311
  /**
1312
1312
  * @public
@@ -1325,14 +1325,14 @@ export interface UpdateStreamProcessorRequest {
1325
1325
  * </p>
1326
1326
  * @public
1327
1327
  */
1328
- SettingsForUpdate?: StreamProcessorSettingsForUpdate;
1328
+ SettingsForUpdate?: StreamProcessorSettingsForUpdate | undefined;
1329
1329
  /**
1330
1330
  * <p>
1331
1331
  * Specifies locations in the frames where Amazon Rekognition checks for objects or people. This is an optional parameter for label detection stream processors.
1332
1332
  * </p>
1333
1333
  * @public
1334
1334
  */
1335
- RegionsOfInterestForUpdate?: RegionOfInterest[];
1335
+ RegionsOfInterestForUpdate?: RegionOfInterest[] | undefined;
1336
1336
  /**
1337
1337
  * <p>
1338
1338
  * Shows whether you are sharing data with Rekognition to improve model performance. You can choose this option at the account level or on a per-stream basis.
@@ -1340,14 +1340,14 @@ export interface UpdateStreamProcessorRequest {
1340
1340
  * </p>
1341
1341
  * @public
1342
1342
  */
1343
- DataSharingPreferenceForUpdate?: StreamProcessorDataSharingPreference;
1343
+ DataSharingPreferenceForUpdate?: StreamProcessorDataSharingPreference | undefined;
1344
1344
  /**
1345
1345
  * <p>
1346
1346
  * A list of parameters you want to delete from the stream processor.
1347
1347
  * </p>
1348
1348
  * @public
1349
1349
  */
1350
- ParametersToDelete?: StreamProcessorParameterToDelete[];
1350
+ ParametersToDelete?: StreamProcessorParameterToDelete[] | undefined;
1351
1351
  }
1352
1352
  /**
1353
1353
  * @public