@aws-sdk/client-mediaconvert 3.379.1 → 3.382.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -7,6 +7,7 @@ import { AccelerationSettings, AccelerationStatus, AudioDescription, AutomatedEn
7
7
  */
8
8
  export interface F4vSettings {
9
9
  /**
10
+ * @public
10
11
  * If set to PROGRESSIVE_DOWNLOAD, the MOOV atom is relocated to the beginning of the archive as required for progressive downloading. Otherwise it is placed normally at the end.
11
12
  */
12
13
  MoovPlacement?: F4vMoovPlacement | string;
@@ -65,14 +66,17 @@ export type M2tsDataPtsControl = (typeof M2tsDataPtsControl)[keyof typeof M2tsDa
65
66
  */
66
67
  export interface DvbNitSettings {
67
68
  /**
69
+ * @public
68
70
  * The numeric value placed in the Network Information Table (NIT).
69
71
  */
70
72
  NetworkId?: number;
71
73
  /**
74
+ * @public
72
75
  * The network name text placed in the network_name_descriptor inside the Network Information Table. Maximum length is 256 characters.
73
76
  */
74
77
  NetworkName?: string;
75
78
  /**
79
+ * @public
76
80
  * The number of milliseconds between instances of this table in the output transport stream.
77
81
  */
78
82
  NitInterval?: number;
@@ -97,18 +101,22 @@ export type OutputSdt = (typeof OutputSdt)[keyof typeof OutputSdt];
97
101
  */
98
102
  export interface DvbSdtSettings {
99
103
  /**
104
+ * @public
100
105
  * Selects method of inserting SDT information into output stream. "Follow input SDT" copies SDT information from input stream to output stream. "Follow input SDT if present" copies SDT information from input stream to output stream if SDT information is present in the input, otherwise it will fall back on the user-defined values. Enter "SDT Manually" means user will enter the SDT information. "No SDT" means output stream will not contain SDT information.
101
106
  */
102
107
  OutputSdt?: OutputSdt | string;
103
108
  /**
109
+ * @public
104
110
  * The number of milliseconds between instances of this table in the output transport stream.
105
111
  */
106
112
  SdtInterval?: number;
107
113
  /**
114
+ * @public
108
115
  * The service name placed in the service_descriptor in the Service Description Table. Maximum length is 256 characters.
109
116
  */
110
117
  ServiceName?: string;
111
118
  /**
119
+ * @public
112
120
  * The service provider name placed in the service_descriptor in the Service Description Table. Maximum length is 256 characters.
113
121
  */
114
122
  ServiceProviderName?: string;
@@ -119,6 +127,7 @@ export interface DvbSdtSettings {
119
127
  */
120
128
  export interface DvbTdtSettings {
121
129
  /**
130
+ * @public
122
131
  * The number of milliseconds between instances of this table in the output transport stream.
123
132
  */
124
133
  TdtInterval?: number;
@@ -225,6 +234,7 @@ export type M2tsRateMode = (typeof M2tsRateMode)[keyof typeof M2tsRateMode];
225
234
  */
226
235
  export interface M2tsScte35Esam {
227
236
  /**
237
+ * @public
228
238
  * Packet Identifier (PID) of the SCTE-35 stream in the transport stream generated by ESAM.
229
239
  */
230
240
  Scte35EsamPid?: number;
@@ -275,158 +285,197 @@ export type M2tsSegmentationStyle = (typeof M2tsSegmentationStyle)[keyof typeof
275
285
  */
276
286
  export interface M2tsSettings {
277
287
  /**
288
+ * @public
278
289
  * Selects between the DVB and ATSC buffer models for Dolby Digital audio.
279
290
  */
280
291
  AudioBufferModel?: M2tsAudioBufferModel | string;
281
292
  /**
293
+ * @public
282
294
  * Specify this setting only when your output will be consumed by a downstream repackaging workflow that is sensitive to very small duration differences between video and audio. For this situation, choose Match video duration. In all other cases, keep the default value, Default codec duration. When you choose Match video duration, MediaConvert pads the output audio streams with silence or trims them to ensure that the total duration of each audio stream is at least as long as the total duration of the video stream. After padding or trimming, the audio stream duration is no more than one frame longer than the video stream. MediaConvert applies audio padding or trimming only to the end of the last segment of the output. For unsegmented outputs, MediaConvert adds padding only to the end of the file. When you keep the default value, any minor discrepancies between audio and video duration will depend on your output audio codec.
283
295
  */
284
296
  AudioDuration?: M2tsAudioDuration | string;
285
297
  /**
298
+ * @public
286
299
  * The number of audio frames to insert for each PES packet.
287
300
  */
288
301
  AudioFramesPerPes?: number;
289
302
  /**
303
+ * @public
290
304
  * Specify the packet identifiers (PIDs) for any elementary audio streams you include in this output. Specify multiple PIDs as a JSON array. Default is the range 482-492.
291
305
  */
292
306
  AudioPids?: number[];
293
307
  /**
308
+ * @public
294
309
  * Specify the output bitrate of the transport stream in bits per second. Setting to 0 lets the muxer automatically determine the appropriate bitrate. Other common values are 3750000, 7500000, and 15000000.
295
310
  */
296
311
  Bitrate?: number;
297
312
  /**
313
+ * @public
298
314
  * Controls what buffer model to use for accurate interleaving. If set to MULTIPLEX, use multiplex buffer model. If set to NONE, this can lead to lower latency, but low-memory devices may not be able to play back the stream without interruptions.
299
315
  */
300
316
  BufferModel?: M2tsBufferModel | string;
301
317
  /**
318
+ * @public
302
319
  * If you select ALIGN_TO_VIDEO, MediaConvert writes captions and data packets with Presentation Timestamp (PTS) values greater than or equal to the first video packet PTS (MediaConvert drops captions and data packets with lesser PTS values). Keep the default value to allow all PTS values.
303
320
  */
304
321
  DataPTSControl?: M2tsDataPtsControl | string;
305
322
  /**
323
+ * @public
306
324
  * Use these settings to insert a DVB Network Information Table (NIT) in the transport stream of this output.
307
325
  */
308
326
  DvbNitSettings?: DvbNitSettings;
309
327
  /**
328
+ * @public
310
329
  * Use these settings to insert a DVB Service Description Table (SDT) in the transport stream of this output.
311
330
  */
312
331
  DvbSdtSettings?: DvbSdtSettings;
313
332
  /**
333
+ * @public
314
334
  * Specify the packet identifiers (PIDs) for DVB subtitle data included in this output. Specify multiple PIDs as a JSON array. Default is the range 460-479.
315
335
  */
316
336
  DvbSubPids?: number[];
317
337
  /**
338
+ * @public
318
339
  * Use these settings to insert a DVB Time and Date Table (TDT) in the transport stream of this output.
319
340
  */
320
341
  DvbTdtSettings?: DvbTdtSettings;
321
342
  /**
343
+ * @public
322
344
  * Specify the packet identifier (PID) for DVB teletext data you include in this output. Default is 499.
323
345
  */
324
346
  DvbTeletextPid?: number;
325
347
  /**
348
+ * @public
326
349
  * When set to VIDEO_AND_FIXED_INTERVALS, audio EBP markers will be added to partitions 3 and 4. The interval between these additional markers will be fixed, and will be slightly shorter than the video EBP marker interval. When set to VIDEO_INTERVAL, these additional markers will not be inserted. Only applicable when EBP segmentation markers are is selected (segmentationMarkers is EBP or EBP_LEGACY).
327
350
  */
328
351
  EbpAudioInterval?: M2tsEbpAudioInterval | string;
329
352
  /**
353
+ * @public
330
354
  * Selects which PIDs to place EBP markers on. They can either be placed only on the video PID, or on both the video PID and all audio PIDs. Only applicable when EBP segmentation markers are is selected (segmentationMarkers is EBP or EBP_LEGACY).
331
355
  */
332
356
  EbpPlacement?: M2tsEbpPlacement | string;
333
357
  /**
358
+ * @public
334
359
  * Controls whether to include the ES Rate field in the PES header.
335
360
  */
336
361
  EsRateInPes?: M2tsEsRateInPes | string;
337
362
  /**
363
+ * @public
338
364
  * Keep the default value unless you know that your audio EBP markers are incorrectly appearing before your video EBP markers. To correct this problem, set this value to Force.
339
365
  */
340
366
  ForceTsVideoEbpOrder?: M2tsForceTsVideoEbpOrder | string;
341
367
  /**
368
+ * @public
342
369
  * The length, in seconds, of each fragment. Only used with EBP markers.
343
370
  */
344
371
  FragmentTime?: number;
345
372
  /**
373
+ * @public
346
374
  * To include key-length-value metadata in this output: Set KLV metadata insertion to Passthrough. MediaConvert reads KLV metadata present in your input and passes it through to the output transport stream. To exclude this KLV metadata: Set KLV metadata insertion to None or leave blank.
347
375
  */
348
376
  KlvMetadata?: M2tsKlvMetadata | string;
349
377
  /**
378
+ * @public
350
379
  * Specify the maximum time, in milliseconds, between Program Clock References (PCRs) inserted into the transport stream.
351
380
  */
352
381
  MaxPcrInterval?: number;
353
382
  /**
383
+ * @public
354
384
  * When set, enforces that Encoder Boundary Points do not come within the specified time interval of each other by looking ahead at input video. If another EBP is going to come in within the specified time interval, the current EBP is not emitted, and the segment is "stretched" to the next marker. The lookahead value does not add latency to the system. The Live Event must be configured elsewhere to create sufficient latency to make the lookahead accurate.
355
385
  */
356
386
  MinEbpInterval?: number;
357
387
  /**
388
+ * @public
358
389
  * If INSERT, Nielsen inaudible tones for media tracking will be detected in the input audio and an equivalent ID3 tag will be inserted in the output.
359
390
  */
360
391
  NielsenId3?: M2tsNielsenId3 | string;
361
392
  /**
393
+ * @public
362
394
  * Value in bits per second of extra null packets to insert into the transport stream. This can be used if a downstream encryption system requires periodic null packets.
363
395
  */
364
396
  NullPacketBitrate?: number;
365
397
  /**
398
+ * @public
366
399
  * The number of milliseconds between instances of this table in the output transport stream.
367
400
  */
368
401
  PatInterval?: number;
369
402
  /**
403
+ * @public
370
404
  * When set to PCR_EVERY_PES_PACKET, a Program Clock Reference value is inserted for every Packetized Elementary Stream (PES) header. This is effective only when the PCR PID is the same as the video or audio elementary stream.
371
405
  */
372
406
  PcrControl?: M2tsPcrControl | string;
373
407
  /**
408
+ * @public
374
409
  * Specify the packet identifier (PID) for the program clock reference (PCR) in this output. If you do not specify a value, the service will use the value for Video PID.
375
410
  */
376
411
  PcrPid?: number;
377
412
  /**
413
+ * @public
378
414
  * Specify the number of milliseconds between instances of the program map table (PMT) in the output transport stream.
379
415
  */
380
416
  PmtInterval?: number;
381
417
  /**
418
+ * @public
382
419
  * Specify the packet identifier (PID) for the program map table (PMT) itself. Default is 480.
383
420
  */
384
421
  PmtPid?: number;
385
422
  /**
423
+ * @public
386
424
  * Specify the packet identifier (PID) of the private metadata stream. Default is 503.
387
425
  */
388
426
  PrivateMetadataPid?: number;
389
427
  /**
428
+ * @public
390
429
  * Use Program number to specify the program number used in the program map table (PMT) for this output. Default is 1. Program numbers and program map tables are parts of MPEG-2 transport stream containers, used for organizing data.
391
430
  */
392
431
  ProgramNumber?: number;
393
432
  /**
433
+ * @public
394
434
  * When set to CBR, inserts null packets into transport stream to fill specified bitrate. When set to VBR, the bitrate setting acts as the maximum bitrate, but the output will not be padded up to that bitrate.
395
435
  */
396
436
  RateMode?: M2tsRateMode | string;
397
437
  /**
438
+ * @public
398
439
  * Include this in your job settings to put SCTE-35 markers in your HLS and transport stream outputs at the insertion points that you specify in an ESAM XML document. Provide the document in the setting SCC XML.
399
440
  */
400
441
  Scte35Esam?: M2tsScte35Esam;
401
442
  /**
443
+ * @public
402
444
  * Specify the packet identifier (PID) of the SCTE-35 stream in the transport stream.
403
445
  */
404
446
  Scte35Pid?: number;
405
447
  /**
448
+ * @public
406
449
  * For SCTE-35 markers from your input-- Choose Passthrough if you want SCTE-35 markers that appear in your input to also appear in this output. Choose None if you don't want SCTE-35 markers in this output. For SCTE-35 markers from an ESAM XML document-- Choose None. Also provide the ESAM XML as a string in the setting Signal processing notification XML. Also enable ESAM SCTE-35 (include the property scte35Esam).
407
450
  */
408
451
  Scte35Source?: M2tsScte35Source | string;
409
452
  /**
453
+ * @public
410
454
  * Inserts segmentation markers at each segmentation_time period. rai_segstart sets the Random Access Indicator bit in the adaptation field. rai_adapt sets the RAI bit and adds the current timecode in the private data bytes. psi_segstart inserts PAT and PMT tables at the start of segments. ebp adds Encoder Boundary Point information to the adaptation field as per OpenCable specification OC-SP-EBP-I01-130118. ebp_legacy adds Encoder Boundary Point information to the adaptation field using a legacy proprietary format.
411
455
  */
412
456
  SegmentationMarkers?: M2tsSegmentationMarkers | string;
413
457
  /**
458
+ * @public
414
459
  * The segmentation style parameter controls how segmentation markers are inserted into the transport stream. With avails, it is possible that segments may be truncated, which can influence where future segmentation markers are inserted. When a segmentation style of "reset_cadence" is selected and a segment is truncated due to an avail, we will reset the segmentation cadence. This means the subsequent segment will have a duration of of $segmentation_time seconds. When a segmentation style of "maintain_cadence" is selected and a segment is truncated due to an avail, we will not reset the segmentation cadence. This means the subsequent segment will likely be truncated as well. However, all segments after that will have a duration of $segmentation_time seconds. Note that EBP lookahead is a slight exception to this rule.
415
460
  */
416
461
  SegmentationStyle?: M2tsSegmentationStyle | string;
417
462
  /**
463
+ * @public
418
464
  * Specify the length, in seconds, of each segment. Required unless markers is set to _none_.
419
465
  */
420
466
  SegmentationTime?: number;
421
467
  /**
468
+ * @public
422
469
  * Packet Identifier (PID) of the ID3 metadata stream in the transport stream.
423
470
  */
424
471
  TimedMetadataPid?: number;
425
472
  /**
473
+ * @public
426
474
  * Specify the ID for the transport stream itself in the program map table for this output. Transport stream IDs and program map tables are parts of MPEG-2 transport stream containers, used for organizing data.
427
475
  */
428
476
  TransportStreamId?: number;
429
477
  /**
478
+ * @public
430
479
  * Specify the packet identifier (PID) of the elementary video stream in the transport stream.
431
480
  */
432
481
  VideoPid?: number;
@@ -509,78 +558,97 @@ export type TimedMetadata = (typeof TimedMetadata)[keyof typeof TimedMetadata];
509
558
  */
510
559
  export interface M3u8Settings {
511
560
  /**
561
+ * @public
512
562
  * Specify this setting only when your output will be consumed by a downstream repackaging workflow that is sensitive to very small duration differences between video and audio. For this situation, choose Match video duration. In all other cases, keep the default value, Default codec duration. When you choose Match video duration, MediaConvert pads the output audio streams with silence or trims them to ensure that the total duration of each audio stream is at least as long as the total duration of the video stream. After padding or trimming, the audio stream duration is no more than one frame longer than the video stream. MediaConvert applies audio padding or trimming only to the end of the last segment of the output. For unsegmented outputs, MediaConvert adds padding only to the end of the file. When you keep the default value, any minor discrepancies between audio and video duration will depend on your output audio codec.
513
563
  */
514
564
  AudioDuration?: M3u8AudioDuration | string;
515
565
  /**
566
+ * @public
516
567
  * The number of audio frames to insert for each PES packet.
517
568
  */
518
569
  AudioFramesPerPes?: number;
519
570
  /**
571
+ * @public
520
572
  * Packet Identifier (PID) of the elementary audio stream(s) in the transport stream. Multiple values are accepted, and can be entered in ranges and/or by comma separation.
521
573
  */
522
574
  AudioPids?: number[];
523
575
  /**
576
+ * @public
524
577
  * If you select ALIGN_TO_VIDEO, MediaConvert writes captions and data packets with Presentation Timestamp (PTS) values greater than or equal to the first video packet PTS (MediaConvert drops captions and data packets with lesser PTS values). Keep the default value AUTO to allow all PTS values.
525
578
  */
526
579
  DataPTSControl?: M3u8DataPtsControl | string;
527
580
  /**
581
+ * @public
528
582
  * Specify the maximum time, in milliseconds, between Program Clock References (PCRs) inserted into the transport stream.
529
583
  */
530
584
  MaxPcrInterval?: number;
531
585
  /**
586
+ * @public
532
587
  * If INSERT, Nielsen inaudible tones for media tracking will be detected in the input audio and an equivalent ID3 tag will be inserted in the output.
533
588
  */
534
589
  NielsenId3?: M3u8NielsenId3 | string;
535
590
  /**
591
+ * @public
536
592
  * The number of milliseconds between instances of this table in the output transport stream.
537
593
  */
538
594
  PatInterval?: number;
539
595
  /**
596
+ * @public
540
597
  * When set to PCR_EVERY_PES_PACKET a Program Clock Reference value is inserted for every Packetized Elementary Stream (PES) header. This parameter is effective only when the PCR PID is the same as the video or audio elementary stream.
541
598
  */
542
599
  PcrControl?: M3u8PcrControl | string;
543
600
  /**
601
+ * @public
544
602
  * Packet Identifier (PID) of the Program Clock Reference (PCR) in the transport stream. When no value is given, the encoder will assign the same value as the Video PID.
545
603
  */
546
604
  PcrPid?: number;
547
605
  /**
606
+ * @public
548
607
  * The number of milliseconds between instances of this table in the output transport stream.
549
608
  */
550
609
  PmtInterval?: number;
551
610
  /**
611
+ * @public
552
612
  * Packet Identifier (PID) for the Program Map Table (PMT) in the transport stream.
553
613
  */
554
614
  PmtPid?: number;
555
615
  /**
616
+ * @public
556
617
  * Packet Identifier (PID) of the private metadata stream in the transport stream.
557
618
  */
558
619
  PrivateMetadataPid?: number;
559
620
  /**
621
+ * @public
560
622
  * The value of the program number field in the Program Map Table.
561
623
  */
562
624
  ProgramNumber?: number;
563
625
  /**
626
+ * @public
564
627
  * Packet Identifier (PID) of the SCTE-35 stream in the transport stream.
565
628
  */
566
629
  Scte35Pid?: number;
567
630
  /**
631
+ * @public
568
632
  * For SCTE-35 markers from your input-- Choose Passthrough if you want SCTE-35 markers that appear in your input to also appear in this output. Choose None if you don't want SCTE-35 markers in this output. For SCTE-35 markers from an ESAM XML document-- Choose None if you don't want manifest conditioning. Choose Passthrough and choose Ad markers if you do want manifest conditioning. In both cases, also provide the ESAM XML as a string in the setting Signal processing notification XML.
569
633
  */
570
634
  Scte35Source?: M3u8Scte35Source | string;
571
635
  /**
636
+ * @public
572
637
  * Set ID3 metadata to Passthrough to include ID3 metadata in this output. This includes ID3 metadata from the following features: ID3 timestamp period, and Custom ID3 metadata inserter. To exclude this ID3 metadata in this output: set ID3 metadata to None or leave blank.
573
638
  */
574
639
  TimedMetadata?: TimedMetadata | string;
575
640
  /**
641
+ * @public
576
642
  * Packet Identifier (PID) of the ID3 metadata stream in the transport stream.
577
643
  */
578
644
  TimedMetadataPid?: number;
579
645
  /**
646
+ * @public
580
647
  * The value of the transport stream ID field in the Program Map Table.
581
648
  */
582
649
  TransportStreamId?: number;
583
650
  /**
651
+ * @public
584
652
  * Packet Identifier (PID) of the elementary video stream in the transport stream.
585
653
  */
586
654
  VideoPid?: number;
@@ -651,22 +719,27 @@ export type MovReference = (typeof MovReference)[keyof typeof MovReference];
651
719
  */
652
720
  export interface MovSettings {
653
721
  /**
722
+ * @public
654
723
  * When enabled, include 'clap' atom if appropriate for the video output settings.
655
724
  */
656
725
  ClapAtom?: MovClapAtom | string;
657
726
  /**
727
+ * @public
658
728
  * When enabled, file composition times will start at zero, composition times in the 'ctts' (composition time to sample) box for B-frames will be negative, and a 'cslg' (composition shift least greatest) box will be included per 14496-1 amendment 1. This improves compatibility with Apple players and tools.
659
729
  */
660
730
  CslgAtom?: MovCslgAtom | string;
661
731
  /**
732
+ * @public
662
733
  * When set to XDCAM, writes MPEG2 video streams into the QuickTime file using XDCAM fourcc codes. This increases compatibility with Apple editors and players, but may decrease compatibility with other players. Only applicable when the video codec is MPEG2.
663
734
  */
664
735
  Mpeg2FourCCControl?: MovMpeg2FourCCControl | string;
665
736
  /**
737
+ * @public
666
738
  * Unless you need Omneon compatibility: Keep the default value, None. To make this output compatible with Omneon: Choose Omneon. When you do, MediaConvert increases the length of the 'elst' edit list atom. Note that this might cause file rejections when a recipient of the output file doesn't expect this extra padding.
667
739
  */
668
740
  PaddingControl?: MovPaddingControl | string;
669
741
  /**
742
+ * @public
670
743
  * Always keep the default value (SELF_CONTAINED) for this setting.
671
744
  */
672
745
  Reference?: MovReference | string;
@@ -713,26 +786,32 @@ export type Mp4MoovPlacement = (typeof Mp4MoovPlacement)[keyof typeof Mp4MoovPla
713
786
  */
714
787
  export interface Mp4Settings {
715
788
  /**
789
+ * @public
716
790
  * Specify this setting only when your output will be consumed by a downstream repackaging workflow that is sensitive to very small duration differences between video and audio. For this situation, choose Match video duration. In all other cases, keep the default value, Default codec duration. When you choose Match video duration, MediaConvert pads the output audio streams with silence or trims them to ensure that the total duration of each audio stream is at least as long as the total duration of the video stream. After padding or trimming, the audio stream duration is no more than one frame longer than the video stream. MediaConvert applies audio padding or trimming only to the end of the last segment of the output. For unsegmented outputs, MediaConvert adds padding only to the end of the file. When you keep the default value, any minor discrepancies between audio and video duration will depend on your output audio codec.
717
791
  */
718
792
  AudioDuration?: CmfcAudioDuration | string;
719
793
  /**
794
+ * @public
720
795
  * When enabled, file composition times will start at zero, composition times in the 'ctts' (composition time to sample) box for B-frames will be negative, and a 'cslg' (composition shift least greatest) box will be included per 14496-1 amendment 1. This improves compatibility with Apple players and tools.
721
796
  */
722
797
  CslgAtom?: Mp4CslgAtom | string;
723
798
  /**
799
+ * @public
724
800
  * Ignore this setting unless compliance to the CTTS box version specification matters in your workflow. Specify a value of 1 to set your CTTS box version to 1 and make your output compliant with the specification. When you specify a value of 1, you must also set CSLG atom to the value INCLUDE. Keep the default value 0 to set your CTTS box version to 0. This can provide backward compatibility for some players and packagers.
725
801
  */
726
802
  CttsVersion?: number;
727
803
  /**
804
+ * @public
728
805
  * Inserts a free-space box immediately after the moov box.
729
806
  */
730
807
  FreeSpaceBox?: Mp4FreeSpaceBox | string;
731
808
  /**
809
+ * @public
732
810
  * If set to PROGRESSIVE_DOWNLOAD, the MOOV atom is relocated to the beginning of the archive as required for progressive downloading. Otherwise it is placed normally at the end.
733
811
  */
734
812
  MoovPlacement?: Mp4MoovPlacement | string;
735
813
  /**
814
+ * @public
736
815
  * Overrides the "Major Brand" field in the output file. Usually not necessary to specify.
737
816
  */
738
817
  Mp4MajorBrand?: string;
@@ -851,38 +930,47 @@ export type MpdTimedMetadataBoxVersion = (typeof MpdTimedMetadataBoxVersion)[key
851
930
  */
852
931
  export interface MpdSettings {
853
932
  /**
933
+ * @public
854
934
  * Optional. Choose Include to have MediaConvert mark up your DASH manifest with <Accessibility> elements for embedded 608 captions. This markup isn't generally required, but some video players require it to discover and play embedded 608 captions. Keep the default value, Exclude, to leave these elements out. When you enable this setting, this is the markup that MediaConvert includes in your manifest: <Accessibility schemeIdUri="urn:scte:dash:cc:cea-608:2015" value="CC1=eng"/>
855
935
  */
856
936
  AccessibilityCaptionHints?: MpdAccessibilityCaptionHints | string;
857
937
  /**
938
+ * @public
858
939
  * Specify this setting only when your output will be consumed by a downstream repackaging workflow that is sensitive to very small duration differences between video and audio. For this situation, choose Match video duration. In all other cases, keep the default value, Default codec duration. When you choose Match video duration, MediaConvert pads the output audio streams with silence or trims them to ensure that the total duration of each audio stream is at least as long as the total duration of the video stream. After padding or trimming, the audio stream duration is no more than one frame longer than the video stream. MediaConvert applies audio padding or trimming only to the end of the last segment of the output. For unsegmented outputs, MediaConvert adds padding only to the end of the file. When you keep the default value, any minor discrepancies between audio and video duration will depend on your output audio codec.
859
940
  */
860
941
  AudioDuration?: MpdAudioDuration | string;
861
942
  /**
943
+ * @public
862
944
  * Use this setting only in DASH output groups that include sidecar TTML or IMSC captions. You specify sidecar captions in a separate output from your audio and video. Choose Raw for captions in a single XML file in a raw container. Choose Fragmented MPEG-4 for captions in XML format contained within fragmented MP4 files. This set of fragmented MP4 files is separate from your video and audio fragmented MP4 files.
863
945
  */
864
946
  CaptionContainerType?: MpdCaptionContainerType | string;
865
947
  /**
948
+ * @public
866
949
  * To include key-length-value metadata in this output: Set KLV metadata insertion to Passthrough. MediaConvert reads KLV metadata present in your input and writes each instance to a separate event message box in the output, according to MISB ST1910.1. To exclude this KLV metadata: Set KLV metadata insertion to None or leave blank.
867
950
  */
868
951
  KlvMetadata?: MpdKlvMetadata | string;
869
952
  /**
953
+ * @public
870
954
  * To add an InbandEventStream element in your output MPD manifest for each type of event message, set Manifest metadata signaling to Enabled. For ID3 event messages, the InbandEventStream element schemeIdUri will be same value that you specify for ID3 metadata scheme ID URI. For SCTE35 event messages, the InbandEventStream element schemeIdUri will be "urn:scte:scte35:2013:bin". To leave these elements out of your output MPD manifest, set Manifest metadata signaling to Disabled. To enable Manifest metadata signaling, you must also set SCTE-35 source to Passthrough, ESAM SCTE-35 to insert, or ID3 metadata to Passthrough.
871
955
  */
872
956
  ManifestMetadataSignaling?: MpdManifestMetadataSignaling | string;
873
957
  /**
958
+ * @public
874
959
  * Use this setting only when you specify SCTE-35 markers from ESAM. Choose INSERT to put SCTE-35 markers in this output at the insertion points that you specify in an ESAM XML document. Provide the document in the setting SCC XML.
875
960
  */
876
961
  Scte35Esam?: MpdScte35Esam | string;
877
962
  /**
963
+ * @public
878
964
  * Ignore this setting unless you have SCTE-35 markers in your input video file. Choose Passthrough if you want SCTE-35 markers that appear in your input to also appear in this output. Choose None if you don't want those SCTE-35 markers in this output.
879
965
  */
880
966
  Scte35Source?: MpdScte35Source | string;
881
967
  /**
968
+ * @public
882
969
  * To include ID3 metadata in this output: Set ID3 metadata to Passthrough. Specify this ID3 metadata in Custom ID3 metadata inserter. MediaConvert writes each instance of ID3 metadata in a separate Event Message (eMSG) box. To exclude this ID3 metadata: Set ID3 metadata to None or leave blank.
883
970
  */
884
971
  TimedMetadata?: MpdTimedMetadata | string;
885
972
  /**
973
+ * @public
886
974
  * Specify the event message box (eMSG) version for ID3 timed metadata in your output.
887
975
  * For more information, see ISO/IEC 23009-1:2022 section 5.10.3.3.3 Syntax.
888
976
  * Leave blank to use the default value Version 0.
@@ -890,10 +978,12 @@ export interface MpdSettings {
890
978
  */
891
979
  TimedMetadataBoxVersion?: MpdTimedMetadataBoxVersion | string;
892
980
  /**
981
+ * @public
893
982
  * Specify the event message box (eMSG) scheme ID URI for ID3 timed metadata in your output. For more information, see ISO/IEC 23009-1:2022 section 5.10.3.3.4 Semantics. Leave blank to use the default value: https://aomedia.org/emsg/ID3 When you specify a value for ID3 metadata scheme ID URI, you must also set ID3 metadata to Passthrough.
894
983
  */
895
984
  TimedMetadataSchemeIdUri?: string;
896
985
  /**
986
+ * @public
897
987
  * Specify the event message box (eMSG) value for ID3 timed metadata in your output. For more information, see ISO/IEC 23009-1:2022 section 5.10.3.3.4 Semantics. When you specify a value for ID3 Metadata Value, you must also set ID3 metadata to Passthrough.
898
988
  */
899
989
  TimedMetadataValue?: string;
@@ -943,10 +1033,12 @@ export type MxfXavcDurationMode = (typeof MxfXavcDurationMode)[keyof typeof MxfX
943
1033
  */
944
1034
  export interface MxfXavcProfileSettings {
945
1035
  /**
1036
+ * @public
946
1037
  * To create an output that complies with the XAVC file format guidelines for interoperability, keep the default value, Drop frames for compliance. To include all frames from your input in this output, keep the default setting, Allow any duration. The number of frames that MediaConvert excludes when you set this to Drop frames for compliance depends on the output frame rate and duration.
947
1038
  */
948
1039
  DurationMode?: MxfXavcDurationMode | string;
949
1040
  /**
1041
+ * @public
950
1042
  * Specify a value for this setting only for outputs that you set up with one of these two XAVC profiles: XAVC HD Intra CBG or XAVC 4K Intra CBG. Specify the amount of space in each frame that the service reserves for ancillary data, such as teletext captions. The default value for this setting is 1492 bytes per frame. This should be sufficient to prevent overflow unless you have multiple pages of teletext captions data. If you have a large amount of teletext data, specify a larger number.
951
1043
  */
952
1044
  MaxAncDataSize?: number;
@@ -957,14 +1049,17 @@ export interface MxfXavcProfileSettings {
957
1049
  */
958
1050
  export interface MxfSettings {
959
1051
  /**
1052
+ * @public
960
1053
  * Optional. When you have AFD signaling set up in your output video stream, use this setting to choose whether to also include it in the MXF wrapper. Choose Don't copy to exclude AFD signaling from the MXF wrapper. Choose Copy from video stream to copy the AFD values from the video stream for this output to the MXF wrapper. Regardless of which option you choose, the AFD values remain in the video stream. Related settings: To set up your output to include or exclude AFD values, see AfdSignaling, under VideoDescription. On the console, find AFD signaling under the output's video encoding settings.
961
1054
  */
962
1055
  AfdSignaling?: MxfAfdSignaling | string;
963
1056
  /**
1057
+ * @public
964
1058
  * Specify the MXF profile, also called shim, for this output. To automatically select a profile according to your output video codec and resolution, leave blank. For a list of codecs supported with each MXF profile, see https://docs.aws.amazon.com/mediaconvert/latest/ug/codecs-supported-with-each-mxf-profile.html. For more information about the automatic selection behavior, see https://docs.aws.amazon.com/mediaconvert/latest/ug/default-automatic-selection-of-mxf-profiles.html.
965
1059
  */
966
1060
  Profile?: MxfProfile | string;
967
1061
  /**
1062
+ * @public
968
1063
  * Specify the XAVC profile settings for MXF outputs when you set your MXF profile to XAVC.
969
1064
  */
970
1065
  XavcProfileSettings?: MxfXavcProfileSettings;
@@ -975,38 +1070,47 @@ export interface MxfSettings {
975
1070
  */
976
1071
  export interface ContainerSettings {
977
1072
  /**
1073
+ * @public
978
1074
  * These settings relate to the fragmented MP4 container for the segments in your CMAF outputs.
979
1075
  */
980
1076
  CmfcSettings?: CmfcSettings;
981
1077
  /**
1078
+ * @public
982
1079
  * Container for this output. Some containers require a container settings object. If not specified, the default object will be created.
983
1080
  */
984
1081
  Container?: ContainerType | string;
985
1082
  /**
1083
+ * @public
986
1084
  * Settings for F4v container
987
1085
  */
988
1086
  F4vSettings?: F4vSettings;
989
1087
  /**
1088
+ * @public
990
1089
  * MPEG-2 TS container settings. These apply to outputs in a File output group when the output's container is MPEG-2 Transport Stream (M2TS). In these assets, data is organized by the program map table (PMT). Each transport stream program contains subsets of data, including audio, video, and metadata. Each of these subsets of data has a numerical label called a packet identifier (PID). Each transport stream program corresponds to one MediaConvert output. The PMT lists the types of data in a program along with their PID. Downstream systems and players use the program map table to look up the PID for each type of data it accesses and then uses the PIDs to locate specific data within the asset.
991
1090
  */
992
1091
  M2tsSettings?: M2tsSettings;
993
1092
  /**
1093
+ * @public
994
1094
  * These settings relate to the MPEG-2 transport stream (MPEG2-TS) container for the MPEG2-TS segments in your HLS outputs.
995
1095
  */
996
1096
  M3u8Settings?: M3u8Settings;
997
1097
  /**
1098
+ * @public
998
1099
  * These settings relate to your QuickTime MOV output container.
999
1100
  */
1000
1101
  MovSettings?: MovSettings;
1001
1102
  /**
1103
+ * @public
1002
1104
  * These settings relate to your MP4 output container. You can create audio only outputs with this container. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/supported-codecs-containers-audio-only.html#output-codecs-and-containers-supported-for-audio-only.
1003
1105
  */
1004
1106
  Mp4Settings?: Mp4Settings;
1005
1107
  /**
1108
+ * @public
1006
1109
  * These settings relate to the fragmented MP4 container for the segments in your DASH outputs.
1007
1110
  */
1008
1111
  MpdSettings?: MpdSettings;
1009
1112
  /**
1113
+ * @public
1010
1114
  * These settings relate to your MXF output container.
1011
1115
  */
1012
1116
  MxfSettings?: MxfSettings;
@@ -1067,30 +1171,37 @@ export type HlsIFrameOnlyManifest = (typeof HlsIFrameOnlyManifest)[keyof typeof
1067
1171
  */
1068
1172
  export interface HlsSettings {
1069
1173
  /**
1174
+ * @public
1070
1175
  * Specifies the group to which the audio rendition belongs.
1071
1176
  */
1072
1177
  AudioGroupId?: string;
1073
1178
  /**
1179
+ * @public
1074
1180
  * Use this setting only in audio-only outputs. Choose MPEG-2 Transport Stream (M2TS) to create a file in an MPEG2-TS container. Keep the default value Automatic to create an audio-only file in a raw container. Regardless of the value that you specify here, if this output has video, the service will place the output into an MPEG2-TS container.
1075
1181
  */
1076
1182
  AudioOnlyContainer?: HlsAudioOnlyContainer | string;
1077
1183
  /**
1184
+ * @public
1078
1185
  * List all the audio groups that are used with the video output stream. Input all the audio GROUP-IDs that are associated to the video, separate by ','.
1079
1186
  */
1080
1187
  AudioRenditionSets?: string;
1081
1188
  /**
1189
+ * @public
1082
1190
  * Four types of audio-only tracks are supported: Audio-Only Variant Stream The client can play back this audio-only stream instead of video in low-bandwidth scenarios. Represented as an EXT-X-STREAM-INF in the HLS manifest. Alternate Audio, Auto Select, Default Alternate rendition that the client should try to play back by default. Represented as an EXT-X-MEDIA in the HLS manifest with DEFAULT=YES, AUTOSELECT=YES Alternate Audio, Auto Select, Not Default Alternate rendition that the client may try to play back by default. Represented as an EXT-X-MEDIA in the HLS manifest with DEFAULT=NO, AUTOSELECT=YES Alternate Audio, not Auto Select Alternate rendition that the client will not try to play back by default. Represented as an EXT-X-MEDIA in the HLS manifest with DEFAULT=NO, AUTOSELECT=NO
1083
1191
  */
1084
1192
  AudioTrackType?: HlsAudioTrackType | string;
1085
1193
  /**
1194
+ * @public
1086
1195
  * Specify whether to flag this audio track as descriptive video service (DVS) in your HLS parent manifest. When you choose Flag, MediaConvert includes the parameter CHARACTERISTICS="public.accessibility.describes-video" in the EXT-X-MEDIA entry for this track. When you keep the default choice, Don't flag, MediaConvert leaves this parameter out. The DVS flag can help with accessibility on Apple devices. For more information, see the Apple documentation.
1087
1196
  */
1088
1197
  DescriptiveVideoServiceFlag?: HlsDescriptiveVideoServiceFlag | string;
1089
1198
  /**
1199
+ * @public
1090
1200
  * Choose Include to have MediaConvert generate a child manifest that lists only the I-frames for this rendition, in addition to your regular manifest for this rendition. You might use this manifest as part of a workflow that creates preview functions for your video. MediaConvert adds both the I-frame only child manifest and the regular child manifest to the parent manifest. When you don't need the I-frame only child manifest, keep the default value Exclude.
1091
1201
  */
1092
1202
  IFrameOnlyManifest?: HlsIFrameOnlyManifest | string;
1093
1203
  /**
1204
+ * @public
1094
1205
  * Use this setting to add an identifying string to the filename of each segment. The service adds this string between the name modifier and segment index number. You can use format identifiers in the string. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/using-variables-in-your-job-settings.html
1095
1206
  */
1096
1207
  SegmentModifier?: string;
@@ -1101,6 +1212,7 @@ export interface HlsSettings {
1101
1212
  */
1102
1213
  export interface OutputSettings {
1103
1214
  /**
1215
+ * @public
1104
1216
  * Settings for HLS output groups
1105
1217
  */
1106
1218
  HlsSettings?: HlsSettings;
@@ -1189,10 +1301,12 @@ export type Av1FramerateConversionAlgorithm = (typeof Av1FramerateConversionAlgo
1189
1301
  */
1190
1302
  export interface Av1QvbrSettings {
1191
1303
  /**
1304
+ * @public
1192
1305
  * Use this setting only when you set Rate control mode to QVBR. Specify the target quality level for this output. MediaConvert determines the right number of bits to use for each part of the video to maintain the video quality that you specify. When you keep the default value, AUTO, MediaConvert picks a quality level for you, based on characteristics of your input video. If you prefer to specify a quality level, specify a number from 1 through 10. Use higher numbers for greater quality. Level 10 results in nearly lossless compression. The quality level for most broadcast-quality transcodes is between 6 and 9. Optionally, to specify a value between whole numbers, also provide a value for the setting qvbrQualityLevelFineTune. For example, if you want your QVBR quality level to be 7.33, set qvbrQualityLevel to 7 and set qvbrQualityLevelFineTune to .33.
1193
1306
  */
1194
1307
  QvbrQualityLevel?: number;
1195
1308
  /**
1309
+ * @public
1196
1310
  * Optional. Specify a value here to set the QVBR quality to a level that is between whole numbers. For example, if you want your QVBR quality level to be 7.33, set qvbrQualityLevel to 7 and set qvbrQualityLevelFineTune to .33. MediaConvert rounds your QVBR quality level to the nearest third of a whole number. For example, if you set qvbrQualityLevel to 7 and you set qvbrQualityLevelFineTune to .25, your actual QVBR quality level is 7.33.
1197
1311
  */
1198
1312
  QvbrQualityLevelFineTune?: number;
@@ -1226,54 +1340,67 @@ export type Av1SpatialAdaptiveQuantization = (typeof Av1SpatialAdaptiveQuantizat
1226
1340
  */
1227
1341
  export interface Av1Settings {
1228
1342
  /**
1343
+ * @public
1229
1344
  * Specify the strength of any adaptive quantization filters that you enable. The value that you choose here applies to Spatial adaptive quantization.
1230
1345
  */
1231
1346
  AdaptiveQuantization?: Av1AdaptiveQuantization | string;
1232
1347
  /**
1348
+ * @public
1233
1349
  * Specify the Bit depth. You can choose 8-bit or 10-bit.
1234
1350
  */
1235
1351
  BitDepth?: Av1BitDepth | string;
1236
1352
  /**
1353
+ * @public
1237
1354
  * Use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.
1238
1355
  */
1239
1356
  FramerateControl?: Av1FramerateControl | string;
1240
1357
  /**
1358
+ * @public
1241
1359
  * Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate. For numerically simple conversions, such as 60 fps to 30 fps: We recommend that you keep the default value, Drop duplicate. For numerically complex conversions, to avoid stutter: Choose Interpolate. This results in a smooth picture, but might introduce undesirable video artifacts. For complex frame rate conversions, especially if your source video has already been converted from its original cadence: Choose FrameFormer to do motion-compensated interpolation. FrameFormer uses the best conversion method frame by frame. Note that using FrameFormer increases the transcoding time and incurs a significant add-on cost. When you choose FrameFormer, your input video resolution must be at least 128x96.
1242
1360
  */
1243
1361
  FramerateConversionAlgorithm?: Av1FramerateConversionAlgorithm | string;
1244
1362
  /**
1363
+ * @public
1245
1364
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateDenominator to specify the denominator of this fraction. In this example, use 1001 for the value of FramerateDenominator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
1246
1365
  */
1247
1366
  FramerateDenominator?: number;
1248
1367
  /**
1368
+ * @public
1249
1369
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateNumerator to specify the numerator of this fraction. In this example, use 24000 for the value of FramerateNumerator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
1250
1370
  */
1251
1371
  FramerateNumerator?: number;
1252
1372
  /**
1373
+ * @public
1253
1374
  * Specify the GOP length (keyframe interval) in frames. With AV1, MediaConvert doesn't support GOP length in seconds. This value must be greater than zero and preferably equal to 1 + ((numberBFrames + 1) * x), where x is an integer value.
1254
1375
  */
1255
1376
  GopSize?: number;
1256
1377
  /**
1378
+ * @public
1257
1379
  * Maximum bitrate in bits/second. For example, enter five megabits per second as 5000000. Required when Rate control mode is QVBR.
1258
1380
  */
1259
1381
  MaxBitrate?: number;
1260
1382
  /**
1383
+ * @public
1261
1384
  * Specify from the number of B-frames, in the range of 0-15. For AV1 encoding, we recommend using 7 or 15. Choose a larger number for a lower bitrate and smaller file size; choose a smaller number for better video quality.
1262
1385
  */
1263
1386
  NumberBFramesBetweenReferenceFrames?: number;
1264
1387
  /**
1388
+ * @public
1265
1389
  * Settings for quality-defined variable bitrate encoding with the H.265 codec. Use these settings only when you set QVBR for Rate control mode.
1266
1390
  */
1267
1391
  QvbrSettings?: Av1QvbrSettings;
1268
1392
  /**
1393
+ * @public
1269
1394
  * 'With AV1 outputs, for rate control mode, MediaConvert supports only quality-defined variable bitrate (QVBR). You can''t use CBR or VBR.'
1270
1395
  */
1271
1396
  RateControlMode?: Av1RateControlMode | string;
1272
1397
  /**
1398
+ * @public
1273
1399
  * Specify the number of slices per picture. This value must be 1, 2, 4, 8, 16, or 32. For progressive pictures, this value must be less than or equal to the number of macroblock rows. For interlaced pictures, this value must be less than or equal to half the number of macroblock rows.
1274
1400
  */
1275
1401
  Slices?: number;
1276
1402
  /**
1403
+ * @public
1277
1404
  * Keep the default value, Enabled, to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher.
1278
1405
  */
1279
1406
  SpatialAdaptiveQuantization?: Av1SpatialAdaptiveQuantization | string;
@@ -1310,6 +1437,7 @@ export type AvcIntraUhdQualityTuningLevel = (typeof AvcIntraUhdQualityTuningLeve
1310
1437
  */
1311
1438
  export interface AvcIntraUhdSettings {
1312
1439
  /**
1440
+ * @public
1313
1441
  * Optional. Use Quality tuning level to choose how many transcoding passes MediaConvert does with your video. When you choose Multi-pass, your video quality is better and your output bitrate is more accurate. That is, the actual bitrate of your output is closer to the target bitrate defined in the specification. When you choose Single-pass, your encoding time is faster. The default behavior is Single-pass.
1314
1442
  */
1315
1443
  QualityTuningLevel?: AvcIntraUhdQualityTuningLevel | string;
@@ -1396,42 +1524,52 @@ export type AvcIntraTelecine = (typeof AvcIntraTelecine)[keyof typeof AvcIntraTe
1396
1524
  */
1397
1525
  export interface AvcIntraSettings {
1398
1526
  /**
1527
+ * @public
1399
1528
  * Specify the AVC-Intra class of your output. The AVC-Intra class selection determines the output video bit rate depending on the frame rate of the output. Outputs with higher class values have higher bitrates and improved image quality. Note that for Class 4K/2K, MediaConvert supports only 4:2:2 chroma subsampling.
1400
1529
  */
1401
1530
  AvcIntraClass?: AvcIntraClass | string;
1402
1531
  /**
1532
+ * @public
1403
1533
  * Optional when you set AVC-Intra class to Class 4K/2K. When you set AVC-Intra class to a different value, this object isn't allowed.
1404
1534
  */
1405
1535
  AvcIntraUhdSettings?: AvcIntraUhdSettings;
1406
1536
  /**
1537
+ * @public
1407
1538
  * If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.
1408
1539
  */
1409
1540
  FramerateControl?: AvcIntraFramerateControl | string;
1410
1541
  /**
1542
+ * @public
1411
1543
  * Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate. For numerically simple conversions, such as 60 fps to 30 fps: We recommend that you keep the default value, Drop duplicate. For numerically complex conversions, to avoid stutter: Choose Interpolate. This results in a smooth picture, but might introduce undesirable video artifacts. For complex frame rate conversions, especially if your source video has already been converted from its original cadence: Choose FrameFormer to do motion-compensated interpolation. FrameFormer uses the best conversion method frame by frame. Note that using FrameFormer increases the transcoding time and incurs a significant add-on cost. When you choose FrameFormer, your input video resolution must be at least 128x96.
1412
1544
  */
1413
1545
  FramerateConversionAlgorithm?: AvcIntraFramerateConversionAlgorithm | string;
1414
1546
  /**
1547
+ * @public
1415
1548
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateDenominator to specify the denominator of this fraction. In this example, use 1001 for the value of FramerateDenominator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
1416
1549
  */
1417
1550
  FramerateDenominator?: number;
1418
1551
  /**
1552
+ * @public
1419
1553
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateNumerator to specify the numerator of this fraction. In this example, use 24000 for the value of FramerateNumerator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
1420
1554
  */
1421
1555
  FramerateNumerator?: number;
1422
1556
  /**
1557
+ * @public
1423
1558
  * Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.
1424
1559
  */
1425
1560
  InterlaceMode?: AvcIntraInterlaceMode | string;
1426
1561
  /**
1562
+ * @public
1427
1563
  * Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive.
1428
1564
  */
1429
1565
  ScanTypeConversionMode?: AvcIntraScanTypeConversionMode | string;
1430
1566
  /**
1567
+ * @public
1431
1568
  * Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25.
1432
1569
  */
1433
1570
  SlowPal?: AvcIntraSlowPal | string;
1434
1571
  /**
1572
+ * @public
1435
1573
  * When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard telecine to create a smoother picture. When you keep the default value, None, MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture.
1436
1574
  */
1437
1575
  Telecine?: AvcIntraTelecine | string;
@@ -1464,18 +1602,22 @@ export type VideoCodec = (typeof VideoCodec)[keyof typeof VideoCodec];
1464
1602
  */
1465
1603
  export interface FrameCaptureSettings {
1466
1604
  /**
1605
+ * @public
1467
1606
  * Frame capture will encode the first frame of the output stream, then one frame every framerateDenominator/framerateNumerator seconds. For example, settings of framerateNumerator = 1 and framerateDenominator = 3 (a rate of 1/3 frame per second) will capture the first frame, then 1 frame every 3s. Files will be named as filename.n.jpg where n is the 0-based sequence number of each Capture.
1468
1607
  */
1469
1608
  FramerateDenominator?: number;
1470
1609
  /**
1610
+ * @public
1471
1611
  * Frame capture will encode the first frame of the output stream, then one frame every framerateDenominator/framerateNumerator seconds. For example, settings of framerateNumerator = 1 and framerateDenominator = 3 (a rate of 1/3 frame per second) will capture the first frame, then 1 frame every 3s. Files will be named as filename.NNNNNNN.jpg where N is the 0-based frame sequence number zero padded to 7 decimal places.
1472
1612
  */
1473
1613
  FramerateNumerator?: number;
1474
1614
  /**
1615
+ * @public
1475
1616
  * Maximum number of captures (encoded jpg output files).
1476
1617
  */
1477
1618
  MaxCaptures?: number;
1478
1619
  /**
1620
+ * @public
1479
1621
  * JPEG Quality - a higher value equals higher quality.
1480
1622
  */
1481
1623
  Quality?: number;
@@ -1532,10 +1674,12 @@ export type BandwidthReductionFilterStrength = (typeof BandwidthReductionFilterS
1532
1674
  */
1533
1675
  export interface BandwidthReductionFilter {
1534
1676
  /**
1677
+ * @public
1535
1678
  * Optionally specify the level of sharpening to apply when you use the Bandwidth reduction filter. Sharpening adds contrast to the edges of your video content and can reduce softness. Keep the default value Off to apply no sharpening. Set Sharpening strength to Low to apply a minimal amount of sharpening, or High to apply a maximum amount of sharpening.
1536
1679
  */
1537
1680
  Sharpening?: BandwidthReductionFilterSharpening | string;
1538
1681
  /**
1682
+ * @public
1539
1683
  * Specify the strength of the Bandwidth reduction filter. For most workflows, we recommend that you choose Auto to reduce the bandwidth of your output with little to no perceptual decrease in video quality. For high quality and high bitrate outputs, choose Low. For the most bandwidth reduction, choose High. We recommend that you choose High for low bitrate outputs. Note that High may incur a slight increase in the softness of your output.
1540
1684
  */
1541
1685
  Strength?: BandwidthReductionFilterStrength | string;
@@ -1728,14 +1872,17 @@ export type H264QualityTuningLevel = (typeof H264QualityTuningLevel)[keyof typeo
1728
1872
  */
1729
1873
  export interface H264QvbrSettings {
1730
1874
  /**
1875
+ * @public
1731
1876
  * Use this setting only when Rate control mode is QVBR and Quality tuning level is Multi-pass HQ. For Max average bitrate values suited to the complexity of your input video, the service limits the average bitrate of the video part of this output to the value that you choose. That is, the total size of the video element is less than or equal to the value you set multiplied by the number of seconds of encoded output.
1732
1877
  */
1733
1878
  MaxAverageBitrate?: number;
1734
1879
  /**
1880
+ * @public
1735
1881
  * Use this setting only when you set Rate control mode to QVBR. Specify the target quality level for this output. MediaConvert determines the right number of bits to use for each part of the video to maintain the video quality that you specify. When you keep the default value, AUTO, MediaConvert picks a quality level for you, based on characteristics of your input video. If you prefer to specify a quality level, specify a number from 1 through 10. Use higher numbers for greater quality. Level 10 results in nearly lossless compression. The quality level for most broadcast-quality transcodes is between 6 and 9. Optionally, to specify a value between whole numbers, also provide a value for the setting qvbrQualityLevelFineTune. For example, if you want your QVBR quality level to be 7.33, set qvbrQualityLevel to 7 and set qvbrQualityLevelFineTune to .33.
1736
1882
  */
1737
1883
  QvbrQualityLevel?: number;
1738
1884
  /**
1885
+ * @public
1739
1886
  * Optional. Specify a value here to set the QVBR quality to a level that is between whole numbers. For example, if you want your QVBR quality level to be 7.33, set qvbrQualityLevel to 7 and set qvbrQualityLevelFineTune to .33. MediaConvert rounds your QVBR quality level to the nearest third of a whole number. For example, if you set qvbrQualityLevel to 7 and you set qvbrQualityLevelFineTune to .25, your actual QVBR quality level is 7.33.
1740
1887
  */
1741
1888
  QvbrQualityLevelFineTune?: number;
@@ -1869,170 +2016,212 @@ export type H264UnregisteredSeiTimecode = (typeof H264UnregisteredSeiTimecode)[k
1869
2016
  */
1870
2017
  export interface H264Settings {
1871
2018
  /**
2019
+ * @public
1872
2020
  * Keep the default value, Auto, for this setting to have MediaConvert automatically apply the best types of quantization for your video content. When you want to apply your quantization settings manually, you must set H264AdaptiveQuantization to a value other than Auto. Use this setting to specify the strength of any adaptive quantization filters that you enable. If you don't want MediaConvert to do any adaptive quantization in this transcode, set Adaptive quantization to Off. Related settings: The value that you choose here applies to the following settings: H264FlickerAdaptiveQuantization, H264SpatialAdaptiveQuantization, and H264TemporalAdaptiveQuantization.
1873
2021
  */
1874
2022
  AdaptiveQuantization?: H264AdaptiveQuantization | string;
1875
2023
  /**
2024
+ * @public
1876
2025
  * The Bandwidth reduction filter increases the video quality of your output relative to its bitrate. Use to lower the bitrate of your constant quality QVBR output, with little or no perceptual decrease in quality. Or, use to increase the video quality of outputs with other rate control modes relative to the bitrate that you specify. Bandwidth reduction increases further when your input is low quality or noisy. Outputs that use this feature incur pro-tier pricing. When you include Bandwidth reduction filter, you cannot include the Noise reducer preprocessor.
1877
2026
  */
1878
2027
  BandwidthReductionFilter?: BandwidthReductionFilter;
1879
2028
  /**
2029
+ * @public
1880
2030
  * Specify the average bitrate in bits per second. Required for VBR and CBR. For MS Smooth outputs, bitrates must be unique when rounded down to the nearest multiple of 1000.
1881
2031
  */
1882
2032
  Bitrate?: number;
1883
2033
  /**
2034
+ * @public
1884
2035
  * Specify an H.264 level that is consistent with your output video settings. If you aren't sure what level to specify, choose Auto.
1885
2036
  */
1886
2037
  CodecLevel?: H264CodecLevel | string;
1887
2038
  /**
2039
+ * @public
1888
2040
  * H.264 Profile. High 4:2:2 and 10-bit profiles are only available with the AVC-I License.
1889
2041
  */
1890
2042
  CodecProfile?: H264CodecProfile | string;
1891
2043
  /**
2044
+ * @public
1892
2045
  * Specify whether to allow the number of B-frames in your output GOP structure to vary or not depending on your input video content. To improve the subjective video quality of your output that has high-motion content: Leave blank or keep the default value Adaptive. MediaConvert will use fewer B-frames for high-motion video content than low-motion content. The maximum number of B- frames is limited by the value that you choose for B-frames between reference frames. To use the same number B-frames for all types of content: Choose Static.
1893
2046
  */
1894
2047
  DynamicSubGop?: H264DynamicSubGop | string;
1895
2048
  /**
2049
+ * @public
1896
2050
  * Entropy encoding mode. Use CABAC (must be in Main or High profile) or CAVLC.
1897
2051
  */
1898
2052
  EntropyEncoding?: H264EntropyEncoding | string;
1899
2053
  /**
2054
+ * @public
1900
2055
  * The video encoding method for your MPEG-4 AVC output. Keep the default value, PAFF, to have MediaConvert use PAFF encoding for interlaced outputs. Choose Force field to disable PAFF encoding and create separate interlaced fields. Choose MBAFF to disable PAFF and have MediaConvert use MBAFF encoding for interlaced outputs.
1901
2056
  */
1902
2057
  FieldEncoding?: H264FieldEncoding | string;
1903
2058
  /**
2059
+ * @public
1904
2060
  * Only use this setting when you change the default value, AUTO, for the setting H264AdaptiveQuantization. When you keep all defaults, excluding H264AdaptiveQuantization and all other adaptive quantization from your JSON job specification, MediaConvert automatically applies the best types of quantization for your video content. When you set H264AdaptiveQuantization to a value other than AUTO, the default value for H264FlickerAdaptiveQuantization is Disabled. Change this value to Enabled to reduce I-frame pop. I-frame pop appears as a visual flicker that can arise when the encoder saves bits by copying some macroblocks many times from frame to frame, and then refreshes them at the I-frame. When you enable this setting, the encoder updates these macroblocks slightly more often to smooth out the flicker. To manually enable or disable H264FlickerAdaptiveQuantization, you must set Adaptive quantization to a value other than AUTO.
1905
2061
  */
1906
2062
  FlickerAdaptiveQuantization?: H264FlickerAdaptiveQuantization | string;
1907
2063
  /**
2064
+ * @public
1908
2065
  * If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.
1909
2066
  */
1910
2067
  FramerateControl?: H264FramerateControl | string;
1911
2068
  /**
2069
+ * @public
1912
2070
  * Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate. For numerically simple conversions, such as 60 fps to 30 fps: We recommend that you keep the default value, Drop duplicate. For numerically complex conversions, to avoid stutter: Choose Interpolate. This results in a smooth picture, but might introduce undesirable video artifacts. For complex frame rate conversions, especially if your source video has already been converted from its original cadence: Choose FrameFormer to do motion-compensated interpolation. FrameFormer uses the best conversion method frame by frame. Note that using FrameFormer increases the transcoding time and incurs a significant add-on cost. When you choose FrameFormer, your input video resolution must be at least 128x96.
1913
2071
  */
1914
2072
  FramerateConversionAlgorithm?: H264FramerateConversionAlgorithm | string;
1915
2073
  /**
2074
+ * @public
1916
2075
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateDenominator to specify the denominator of this fraction. In this example, use 1001 for the value of FramerateDenominator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
1917
2076
  */
1918
2077
  FramerateDenominator?: number;
1919
2078
  /**
2079
+ * @public
1920
2080
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateNumerator to specify the numerator of this fraction. In this example, use 24000 for the value of FramerateNumerator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
1921
2081
  */
1922
2082
  FramerateNumerator?: number;
1923
2083
  /**
2084
+ * @public
1924
2085
  * Specify whether to allow B-frames to be referenced by other frame types. To use reference B-frames when your GOP structure has 1 or more B-frames: Leave blank or keep the default value Enabled. We recommend that you choose Enabled to help improve the video quality of your output relative to its bitrate. To not use reference B-frames: Choose Disabled.
1925
2086
  */
1926
2087
  GopBReference?: H264GopBReference | string;
1927
2088
  /**
2089
+ * @public
1928
2090
  * Specify the relative frequency of open to closed GOPs in this output. For example, if you want to allow four open GOPs and then require a closed GOP, set this value to 5. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. In the console, do this by keeping the default empty value. If you do explicitly specify a value, for segmented outputs, don't set this value to 0.
1929
2091
  */
1930
2092
  GopClosedCadence?: number;
1931
2093
  /**
2094
+ * @public
1932
2095
  * Use this setting only when you set GOP mode control to Specified, frames or Specified, seconds. Specify the GOP length using a whole number of frames or a decimal value of seconds. MediaConvert will interpret this value as frames or seconds depending on the value you choose for GOP mode control. If you want to allow MediaConvert to automatically determine GOP size, leave GOP size blank and set GOP mode control to Auto. If your output group specifies HLS, DASH, or CMAF, leave GOP size blank and set GOP mode control to Auto in each output in your output group.
1933
2096
  */
1934
2097
  GopSize?: number;
1935
2098
  /**
2099
+ * @public
1936
2100
  * Specify how the transcoder determines GOP size for this output. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, choose Auto and and leave GOP size blank. By default, if you don't specify GOP mode control, MediaConvert will use automatic behavior. If your output group specifies HLS, DASH, or CMAF, set GOP mode control to Auto and leave GOP size blank in each output in your output group. To explicitly specify the GOP length, choose Specified, frames or Specified, seconds and then provide the GOP length in the related setting GOP size.
1937
2101
  */
1938
2102
  GopSizeUnits?: H264GopSizeUnits | string;
1939
2103
  /**
2104
+ * @public
1940
2105
  * If your downstream systems have strict buffer requirements: Specify the minimum percentage of the HRD buffer that's available at the end of each encoded video segment. For the best video quality: Set to 0 or leave blank to automatically determine the final buffer fill percentage.
1941
2106
  */
1942
2107
  HrdBufferFinalFillPercentage?: number;
1943
2108
  /**
2109
+ * @public
1944
2110
  * Percentage of the buffer that should initially be filled (HRD buffer model).
1945
2111
  */
1946
2112
  HrdBufferInitialFillPercentage?: number;
1947
2113
  /**
2114
+ * @public
1948
2115
  * Size of buffer (HRD buffer model) in bits. For example, enter five megabits as 5000000.
1949
2116
  */
1950
2117
  HrdBufferSize?: number;
1951
2118
  /**
2119
+ * @public
1952
2120
  * Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.
1953
2121
  */
1954
2122
  InterlaceMode?: H264InterlaceMode | string;
1955
2123
  /**
2124
+ * @public
1956
2125
  * Maximum bitrate in bits/second. For example, enter five megabits per second as 5000000. Required when Rate control mode is QVBR.
1957
2126
  */
1958
2127
  MaxBitrate?: number;
1959
2128
  /**
2129
+ * @public
1960
2130
  * Use this setting only when you also enable Scene change detection. This setting determines how the encoder manages the spacing between I-frames that it inserts as part of the I-frame cadence and the I-frames that it inserts for Scene change detection. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, do this by keeping the default empty value. When you explicitly specify a value for this setting, the encoder determines whether to skip a cadence-driven I-frame by the value you set. For example, if you set Min I interval to 5 and a cadence-driven I-frame would fall within 5 frames of a scene-change I-frame, then the encoder skips the cadence-driven I-frame. In this way, one GOP is shrunk slightly and one GOP is stretched slightly. When the cadence-driven I-frames are farther from the scene-change I-frame than the value you set, then the encoder leaves all I-frames in place and the GOPs surrounding the scene change are smaller than the usual cadence GOPs.
1961
2131
  */
1962
2132
  MinIInterval?: number;
1963
2133
  /**
2134
+ * @public
1964
2135
  * Specify the number of B-frames between reference frames in this output. For the best video quality: Leave blank. MediaConvert automatically determines the number of B-frames to use based on the characteristics of your input video. To manually specify the number of B-frames between reference frames: Enter an integer from 0 to 7.
1965
2136
  */
1966
2137
  NumberBFramesBetweenReferenceFrames?: number;
1967
2138
  /**
2139
+ * @public
1968
2140
  * Number of reference frames to use. The encoder may use more than requested if using B-frames and/or interlaced encoding.
1969
2141
  */
1970
2142
  NumberReferenceFrames?: number;
1971
2143
  /**
2144
+ * @public
1972
2145
  * Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source, uses the PAR from your input video for your output. To specify a different PAR in the console, choose any value other than Follow source. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings.
1973
2146
  */
1974
2147
  ParControl?: H264ParControl | string;
1975
2148
  /**
2149
+ * @public
1976
2150
  * Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parDenominator is 33.
1977
2151
  */
1978
2152
  ParDenominator?: number;
1979
2153
  /**
2154
+ * @public
1980
2155
  * Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parNumerator is 40.
1981
2156
  */
1982
2157
  ParNumerator?: number;
1983
2158
  /**
2159
+ * @public
1984
2160
  * The Quality tuning level you choose represents a trade-off between the encoding speed of your job and the output video quality. For the fastest encoding speed at the cost of video quality: Choose Single pass. For a good balance between encoding speed and video quality: Leave blank or keep the default value Single pass HQ. For the best video quality, at the cost of encoding speed: Choose Multi pass HQ. MediaConvert performs an analysis pass on your input followed by an encoding pass. Outputs that use this feature incur pro-tier pricing.
1985
2161
  */
1986
2162
  QualityTuningLevel?: H264QualityTuningLevel | string;
1987
2163
  /**
2164
+ * @public
1988
2165
  * Settings for quality-defined variable bitrate encoding with the H.265 codec. Use these settings only when you set QVBR for Rate control mode.
1989
2166
  */
1990
2167
  QvbrSettings?: H264QvbrSettings;
1991
2168
  /**
2169
+ * @public
1992
2170
  * Use this setting to specify whether this output has a variable bitrate (VBR), constant bitrate (CBR) or quality-defined variable bitrate (QVBR).
1993
2171
  */
1994
2172
  RateControlMode?: H264RateControlMode | string;
1995
2173
  /**
2174
+ * @public
1996
2175
  * Places a PPS header on each encoded picture, even if repeated.
1997
2176
  */
1998
2177
  RepeatPps?: H264RepeatPps | string;
1999
2178
  /**
2179
+ * @public
2000
2180
  * Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive.
2001
2181
  */
2002
2182
  ScanTypeConversionMode?: H264ScanTypeConversionMode | string;
2003
2183
  /**
2184
+ * @public
2004
2185
  * Enable this setting to insert I-frames at scene changes that the service automatically detects. This improves video quality and is enabled by default. If this output uses QVBR, choose Transition detection for further video quality improvement. For more information about QVBR, see https://docs.aws.amazon.com/console/mediaconvert/cbr-vbr-qvbr.
2005
2186
  */
2006
2187
  SceneChangeDetect?: H264SceneChangeDetect | string;
2007
2188
  /**
2189
+ * @public
2008
2190
  * Number of slices per picture. Must be less than or equal to the number of macroblock rows for progressive pictures, and less than or equal to half the number of macroblock rows for interlaced pictures.
2009
2191
  */
2010
2192
  Slices?: number;
2011
2193
  /**
2194
+ * @public
2012
2195
  * Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25.
2013
2196
  */
2014
2197
  SlowPal?: H264SlowPal | string;
2015
2198
  /**
2199
+ * @public
2016
2200
  * Ignore this setting unless you need to comply with a specification that requires a specific value. If you don't have a specification requirement, we recommend that you adjust the softness of your output by using a lower value for the setting Sharpness or by enabling a noise reducer filter. The Softness setting specifies the quantization matrices that the encoder uses. Keep the default value, 0, for flat quantization. Choose the value 1 or 16 to use the default JVT softening quantization matricies from the H.264 specification. Choose a value from 17 to 128 to use planar interpolation. Increasing values from 17 to 128 result in increasing reduction of high-frequency data. The value 128 results in the softest video.
2017
2201
  */
2018
2202
  Softness?: number;
2019
2203
  /**
2204
+ * @public
2020
2205
  * Only use this setting when you change the default value, Auto, for the setting H264AdaptiveQuantization. When you keep all defaults, excluding H264AdaptiveQuantization and all other adaptive quantization from your JSON job specification, MediaConvert automatically applies the best types of quantization for your video content. When you set H264AdaptiveQuantization to a value other than AUTO, the default value for H264SpatialAdaptiveQuantization is Enabled. Keep this default value to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to set H264SpatialAdaptiveQuantization to Disabled. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher. To manually enable or disable H264SpatialAdaptiveQuantization, you must set Adaptive quantization to a value other than AUTO.
2021
2206
  */
2022
2207
  SpatialAdaptiveQuantization?: H264SpatialAdaptiveQuantization | string;
2023
2208
  /**
2209
+ * @public
2024
2210
  * Produces a bitstream compliant with SMPTE RP-2027.
2025
2211
  */
2026
2212
  Syntax?: H264Syntax | string;
2027
2213
  /**
2214
+ * @public
2028
2215
  * When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard or soft telecine to create a smoother picture. Hard telecine produces a 29.97i output. Soft telecine produces an output with a 23.976 output that signals to the video player device to do the conversion during play back. When you keep the default value, None, MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture.
2029
2216
  */
2030
2217
  Telecine?: H264Telecine | string;
2031
2218
  /**
2219
+ * @public
2032
2220
  * Only use this setting when you change the default value, AUTO, for the setting H264AdaptiveQuantization. When you keep all defaults, excluding H264AdaptiveQuantization and all other adaptive quantization from your JSON job specification, MediaConvert automatically applies the best types of quantization for your video content. When you set H264AdaptiveQuantization to a value other than AUTO, the default value for H264TemporalAdaptiveQuantization is Enabled. Keep this default value to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to set H264TemporalAdaptiveQuantization to Disabled. Related setting: When you enable temporal quantization, adjust the strength of the filter with the setting Adaptive quantization. To manually enable or disable H264TemporalAdaptiveQuantization, you must set Adaptive quantization to a value other than AUTO.
2033
2221
  */
2034
2222
  TemporalAdaptiveQuantization?: H264TemporalAdaptiveQuantization | string;
2035
2223
  /**
2224
+ * @public
2036
2225
  * Inserts timecode for each frame as 4 bytes of an unregistered SEI message.
2037
2226
  */
2038
2227
  UnregisteredSeiTimecode?: H264UnregisteredSeiTimecode | string;
@@ -2228,14 +2417,17 @@ export type H265QualityTuningLevel = (typeof H265QualityTuningLevel)[keyof typeo
2228
2417
  */
2229
2418
  export interface H265QvbrSettings {
2230
2419
  /**
2420
+ * @public
2231
2421
  * Use this setting only when Rate control mode is QVBR and Quality tuning level is Multi-pass HQ. For Max average bitrate values suited to the complexity of your input video, the service limits the average bitrate of the video part of this output to the value that you choose. That is, the total size of the video element is less than or equal to the value you set multiplied by the number of seconds of encoded output.
2232
2422
  */
2233
2423
  MaxAverageBitrate?: number;
2234
2424
  /**
2425
+ * @public
2235
2426
  * Use this setting only when you set Rate control mode to QVBR. Specify the target quality level for this output. MediaConvert determines the right number of bits to use for each part of the video to maintain the video quality that you specify. When you keep the default value, AUTO, MediaConvert picks a quality level for you, based on characteristics of your input video. If you prefer to specify a quality level, specify a number from 1 through 10. Use higher numbers for greater quality. Level 10 results in nearly lossless compression. The quality level for most broadcast-quality transcodes is between 6 and 9. Optionally, to specify a value between whole numbers, also provide a value for the setting qvbrQualityLevelFineTune. For example, if you want your QVBR quality level to be 7.33, set qvbrQualityLevel to 7 and set qvbrQualityLevelFineTune to .33.
2236
2427
  */
2237
2428
  QvbrQualityLevel?: number;
2238
2429
  /**
2430
+ * @public
2239
2431
  * Optional. Specify a value here to set the QVBR quality to a level that is between whole numbers. For example, if you want your QVBR quality level to be 7.33, set qvbrQualityLevel to 7 and set qvbrQualityLevelFineTune to .33. MediaConvert rounds your QVBR quality level to the nearest third of a whole number. For example, if you set qvbrQualityLevel to 7 and you set qvbrQualityLevelFineTune to .25, your actual QVBR quality level is 7.33.
2240
2432
  */
2241
2433
  QvbrQualityLevelFineTune?: number;
@@ -2394,170 +2586,212 @@ export type H265WriteMp4PackagingType = (typeof H265WriteMp4PackagingType)[keyof
2394
2586
  */
2395
2587
  export interface H265Settings {
2396
2588
  /**
2589
+ * @public
2397
2590
  * When you set Adaptive Quantization to Auto, or leave blank, MediaConvert automatically applies quantization to improve the video quality of your output. Set Adaptive Quantization to Low, Medium, High, Higher, or Max to manually control the strength of the quantization filter. When you do, you can specify a value for Spatial Adaptive Quantization, Temporal Adaptive Quantization, and Flicker Adaptive Quantization, to further control the quantization filter. Set Adaptive Quantization to Off to apply no quantization to your output.
2398
2591
  */
2399
2592
  AdaptiveQuantization?: H265AdaptiveQuantization | string;
2400
2593
  /**
2594
+ * @public
2401
2595
  * Enables Alternate Transfer Function SEI message for outputs using Hybrid Log Gamma (HLG) Electro-Optical Transfer Function (EOTF).
2402
2596
  */
2403
2597
  AlternateTransferFunctionSei?: H265AlternateTransferFunctionSei | string;
2404
2598
  /**
2599
+ * @public
2405
2600
  * The Bandwidth reduction filter increases the video quality of your output relative to its bitrate. Use to lower the bitrate of your constant quality QVBR output, with little or no perceptual decrease in quality. Or, use to increase the video quality of outputs with other rate control modes relative to the bitrate that you specify. Bandwidth reduction increases further when your input is low quality or noisy. Outputs that use this feature incur pro-tier pricing. When you include Bandwidth reduction filter, you cannot include the Noise reducer preprocessor.
2406
2601
  */
2407
2602
  BandwidthReductionFilter?: BandwidthReductionFilter;
2408
2603
  /**
2604
+ * @public
2409
2605
  * Specify the average bitrate in bits per second. Required for VBR and CBR. For MS Smooth outputs, bitrates must be unique when rounded down to the nearest multiple of 1000.
2410
2606
  */
2411
2607
  Bitrate?: number;
2412
2608
  /**
2609
+ * @public
2413
2610
  * H.265 Level.
2414
2611
  */
2415
2612
  CodecLevel?: H265CodecLevel | string;
2416
2613
  /**
2614
+ * @public
2417
2615
  * Represents the Profile and Tier, per the HEVC (H.265) specification. Selections are grouped as [Profile] / [Tier], so "Main/High" represents Main Profile with High Tier. 4:2:2 profiles are only available with the HEVC 4:2:2 License.
2418
2616
  */
2419
2617
  CodecProfile?: H265CodecProfile | string;
2420
2618
  /**
2619
+ * @public
2421
2620
  * Specify whether to allow the number of B-frames in your output GOP structure to vary or not depending on your input video content. To improve the subjective video quality of your output that has high-motion content: Leave blank or keep the default value Adaptive. MediaConvert will use fewer B-frames for high-motion video content than low-motion content. The maximum number of B- frames is limited by the value that you choose for B-frames between reference frames. To use the same number B-frames for all types of content: Choose Static.
2422
2621
  */
2423
2622
  DynamicSubGop?: H265DynamicSubGop | string;
2424
2623
  /**
2624
+ * @public
2425
2625
  * Enable this setting to have the encoder reduce I-frame pop. I-frame pop appears as a visual flicker that can arise when the encoder saves bits by copying some macroblocks many times from frame to frame, and then refreshes them at the I-frame. When you enable this setting, the encoder updates these macroblocks slightly more often to smooth out the flicker. This setting is disabled by default. Related setting: In addition to enabling this setting, you must also set adaptiveQuantization to a value other than Off.
2426
2626
  */
2427
2627
  FlickerAdaptiveQuantization?: H265FlickerAdaptiveQuantization | string;
2428
2628
  /**
2629
+ * @public
2429
2630
  * Use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.
2430
2631
  */
2431
2632
  FramerateControl?: H265FramerateControl | string;
2432
2633
  /**
2634
+ * @public
2433
2635
  * Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate. For numerically simple conversions, such as 60 fps to 30 fps: We recommend that you keep the default value, Drop duplicate. For numerically complex conversions, to avoid stutter: Choose Interpolate. This results in a smooth picture, but might introduce undesirable video artifacts. For complex frame rate conversions, especially if your source video has already been converted from its original cadence: Choose FrameFormer to do motion-compensated interpolation. FrameFormer uses the best conversion method frame by frame. Note that using FrameFormer increases the transcoding time and incurs a significant add-on cost. When you choose FrameFormer, your input video resolution must be at least 128x96.
2434
2636
  */
2435
2637
  FramerateConversionAlgorithm?: H265FramerateConversionAlgorithm | string;
2436
2638
  /**
2639
+ * @public
2437
2640
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateDenominator to specify the denominator of this fraction. In this example, use 1001 for the value of FramerateDenominator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
2438
2641
  */
2439
2642
  FramerateDenominator?: number;
2440
2643
  /**
2644
+ * @public
2441
2645
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateNumerator to specify the numerator of this fraction. In this example, use 24000 for the value of FramerateNumerator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
2442
2646
  */
2443
2647
  FramerateNumerator?: number;
2444
2648
  /**
2649
+ * @public
2445
2650
  * Specify whether to allow B-frames to be referenced by other frame types. To use reference B-frames when your GOP structure has 1 or more B-frames: Leave blank or keep the default value Enabled. We recommend that you choose Enabled to help improve the video quality of your output relative to its bitrate. To not use reference B-frames: Choose Disabled.
2446
2651
  */
2447
2652
  GopBReference?: H265GopBReference | string;
2448
2653
  /**
2654
+ * @public
2449
2655
  * Specify the relative frequency of open to closed GOPs in this output. For example, if you want to allow four open GOPs and then require a closed GOP, set this value to 5. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, do this by keeping the default empty value. If you do explicitly specify a value, for segmented outputs, don't set this value to 0.
2450
2656
  */
2451
2657
  GopClosedCadence?: number;
2452
2658
  /**
2659
+ * @public
2453
2660
  * Use this setting only when you set GOP mode control to Specified, frames or Specified, seconds. Specify the GOP length using a whole number of frames or a decimal value of seconds. MediaConvert will interpret this value as frames or seconds depending on the value you choose for GOP mode control. If you want to allow MediaConvert to automatically determine GOP size, leave GOP size blank and set GOP mode control to Auto. If your output group specifies HLS, DASH, or CMAF, leave GOP size blank and set GOP mode control to Auto in each output in your output group.
2454
2661
  */
2455
2662
  GopSize?: number;
2456
2663
  /**
2664
+ * @public
2457
2665
  * Specify how the transcoder determines GOP size for this output. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, choose Auto and and leave GOP size blank. By default, if you don't specify GOP mode control, MediaConvert will use automatic behavior. If your output group specifies HLS, DASH, or CMAF, set GOP mode control to Auto and leave GOP size blank in each output in your output group. To explicitly specify the GOP length, choose Specified, frames or Specified, seconds and then provide the GOP length in the related setting GOP size.
2458
2666
  */
2459
2667
  GopSizeUnits?: H265GopSizeUnits | string;
2460
2668
  /**
2669
+ * @public
2461
2670
  * If your downstream systems have strict buffer requirements: Specify the minimum percentage of the HRD buffer that's available at the end of each encoded video segment. For the best video quality: Set to 0 or leave blank to automatically determine the final buffer fill percentage.
2462
2671
  */
2463
2672
  HrdBufferFinalFillPercentage?: number;
2464
2673
  /**
2674
+ * @public
2465
2675
  * Percentage of the buffer that should initially be filled (HRD buffer model).
2466
2676
  */
2467
2677
  HrdBufferInitialFillPercentage?: number;
2468
2678
  /**
2679
+ * @public
2469
2680
  * Size of buffer (HRD buffer model) in bits. For example, enter five megabits as 5000000.
2470
2681
  */
2471
2682
  HrdBufferSize?: number;
2472
2683
  /**
2684
+ * @public
2473
2685
  * Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.
2474
2686
  */
2475
2687
  InterlaceMode?: H265InterlaceMode | string;
2476
2688
  /**
2689
+ * @public
2477
2690
  * Maximum bitrate in bits/second. For example, enter five megabits per second as 5000000. Required when Rate control mode is QVBR.
2478
2691
  */
2479
2692
  MaxBitrate?: number;
2480
2693
  /**
2694
+ * @public
2481
2695
  * Use this setting only when you also enable Scene change detection. This setting determines how the encoder manages the spacing between I-frames that it inserts as part of the I-frame cadence and the I-frames that it inserts for Scene change detection. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, do this by keeping the default empty value. When you explicitly specify a value for this setting, the encoder determines whether to skip a cadence-driven I-frame by the value you set. For example, if you set Min I interval to 5 and a cadence-driven I-frame would fall within 5 frames of a scene-change I-frame, then the encoder skips the cadence-driven I-frame. In this way, one GOP is shrunk slightly and one GOP is stretched slightly. When the cadence-driven I-frames are farther from the scene-change I-frame than the value you set, then the encoder leaves all I-frames in place and the GOPs surrounding the scene change are smaller than the usual cadence GOPs.
2482
2696
  */
2483
2697
  MinIInterval?: number;
2484
2698
  /**
2699
+ * @public
2485
2700
  * Specify the number of B-frames between reference frames in this output. For the best video quality: Leave blank. MediaConvert automatically determines the number of B-frames to use based on the characteristics of your input video. To manually specify the number of B-frames between reference frames: Enter an integer from 0 to 7.
2486
2701
  */
2487
2702
  NumberBFramesBetweenReferenceFrames?: number;
2488
2703
  /**
2704
+ * @public
2489
2705
  * Number of reference frames to use. The encoder may use more than requested if using B-frames and/or interlaced encoding.
2490
2706
  */
2491
2707
  NumberReferenceFrames?: number;
2492
2708
  /**
2709
+ * @public
2493
2710
  * Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source, uses the PAR from your input video for your output. To specify a different PAR, choose any value other than Follow source. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings.
2494
2711
  */
2495
2712
  ParControl?: H265ParControl | string;
2496
2713
  /**
2714
+ * @public
2497
2715
  * Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parDenominator is 33.
2498
2716
  */
2499
2717
  ParDenominator?: number;
2500
2718
  /**
2719
+ * @public
2501
2720
  * Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parNumerator is 40.
2502
2721
  */
2503
2722
  ParNumerator?: number;
2504
2723
  /**
2724
+ * @public
2505
2725
  * Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding.
2506
2726
  */
2507
2727
  QualityTuningLevel?: H265QualityTuningLevel | string;
2508
2728
  /**
2729
+ * @public
2509
2730
  * Settings for quality-defined variable bitrate encoding with the H.265 codec. Use these settings only when you set QVBR for Rate control mode.
2510
2731
  */
2511
2732
  QvbrSettings?: H265QvbrSettings;
2512
2733
  /**
2734
+ * @public
2513
2735
  * Use this setting to specify whether this output has a variable bitrate (VBR), constant bitrate (CBR) or quality-defined variable bitrate (QVBR).
2514
2736
  */
2515
2737
  RateControlMode?: H265RateControlMode | string;
2516
2738
  /**
2739
+ * @public
2517
2740
  * Specify Sample Adaptive Offset (SAO) filter strength. Adaptive mode dynamically selects best strength based on content
2518
2741
  */
2519
2742
  SampleAdaptiveOffsetFilterMode?: H265SampleAdaptiveOffsetFilterMode | string;
2520
2743
  /**
2744
+ * @public
2521
2745
  * Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive.
2522
2746
  */
2523
2747
  ScanTypeConversionMode?: H265ScanTypeConversionMode | string;
2524
2748
  /**
2749
+ * @public
2525
2750
  * Enable this setting to insert I-frames at scene changes that the service automatically detects. This improves video quality and is enabled by default. If this output uses QVBR, choose Transition detection for further video quality improvement. For more information about QVBR, see https://docs.aws.amazon.com/console/mediaconvert/cbr-vbr-qvbr.
2526
2751
  */
2527
2752
  SceneChangeDetect?: H265SceneChangeDetect | string;
2528
2753
  /**
2754
+ * @public
2529
2755
  * Number of slices per picture. Must be less than or equal to the number of macroblock rows for progressive pictures, and less than or equal to half the number of macroblock rows for interlaced pictures.
2530
2756
  */
2531
2757
  Slices?: number;
2532
2758
  /**
2759
+ * @public
2533
2760
  * Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25.
2534
2761
  */
2535
2762
  SlowPal?: H265SlowPal | string;
2536
2763
  /**
2764
+ * @public
2537
2765
  * Keep the default value, Enabled, to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher.
2538
2766
  */
2539
2767
  SpatialAdaptiveQuantization?: H265SpatialAdaptiveQuantization | string;
2540
2768
  /**
2769
+ * @public
2541
2770
  * This field applies only if the Streams > Advanced > Framerate field is set to 29.970. This field works with the Streams > Advanced > Preprocessors > Deinterlacer field and the Streams > Advanced > Interlaced Mode field to identify the scan type for the output: Progressive, Interlaced, Hard Telecine or Soft Telecine. - Hard: produces 29.97i output from 23.976 input. - Soft: produces 23.976; the player converts this output to 29.97i.
2542
2771
  */
2543
2772
  Telecine?: H265Telecine | string;
2544
2773
  /**
2774
+ * @public
2545
2775
  * Keep the default value, Enabled, to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to disable this feature. Related setting: When you enable temporal quantization, adjust the strength of the filter with the setting Adaptive quantization.
2546
2776
  */
2547
2777
  TemporalAdaptiveQuantization?: H265TemporalAdaptiveQuantization | string;
2548
2778
  /**
2779
+ * @public
2549
2780
  * Enables temporal layer identifiers in the encoded bitstream. Up to 3 layers are supported depending on GOP structure: I- and P-frames form one layer, reference B-frames can form a second layer and non-reference b-frames can form a third layer. Decoders can optionally decode only the lower temporal layers to generate a lower frame rate output. For example, given a bitstream with temporal IDs and with b-frames = 1 (i.e. IbPbPb display order), a decoder could decode all the frames for full frame rate output or only the I and P frames (lowest temporal layer) for a half frame rate output.
2550
2781
  */
2551
2782
  TemporalIds?: H265TemporalIds | string;
2552
2783
  /**
2784
+ * @public
2553
2785
  * Enable use of tiles, allowing horizontal as well as vertical subdivision of the encoded pictures.
2554
2786
  */
2555
2787
  Tiles?: H265Tiles | string;
2556
2788
  /**
2789
+ * @public
2557
2790
  * Inserts timecode for each frame as 4 bytes of an unregistered SEI message.
2558
2791
  */
2559
2792
  UnregisteredSeiTimecode?: H265UnregisteredSeiTimecode | string;
2560
2793
  /**
2794
+ * @public
2561
2795
  * If the location of parameter set NAL units doesn't matter in your workflow, ignore this setting. Use this setting only with CMAF or DASH outputs, or with standalone file outputs in an MPEG-4 container (MP4 outputs). Choose HVC1 to mark your output as HVC1. This makes your output compliant with the following specification: ISO IECJTC1 SC29 N13798 Text ISO/IEC FDIS 14496-15 3rd Edition. For these outputs, the service stores parameter set NAL units in the sample headers but not in the samples directly. For MP4 outputs, when you choose HVC1, your output video might not work properly with some downstream systems and video players. The service defaults to marking your output as HEV1. For these outputs, the service writes parameter set NAL units directly into the samples.
2562
2796
  */
2563
2797
  WriteMp4PackagingType?: H265WriteMp4PackagingType | string;
@@ -2809,134 +3043,167 @@ export type Mpeg2TemporalAdaptiveQuantization = (typeof Mpeg2TemporalAdaptiveQua
2809
3043
  */
2810
3044
  export interface Mpeg2Settings {
2811
3045
  /**
3046
+ * @public
2812
3047
  * Specify the strength of any adaptive quantization filters that you enable. The value that you choose here applies to the following settings: Spatial adaptive quantization, and Temporal adaptive quantization.
2813
3048
  */
2814
3049
  AdaptiveQuantization?: Mpeg2AdaptiveQuantization | string;
2815
3050
  /**
3051
+ * @public
2816
3052
  * Specify the average bitrate in bits per second. Required for VBR and CBR. For MS Smooth outputs, bitrates must be unique when rounded down to the nearest multiple of 1000.
2817
3053
  */
2818
3054
  Bitrate?: number;
2819
3055
  /**
3056
+ * @public
2820
3057
  * Use Level to set the MPEG-2 level for the video output.
2821
3058
  */
2822
3059
  CodecLevel?: Mpeg2CodecLevel | string;
2823
3060
  /**
3061
+ * @public
2824
3062
  * Use Profile to set the MPEG-2 profile for the video output.
2825
3063
  */
2826
3064
  CodecProfile?: Mpeg2CodecProfile | string;
2827
3065
  /**
3066
+ * @public
2828
3067
  * Choose Adaptive to improve subjective video quality for high-motion content. This will cause the service to use fewer B-frames (which infer information based on other frames) for high-motion portions of the video and more B-frames for low-motion portions. The maximum number of B-frames is limited by the value you provide for the setting B frames between reference frames.
2829
3068
  */
2830
3069
  DynamicSubGop?: Mpeg2DynamicSubGop | string;
2831
3070
  /**
3071
+ * @public
2832
3072
  * If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.
2833
3073
  */
2834
3074
  FramerateControl?: Mpeg2FramerateControl | string;
2835
3075
  /**
3076
+ * @public
2836
3077
  * Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate. For numerically simple conversions, such as 60 fps to 30 fps: We recommend that you keep the default value, Drop duplicate. For numerically complex conversions, to avoid stutter: Choose Interpolate. This results in a smooth picture, but might introduce undesirable video artifacts. For complex frame rate conversions, especially if your source video has already been converted from its original cadence: Choose FrameFormer to do motion-compensated interpolation. FrameFormer uses the best conversion method frame by frame. Note that using FrameFormer increases the transcoding time and incurs a significant add-on cost. When you choose FrameFormer, your input video resolution must be at least 128x96.
2837
3078
  */
2838
3079
  FramerateConversionAlgorithm?: Mpeg2FramerateConversionAlgorithm | string;
2839
3080
  /**
3081
+ * @public
2840
3082
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateDenominator to specify the denominator of this fraction. In this example, use 1001 for the value of FramerateDenominator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
2841
3083
  */
2842
3084
  FramerateDenominator?: number;
2843
3085
  /**
3086
+ * @public
2844
3087
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateNumerator to specify the numerator of this fraction. In this example, use 24000 for the value of FramerateNumerator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
2845
3088
  */
2846
3089
  FramerateNumerator?: number;
2847
3090
  /**
3091
+ * @public
2848
3092
  * Specify the relative frequency of open to closed GOPs in this output. For example, if you want to allow four open GOPs and then require a closed GOP, set this value to 5. When you create a streaming output, we recommend that you keep the default value, 1, so that players starting mid-stream receive an IDR frame as quickly as possible. Don't set this value to 0; that would break output segmenting.
2849
3093
  */
2850
3094
  GopClosedCadence?: number;
2851
3095
  /**
3096
+ * @public
2852
3097
  * Specify the interval between keyframes, in seconds or frames, for this output. Default: 12 Related settings: When you specify the GOP size in seconds, set GOP mode control to Specified, seconds. The default value for GOP mode control is Frames.
2853
3098
  */
2854
3099
  GopSize?: number;
2855
3100
  /**
3101
+ * @public
2856
3102
  * Specify the units for GOP size. If you don't specify a value here, by default the encoder measures GOP size in frames.
2857
3103
  */
2858
3104
  GopSizeUnits?: Mpeg2GopSizeUnits | string;
2859
3105
  /**
3106
+ * @public
2860
3107
  * If your downstream systems have strict buffer requirements: Specify the minimum percentage of the HRD buffer that's available at the end of each encoded video segment. For the best video quality: Set to 0 or leave blank to automatically determine the final buffer fill percentage.
2861
3108
  */
2862
3109
  HrdBufferFinalFillPercentage?: number;
2863
3110
  /**
3111
+ * @public
2864
3112
  * Percentage of the buffer that should initially be filled (HRD buffer model).
2865
3113
  */
2866
3114
  HrdBufferInitialFillPercentage?: number;
2867
3115
  /**
3116
+ * @public
2868
3117
  * Size of buffer (HRD buffer model) in bits. For example, enter five megabits as 5000000.
2869
3118
  */
2870
3119
  HrdBufferSize?: number;
2871
3120
  /**
3121
+ * @public
2872
3122
  * Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.
2873
3123
  */
2874
3124
  InterlaceMode?: Mpeg2InterlaceMode | string;
2875
3125
  /**
3126
+ * @public
2876
3127
  * Use Intra DC precision to set quantization precision for intra-block DC coefficients. If you choose the value auto, the service will automatically select the precision based on the per-frame compression ratio.
2877
3128
  */
2878
3129
  IntraDcPrecision?: Mpeg2IntraDcPrecision | string;
2879
3130
  /**
3131
+ * @public
2880
3132
  * Maximum bitrate in bits/second. For example, enter five megabits per second as 5000000.
2881
3133
  */
2882
3134
  MaxBitrate?: number;
2883
3135
  /**
3136
+ * @public
2884
3137
  * Use this setting only when you also enable Scene change detection. This setting determines how the encoder manages the spacing between I-frames that it inserts as part of the I-frame cadence and the I-frames that it inserts for Scene change detection. When you specify a value for this setting, the encoder determines whether to skip a cadence-driven I-frame by the value you set. For example, if you set Min I interval to 5 and a cadence-driven I-frame would fall within 5 frames of a scene-change I-frame, then the encoder skips the cadence-driven I-frame. In this way, one GOP is shrunk slightly and one GOP is stretched slightly. When the cadence-driven I-frames are farther from the scene-change I-frame than the value you set, then the encoder leaves all I-frames in place and the GOPs surrounding the scene change are smaller than the usual cadence GOPs.
2885
3138
  */
2886
3139
  MinIInterval?: number;
2887
3140
  /**
3141
+ * @public
2888
3142
  * Specify the number of B-frames that MediaConvert puts between reference frames in this output. Valid values are whole numbers from 0 through 7. When you don't specify a value, MediaConvert defaults to 2.
2889
3143
  */
2890
3144
  NumberBFramesBetweenReferenceFrames?: number;
2891
3145
  /**
3146
+ * @public
2892
3147
  * Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source, uses the PAR from your input video for your output. To specify a different PAR in the console, choose any value other than Follow source. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings.
2893
3148
  */
2894
3149
  ParControl?: Mpeg2ParControl | string;
2895
3150
  /**
3151
+ * @public
2896
3152
  * Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parDenominator is 33.
2897
3153
  */
2898
3154
  ParDenominator?: number;
2899
3155
  /**
3156
+ * @public
2900
3157
  * Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parNumerator is 40.
2901
3158
  */
2902
3159
  ParNumerator?: number;
2903
3160
  /**
3161
+ * @public
2904
3162
  * Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding.
2905
3163
  */
2906
3164
  QualityTuningLevel?: Mpeg2QualityTuningLevel | string;
2907
3165
  /**
3166
+ * @public
2908
3167
  * Use Rate control mode to specify whether the bitrate is variable (vbr) or constant (cbr).
2909
3168
  */
2910
3169
  RateControlMode?: Mpeg2RateControlMode | string;
2911
3170
  /**
3171
+ * @public
2912
3172
  * Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive.
2913
3173
  */
2914
3174
  ScanTypeConversionMode?: Mpeg2ScanTypeConversionMode | string;
2915
3175
  /**
3176
+ * @public
2916
3177
  * Enable this setting to insert I-frames at scene changes that the service automatically detects. This improves video quality and is enabled by default.
2917
3178
  */
2918
3179
  SceneChangeDetect?: Mpeg2SceneChangeDetect | string;
2919
3180
  /**
3181
+ * @public
2920
3182
  * Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25.
2921
3183
  */
2922
3184
  SlowPal?: Mpeg2SlowPal | string;
2923
3185
  /**
3186
+ * @public
2924
3187
  * Ignore this setting unless you need to comply with a specification that requires a specific value. If you don't have a specification requirement, we recommend that you adjust the softness of your output by using a lower value for the setting Sharpness or by enabling a noise reducer filter. The Softness setting specifies the quantization matrices that the encoder uses. Keep the default value, 0, to use the AWS Elemental default matrices. Choose a value from 17 to 128 to use planar interpolation. Increasing values from 17 to 128 result in increasing reduction of high-frequency data. The value 128 results in the softest video.
2925
3188
  */
2926
3189
  Softness?: number;
2927
3190
  /**
3191
+ * @public
2928
3192
  * Keep the default value, Enabled, to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher.
2929
3193
  */
2930
3194
  SpatialAdaptiveQuantization?: Mpeg2SpatialAdaptiveQuantization | string;
2931
3195
  /**
3196
+ * @public
2932
3197
  * Specify whether this output's video uses the D10 syntax. Keep the default value to not use the syntax. Related settings: When you choose D10 for your MXF profile, you must also set this value to D10.
2933
3198
  */
2934
3199
  Syntax?: Mpeg2Syntax | string;
2935
3200
  /**
3201
+ * @public
2936
3202
  * When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard or soft telecine to create a smoother picture. Hard telecine produces a 29.97i output. Soft telecine produces an output with a 23.976 output that signals to the video player device to do the conversion during play back. When you keep the default value, None, MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture.
2937
3203
  */
2938
3204
  Telecine?: Mpeg2Telecine | string;
2939
3205
  /**
3206
+ * @public
2940
3207
  * Keep the default value, Enabled, to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to disable this feature. Related setting: When you enable temporal quantization, adjust the strength of the filter with the setting Adaptive quantization.
2941
3208
  */
2942
3209
  TemporalAdaptiveQuantization?: Mpeg2TemporalAdaptiveQuantization | string;
@@ -3063,54 +3330,67 @@ export type ProresTelecine = (typeof ProresTelecine)[keyof typeof ProresTelecine
3063
3330
  */
3064
3331
  export interface ProresSettings {
3065
3332
  /**
3333
+ * @public
3066
3334
  * This setting applies only to ProRes 4444 and ProRes 4444 XQ outputs that you create from inputs that use 4:4:4 chroma sampling. Set Preserve 4:4:4 sampling to allow outputs to also use 4:4:4 chroma sampling. You must specify a value for this setting when your output codec profile supports 4:4:4 chroma sampling. Related Settings: For Apple ProRes outputs with 4:4:4 chroma sampling: Choose Preserve 4:4:4 sampling. Use when your input has 4:4:4 chroma sampling and your output codec Profile is Apple ProRes 4444 or 4444 XQ. Note that when you choose Preserve 4:4:4 sampling, you cannot include any of the following Preprocessors: Dolby Vision, HDR10+, or Noise reducer.
3067
3335
  */
3068
3336
  ChromaSampling?: ProresChromaSampling | string;
3069
3337
  /**
3338
+ * @public
3070
3339
  * Use Profile to specify the type of Apple ProRes codec to use for this output.
3071
3340
  */
3072
3341
  CodecProfile?: ProresCodecProfile | string;
3073
3342
  /**
3343
+ * @public
3074
3344
  * If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.
3075
3345
  */
3076
3346
  FramerateControl?: ProresFramerateControl | string;
3077
3347
  /**
3348
+ * @public
3078
3349
  * Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate. For numerically simple conversions, such as 60 fps to 30 fps: We recommend that you keep the default value, Drop duplicate. For numerically complex conversions, to avoid stutter: Choose Interpolate. This results in a smooth picture, but might introduce undesirable video artifacts. For complex frame rate conversions, especially if your source video has already been converted from its original cadence: Choose FrameFormer to do motion-compensated interpolation. FrameFormer uses the best conversion method frame by frame. Note that using FrameFormer increases the transcoding time and incurs a significant add-on cost. When you choose FrameFormer, your input video resolution must be at least 128x96.
3079
3350
  */
3080
3351
  FramerateConversionAlgorithm?: ProresFramerateConversionAlgorithm | string;
3081
3352
  /**
3353
+ * @public
3082
3354
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateDenominator to specify the denominator of this fraction. In this example, use 1001 for the value of FramerateDenominator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
3083
3355
  */
3084
3356
  FramerateDenominator?: number;
3085
3357
  /**
3358
+ * @public
3086
3359
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateNumerator to specify the numerator of this fraction. In this example, use 24000 for the value of FramerateNumerator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
3087
3360
  */
3088
3361
  FramerateNumerator?: number;
3089
3362
  /**
3363
+ * @public
3090
3364
  * Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.
3091
3365
  */
3092
3366
  InterlaceMode?: ProresInterlaceMode | string;
3093
3367
  /**
3368
+ * @public
3094
3369
  * Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source, uses the PAR from your input video for your output. To specify a different PAR, choose any value other than Follow source. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings.
3095
3370
  */
3096
3371
  ParControl?: ProresParControl | string;
3097
3372
  /**
3373
+ * @public
3098
3374
  * Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parDenominator is 33.
3099
3375
  */
3100
3376
  ParDenominator?: number;
3101
3377
  /**
3378
+ * @public
3102
3379
  * Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parNumerator is 40.
3103
3380
  */
3104
3381
  ParNumerator?: number;
3105
3382
  /**
3383
+ * @public
3106
3384
  * Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive.
3107
3385
  */
3108
3386
  ScanTypeConversionMode?: ProresScanTypeConversionMode | string;
3109
3387
  /**
3388
+ * @public
3110
3389
  * Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25.
3111
3390
  */
3112
3391
  SlowPal?: ProresSlowPal | string;
3113
3392
  /**
3393
+ * @public
3114
3394
  * When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard telecine to create a smoother picture. When you keep the default value, None, MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture.
3115
3395
  */
3116
3396
  Telecine?: ProresTelecine | string;
@@ -3207,38 +3487,47 @@ export type Vc3Class = (typeof Vc3Class)[keyof typeof Vc3Class];
3207
3487
  */
3208
3488
  export interface Vc3Settings {
3209
3489
  /**
3490
+ * @public
3210
3491
  * If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.
3211
3492
  */
3212
3493
  FramerateControl?: Vc3FramerateControl | string;
3213
3494
  /**
3495
+ * @public
3214
3496
  * Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate. For numerically simple conversions, such as 60 fps to 30 fps: We recommend that you keep the default value, Drop duplicate. For numerically complex conversions, to avoid stutter: Choose Interpolate. This results in a smooth picture, but might introduce undesirable video artifacts. For complex frame rate conversions, especially if your source video has already been converted from its original cadence: Choose FrameFormer to do motion-compensated interpolation. FrameFormer uses the best conversion method frame by frame. Note that using FrameFormer increases the transcoding time and incurs a significant add-on cost. When you choose FrameFormer, your input video resolution must be at least 128x96.
3215
3497
  */
3216
3498
  FramerateConversionAlgorithm?: Vc3FramerateConversionAlgorithm | string;
3217
3499
  /**
3500
+ * @public
3218
3501
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateDenominator to specify the denominator of this fraction. In this example, use 1001 for the value of FramerateDenominator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
3219
3502
  */
3220
3503
  FramerateDenominator?: number;
3221
3504
  /**
3505
+ * @public
3222
3506
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateNumerator to specify the numerator of this fraction. In this example, use 24000 for the value of FramerateNumerator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
3223
3507
  */
3224
3508
  FramerateNumerator?: number;
3225
3509
  /**
3510
+ * @public
3226
3511
  * Optional. Choose the scan line type for this output. If you don't specify a value, MediaConvert will create a progressive output.
3227
3512
  */
3228
3513
  InterlaceMode?: Vc3InterlaceMode | string;
3229
3514
  /**
3515
+ * @public
3230
3516
  * Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive.
3231
3517
  */
3232
3518
  ScanTypeConversionMode?: Vc3ScanTypeConversionMode | string;
3233
3519
  /**
3520
+ * @public
3234
3521
  * Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output by relabeling the video frames and resampling your audio. Note that enabling this setting will slightly reduce the duration of your video. Related settings: You must also set Framerate to 25.
3235
3522
  */
3236
3523
  SlowPal?: Vc3SlowPal | string;
3237
3524
  /**
3525
+ * @public
3238
3526
  * When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard telecine to create a smoother picture. When you keep the default value, None, MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture.
3239
3527
  */
3240
3528
  Telecine?: Vc3Telecine | string;
3241
3529
  /**
3530
+ * @public
3242
3531
  * Specify the VC3 class to choose the quality characteristics for this output. VC3 class, together with the settings Framerate (framerateNumerator and framerateDenominator) and Resolution (height and width), determine your output bitrate. For example, say that your video resolution is 1920x1080 and your framerate is 29.97. Then Class 145 gives you an output with a bitrate of approximately 145 Mbps and Class 220 gives you and output with a bitrate of approximately 220 Mbps. VC3 class also specifies the color bit depth of your output.
3243
3532
  */
3244
3533
  Vc3Class?: Vc3Class | string;
@@ -3309,54 +3598,67 @@ export type Vp8RateControlMode = (typeof Vp8RateControlMode)[keyof typeof Vp8Rat
3309
3598
  */
3310
3599
  export interface Vp8Settings {
3311
3600
  /**
3601
+ * @public
3312
3602
  * Target bitrate in bits/second. For example, enter five megabits per second as 5000000.
3313
3603
  */
3314
3604
  Bitrate?: number;
3315
3605
  /**
3606
+ * @public
3316
3607
  * If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.
3317
3608
  */
3318
3609
  FramerateControl?: Vp8FramerateControl | string;
3319
3610
  /**
3611
+ * @public
3320
3612
  * Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate. For numerically simple conversions, such as 60 fps to 30 fps: We recommend that you keep the default value, Drop duplicate. For numerically complex conversions, to avoid stutter: Choose Interpolate. This results in a smooth picture, but might introduce undesirable video artifacts. For complex frame rate conversions, especially if your source video has already been converted from its original cadence: Choose FrameFormer to do motion-compensated interpolation. FrameFormer uses the best conversion method frame by frame. Note that using FrameFormer increases the transcoding time and incurs a significant add-on cost. When you choose FrameFormer, your input video resolution must be at least 128x96.
3321
3613
  */
3322
3614
  FramerateConversionAlgorithm?: Vp8FramerateConversionAlgorithm | string;
3323
3615
  /**
3616
+ * @public
3324
3617
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateDenominator to specify the denominator of this fraction. In this example, use 1001 for the value of FramerateDenominator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
3325
3618
  */
3326
3619
  FramerateDenominator?: number;
3327
3620
  /**
3621
+ * @public
3328
3622
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateNumerator to specify the numerator of this fraction. In this example, use 24000 for the value of FramerateNumerator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
3329
3623
  */
3330
3624
  FramerateNumerator?: number;
3331
3625
  /**
3626
+ * @public
3332
3627
  * GOP Length (keyframe interval) in frames. Must be greater than zero.
3333
3628
  */
3334
3629
  GopSize?: number;
3335
3630
  /**
3631
+ * @public
3336
3632
  * Optional. Size of buffer (HRD buffer model) in bits. For example, enter five megabits as 5000000.
3337
3633
  */
3338
3634
  HrdBufferSize?: number;
3339
3635
  /**
3636
+ * @public
3340
3637
  * Ignore this setting unless you set qualityTuningLevel to MULTI_PASS. Optional. Specify the maximum bitrate in bits/second. For example, enter five megabits per second as 5000000. The default behavior uses twice the target bitrate as the maximum bitrate.
3341
3638
  */
3342
3639
  MaxBitrate?: number;
3343
3640
  /**
3641
+ * @public
3344
3642
  * Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source, uses the PAR from your input video for your output. To specify a different PAR in the console, choose any value other than Follow source. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings.
3345
3643
  */
3346
3644
  ParControl?: Vp8ParControl | string;
3347
3645
  /**
3646
+ * @public
3348
3647
  * Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parDenominator is 33.
3349
3648
  */
3350
3649
  ParDenominator?: number;
3351
3650
  /**
3651
+ * @public
3352
3652
  * Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parNumerator is 40.
3353
3653
  */
3354
3654
  ParNumerator?: number;
3355
3655
  /**
3656
+ * @public
3356
3657
  * Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, multi-pass encoding.
3357
3658
  */
3358
3659
  QualityTuningLevel?: Vp8QualityTuningLevel | string;
3359
3660
  /**
3661
+ * @public
3360
3662
  * With the VP8 codec, you can use only the variable bitrate (VBR) rate control mode.
3361
3663
  */
3362
3664
  RateControlMode?: Vp8RateControlMode | string;
@@ -3427,54 +3729,67 @@ export type Vp9RateControlMode = (typeof Vp9RateControlMode)[keyof typeof Vp9Rat
3427
3729
  */
3428
3730
  export interface Vp9Settings {
3429
3731
  /**
3732
+ * @public
3430
3733
  * Target bitrate in bits/second. For example, enter five megabits per second as 5000000.
3431
3734
  */
3432
3735
  Bitrate?: number;
3433
3736
  /**
3737
+ * @public
3434
3738
  * If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.
3435
3739
  */
3436
3740
  FramerateControl?: Vp9FramerateControl | string;
3437
3741
  /**
3742
+ * @public
3438
3743
  * Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate. For numerically simple conversions, such as 60 fps to 30 fps: We recommend that you keep the default value, Drop duplicate. For numerically complex conversions, to avoid stutter: Choose Interpolate. This results in a smooth picture, but might introduce undesirable video artifacts. For complex frame rate conversions, especially if your source video has already been converted from its original cadence: Choose FrameFormer to do motion-compensated interpolation. FrameFormer uses the best conversion method frame by frame. Note that using FrameFormer increases the transcoding time and incurs a significant add-on cost. When you choose FrameFormer, your input video resolution must be at least 128x96.
3439
3744
  */
3440
3745
  FramerateConversionAlgorithm?: Vp9FramerateConversionAlgorithm | string;
3441
3746
  /**
3747
+ * @public
3442
3748
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateDenominator to specify the denominator of this fraction. In this example, use 1001 for the value of FramerateDenominator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
3443
3749
  */
3444
3750
  FramerateDenominator?: number;
3445
3751
  /**
3752
+ * @public
3446
3753
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateNumerator to specify the numerator of this fraction. In this example, use 24000 for the value of FramerateNumerator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
3447
3754
  */
3448
3755
  FramerateNumerator?: number;
3449
3756
  /**
3757
+ * @public
3450
3758
  * GOP Length (keyframe interval) in frames. Must be greater than zero.
3451
3759
  */
3452
3760
  GopSize?: number;
3453
3761
  /**
3762
+ * @public
3454
3763
  * Size of buffer (HRD buffer model) in bits. For example, enter five megabits as 5000000.
3455
3764
  */
3456
3765
  HrdBufferSize?: number;
3457
3766
  /**
3767
+ * @public
3458
3768
  * Ignore this setting unless you set qualityTuningLevel to MULTI_PASS. Optional. Specify the maximum bitrate in bits/second. For example, enter five megabits per second as 5000000. The default behavior uses twice the target bitrate as the maximum bitrate.
3459
3769
  */
3460
3770
  MaxBitrate?: number;
3461
3771
  /**
3772
+ * @public
3462
3773
  * Optional. Specify how the service determines the pixel aspect ratio for this output. The default behavior is to use the same pixel aspect ratio as your input video.
3463
3774
  */
3464
3775
  ParControl?: Vp9ParControl | string;
3465
3776
  /**
3777
+ * @public
3466
3778
  * Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parDenominator is 33.
3467
3779
  */
3468
3780
  ParDenominator?: number;
3469
3781
  /**
3782
+ * @public
3470
3783
  * Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parNumerator is 40.
3471
3784
  */
3472
3785
  ParNumerator?: number;
3473
3786
  /**
3787
+ * @public
3474
3788
  * Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, multi-pass encoding.
3475
3789
  */
3476
3790
  QualityTuningLevel?: Vp9QualityTuningLevel | string;
3477
3791
  /**
3792
+ * @public
3478
3793
  * With the VP9 codec, you can use only the variable bitrate (VBR) rate control mode.
3479
3794
  */
3480
3795
  RateControlMode?: Vp9RateControlMode | string;
@@ -3604,6 +3919,7 @@ export type Xavc4kIntraCbgProfileClass = (typeof Xavc4kIntraCbgProfileClass)[key
3604
3919
  */
3605
3920
  export interface Xavc4kIntraCbgProfileSettings {
3606
3921
  /**
3922
+ * @public
3607
3923
  * Specify the XAVC Intra 4k (CBG) Class to set the bitrate of your output. Outputs of the same class have similar image quality over the operating points that are valid for that class.
3608
3924
  */
3609
3925
  XavcClass?: Xavc4kIntraCbgProfileClass | string;
@@ -3627,6 +3943,7 @@ export type Xavc4kIntraVbrProfileClass = (typeof Xavc4kIntraVbrProfileClass)[key
3627
3943
  */
3628
3944
  export interface Xavc4kIntraVbrProfileSettings {
3629
3945
  /**
3946
+ * @public
3630
3947
  * Specify the XAVC Intra 4k (VBR) Class to set the bitrate of your output. Outputs of the same class have similar image quality over the operating points that are valid for that class.
3631
3948
  */
3632
3949
  XavcClass?: Xavc4kIntraVbrProfileClass | string;
@@ -3699,34 +4016,42 @@ export type Xavc4kProfileQualityTuningLevel = (typeof Xavc4kProfileQualityTuning
3699
4016
  */
3700
4017
  export interface Xavc4kProfileSettings {
3701
4018
  /**
4019
+ * @public
3702
4020
  * Specify the XAVC 4k (Long GOP) Bitrate Class to set the bitrate of your output. Outputs of the same class have similar image quality over the operating points that are valid for that class.
3703
4021
  */
3704
4022
  BitrateClass?: Xavc4kProfileBitrateClass | string;
3705
4023
  /**
4024
+ * @public
3706
4025
  * Specify the codec profile for this output. Choose High, 8-bit, 4:2:0 (HIGH) or High, 10-bit, 4:2:2 (HIGH_422). These profiles are specified in ITU-T H.264.
3707
4026
  */
3708
4027
  CodecProfile?: Xavc4kProfileCodecProfile | string;
3709
4028
  /**
4029
+ * @public
3710
4030
  * The best way to set up adaptive quantization is to keep the default value, Auto, for the setting Adaptive quantization. When you do so, MediaConvert automatically applies the best types of quantization for your video content. Include this setting in your JSON job specification only when you choose to change the default value for Adaptive quantization. Enable this setting to have the encoder reduce I-frame pop. I-frame pop appears as a visual flicker that can arise when the encoder saves bits by copying some macroblocks many times from frame to frame, and then refreshes them at the I-frame. When you enable this setting, the encoder updates these macroblocks slightly more often to smooth out the flicker. This setting is disabled by default. Related setting: In addition to enabling this setting, you must also set Adaptive quantization to a value other than Off or Auto. Use Adaptive quantization to adjust the degree of smoothing that Flicker adaptive quantization provides.
3711
4031
  */
3712
4032
  FlickerAdaptiveQuantization?: XavcFlickerAdaptiveQuantization | string;
3713
4033
  /**
4034
+ * @public
3714
4035
  * Specify whether the encoder uses B-frames as reference frames for other pictures in the same GOP. Choose Allow to allow the encoder to use B-frames as reference frames. Choose Don't allow to prevent the encoder from using B-frames as reference frames.
3715
4036
  */
3716
4037
  GopBReference?: XavcGopBReference | string;
3717
4038
  /**
4039
+ * @public
3718
4040
  * Frequency of closed GOPs. In streaming applications, it is recommended that this be set to 1 so a decoder joining mid-stream will receive an IDR frame as quickly as possible. Setting this value to 0 will break output segmenting.
3719
4041
  */
3720
4042
  GopClosedCadence?: number;
3721
4043
  /**
4044
+ * @public
3722
4045
  * Specify the size of the buffer that MediaConvert uses in the HRD buffer model for this output. Specify this value in bits; for example, enter five megabits as 5000000. When you don't set this value, or you set it to zero, MediaConvert calculates the default by doubling the bitrate of this output point.
3723
4046
  */
3724
4047
  HrdBufferSize?: number;
3725
4048
  /**
4049
+ * @public
3726
4050
  * Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding.
3727
4051
  */
3728
4052
  QualityTuningLevel?: Xavc4kProfileQualityTuningLevel | string;
3729
4053
  /**
4054
+ * @public
3730
4055
  * Number of slices per picture. Must be less than or equal to the number of macroblock rows for progressive pictures, and less than or equal to half the number of macroblock rows for interlaced pictures.
3731
4056
  */
3732
4057
  Slices?: number;
@@ -3750,6 +4075,7 @@ export type XavcHdIntraCbgProfileClass = (typeof XavcHdIntraCbgProfileClass)[key
3750
4075
  */
3751
4076
  export interface XavcHdIntraCbgProfileSettings {
3752
4077
  /**
4078
+ * @public
3753
4079
  * Specify the XAVC Intra HD (CBG) Class to set the bitrate of your output. Outputs of the same class have similar image quality over the operating points that are valid for that class.
3754
4080
  */
3755
4081
  XavcClass?: XavcHdIntraCbgProfileClass | string;
@@ -3813,38 +4139,47 @@ export type XavcHdProfileTelecine = (typeof XavcHdProfileTelecine)[keyof typeof
3813
4139
  */
3814
4140
  export interface XavcHdProfileSettings {
3815
4141
  /**
4142
+ * @public
3816
4143
  * Specify the XAVC HD (Long GOP) Bitrate Class to set the bitrate of your output. Outputs of the same class have similar image quality over the operating points that are valid for that class.
3817
4144
  */
3818
4145
  BitrateClass?: XavcHdProfileBitrateClass | string;
3819
4146
  /**
4147
+ * @public
3820
4148
  * The best way to set up adaptive quantization is to keep the default value, Auto, for the setting Adaptive quantization. When you do so, MediaConvert automatically applies the best types of quantization for your video content. Include this setting in your JSON job specification only when you choose to change the default value for Adaptive quantization. Enable this setting to have the encoder reduce I-frame pop. I-frame pop appears as a visual flicker that can arise when the encoder saves bits by copying some macroblocks many times from frame to frame, and then refreshes them at the I-frame. When you enable this setting, the encoder updates these macroblocks slightly more often to smooth out the flicker. This setting is disabled by default. Related setting: In addition to enabling this setting, you must also set Adaptive quantization to a value other than Off or Auto. Use Adaptive quantization to adjust the degree of smoothing that Flicker adaptive quantization provides.
3821
4149
  */
3822
4150
  FlickerAdaptiveQuantization?: XavcFlickerAdaptiveQuantization | string;
3823
4151
  /**
4152
+ * @public
3824
4153
  * Specify whether the encoder uses B-frames as reference frames for other pictures in the same GOP. Choose Allow to allow the encoder to use B-frames as reference frames. Choose Don't allow to prevent the encoder from using B-frames as reference frames.
3825
4154
  */
3826
4155
  GopBReference?: XavcGopBReference | string;
3827
4156
  /**
4157
+ * @public
3828
4158
  * Frequency of closed GOPs. In streaming applications, it is recommended that this be set to 1 so a decoder joining mid-stream will receive an IDR frame as quickly as possible. Setting this value to 0 will break output segmenting.
3829
4159
  */
3830
4160
  GopClosedCadence?: number;
3831
4161
  /**
4162
+ * @public
3832
4163
  * Specify the size of the buffer that MediaConvert uses in the HRD buffer model for this output. Specify this value in bits; for example, enter five megabits as 5000000. When you don't set this value, or you set it to zero, MediaConvert calculates the default by doubling the bitrate of this output point.
3833
4164
  */
3834
4165
  HrdBufferSize?: number;
3835
4166
  /**
4167
+ * @public
3836
4168
  * Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.
3837
4169
  */
3838
4170
  InterlaceMode?: XavcInterlaceMode | string;
3839
4171
  /**
4172
+ * @public
3840
4173
  * Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding.
3841
4174
  */
3842
4175
  QualityTuningLevel?: XavcHdProfileQualityTuningLevel | string;
3843
4176
  /**
4177
+ * @public
3844
4178
  * Number of slices per picture. Must be less than or equal to the number of macroblock rows for progressive pictures, and less than or equal to half the number of macroblock rows for interlaced pictures.
3845
4179
  */
3846
4180
  Slices?: number;
3847
4181
  /**
4182
+ * @public
3848
4183
  * Ignore this setting unless you set Frame rate (framerateNumerator divided by framerateDenominator) to 29.970. If your input framerate is 23.976, choose Hard. Otherwise, keep the default value None. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/working-with-telecine-and-inverse-telecine.html.
3849
4184
  */
3850
4185
  Telecine?: XavcHdProfileTelecine | string;
@@ -3855,66 +4190,82 @@ export interface XavcHdProfileSettings {
3855
4190
  */
3856
4191
  export interface XavcSettings {
3857
4192
  /**
4193
+ * @public
3858
4194
  * Keep the default value, Auto, for this setting to have MediaConvert automatically apply the best types of quantization for your video content. When you want to apply your quantization settings manually, you must set Adaptive quantization to a value other than Auto. Use this setting to specify the strength of any adaptive quantization filters that you enable. If you don't want MediaConvert to do any adaptive quantization in this transcode, set Adaptive quantization to Off. Related settings: The value that you choose here applies to the following settings: Flicker adaptive quantization (flickerAdaptiveQuantization), Spatial adaptive quantization, and Temporal adaptive quantization.
3859
4195
  */
3860
4196
  AdaptiveQuantization?: XavcAdaptiveQuantization | string;
3861
4197
  /**
4198
+ * @public
3862
4199
  * Optional. Choose a specific entropy encoding mode only when you want to override XAVC recommendations. If you choose the value auto, MediaConvert uses the mode that the XAVC file format specifies given this output's operating point.
3863
4200
  */
3864
4201
  EntropyEncoding?: XavcEntropyEncoding | string;
3865
4202
  /**
4203
+ * @public
3866
4204
  * If you are using the console, use the Frame rate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list. The framerates shown in the dropdown list are decimal approximations of fractions.
3867
4205
  */
3868
4206
  FramerateControl?: XavcFramerateControl | string;
3869
4207
  /**
4208
+ * @public
3870
4209
  * Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate. For numerically simple conversions, such as 60 fps to 30 fps: We recommend that you keep the default value, Drop duplicate. For numerically complex conversions, to avoid stutter: Choose Interpolate. This results in a smooth picture, but might introduce undesirable video artifacts. For complex frame rate conversions, especially if your source video has already been converted from its original cadence: Choose FrameFormer to do motion-compensated interpolation. FrameFormer uses the best conversion method frame by frame. Note that using FrameFormer increases the transcoding time and incurs a significant add-on cost. When you choose FrameFormer, your input video resolution must be at least 128x96.
3871
4210
  */
3872
4211
  FramerateConversionAlgorithm?: XavcFramerateConversionAlgorithm | string;
3873
4212
  /**
4213
+ * @public
3874
4214
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateDenominator to specify the denominator of this fraction. In this example, use 1001 for the value of FramerateDenominator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Frame rate. In this example, specify 23.976.
3875
4215
  */
3876
4216
  FramerateDenominator?: number;
3877
4217
  /**
4218
+ * @public
3878
4219
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateNumerator to specify the numerator of this fraction. In this example, use 24000 for the value of FramerateNumerator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
3879
4220
  */
3880
4221
  FramerateNumerator?: number;
3881
4222
  /**
4223
+ * @public
3882
4224
  * Specify the XAVC profile for this output. For more information, see the Sony documentation at https://www.xavc-info.org/. Note that MediaConvert doesn't support the interlaced video XAVC operating points for XAVC_HD_INTRA_CBG. To create an interlaced XAVC output, choose the profile XAVC_HD.
3883
4225
  */
3884
4226
  Profile?: XavcProfile | string;
3885
4227
  /**
4228
+ * @public
3886
4229
  * Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output by relabeling the video frames and resampling your audio. Note that enabling this setting will slightly reduce the duration of your video. Related settings: You must also set Frame rate to 25.
3887
4230
  */
3888
4231
  SlowPal?: XavcSlowPal | string;
3889
4232
  /**
4233
+ * @public
3890
4234
  * Ignore this setting unless your downstream workflow requires that you specify it explicitly. Otherwise, we recommend that you adjust the softness of your output by using a lower value for the setting Sharpness or by enabling a noise reducer filter. The Softness setting specifies the quantization matrices that the encoder uses. Keep the default value, 0, for flat quantization. Choose the value 1 or 16 to use the default JVT softening quantization matricies from the H.264 specification. Choose a value from 17 to 128 to use planar interpolation. Increasing values from 17 to 128 result in increasing reduction of high-frequency data. The value 128 results in the softest video.
3891
4235
  */
3892
4236
  Softness?: number;
3893
4237
  /**
4238
+ * @public
3894
4239
  * The best way to set up adaptive quantization is to keep the default value, Auto, for the setting Adaptive quantization. When you do so, MediaConvert automatically applies the best types of quantization for your video content. Include this setting in your JSON job specification only when you choose to change the default value for Adaptive quantization. For this setting, keep the default value, Enabled, to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher.
3895
4240
  */
3896
4241
  SpatialAdaptiveQuantization?: XavcSpatialAdaptiveQuantization | string;
3897
4242
  /**
4243
+ * @public
3898
4244
  * The best way to set up adaptive quantization is to keep the default value, Auto, for the setting Adaptive quantization. When you do so, MediaConvert automatically applies the best types of quantization for your video content. Include this setting in your JSON job specification only when you choose to change the default value for Adaptive quantization. For this setting, keep the default value, Enabled, to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to disable this feature. Related setting: When you enable temporal adaptive quantization, adjust the strength of the filter with the setting Adaptive quantization.
3899
4245
  */
3900
4246
  TemporalAdaptiveQuantization?: XavcTemporalAdaptiveQuantization | string;
3901
4247
  /**
4248
+ * @public
3902
4249
  * Required when you set Profile to the value XAVC_4K_INTRA_CBG.
3903
4250
  */
3904
4251
  Xavc4kIntraCbgProfileSettings?: Xavc4kIntraCbgProfileSettings;
3905
4252
  /**
4253
+ * @public
3906
4254
  * Required when you set Profile to the value XAVC_4K_INTRA_VBR.
3907
4255
  */
3908
4256
  Xavc4kIntraVbrProfileSettings?: Xavc4kIntraVbrProfileSettings;
3909
4257
  /**
4258
+ * @public
3910
4259
  * Required when you set Profile to the value XAVC_4K.
3911
4260
  */
3912
4261
  Xavc4kProfileSettings?: Xavc4kProfileSettings;
3913
4262
  /**
4263
+ * @public
3914
4264
  * Required when you set Profile to the value XAVC_HD_INTRA_CBG.
3915
4265
  */
3916
4266
  XavcHdIntraCbgProfileSettings?: XavcHdIntraCbgProfileSettings;
3917
4267
  /**
4268
+ * @public
3918
4269
  * Required when you set Profile to the value XAVC_HD.
3919
4270
  */
3920
4271
  XavcHdProfileSettings?: XavcHdProfileSettings;
@@ -3925,50 +4276,62 @@ export interface XavcSettings {
3925
4276
  */
3926
4277
  export interface VideoCodecSettings {
3927
4278
  /**
4279
+ * @public
3928
4280
  * Required when you set Codec, under VideoDescription>CodecSettings to the value AV1.
3929
4281
  */
3930
4282
  Av1Settings?: Av1Settings;
3931
4283
  /**
4284
+ * @public
3932
4285
  * Required when you choose AVC-Intra for your output video codec. For more information about the AVC-Intra settings, see the relevant specification. For detailed information about SD and HD in AVC-Intra, see https://ieeexplore.ieee.org/document/7290936. For information about 4K/2K in AVC-Intra, see https://pro-av.panasonic.net/en/avc-ultra/AVC-ULTRAoverview.pdf.
3933
4286
  */
3934
4287
  AvcIntraSettings?: AvcIntraSettings;
3935
4288
  /**
4289
+ * @public
3936
4290
  * Specifies the video codec. This must be equal to one of the enum values defined by the object VideoCodec. To passthrough the video stream of your input JPEG2000, VC-3, AVC-INTRA or Apple ProRes video without any video encoding: Choose Passthrough. If you have multiple input videos, note that they must have identical encoding attributes. When you choose Passthrough, your output container must be MXF or QuickTime MOV.
3937
4291
  */
3938
4292
  Codec?: VideoCodec | string;
3939
4293
  /**
4294
+ * @public
3940
4295
  * Required when you set Codec to the value FRAME_CAPTURE.
3941
4296
  */
3942
4297
  FrameCaptureSettings?: FrameCaptureSettings;
3943
4298
  /**
4299
+ * @public
3944
4300
  * Required when you set Codec to the value H_264.
3945
4301
  */
3946
4302
  H264Settings?: H264Settings;
3947
4303
  /**
4304
+ * @public
3948
4305
  * Settings for H265 codec
3949
4306
  */
3950
4307
  H265Settings?: H265Settings;
3951
4308
  /**
4309
+ * @public
3952
4310
  * Required when you set Codec to the value MPEG2.
3953
4311
  */
3954
4312
  Mpeg2Settings?: Mpeg2Settings;
3955
4313
  /**
4314
+ * @public
3956
4315
  * Required when you set Codec to the value PRORES.
3957
4316
  */
3958
4317
  ProresSettings?: ProresSettings;
3959
4318
  /**
4319
+ * @public
3960
4320
  * Required when you set Codec to the value VC3
3961
4321
  */
3962
4322
  Vc3Settings?: Vc3Settings;
3963
4323
  /**
4324
+ * @public
3964
4325
  * Required when you set Codec to the value VP8.
3965
4326
  */
3966
4327
  Vp8Settings?: Vp8Settings;
3967
4328
  /**
4329
+ * @public
3968
4330
  * Required when you set Codec to the value VP9.
3969
4331
  */
3970
4332
  Vp9Settings?: Vp9Settings;
3971
4333
  /**
4334
+ * @public
3972
4335
  * Required when you set Codec to the value XAVC.
3973
4336
  */
3974
4337
  XavcSettings?: XavcSettings;
@@ -4040,18 +4403,22 @@ export type VideoTimecodeInsertion = (typeof VideoTimecodeInsertion)[keyof typeo
4040
4403
  */
4041
4404
  export interface ClipLimits {
4042
4405
  /**
4406
+ * @public
4043
4407
  * Specify the Maximum RGB color sample range tolerance for your output. MediaConvert corrects any YUV values that, when converted to RGB, would be outside the upper tolerance that you specify. Enter an integer from 90 to 105 as an offset percentage to the maximum possible value. Leave blank to use the default value 100. When you specify a value for Maximum RGB tolerance, you must set Sample range conversion to Limited range clip.
4044
4408
  */
4045
4409
  MaximumRGBTolerance?: number;
4046
4410
  /**
4411
+ * @public
4047
4412
  * Specify the Maximum YUV color sample limit. MediaConvert conforms any pixels in your input above the value that you specify to typical limited range bounds. Enter an integer from 920 to 1023. Leave blank to use the default value 940. The value that you enter applies to 10-bit ranges. For 8-bit ranges, MediaConvert automatically scales this value down. When you specify a value for Maximum YUV, you must set Sample range conversion to Limited range clip.
4048
4413
  */
4049
4414
  MaximumYUV?: number;
4050
4415
  /**
4416
+ * @public
4051
4417
  * Specify the Minimum RGB color sample range tolerance for your output. MediaConvert corrects any YUV values that, when converted to RGB, would be outside the lower tolerance that you specify. Enter an integer from -5 to 10 as an offset percentage to the minimum possible value. Leave blank to use the default value 0. When you specify a value for Minimum RGB tolerance, you must set Sample range conversion to Limited range clip.
4052
4418
  */
4053
4419
  MinimumRGBTolerance?: number;
4054
4420
  /**
4421
+ * @public
4055
4422
  * Specify the Minimum YUV color sample limit. MediaConvert conforms any pixels in your input below the value that you specify to typical limited range bounds. Enter an integer from 0 to 128. Leave blank to use the default value 64. The value that you enter applies to 10-bit ranges. For 8-bit ranges, MediaConvert automatically scales this value down. When you specify a value for Minumum YUV, you must set Sample range conversion to Limited range clip.
4056
4423
  */
4057
4424
  MinimumYUV?: number;
@@ -4105,14 +4472,17 @@ export type SampleRangeConversion = (typeof SampleRangeConversion)[keyof typeof
4105
4472
  */
4106
4473
  export interface ColorCorrector {
4107
4474
  /**
4475
+ * @public
4108
4476
  * Brightness level.
4109
4477
  */
4110
4478
  Brightness?: number;
4111
4479
  /**
4480
+ * @public
4112
4481
  * Specify YUV limits and RGB tolerances when you set Sample range conversion to Limited range clip.
4113
4482
  */
4114
4483
  ClipLimits?: ClipLimits;
4115
4484
  /**
4485
+ * @public
4116
4486
  * Specify the color space you want for this output. The service supports conversion between HDR formats, between SDR formats, from SDR to HDR, and from HDR to SDR. SDR to HDR conversion doesn't upgrade the dynamic range. The converted video has an HDR format, but visually appears the same as an unconverted output. HDR to SDR conversion uses tone mapping to approximate the outcome of manually regrading from HDR to SDR. When you specify an output color space, MediaConvert uses the following color space metadata, which includes color primaries, transfer characteristics, and matrix coefficients:
4117
4487
  * * HDR 10: BT.2020, PQ, BT.2020 non-constant
4118
4488
  * * HLG 2020: BT.2020, HLG, BT.2020 non-constant
@@ -4122,30 +4492,37 @@ export interface ColorCorrector {
4122
4492
  */
4123
4493
  ColorSpaceConversion?: ColorSpaceConversion | string;
4124
4494
  /**
4495
+ * @public
4125
4496
  * Contrast level.
4126
4497
  */
4127
4498
  Contrast?: number;
4128
4499
  /**
4500
+ * @public
4129
4501
  * Use these settings when you convert to the HDR 10 color space. Specify the SMPTE ST 2086 Mastering Display Color Volume static metadata that you want signaled in the output. These values don't affect the pixel values that are encoded in the video stream. They are intended to help the downstream video player display content in a way that reflects the intentions of the the content creator. When you set Color space conversion to HDR 10, these settings are required. You must set values for Max frame average light level and Max content light level; these settings don't have a default value. The default values for the other HDR 10 metadata settings are defined by the P3D65 color space. For more information about MediaConvert HDR jobs, see https://docs.aws.amazon.com/console/mediaconvert/hdr.
4130
4502
  */
4131
4503
  Hdr10Metadata?: Hdr10Metadata;
4132
4504
  /**
4505
+ * @public
4133
4506
  * Specify how MediaConvert maps brightness and colors from your HDR input to your SDR output. The mode that you select represents a creative choice, with different tradeoffs in the details and tones of your output. To maintain details in bright or saturated areas of your output: Choose Preserve details. For some sources, your SDR output may look less bright and less saturated when compared to your HDR source. MediaConvert automatically applies this mode for HLG sources, regardless of your choice. For a bright and saturated output: Choose Vibrant. We recommend that you choose this mode when any of your source content is HDR10, and for the best results when it is mastered for 1000 nits. You may notice loss of details in bright or saturated areas of your output. HDR to SDR tone mapping has no effect when your input is SDR.
4134
4507
  */
4135
4508
  HdrToSdrToneMapper?: HDRToSDRToneMapper | string;
4136
4509
  /**
4510
+ * @public
4137
4511
  * Hue in degrees.
4138
4512
  */
4139
4513
  Hue?: number;
4140
4514
  /**
4515
+ * @public
4141
4516
  * Specify how MediaConvert limits the color sample range for this output. To create a limited range output from a full range input: Choose Limited range squeeze. For full range inputs, MediaConvert performs a linear offset to color samples equally across all pixels and frames. Color samples in 10-bit outputs are limited to 64 through 940, and 8-bit outputs are limited to 16 through 235. Note: For limited range inputs, values for color samples are passed through to your output unchanged. MediaConvert does not limit the sample range. To correct pixels in your input that are out of range or out of gamut: Choose Limited range clip. Use for broadcast applications. MediaConvert conforms any pixels outside of the values that you specify under Minimum YUV and Maximum YUV to limited range bounds. MediaConvert also corrects any YUV values that, when converted to RGB, would be outside the bounds you specify under Minimum RGB tolerance and Maximum RGB tolerance. With either limited range conversion, MediaConvert writes the sample range metadata in the output.
4142
4517
  */
4143
4518
  SampleRangeConversion?: SampleRangeConversion | string;
4144
4519
  /**
4520
+ * @public
4145
4521
  * Saturation level.
4146
4522
  */
4147
4523
  Saturation?: number;
4148
4524
  /**
4525
+ * @public
4149
4526
  * Specify the reference white level, in nits, for all of your SDR inputs. Use to correct brightness levels within HDR10 outputs. The following color metadata must be present in your SDR input: color primaries, transfer characteristics, and matrix coefficients. If your SDR input has missing color metadata, or if you want to correct input color metadata, manually specify a color space in the input video selector. For 1,000 nit peak brightness displays, we recommend that you set SDR reference white level to 203 (according to ITU-R BT.2408). Leave blank to use the default value of 100, or specify an integer from 100 to 1000.
4150
4527
  */
4151
4528
  SdrReferenceWhiteLevel?: number;
@@ -4196,14 +4573,17 @@ export type DeinterlacerMode = (typeof DeinterlacerMode)[keyof typeof Deinterlac
4196
4573
  */
4197
4574
  export interface Deinterlacer {
4198
4575
  /**
4576
+ * @public
4199
4577
  * Only applies when you set Deinterlace mode to Deinterlace or Adaptive. Interpolate produces sharper pictures, while blend produces smoother motion. If your source file includes a ticker, such as a scrolling headline at the bottom of the frame: Choose Interpolate ticker or Blend ticker. To apply field doubling: Choose Linear interpolation. Note that Linear interpolation may introduce video artifacts into your output.
4200
4578
  */
4201
4579
  Algorithm?: DeinterlaceAlgorithm | string;
4202
4580
  /**
4581
+ * @public
4203
4582
  * - When set to NORMAL (default), the deinterlacer does not convert frames that are tagged in metadata as progressive. It will only convert those that are tagged as some other type. - When set to FORCE_ALL_FRAMES, the deinterlacer converts every frame to progressive - even those that are already tagged as progressive. Turn Force mode on only if there is a good chance that the metadata has tagged frames as progressive when they are not progressive. Do not turn on otherwise; processing frames that are already progressive into progressive will probably result in lower quality video.
4204
4583
  */
4205
4584
  Control?: DeinterlacerControl | string;
4206
4585
  /**
4586
+ * @public
4207
4587
  * Use Deinterlacer to choose how the service will do deinterlacing. Default is Deinterlace.
4208
4588
  * - Deinterlace converts interlaced to progressive.
4209
4589
  * - Inverse telecine converts Hard Telecine 29.97i to progressive 23.976p.
@@ -4217,10 +4597,12 @@ export interface Deinterlacer {
4217
4597
  */
4218
4598
  export interface DolbyVisionLevel6Metadata {
4219
4599
  /**
4600
+ * @public
4220
4601
  * Maximum Content Light Level. Static HDR metadata that corresponds to the brightest pixel in the entire stream. Measured in nits.
4221
4602
  */
4222
4603
  MaxCll?: number;
4223
4604
  /**
4605
+ * @public
4224
4606
  * Maximum Frame-Average Light Level. Static HDR metadata that corresponds to the highest frame-average brightness in the entire stream. Measured in nits.
4225
4607
  */
4226
4608
  MaxFall?: number;
@@ -4268,18 +4650,22 @@ export type DolbyVisionProfile = (typeof DolbyVisionProfile)[keyof typeof DolbyV
4268
4650
  */
4269
4651
  export interface DolbyVision {
4270
4652
  /**
4653
+ * @public
4271
4654
  * Use these settings when you set DolbyVisionLevel6Mode to SPECIFY to override the MaxCLL and MaxFALL values in your input with new values.
4272
4655
  */
4273
4656
  L6Metadata?: DolbyVisionLevel6Metadata;
4274
4657
  /**
4658
+ * @public
4275
4659
  * Use Dolby Vision Mode to choose how the service will handle Dolby Vision MaxCLL and MaxFALL properies.
4276
4660
  */
4277
4661
  L6Mode?: DolbyVisionLevel6Mode | string;
4278
4662
  /**
4663
+ * @public
4279
4664
  * Required when you set Dolby Vision Profile to Profile 8.1. When you set Content mapping to None, content mapping is not applied to the HDR10-compatible signal. Depending on the source peak nit level, clipping might occur on HDR devices without Dolby Vision. When you set Content mapping to HDR10 1000, the transcoder creates a 1,000 nits peak HDR10-compatible signal by applying static content mapping to the source. This mode is speed-optimized for PQ10 sources with metadata that is created from analysis. For graded Dolby Vision content, be aware that creative intent might not be guaranteed with extreme 1,000 nits trims.
4280
4665
  */
4281
4666
  Mapping?: DolbyVisionMapping | string;
4282
4667
  /**
4668
+ * @public
4283
4669
  * Required when you enable Dolby Vision. Use Profile 5 to include frame-interleaved Dolby Vision metadata in your output. Your input must include Dolby Vision metadata or an HDR10 YUV color space. Use Profile 8.1 to include frame-interleaved Dolby Vision metadata and HDR10 metadata in your output. Your input must include Dolby Vision metadata.
4284
4670
  */
4285
4671
  Profile?: DolbyVisionProfile | string;
@@ -4290,10 +4676,12 @@ export interface DolbyVision {
4290
4676
  */
4291
4677
  export interface Hdr10Plus {
4292
4678
  /**
4679
+ * @public
4293
4680
  * Specify the HDR10+ mastering display normalized peak luminance, in nits. This is the normalized actual peak luminance of the mastering display, as defined by ST 2094-40.
4294
4681
  */
4295
4682
  MasteringMonitorNits?: number;
4296
4683
  /**
4684
+ * @public
4297
4685
  * Specify the HDR10+ target display nominal peak luminance, in nits. This is the nominal maximum luminance of the target display as defined by ST 2094-40.
4298
4686
  */
4299
4687
  TargetMonitorNits?: number;
@@ -4322,6 +4710,7 @@ export type NoiseReducerFilter = (typeof NoiseReducerFilter)[keyof typeof NoiseR
4322
4710
  */
4323
4711
  export interface NoiseReducerFilterSettings {
4324
4712
  /**
4713
+ * @public
4325
4714
  * Relative strength of noise reducing filter. Higher values produce stronger filtering.
4326
4715
  */
4327
4716
  Strength?: number;
@@ -4332,14 +4721,17 @@ export interface NoiseReducerFilterSettings {
4332
4721
  */
4333
4722
  export interface NoiseReducerSpatialFilterSettings {
4334
4723
  /**
4724
+ * @public
4335
4725
  * Specify strength of post noise reduction sharpening filter, with 0 disabling the filter and 3 enabling it at maximum strength.
4336
4726
  */
4337
4727
  PostFilterSharpenStrength?: number;
4338
4728
  /**
4729
+ * @public
4339
4730
  * The speed of the filter, from -2 (lower speed) to 3 (higher speed), with 0 being the nominal value.
4340
4731
  */
4341
4732
  Speed?: number;
4342
4733
  /**
4734
+ * @public
4343
4735
  * Relative strength of noise reducing filter. Higher values produce stronger filtering.
4344
4736
  */
4345
4737
  Strength?: number;
@@ -4376,22 +4768,27 @@ export type NoiseFilterPostTemporalSharpeningStrength = (typeof NoiseFilterPostT
4376
4768
  */
4377
4769
  export interface NoiseReducerTemporalFilterSettings {
4378
4770
  /**
4771
+ * @public
4379
4772
  * Use Aggressive mode for content that has complex motion. Higher values produce stronger temporal filtering. This filters highly complex scenes more aggressively and creates better VQ for low bitrate outputs.
4380
4773
  */
4381
4774
  AggressiveMode?: number;
4382
4775
  /**
4776
+ * @public
4383
4777
  * When you set Noise reducer to Temporal, the bandwidth and sharpness of your output is reduced. You can optionally use Post temporal sharpening to apply sharpening to the edges of your output. Note that Post temporal sharpening will also make the bandwidth reduction from the Noise reducer smaller. The default behavior, Auto, allows the transcoder to determine whether to apply sharpening, depending on your input type and quality. When you set Post temporal sharpening to Enabled, specify how much sharpening is applied using Post temporal sharpening strength. Set Post temporal sharpening to Disabled to not apply sharpening.
4384
4778
  */
4385
4779
  PostTemporalSharpening?: NoiseFilterPostTemporalSharpening | string;
4386
4780
  /**
4781
+ * @public
4387
4782
  * Use Post temporal sharpening strength to define the amount of sharpening the transcoder applies to your output. Set Post temporal sharpening strength to Low, Medium, or High to indicate the amount of sharpening.
4388
4783
  */
4389
4784
  PostTemporalSharpeningStrength?: NoiseFilterPostTemporalSharpeningStrength | string;
4390
4785
  /**
4786
+ * @public
4391
4787
  * The speed of the filter (higher number is faster). Low setting reduces bit rate at the cost of transcode time, high setting improves transcode time at the cost of bit rate.
4392
4788
  */
4393
4789
  Speed?: number;
4394
4790
  /**
4791
+ * @public
4395
4792
  * Specify the strength of the noise reducing filter on this output. Higher values produce stronger filtering. We recommend the following value ranges, depending on the result that you want: * 0-2 for complexity reduction with minimal sharpness loss * 2-8 for complexity reduction with image preservation * 8-16 for a high level of complexity reduction
4396
4793
  */
4397
4794
  Strength?: number;
@@ -4402,18 +4799,22 @@ export interface NoiseReducerTemporalFilterSettings {
4402
4799
  */
4403
4800
  export interface NoiseReducer {
4404
4801
  /**
4802
+ * @public
4405
4803
  * Use Noise reducer filter to select one of the following spatial image filtering functions. To use this setting, you must also enable Noise reducer. * Bilateral preserves edges while reducing noise. * Mean (softest), Gaussian, Lanczos, and Sharpen (sharpest) do convolution filtering. * Conserve does min/max noise reduction. * Spatial does frequency-domain filtering based on JND principles. * Temporal optimizes video quality for complex motion.
4406
4804
  */
4407
4805
  Filter?: NoiseReducerFilter | string;
4408
4806
  /**
4807
+ * @public
4409
4808
  * Settings for a noise reducer filter
4410
4809
  */
4411
4810
  FilterSettings?: NoiseReducerFilterSettings;
4412
4811
  /**
4812
+ * @public
4413
4813
  * Noise reducer filter settings for spatial filter.
4414
4814
  */
4415
4815
  SpatialFilterSettings?: NoiseReducerSpatialFilterSettings;
4416
4816
  /**
4817
+ * @public
4417
4818
  * Noise reducer filter settings for temporal filter.
4418
4819
  */
4419
4820
  TemporalFilterSettings?: NoiseReducerTemporalFilterSettings;
@@ -4439,18 +4840,22 @@ export type WatermarkingStrength = (typeof WatermarkingStrength)[keyof typeof Wa
4439
4840
  */
4440
4841
  export interface NexGuardFileMarkerSettings {
4441
4842
  /**
4843
+ * @public
4442
4844
  * Use the base64 license string that Nagra provides you. Enter it directly in your JSON job specification or in the console. Required when you include Nagra NexGuard File Marker watermarking in your job.
4443
4845
  */
4444
4846
  License?: string;
4445
4847
  /**
4848
+ * @public
4446
4849
  * Specify the payload ID that you want associated with this output. Valid values vary depending on your Nagra NexGuard forensic watermarking workflow. Required when you include Nagra NexGuard File Marker watermarking in your job. For PreRelease Content (NGPR/G2), specify an integer from 1 through 4,194,303. You must generate a unique ID for each asset you watermark, and keep a record of which ID you have assigned to each asset. Neither Nagra nor MediaConvert keep track of the relationship between output files and your IDs. For OTT Streaming, create two adaptive bitrate (ABR) stacks for each asset. Do this by setting up two output groups. For one output group, set the value of Payload ID to 0 in every output. For the other output group, set Payload ID to 1 in every output.
4447
4850
  */
4448
4851
  Payload?: number;
4449
4852
  /**
4853
+ * @public
4450
4854
  * Enter one of the watermarking preset strings that Nagra provides you. Required when you include Nagra NexGuard File Marker watermarking in your job.
4451
4855
  */
4452
4856
  Preset?: string;
4453
4857
  /**
4858
+ * @public
4454
4859
  * Optional. Ignore this setting unless Nagra support directs you to specify a value. When you don't specify a value here, the Nagra NexGuard library uses its default value.
4455
4860
  */
4456
4861
  Strength?: WatermarkingStrength | string;
@@ -4461,6 +4866,7 @@ export interface NexGuardFileMarkerSettings {
4461
4866
  */
4462
4867
  export interface PartnerWatermarking {
4463
4868
  /**
4869
+ * @public
4464
4870
  * For forensic video watermarking, MediaConvert supports Nagra NexGuard File Marker watermarking. MediaConvert supports both PreRelease Content (NGPR/G2) and OTT Streaming workflows.
4465
4871
  */
4466
4872
  NexguardFileMarkerSettings?: NexGuardFileMarkerSettings;
@@ -4490,14 +4896,17 @@ export type TimecodeBurninPosition = (typeof TimecodeBurninPosition)[keyof typeo
4490
4896
  */
4491
4897
  export interface TimecodeBurnin {
4492
4898
  /**
4899
+ * @public
4493
4900
  * Use Font size to set the font size of any burned-in timecode. Valid values are 10, 16, 32, 48.
4494
4901
  */
4495
4902
  FontSize?: number;
4496
4903
  /**
4904
+ * @public
4497
4905
  * Use Position under Timecode burn-in to specify the location the burned-in timecode on output video.
4498
4906
  */
4499
4907
  Position?: TimecodeBurninPosition | string;
4500
4908
  /**
4909
+ * @public
4501
4910
  * Use Prefix to place ASCII characters before any burned-in timecode. For example, a prefix of "EZ-" will result in the timecode "EZ-00:00:00:00". Provide either the characters themselves or the ASCII code equivalents. The supported range of characters is 0x20 through 0x7e. This includes letters, numbers, and all special characters represented on a standard English keyboard.
4502
4911
  */
4503
4912
  Prefix?: string;
@@ -4508,34 +4917,42 @@ export interface TimecodeBurnin {
4508
4917
  */
4509
4918
  export interface VideoPreprocessor {
4510
4919
  /**
4920
+ * @public
4511
4921
  * Use these settings to convert the color space or to modify properties such as hue and contrast for this output. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/converting-the-color-space.html.
4512
4922
  */
4513
4923
  ColorCorrector?: ColorCorrector;
4514
4924
  /**
4925
+ * @public
4515
4926
  * Use the deinterlacer to produce smoother motion and a clearer picture. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/working-with-scan-type.html.
4516
4927
  */
4517
4928
  Deinterlacer?: Deinterlacer;
4518
4929
  /**
4930
+ * @public
4519
4931
  * Enable Dolby Vision feature to produce Dolby Vision compatible video output.
4520
4932
  */
4521
4933
  DolbyVision?: DolbyVision;
4522
4934
  /**
4935
+ * @public
4523
4936
  * Enable HDR10+ analysis and metadata injection. Compatible with HEVC only.
4524
4937
  */
4525
4938
  Hdr10Plus?: Hdr10Plus;
4526
4939
  /**
4940
+ * @public
4527
4941
  * Enable the Image inserter feature to include a graphic overlay on your video. Enable or disable this feature for each output individually. This setting is disabled by default.
4528
4942
  */
4529
4943
  ImageInserter?: ImageInserter;
4530
4944
  /**
4945
+ * @public
4531
4946
  * Enable the Noise reducer feature to remove noise from your video output if necessary. Enable or disable this feature for each output individually. This setting is disabled by default. When you enable Noise reducer, you must also select a value for Noise reducer filter. For AVC outputs, when you include Noise reducer, you cannot include the Bandwidth reduction filter.
4532
4947
  */
4533
4948
  NoiseReducer?: NoiseReducer;
4534
4949
  /**
4950
+ * @public
4535
4951
  * If you work with a third party video watermarking partner, use the group of settings that correspond with your watermarking partner to include watermarks in your output.
4536
4952
  */
4537
4953
  PartnerWatermarking?: PartnerWatermarking;
4538
4954
  /**
4955
+ * @public
4539
4956
  * Settings for burning the output timecode and specified prefix into the output.
4540
4957
  */
4541
4958
  TimecodeBurnin?: TimecodeBurnin;
@@ -4546,62 +4963,77 @@ export interface VideoPreprocessor {
4546
4963
  */
4547
4964
  export interface VideoDescription {
4548
4965
  /**
4966
+ * @public
4549
4967
  * This setting only applies to H.264, H.265, and MPEG2 outputs. Use Insert AFD signaling to specify whether the service includes AFD values in the output video data and what those values are. * Choose None to remove all AFD values from this output. * Choose Fixed to ignore input AFD values and instead encode the value specified in the job. * Choose Auto to calculate output AFD values based on the input AFD scaler data.
4550
4968
  */
4551
4969
  AfdSignaling?: AfdSignaling | string;
4552
4970
  /**
4971
+ * @public
4553
4972
  * The anti-alias filter is automatically applied to all outputs. The service no longer accepts the value DISABLED for AntiAlias. If you specify that in your job, the service will ignore the setting.
4554
4973
  */
4555
4974
  AntiAlias?: AntiAlias | string;
4556
4975
  /**
4976
+ * @public
4557
4977
  * Video codec settings contains the group of settings related to video encoding. The settings in this group vary depending on the value that you choose for Video codec. For each codec enum that you choose, define the corresponding settings object. The following lists the codec enum, settings object pairs. * AV1, Av1Settings * AVC_INTRA, AvcIntraSettings * FRAME_CAPTURE, FrameCaptureSettings * H_264, H264Settings * H_265, H265Settings * MPEG2, Mpeg2Settings * PRORES, ProresSettings * VC3, Vc3Settings * VP8, Vp8Settings * VP9, Vp9Settings * XAVC, XavcSettings
4558
4978
  */
4559
4979
  CodecSettings?: VideoCodecSettings;
4560
4980
  /**
4981
+ * @public
4561
4982
  * Choose Insert for this setting to include color metadata in this output. Choose Ignore to exclude color metadata from this output. If you don't specify a value, the service sets this to Insert by default.
4562
4983
  */
4563
4984
  ColorMetadata?: ColorMetadata | string;
4564
4985
  /**
4986
+ * @public
4565
4987
  * Use Cropping selection to specify the video area that the service will include in the output video frame.
4566
4988
  */
4567
4989
  Crop?: Rectangle;
4568
4990
  /**
4991
+ * @public
4569
4992
  * Applies only to 29.97 fps outputs. When this feature is enabled, the service will use drop-frame timecode on outputs. If it is not possible to use drop-frame timecode, the system will fall back to non-drop-frame. This setting is enabled by default when Timecode insertion is enabled.
4570
4993
  */
4571
4994
  DropFrameTimecode?: DropFrameTimecode | string;
4572
4995
  /**
4996
+ * @public
4573
4997
  * Applies only if you set AFD Signaling to Fixed. Use Fixed to specify a four-bit AFD value which the service will write on all frames of this video output.
4574
4998
  */
4575
4999
  FixedAfd?: number;
4576
5000
  /**
5001
+ * @public
4577
5002
  * Use Height to define the video resolution height, in pixels, for this output. To use the same resolution as your input: Leave both Width and Height blank. To evenly scale from your input resolution: Leave Height blank and enter a value for Width. For example, if your input is 1920x1080 and you set Width to 1280, your output will be 1280x720.
4578
5003
  */
4579
5004
  Height?: number;
4580
5005
  /**
5006
+ * @public
4581
5007
  * Use Selection placement to define the video area in your output frame. The area outside of the rectangle that you specify here is black.
4582
5008
  */
4583
5009
  Position?: Rectangle;
4584
5010
  /**
5011
+ * @public
4585
5012
  * Use Respond to AFD to specify how the service changes the video itself in response to AFD values in the input. * Choose Respond to clip the input video frame according to the AFD value, input display aspect ratio, and output display aspect ratio. * Choose Passthrough to include the input AFD values. Do not choose this when AfdSignaling is set to NONE. A preferred implementation of this workflow is to set RespondToAfd to and set AfdSignaling to AUTO. * Choose None to remove all input AFD values from this output.
4586
5013
  */
4587
5014
  RespondToAfd?: RespondToAfd | string;
4588
5015
  /**
5016
+ * @public
4589
5017
  * Specify how the service handles outputs that have a different aspect ratio from the input aspect ratio. Choose Stretch to output to have the service stretch your video image to fit. Keep the setting Default to have the service letterbox your video instead. This setting overrides any value that you specify for the setting Selection placement in this output.
4590
5018
  */
4591
5019
  ScalingBehavior?: ScalingBehavior | string;
4592
5020
  /**
5021
+ * @public
4593
5022
  * Use Sharpness setting to specify the strength of anti-aliasing. This setting changes the width of the anti-alias filter kernel used for scaling. Sharpness only applies if your output resolution is different from your input resolution. 0 is the softest setting, 100 the sharpest, and 50 recommended for most content.
4594
5023
  */
4595
5024
  Sharpness?: number;
4596
5025
  /**
5026
+ * @public
4597
5027
  * Applies only to H.264, H.265, MPEG2, and ProRes outputs. Only enable Timecode insertion when the input frame rate is identical to the output frame rate. To include timecodes in this output, set Timecode insertion to PIC_TIMING_SEI. To leave them out, set it to DISABLED. Default is DISABLED. When the service inserts timecodes in an output, by default, it uses any embedded timecodes from the input. If none are present, the service will set the timecode for the first output frame to zero. To change this default behavior, adjust the settings under Timecode configuration. In the console, these settings are located under Job > Job settings > Timecode configuration. Note - Timecode source under input settings does not affect the timecodes that are inserted in the output. Source under Job settings > Timecode configuration does.
4598
5028
  */
4599
5029
  TimecodeInsertion?: VideoTimecodeInsertion | string;
4600
5030
  /**
5031
+ * @public
4601
5032
  * Find additional transcoding features under Preprocessors. Enable the features at each output individually. These features are disabled by default.
4602
5033
  */
4603
5034
  VideoPreprocessors?: VideoPreprocessor;
4604
5035
  /**
5036
+ * @public
4605
5037
  * Use Width to define the video resolution width, in pixels, for this output. To use the same resolution as your input: Leave both Width and Height blank. To evenly scale from your input resolution: Leave Width blank and enter a value for Height. For example, if your input is 1920x1080 and you set Height to 720, your output will be 1280x720.
4606
5038
  */
4607
5039
  Width?: number;
@@ -4612,34 +5044,42 @@ export interface VideoDescription {
4612
5044
  */
4613
5045
  export interface Output {
4614
5046
  /**
5047
+ * @public
4615
5048
  * Contains groups of audio encoding settings organized by audio codec. Include one instance of per output. Can contain multiple groups of encoding settings.
4616
5049
  */
4617
5050
  AudioDescriptions?: AudioDescription[];
4618
5051
  /**
5052
+ * @public
4619
5053
  * Contains groups of captions settings. For each output that has captions, include one instance of CaptionDescriptions. Can contain multiple groups of captions settings.
4620
5054
  */
4621
5055
  CaptionDescriptions?: CaptionDescription[];
4622
5056
  /**
5057
+ * @public
4623
5058
  * Container specific settings.
4624
5059
  */
4625
5060
  ContainerSettings?: ContainerSettings;
4626
5061
  /**
5062
+ * @public
4627
5063
  * Use Extension to specify the file extension for outputs in File output groups. If you do not specify a value, the service will use default extensions by container type as follows * MPEG-2 transport stream, m2ts * Quicktime, mov * MXF container, mxf * MPEG-4 container, mp4 * WebM container, webm * No Container, the service will use codec extensions (e.g. AAC, H265, H265, AC3)
4628
5064
  */
4629
5065
  Extension?: string;
4630
5066
  /**
5067
+ * @public
4631
5068
  * Use Name modifier to have the service add a string to the end of each output filename. You specify the base filename as part of your destination URI. When you create multiple outputs in the same output group, Name modifier is required. Name modifier also accepts format identifiers. For DASH ISO outputs, if you use the format identifiers $Number$ or $Time$ in one output, you must use them in the same way in all outputs of the output group.
4632
5069
  */
4633
5070
  NameModifier?: string;
4634
5071
  /**
5072
+ * @public
4635
5073
  * Specific settings for this type of output.
4636
5074
  */
4637
5075
  OutputSettings?: OutputSettings;
4638
5076
  /**
5077
+ * @public
4639
5078
  * Use Preset to specify a preset for your transcoding settings. Provide the system or custom preset name. You can specify either Preset or Container settings, but not both.
4640
5079
  */
4641
5080
  Preset?: string;
4642
5081
  /**
5082
+ * @public
4643
5083
  * VideoDescription contains a group of video encoding settings. The specific video settings depend on the video codec that you choose for the property codec. Include one instance of VideoDescription per output.
4644
5084
  */
4645
5085
  VideoDescription?: VideoDescription;
@@ -4650,22 +5090,27 @@ export interface Output {
4650
5090
  */
4651
5091
  export interface OutputGroup {
4652
5092
  /**
5093
+ * @public
4653
5094
  * Use automated encoding to have MediaConvert choose your encoding settings for you, based on characteristics of your input video.
4654
5095
  */
4655
5096
  AutomatedEncodingSettings?: AutomatedEncodingSettings;
4656
5097
  /**
5098
+ * @public
4657
5099
  * Use Custom Group Name to specify a name for the output group. This value is displayed on the console and can make your job settings JSON more human-readable. It does not affect your outputs. Use up to twelve characters that are either letters, numbers, spaces, or underscores.
4658
5100
  */
4659
5101
  CustomName?: string;
4660
5102
  /**
5103
+ * @public
4661
5104
  * Name of the output group
4662
5105
  */
4663
5106
  Name?: string;
4664
5107
  /**
5108
+ * @public
4665
5109
  * Output Group settings, including type
4666
5110
  */
4667
5111
  OutputGroupSettings?: OutputGroupSettings;
4668
5112
  /**
5113
+ * @public
4669
5114
  * This object holds groups of encoding settings, one group of settings per output.
4670
5115
  */
4671
5116
  Outputs?: Output[];
@@ -4689,18 +5134,22 @@ export type TimecodeSource = (typeof TimecodeSource)[keyof typeof TimecodeSource
4689
5134
  */
4690
5135
  export interface TimecodeConfig {
4691
5136
  /**
5137
+ * @public
4692
5138
  * If you use an editing platform that relies on an anchor timecode, use Anchor Timecode to specify a timecode that will match the input video frame to the output video frame. Use 24-hour format with frame number, (HH:MM:SS:FF) or (HH:MM:SS;FF). This setting ignores frame rate conversion. System behavior for Anchor Timecode varies depending on your setting for Source. * If Source is set to Specified Start, the first input frame is the specified value in Start Timecode. Anchor Timecode and Start Timecode are used calculate output timecode. * If Source is set to Start at 0 the first frame is 00:00:00:00. * If Source is set to Embedded, the first frame is the timecode value on the first input frame of the input.
4693
5139
  */
4694
5140
  Anchor?: string;
4695
5141
  /**
5142
+ * @public
4696
5143
  * Use Source to set how timecodes are handled within this job. To make sure that your video, audio, captions, and markers are synchronized and that time-based features, such as image inserter, work correctly, choose the Timecode source option that matches your assets. All timecodes are in a 24-hour format with frame number (HH:MM:SS:FF). * Embedded - Use the timecode that is in the input video. If no embedded timecode is in the source, the service will use Start at 0 instead. * Start at 0 - Set the timecode of the initial frame to 00:00:00:00. * Specified Start - Set the timecode of the initial frame to a value other than zero. You use Start timecode to provide this value.
4697
5144
  */
4698
5145
  Source?: TimecodeSource | string;
4699
5146
  /**
5147
+ * @public
4700
5148
  * Only use when you set Source to Specified start. Use Start timecode to specify the timecode for the initial frame. Use 24-hour format with frame number, (HH:MM:SS:FF) or (HH:MM:SS;FF).
4701
5149
  */
4702
5150
  Start?: string;
4703
5151
  /**
5152
+ * @public
4704
5153
  * Only applies to outputs that support program-date-time stamp. Use Timestamp offset to overwrite the timecode date without affecting the time and frame number. Provide the new date as a string in the format "yyyy-mm-dd". To use Timestamp offset, you must also enable Insert program-date-time in the output settings. For example, if the date part of your timecodes is 2002-1-25 and you want to change it to one year later, set Timestamp offset to 2003-1-25.
4705
5154
  */
4706
5155
  TimestampOffset?: string;
@@ -4711,6 +5160,7 @@ export interface TimecodeConfig {
4711
5160
  */
4712
5161
  export interface TimedMetadataInsertion {
4713
5162
  /**
5163
+ * @public
4714
5164
  * Id3Insertions contains the array of Id3Insertion instances.
4715
5165
  */
4716
5166
  Id3Insertions?: Id3Insertion[];
@@ -4721,50 +5171,62 @@ export interface TimedMetadataInsertion {
4721
5171
  */
4722
5172
  export interface JobSettings {
4723
5173
  /**
5174
+ * @public
4724
5175
  * When specified, this offset (in milliseconds) is added to the input Ad Avail PTS time.
4725
5176
  */
4726
5177
  AdAvailOffset?: number;
4727
5178
  /**
5179
+ * @public
4728
5180
  * Settings for ad avail blanking. Video can be blanked or overlaid with an image, and audio muted during SCTE-35 triggered ad avails.
4729
5181
  */
4730
5182
  AvailBlanking?: AvailBlanking;
4731
5183
  /**
5184
+ * @public
4732
5185
  * Settings for Event Signaling And Messaging (ESAM). If you don't do ad insertion, you can ignore these settings.
4733
5186
  */
4734
5187
  Esam?: EsamSettings;
4735
5188
  /**
5189
+ * @public
4736
5190
  * If your source content has EIA-608 Line 21 Data Services, enable this feature to specify what MediaConvert does with the Extended Data Services (XDS) packets. You can choose to pass through XDS packets, or remove them from the output. For more information about XDS, see EIA-608 Line Data Services, section 9.5.1.5 05h Content Advisory.
4737
5191
  */
4738
5192
  ExtendedDataServices?: ExtendedDataServices;
4739
5193
  /**
5194
+ * @public
4740
5195
  * Use Inputs to define source file used in the transcode job. There can be multiple inputs add in a job. These inputs will be concantenated together to create the output.
4741
5196
  */
4742
5197
  Inputs?: Input[];
4743
5198
  /**
5199
+ * @public
4744
5200
  * Use these settings only when you use Kantar watermarking. Specify the values that MediaConvert uses to generate and place Kantar watermarks in your output audio. These settings apply to every output in your job. In addition to specifying these values, you also need to store your Kantar credentials in AWS Secrets Manager. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/kantar-watermarking.html.
4745
5201
  */
4746
5202
  KantarWatermark?: KantarWatermarkSettings;
4747
5203
  /**
5204
+ * @public
4748
5205
  * Overlay motion graphics on top of your video. The motion graphics that you specify here appear on all outputs in all output groups. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/motion-graphic-overlay.html.
4749
5206
  */
4750
5207
  MotionImageInserter?: MotionImageInserter;
4751
5208
  /**
5209
+ * @public
4752
5210
  * Settings for your Nielsen configuration. If you don't do Nielsen measurement and analytics, ignore these settings. When you enable Nielsen configuration, MediaConvert enables PCM to ID3 tagging for all outputs in the job.
4753
5211
  */
4754
5212
  NielsenConfiguration?: NielsenConfiguration;
4755
5213
  /**
5214
+ * @public
4756
5215
  * Ignore these settings unless you are using Nielsen non-linear watermarking. Specify the values that MediaConvert uses to generate and place Nielsen watermarks in your output audio. In addition to specifying these values, you also need to set up your cloud TIC server. These settings apply to every output in your job. The MediaConvert implementation is currently with the following Nielsen versions: Nielsen Watermark SDK Version 5.2.1 Nielsen NLM Watermark Engine Version 1.2.7 Nielsen Watermark Authenticator [SID_TIC] Version [5.0.0]
4757
5216
  */
4758
5217
  NielsenNonLinearWatermark?: NielsenNonLinearWatermarkSettings;
4759
5218
  /**
5219
+ * @public
4760
5220
  * Contains one group of settings for each set of outputs that share a common package type. All unpackaged files (MPEG-4, MPEG-2 TS, Quicktime, MXF, and no container) are grouped in a single output group as well. Required in is a group of settings that apply to the whole group. This required object depends on the value you set for Type. Type, settings object pairs are as follows. * FILE_GROUP_SETTINGS, FileGroupSettings * HLS_GROUP_SETTINGS, HlsGroupSettings * DASH_ISO_GROUP_SETTINGS, DashIsoGroupSettings * MS_SMOOTH_GROUP_SETTINGS, MsSmoothGroupSettings * CMAF_GROUP_SETTINGS, CmafGroupSettings
4761
5221
  */
4762
5222
  OutputGroups?: OutputGroup[];
4763
5223
  /**
5224
+ * @public
4764
5225
  * These settings control how the service handles timecodes throughout the job. These settings don't affect input clipping.
4765
5226
  */
4766
5227
  TimecodeConfig?: TimecodeConfig;
4767
5228
  /**
5229
+ * @public
4768
5230
  * Insert user-defined custom ID3 metadata at timecodes that you specify. In each output that you want to include this metadata, you must set ID3 metadata to Passthrough.
4769
5231
  */
4770
5232
  TimedMetadataInsertion?: TimedMetadataInsertion;
@@ -4827,14 +5289,17 @@ export type StatusUpdateInterval = (typeof StatusUpdateInterval)[keyof typeof St
4827
5289
  */
4828
5290
  export interface Timing {
4829
5291
  /**
5292
+ * @public
4830
5293
  * The time, in Unix epoch format, that the transcoding job finished
4831
5294
  */
4832
5295
  FinishTime?: Date;
4833
5296
  /**
5297
+ * @public
4834
5298
  * The time, in Unix epoch format, that transcoding for the job began.
4835
5299
  */
4836
5300
  StartTime?: Date;
4837
5301
  /**
5302
+ * @public
4838
5303
  * The time, in Unix epoch format, that you submitted the job.
4839
5304
  */
4840
5305
  SubmitTime?: Date;
@@ -4845,10 +5310,12 @@ export interface Timing {
4845
5310
  */
4846
5311
  export interface WarningGroup {
4847
5312
  /**
5313
+ * @public
4848
5314
  * Warning code that identifies a specific warning in the job. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/warning_codes.html
4849
5315
  */
4850
5316
  Code: number | undefined;
4851
5317
  /**
5318
+ * @public
4852
5319
  * The number of times this warning occurred in the job.
4853
5320
  */
4854
5321
  Count: number | undefined;
@@ -4859,110 +5326,137 @@ export interface WarningGroup {
4859
5326
  */
4860
5327
  export interface Job {
4861
5328
  /**
5329
+ * @public
4862
5330
  * Accelerated transcoding can significantly speed up jobs with long, visually complex content.
4863
5331
  */
4864
5332
  AccelerationSettings?: AccelerationSettings;
4865
5333
  /**
5334
+ * @public
4866
5335
  * Describes whether the current job is running with accelerated transcoding. For jobs that have Acceleration (AccelerationMode) set to DISABLED, AccelerationStatus is always NOT_APPLICABLE. For jobs that have Acceleration (AccelerationMode) set to ENABLED or PREFERRED, AccelerationStatus is one of the other states. AccelerationStatus is IN_PROGRESS initially, while the service determines whether the input files and job settings are compatible with accelerated transcoding. If they are, AcclerationStatus is ACCELERATED. If your input files and job settings aren't compatible with accelerated transcoding, the service either fails your job or runs it without accelerated transcoding, depending on how you set Acceleration (AccelerationMode). When the service runs your job without accelerated transcoding, AccelerationStatus is NOT_ACCELERATED.
4867
5336
  */
4868
5337
  AccelerationStatus?: AccelerationStatus | string;
4869
5338
  /**
5339
+ * @public
4870
5340
  * An identifier for this resource that is unique within all of AWS.
4871
5341
  */
4872
5342
  Arn?: string;
4873
5343
  /**
5344
+ * @public
4874
5345
  * The tag type that AWS Billing and Cost Management will use to sort your AWS Elemental MediaConvert costs on any billing report that you set up.
4875
5346
  */
4876
5347
  BillingTagsSource?: BillingTagsSource | string;
4877
5348
  /**
5349
+ * @public
4878
5350
  * Prevent duplicate jobs from being created and ensure idempotency for your requests. A client request token can be any string that includes up to 64 ASCII characters. If you reuse a client request token within one minute of a successful request, the API returns the job details of the original request instead. For more information see https://docs.aws.amazon.com/mediaconvert/latest/apireference/idempotency.html.
4879
5351
  */
4880
5352
  ClientRequestToken?: string;
4881
5353
  /**
5354
+ * @public
4882
5355
  * The time, in Unix epoch format in seconds, when the job got created.
4883
5356
  */
4884
5357
  CreatedAt?: Date;
4885
5358
  /**
5359
+ * @public
4886
5360
  * A job's phase can be PROBING, TRANSCODING OR UPLOADING
4887
5361
  */
4888
5362
  CurrentPhase?: JobPhase | string;
4889
5363
  /**
5364
+ * @public
4890
5365
  * Error code for the job
4891
5366
  */
4892
5367
  ErrorCode?: number;
4893
5368
  /**
5369
+ * @public
4894
5370
  * Error message of Job
4895
5371
  */
4896
5372
  ErrorMessage?: string;
4897
5373
  /**
5374
+ * @public
4898
5375
  * Optional list of hop destinations.
4899
5376
  */
4900
5377
  HopDestinations?: HopDestination[];
4901
5378
  /**
5379
+ * @public
4902
5380
  * A portion of the job's ARN, unique within your AWS Elemental MediaConvert resources
4903
5381
  */
4904
5382
  Id?: string;
4905
5383
  /**
5384
+ * @public
4906
5385
  * An estimate of how far your job has progressed. This estimate is shown as a percentage of the total time from when your job leaves its queue to when your output files appear in your output Amazon S3 bucket. AWS Elemental MediaConvert provides jobPercentComplete in CloudWatch STATUS_UPDATE events and in the response to GetJob and ListJobs requests. The jobPercentComplete estimate is reliable for the following input containers: Quicktime, Transport Stream, MP4, and MXF. For some jobs, the service can't provide information about job progress. In those cases, jobPercentComplete returns a null value.
4907
5386
  */
4908
5387
  JobPercentComplete?: number;
4909
5388
  /**
5389
+ * @public
4910
5390
  * The job template that the job is created from, if it is created from a job template.
4911
5391
  */
4912
5392
  JobTemplate?: string;
4913
5393
  /**
5394
+ * @public
4914
5395
  * Provides messages from the service about jobs that you have already successfully submitted.
4915
5396
  */
4916
5397
  Messages?: JobMessages;
4917
5398
  /**
5399
+ * @public
4918
5400
  * List of output group details
4919
5401
  */
4920
5402
  OutputGroupDetails?: OutputGroupDetail[];
4921
5403
  /**
5404
+ * @public
4922
5405
  * Relative priority on the job.
4923
5406
  */
4924
5407
  Priority?: number;
4925
5408
  /**
5409
+ * @public
4926
5410
  * When you create a job, you can specify a queue to send it to. If you don't specify, the job will go to the default queue. For more about queues, see the User Guide topic at https://docs.aws.amazon.com/mediaconvert/latest/ug/what-is.html
4927
5411
  */
4928
5412
  Queue?: string;
4929
5413
  /**
5414
+ * @public
4930
5415
  * The job's queue hopping history.
4931
5416
  */
4932
5417
  QueueTransitions?: QueueTransition[];
4933
5418
  /**
5419
+ * @public
4934
5420
  * The number of times that the service automatically attempted to process your job after encountering an error.
4935
5421
  */
4936
5422
  RetryCount?: number;
4937
5423
  /**
5424
+ * @public
4938
5425
  * The IAM role you use for creating this job. For details about permissions, see the User Guide topic at the User Guide at https://docs.aws.amazon.com/mediaconvert/latest/ug/iam-role.html
4939
5426
  */
4940
5427
  Role: string | undefined;
4941
5428
  /**
5429
+ * @public
4942
5430
  * JobSettings contains all the transcode settings for a job.
4943
5431
  */
4944
5432
  Settings: JobSettings | undefined;
4945
5433
  /**
5434
+ * @public
4946
5435
  * Enable this setting when you run a test job to estimate how many reserved transcoding slots (RTS) you need. When this is enabled, MediaConvert runs your job from an on-demand queue with similar performance to what you will see with one RTS in a reserved queue. This setting is disabled by default.
4947
5436
  */
4948
5437
  SimulateReservedQueue?: SimulateReservedQueue | string;
4949
5438
  /**
5439
+ * @public
4950
5440
  * A job's status can be SUBMITTED, PROGRESSING, COMPLETE, CANCELED, or ERROR.
4951
5441
  */
4952
5442
  Status?: JobStatus | string;
4953
5443
  /**
5444
+ * @public
4954
5445
  * Specify how often MediaConvert sends STATUS_UPDATE events to Amazon CloudWatch Events. Set the interval, in seconds, between status updates. MediaConvert sends an update at this interval from the time the service begins processing your job to the time it completes the transcode or encounters an error.
4955
5446
  */
4956
5447
  StatusUpdateInterval?: StatusUpdateInterval | string;
4957
5448
  /**
5449
+ * @public
4958
5450
  * Information about when jobs are submitted, started, and finished is specified in Unix epoch format in seconds.
4959
5451
  */
4960
5452
  Timing?: Timing;
4961
5453
  /**
5454
+ * @public
4962
5455
  * User-defined metadata that you want to associate with an MediaConvert job. You specify metadata in key/value pairs.
4963
5456
  */
4964
5457
  UserMetadata?: Record<string, string>;
4965
5458
  /**
5459
+ * @public
4966
5460
  * Contains any warning messages for the job. Use to help identify potential issues with your input, output, or job. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/warning_codes.html
4967
5461
  */
4968
5462
  Warnings?: WarningGroup[];
@@ -4973,50 +5467,62 @@ export interface Job {
4973
5467
  */
4974
5468
  export interface JobTemplateSettings {
4975
5469
  /**
5470
+ * @public
4976
5471
  * When specified, this offset (in milliseconds) is added to the input Ad Avail PTS time.
4977
5472
  */
4978
5473
  AdAvailOffset?: number;
4979
5474
  /**
5475
+ * @public
4980
5476
  * Settings for ad avail blanking. Video can be blanked or overlaid with an image, and audio muted during SCTE-35 triggered ad avails.
4981
5477
  */
4982
5478
  AvailBlanking?: AvailBlanking;
4983
5479
  /**
5480
+ * @public
4984
5481
  * Settings for Event Signaling And Messaging (ESAM). If you don't do ad insertion, you can ignore these settings.
4985
5482
  */
4986
5483
  Esam?: EsamSettings;
4987
5484
  /**
5485
+ * @public
4988
5486
  * If your source content has EIA-608 Line 21 Data Services, enable this feature to specify what MediaConvert does with the Extended Data Services (XDS) packets. You can choose to pass through XDS packets, or remove them from the output. For more information about XDS, see EIA-608 Line Data Services, section 9.5.1.5 05h Content Advisory.
4989
5487
  */
4990
5488
  ExtendedDataServices?: ExtendedDataServices;
4991
5489
  /**
5490
+ * @public
4992
5491
  * Use Inputs to define the source file used in the transcode job. There can only be one input in a job template. Using the API, you can include multiple inputs when referencing a job template.
4993
5492
  */
4994
5493
  Inputs?: InputTemplate[];
4995
5494
  /**
5495
+ * @public
4996
5496
  * Use these settings only when you use Kantar watermarking. Specify the values that MediaConvert uses to generate and place Kantar watermarks in your output audio. These settings apply to every output in your job. In addition to specifying these values, you also need to store your Kantar credentials in AWS Secrets Manager. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/kantar-watermarking.html.
4997
5497
  */
4998
5498
  KantarWatermark?: KantarWatermarkSettings;
4999
5499
  /**
5500
+ * @public
5000
5501
  * Overlay motion graphics on top of your video. The motion graphics that you specify here appear on all outputs in all output groups. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/motion-graphic-overlay.html.
5001
5502
  */
5002
5503
  MotionImageInserter?: MotionImageInserter;
5003
5504
  /**
5505
+ * @public
5004
5506
  * Settings for your Nielsen configuration. If you don't do Nielsen measurement and analytics, ignore these settings. When you enable Nielsen configuration, MediaConvert enables PCM to ID3 tagging for all outputs in the job.
5005
5507
  */
5006
5508
  NielsenConfiguration?: NielsenConfiguration;
5007
5509
  /**
5510
+ * @public
5008
5511
  * Ignore these settings unless you are using Nielsen non-linear watermarking. Specify the values that MediaConvert uses to generate and place Nielsen watermarks in your output audio. In addition to specifying these values, you also need to set up your cloud TIC server. These settings apply to every output in your job. The MediaConvert implementation is currently with the following Nielsen versions: Nielsen Watermark SDK Version 5.2.1 Nielsen NLM Watermark Engine Version 1.2.7 Nielsen Watermark Authenticator [SID_TIC] Version [5.0.0]
5009
5512
  */
5010
5513
  NielsenNonLinearWatermark?: NielsenNonLinearWatermarkSettings;
5011
5514
  /**
5515
+ * @public
5012
5516
  * Contains one group of settings for each set of outputs that share a common package type. All unpackaged files (MPEG-4, MPEG-2 TS, Quicktime, MXF, and no container) are grouped in a single output group as well. Required in is a group of settings that apply to the whole group. This required object depends on the value you set for Type. Type, settings object pairs are as follows. * FILE_GROUP_SETTINGS, FileGroupSettings * HLS_GROUP_SETTINGS, HlsGroupSettings * DASH_ISO_GROUP_SETTINGS, DashIsoGroupSettings * MS_SMOOTH_GROUP_SETTINGS, MsSmoothGroupSettings * CMAF_GROUP_SETTINGS, CmafGroupSettings
5013
5517
  */
5014
5518
  OutputGroups?: OutputGroup[];
5015
5519
  /**
5520
+ * @public
5016
5521
  * These settings control how the service handles timecodes throughout the job. These settings don't affect input clipping.
5017
5522
  */
5018
5523
  TimecodeConfig?: TimecodeConfig;
5019
5524
  /**
5525
+ * @public
5020
5526
  * Insert user-defined custom ID3 metadata at timecodes that you specify. In each output that you want to include this metadata, you must set ID3 metadata to Passthrough.
5021
5527
  */
5022
5528
  TimedMetadataInsertion?: TimedMetadataInsertion;
@@ -5039,54 +5545,67 @@ export type Type = (typeof Type)[keyof typeof Type];
5039
5545
  */
5040
5546
  export interface JobTemplate {
5041
5547
  /**
5548
+ * @public
5042
5549
  * Accelerated transcoding can significantly speed up jobs with long, visually complex content.
5043
5550
  */
5044
5551
  AccelerationSettings?: AccelerationSettings;
5045
5552
  /**
5553
+ * @public
5046
5554
  * An identifier for this resource that is unique within all of AWS.
5047
5555
  */
5048
5556
  Arn?: string;
5049
5557
  /**
5558
+ * @public
5050
5559
  * An optional category you create to organize your job templates.
5051
5560
  */
5052
5561
  Category?: string;
5053
5562
  /**
5563
+ * @public
5054
5564
  * The timestamp in epoch seconds for Job template creation.
5055
5565
  */
5056
5566
  CreatedAt?: Date;
5057
5567
  /**
5568
+ * @public
5058
5569
  * An optional description you create for each job template.
5059
5570
  */
5060
5571
  Description?: string;
5061
5572
  /**
5573
+ * @public
5062
5574
  * Optional list of hop destinations.
5063
5575
  */
5064
5576
  HopDestinations?: HopDestination[];
5065
5577
  /**
5578
+ * @public
5066
5579
  * The timestamp in epoch seconds when the Job template was last updated.
5067
5580
  */
5068
5581
  LastUpdated?: Date;
5069
5582
  /**
5583
+ * @public
5070
5584
  * A name you create for each job template. Each name must be unique within your account.
5071
5585
  */
5072
5586
  Name: string | undefined;
5073
5587
  /**
5588
+ * @public
5074
5589
  * Relative priority on the job.
5075
5590
  */
5076
5591
  Priority?: number;
5077
5592
  /**
5593
+ * @public
5078
5594
  * Optional. The queue that jobs created from this template are assigned to. If you don't specify this, jobs will go to the default queue.
5079
5595
  */
5080
5596
  Queue?: string;
5081
5597
  /**
5598
+ * @public
5082
5599
  * JobTemplateSettings contains all the transcode settings saved in the template that will be applied to jobs created from it.
5083
5600
  */
5084
5601
  Settings: JobTemplateSettings | undefined;
5085
5602
  /**
5603
+ * @public
5086
5604
  * Specify how often MediaConvert sends STATUS_UPDATE events to Amazon CloudWatch Events. Set the interval, in seconds, between status updates. MediaConvert sends an update at this interval from the time the service begins processing your job to the time it completes the transcode or encounters an error.
5087
5605
  */
5088
5606
  StatusUpdateInterval?: StatusUpdateInterval | string;
5089
5607
  /**
5608
+ * @public
5090
5609
  * A job template can be of two types: system or custom. System or built-in job templates can't be modified or deleted by the user.
5091
5610
  */
5092
5611
  Type?: Type | string;
@@ -5097,18 +5616,22 @@ export interface JobTemplate {
5097
5616
  */
5098
5617
  export interface PresetSettings {
5099
5618
  /**
5619
+ * @public
5100
5620
  * Contains groups of audio encoding settings organized by audio codec. Include one instance of per output. Can contain multiple groups of encoding settings.
5101
5621
  */
5102
5622
  AudioDescriptions?: AudioDescription[];
5103
5623
  /**
5624
+ * @public
5104
5625
  * This object holds groups of settings related to captions for one output. For each output that has captions, include one instance of CaptionDescriptions.
5105
5626
  */
5106
5627
  CaptionDescriptions?: CaptionDescriptionPreset[];
5107
5628
  /**
5629
+ * @public
5108
5630
  * Container specific settings.
5109
5631
  */
5110
5632
  ContainerSettings?: ContainerSettings;
5111
5633
  /**
5634
+ * @public
5112
5635
  * VideoDescription contains a group of video encoding settings. The specific video settings depend on the video codec that you choose for the property codec. Include one instance of VideoDescription per output.
5113
5636
  */
5114
5637
  VideoDescription?: VideoDescription;
@@ -5119,34 +5642,42 @@ export interface PresetSettings {
5119
5642
  */
5120
5643
  export interface Preset {
5121
5644
  /**
5645
+ * @public
5122
5646
  * An identifier for this resource that is unique within all of AWS.
5123
5647
  */
5124
5648
  Arn?: string;
5125
5649
  /**
5650
+ * @public
5126
5651
  * An optional category you create to organize your presets.
5127
5652
  */
5128
5653
  Category?: string;
5129
5654
  /**
5655
+ * @public
5130
5656
  * The timestamp in epoch seconds for preset creation.
5131
5657
  */
5132
5658
  CreatedAt?: Date;
5133
5659
  /**
5660
+ * @public
5134
5661
  * An optional description you create for each preset.
5135
5662
  */
5136
5663
  Description?: string;
5137
5664
  /**
5665
+ * @public
5138
5666
  * The timestamp in epoch seconds when the preset was last updated.
5139
5667
  */
5140
5668
  LastUpdated?: Date;
5141
5669
  /**
5670
+ * @public
5142
5671
  * A name you create for each preset. Each name must be unique within your account.
5143
5672
  */
5144
5673
  Name: string | undefined;
5145
5674
  /**
5675
+ * @public
5146
5676
  * Settings for preset
5147
5677
  */
5148
5678
  Settings: PresetSettings | undefined;
5149
5679
  /**
5680
+ * @public
5150
5681
  * A preset can be of two types: system or custom. System or built-in preset can't be modified or deleted by the user.
5151
5682
  */
5152
5683
  Type?: Type | string;
@@ -5204,26 +5735,32 @@ export type ReservationPlanStatus = (typeof ReservationPlanStatus)[keyof typeof
5204
5735
  */
5205
5736
  export interface ReservationPlan {
5206
5737
  /**
5738
+ * @public
5207
5739
  * The length of the term of your reserved queue pricing plan commitment.
5208
5740
  */
5209
5741
  Commitment?: Commitment | string;
5210
5742
  /**
5743
+ * @public
5211
5744
  * The timestamp in epoch seconds for when the current pricing plan term for this reserved queue expires.
5212
5745
  */
5213
5746
  ExpiresAt?: Date;
5214
5747
  /**
5748
+ * @public
5215
5749
  * The timestamp in epoch seconds for when you set up the current pricing plan for this reserved queue.
5216
5750
  */
5217
5751
  PurchasedAt?: Date;
5218
5752
  /**
5753
+ * @public
5219
5754
  * Specifies whether the term of your reserved queue pricing plan is automatically extended (AUTO_RENEW) or expires (EXPIRE) at the end of the term.
5220
5755
  */
5221
5756
  RenewalType?: RenewalType | string;
5222
5757
  /**
5758
+ * @public
5223
5759
  * Specifies the number of reserved transcode slots (RTS) for this queue. The number of RTS determines how many jobs the queue can process in parallel; each RTS can process one job at a time. When you increase this number, you extend your existing commitment with a new 12-month commitment for a larger number of RTS. The new commitment begins when you purchase the additional capacity. You can't decrease the number of RTS in your reserved queue.
5224
5760
  */
5225
5761
  ReservedSlots?: number;
5226
5762
  /**
5763
+ * @public
5227
5764
  * Specifies whether the pricing plan for your reserved queue is ACTIVE or EXPIRED.
5228
5765
  */
5229
5766
  Status?: ReservationPlanStatus | string;
@@ -5246,46 +5783,57 @@ export type QueueStatus = (typeof QueueStatus)[keyof typeof QueueStatus];
5246
5783
  */
5247
5784
  export interface Queue {
5248
5785
  /**
5786
+ * @public
5249
5787
  * An identifier for this resource that is unique within all of AWS.
5250
5788
  */
5251
5789
  Arn?: string;
5252
5790
  /**
5791
+ * @public
5253
5792
  * The timestamp in epoch seconds for when you created the queue.
5254
5793
  */
5255
5794
  CreatedAt?: Date;
5256
5795
  /**
5796
+ * @public
5257
5797
  * An optional description that you create for each queue.
5258
5798
  */
5259
5799
  Description?: string;
5260
5800
  /**
5801
+ * @public
5261
5802
  * The timestamp in epoch seconds for when you most recently updated the queue.
5262
5803
  */
5263
5804
  LastUpdated?: Date;
5264
5805
  /**
5806
+ * @public
5265
5807
  * A name that you create for each queue. Each name must be unique within your account.
5266
5808
  */
5267
5809
  Name: string | undefined;
5268
5810
  /**
5811
+ * @public
5269
5812
  * Specifies whether the pricing plan for the queue is on-demand or reserved. For on-demand, you pay per minute, billed in increments of .01 minute. For reserved, you pay for the transcoding capacity of the entire queue, regardless of how much or how little you use it. Reserved pricing requires a 12-month commitment.
5270
5813
  */
5271
5814
  PricingPlan?: PricingPlan | string;
5272
5815
  /**
5816
+ * @public
5273
5817
  * The estimated number of jobs with a PROGRESSING status.
5274
5818
  */
5275
5819
  ProgressingJobsCount?: number;
5276
5820
  /**
5821
+ * @public
5277
5822
  * Details about the pricing plan for your reserved queue. Required for reserved queues and not applicable to on-demand queues.
5278
5823
  */
5279
5824
  ReservationPlan?: ReservationPlan;
5280
5825
  /**
5826
+ * @public
5281
5827
  * Queues can be ACTIVE or PAUSED. If you pause a queue, the service won't begin processing jobs in that queue. Jobs that are running when you pause the queue continue to run until they finish or result in an error.
5282
5828
  */
5283
5829
  Status?: QueueStatus | string;
5284
5830
  /**
5831
+ * @public
5285
5832
  * The estimated number of jobs with a SUBMITTED status.
5286
5833
  */
5287
5834
  SubmittedJobsCount?: number;
5288
5835
  /**
5836
+ * @public
5289
5837
  * Specifies whether this on-demand queue is system or custom. System queues are built in. You can't modify or delete system queues. You can create and modify custom queues.
5290
5838
  */
5291
5839
  Type?: Type | string;
@@ -5295,6 +5843,7 @@ export interface Queue {
5295
5843
  */
5296
5844
  export interface AssociateCertificateRequest {
5297
5845
  /**
5846
+ * @public
5298
5847
  * The ARN of the ACM certificate that you want to associate with your MediaConvert resource.
5299
5848
  */
5300
5849
  Arn: string | undefined;
@@ -5387,6 +5936,7 @@ export declare class TooManyRequestsException extends __BaseException {
5387
5936
  */
5388
5937
  export interface CancelJobRequest {
5389
5938
  /**
5939
+ * @public
5390
5940
  * The Job ID of the job to be cancelled.
5391
5941
  */
5392
5942
  Id: string | undefined;
@@ -5401,54 +5951,67 @@ export interface CancelJobResponse {
5401
5951
  */
5402
5952
  export interface CreateJobRequest {
5403
5953
  /**
5954
+ * @public
5404
5955
  * Optional. Accelerated transcoding can significantly speed up jobs with long, visually complex content. Outputs that use this feature incur pro-tier pricing. For information about feature limitations, see the AWS Elemental MediaConvert User Guide.
5405
5956
  */
5406
5957
  AccelerationSettings?: AccelerationSettings;
5407
5958
  /**
5959
+ * @public
5408
5960
  * Optional. Choose a tag type that AWS Billing and Cost Management will use to sort your AWS Elemental MediaConvert costs on any billing report that you set up. Any transcoding outputs that don't have an associated tag will appear in your billing report unsorted. If you don't choose a valid value for this field, your job outputs will appear on the billing report unsorted.
5409
5961
  */
5410
5962
  BillingTagsSource?: BillingTagsSource | string;
5411
5963
  /**
5964
+ * @public
5412
5965
  * Prevent duplicate jobs from being created and ensure idempotency for your requests. A client request token can be any string that includes up to 64 ASCII characters. If you reuse a client request token within one minute of a successful request, the API returns the job details of the original request instead. For more information see https://docs.aws.amazon.com/mediaconvert/latest/apireference/idempotency.html.
5413
5966
  */
5414
5967
  ClientRequestToken?: string;
5415
5968
  /**
5969
+ * @public
5416
5970
  * Optional. Use queue hopping to avoid overly long waits in the backlog of the queue that you submit your job to. Specify an alternate queue and the maximum time that your job will wait in the initial queue before hopping. For more information about this feature, see the AWS Elemental MediaConvert User Guide.
5417
5971
  */
5418
5972
  HopDestinations?: HopDestination[];
5419
5973
  /**
5974
+ * @public
5420
5975
  * Optional. When you create a job, you can either specify a job template or specify the transcoding settings individually.
5421
5976
  */
5422
5977
  JobTemplate?: string;
5423
5978
  /**
5979
+ * @public
5424
5980
  * Optional. Specify the relative priority for this job. In any given queue, the service begins processing the job with the highest value first. When more than one job has the same priority, the service begins processing the job that you submitted first. If you don't specify a priority, the service uses the default value 0.
5425
5981
  */
5426
5982
  Priority?: number;
5427
5983
  /**
5984
+ * @public
5428
5985
  * Optional. When you create a job, you can specify a queue to send it to. If you don't specify, the job will go to the default queue. For more about queues, see the User Guide topic at https://docs.aws.amazon.com/mediaconvert/latest/ug/what-is.html.
5429
5986
  */
5430
5987
  Queue?: string;
5431
5988
  /**
5989
+ * @public
5432
5990
  * Required. The IAM role you use for creating this job. For details about permissions, see the User Guide topic at the User Guide at https://docs.aws.amazon.com/mediaconvert/latest/ug/iam-role.html.
5433
5991
  */
5434
5992
  Role: string | undefined;
5435
5993
  /**
5994
+ * @public
5436
5995
  * JobSettings contains all the transcode settings for a job.
5437
5996
  */
5438
5997
  Settings: JobSettings | undefined;
5439
5998
  /**
5999
+ * @public
5440
6000
  * Optional. Enable this setting when you run a test job to estimate how many reserved transcoding slots (RTS) you need. When this is enabled, MediaConvert runs your job from an on-demand queue with similar performance to what you will see with one RTS in a reserved queue. This setting is disabled by default.
5441
6001
  */
5442
6002
  SimulateReservedQueue?: SimulateReservedQueue | string;
5443
6003
  /**
6004
+ * @public
5444
6005
  * Optional. Specify how often MediaConvert sends STATUS_UPDATE events to Amazon CloudWatch Events. Set the interval, in seconds, between status updates. MediaConvert sends an update at this interval from the time the service begins processing your job to the time it completes the transcode or encounters an error.
5445
6006
  */
5446
6007
  StatusUpdateInterval?: StatusUpdateInterval | string;
5447
6008
  /**
6009
+ * @public
5448
6010
  * Optional. The tags that you want to add to the resource. You can tag resources with a key-value pair or with only a key. Use standard AWS tags on your job for automatic integration with AWS services and for custom integrations and workflows.
5449
6011
  */
5450
6012
  Tags?: Record<string, string>;
5451
6013
  /**
6014
+ * @public
5452
6015
  * Optional. User-defined metadata that you want to associate with an MediaConvert job. You specify metadata in key/value pairs. Use only for existing integrations or workflows that rely on job metadata tags. Otherwise, we recommend that you use standard AWS tags.
5453
6016
  */
5454
6017
  UserMetadata?: Record<string, string>;
@@ -5458,6 +6021,7 @@ export interface CreateJobRequest {
5458
6021
  */
5459
6022
  export interface CreateJobResponse {
5460
6023
  /**
6024
+ * @public
5461
6025
  * Each job converts an input file into an output file or files. For more information, see the User Guide at https://docs.aws.amazon.com/mediaconvert/latest/ug/what-is.html
5462
6026
  */
5463
6027
  Job?: Job;
@@ -5467,42 +6031,52 @@ export interface CreateJobResponse {
5467
6031
  */
5468
6032
  export interface CreateJobTemplateRequest {
5469
6033
  /**
6034
+ * @public
5470
6035
  * Accelerated transcoding can significantly speed up jobs with long, visually complex content. Outputs that use this feature incur pro-tier pricing. For information about feature limitations, see the AWS Elemental MediaConvert User Guide.
5471
6036
  */
5472
6037
  AccelerationSettings?: AccelerationSettings;
5473
6038
  /**
6039
+ * @public
5474
6040
  * Optional. A category for the job template you are creating
5475
6041
  */
5476
6042
  Category?: string;
5477
6043
  /**
6044
+ * @public
5478
6045
  * Optional. A description of the job template you are creating.
5479
6046
  */
5480
6047
  Description?: string;
5481
6048
  /**
6049
+ * @public
5482
6050
  * Optional. Use queue hopping to avoid overly long waits in the backlog of the queue that you submit your job to. Specify an alternate queue and the maximum time that your job will wait in the initial queue before hopping. For more information about this feature, see the AWS Elemental MediaConvert User Guide.
5483
6051
  */
5484
6052
  HopDestinations?: HopDestination[];
5485
6053
  /**
6054
+ * @public
5486
6055
  * The name of the job template you are creating.
5487
6056
  */
5488
6057
  Name: string | undefined;
5489
6058
  /**
6059
+ * @public
5490
6060
  * Specify the relative priority for this job. In any given queue, the service begins processing the job with the highest value first. When more than one job has the same priority, the service begins processing the job that you submitted first. If you don't specify a priority, the service uses the default value 0.
5491
6061
  */
5492
6062
  Priority?: number;
5493
6063
  /**
6064
+ * @public
5494
6065
  * Optional. The queue that jobs created from this template are assigned to. If you don't specify this, jobs will go to the default queue.
5495
6066
  */
5496
6067
  Queue?: string;
5497
6068
  /**
6069
+ * @public
5498
6070
  * JobTemplateSettings contains all the transcode settings saved in the template that will be applied to jobs created from it.
5499
6071
  */
5500
6072
  Settings: JobTemplateSettings | undefined;
5501
6073
  /**
6074
+ * @public
5502
6075
  * Specify how often MediaConvert sends STATUS_UPDATE events to Amazon CloudWatch Events. Set the interval, in seconds, between status updates. MediaConvert sends an update at this interval from the time the service begins processing your job to the time it completes the transcode or encounters an error.
5503
6076
  */
5504
6077
  StatusUpdateInterval?: StatusUpdateInterval | string;
5505
6078
  /**
6079
+ * @public
5506
6080
  * The tags that you want to add to the resource. You can tag resources with a key-value pair or with only a key.
5507
6081
  */
5508
6082
  Tags?: Record<string, string>;
@@ -5512,6 +6086,7 @@ export interface CreateJobTemplateRequest {
5512
6086
  */
5513
6087
  export interface CreateJobTemplateResponse {
5514
6088
  /**
6089
+ * @public
5515
6090
  * A job template is a pre-made set of encoding instructions that you can use to quickly create a job.
5516
6091
  */
5517
6092
  JobTemplate?: JobTemplate;
@@ -5521,22 +6096,27 @@ export interface CreateJobTemplateResponse {
5521
6096
  */
5522
6097
  export interface CreatePresetRequest {
5523
6098
  /**
6099
+ * @public
5524
6100
  * Optional. A category for the preset you are creating.
5525
6101
  */
5526
6102
  Category?: string;
5527
6103
  /**
6104
+ * @public
5528
6105
  * Optional. A description of the preset you are creating.
5529
6106
  */
5530
6107
  Description?: string;
5531
6108
  /**
6109
+ * @public
5532
6110
  * The name of the preset you are creating.
5533
6111
  */
5534
6112
  Name: string | undefined;
5535
6113
  /**
6114
+ * @public
5536
6115
  * Settings for preset
5537
6116
  */
5538
6117
  Settings: PresetSettings | undefined;
5539
6118
  /**
6119
+ * @public
5540
6120
  * The tags that you want to add to the resource. You can tag resources with a key-value pair or with only a key.
5541
6121
  */
5542
6122
  Tags?: Record<string, string>;
@@ -5546,6 +6126,7 @@ export interface CreatePresetRequest {
5546
6126
  */
5547
6127
  export interface CreatePresetResponse {
5548
6128
  /**
6129
+ * @public
5549
6130
  * A preset is a collection of preconfigured media conversion settings that you want MediaConvert to apply to the output during the conversion process.
5550
6131
  */
5551
6132
  Preset?: Preset;
@@ -5556,14 +6137,17 @@ export interface CreatePresetResponse {
5556
6137
  */
5557
6138
  export interface ReservationPlanSettings {
5558
6139
  /**
6140
+ * @public
5559
6141
  * The length of the term of your reserved queue pricing plan commitment.
5560
6142
  */
5561
6143
  Commitment: Commitment | string | undefined;
5562
6144
  /**
6145
+ * @public
5563
6146
  * Specifies whether the term of your reserved queue pricing plan is automatically extended (AUTO_RENEW) or expires (EXPIRE) at the end of the term. When your term is auto renewed, you extend your commitment by 12 months from the auto renew date. You can cancel this commitment.
5564
6147
  */
5565
6148
  RenewalType: RenewalType | string | undefined;
5566
6149
  /**
6150
+ * @public
5567
6151
  * Specifies the number of reserved transcode slots (RTS) for this queue. The number of RTS determines how many jobs the queue can process in parallel; each RTS can process one job at a time. You can't decrease the number of RTS in your reserved queue. You can increase the number of RTS by extending your existing commitment with a new 12-month commitment for the larger number. The new commitment begins when you purchase the additional capacity. You can't cancel your commitment or revert to your original commitment after you increase the capacity.
5568
6152
  */
5569
6153
  ReservedSlots: number | undefined;
@@ -5573,26 +6157,32 @@ export interface ReservationPlanSettings {
5573
6157
  */
5574
6158
  export interface CreateQueueRequest {
5575
6159
  /**
6160
+ * @public
5576
6161
  * Optional. A description of the queue that you are creating.
5577
6162
  */
5578
6163
  Description?: string;
5579
6164
  /**
6165
+ * @public
5580
6166
  * The name of the queue that you are creating.
5581
6167
  */
5582
6168
  Name: string | undefined;
5583
6169
  /**
6170
+ * @public
5584
6171
  * Specifies whether the pricing plan for the queue is on-demand or reserved. For on-demand, you pay per minute, billed in increments of .01 minute. For reserved, you pay for the transcoding capacity of the entire queue, regardless of how much or how little you use it. Reserved pricing requires a 12-month commitment. When you use the API to create a queue, the default is on-demand.
5585
6172
  */
5586
6173
  PricingPlan?: PricingPlan | string;
5587
6174
  /**
6175
+ * @public
5588
6176
  * Details about the pricing plan for your reserved queue. Required for reserved queues and not applicable to on-demand queues.
5589
6177
  */
5590
6178
  ReservationPlanSettings?: ReservationPlanSettings;
5591
6179
  /**
6180
+ * @public
5592
6181
  * Initial state of the queue. If you create a paused queue, then jobs in that queue won't begin.
5593
6182
  */
5594
6183
  Status?: QueueStatus | string;
5595
6184
  /**
6185
+ * @public
5596
6186
  * The tags that you want to add to the resource. You can tag resources with a key-value pair or with only a key.
5597
6187
  */
5598
6188
  Tags?: Record<string, string>;
@@ -5602,6 +6192,7 @@ export interface CreateQueueRequest {
5602
6192
  */
5603
6193
  export interface CreateQueueResponse {
5604
6194
  /**
6195
+ * @public
5605
6196
  * You can use queues to manage the resources that are available to your AWS account for running multiple transcoding jobs at the same time. If you don't specify a queue, the service sends all jobs through the default queue. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/working-with-queues.html.
5606
6197
  */
5607
6198
  Queue?: Queue;
@@ -5611,6 +6202,7 @@ export interface CreateQueueResponse {
5611
6202
  */
5612
6203
  export interface DeleteJobTemplateRequest {
5613
6204
  /**
6205
+ * @public
5614
6206
  * The name of the job template to be deleted.
5615
6207
  */
5616
6208
  Name: string | undefined;
@@ -5635,6 +6227,7 @@ export interface DeletePolicyResponse {
5635
6227
  */
5636
6228
  export interface DeletePresetRequest {
5637
6229
  /**
6230
+ * @public
5638
6231
  * The name of the preset to be deleted.
5639
6232
  */
5640
6233
  Name: string | undefined;
@@ -5649,6 +6242,7 @@ export interface DeletePresetResponse {
5649
6242
  */
5650
6243
  export interface DeleteQueueRequest {
5651
6244
  /**
6245
+ * @public
5652
6246
  * The name of the queue that you want to delete.
5653
6247
  */
5654
6248
  Name: string | undefined;