@aws-sdk/client-mediaconvert 3.379.1 → 3.385.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -16,14 +16,17 @@ export type RequiredFlag = (typeof RequiredFlag)[keyof typeof RequiredFlag];
16
16
  */
17
17
  export interface AllowedRenditionSize {
18
18
  /**
19
+ * @public
19
20
  * Use Height to define the video resolution height, in pixels, for this rule.
20
21
  */
21
22
  Height?: number;
22
23
  /**
24
+ * @public
23
25
  * Set to ENABLED to force a rendition to be included.
24
26
  */
25
27
  Required?: RequiredFlag | string;
26
28
  /**
29
+ * @public
27
30
  * Use Width to define the video resolution width, in pixels, for this rule.
28
31
  */
29
32
  Width?: number;
@@ -59,6 +62,7 @@ export type AudioChannelTag = (typeof AudioChannelTag)[keyof typeof AudioChannel
59
62
  */
60
63
  export interface AudioChannelTaggingSettings {
61
64
  /**
65
+ * @public
62
66
  * You can add a tag for this mono-channel audio track to mimic its placement in a multi-channel layout. For example, if this track is the left surround channel, choose Left surround (LS).
63
67
  */
64
68
  ChannelTag?: AudioChannelTag | string;
@@ -119,30 +123,37 @@ export type AudioNormalizationPeakCalculation = (typeof AudioNormalizationPeakCa
119
123
  */
120
124
  export interface AudioNormalizationSettings {
121
125
  /**
126
+ * @public
122
127
  * Choose one of the following audio normalization algorithms: ITU-R BS.1770-1: Ungated loudness. A measurement of ungated average loudness for an entire piece of content, suitable for measurement of short-form content under ATSC recommendation A/85. Supports up to 5.1 audio channels. ITU-R BS.1770-2: Gated loudness. A measurement of gated average loudness compliant with the requirements of EBU-R128. Supports up to 5.1 audio channels. ITU-R BS.1770-3: Modified peak. The same loudness measurement algorithm as 1770-2, with an updated true peak measurement. ITU-R BS.1770-4: Higher channel count. Allows for more audio channels than the other algorithms, including configurations such as 7.1.
123
128
  */
124
129
  Algorithm?: AudioNormalizationAlgorithm | string;
125
130
  /**
131
+ * @public
126
132
  * When enabled the output audio is corrected using the chosen algorithm. If disabled, the audio will be measured but not adjusted.
127
133
  */
128
134
  AlgorithmControl?: AudioNormalizationAlgorithmControl | string;
129
135
  /**
136
+ * @public
130
137
  * Content measuring above this level will be corrected to the target level. Content measuring below this level will not be corrected.
131
138
  */
132
139
  CorrectionGateLevel?: number;
133
140
  /**
141
+ * @public
134
142
  * If set to LOG, log each output's audio track loudness to a CSV file.
135
143
  */
136
144
  LoudnessLogging?: AudioNormalizationLoudnessLogging | string;
137
145
  /**
146
+ * @public
138
147
  * If set to TRUE_PEAK, calculate and log the TruePeak for each output's audio track loudness.
139
148
  */
140
149
  PeakCalculation?: AudioNormalizationPeakCalculation | string;
141
150
  /**
151
+ * @public
142
152
  * When you use Audio normalization, optionally use this setting to specify a target loudness. If you don't specify a value here, the encoder chooses a value for you, based on the algorithm that you choose for Algorithm. If you choose algorithm 1770-1, the encoder will choose -24 LKFS; otherwise, the encoder will choose -23 LKFS.
143
153
  */
144
154
  TargetLkfs?: number;
145
155
  /**
156
+ * @public
146
157
  * Specify the True-peak limiter threshold in decibels relative to full scale (dBFS). The peak inter-audio sample loudness in your output will be limited to the value that you specify, without affecting the overall target LKFS. Enter a value from 0 to -8. Leave blank to use the default value 0.
147
158
  */
148
159
  TruePeakLimiterThreshold?: number;
@@ -255,38 +266,47 @@ export type AacVbrQuality = (typeof AacVbrQuality)[keyof typeof AacVbrQuality];
255
266
  */
256
267
  export interface AacSettings {
257
268
  /**
269
+ * @public
258
270
  * Choose BROADCASTER_MIXED_AD when the input contains pre-mixed main audio + audio description (AD) as a stereo pair. The value for AudioType will be set to 3, which signals to downstream systems that this stream contains "broadcaster mixed AD". Note that the input received by the encoder must contain pre-mixed audio; the encoder does not perform the mixing. When you choose BROADCASTER_MIXED_AD, the encoder ignores any values you provide in AudioType and FollowInputAudioType. Choose NORMAL when the input does not contain pre-mixed audio + audio description (AD). In this case, the encoder will use any values you provide for AudioType and FollowInputAudioType.
259
271
  */
260
272
  AudioDescriptionBroadcasterMix?: AacAudioDescriptionBroadcasterMix | string;
261
273
  /**
274
+ * @public
262
275
  * Specify the average bitrate in bits per second. The set of valid values for this setting is: 6000, 8000, 10000, 12000, 14000, 16000, 20000, 24000, 28000, 32000, 40000, 48000, 56000, 64000, 80000, 96000, 112000, 128000, 160000, 192000, 224000, 256000, 288000, 320000, 384000, 448000, 512000, 576000, 640000, 768000, 896000, 1024000. The value you set is also constrained by the values that you choose for Profile, Bitrate control mode, and Sample rate. Default values depend on Bitrate control mode and Profile.
263
276
  */
264
277
  Bitrate?: number;
265
278
  /**
279
+ * @public
266
280
  * AAC Profile.
267
281
  */
268
282
  CodecProfile?: AacCodecProfile | string;
269
283
  /**
284
+ * @public
270
285
  * The Coding mode that you specify determines the number of audio channels and the audio channel layout metadata in your AAC output. Valid coding modes depend on the Rate control mode and Profile that you select. The following list shows the number of audio channels and channel layout for each coding mode. * 1.0 Audio Description (Receiver Mix): One channel, C. Includes audio description data from your stereo input. For more information see ETSI TS 101 154 Annex E. * 1.0 Mono: One channel, C. * 2.0 Stereo: Two channels, L, R. * 5.1 Surround: Six channels, C, L, R, Ls, Rs, LFE.
271
286
  */
272
287
  CodingMode?: AacCodingMode | string;
273
288
  /**
289
+ * @public
274
290
  * Rate Control Mode.
275
291
  */
276
292
  RateControlMode?: AacRateControlMode | string;
277
293
  /**
294
+ * @public
278
295
  * Enables LATM/LOAS AAC output. Note that if you use LATM/LOAS AAC in an output, you must choose "No container" for the output container.
279
296
  */
280
297
  RawFormat?: AacRawFormat | string;
281
298
  /**
299
+ * @public
282
300
  * Specify the Sample rate in Hz. Valid sample rates depend on the Profile and Coding mode that you select. The following list shows valid sample rates for each Profile and Coding mode. * LC Profile, Coding mode 1.0, 2.0, and Receiver Mix: 8000, 12000, 16000, 22050, 24000, 32000, 44100, 48000, 88200, 96000. * LC Profile, Coding mode 5.1: 32000, 44100, 48000, 96000. * HEV1 Profile, Coding mode 1.0 and Receiver Mix: 22050, 24000, 32000, 44100, 48000. * HEV1 Profile, Coding mode 2.0 and 5.1: 32000, 44100, 48000, 96000. * HEV2 Profile, Coding mode 2.0: 22050, 24000, 32000, 44100, 48000.
283
301
  */
284
302
  SampleRate?: number;
285
303
  /**
304
+ * @public
286
305
  * Use MPEG-2 AAC instead of MPEG-4 AAC audio for raw or MPEG-2 Transport Stream containers.
287
306
  */
288
307
  Specification?: AacSpecification | string;
289
308
  /**
309
+ * @public
290
310
  * VBR Quality Level - Only used if rate_control_mode is VBR.
291
311
  */
292
312
  VbrQuality?: AacVbrQuality | string;
@@ -397,42 +417,52 @@ export type Ac3MetadataControl = (typeof Ac3MetadataControl)[keyof typeof Ac3Met
397
417
  */
398
418
  export interface Ac3Settings {
399
419
  /**
420
+ * @public
400
421
  * Specify the average bitrate in bits per second. The bitrate that you specify must be a multiple of 8000 within the allowed minimum and maximum values. Leave blank to use the default bitrate for the coding mode you select according ETSI TS 102 366. Valid bitrates for coding mode 1/0: Default: 96000. Minimum: 64000. Maximum: 128000. Valid bitrates for coding mode 1/1: Default: 192000. Minimum: 128000. Maximum: 384000. Valid bitrates for coding mode 2/0: Default: 192000. Minimum: 128000. Maximum: 384000. Valid bitrates for coding mode 3/2 with FLE: Default: 384000. Minimum: 384000. Maximum: 640000.
401
422
  */
402
423
  Bitrate?: number;
403
424
  /**
425
+ * @public
404
426
  * Specify the bitstream mode for the AC-3 stream that the encoder emits. For more information about the AC3 bitstream mode, see ATSC A/52-2012 (Annex E).
405
427
  */
406
428
  BitstreamMode?: Ac3BitstreamMode | string;
407
429
  /**
430
+ * @public
408
431
  * Dolby Digital coding mode. Determines number of channels.
409
432
  */
410
433
  CodingMode?: Ac3CodingMode | string;
411
434
  /**
435
+ * @public
412
436
  * Sets the dialnorm for the output. If blank and input audio is Dolby Digital, dialnorm will be passed through.
413
437
  */
414
438
  Dialnorm?: number;
415
439
  /**
440
+ * @public
416
441
  * Choose the Dolby Digital dynamic range control (DRC) profile that MediaConvert uses when encoding the metadata in the Dolby Digital stream for the line operating mode. Related setting: When you use this setting, MediaConvert ignores any value you provide for Dynamic range compression profile. For information about the Dolby Digital DRC operating modes and profiles, see the Dynamic Range Control chapter of the Dolby Metadata Guide at https://developer.dolby.com/globalassets/professional/documents/dolby-metadata-guide.pdf.
417
442
  */
418
443
  DynamicRangeCompressionLine?: Ac3DynamicRangeCompressionLine | string;
419
444
  /**
445
+ * @public
420
446
  * When you want to add Dolby dynamic range compression (DRC) signaling to your output stream, we recommend that you use the mode-specific settings instead of Dynamic range compression profile. The mode-specific settings are Dynamic range compression profile, line mode and Dynamic range compression profile, RF mode. Note that when you specify values for all three settings, MediaConvert ignores the value of this setting in favor of the mode-specific settings. If you do use this setting instead of the mode-specific settings, choose None to leave out DRC signaling. Keep the default Film standard to set the profile to Dolby's film standard profile for all operating modes.
421
447
  */
422
448
  DynamicRangeCompressionProfile?: Ac3DynamicRangeCompressionProfile | string;
423
449
  /**
450
+ * @public
424
451
  * Choose the Dolby Digital dynamic range control (DRC) profile that MediaConvert uses when encoding the metadata in the Dolby Digital stream for the RF operating mode. Related setting: When you use this setting, MediaConvert ignores any value you provide for Dynamic range compression profile. For information about the Dolby Digital DRC operating modes and profiles, see the Dynamic Range Control chapter of the Dolby Metadata Guide at https://developer.dolby.com/globalassets/professional/documents/dolby-metadata-guide.pdf.
425
452
  */
426
453
  DynamicRangeCompressionRf?: Ac3DynamicRangeCompressionRf | string;
427
454
  /**
455
+ * @public
428
456
  * Applies a 120Hz lowpass filter to the LFE channel prior to encoding. Only valid with 3_2_LFE coding mode.
429
457
  */
430
458
  LfeFilter?: Ac3LfeFilter | string;
431
459
  /**
460
+ * @public
432
461
  * When set to FOLLOW_INPUT, encoder metadata will be sourced from the DD, DD+, or DolbyE decoder that supplied this audio data. If audio was not supplied from one of these streams, then the static metadata settings will be used.
433
462
  */
434
463
  MetadataControl?: Ac3MetadataControl | string;
435
464
  /**
465
+ * @public
436
466
  * This value is always 48000. It represents the sample rate in Hz.
437
467
  */
438
468
  SampleRate?: number;
@@ -443,14 +473,17 @@ export interface Ac3Settings {
443
473
  */
444
474
  export interface AiffSettings {
445
475
  /**
476
+ * @public
446
477
  * Specify Bit depth, in bits per sample, to choose the encoding quality for this audio track.
447
478
  */
448
479
  BitDepth?: number;
449
480
  /**
481
+ * @public
450
482
  * Specify the number of channels in this output audio track. Valid values are 1 and even numbers up to 64. For example, 1, 2, 4, 6, and so on, up to 64.
451
483
  */
452
484
  Channels?: number;
453
485
  /**
486
+ * @public
454
487
  * Sample rate in hz.
455
488
  */
456
489
  SampleRate?: number;
@@ -617,70 +650,87 @@ export type Eac3AtmosSurroundExMode = (typeof Eac3AtmosSurroundExMode)[keyof typ
617
650
  */
618
651
  export interface Eac3AtmosSettings {
619
652
  /**
653
+ * @public
620
654
  * Specify the average bitrate for this output in bits per second. Valid values: 384k, 448k, 576k, 640k, 768k, 1024k Default value: 448k Note that MediaConvert supports 384k only with channel-based immersive (CBI) 7.1.4 and 5.1.4 inputs. For CBI 9.1.6 and other input types, MediaConvert automatically increases your output bitrate to 448k.
621
655
  */
622
656
  Bitrate?: number;
623
657
  /**
658
+ * @public
624
659
  * Specify the bitstream mode for the E-AC-3 stream that the encoder emits. For more information about the EAC3 bitstream mode, see ATSC A/52-2012 (Annex E).
625
660
  */
626
661
  BitstreamMode?: Eac3AtmosBitstreamMode | string;
627
662
  /**
663
+ * @public
628
664
  * The coding mode for Dolby Digital Plus JOC (Atmos).
629
665
  */
630
666
  CodingMode?: Eac3AtmosCodingMode | string;
631
667
  /**
668
+ * @public
632
669
  * Enable Dolby Dialogue Intelligence to adjust loudness based on dialogue analysis.
633
670
  */
634
671
  DialogueIntelligence?: Eac3AtmosDialogueIntelligence | string;
635
672
  /**
673
+ * @public
636
674
  * Specify whether MediaConvert should use any downmix metadata from your input file. Keep the default value, Custom to provide downmix values in your job settings. Choose Follow source to use the metadata from your input. Related settings--Use these settings to specify your downmix values: Left only/Right only surround, Left total/Right total surround, Left total/Right total center, Left only/Right only center, and Stereo downmix. When you keep Custom for Downmix control and you don't specify values for the related settings, MediaConvert uses default values for those settings.
637
675
  */
638
676
  DownmixControl?: Eac3AtmosDownmixControl | string;
639
677
  /**
678
+ * @public
640
679
  * Choose the Dolby dynamic range control (DRC) profile that MediaConvert uses when encoding the metadata in the Dolby stream for the line operating mode. Default value: Film light Related setting: To have MediaConvert use the value you specify here, keep the default value, Custom for the setting Dynamic range control. Otherwise, MediaConvert ignores Dynamic range compression line. For information about the Dolby DRC operating modes and profiles, see the Dynamic Range Control chapter of the Dolby Metadata Guide at https://developer.dolby.com/globalassets/professional/documents/dolby-metadata-guide.pdf.
641
680
  */
642
681
  DynamicRangeCompressionLine?: Eac3AtmosDynamicRangeCompressionLine | string;
643
682
  /**
683
+ * @public
644
684
  * Choose the Dolby dynamic range control (DRC) profile that MediaConvert uses when encoding the metadata in the Dolby stream for the RF operating mode. Default value: Film light Related setting: To have MediaConvert use the value you specify here, keep the default value, Custom for the setting Dynamic range control. Otherwise, MediaConvert ignores Dynamic range compression RF. For information about the Dolby DRC operating modes and profiles, see the Dynamic Range Control chapter of the Dolby Metadata Guide at https://developer.dolby.com/globalassets/professional/documents/dolby-metadata-guide.pdf.
645
685
  */
646
686
  DynamicRangeCompressionRf?: Eac3AtmosDynamicRangeCompressionRf | string;
647
687
  /**
688
+ * @public
648
689
  * Specify whether MediaConvert should use any dynamic range control metadata from your input file. Keep the default value, Custom, to provide dynamic range control values in your job settings. Choose Follow source to use the metadata from your input. Related settings--Use these settings to specify your dynamic range control values: Dynamic range compression line and Dynamic range compression RF. When you keep the value Custom for Dynamic range control and you don't specify values for the related settings, MediaConvert uses default values for those settings.
649
690
  */
650
691
  DynamicRangeControl?: Eac3AtmosDynamicRangeControl | string;
651
692
  /**
693
+ * @public
652
694
  * Specify a value for the following Dolby Atmos setting: Left only/Right only center mix (Lo/Ro center). MediaConvert uses this value for downmixing. Default value: -3 dB. Valid values: 3.0, 1.5, 0.0, -1.5, -3.0, -4.5, and -6.0. Related setting: How the service uses this value depends on the value that you choose for Stereo downmix. Related setting: To have MediaConvert use this value, keep the default value, Custom for the setting Downmix control. Otherwise, MediaConvert ignores Left only/Right only center.
653
695
  */
654
696
  LoRoCenterMixLevel?: number;
655
697
  /**
698
+ * @public
656
699
  * Specify a value for the following Dolby Atmos setting: Left only/Right only. MediaConvert uses this value for downmixing. Default value: -3 dB. Valid values: -1.5, -3.0, -4.5, -6.0, and -60. The value -60 mutes the channel. Related setting: How the service uses this value depends on the value that you choose for Stereo downmix. Related setting: To have MediaConvert use this value, keep the default value, Custom for the setting Downmix control. Otherwise, MediaConvert ignores Left only/Right only surround.
657
700
  */
658
701
  LoRoSurroundMixLevel?: number;
659
702
  /**
703
+ * @public
660
704
  * Specify a value for the following Dolby Atmos setting: Left total/Right total center mix (Lt/Rt center). MediaConvert uses this value for downmixing. Default value: -3 dB Valid values: 3.0, 1.5, 0.0, -1.5, -3.0, -4.5, and -6.0. Related setting: How the service uses this value depends on the value that you choose for Stereo downmix. Related setting: To have MediaConvert use this value, keep the default value, Custom for the setting Downmix control. Otherwise, MediaConvert ignores Left total/Right total center.
661
705
  */
662
706
  LtRtCenterMixLevel?: number;
663
707
  /**
708
+ * @public
664
709
  * Specify a value for the following Dolby Atmos setting: Left total/Right total surround mix (Lt/Rt surround). MediaConvert uses this value for downmixing. Default value: -3 dB Valid values: -1.5, -3.0, -4.5, -6.0, and -60. The value -60 mutes the channel. Related setting: How the service uses this value depends on the value that you choose for Stereo downmix. Related setting: To have MediaConvert use this value, keep the default value, Custom for the setting Downmix control. Otherwise, the service ignores Left total/Right total surround.
665
710
  */
666
711
  LtRtSurroundMixLevel?: number;
667
712
  /**
713
+ * @public
668
714
  * Choose how the service meters the loudness of your audio.
669
715
  */
670
716
  MeteringMode?: Eac3AtmosMeteringMode | string;
671
717
  /**
718
+ * @public
672
719
  * This value is always 48000. It represents the sample rate in Hz.
673
720
  */
674
721
  SampleRate?: number;
675
722
  /**
723
+ * @public
676
724
  * Specify the percentage of audio content, from 0% to 100%, that must be speech in order for the encoder to use the measured speech loudness as the overall program loudness. Default value: 15%
677
725
  */
678
726
  SpeechThreshold?: number;
679
727
  /**
728
+ * @public
680
729
  * Choose how the service does stereo downmixing. Default value: Not indicated Related setting: To have MediaConvert use this value, keep the default value, Custom for the setting Downmix control. Otherwise, MediaConvert ignores Stereo downmix.
681
730
  */
682
731
  StereoDownmix?: Eac3AtmosStereoDownmix | string;
683
732
  /**
733
+ * @public
684
734
  * Specify whether your input audio has an additional center rear surround channel matrix encoded into your left and right surround channels.
685
735
  */
686
736
  SurroundExMode?: Eac3AtmosSurroundExMode | string;
@@ -875,86 +925,107 @@ export type Eac3SurroundMode = (typeof Eac3SurroundMode)[keyof typeof Eac3Surrou
875
925
  */
876
926
  export interface Eac3Settings {
877
927
  /**
928
+ * @public
878
929
  * If set to ATTENUATE_3_DB, applies a 3 dB attenuation to the surround channels. Only used for 3/2 coding mode.
879
930
  */
880
931
  AttenuationControl?: Eac3AttenuationControl | string;
881
932
  /**
933
+ * @public
882
934
  * Specify the average bitrate in bits per second. The bitrate that you specify must be a multiple of 8000 within the allowed minimum and maximum values. Leave blank to use the default bitrate for the coding mode you select according ETSI TS 102 366. Valid bitrates for coding mode 1/0: Default: 96000. Minimum: 32000. Maximum: 3024000. Valid bitrates for coding mode 2/0: Default: 192000. Minimum: 96000. Maximum: 3024000. Valid bitrates for coding mode 3/2: Default: 384000. Minimum: 192000. Maximum: 3024000.
883
935
  */
884
936
  Bitrate?: number;
885
937
  /**
938
+ * @public
886
939
  * Specify the bitstream mode for the E-AC-3 stream that the encoder emits. For more information about the EAC3 bitstream mode, see ATSC A/52-2012 (Annex E).
887
940
  */
888
941
  BitstreamMode?: Eac3BitstreamMode | string;
889
942
  /**
943
+ * @public
890
944
  * Dolby Digital Plus coding mode. Determines number of channels.
891
945
  */
892
946
  CodingMode?: Eac3CodingMode | string;
893
947
  /**
948
+ * @public
894
949
  * Activates a DC highpass filter for all input channels.
895
950
  */
896
951
  DcFilter?: Eac3DcFilter | string;
897
952
  /**
953
+ * @public
898
954
  * Sets the dialnorm for the output. If blank and input audio is Dolby Digital Plus, dialnorm will be passed through.
899
955
  */
900
956
  Dialnorm?: number;
901
957
  /**
958
+ * @public
902
959
  * Choose the Dolby Digital dynamic range control (DRC) profile that MediaConvert uses when encoding the metadata in the Dolby Digital stream for the line operating mode. Related setting: When you use this setting, MediaConvert ignores any value you provide for Dynamic range compression profile. For information about the Dolby Digital DRC operating modes and profiles, see the Dynamic Range Control chapter of the Dolby Metadata Guide at https://developer.dolby.com/globalassets/professional/documents/dolby-metadata-guide.pdf.
903
960
  */
904
961
  DynamicRangeCompressionLine?: Eac3DynamicRangeCompressionLine | string;
905
962
  /**
963
+ * @public
906
964
  * Choose the Dolby Digital dynamic range control (DRC) profile that MediaConvert uses when encoding the metadata in the Dolby Digital stream for the RF operating mode. Related setting: When you use this setting, MediaConvert ignores any value you provide for Dynamic range compression profile. For information about the Dolby Digital DRC operating modes and profiles, see the Dynamic Range Control chapter of the Dolby Metadata Guide at https://developer.dolby.com/globalassets/professional/documents/dolby-metadata-guide.pdf.
907
965
  */
908
966
  DynamicRangeCompressionRf?: Eac3DynamicRangeCompressionRf | string;
909
967
  /**
968
+ * @public
910
969
  * When encoding 3/2 audio, controls whether the LFE channel is enabled
911
970
  */
912
971
  LfeControl?: Eac3LfeControl | string;
913
972
  /**
973
+ * @public
914
974
  * Applies a 120Hz lowpass filter to the LFE channel prior to encoding. Only valid with 3_2_LFE coding mode.
915
975
  */
916
976
  LfeFilter?: Eac3LfeFilter | string;
917
977
  /**
978
+ * @public
918
979
  * Specify a value for the following Dolby Digital Plus setting: Left only/Right only center mix. MediaConvert uses this value for downmixing. How the service uses this value depends on the value that you choose for Stereo downmix. Valid values: 3.0, 1.5, 0.0, -1.5, -3.0, -4.5, -6.0, and -60. The value -60 mutes the channel. This setting applies only if you keep the default value of 3/2 - L, R, C, Ls, Rs for the setting Coding mode. If you choose a different value for Coding mode, the service ignores Left only/Right only center.
919
980
  */
920
981
  LoRoCenterMixLevel?: number;
921
982
  /**
983
+ * @public
922
984
  * Specify a value for the following Dolby Digital Plus setting: Left only/Right only. MediaConvert uses this value for downmixing. How the service uses this value depends on the value that you choose for Stereo downmix. Valid values: -1.5, -3.0, -4.5, -6.0, and -60. The value -60 mutes the channel. This setting applies only if you keep the default value of 3/2 - L, R, C, Ls, Rs for the setting Coding mode. If you choose a different value for Coding mode, the service ignores Left only/Right only surround.
923
985
  */
924
986
  LoRoSurroundMixLevel?: number;
925
987
  /**
988
+ * @public
926
989
  * Specify a value for the following Dolby Digital Plus setting: Left total/Right total center mix. MediaConvert uses this value for downmixing. How the service uses this value depends on the value that you choose for Stereo downmix. Valid values: 3.0, 1.5, 0.0, -1.5, -3.0, -4.5, -6.0, and -60. The value -60 mutes the channel. This setting applies only if you keep the default value of 3/2 - L, R, C, Ls, Rs for the setting Coding mode. If you choose a different value for Coding mode, the service ignores Left total/Right total center.
927
990
  */
928
991
  LtRtCenterMixLevel?: number;
929
992
  /**
993
+ * @public
930
994
  * Specify a value for the following Dolby Digital Plus setting: Left total/Right total surround mix. MediaConvert uses this value for downmixing. How the service uses this value depends on the value that you choose for Stereo downmix. Valid values: -1.5, -3.0, -4.5, -6.0, and -60. The value -60 mutes the channel. This setting applies only if you keep the default value of 3/2 - L, R, C, Ls, Rs for the setting Coding mode. If you choose a different value for Coding mode, the service ignores Left total/Right total surround.
931
995
  */
932
996
  LtRtSurroundMixLevel?: number;
933
997
  /**
998
+ * @public
934
999
  * When set to FOLLOW_INPUT, encoder metadata will be sourced from the DD, DD+, or DolbyE decoder that supplied this audio data. If audio was not supplied from one of these streams, then the static metadata settings will be used.
935
1000
  */
936
1001
  MetadataControl?: Eac3MetadataControl | string;
937
1002
  /**
1003
+ * @public
938
1004
  * When set to WHEN_POSSIBLE, input DD+ audio will be passed through if it is present on the input. this detection is dynamic over the life of the transcode. Inputs that alternate between DD+ and non-DD+ content will have a consistent DD+ output as the system alternates between passthrough and encoding.
939
1005
  */
940
1006
  PassthroughControl?: Eac3PassthroughControl | string;
941
1007
  /**
1008
+ * @public
942
1009
  * Controls the amount of phase-shift applied to the surround channels. Only used for 3/2 coding mode.
943
1010
  */
944
1011
  PhaseControl?: Eac3PhaseControl | string;
945
1012
  /**
1013
+ * @public
946
1014
  * This value is always 48000. It represents the sample rate in Hz.
947
1015
  */
948
1016
  SampleRate?: number;
949
1017
  /**
1018
+ * @public
950
1019
  * Choose how the service does stereo downmixing. This setting only applies if you keep the default value of 3/2 - L, R, C, Ls, Rs for the setting Coding mode. If you choose a different value for Coding mode, the service ignores Stereo downmix.
951
1020
  */
952
1021
  StereoDownmix?: Eac3StereoDownmix | string;
953
1022
  /**
1023
+ * @public
954
1024
  * When encoding 3/2 audio, sets whether an extra center back surround channel is matrix encoded into the left and right surround channels.
955
1025
  */
956
1026
  SurroundExMode?: Eac3SurroundExMode | string;
957
1027
  /**
1028
+ * @public
958
1029
  * When encoding 2/0 audio, sets whether Dolby Surround is matrix encoded into the two channels.
959
1030
  */
960
1031
  SurroundMode?: Eac3SurroundMode | string;
@@ -965,14 +1036,17 @@ export interface Eac3Settings {
965
1036
  */
966
1037
  export interface Mp2Settings {
967
1038
  /**
1039
+ * @public
968
1040
  * Specify the average bitrate in bits per second.
969
1041
  */
970
1042
  Bitrate?: number;
971
1043
  /**
1044
+ * @public
972
1045
  * Set Channels to specify the number of channels in this output audio track. Choosing Mono in will give you 1 output channel; choosing Stereo will give you 2. In the API, valid values are 1 and 2.
973
1046
  */
974
1047
  Channels?: number;
975
1048
  /**
1049
+ * @public
976
1050
  * Sample rate in hz.
977
1051
  */
978
1052
  SampleRate?: number;
@@ -995,22 +1069,27 @@ export type Mp3RateControlMode = (typeof Mp3RateControlMode)[keyof typeof Mp3Rat
995
1069
  */
996
1070
  export interface Mp3Settings {
997
1071
  /**
1072
+ * @public
998
1073
  * Specify the average bitrate in bits per second.
999
1074
  */
1000
1075
  Bitrate?: number;
1001
1076
  /**
1077
+ * @public
1002
1078
  * Specify the number of channels in this output audio track. Choosing Mono gives you 1 output channel; choosing Stereo gives you 2. In the API, valid values are 1 and 2.
1003
1079
  */
1004
1080
  Channels?: number;
1005
1081
  /**
1082
+ * @public
1006
1083
  * Specify whether the service encodes this MP3 audio output with a constant bitrate (CBR) or a variable bitrate (VBR).
1007
1084
  */
1008
1085
  RateControlMode?: Mp3RateControlMode | string;
1009
1086
  /**
1087
+ * @public
1010
1088
  * Sample rate in hz.
1011
1089
  */
1012
1090
  SampleRate?: number;
1013
1091
  /**
1092
+ * @public
1014
1093
  * Required when you set Bitrate control mode to VBR. Specify the audio quality of this MP3 output from 0 (highest quality) to 9 (lowest quality).
1015
1094
  */
1016
1095
  VbrQuality?: number;
@@ -1021,14 +1100,17 @@ export interface Mp3Settings {
1021
1100
  */
1022
1101
  export interface OpusSettings {
1023
1102
  /**
1103
+ * @public
1024
1104
  * Optional. Specify the average bitrate in bits per second. Valid values are multiples of 8000, from 32000 through 192000. The default value is 96000, which we recommend for quality and bandwidth.
1025
1105
  */
1026
1106
  Bitrate?: number;
1027
1107
  /**
1108
+ * @public
1028
1109
  * Specify the number of channels in this output audio track. Choosing Mono on gives you 1 output channel; choosing Stereo gives you 2. In the API, valid values are 1 and 2.
1029
1110
  */
1030
1111
  Channels?: number;
1031
1112
  /**
1113
+ * @public
1032
1114
  * Optional. Sample rate in hz. Valid values are 16000, 24000, and 48000. The default value is 48000.
1033
1115
  */
1034
1116
  SampleRate?: number;
@@ -1039,14 +1121,17 @@ export interface OpusSettings {
1039
1121
  */
1040
1122
  export interface VorbisSettings {
1041
1123
  /**
1124
+ * @public
1042
1125
  * Optional. Specify the number of channels in this output audio track. Choosing Mono on the console gives you 1 output channel; choosing Stereo gives you 2. In the API, valid values are 1 and 2. The default value is 2.
1043
1126
  */
1044
1127
  Channels?: number;
1045
1128
  /**
1129
+ * @public
1046
1130
  * Optional. Specify the audio sample rate in Hz. Valid values are 22050, 32000, 44100, and 48000. The default value is 48000.
1047
1131
  */
1048
1132
  SampleRate?: number;
1049
1133
  /**
1134
+ * @public
1050
1135
  * Optional. Specify the variable audio quality of this Vorbis output from -1 (lowest quality, ~45 kbit/s) to 10 (highest quality, ~500 kbit/s). The default value is 4 (~128 kbit/s). Values 5 and 6 are approximately 160 and 192 kbit/s, respectively.
1051
1136
  */
1052
1137
  VbrQuality?: number;
@@ -1069,18 +1154,22 @@ export type WavFormat = (typeof WavFormat)[keyof typeof WavFormat];
1069
1154
  */
1070
1155
  export interface WavSettings {
1071
1156
  /**
1157
+ * @public
1072
1158
  * Specify Bit depth, in bits per sample, to choose the encoding quality for this audio track.
1073
1159
  */
1074
1160
  BitDepth?: number;
1075
1161
  /**
1162
+ * @public
1076
1163
  * Specify the number of channels in this output audio track. Valid values are 1 and even numbers up to 64. For example, 1, 2, 4, 6, and so on, up to 64.
1077
1164
  */
1078
1165
  Channels?: number;
1079
1166
  /**
1167
+ * @public
1080
1168
  * The service defaults to using RIFF for WAV outputs. If your output audio is likely to exceed 4 GB in file size, or if you otherwise need the extended support of the RF64 format, set your output WAV file format to RF64.
1081
1169
  */
1082
1170
  Format?: WavFormat | string;
1083
1171
  /**
1172
+ * @public
1084
1173
  * Sample rate in Hz.
1085
1174
  */
1086
1175
  SampleRate?: number;
@@ -1091,46 +1180,57 @@ export interface WavSettings {
1091
1180
  */
1092
1181
  export interface AudioCodecSettings {
1093
1182
  /**
1183
+ * @public
1094
1184
  * Required when you set Codec to the value AAC. The service accepts one of two mutually exclusive groups of AAC settings--VBR and CBR. To select one of these modes, set the value of Bitrate control mode to "VBR" or "CBR". In VBR mode, you control the audio quality with the setting VBR quality. In CBR mode, you use the setting Bitrate. Defaults and valid values depend on the rate control mode.
1095
1185
  */
1096
1186
  AacSettings?: AacSettings;
1097
1187
  /**
1188
+ * @public
1098
1189
  * Required when you set Codec to the value AC3.
1099
1190
  */
1100
1191
  Ac3Settings?: Ac3Settings;
1101
1192
  /**
1193
+ * @public
1102
1194
  * Required when you set Codec to the value AIFF.
1103
1195
  */
1104
1196
  AiffSettings?: AiffSettings;
1105
1197
  /**
1198
+ * @public
1106
1199
  * Choose the audio codec for this output. Note that the option Dolby Digital passthrough applies only to Dolby Digital and Dolby Digital Plus audio inputs. Make sure that you choose a codec that's supported with your output container: https://docs.aws.amazon.com/mediaconvert/latest/ug/reference-codecs-containers.html#reference-codecs-containers-output-audio For audio-only outputs, make sure that both your input audio codec and your output audio codec are supported for audio-only workflows. For more information, see: https://docs.aws.amazon.com/mediaconvert/latest/ug/reference-codecs-containers-input.html#reference-codecs-containers-input-audio-only and https://docs.aws.amazon.com/mediaconvert/latest/ug/reference-codecs-containers.html#audio-only-output
1107
1200
  */
1108
1201
  Codec?: AudioCodec | string;
1109
1202
  /**
1203
+ * @public
1110
1204
  * Required when you set Codec to the value EAC3_ATMOS.
1111
1205
  */
1112
1206
  Eac3AtmosSettings?: Eac3AtmosSettings;
1113
1207
  /**
1208
+ * @public
1114
1209
  * Required when you set Codec to the value EAC3.
1115
1210
  */
1116
1211
  Eac3Settings?: Eac3Settings;
1117
1212
  /**
1213
+ * @public
1118
1214
  * Required when you set Codec to the value MP2.
1119
1215
  */
1120
1216
  Mp2Settings?: Mp2Settings;
1121
1217
  /**
1218
+ * @public
1122
1219
  * Required when you set Codec, under AudioDescriptions>CodecSettings, to the value MP3.
1123
1220
  */
1124
1221
  Mp3Settings?: Mp3Settings;
1125
1222
  /**
1223
+ * @public
1126
1224
  * Required when you set Codec, under AudioDescriptions>CodecSettings, to the value OPUS.
1127
1225
  */
1128
1226
  OpusSettings?: OpusSettings;
1129
1227
  /**
1228
+ * @public
1130
1229
  * Required when you set Codec, under AudioDescriptions>CodecSettings, to the value Vorbis.
1131
1230
  */
1132
1231
  VorbisSettings?: VorbisSettings;
1133
1232
  /**
1233
+ * @public
1134
1234
  * Required when you set Codec to the value WAV.
1135
1235
  */
1136
1236
  WavSettings?: WavSettings;
@@ -1355,10 +1455,12 @@ export type AudioLanguageCodeControl = (typeof AudioLanguageCodeControl)[keyof t
1355
1455
  */
1356
1456
  export interface OutputChannelMapping {
1357
1457
  /**
1458
+ * @public
1358
1459
  * Use this setting to specify your remix values when they are integers, such as -10, 0, or 4.
1359
1460
  */
1360
1461
  InputChannels?: number[];
1361
1462
  /**
1463
+ * @public
1362
1464
  * Use this setting to specify your remix values when they have a decimal component, such as -10.312, 0.08, or 4.9. MediaConvert rounds your remixing values to the nearest thousandth.
1363
1465
  */
1364
1466
  InputChannelsFineTune?: number[];
@@ -1369,6 +1471,7 @@ export interface OutputChannelMapping {
1369
1471
  */
1370
1472
  export interface ChannelMapping {
1371
1473
  /**
1474
+ * @public
1372
1475
  * In your JSON job specification, include one child of OutputChannels for each audio channel that you want in your output. Each child should contain one instance of InputChannels or InputChannelsFineTune.
1373
1476
  */
1374
1477
  OutputChannels?: OutputChannelMapping[];
@@ -1379,14 +1482,17 @@ export interface ChannelMapping {
1379
1482
  */
1380
1483
  export interface RemixSettings {
1381
1484
  /**
1485
+ * @public
1382
1486
  * Channel mapping contains the group of fields that hold the remixing value for each channel, in dB. Specify remix values to indicate how much of the content from your input audio channel you want in your output audio channels. Each instance of the InputChannels or InputChannelsFineTune array specifies these values for one output channel. Use one instance of this array for each output channel. In the console, each array corresponds to a column in the graphical depiction of the mapping matrix. The rows of the graphical matrix correspond to input channels. Valid values are within the range from -60 (mute) through 6. A setting of 0 passes the input channel unchanged to the output channel (no attenuation or amplification). Use InputChannels or InputChannelsFineTune to specify your remix values. Don't use both.
1383
1487
  */
1384
1488
  ChannelMapping?: ChannelMapping;
1385
1489
  /**
1490
+ * @public
1386
1491
  * Specify the number of audio channels from your input that you want to use in your output. With remixing, you might combine or split the data in these channels, so the number of channels in your final output might be different. If you are doing both input channel mapping and output channel mapping, the number of output channels in your input mapping must be the same as the number of input channels in your output mapping.
1387
1492
  */
1388
1493
  ChannelsIn?: number;
1389
1494
  /**
1495
+ * @public
1390
1496
  * Specify the number of channels in this output after remixing. Valid values: 1, 2, 4, 6, 8... 64. (1 and even numbers to 64.) If you are doing both input channel mapping and output channel mapping, the number of output channels in your input mapping must be the same as the number of input channels in your output mapping.
1391
1497
  */
1392
1498
  ChannelsOut?: number;
@@ -1397,46 +1503,57 @@ export interface RemixSettings {
1397
1503
  */
1398
1504
  export interface AudioDescription {
1399
1505
  /**
1506
+ * @public
1400
1507
  * When you mimic a multi-channel audio layout with multiple mono-channel tracks, you can tag each channel layout manually. For example, you would tag the tracks that contain your left, right, and center audio with Left (L), Right (R), and Center (C), respectively. When you don't specify a value, MediaConvert labels your track as Center (C) by default. To use audio layout tagging, your output must be in a QuickTime (.mov) container; your audio codec must be AAC, WAV, or AIFF; and you must set up your audio track to have only one channel.
1401
1508
  */
1402
1509
  AudioChannelTaggingSettings?: AudioChannelTaggingSettings;
1403
1510
  /**
1511
+ * @public
1404
1512
  * Advanced audio normalization settings. Ignore these settings unless you need to comply with a loudness standard.
1405
1513
  */
1406
1514
  AudioNormalizationSettings?: AudioNormalizationSettings;
1407
1515
  /**
1516
+ * @public
1408
1517
  * Specifies which audio data to use from each input. In the simplest case, specify an "Audio Selector":#inputs-audio_selector by name based on its order within each input. For example if you specify "Audio Selector 3", then the third audio selector will be used from each input. If an input does not have an "Audio Selector 3", then the audio selector marked as "default" in that input will be used. If there is no audio selector marked as "default", silence will be inserted for the duration of that input. Alternatively, an "Audio Selector Group":#inputs-audio_selector_group name may be specified, with similar default/silence behavior. If no audio_source_name is specified, then "Audio Selector 1" will be chosen automatically.
1409
1518
  */
1410
1519
  AudioSourceName?: string;
1411
1520
  /**
1521
+ * @public
1412
1522
  * Applies only if Follow Input Audio Type is unchecked (false). A number between 0 and 255. The following are defined in ISO-IEC 13818-1: 0 = Undefined, 1 = Clean Effects, 2 = Hearing Impaired, 3 = Visually Impaired Commentary, 4-255 = Reserved.
1413
1523
  */
1414
1524
  AudioType?: number;
1415
1525
  /**
1526
+ * @public
1416
1527
  * When set to FOLLOW_INPUT, if the input contains an ISO 639 audio_type, then that value is passed through to the output. If the input contains no ISO 639 audio_type, the value in Audio Type is included in the output. Otherwise the value in Audio Type is included in the output. Note that this field and audioType are both ignored if audioDescriptionBroadcasterMix is set to BROADCASTER_MIXED_AD.
1417
1528
  */
1418
1529
  AudioTypeControl?: AudioTypeControl | string;
1419
1530
  /**
1531
+ * @public
1420
1532
  * Settings related to audio encoding. The settings in this group vary depending on the value that you choose for your audio codec.
1421
1533
  */
1422
1534
  CodecSettings?: AudioCodecSettings;
1423
1535
  /**
1536
+ * @public
1424
1537
  * Specify the language for this audio output track. The service puts this language code into your output audio track when you set Language code control to Use configured. The service also uses your specified custom language code when you set Language code control to Follow input, but your input file doesn't specify a language code. For all outputs, you can use an ISO 639-2 or ISO 639-3 code. For streaming outputs, you can also use any other code in the full RFC-5646 specification. Streaming outputs are those that are in one of the following output groups: CMAF, DASH ISO, Apple HLS, or Microsoft Smooth Streaming.
1425
1538
  */
1426
1539
  CustomLanguageCode?: string;
1427
1540
  /**
1541
+ * @public
1428
1542
  * Indicates the language of the audio output track. The ISO 639 language specified in the 'Language Code' drop down will be used when 'Follow Input Language Code' is not selected or when 'Follow Input Language Code' is selected but there is no ISO 639 language code specified by the input.
1429
1543
  */
1430
1544
  LanguageCode?: LanguageCode | string;
1431
1545
  /**
1546
+ * @public
1432
1547
  * Specify which source for language code takes precedence for this audio track. When you choose Follow input, the service uses the language code from the input track if it's present. If there's no languge code on the input track, the service uses the code that you specify in the setting Language code. When you choose Use configured, the service uses the language code that you specify.
1433
1548
  */
1434
1549
  LanguageCodeControl?: AudioLanguageCodeControl | string;
1435
1550
  /**
1551
+ * @public
1436
1552
  * Advanced audio remixing settings.
1437
1553
  */
1438
1554
  RemixSettings?: RemixSettings;
1439
1555
  /**
1556
+ * @public
1440
1557
  * Specify a label for this output audio stream. For example, "English", "Director commentary", or "track_2". For streaming outputs, MediaConvert passes this information into destination manifests for display on the end-viewer's player device. For outputs in other output groups, the service ignores this setting.
1441
1558
  */
1442
1559
  StreamName?: string;
@@ -1447,10 +1564,12 @@ export interface AudioDescription {
1447
1564
  */
1448
1565
  export interface ForceIncludeRenditionSize {
1449
1566
  /**
1567
+ * @public
1450
1568
  * Use Height to define the video resolution height, in pixels, for this rule.
1451
1569
  */
1452
1570
  Height?: number;
1453
1571
  /**
1572
+ * @public
1454
1573
  * Use Width to define the video resolution width, in pixels, for this rule.
1455
1574
  */
1456
1575
  Width?: number;
@@ -1461,10 +1580,12 @@ export interface ForceIncludeRenditionSize {
1461
1580
  */
1462
1581
  export interface MinBottomRenditionSize {
1463
1582
  /**
1583
+ * @public
1464
1584
  * Use Height to define the video resolution height, in pixels, for this rule.
1465
1585
  */
1466
1586
  Height?: number;
1467
1587
  /**
1588
+ * @public
1468
1589
  * Use Width to define the video resolution width, in pixels, for this rule.
1469
1590
  */
1470
1591
  Width?: number;
@@ -1475,10 +1596,12 @@ export interface MinBottomRenditionSize {
1475
1596
  */
1476
1597
  export interface MinTopRenditionSize {
1477
1598
  /**
1599
+ * @public
1478
1600
  * Use Height to define the video resolution height, in pixels, for this rule.
1479
1601
  */
1480
1602
  Height?: number;
1481
1603
  /**
1604
+ * @public
1482
1605
  * Use Width to define the video resolution width, in pixels, for this rule.
1483
1606
  */
1484
1607
  Width?: number;
@@ -1503,22 +1626,27 @@ export type RuleType = (typeof RuleType)[keyof typeof RuleType];
1503
1626
  */
1504
1627
  export interface AutomatedAbrRule {
1505
1628
  /**
1629
+ * @public
1506
1630
  * When customer adds the allowed renditions rule for auto ABR ladder, they are required to add at leat one rendition to allowedRenditions list
1507
1631
  */
1508
1632
  AllowedRenditions?: AllowedRenditionSize[];
1509
1633
  /**
1634
+ * @public
1510
1635
  * When customer adds the force include renditions rule for auto ABR ladder, they are required to add at leat one rendition to forceIncludeRenditions list
1511
1636
  */
1512
1637
  ForceIncludeRenditions?: ForceIncludeRenditionSize[];
1513
1638
  /**
1639
+ * @public
1514
1640
  * Use Min bottom rendition size to specify a minimum size for the lowest resolution in your ABR stack. * The lowest resolution in your ABR stack will be equal to or greater than the value that you enter. For example: If you specify 640x360 the lowest resolution in your ABR stack will be equal to or greater than to 640x360. * If you specify a Min top rendition size rule, the value that you specify for Min bottom rendition size must be less than, or equal to, Min top rendition size.
1515
1641
  */
1516
1642
  MinBottomRenditionSize?: MinBottomRenditionSize;
1517
1643
  /**
1644
+ * @public
1518
1645
  * Use Min top rendition size to specify a minimum size for the highest resolution in your ABR stack. * The highest resolution in your ABR stack will be equal to or greater than the value that you enter. For example: If you specify 1280x720 the highest resolution in your ABR stack will be equal to or greater than 1280x720. * If you specify a value for Max resolution, the value that you specify for Min top rendition size must be less than, or equal to, Max resolution.
1519
1646
  */
1520
1647
  MinTopRenditionSize?: MinTopRenditionSize;
1521
1648
  /**
1649
+ * @public
1522
1650
  * Use Min top rendition size to specify a minimum size for the highest resolution in your ABR stack. * The highest resolution in your ABR stack will be equal to or greater than the value that you enter. For example: If you specify 1280x720 the highest resolution in your ABR stack will be equal to or greater than 1280x720. * If you specify a value for Max resolution, the value that you specify for Min top rendition size must be less than, or equal to, Max resolution. Use Min bottom rendition size to specify a minimum size for the lowest resolution in your ABR stack. * The lowest resolution in your ABR stack will be equal to or greater than the value that you enter. For example: If you specify 640x360 the lowest resolution in your ABR stack will be equal to or greater than to 640x360. * If you specify a Min top rendition size rule, the value that you specify for Min bottom rendition size must be less than, or equal to, Min top rendition size. Use Force include renditions to specify one or more resolutions to include your ABR stack. * (Recommended) To optimize automated ABR, specify as few resolutions as possible. * (Required) The number of resolutions that you specify must be equal to, or less than, the Max renditions setting. * If you specify a Min top rendition size rule, specify at least one resolution that is equal to, or greater than, Min top rendition size. * If you specify a Min bottom rendition size rule, only specify resolutions that are equal to, or greater than, Min bottom rendition size. * If you specify a Force include renditions rule, do not specify a separate rule for Allowed renditions. * Note: The ABR stack may include other resolutions that you do not specify here, depending on the Max renditions setting. Use Allowed renditions to specify a list of possible resolutions in your ABR stack. * (Required) The number of resolutions that you specify must be equal to, or greater than, the Max renditions setting. * MediaConvert will create an ABR stack exclusively from the list of resolutions that you specify. * Some resolutions in the Allowed renditions list may not be included, however you can force a resolution to be included by setting Required to ENABLED. * You must specify at least one resolution that is greater than or equal to any resolutions that you specify in Min top rendition size or Min bottom rendition size. * If you specify Allowed renditions, you must not specify a separate rule for Force include renditions.
1523
1651
  */
1524
1652
  Type?: RuleType | string;
@@ -1670,86 +1798,107 @@ export type BurninSubtitleTeletextSpacing = (typeof BurninSubtitleTeletextSpacin
1670
1798
  */
1671
1799
  export interface BurninDestinationSettings {
1672
1800
  /**
1801
+ * @public
1673
1802
  * Specify the alignment of your captions. If no explicit x_position is provided, setting alignment to centered will placethe captions at the bottom center of the output. Similarly, setting a left alignment willalign captions to the bottom left of the output. If x and y positions are given in conjunction with the alignment parameter, the font will be justified (either left or centered) relative to those coordinates.
1674
1803
  */
1675
1804
  Alignment?: BurninSubtitleAlignment | string;
1676
1805
  /**
1806
+ * @public
1677
1807
  * Ignore this setting unless Style passthrough is set to Enabled and Font color set to Black, Yellow, Red, Green, Blue, or Hex. Use Apply font color for additional font color controls. When you choose White text only, or leave blank, your font color setting only applies to white text in your input captions. For example, if your font color setting is Yellow, and your input captions have red and white text, your output captions will have red and yellow text. When you choose ALL_TEXT, your font color setting applies to all of your output captions text.
1678
1808
  */
1679
1809
  ApplyFontColor?: BurninSubtitleApplyFontColor | string;
1680
1810
  /**
1811
+ * @public
1681
1812
  * Specify the color of the rectangle behind the captions. Leave background color blank and set Style passthrough to enabled to use the background color data from your input captions, if present.
1682
1813
  */
1683
1814
  BackgroundColor?: BurninSubtitleBackgroundColor | string;
1684
1815
  /**
1816
+ * @public
1685
1817
  * Specify the opacity of the background rectangle. Enter a value from 0 to 255, where 0 is transparent and 255 is opaque. If Style passthrough is set to enabled, leave blank to pass through the background style information in your input captions to your output captions. If Style passthrough is set to disabled, leave blank to use a value of 0 and remove all backgrounds from your output captions.
1686
1818
  */
1687
1819
  BackgroundOpacity?: number;
1688
1820
  /**
1821
+ * @public
1689
1822
  * Specify the font that you want the service to use for your burn in captions when your input captions specify a font that MediaConvert doesn't support. When you set Fallback font to best match, or leave blank, MediaConvert uses a supported font that most closely matches the font that your input captions specify. When there are multiple unsupported fonts in your input captions, MediaConvert matches each font with the supported font that matches best. When you explicitly choose a replacement font, MediaConvert uses that font to replace all unsupported fonts from your input.
1690
1823
  */
1691
1824
  FallbackFont?: BurninSubtitleFallbackFont | string;
1692
1825
  /**
1826
+ * @public
1693
1827
  * Specify the color of the burned-in captions text. Leave Font color blank and set Style passthrough to enabled to use the font color data from your input captions, if present.
1694
1828
  */
1695
1829
  FontColor?: BurninSubtitleFontColor | string;
1696
1830
  /**
1831
+ * @public
1697
1832
  * Specify the opacity of the burned-in captions. 255 is opaque; 0 is transparent.
1698
1833
  */
1699
1834
  FontOpacity?: number;
1700
1835
  /**
1836
+ * @public
1701
1837
  * Specify the Font resolution in DPI (dots per inch).
1702
1838
  */
1703
1839
  FontResolution?: number;
1704
1840
  /**
1841
+ * @public
1705
1842
  * Set Font script to Automatically determined, or leave blank, to automatically determine the font script in your input captions. Otherwise, set to Simplified Chinese (HANS) or Traditional Chinese (HANT) if your input font script uses Simplified or Traditional Chinese.
1706
1843
  */
1707
1844
  FontScript?: FontScript | string;
1708
1845
  /**
1846
+ * @public
1709
1847
  * Specify the Font size in pixels. Must be a positive integer. Set to 0, or leave blank, for automatic font size.
1710
1848
  */
1711
1849
  FontSize?: number;
1712
1850
  /**
1851
+ * @public
1713
1852
  * Ignore this setting unless your Font color is set to Hex. Enter either six or eight hexidecimal digits, representing red, green, and blue, with two optional extra digits for alpha. For example a value of 1122AABB is a red value of 0x11, a green value of 0x22, a blue value of 0xAA, and an alpha value of 0xBB.
1714
1853
  */
1715
1854
  HexFontColor?: string;
1716
1855
  /**
1856
+ * @public
1717
1857
  * Specify font outline color. Leave Outline color blank and set Style passthrough to enabled to use the font outline color data from your input captions, if present.
1718
1858
  */
1719
1859
  OutlineColor?: BurninSubtitleOutlineColor | string;
1720
1860
  /**
1861
+ * @public
1721
1862
  * Specify the Outline size of the caption text, in pixels. Leave Outline size blank and set Style passthrough to enabled to use the outline size data from your input captions, if present.
1722
1863
  */
1723
1864
  OutlineSize?: number;
1724
1865
  /**
1866
+ * @public
1725
1867
  * Specify the color of the shadow cast by the captions. Leave Shadow color blank and set Style passthrough to enabled to use the shadow color data from your input captions, if present.
1726
1868
  */
1727
1869
  ShadowColor?: BurninSubtitleShadowColor | string;
1728
1870
  /**
1871
+ * @public
1729
1872
  * Specify the opacity of the shadow. Enter a value from 0 to 255, where 0 is transparent and 255 is opaque. If Style passthrough is set to Enabled, leave Shadow opacity blank to pass through the shadow style information in your input captions to your output captions. If Style passthrough is set to disabled, leave blank to use a value of 0 and remove all shadows from your output captions.
1730
1873
  */
1731
1874
  ShadowOpacity?: number;
1732
1875
  /**
1876
+ * @public
1733
1877
  * Specify the horizontal offset of the shadow, relative to the captions in pixels. A value of -2 would result in a shadow offset 2 pixels to the left.
1734
1878
  */
1735
1879
  ShadowXOffset?: number;
1736
1880
  /**
1881
+ * @public
1737
1882
  * Specify the vertical offset of the shadow relative to the captions in pixels. A value of -2 would result in a shadow offset 2 pixels above the text. Leave Shadow y-offset blank and set Style passthrough to enabled to use the shadow y-offset data from your input captions, if present.
1738
1883
  */
1739
1884
  ShadowYOffset?: number;
1740
1885
  /**
1886
+ * @public
1741
1887
  * Set Style passthrough to ENABLED to use the available style, color, and position information from your input captions. MediaConvert uses default settings for any missing style and position information in your input captions. Set Style passthrough to DISABLED, or leave blank, to ignore the style and position information from your input captions and use default settings: white text with black outlining, bottom-center positioning, and automatic sizing. Whether you set Style passthrough to enabled or not, you can also choose to manually override any of the individual style and position settings.
1742
1888
  */
1743
1889
  StylePassthrough?: BurnInSubtitleStylePassthrough | string;
1744
1890
  /**
1891
+ * @public
1745
1892
  * Specify whether the text spacing in your captions is set by the captions grid, or varies depending on letter width. Choose fixed grid to conform to the spacing specified in the captions file more accurately. Choose proportional to make the text easier to read for closed captions.
1746
1893
  */
1747
1894
  TeletextSpacing?: BurninSubtitleTeletextSpacing | string;
1748
1895
  /**
1896
+ * @public
1749
1897
  * Specify the horizontal position of the captions, relative to the left side of the output in pixels. A value of 10 would result in the captions starting 10 pixels from the left of the output. If no explicit x_position is provided, the horizontal caption position will be determined by the alignment parameter.
1750
1898
  */
1751
1899
  XPosition?: number;
1752
1900
  /**
1901
+ * @public
1753
1902
  * Specify the vertical position of the captions, relative to the top of the output in pixels. A value of 10 would result in the captions starting 10 pixels from the top of the output. If no explicit y_position is provided, the caption will be positioned towards the bottom of the output.
1754
1903
  */
1755
1904
  YPosition?: number;
@@ -1935,112 +2084,139 @@ export type DvbSubtitleTeletextSpacing = (typeof DvbSubtitleTeletextSpacing)[key
1935
2084
  */
1936
2085
  export interface DvbSubDestinationSettings {
1937
2086
  /**
2087
+ * @public
1938
2088
  * Specify the alignment of your captions. If no explicit x_position is provided, setting alignment to centered will placethe captions at the bottom center of the output. Similarly, setting a left alignment willalign captions to the bottom left of the output. If x and y positions are given in conjunction with the alignment parameter, the font will be justified (either left or centered) relative to those coordinates. Within your job settings, all of your DVB-Sub settings must be identical.
1939
2089
  */
1940
2090
  Alignment?: DvbSubtitleAlignment | string;
1941
2091
  /**
2092
+ * @public
1942
2093
  * Ignore this setting unless Style Passthrough is set to Enabled and Font color set to Black, Yellow, Red, Green, Blue, or Hex. Use Apply font color for additional font color controls. When you choose White text only, or leave blank, your font color setting only applies to white text in your input captions. For example, if your font color setting is Yellow, and your input captions have red and white text, your output captions will have red and yellow text. When you choose ALL_TEXT, your font color setting applies to all of your output captions text.
1943
2094
  */
1944
2095
  ApplyFontColor?: DvbSubtitleApplyFontColor | string;
1945
2096
  /**
2097
+ * @public
1946
2098
  * Specify the color of the rectangle behind the captions. Leave background color blank and set Style passthrough to enabled to use the background color data from your input captions, if present.
1947
2099
  */
1948
2100
  BackgroundColor?: DvbSubtitleBackgroundColor | string;
1949
2101
  /**
2102
+ * @public
1950
2103
  * Specify the opacity of the background rectangle. Enter a value from 0 to 255, where 0 is transparent and 255 is opaque. If Style passthrough is set to enabled, leave blank to pass through the background style information in your input captions to your output captions. If Style passthrough is set to disabled, leave blank to use a value of 0 and remove all backgrounds from your output captions. Within your job settings, all of your DVB-Sub settings must be identical.
1951
2104
  */
1952
2105
  BackgroundOpacity?: number;
1953
2106
  /**
2107
+ * @public
1954
2108
  * Specify how MediaConvert handles the display definition segment (DDS). To exclude the DDS from this set of captions: Keep the default, None. To include the DDS: Choose Specified. When you do, also specify the offset coordinates of the display window with DDS x-coordinate and DDS y-coordinate. To include the DDS, but not include display window data: Choose No display window. When you do, you can write position metadata to the page composition segment (PCS) with DDS x-coordinate and DDS y-coordinate. For video resolutions with a height of 576 pixels or less, MediaConvert doesn't include the DDS, regardless of the value you choose for DDS handling. All burn-in and DVB-Sub font settings must match.
1955
2109
  */
1956
2110
  DdsHandling?: DvbddsHandling | string;
1957
2111
  /**
2112
+ * @public
1958
2113
  * Use this setting, along with DDS y-coordinate, to specify the upper left corner of the display definition segment (DDS) display window. With this setting, specify the distance, in pixels, between the left side of the frame and the left side of the DDS display window. Keep the default value, 0, to have MediaConvert automatically choose this offset. Related setting: When you use this setting, you must set DDS handling to a value other than None. MediaConvert uses these values to determine whether to write page position data to the DDS or to the page composition segment. All burn-in and DVB-Sub font settings must match.
1959
2114
  */
1960
2115
  DdsXCoordinate?: number;
1961
2116
  /**
2117
+ * @public
1962
2118
  * Use this setting, along with DDS x-coordinate, to specify the upper left corner of the display definition segment (DDS) display window. With this setting, specify the distance, in pixels, between the top of the frame and the top of the DDS display window. Keep the default value, 0, to have MediaConvert automatically choose this offset. Related setting: When you use this setting, you must set DDS handling to a value other than None. MediaConvert uses these values to determine whether to write page position data to the DDS or to the page composition segment (PCS). All burn-in and DVB-Sub font settings must match.
1963
2119
  */
1964
2120
  DdsYCoordinate?: number;
1965
2121
  /**
2122
+ * @public
1966
2123
  * Specify the font that you want the service to use for your burn in captions when your input captions specify a font that MediaConvert doesn't support. When you set Fallback font to best match, or leave blank, MediaConvert uses a supported font that most closely matches the font that your input captions specify. When there are multiple unsupported fonts in your input captions, MediaConvert matches each font with the supported font that matches best. When you explicitly choose a replacement font, MediaConvert uses that font to replace all unsupported fonts from your input.
1967
2124
  */
1968
2125
  FallbackFont?: DvbSubSubtitleFallbackFont | string;
1969
2126
  /**
2127
+ * @public
1970
2128
  * Specify the color of the captions text. Leave Font color blank and set Style passthrough to enabled to use the font color data from your input captions, if present. Within your job settings, all of your DVB-Sub settings must be identical.
1971
2129
  */
1972
2130
  FontColor?: DvbSubtitleFontColor | string;
1973
2131
  /**
2132
+ * @public
1974
2133
  * Specify the opacity of the burned-in captions. 255 is opaque; 0 is transparent.
1975
2134
  * Within your job settings, all of your DVB-Sub settings must be identical.
1976
2135
  */
1977
2136
  FontOpacity?: number;
1978
2137
  /**
2138
+ * @public
1979
2139
  * Specify the Font resolution in DPI (dots per inch).
1980
2140
  * Within your job settings, all of your DVB-Sub settings must be identical.
1981
2141
  */
1982
2142
  FontResolution?: number;
1983
2143
  /**
2144
+ * @public
1984
2145
  * Set Font script to Automatically determined, or leave blank, to automatically determine the font script in your input captions. Otherwise, set to Simplified Chinese (HANS) or Traditional Chinese (HANT) if your input font script uses Simplified or Traditional Chinese. Within your job settings, all of your DVB-Sub settings must be identical.
1985
2146
  */
1986
2147
  FontScript?: FontScript | string;
1987
2148
  /**
2149
+ * @public
1988
2150
  * Specify the Font size in pixels. Must be a positive integer. Set to 0, or leave blank, for automatic font size. Within your job settings, all of your DVB-Sub settings must be identical.
1989
2151
  */
1990
2152
  FontSize?: number;
1991
2153
  /**
2154
+ * @public
1992
2155
  * Specify the height, in pixels, of this set of DVB-Sub captions. The default value is 576 pixels. Related setting: When you use this setting, you must set DDS handling to a value other than None. All burn-in and DVB-Sub font settings must match.
1993
2156
  */
1994
2157
  Height?: number;
1995
2158
  /**
2159
+ * @public
1996
2160
  * Ignore this setting unless your Font color is set to Hex. Enter either six or eight hexidecimal digits, representing red, green, and blue, with two optional extra digits for alpha. For example a value of 1122AABB is a red value of 0x11, a green value of 0x22, a blue value of 0xAA, and an alpha value of 0xBB.
1997
2161
  */
1998
2162
  HexFontColor?: string;
1999
2163
  /**
2164
+ * @public
2000
2165
  * Specify font outline color. Leave Outline color blank and set Style passthrough to enabled to use the font outline color data from your input captions, if present. Within your job settings, all of your DVB-Sub settings must be identical.
2001
2166
  */
2002
2167
  OutlineColor?: DvbSubtitleOutlineColor | string;
2003
2168
  /**
2169
+ * @public
2004
2170
  * Specify the Outline size of the caption text, in pixels. Leave Outline size blank and set Style passthrough to enabled to use the outline size data from your input captions, if present. Within your job settings, all of your DVB-Sub settings must be identical.
2005
2171
  */
2006
2172
  OutlineSize?: number;
2007
2173
  /**
2174
+ * @public
2008
2175
  * Specify the color of the shadow cast by the captions. Leave Shadow color blank and set Style passthrough to enabled to use the shadow color data from your input captions, if present. Within your job settings, all of your DVB-Sub settings must be identical.
2009
2176
  */
2010
2177
  ShadowColor?: DvbSubtitleShadowColor | string;
2011
2178
  /**
2179
+ * @public
2012
2180
  * Specify the opacity of the shadow. Enter a value from 0 to 255, where 0 is transparent and 255 is opaque. If Style passthrough is set to Enabled, leave Shadow opacity blank to pass through the shadow style information in your input captions to your output captions. If Style passthrough is set to disabled, leave blank to use a value of 0 and remove all shadows from your output captions. Within your job settings, all of your DVB-Sub settings must be identical.
2013
2181
  */
2014
2182
  ShadowOpacity?: number;
2015
2183
  /**
2184
+ * @public
2016
2185
  * Specify the horizontal offset of the shadow, relative to the captions in pixels. A value of -2 would result in a shadow offset 2 pixels to the left. Within your job settings, all of your DVB-Sub settings must be identical.
2017
2186
  */
2018
2187
  ShadowXOffset?: number;
2019
2188
  /**
2189
+ * @public
2020
2190
  * Specify the vertical offset of the shadow relative to the captions in pixels. A value of -2 would result in a shadow offset 2 pixels above the text. Leave Shadow y-offset blank and set Style passthrough to enabled to use the shadow y-offset data from your input captions, if present. Within your job settings, all of your DVB-Sub settings must be identical.
2021
2191
  */
2022
2192
  ShadowYOffset?: number;
2023
2193
  /**
2194
+ * @public
2024
2195
  * Set Style passthrough to ENABLED to use the available style, color, and position information from your input captions. MediaConvert uses default settings for any missing style and position information in your input captions. Set Style passthrough to DISABLED, or leave blank, to ignore the style and position information from your input captions and use default settings: white text with black outlining, bottom-center positioning, and automatic sizing. Whether you set Style passthrough to enabled or not, you can also choose to manually override any of the individual style and position settings.
2025
2196
  */
2026
2197
  StylePassthrough?: DvbSubtitleStylePassthrough | string;
2027
2198
  /**
2199
+ * @public
2028
2200
  * Specify whether your DVB subtitles are standard or for hearing impaired. Choose hearing impaired if your subtitles include audio descriptions and dialogue. Choose standard if your subtitles include only dialogue.
2029
2201
  */
2030
2202
  SubtitlingType?: DvbSubtitlingType | string;
2031
2203
  /**
2204
+ * @public
2032
2205
  * Specify whether the Text spacing in your captions is set by the captions grid, or varies depending on letter width. Choose fixed grid to conform to the spacing specified in the captions file more accurately. Choose proportional to make the text easier to read for closed captions. Within your job settings, all of your DVB-Sub settings must be identical.
2033
2206
  */
2034
2207
  TeletextSpacing?: DvbSubtitleTeletextSpacing | string;
2035
2208
  /**
2209
+ * @public
2036
2210
  * Specify the width, in pixels, of this set of DVB-Sub captions. The default value is 720 pixels. Related setting: When you use this setting, you must set DDS handling to a value other than None. All burn-in and DVB-Sub font settings must match.
2037
2211
  */
2038
2212
  Width?: number;
2039
2213
  /**
2214
+ * @public
2040
2215
  * Specify the horizontal position of the captions, relative to the left side of the output in pixels. A value of 10 would result in the captions starting 10 pixels from the left of the output. If no explicit x_position is provided, the horizontal caption position will be determined by the alignment parameter. Within your job settings, all of your DVB-Sub settings must be identical.
2041
2216
  */
2042
2217
  XPosition?: number;
2043
2218
  /**
2219
+ * @public
2044
2220
  * Specify the vertical position of the captions, relative to the top of the output in pixels. A value of 10 would result in the captions starting 10 pixels from the top of the output. If no explicit y_position is provided, the caption will be positioned towards the bottom of the output. Within your job settings, all of your DVB-Sub settings must be identical.
2045
2221
  */
2046
2222
  YPosition?: number;
@@ -2051,10 +2227,12 @@ export interface DvbSubDestinationSettings {
2051
2227
  */
2052
2228
  export interface EmbeddedDestinationSettings {
2053
2229
  /**
2230
+ * @public
2054
2231
  * Ignore this setting unless your input captions are SCC format and your output captions are embedded in the video stream. Specify a CC number for each captions channel in this output. If you have two channels, choose CC numbers that aren't in the same field. For example, choose 1 and 3. For more information, see https://docs.aws.amazon.com/console/mediaconvert/dual-scc-to-embedded.
2055
2232
  */
2056
2233
  Destination608ChannelNumber?: number;
2057
2234
  /**
2235
+ * @public
2058
2236
  * Ignore this setting unless your input captions are SCC format and you want both 608 and 708 captions embedded in your output stream. Optionally, specify the 708 service number for each output captions channel. Choose a different number for each channel. To use this setting, also set Force 608 to 708 upconvert to Upconvert in your input captions selector settings. If you choose to upconvert but don't specify a 708 service number, MediaConvert uses the number that you specify for CC channel number for the 708 service number. For more information, see https://docs.aws.amazon.com/console/mediaconvert/dual-scc-to-embedded.
2059
2237
  */
2060
2238
  Destination708ServiceNumber?: number;
@@ -2089,10 +2267,12 @@ export type ImscStylePassthrough = (typeof ImscStylePassthrough)[keyof typeof Im
2089
2267
  */
2090
2268
  export interface ImscDestinationSettings {
2091
2269
  /**
2270
+ * @public
2092
2271
  * If the IMSC captions track is intended to provide accessibility for people who are deaf or hard of hearing: Set Accessibility subtitles to Enabled. When you do, MediaConvert adds accessibility attributes to your output HLS or DASH manifest. For HLS manifests, MediaConvert adds the following accessibility attributes under EXT-X-MEDIA for this track: CHARACTERISTICS="public.accessibility.describes-spoken-dialog,public.accessibility.describes-music-and-sound" and AUTOSELECT="YES". For DASH manifests, MediaConvert adds the following in the adaptation set for this track: <Accessibility schemeIdUri="urn:mpeg:dash:role:2011" value="caption"/>. If the captions track is not intended to provide such accessibility: Keep the default value, Disabled. When you do, for DASH manifests, MediaConvert instead adds the following in the adaptation set for this track: <Role schemeIDUri="urn:mpeg:dash:role:2011" value="subtitle"/>.
2093
2272
  */
2094
2273
  Accessibility?: ImscAccessibilitySubs | string;
2095
2274
  /**
2275
+ * @public
2096
2276
  * Keep this setting enabled to have MediaConvert use the font style and position information from the captions source in the output. This option is available only when your input captions are IMSC, SMPTE-TT, or TTML. Disable this setting for simplified output captions.
2097
2277
  */
2098
2278
  StylePassthrough?: ImscStylePassthrough | string;
@@ -2118,6 +2298,7 @@ export type SccDestinationFramerate = (typeof SccDestinationFramerate)[keyof typ
2118
2298
  */
2119
2299
  export interface SccDestinationSettings {
2120
2300
  /**
2301
+ * @public
2121
2302
  * Set Framerate to make sure that the captions and the video are synchronized in the output. Specify a frame rate that matches the frame rate of the associated video. If the video frame rate is 29.97, choose 29.97 dropframe only if the video has video_insertion=true and drop_frame_timecode=true; otherwise, choose 29.97 non-dropframe.
2122
2303
  */
2123
2304
  Framerate?: SccDestinationFramerate | string;
@@ -2140,6 +2321,7 @@ export type SrtStylePassthrough = (typeof SrtStylePassthrough)[keyof typeof SrtS
2140
2321
  */
2141
2322
  export interface SrtDestinationSettings {
2142
2323
  /**
2324
+ * @public
2143
2325
  * Set Style passthrough to ENABLED to use the available style, color, and position information from your input captions. MediaConvert uses default settings for any missing style and position information in your input captions. Set Style passthrough to DISABLED, or leave blank, to ignore the style and position information from your input captions and use simplified output captions.
2144
2326
  */
2145
2327
  StylePassthrough?: SrtStylePassthrough | string;
@@ -2165,10 +2347,12 @@ export type TeletextPageType = (typeof TeletextPageType)[keyof typeof TeletextPa
2165
2347
  */
2166
2348
  export interface TeletextDestinationSettings {
2167
2349
  /**
2350
+ * @public
2168
2351
  * Set pageNumber to the Teletext page number for the destination captions for this output. This value must be a three-digit hexadecimal string; strings ending in -FF are invalid. If you are passing through the entire set of Teletext data, do not use this field.
2169
2352
  */
2170
2353
  PageNumber?: string;
2171
2354
  /**
2355
+ * @public
2172
2356
  * Specify the page types for this Teletext page. If you don't specify a value here, the service sets the page type to the default value Subtitle. If you pass through the entire set of Teletext data, don't use this field. When you pass through a set of Teletext pages, your output has the same page types as your input.
2173
2357
  */
2174
2358
  PageTypes?: (TeletextPageType | string)[];
@@ -2191,6 +2375,7 @@ export type TtmlStylePassthrough = (typeof TtmlStylePassthrough)[keyof typeof Tt
2191
2375
  */
2192
2376
  export interface TtmlDestinationSettings {
2193
2377
  /**
2378
+ * @public
2194
2379
  * Pass through style and position information from a TTML-like input source (TTML, IMSC, SMPTE-TT) to the TTML output.
2195
2380
  */
2196
2381
  StylePassthrough?: TtmlStylePassthrough | string;
@@ -2226,10 +2411,12 @@ export type WebvttStylePassthrough = (typeof WebvttStylePassthrough)[keyof typeo
2226
2411
  */
2227
2412
  export interface WebvttDestinationSettings {
2228
2413
  /**
2414
+ * @public
2229
2415
  * If the WebVTT captions track is intended to provide accessibility for people who are deaf or hard of hearing: Set Accessibility subtitles to Enabled. When you do, MediaConvert adds accessibility attributes to your output HLS or DASH manifest. For HLS manifests, MediaConvert adds the following accessibility attributes under EXT-X-MEDIA for this track: CHARACTERISTICS="public.accessibility.describes-spoken-dialog,public.accessibility.describes-music-and-sound" and AUTOSELECT="YES". For DASH manifests, MediaConvert adds the following in the adaptation set for this track: <Accessibility schemeIdUri="urn:mpeg:dash:role:2011" value="caption"/>. If the captions track is not intended to provide such accessibility: Keep the default value, Disabled. When you do, for DASH manifests, MediaConvert instead adds the following in the adaptation set for this track: <Role schemeIDUri="urn:mpeg:dash:role:2011" value="subtitle"/>.
2230
2416
  */
2231
2417
  Accessibility?: WebvttAccessibilitySubs | string;
2232
2418
  /**
2419
+ * @public
2233
2420
  * To use the available style, color, and position information from your input captions: Set Style passthrough to Enabled. MediaConvert uses default settings when style and position information is missing from your input captions. To recreate the input captions exactly: Set Style passthrough to Strict. MediaConvert automatically applies timing adjustments, including adjustments for frame rate conversion, ad avails, and input clipping. Your input captions format must be WebVTT. To ignore the style and position information from your input captions and use simplified output captions: Set Style passthrough to Disabled, or leave blank.
2234
2421
  */
2235
2422
  StylePassthrough?: WebvttStylePassthrough | string;
@@ -2240,42 +2427,52 @@ export interface WebvttDestinationSettings {
2240
2427
  */
2241
2428
  export interface CaptionDestinationSettings {
2242
2429
  /**
2430
+ * @public
2243
2431
  * Burn-in is a captions delivery method, rather than a captions format. Burn-in writes the captions directly on your video frames, replacing pixels of video content with the captions. Set up burn-in captions in the same output as your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/burn-in-output-captions.html.
2244
2432
  */
2245
2433
  BurninDestinationSettings?: BurninDestinationSettings;
2246
2434
  /**
2435
+ * @public
2247
2436
  * Specify the format for this set of captions on this output. The default format is embedded without SCTE-20. Note that your choice of video output container constrains your choice of output captions format. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/captions-support-tables.html. If you are using SCTE-20 and you want to create an output that complies with the SCTE-43 spec, choose SCTE-20 plus embedded. To create a non-compliant output where the embedded captions come first, choose Embedded plus SCTE-20.
2248
2437
  */
2249
2438
  DestinationType?: CaptionDestinationType | string;
2250
2439
  /**
2440
+ * @public
2251
2441
  * Settings related to DVB-Sub captions. Set up DVB-Sub captions in the same output as your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/dvb-sub-output-captions.html.
2252
2442
  */
2253
2443
  DvbSubDestinationSettings?: DvbSubDestinationSettings;
2254
2444
  /**
2445
+ * @public
2255
2446
  * Settings related to CEA/EIA-608 and CEA/EIA-708 (also called embedded or ancillary) captions. Set up embedded captions in the same output as your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/embedded-output-captions.html.
2256
2447
  */
2257
2448
  EmbeddedDestinationSettings?: EmbeddedDestinationSettings;
2258
2449
  /**
2450
+ * @public
2259
2451
  * Settings related to IMSC captions. IMSC is a sidecar format that holds captions in a file that is separate from the video container. Set up sidecar captions in the same output group, but different output from your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/ttml-and-webvtt-output-captions.html.
2260
2452
  */
2261
2453
  ImscDestinationSettings?: ImscDestinationSettings;
2262
2454
  /**
2455
+ * @public
2263
2456
  * Settings related to SCC captions. SCC is a sidecar format that holds captions in a file that is separate from the video container. Set up sidecar captions in the same output group, but different output from your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/scc-srt-output-captions.html.
2264
2457
  */
2265
2458
  SccDestinationSettings?: SccDestinationSettings;
2266
2459
  /**
2460
+ * @public
2267
2461
  * Settings related to SRT captions. SRT is a sidecar format that holds captions in a file that is separate from the video container. Set up sidecar captions in the same output group, but different output from your video.
2268
2462
  */
2269
2463
  SrtDestinationSettings?: SrtDestinationSettings;
2270
2464
  /**
2465
+ * @public
2271
2466
  * Settings related to teletext captions. Set up teletext captions in the same output as your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/teletext-output-captions.html.
2272
2467
  */
2273
2468
  TeletextDestinationSettings?: TeletextDestinationSettings;
2274
2469
  /**
2470
+ * @public
2275
2471
  * Settings related to TTML captions. TTML is a sidecar format that holds captions in a file that is separate from the video container. Set up sidecar captions in the same output group, but different output from your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/ttml-and-webvtt-output-captions.html.
2276
2472
  */
2277
2473
  TtmlDestinationSettings?: TtmlDestinationSettings;
2278
2474
  /**
2475
+ * @public
2279
2476
  * Settings related to WebVTT captions. WebVTT is a sidecar format that holds captions in a file that is separate from the video container. Set up sidecar captions in the same output group, but different output from your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/ttml-and-webvtt-output-captions.html.
2280
2477
  */
2281
2478
  WebvttDestinationSettings?: WebvttDestinationSettings;
@@ -2286,22 +2483,27 @@ export interface CaptionDestinationSettings {
2286
2483
  */
2287
2484
  export interface CaptionDescription {
2288
2485
  /**
2486
+ * @public
2289
2487
  * Specifies which "Caption Selector":#inputs-caption_selector to use from each input when generating captions. The name should be of the format "Caption Selector <N>", which denotes that the Nth Caption Selector will be used from each input.
2290
2488
  */
2291
2489
  CaptionSelectorName?: string;
2292
2490
  /**
2491
+ * @public
2293
2492
  * Specify the language for this captions output track. For most captions output formats, the encoder puts this language information in the output captions metadata. If your output captions format is DVB-Sub or Burn in, the encoder uses this language information when automatically selecting the font script for rendering the captions text. For all outputs, you can use an ISO 639-2 or ISO 639-3 code. For streaming outputs, you can also use any other code in the full RFC-5646 specification. Streaming outputs are those that are in one of the following output groups: CMAF, DASH ISO, Apple HLS, or Microsoft Smooth Streaming.
2294
2493
  */
2295
2494
  CustomLanguageCode?: string;
2296
2495
  /**
2496
+ * @public
2297
2497
  * Settings related to one captions tab on the MediaConvert console. Usually, one captions tab corresponds to one output captions track. Depending on your output captions format, one tab might correspond to a set of output captions tracks. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/including-captions.html.
2298
2498
  */
2299
2499
  DestinationSettings?: CaptionDestinationSettings;
2300
2500
  /**
2501
+ * @public
2301
2502
  * Specify the language of this captions output track. For most captions output formats, the encoder puts this language information in the output captions metadata. If your output captions format is DVB-Sub or Burn in, the encoder uses this language information to choose the font language for rendering the captions text.
2302
2503
  */
2303
2504
  LanguageCode?: LanguageCode | string;
2304
2505
  /**
2506
+ * @public
2305
2507
  * Specify a label for this set of output captions. For example, "English", "Director commentary", or "track_2". For streaming outputs, MediaConvert passes this information into destination manifests for display on the end-viewer's player device. For outputs in other output groups, the service ignores this setting.
2306
2508
  */
2307
2509
  LanguageDescription?: string;
@@ -2312,18 +2514,22 @@ export interface CaptionDescription {
2312
2514
  */
2313
2515
  export interface CaptionDescriptionPreset {
2314
2516
  /**
2517
+ * @public
2315
2518
  * Specify the language for this captions output track. For most captions output formats, the encoder puts this language information in the output captions metadata. If your output captions format is DVB-Sub or Burn in, the encoder uses this language information when automatically selecting the font script for rendering the captions text. For all outputs, you can use an ISO 639-2 or ISO 639-3 code. For streaming outputs, you can also use any other code in the full RFC-5646 specification. Streaming outputs are those that are in one of the following output groups: CMAF, DASH ISO, Apple HLS, or Microsoft Smooth Streaming.
2316
2519
  */
2317
2520
  CustomLanguageCode?: string;
2318
2521
  /**
2522
+ * @public
2319
2523
  * Settings related to one captions tab on the MediaConvert console. Usually, one captions tab corresponds to one output captions track. Depending on your output captions format, one tab might correspond to a set of output captions tracks. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/including-captions.html.
2320
2524
  */
2321
2525
  DestinationSettings?: CaptionDestinationSettings;
2322
2526
  /**
2527
+ * @public
2323
2528
  * Specify the language of this captions output track. For most captions output formats, the encoder puts this language information in the output captions metadata. If your output captions format is DVB-Sub or Burn in, the encoder uses this language information to choose the font language for rendering the captions text.
2324
2529
  */
2325
2530
  LanguageCode?: LanguageCode | string;
2326
2531
  /**
2532
+ * @public
2327
2533
  * Specify a label for this set of output captions. For example, "English", "Director commentary", or "track_2". For streaming outputs, MediaConvert passes this information into destination manifests for display on the end-viewer's player device. For outputs in other output groups, the service ignores this setting.
2328
2534
  */
2329
2535
  LanguageDescription?: string;
@@ -2334,10 +2540,12 @@ export interface CaptionDescriptionPreset {
2334
2540
  */
2335
2541
  export interface CmafAdditionalManifest {
2336
2542
  /**
2543
+ * @public
2337
2544
  * Specify a name modifier that the service adds to the name of this manifest to make it different from the file names of the other main manifests in the output group. For example, say that the default main manifest for your HLS group is film-name.m3u8. If you enter "-no-premium" for this setting, then the file name the service generates for this top-level manifest is film-name-no-premium.m3u8. For HLS output groups, specify a manifestNameModifier that is different from the nameModifier of the output. The service uses the output name modifier to create unique names for the individual variant manifests.
2338
2545
  */
2339
2546
  ManifestNameModifier?: string;
2340
2547
  /**
2548
+ * @public
2341
2549
  * Specify the outputs that you want this additional top-level manifest to reference.
2342
2550
  */
2343
2551
  SelectedOutputs?: string[];
@@ -2348,10 +2556,12 @@ export interface CmafAdditionalManifest {
2348
2556
  */
2349
2557
  export interface DashAdditionalManifest {
2350
2558
  /**
2559
+ * @public
2351
2560
  * Specify a name modifier that the service adds to the name of this manifest to make it different from the file names of the other main manifests in the output group. For example, say that the default main manifest for your DASH group is film-name.mpd. If you enter "-no-premium" for this setting, then the file name the service generates for this top-level manifest is film-name-no-premium.mpd.
2352
2561
  */
2353
2562
  ManifestNameModifier?: string;
2354
2563
  /**
2564
+ * @public
2355
2565
  * Specify the outputs that you want this additional top-level manifest to reference.
2356
2566
  */
2357
2567
  SelectedOutputs?: string[];
@@ -2362,6 +2572,7 @@ export interface DashAdditionalManifest {
2362
2572
  */
2363
2573
  export interface Endpoint {
2364
2574
  /**
2575
+ * @public
2365
2576
  * URL of endpoint
2366
2577
  */
2367
2578
  Url?: string;
@@ -2372,10 +2583,12 @@ export interface Endpoint {
2372
2583
  */
2373
2584
  export interface HlsAdditionalManifest {
2374
2585
  /**
2586
+ * @public
2375
2587
  * Specify a name modifier that the service adds to the name of this manifest to make it different from the file names of the other main manifests in the output group. For example, say that the default main manifest for your HLS group is film-name.m3u8. If you enter "-no-premium" for this setting, then the file name the service generates for this top-level manifest is film-name-no-premium.m3u8. For HLS output groups, specify a manifestNameModifier that is different from the nameModifier of the output. The service uses the output name modifier to create unique names for the individual variant manifests.
2376
2588
  */
2377
2589
  ManifestNameModifier?: string;
2378
2590
  /**
2591
+ * @public
2379
2592
  * Specify the outputs that you want this additional top-level manifest to reference.
2380
2593
  */
2381
2594
  SelectedOutputs?: string[];
@@ -2398,18 +2611,22 @@ export type HlsAdMarkers = (typeof HlsAdMarkers)[keyof typeof HlsAdMarkers];
2398
2611
  */
2399
2612
  export interface HlsCaptionLanguageMapping {
2400
2613
  /**
2614
+ * @public
2401
2615
  * Caption channel.
2402
2616
  */
2403
2617
  CaptionChannel?: number;
2404
2618
  /**
2619
+ * @public
2405
2620
  * Specify the language for this captions channel, using the ISO 639-2 or ISO 639-3 three-letter language code
2406
2621
  */
2407
2622
  CustomLanguageCode?: string;
2408
2623
  /**
2624
+ * @public
2409
2625
  * Specify the language, using the ISO 639-2 three-letter code listed at https://www.loc.gov/standards/iso639-2/php/code_list.php.
2410
2626
  */
2411
2627
  LanguageCode?: LanguageCode | string;
2412
2628
  /**
2629
+ * @public
2413
2630
  * Caption language description.
2414
2631
  */
2415
2632
  LanguageDescription?: string;
@@ -2420,14 +2637,17 @@ export interface HlsCaptionLanguageMapping {
2420
2637
  */
2421
2638
  export interface HopDestination {
2422
2639
  /**
2640
+ * @public
2423
2641
  * Optional. When you set up a job to use queue hopping, you can specify a different relative priority for the job in the destination queue. If you don't specify, the relative priority will remain the same as in the previous queue.
2424
2642
  */
2425
2643
  Priority?: number;
2426
2644
  /**
2645
+ * @public
2427
2646
  * Optional unless the job is submitted on the default queue. When you set up a job to use queue hopping, you can specify a destination queue. This queue cannot be the original queue to which the job is submitted. If the original queue isn't the default queue and you don't specify the destination queue, the job will move to the default queue.
2428
2647
  */
2429
2648
  Queue?: string;
2430
2649
  /**
2650
+ * @public
2431
2651
  * Required for setting up a job to use queue hopping. Minimum wait time in minutes until the job can hop to the destination queue. Valid range is 1 to 4320 minutes, inclusive.
2432
2652
  */
2433
2653
  WaitMinutes?: number;
@@ -2438,10 +2658,12 @@ export interface HopDestination {
2438
2658
  */
2439
2659
  export interface Id3Insertion {
2440
2660
  /**
2661
+ * @public
2441
2662
  * Use ID3 tag to provide a fully formed ID3 tag in base64-encode format.
2442
2663
  */
2443
2664
  Id3?: string;
2444
2665
  /**
2666
+ * @public
2445
2667
  * Provide a Timecode in HH:MM:SS:FF or HH:MM:SS;FF format.
2446
2668
  */
2447
2669
  Timecode?: string;
@@ -2489,10 +2711,12 @@ export type AdvancedInputFilterSharpen = (typeof AdvancedInputFilterSharpen)[key
2489
2711
  */
2490
2712
  export interface AdvancedInputFilterSettings {
2491
2713
  /**
2714
+ * @public
2492
2715
  * Add texture and detail to areas of your input video content that were lost after applying the Advanced input filter. To adaptively add texture and reduce softness: Choose Enabled. To not add any texture: Keep the default value, Disabled. We recommend that you choose Disabled for input video content that doesn't have texture, including screen recordings, computer graphics, or cartoons.
2493
2716
  */
2494
2717
  AddTexture?: AdvancedInputFilterAddTexture | string;
2495
2718
  /**
2719
+ * @public
2496
2720
  * Optionally specify the amount of sharpening to apply when you use the Advanced input filter. Sharpening adds contrast to the edges of your video content and can reduce softness. To apply no sharpening: Keep the default value, Off. To apply a minimal amount of sharpening choose Low, or for the maximum choose High.
2497
2721
  */
2498
2722
  Sharpening?: AdvancedInputFilterSharpen | string;
@@ -2503,6 +2727,7 @@ export interface AdvancedInputFilterSettings {
2503
2727
  */
2504
2728
  export interface AudioSelectorGroup {
2505
2729
  /**
2730
+ * @public
2506
2731
  * Name of an Audio Selector within the same input to include in the group. Audio selector names are standardized, based on their order within the input (e.g., "Audio Selector 1"). The audio selector name parameter can be repeated to add any number of audio selectors to the group.
2507
2732
  */
2508
2733
  AudioSelectorNames?: string[];
@@ -2539,14 +2764,17 @@ export type AudioDefaultSelection = (typeof AudioDefaultSelection)[keyof typeof
2539
2764
  */
2540
2765
  export interface HlsRenditionGroupSettings {
2541
2766
  /**
2767
+ * @public
2542
2768
  * Optional. Specify alternative group ID
2543
2769
  */
2544
2770
  RenditionGroupId?: string;
2545
2771
  /**
2772
+ * @public
2546
2773
  * Optional. Specify ISO 639-2 or ISO 639-3 code in the language property
2547
2774
  */
2548
2775
  RenditionLanguageCode?: LanguageCode | string;
2549
2776
  /**
2777
+ * @public
2550
2778
  * Optional. Specify media name
2551
2779
  */
2552
2780
  RenditionName?: string;
@@ -2571,50 +2799,62 @@ export type AudioSelectorType = (typeof AudioSelectorType)[keyof typeof AudioSel
2571
2799
  */
2572
2800
  export interface AudioSelector {
2573
2801
  /**
2802
+ * @public
2574
2803
  * Apply audio timing corrections to help synchronize audio and video in your output. To apply timing corrections, your input must meet the following requirements: * Container: MP4, or MOV, with an accurate time-to-sample (STTS) table. * Audio track: AAC. Choose from the following audio timing correction settings: * Disabled (Default): Apply no correction. * Auto: Recommended for most inputs. MediaConvert analyzes the audio timing in your input and determines which correction setting to use, if needed. * Track: Adjust the duration of each audio frame by a constant amount to align the audio track length with STTS duration. Track-level correction does not affect pitch, and is recommended for tonal audio content such as music. * Frame: Adjust the duration of each audio frame by a variable amount to align audio frames with STTS timestamps. No corrections are made to already-aligned frames. Frame-level correction may affect the pitch of corrected frames, and is recommended for atonal audio content such as speech or percussion.
2575
2804
  */
2576
2805
  AudioDurationCorrection?: AudioDurationCorrection | string;
2577
2806
  /**
2807
+ * @public
2578
2808
  * Selects a specific language code from within an audio source, using the ISO 639-2 or ISO 639-3 three-letter language code
2579
2809
  */
2580
2810
  CustomLanguageCode?: string;
2581
2811
  /**
2812
+ * @public
2582
2813
  * Enable this setting on one audio selector to set it as the default for the job. The service uses this default for outputs where it can't find the specified input audio. If you don't set a default, those outputs have no audio.
2583
2814
  */
2584
2815
  DefaultSelection?: AudioDefaultSelection | string;
2585
2816
  /**
2817
+ * @public
2586
2818
  * Specifies audio data from an external file source.
2587
2819
  */
2588
2820
  ExternalAudioFileInput?: string;
2589
2821
  /**
2822
+ * @public
2590
2823
  * Settings specific to audio sources in an HLS alternate rendition group. Specify the properties (renditionGroupId, renditionName or renditionLanguageCode) to identify the unique audio track among the alternative rendition groups present in the HLS manifest. If no unique track is found, or multiple tracks match the properties provided, the job fails. If no properties in hlsRenditionGroupSettings are specified, the default audio track within the video segment is chosen. If there is no audio within video segment, the alternative audio with DEFAULT=YES is chosen instead.
2591
2824
  */
2592
2825
  HlsRenditionGroupSettings?: HlsRenditionGroupSettings;
2593
2826
  /**
2827
+ * @public
2594
2828
  * Selects a specific language code from within an audio source.
2595
2829
  */
2596
2830
  LanguageCode?: LanguageCode | string;
2597
2831
  /**
2832
+ * @public
2598
2833
  * Specifies a time delta in milliseconds to offset the audio from the input video.
2599
2834
  */
2600
2835
  Offset?: number;
2601
2836
  /**
2837
+ * @public
2602
2838
  * Selects a specific PID from within an audio source (e.g. 257 selects PID 0x101).
2603
2839
  */
2604
2840
  Pids?: number[];
2605
2841
  /**
2842
+ * @public
2606
2843
  * Use this setting for input streams that contain Dolby E, to have the service extract specific program data from the track. To select multiple programs, create multiple selectors with the same Track and different Program numbers. In the console, this setting is visible when you set Selector type to Track. Choose the program number from the dropdown list. If your input file has incorrect metadata, you can choose All channels instead of a program number to have the service ignore the program IDs and include all the programs in the track.
2607
2844
  */
2608
2845
  ProgramSelection?: number;
2609
2846
  /**
2847
+ * @public
2610
2848
  * Use these settings to reorder the audio channels of one input to match those of another input. This allows you to combine the two files into a single output, one after the other.
2611
2849
  */
2612
2850
  RemixSettings?: RemixSettings;
2613
2851
  /**
2852
+ * @public
2614
2853
  * Specifies the type of the audio selector.
2615
2854
  */
2616
2855
  SelectorType?: AudioSelectorType | string;
2617
2856
  /**
2857
+ * @public
2618
2858
  * Identify a track from the input audio to include in this selector by entering the track index number. To include several tracks in a single audio selector, specify multiple tracks as follows. Using the console, enter a comma-separated list. For example, type "1,2,3" to include tracks 1 through 3.
2619
2859
  */
2620
2860
  Tracks?: number[];
@@ -2649,14 +2889,17 @@ export type AncillaryTerminateCaptions = (typeof AncillaryTerminateCaptions)[key
2649
2889
  */
2650
2890
  export interface AncillarySourceSettings {
2651
2891
  /**
2892
+ * @public
2652
2893
  * Specify whether this set of input captions appears in your outputs in both 608 and 708 format. If you choose Upconvert, MediaConvert includes the captions data in two ways: it passes the 608 data through using the 608 compatibility bytes fields of the 708 wrapper, and it also translates the 608 data into 708.
2653
2894
  */
2654
2895
  Convert608To708?: AncillaryConvert608To708 | string;
2655
2896
  /**
2897
+ * @public
2656
2898
  * Specifies the 608 channel number in the ancillary data track from which to extract captions. Unused for passthrough.
2657
2899
  */
2658
2900
  SourceAncillaryChannelNumber?: number;
2659
2901
  /**
2902
+ * @public
2660
2903
  * By default, the service terminates any unterminated captions at the end of each input. If you want the caption to continue onto your next input, disable this setting.
2661
2904
  */
2662
2905
  TerminateCaptions?: AncillaryTerminateCaptions | string;
@@ -2667,6 +2910,7 @@ export interface AncillarySourceSettings {
2667
2910
  */
2668
2911
  export interface DvbSubSourceSettings {
2669
2912
  /**
2913
+ * @public
2670
2914
  * When using DVB-Sub with Burn-in, use this PID for the source content. Unused for DVB-Sub passthrough. All DVB-Sub content is passed through, regardless of selectors.
2671
2915
  */
2672
2916
  Pid?: number;
@@ -2701,18 +2945,22 @@ export type EmbeddedTerminateCaptions = (typeof EmbeddedTerminateCaptions)[keyof
2701
2945
  */
2702
2946
  export interface EmbeddedSourceSettings {
2703
2947
  /**
2948
+ * @public
2704
2949
  * Specify whether this set of input captions appears in your outputs in both 608 and 708 format. If you choose Upconvert, MediaConvert includes the captions data in two ways: it passes the 608 data through using the 608 compatibility bytes fields of the 708 wrapper, and it also translates the 608 data into 708.
2705
2950
  */
2706
2951
  Convert608To708?: EmbeddedConvert608To708 | string;
2707
2952
  /**
2953
+ * @public
2708
2954
  * Specifies the 608/708 channel number within the video track from which to extract captions. Unused for passthrough.
2709
2955
  */
2710
2956
  Source608ChannelNumber?: number;
2711
2957
  /**
2958
+ * @public
2712
2959
  * Specifies the video track index used for extracting captions. The system only supports one input video track, so this should always be set to '1'.
2713
2960
  */
2714
2961
  Source608TrackNumber?: number;
2715
2962
  /**
2963
+ * @public
2716
2964
  * By default, the service terminates any unterminated captions at the end of each input. If you want the caption to continue onto your next input, disable this setting.
2717
2965
  */
2718
2966
  TerminateCaptions?: EmbeddedTerminateCaptions | string;
@@ -2747,10 +2995,12 @@ export type CaptionSourceConvertPaintOnToPopOn = (typeof CaptionSourceConvertPai
2747
2995
  */
2748
2996
  export interface CaptionSourceFramerate {
2749
2997
  /**
2998
+ * @public
2750
2999
  * Specify the denominator of the fraction that represents the frame rate for the setting Caption source frame rate. Use this setting along with the setting Framerate numerator.
2751
3000
  */
2752
3001
  FramerateDenominator?: number;
2753
3002
  /**
3003
+ * @public
2754
3004
  * Specify the numerator of the fraction that represents the frame rate for the setting Caption source frame rate. Use this setting along with the setting Framerate denominator.
2755
3005
  */
2756
3006
  FramerateNumerator?: number;
@@ -2773,26 +3023,32 @@ export type FileSourceTimeDeltaUnits = (typeof FileSourceTimeDeltaUnits)[keyof t
2773
3023
  */
2774
3024
  export interface FileSourceSettings {
2775
3025
  /**
3026
+ * @public
2776
3027
  * Specify whether this set of input captions appears in your outputs in both 608 and 708 format. If you choose Upconvert, MediaConvert includes the captions data in two ways: it passes the 608 data through using the 608 compatibility bytes fields of the 708 wrapper, and it also translates the 608 data into 708.
2777
3028
  */
2778
3029
  Convert608To708?: FileSourceConvert608To708 | string;
2779
3030
  /**
3031
+ * @public
2780
3032
  * Choose the presentation style of your input SCC captions. To use the same presentation style as your input: Keep the default value, Disabled. To convert paint-on captions to pop-on: Choose Enabled. We also recommend that you choose Enabled if you notice additional repeated lines in your output captions.
2781
3033
  */
2782
3034
  ConvertPaintToPop?: CaptionSourceConvertPaintOnToPopOn | string;
2783
3035
  /**
3036
+ * @public
2784
3037
  * Ignore this setting unless your input captions format is SCC. To have the service compensate for differing frame rates between your input captions and input video, specify the frame rate of the captions file. Specify this value as a fraction. For example, you might specify 24 / 1 for 24 fps, 25 / 1 for 25 fps, 24000 / 1001 for 23.976 fps, or 30000 / 1001 for 29.97 fps.
2785
3038
  */
2786
3039
  Framerate?: CaptionSourceFramerate;
2787
3040
  /**
3041
+ * @public
2788
3042
  * External caption file used for loading captions. Accepted file extensions are 'scc', 'ttml', 'dfxp', 'stl', 'srt', 'xml', 'smi', 'webvtt', and 'vtt'.
2789
3043
  */
2790
3044
  SourceFile?: string;
2791
3045
  /**
3046
+ * @public
2792
3047
  * Optional. Use this setting when you need to adjust the sync between your sidecar captions and your video. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/time-delta-use-cases.html. Enter a positive or negative number to modify the times in the captions file. For example, type 15 to add 15 seconds to all the times in the captions file. Type -5 to subtract 5 seconds from the times in the captions file. You can optionally specify your time delta in milliseconds instead of seconds. When you do so, set the related setting, Time delta units to Milliseconds. Note that, when you specify a time delta for timecode-based caption sources, such as SCC and STL, and your time delta isn't a multiple of the input frame rate, MediaConvert snaps the captions to the nearest frame. For example, when your input video frame rate is 25 fps and you specify 1010ms for time delta, MediaConvert delays your captions by 1000 ms.
2793
3048
  */
2794
3049
  TimeDelta?: number;
2795
3050
  /**
3051
+ * @public
2796
3052
  * When you use the setting Time delta to adjust the sync between your sidecar captions and your video, use this setting to specify the units for the delta that you specify. When you don't specify a value for Time delta units, MediaConvert uses seconds by default.
2797
3053
  */
2798
3054
  TimeDeltaUnits?: FileSourceTimeDeltaUnits | string;
@@ -2827,6 +3083,7 @@ export type CaptionSourceType = (typeof CaptionSourceType)[keyof typeof CaptionS
2827
3083
  */
2828
3084
  export interface TeletextSourceSettings {
2829
3085
  /**
3086
+ * @public
2830
3087
  * Use Page Number to specify the three-digit hexadecimal page number that will be used for Teletext captions. Do not use this setting if you are passing through teletext from the input source to output.
2831
3088
  */
2832
3089
  PageNumber?: string;
@@ -2837,6 +3094,7 @@ export interface TeletextSourceSettings {
2837
3094
  */
2838
3095
  export interface TrackSourceSettings {
2839
3096
  /**
3097
+ * @public
2840
3098
  * Use this setting to select a single captions track from a source. Track numbers correspond to the order in the captions source file. For IMF sources, track numbering is based on the order that the captions appear in the CPL. For example, use 1 to select the captions asset that is listed first in the CPL. To include more than one captions track in your job outputs, create multiple input captions selectors. Specify one track per selector.
2841
3099
  */
2842
3100
  TrackNumber?: number;
@@ -2847,14 +3105,17 @@ export interface TrackSourceSettings {
2847
3105
  */
2848
3106
  export interface WebvttHlsSourceSettings {
2849
3107
  /**
3108
+ * @public
2850
3109
  * Optional. Specify alternative group ID
2851
3110
  */
2852
3111
  RenditionGroupId?: string;
2853
3112
  /**
3113
+ * @public
2854
3114
  * Optional. Specify ISO 639-2 or ISO 639-3 code in the language property
2855
3115
  */
2856
3116
  RenditionLanguageCode?: LanguageCode | string;
2857
3117
  /**
3118
+ * @public
2858
3119
  * Optional. Specify media name
2859
3120
  */
2860
3121
  RenditionName?: string;
@@ -2865,34 +3126,42 @@ export interface WebvttHlsSourceSettings {
2865
3126
  */
2866
3127
  export interface CaptionSourceSettings {
2867
3128
  /**
3129
+ * @public
2868
3130
  * Settings for ancillary captions source.
2869
3131
  */
2870
3132
  AncillarySourceSettings?: AncillarySourceSettings;
2871
3133
  /**
3134
+ * @public
2872
3135
  * DVB Sub Source Settings
2873
3136
  */
2874
3137
  DvbSubSourceSettings?: DvbSubSourceSettings;
2875
3138
  /**
3139
+ * @public
2876
3140
  * Settings for embedded captions Source
2877
3141
  */
2878
3142
  EmbeddedSourceSettings?: EmbeddedSourceSettings;
2879
3143
  /**
3144
+ * @public
2880
3145
  * If your input captions are SCC, SMI, SRT, STL, TTML, WebVTT, or IMSC 1.1 in an xml file, specify the URI of the input caption source file. If your caption source is IMSC in an IMF package, use TrackSourceSettings instead of FileSoureSettings.
2881
3146
  */
2882
3147
  FileSourceSettings?: FileSourceSettings;
2883
3148
  /**
3149
+ * @public
2884
3150
  * Use Source to identify the format of your input captions. The service cannot auto-detect caption format.
2885
3151
  */
2886
3152
  SourceType?: CaptionSourceType | string;
2887
3153
  /**
3154
+ * @public
2888
3155
  * Settings specific to Teletext caption sources, including Page number.
2889
3156
  */
2890
3157
  TeletextSourceSettings?: TeletextSourceSettings;
2891
3158
  /**
3159
+ * @public
2892
3160
  * Settings specific to caption sources that are specified by track number. Currently, this is only IMSC captions in an IMF package. If your caption source is IMSC 1.1 in a separate xml file, use FileSourceSettings instead of TrackSourceSettings.
2893
3161
  */
2894
3162
  TrackSourceSettings?: TrackSourceSettings;
2895
3163
  /**
3164
+ * @public
2896
3165
  * Settings specific to WebVTT sources in HLS alternative rendition group. Specify the properties (renditionGroupId, renditionName or renditionLanguageCode) to identify the unique subtitle track among the alternative rendition groups present in the HLS manifest. If no unique track is found, or multiple tracks match the specified properties, the job fails. If there is only one subtitle track in the rendition group, the settings can be left empty and the default subtitle track will be chosen. If your caption source is a sidecar file, use FileSourceSettings instead of WebvttHlsSourceSettings.
2897
3166
  */
2898
3167
  WebvttHlsSourceSettings?: WebvttHlsSourceSettings;
@@ -2903,14 +3172,17 @@ export interface CaptionSourceSettings {
2903
3172
  */
2904
3173
  export interface CaptionSelector {
2905
3174
  /**
3175
+ * @public
2906
3176
  * The specific language to extract from source, using the ISO 639-2 or ISO 639-3 three-letter language code. If input is SCTE-27, complete this field and/or PID to select the caption language to extract. If input is DVB-Sub and output is Burn-in, complete this field and/or PID to select the caption language to extract. If input is DVB-Sub that is being passed through, omit this field (and PID field); there is no way to extract a specific language with pass-through captions.
2907
3177
  */
2908
3178
  CustomLanguageCode?: string;
2909
3179
  /**
3180
+ * @public
2910
3181
  * The specific language to extract from source. If input is SCTE-27, complete this field and/or PID to select the caption language to extract. If input is DVB-Sub and output is Burn-in, complete this field and/or PID to select the caption language to extract. If input is DVB-Sub that is being passed through, omit this field (and PID field); there is no way to extract a specific language with pass-through captions.
2911
3182
  */
2912
3183
  LanguageCode?: LanguageCode | string;
2913
3184
  /**
3185
+ * @public
2914
3186
  * If your input captions are SCC, TTML, STL, SMI, SRT, or IMSC in an xml file, specify the URI of the input captions source file. If your input captions are IMSC in an IMF package, use TrackSourceSettings instead of FileSoureSettings.
2915
3187
  */
2916
3188
  SourceSettings?: CaptionSourceSettings;
@@ -2921,18 +3193,22 @@ export interface CaptionSelector {
2921
3193
  */
2922
3194
  export interface Rectangle {
2923
3195
  /**
3196
+ * @public
2924
3197
  * Height of rectangle in pixels. Specify only even numbers.
2925
3198
  */
2926
3199
  Height?: number;
2927
3200
  /**
3201
+ * @public
2928
3202
  * Width of rectangle in pixels. Specify only even numbers.
2929
3203
  */
2930
3204
  Width?: number;
2931
3205
  /**
3206
+ * @public
2932
3207
  * The distance, in pixels, between the rectangle and the left edge of the video frame. Specify only even numbers.
2933
3208
  */
2934
3209
  X?: number;
2935
3210
  /**
3211
+ * @public
2936
3212
  * The distance, in pixels, between the rectangle and the top edge of the video frame. Specify only even numbers.
2937
3213
  */
2938
3214
  Y?: number;
@@ -2968,18 +3244,22 @@ export type DecryptionMode = (typeof DecryptionMode)[keyof typeof DecryptionMode
2968
3244
  */
2969
3245
  export interface InputDecryptionSettings {
2970
3246
  /**
3247
+ * @public
2971
3248
  * Specify the encryption mode that you used to encrypt your input files.
2972
3249
  */
2973
3250
  DecryptionMode?: DecryptionMode | string;
2974
3251
  /**
3252
+ * @public
2975
3253
  * Warning! Don't provide your encryption key in plaintext. Your job settings could be intercepted, making your encrypted content vulnerable. Specify the encrypted version of the data key that you used to encrypt your content. The data key must be encrypted by AWS Key Management Service (KMS). The key can be 128, 192, or 256 bits.
2976
3254
  */
2977
3255
  EncryptedDecryptionKey?: string;
2978
3256
  /**
3257
+ * @public
2979
3258
  * Specify the initialization vector that you used when you encrypted your content before uploading it to Amazon S3. You can use a 16-byte initialization vector with any encryption mode. Or, you can use a 12-byte initialization vector with GCM or CTR. MediaConvert accepts only initialization vectors that are base64-encoded.
2980
3259
  */
2981
3260
  InitializationVector?: string;
2982
3261
  /**
3262
+ * @public
2983
3263
  * Specify the AWS Region for AWS Key Management Service (KMS) that you used to encrypt your data key, if that Region is different from the one you are using for AWS Elemental MediaConvert.
2984
3264
  */
2985
3265
  KmsKeyRegion?: string;
@@ -3015,46 +3295,57 @@ export type InputFilterEnable = (typeof InputFilterEnable)[keyof typeof InputFil
3015
3295
  */
3016
3296
  export interface InsertableImage {
3017
3297
  /**
3298
+ * @public
3018
3299
  * Specify the time, in milliseconds, for the image to remain on the output video. This duration includes fade-in time but not fade-out time.
3019
3300
  */
3020
3301
  Duration?: number;
3021
3302
  /**
3303
+ * @public
3022
3304
  * Specify the length of time, in milliseconds, between the Start time that you specify for the image insertion and the time that the image appears at full opacity. Full opacity is the level that you specify for the opacity setting. If you don't specify a value for Fade-in, the image will appear abruptly at the overlay start time.
3023
3305
  */
3024
3306
  FadeIn?: number;
3025
3307
  /**
3308
+ * @public
3026
3309
  * Specify the length of time, in milliseconds, between the end of the time that you have specified for the image overlay Duration and when the overlaid image has faded to total transparency. If you don't specify a value for Fade-out, the image will disappear abruptly at the end of the inserted image duration.
3027
3310
  */
3028
3311
  FadeOut?: number;
3029
3312
  /**
3313
+ * @public
3030
3314
  * Specify the height of the inserted image in pixels. If you specify a value that's larger than the video resolution height, the service will crop your overlaid image to fit. To use the native height of the image, keep this setting blank.
3031
3315
  */
3032
3316
  Height?: number;
3033
3317
  /**
3318
+ * @public
3034
3319
  * Specify the HTTP, HTTPS, or Amazon S3 location of the image that you want to overlay on the video. Use a PNG or TGA file.
3035
3320
  */
3036
3321
  ImageInserterInput?: string;
3037
3322
  /**
3323
+ * @public
3038
3324
  * Specify the distance, in pixels, between the inserted image and the left edge of the video frame. Required for any image overlay that you specify.
3039
3325
  */
3040
3326
  ImageX?: number;
3041
3327
  /**
3328
+ * @public
3042
3329
  * Specify the distance, in pixels, between the overlaid image and the top edge of the video frame. Required for any image overlay that you specify.
3043
3330
  */
3044
3331
  ImageY?: number;
3045
3332
  /**
3333
+ * @public
3046
3334
  * Specify how overlapping inserted images appear. Images with higher values for Layer appear on top of images with lower values for Layer.
3047
3335
  */
3048
3336
  Layer?: number;
3049
3337
  /**
3338
+ * @public
3050
3339
  * Use Opacity to specify how much of the underlying video shows through the inserted image. 0 is transparent and 100 is fully opaque. Default is 50.
3051
3340
  */
3052
3341
  Opacity?: number;
3053
3342
  /**
3343
+ * @public
3054
3344
  * Specify the timecode of the frame that you want the overlay to first appear on. This must be in timecode (HH:MM:SS:FF or HH:MM:SS;FF) format. Remember to take into account your timecode source settings.
3055
3345
  */
3056
3346
  StartTime?: string;
3057
3347
  /**
3348
+ * @public
3058
3349
  * Specify the width of the inserted image in pixels. If you specify a value that's larger than the video resolution width, the service will crop your overlaid image to fit. To use the native width of the image, keep this setting blank.
3059
3350
  */
3060
3351
  Width?: number;
@@ -3065,10 +3356,12 @@ export interface InsertableImage {
3065
3356
  */
3066
3357
  export interface ImageInserter {
3067
3358
  /**
3359
+ * @public
3068
3360
  * Specify the images that you want to overlay on your video. The images must be PNG or TGA files.
3069
3361
  */
3070
3362
  InsertableImages?: InsertableImage[];
3071
3363
  /**
3364
+ * @public
3072
3365
  * Specify the reference white level, in nits, for all of your image inserter images. Use to correct brightness levels within HDR10 outputs. For 1,000 nit peak brightness displays, we recommend that you set SDR reference white level to 203 (according to ITU-R BT.2408). Leave blank to use the default value of 100, or specify an integer from 100 to 1000.
3073
3366
  */
3074
3367
  SdrReferenceWhiteLevel?: number;
@@ -3079,10 +3372,12 @@ export interface ImageInserter {
3079
3372
  */
3080
3373
  export interface InputClipping {
3081
3374
  /**
3375
+ * @public
3082
3376
  * Set End timecode to the end of the portion of the input you are clipping. The frame corresponding to the End timecode value is included in the clip. Start timecode or End timecode may be left blank, but not both. Use the format HH:MM:SS:FF or HH:MM:SS;FF, where HH is the hour, MM is the minute, SS is the second, and FF is the frame number. When choosing this value, take into account your setting for timecode source under input settings. For example, if you have embedded timecodes that start at 01:00:00:00 and you want your clip to end six minutes into the video, use 01:06:00:00.
3083
3377
  */
3084
3378
  EndTimecode?: string;
3085
3379
  /**
3380
+ * @public
3086
3381
  * Set Start timecode to the beginning of the portion of the input you are clipping. The frame corresponding to the Start timecode value is included in the clip. Start timecode or End timecode may be left blank, but not both. Use the format HH:MM:SS:FF or HH:MM:SS;FF, where HH is the hour, MM is the minute, SS is the second, and FF is the frame number. When choosing this value, take into account your setting for Input timecode source. For example, if you have embedded timecodes that start at 01:00:00:00 and you want your clip to begin five minutes into the video, use 01:05:00:00.
3087
3382
  */
3088
3383
  StartTimecode?: string;
@@ -3130,6 +3425,7 @@ export type InputTimecodeSource = (typeof InputTimecodeSource)[keyof typeof Inpu
3130
3425
  */
3131
3426
  export interface InputVideoGenerator {
3132
3427
  /**
3428
+ * @public
3133
3429
  * Specify an integer value for Black video duration from 50 to 86400000 to generate a black video input for that many milliseconds. Required when you include Video generator.
3134
3430
  */
3135
3431
  Duration?: number;
@@ -3194,50 +3490,62 @@ export type EmbeddedTimecodeOverride = (typeof EmbeddedTimecodeOverride)[keyof t
3194
3490
  */
3195
3491
  export interface Hdr10Metadata {
3196
3492
  /**
3493
+ * @public
3197
3494
  * HDR Master Display Information must be provided by a color grader, using color grading tools. Range is 0 to 50,000, each increment represents 0.00002 in CIE1931 color coordinate. Note that this setting is not for color correction.
3198
3495
  */
3199
3496
  BluePrimaryX?: number;
3200
3497
  /**
3498
+ * @public
3201
3499
  * HDR Master Display Information must be provided by a color grader, using color grading tools. Range is 0 to 50,000, each increment represents 0.00002 in CIE1931 color coordinate. Note that this setting is not for color correction.
3202
3500
  */
3203
3501
  BluePrimaryY?: number;
3204
3502
  /**
3503
+ * @public
3205
3504
  * HDR Master Display Information must be provided by a color grader, using color grading tools. Range is 0 to 50,000, each increment represents 0.00002 in CIE1931 color coordinate. Note that this setting is not for color correction.
3206
3505
  */
3207
3506
  GreenPrimaryX?: number;
3208
3507
  /**
3508
+ * @public
3209
3509
  * HDR Master Display Information must be provided by a color grader, using color grading tools. Range is 0 to 50,000, each increment represents 0.00002 in CIE1931 color coordinate. Note that this setting is not for color correction.
3210
3510
  */
3211
3511
  GreenPrimaryY?: number;
3212
3512
  /**
3513
+ * @public
3213
3514
  * Maximum light level among all samples in the coded video sequence, in units of candelas per square meter. This setting doesn't have a default value; you must specify a value that is suitable for the content.
3214
3515
  */
3215
3516
  MaxContentLightLevel?: number;
3216
3517
  /**
3518
+ * @public
3217
3519
  * Maximum average light level of any frame in the coded video sequence, in units of candelas per square meter. This setting doesn't have a default value; you must specify a value that is suitable for the content.
3218
3520
  */
3219
3521
  MaxFrameAverageLightLevel?: number;
3220
3522
  /**
3523
+ * @public
3221
3524
  * Nominal maximum mastering display luminance in units of of 0.0001 candelas per square meter.
3222
3525
  */
3223
3526
  MaxLuminance?: number;
3224
3527
  /**
3528
+ * @public
3225
3529
  * Nominal minimum mastering display luminance in units of of 0.0001 candelas per square meter
3226
3530
  */
3227
3531
  MinLuminance?: number;
3228
3532
  /**
3533
+ * @public
3229
3534
  * HDR Master Display Information must be provided by a color grader, using color grading tools. Range is 0 to 50,000, each increment represents 0.00002 in CIE1931 color coordinate. Note that this setting is not for color correction.
3230
3535
  */
3231
3536
  RedPrimaryX?: number;
3232
3537
  /**
3538
+ * @public
3233
3539
  * HDR Master Display Information must be provided by a color grader, using color grading tools. Range is 0 to 50,000, each increment represents 0.00002 in CIE1931 color coordinate. Note that this setting is not for color correction.
3234
3540
  */
3235
3541
  RedPrimaryY?: number;
3236
3542
  /**
3543
+ * @public
3237
3544
  * HDR Master Display Information must be provided by a color grader, using color grading tools. Range is 0 to 50,000, each increment represents 0.00002 in CIE1931 color coordinate. Note that this setting is not for color correction.
3238
3545
  */
3239
3546
  WhitePointX?: number;
3240
3547
  /**
3548
+ * @public
3241
3549
  * HDR Master Display Information must be provided by a color grader, using color grading tools. Range is 0 to 50,000, each increment represents 0.00002 in CIE1931 color coordinate. Note that this setting is not for color correction.
3242
3550
  */
3243
3551
  WhitePointY?: number;
@@ -3288,10 +3596,12 @@ export type InputSampleRange = (typeof InputSampleRange)[keyof typeof InputSampl
3288
3596
  */
3289
3597
  export interface VideoSelector {
3290
3598
  /**
3599
+ * @public
3291
3600
  * Ignore this setting unless this input is a QuickTime animation with an alpha channel. Use this setting to create separate Key and Fill outputs. In each output, specify which part of the input MediaConvert uses. Leave this setting at the default value DISCARD to delete the alpha channel and preserve the video. Set it to REMAP_TO_LUMA to delete the video and map the alpha channel to the luma channel of your outputs.
3292
3601
  */
3293
3602
  AlphaBehavior?: AlphaBehavior | string;
3294
3603
  /**
3604
+ * @public
3295
3605
  * If your input video has accurate color space metadata, or if you don't know about color space: Keep the default value, Follow. MediaConvert will automatically detect your input color space. If your input video has metadata indicating the wrong color space, or has missing metadata: Specify the accurate color space here. If your input video is HDR 10 and the SMPTE ST 2086 Mastering Display Color Volume static metadata isn't present in your video stream, or if that metadata is present but not accurate: Choose Force HDR 10. Specify correct values in the input HDR 10 metadata settings. For more information about HDR jobs, see https://docs.aws.amazon.com/console/mediaconvert/hdr. When you specify an input color space, MediaConvert uses the following color space metadata, which includes color primaries, transfer characteristics, and matrix coefficients:
3296
3606
  * * HDR 10: BT.2020, PQ, BT.2020 non-constant
3297
3607
  * * HLG 2020: BT.2020, HLG, BT.2020 non-constant
@@ -3301,34 +3611,42 @@ export interface VideoSelector {
3301
3611
  */
3302
3612
  ColorSpace?: ColorSpace | string;
3303
3613
  /**
3614
+ * @public
3304
3615
  * There are two sources for color metadata, the input file and the job input settings Color space and HDR master display information settings. The Color space usage setting determines which takes precedence. Choose Force to use color metadata from the input job settings. If you don't specify values for those settings, the service defaults to using metadata from your input. FALLBACK - Choose Fallback to use color metadata from the source when it is present. If there's no color metadata in your input file, the service defaults to using values you specify in the input settings.
3305
3616
  */
3306
3617
  ColorSpaceUsage?: ColorSpaceUsage | string;
3307
3618
  /**
3619
+ * @public
3308
3620
  * Set Embedded timecode override to Use MDPM when your AVCHD input contains timecode tag data in the Modified Digital Video Pack Metadata. When you do, we recommend you also set Timecode source to Embedded. Leave Embedded timecode override blank, or set to None, when your input does not contain MDPM timecode.
3309
3621
  */
3310
3622
  EmbeddedTimecodeOverride?: EmbeddedTimecodeOverride | string;
3311
3623
  /**
3624
+ * @public
3312
3625
  * Use these settings to provide HDR 10 metadata that is missing or inaccurate in your input video. Appropriate values vary depending on the input video and must be provided by a color grader. The color grader generates these values during the HDR 10 mastering process. The valid range for each of these settings is 0 to 50,000. Each increment represents 0.00002 in CIE1931 color coordinate. Related settings - When you specify these values, you must also set Color space to HDR 10. To specify whether the the values you specify here take precedence over the values in the metadata of your input file, set Color space usage. To specify whether color metadata is included in an output, set Color metadata. For more information about MediaConvert HDR jobs, see https://docs.aws.amazon.com/console/mediaconvert/hdr.
3313
3626
  */
3314
3627
  Hdr10Metadata?: Hdr10Metadata;
3315
3628
  /**
3629
+ * @public
3316
3630
  * Use this setting if your input has video and audio durations that don't align, and your output or player has strict alignment requirements. Examples: Input audio track has a delayed start. Input video track ends before audio ends. When you set Pad video to Black, MediaConvert generates black video frames so that output video and audio durations match. Black video frames are added at the beginning or end, depending on your input. To keep the default behavior and not generate black video, set Pad video to Disabled or leave blank.
3317
3631
  */
3318
3632
  PadVideo?: PadVideo | string;
3319
3633
  /**
3634
+ * @public
3320
3635
  * Use PID to select specific video data from an input file. Specify this value as an integer; the system automatically converts it to the hexidecimal value. For example, 257 selects PID 0x101. A PID, or packet identifier, is an identifier for a set of data in an MPEG-2 transport stream container.
3321
3636
  */
3322
3637
  Pid?: number;
3323
3638
  /**
3639
+ * @public
3324
3640
  * Selects a specific program from within a multi-program transport stream. Note that Quad 4K is not currently supported.
3325
3641
  */
3326
3642
  ProgramNumber?: number;
3327
3643
  /**
3644
+ * @public
3328
3645
  * Use Rotate to specify how the service rotates your video. You can choose automatic rotation or specify a rotation. You can specify a clockwise rotation of 0, 90, 180, or 270 degrees. If your input video container is .mov or .mp4 and your input has rotation metadata, you can choose Automatic to have the service rotate your video according to the rotation specified in the metadata. The rotation must be within one degree of 90, 180, or 270 degrees. If the rotation metadata specifies any other rotation, the service will default to no rotation. By default, the service does no rotation, even if your input video has rotation metadata. The service doesn't pass through rotation metadata.
3329
3646
  */
3330
3647
  Rotate?: InputRotate | string;
3331
3648
  /**
3649
+ * @public
3332
3650
  * If the sample range metadata in your input video is accurate, or if you don't know about sample range, keep the default value, Follow, for this setting. When you do, the service automatically detects your input sample range. If your input video has metadata indicating the wrong sample range, specify the accurate sample range here. When you do, MediaConvert ignores any sample range information in the input metadata. Regardless of whether MediaConvert uses the input sample range or the sample range that you specify, MediaConvert uses the sample range for transcoding and also writes it to the output metadata.
3333
3651
  */
3334
3652
  SampleRange?: InputSampleRange | string;
@@ -3339,100 +3657,124 @@ export interface VideoSelector {
3339
3657
  */
3340
3658
  export interface Input {
3341
3659
  /**
3660
+ * @public
3342
3661
  * Use to remove noise, blocking, blurriness, or ringing from your input as a pre-filter step before encoding. The Advanced input filter removes more types of compression artifacts and is an improvement when compared to basic Deblock and Denoise filters. To remove video compression artifacts from your input and improve the video quality: Choose Enabled. Additionally, this filter can help increase the video quality of your output relative to its bitrate, since noisy inputs are more complex and require more bits to encode. To help restore loss of detail after applying the filter, you can optionally add texture or sharpening as an additional step. Jobs that use this feature incur pro-tier pricing. To not apply advanced input filtering: Choose Disabled. Note that you can still apply basic filtering with Deblock and Denoise.
3343
3662
  */
3344
3663
  AdvancedInputFilter?: AdvancedInputFilter | string;
3345
3664
  /**
3665
+ * @public
3346
3666
  * Optional settings for Advanced input filter when you set Advanced input filter to Enabled.
3347
3667
  */
3348
3668
  AdvancedInputFilterSettings?: AdvancedInputFilterSettings;
3349
3669
  /**
3670
+ * @public
3350
3671
  * Use audio selector groups to combine multiple sidecar audio inputs so that you can assign them to a single output audio tab. Note that, if you're working with embedded audio, it's simpler to assign multiple input tracks into a single audio selector rather than use an audio selector group.
3351
3672
  */
3352
3673
  AudioSelectorGroups?: Record<string, AudioSelectorGroup>;
3353
3674
  /**
3675
+ * @public
3354
3676
  * Use Audio selectors to specify a track or set of tracks from the input that you will use in your outputs. You can use multiple Audio selectors per input.
3355
3677
  */
3356
3678
  AudioSelectors?: Record<string, AudioSelector>;
3357
3679
  /**
3680
+ * @public
3358
3681
  * Use captions selectors to specify the captions data from your input that you use in your outputs. You can use up to 100 captions selectors per input.
3359
3682
  */
3360
3683
  CaptionSelectors?: Record<string, CaptionSelector>;
3361
3684
  /**
3685
+ * @public
3362
3686
  * Use Cropping selection to specify the video area that the service will include in the output video frame. If you specify a value here, it will override any value that you specify in the output setting Cropping selection.
3363
3687
  */
3364
3688
  Crop?: Rectangle;
3365
3689
  /**
3690
+ * @public
3366
3691
  * Enable Deblock to produce smoother motion in the output. Default is disabled. Only manually controllable for MPEG2 and uncompressed video inputs.
3367
3692
  */
3368
3693
  DeblockFilter?: InputDeblockFilter | string;
3369
3694
  /**
3695
+ * @public
3370
3696
  * Settings for decrypting any input files that you encrypt before you upload them to Amazon S3. MediaConvert can decrypt files only when you use AWS Key Management Service (KMS) to encrypt the data key that you use to encrypt your content.
3371
3697
  */
3372
3698
  DecryptionSettings?: InputDecryptionSettings;
3373
3699
  /**
3700
+ * @public
3374
3701
  * Enable Denoise to filter noise from the input. Default is disabled. Only applicable to MPEG2, H.264, H.265, and uncompressed video inputs.
3375
3702
  */
3376
3703
  DenoiseFilter?: InputDenoiseFilter | string;
3377
3704
  /**
3705
+ * @public
3378
3706
  * Use this setting only when your video source has Dolby Vision studio mastering metadata that is carried in a separate XML file. Specify the Amazon S3 location for the metadata XML file. MediaConvert uses this file to provide global and frame-level metadata for Dolby Vision preprocessing. When you specify a file here and your input also has interleaved global and frame level metadata, MediaConvert ignores the interleaved metadata and uses only the the metadata from this external XML file. Note that your IAM service role must grant MediaConvert read permissions to this file. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/iam-role.html.
3379
3707
  */
3380
3708
  DolbyVisionMetadataXml?: string;
3381
3709
  /**
3710
+ * @public
3382
3711
  * Specify the source file for your transcoding job. You can use multiple inputs in a single job. The service concatenates these inputs, in the order that you specify them in the job, to create the outputs. If your input format is IMF, specify your input by providing the path to your CPL. For example, "s3://bucket/vf/cpl.xml". If the CPL is in an incomplete IMP, make sure to use *Supplemental IMPs* to specify any supplemental IMPs that contain assets referenced by the CPL.
3383
3712
  */
3384
3713
  FileInput?: string;
3385
3714
  /**
3715
+ * @public
3386
3716
  * Specify whether to apply input filtering to improve the video quality of your input. To apply filtering depending on your input type and quality: Choose Auto. To apply no filtering: Choose Disable. To apply filtering regardless of your input type and quality: Choose Force. When you do, you must also specify a value for Filter strength.
3387
3717
  */
3388
3718
  FilterEnable?: InputFilterEnable | string;
3389
3719
  /**
3720
+ * @public
3390
3721
  * Specify the strength of the input filter. To apply an automatic amount of filtering based the compression artifacts measured in your input: We recommend that you leave Filter strength blank and set Filter enable to Auto. To manually apply filtering: Enter a value from 1 to 5, where 1 is the least amount of filtering and 5 is the most. The value that you enter applies to the strength of the Deblock or Denoise filters, or to the strength of the Advanced input filter.
3391
3722
  */
3392
3723
  FilterStrength?: number;
3393
3724
  /**
3725
+ * @public
3394
3726
  * Enable the image inserter feature to include a graphic overlay on your video. Enable or disable this feature for each input individually. This setting is disabled by default.
3395
3727
  */
3396
3728
  ImageInserter?: ImageInserter;
3397
3729
  /**
3730
+ * @public
3398
3731
  * Contains sets of start and end times that together specify a portion of the input to be used in the outputs. If you provide only a start time, the clip will be the entire input from that point to the end. If you provide only an end time, it will be the entire input up to that point. When you specify more than one input clip, the transcoding service creates the job outputs by stringing the clips together in the order you specify them.
3399
3732
  */
3400
3733
  InputClippings?: InputClipping[];
3401
3734
  /**
3735
+ * @public
3402
3736
  * When you have a progressive segmented frame (PsF) input, use this setting to flag the input as PsF. MediaConvert doesn't automatically detect PsF. Therefore, flagging your input as PsF results in better preservation of video quality when you do deinterlacing and frame rate conversion. If you don't specify, the default value is Auto. Auto is the correct setting for all inputs that are not PsF. Don't set this value to PsF when your input is interlaced. Doing so creates horizontal interlacing artifacts.
3403
3737
  */
3404
3738
  InputScanType?: InputScanType | string;
3405
3739
  /**
3740
+ * @public
3406
3741
  * Use Selection placement to define the video area in your output frame. The area outside of the rectangle that you specify here is black. If you specify a value here, it will override any value that you specify in the output setting Selection placement. If you specify a value here, this will override any AFD values in your input, even if you set Respond to AFD to Respond. If you specify a value here, this will ignore anything that you specify for the setting Scaling Behavior.
3407
3742
  */
3408
3743
  Position?: Rectangle;
3409
3744
  /**
3745
+ * @public
3410
3746
  * Use Program to select a specific program from within a multi-program transport stream. Note that Quad 4K is not currently supported. Default is the first program within the transport stream. If the program you specify doesn't exist, the transcoding service will use this default.
3411
3747
  */
3412
3748
  ProgramNumber?: number;
3413
3749
  /**
3750
+ * @public
3414
3751
  * Set PSI control for transport stream inputs to specify which data the demux process to scans.
3415
3752
  * * Ignore PSI - Scan all PIDs for audio and video.
3416
3753
  * * Use PSI - Scan only PSI data.
3417
3754
  */
3418
3755
  PsiControl?: InputPsiControl | string;
3419
3756
  /**
3757
+ * @public
3420
3758
  * Provide a list of any necessary supplemental IMPs. You need supplemental IMPs if the CPL that you're using for your input is in an incomplete IMP. Specify either the supplemental IMP directories with a trailing slash or the ASSETMAP.xml files. For example ["s3://bucket/ov/", "s3://bucket/vf2/ASSETMAP.xml"]. You don't need to specify the IMP that contains your input CPL, because the service automatically detects it.
3421
3759
  */
3422
3760
  SupplementalImps?: string[];
3423
3761
  /**
3762
+ * @public
3424
3763
  * Use this Timecode source setting, located under the input settings, to specify how the service counts input video frames. This input frame count affects only the behavior of features that apply to a single input at a time, such as input clipping and synchronizing some captions formats. Choose Embedded to use the timecodes in your input video. Choose Start at zero to start the first frame at zero. Choose Specified start to start the first frame at the timecode that you specify in the setting Start timecode. If you don't specify a value for Timecode source, the service will use Embedded by default. For more information about timecodes, see https://docs.aws.amazon.com/console/mediaconvert/timecode.
3425
3764
  */
3426
3765
  TimecodeSource?: InputTimecodeSource | string;
3427
3766
  /**
3767
+ * @public
3428
3768
  * Specify the timecode that you want the service to use for this input's initial frame. To use this setting, you must set the Timecode source setting, located under the input settings, to Specified start. For more information about timecodes, see https://docs.aws.amazon.com/console/mediaconvert/timecode.
3429
3769
  */
3430
3770
  TimecodeStart?: string;
3431
3771
  /**
3772
+ * @public
3432
3773
  * When you include Video generator, MediaConvert creates a video input with black frames. Use this setting if you do not have a video input or if you want to add black video frames before, or after, other inputs. You can specify Video generator, or you can specify an Input file, but you cannot specify both. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/video-generator.html
3433
3774
  */
3434
3775
  VideoGenerator?: InputVideoGenerator;
3435
3776
  /**
3777
+ * @public
3436
3778
  * Input video selectors contain the video settings for the input. Each of your inputs can have up to one video selector.
3437
3779
  */
3438
3780
  VideoSelector?: VideoSelector;
@@ -3443,84 +3785,104 @@ export interface Input {
3443
3785
  */
3444
3786
  export interface InputTemplate {
3445
3787
  /**
3788
+ * @public
3446
3789
  * Use to remove noise, blocking, blurriness, or ringing from your input as a pre-filter step before encoding. The Advanced input filter removes more types of compression artifacts and is an improvement when compared to basic Deblock and Denoise filters. To remove video compression artifacts from your input and improve the video quality: Choose Enabled. Additionally, this filter can help increase the video quality of your output relative to its bitrate, since noisy inputs are more complex and require more bits to encode. To help restore loss of detail after applying the filter, you can optionally add texture or sharpening as an additional step. Jobs that use this feature incur pro-tier pricing. To not apply advanced input filtering: Choose Disabled. Note that you can still apply basic filtering with Deblock and Denoise.
3447
3790
  */
3448
3791
  AdvancedInputFilter?: AdvancedInputFilter | string;
3449
3792
  /**
3793
+ * @public
3450
3794
  * Optional settings for Advanced input filter when you set Advanced input filter to Enabled.
3451
3795
  */
3452
3796
  AdvancedInputFilterSettings?: AdvancedInputFilterSettings;
3453
3797
  /**
3798
+ * @public
3454
3799
  * Use audio selector groups to combine multiple sidecar audio inputs so that you can assign them to a single output audio tab. Note that, if you're working with embedded audio, it's simpler to assign multiple input tracks into a single audio selector rather than use an audio selector group.
3455
3800
  */
3456
3801
  AudioSelectorGroups?: Record<string, AudioSelectorGroup>;
3457
3802
  /**
3803
+ * @public
3458
3804
  * Use Audio selectors to specify a track or set of tracks from the input that you will use in your outputs. You can use multiple Audio selectors per input.
3459
3805
  */
3460
3806
  AudioSelectors?: Record<string, AudioSelector>;
3461
3807
  /**
3808
+ * @public
3462
3809
  * Use captions selectors to specify the captions data from your input that you use in your outputs. You can use up to 100 captions selectors per input.
3463
3810
  */
3464
3811
  CaptionSelectors?: Record<string, CaptionSelector>;
3465
3812
  /**
3813
+ * @public
3466
3814
  * Use Cropping selection to specify the video area that the service will include in the output video frame. If you specify a value here, it will override any value that you specify in the output setting Cropping selection.
3467
3815
  */
3468
3816
  Crop?: Rectangle;
3469
3817
  /**
3818
+ * @public
3470
3819
  * Enable Deblock to produce smoother motion in the output. Default is disabled. Only manually controllable for MPEG2 and uncompressed video inputs.
3471
3820
  */
3472
3821
  DeblockFilter?: InputDeblockFilter | string;
3473
3822
  /**
3823
+ * @public
3474
3824
  * Enable Denoise to filter noise from the input. Default is disabled. Only applicable to MPEG2, H.264, H.265, and uncompressed video inputs.
3475
3825
  */
3476
3826
  DenoiseFilter?: InputDenoiseFilter | string;
3477
3827
  /**
3828
+ * @public
3478
3829
  * Use this setting only when your video source has Dolby Vision studio mastering metadata that is carried in a separate XML file. Specify the Amazon S3 location for the metadata XML file. MediaConvert uses this file to provide global and frame-level metadata for Dolby Vision preprocessing. When you specify a file here and your input also has interleaved global and frame level metadata, MediaConvert ignores the interleaved metadata and uses only the the metadata from this external XML file. Note that your IAM service role must grant MediaConvert read permissions to this file. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/iam-role.html.
3479
3830
  */
3480
3831
  DolbyVisionMetadataXml?: string;
3481
3832
  /**
3833
+ * @public
3482
3834
  * Specify whether to apply input filtering to improve the video quality of your input. To apply filtering depending on your input type and quality: Choose Auto. To apply no filtering: Choose Disable. To apply filtering regardless of your input type and quality: Choose Force. When you do, you must also specify a value for Filter strength.
3483
3835
  */
3484
3836
  FilterEnable?: InputFilterEnable | string;
3485
3837
  /**
3838
+ * @public
3486
3839
  * Specify the strength of the input filter. To apply an automatic amount of filtering based the compression artifacts measured in your input: We recommend that you leave Filter strength blank and set Filter enable to Auto. To manually apply filtering: Enter a value from 1 to 5, where 1 is the least amount of filtering and 5 is the most. The value that you enter applies to the strength of the Deblock or Denoise filters, or to the strength of the Advanced input filter.
3487
3840
  */
3488
3841
  FilterStrength?: number;
3489
3842
  /**
3843
+ * @public
3490
3844
  * Enable the image inserter feature to include a graphic overlay on your video. Enable or disable this feature for each input individually. This setting is disabled by default.
3491
3845
  */
3492
3846
  ImageInserter?: ImageInserter;
3493
3847
  /**
3848
+ * @public
3494
3849
  * Contains sets of start and end times that together specify a portion of the input to be used in the outputs. If you provide only a start time, the clip will be the entire input from that point to the end. If you provide only an end time, it will be the entire input up to that point. When you specify more than one input clip, the transcoding service creates the job outputs by stringing the clips together in the order you specify them.
3495
3850
  */
3496
3851
  InputClippings?: InputClipping[];
3497
3852
  /**
3853
+ * @public
3498
3854
  * When you have a progressive segmented frame (PsF) input, use this setting to flag the input as PsF. MediaConvert doesn't automatically detect PsF. Therefore, flagging your input as PsF results in better preservation of video quality when you do deinterlacing and frame rate conversion. If you don't specify, the default value is Auto. Auto is the correct setting for all inputs that are not PsF. Don't set this value to PsF when your input is interlaced. Doing so creates horizontal interlacing artifacts.
3499
3855
  */
3500
3856
  InputScanType?: InputScanType | string;
3501
3857
  /**
3858
+ * @public
3502
3859
  * Use Selection placement to define the video area in your output frame. The area outside of the rectangle that you specify here is black. If you specify a value here, it will override any value that you specify in the output setting Selection placement. If you specify a value here, this will override any AFD values in your input, even if you set Respond to AFD to Respond. If you specify a value here, this will ignore anything that you specify for the setting Scaling Behavior.
3503
3860
  */
3504
3861
  Position?: Rectangle;
3505
3862
  /**
3863
+ * @public
3506
3864
  * Use Program to select a specific program from within a multi-program transport stream. Note that Quad 4K is not currently supported. Default is the first program within the transport stream. If the program you specify doesn't exist, the transcoding service will use this default.
3507
3865
  */
3508
3866
  ProgramNumber?: number;
3509
3867
  /**
3868
+ * @public
3510
3869
  * Set PSI control for transport stream inputs to specify which data the demux process to scans.
3511
3870
  * * Ignore PSI - Scan all PIDs for audio and video.
3512
3871
  * * Use PSI - Scan only PSI data.
3513
3872
  */
3514
3873
  PsiControl?: InputPsiControl | string;
3515
3874
  /**
3875
+ * @public
3516
3876
  * Use this Timecode source setting, located under the input settings, to specify how the service counts input video frames. This input frame count affects only the behavior of features that apply to a single input at a time, such as input clipping and synchronizing some captions formats. Choose Embedded to use the timecodes in your input video. Choose Start at zero to start the first frame at zero. Choose Specified start to start the first frame at the timecode that you specify in the setting Start timecode. If you don't specify a value for Timecode source, the service will use Embedded by default. For more information about timecodes, see https://docs.aws.amazon.com/console/mediaconvert/timecode.
3517
3877
  */
3518
3878
  TimecodeSource?: InputTimecodeSource | string;
3519
3879
  /**
3880
+ * @public
3520
3881
  * Specify the timecode that you want the service to use for this input's initial frame. To use this setting, you must set the Timecode source setting, located under the input settings, to Specified start. For more information about timecodes, see https://docs.aws.amazon.com/console/mediaconvert/timecode.
3521
3882
  */
3522
3883
  TimecodeStart?: string;
3523
3884
  /**
3885
+ * @public
3524
3886
  * Input video selectors contain the video settings for the input. Each of your inputs can have up to one video selector.
3525
3887
  */
3526
3888
  VideoSelector?: VideoSelector;
@@ -3544,6 +3906,7 @@ export type AccelerationMode = (typeof AccelerationMode)[keyof typeof Accelerati
3544
3906
  */
3545
3907
  export interface AccelerationSettings {
3546
3908
  /**
3909
+ * @public
3547
3910
  * Specify the conditions when the service will run your job with accelerated transcoding.
3548
3911
  */
3549
3912
  Mode: AccelerationMode | string | undefined;
@@ -3595,10 +3958,12 @@ export type JobPhase = (typeof JobPhase)[keyof typeof JobPhase];
3595
3958
  */
3596
3959
  export interface JobMessages {
3597
3960
  /**
3961
+ * @public
3598
3962
  * List of messages that are informational only and don't indicate a problem with your job.
3599
3963
  */
3600
3964
  Info?: string[];
3601
3965
  /**
3966
+ * @public
3602
3967
  * List of messages that warn about conditions that might cause your job not to run or to fail.
3603
3968
  */
3604
3969
  Warning?: string[];
@@ -3609,10 +3974,12 @@ export interface JobMessages {
3609
3974
  */
3610
3975
  export interface VideoDetail {
3611
3976
  /**
3977
+ * @public
3612
3978
  * Height in pixels for the output
3613
3979
  */
3614
3980
  HeightInPx?: number;
3615
3981
  /**
3982
+ * @public
3616
3983
  * Width in pixels for the output
3617
3984
  */
3618
3985
  WidthInPx?: number;
@@ -3623,10 +3990,12 @@ export interface VideoDetail {
3623
3990
  */
3624
3991
  export interface OutputDetail {
3625
3992
  /**
3993
+ * @public
3626
3994
  * Duration in milliseconds
3627
3995
  */
3628
3996
  DurationInMs?: number;
3629
3997
  /**
3998
+ * @public
3630
3999
  * Contains details about the output's video stream
3631
4000
  */
3632
4001
  VideoDetails?: VideoDetail;
@@ -3637,6 +4006,7 @@ export interface OutputDetail {
3637
4006
  */
3638
4007
  export interface OutputGroupDetail {
3639
4008
  /**
4009
+ * @public
3640
4010
  * Details about the output
3641
4011
  */
3642
4012
  OutputDetails?: OutputDetail[];
@@ -3647,14 +4017,17 @@ export interface OutputGroupDetail {
3647
4017
  */
3648
4018
  export interface QueueTransition {
3649
4019
  /**
4020
+ * @public
3650
4021
  * The queue that the job was on after the transition.
3651
4022
  */
3652
4023
  DestinationQueue?: string;
3653
4024
  /**
4025
+ * @public
3654
4026
  * The queue that the job was on before the transition.
3655
4027
  */
3656
4028
  SourceQueue?: string;
3657
4029
  /**
4030
+ * @public
3658
4031
  * The time, in Unix epoch format, that the job moved from the source queue to the destination queue.
3659
4032
  */
3660
4033
  Timestamp?: Date;
@@ -3665,6 +4038,7 @@ export interface QueueTransition {
3665
4038
  */
3666
4039
  export interface AvailBlanking {
3667
4040
  /**
4041
+ * @public
3668
4042
  * Blanking image to be used. Leave empty for solid black. Only bmp and png images are supported.
3669
4043
  */
3670
4044
  AvailBlankingImage?: string;
@@ -3675,6 +4049,7 @@ export interface AvailBlanking {
3675
4049
  */
3676
4050
  export interface EsamManifestConfirmConditionNotification {
3677
4051
  /**
4052
+ * @public
3678
4053
  * Provide your ESAM ManifestConfirmConditionNotification XML document inside your JSON job settings. Form the XML document as per OC-SP-ESAM-API-I03-131025. The transcoder will use the Manifest Conditioning instructions in the message that you supply.
3679
4054
  */
3680
4055
  MccXml?: string;
@@ -3685,6 +4060,7 @@ export interface EsamManifestConfirmConditionNotification {
3685
4060
  */
3686
4061
  export interface EsamSignalProcessingNotification {
3687
4062
  /**
4063
+ * @public
3688
4064
  * Provide your ESAM SignalProcessingNotification XML document inside your JSON job settings. Form the XML document as per OC-SP-ESAM-API-I03-131025. The transcoder will use the signal processing instructions in the message that you supply. For your MPEG2-TS file outputs, if you want the service to place SCTE-35 markers at the insertion points you specify in the XML document, you must also enable SCTE-35 ESAM. Note that you can either specify an ESAM XML document or enable SCTE-35 passthrough. You can't do both.
3689
4065
  */
3690
4066
  SccXml?: string;
@@ -3695,14 +4071,17 @@ export interface EsamSignalProcessingNotification {
3695
4071
  */
3696
4072
  export interface EsamSettings {
3697
4073
  /**
4074
+ * @public
3698
4075
  * Specifies an ESAM ManifestConfirmConditionNotification XML as per OC-SP-ESAM-API-I03-131025. The transcoder uses the manifest conditioning instructions that you provide in the setting MCC XML.
3699
4076
  */
3700
4077
  ManifestConfirmConditionNotification?: EsamManifestConfirmConditionNotification;
3701
4078
  /**
4079
+ * @public
3702
4080
  * Specifies the stream distance, in milliseconds, between the SCTE 35 messages that the transcoder places and the splice points that they refer to. If the time between the start of the asset and the SCTE-35 message is less than this value, then the transcoder places the SCTE-35 marker at the beginning of the stream.
3703
4081
  */
3704
4082
  ResponseSignalPreroll?: number;
3705
4083
  /**
4084
+ * @public
3706
4085
  * Specifies an ESAM SignalProcessingNotification XML as per OC-SP-ESAM-API-I03-131025. The transcoder uses the signal processing instructions that you provide in the setting SCC XML.
3707
4086
  */
3708
4087
  SignalProcessingNotification?: EsamSignalProcessingNotification;
@@ -3737,10 +4116,12 @@ export type VchipAction = (typeof VchipAction)[keyof typeof VchipAction];
3737
4116
  */
3738
4117
  export interface ExtendedDataServices {
3739
4118
  /**
4119
+ * @public
3740
4120
  * The action to take on copy and redistribution control XDS packets. If you select PASSTHROUGH, packets will not be changed. If you select STRIP, any packets will be removed in output captions.
3741
4121
  */
3742
4122
  CopyProtectionAction?: CopyProtectionAction | string;
3743
4123
  /**
4124
+ * @public
3744
4125
  * The action to take on content advisory XDS packets. If you select PASSTHROUGH, packets will not be changed. If you select STRIP, any packets will be removed in output captions.
3745
4126
  */
3746
4127
  VchipAction?: VchipAction | string;
@@ -3751,54 +4132,67 @@ export interface ExtendedDataServices {
3751
4132
  */
3752
4133
  export interface KantarWatermarkSettings {
3753
4134
  /**
4135
+ * @public
3754
4136
  * Provide an audio channel name from your Kantar audio license.
3755
4137
  */
3756
4138
  ChannelName?: string;
3757
4139
  /**
4140
+ * @public
3758
4141
  * Specify a unique identifier for Kantar to use for this piece of content.
3759
4142
  */
3760
4143
  ContentReference?: string;
3761
4144
  /**
4145
+ * @public
3762
4146
  * Provide the name of the AWS Secrets Manager secret where your Kantar credentials are stored. Note that your MediaConvert service role must provide access to this secret. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/granting-permissions-for-mediaconvert-to-access-secrets-manager-secret.html. For instructions on creating a secret, see https://docs.aws.amazon.com/secretsmanager/latest/userguide/tutorials_basic.html, in the AWS Secrets Manager User Guide.
3763
4147
  */
3764
4148
  CredentialsSecretName?: string;
3765
4149
  /**
4150
+ * @public
3766
4151
  * Optional. Specify an offset, in whole seconds, from the start of your output and the beginning of the watermarking. When you don't specify an offset, Kantar defaults to zero.
3767
4152
  */
3768
4153
  FileOffset?: number;
3769
4154
  /**
4155
+ * @public
3770
4156
  * Provide your Kantar license ID number. You should get this number from Kantar.
3771
4157
  */
3772
4158
  KantarLicenseId?: number;
3773
4159
  /**
4160
+ * @public
3774
4161
  * Provide the HTTPS endpoint to the Kantar server. You should get this endpoint from Kantar.
3775
4162
  */
3776
4163
  KantarServerUrl?: string;
3777
4164
  /**
4165
+ * @public
3778
4166
  * Optional. Specify the Amazon S3 bucket where you want MediaConvert to store your Kantar watermark XML logs. When you don't specify a bucket, MediaConvert doesn't save these logs. Note that your MediaConvert service role must provide access to this location. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/iam-role.html
3779
4167
  */
3780
4168
  LogDestination?: string;
3781
4169
  /**
4170
+ * @public
3782
4171
  * You can optionally use this field to specify the first timestamp that Kantar embeds during watermarking. Kantar suggests that you be very cautious when using this Kantar feature, and that you use it only on channels that are managed specifically for use with this feature by your Audience Measurement Operator. For more information about this feature, contact Kantar technical support.
3783
4172
  */
3784
4173
  Metadata3?: string;
3785
4174
  /**
4175
+ * @public
3786
4176
  * Additional metadata that MediaConvert sends to Kantar. Maximum length is 50 characters.
3787
4177
  */
3788
4178
  Metadata4?: string;
3789
4179
  /**
4180
+ * @public
3790
4181
  * Additional metadata that MediaConvert sends to Kantar. Maximum length is 50 characters.
3791
4182
  */
3792
4183
  Metadata5?: string;
3793
4184
  /**
4185
+ * @public
3794
4186
  * Additional metadata that MediaConvert sends to Kantar. Maximum length is 50 characters.
3795
4187
  */
3796
4188
  Metadata6?: string;
3797
4189
  /**
4190
+ * @public
3798
4191
  * Additional metadata that MediaConvert sends to Kantar. Maximum length is 50 characters.
3799
4192
  */
3800
4193
  Metadata7?: string;
3801
4194
  /**
4195
+ * @public
3802
4196
  * Additional metadata that MediaConvert sends to Kantar. Maximum length is 50 characters.
3803
4197
  */
3804
4198
  Metadata8?: string;
@@ -3809,10 +4203,12 @@ export interface KantarWatermarkSettings {
3809
4203
  */
3810
4204
  export interface MotionImageInsertionFramerate {
3811
4205
  /**
4206
+ * @public
3812
4207
  * The bottom of the fraction that expresses your overlay frame rate. For example, if your frame rate is 24 fps, set this value to 1.
3813
4208
  */
3814
4209
  FramerateDenominator?: number;
3815
4210
  /**
4211
+ * @public
3816
4212
  * The top of the fraction that expresses your overlay frame rate. For example, if your frame rate is 24 fps, set this value to 24.
3817
4213
  */
3818
4214
  FramerateNumerator?: number;
@@ -3835,10 +4231,12 @@ export type MotionImageInsertionMode = (typeof MotionImageInsertionMode)[keyof t
3835
4231
  */
3836
4232
  export interface MotionImageInsertionOffset {
3837
4233
  /**
4234
+ * @public
3838
4235
  * Set the distance, in pixels, between the overlay and the left edge of the video frame.
3839
4236
  */
3840
4237
  ImageX?: number;
3841
4238
  /**
4239
+ * @public
3842
4240
  * Set the distance, in pixels, between the overlay and the top edge of the video frame.
3843
4241
  */
3844
4242
  ImageY?: number;
@@ -3861,26 +4259,32 @@ export type MotionImagePlayback = (typeof MotionImagePlayback)[keyof typeof Moti
3861
4259
  */
3862
4260
  export interface MotionImageInserter {
3863
4261
  /**
4262
+ * @public
3864
4263
  * If your motion graphic asset is a .mov file, keep this setting unspecified. If your motion graphic asset is a series of .png files, specify the frame rate of the overlay in frames per second, as a fraction. For example, specify 24 fps as 24/1. Make sure that the number of images in your series matches the frame rate and your intended overlay duration. For example, if you want a 30-second overlay at 30 fps, you should have 900 .png images. This overlay frame rate doesn't need to match the frame rate of the underlying video.
3865
4264
  */
3866
4265
  Framerate?: MotionImageInsertionFramerate;
3867
4266
  /**
4267
+ * @public
3868
4268
  * Specify the .mov file or series of .png files that you want to overlay on your video. For .png files, provide the file name of the first file in the series. Make sure that the names of the .png files end with sequential numbers that specify the order that they are played in. For example, overlay_000.png, overlay_001.png, overlay_002.png, and so on. The sequence must start at zero, and each image file name must have the same number of digits. Pad your initial file names with enough zeros to complete the sequence. For example, if the first image is overlay_0.png, there can be only 10 images in the sequence, with the last image being overlay_9.png. But if the first image is overlay_00.png, there can be 100 images in the sequence.
3869
4269
  */
3870
4270
  Input?: string;
3871
4271
  /**
4272
+ * @public
3872
4273
  * Choose the type of motion graphic asset that you are providing for your overlay. You can choose either a .mov file or a series of .png files.
3873
4274
  */
3874
4275
  InsertionMode?: MotionImageInsertionMode | string;
3875
4276
  /**
4277
+ * @public
3876
4278
  * Use Offset to specify the placement of your motion graphic overlay on the video frame. Specify in pixels, from the upper-left corner of the frame. If you don't specify an offset, the service scales your overlay to the full size of the frame. Otherwise, the service inserts the overlay at its native resolution and scales the size up or down with any video scaling.
3877
4279
  */
3878
4280
  Offset?: MotionImageInsertionOffset;
3879
4281
  /**
4282
+ * @public
3880
4283
  * Specify whether your motion graphic overlay repeats on a loop or plays only once.
3881
4284
  */
3882
4285
  Playback?: MotionImagePlayback | string;
3883
4286
  /**
4287
+ * @public
3884
4288
  * Specify when the motion overlay begins. Use timecode format (HH:MM:SS:FF or HH:MM:SS;FF). Make sure that the timecode you provide here takes into account how you have set up your timecode configuration under both job settings and input settings. The simplest way to do that is to set both to start at 0. If you need to set up your job to follow timecodes embedded in your source that don't start at zero, make sure that you specify a start time that is after the first embedded timecode. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/setting-up-timecode.html
3885
4289
  */
3886
4290
  StartTime?: string;
@@ -3891,10 +4295,12 @@ export interface MotionImageInserter {
3891
4295
  */
3892
4296
  export interface NielsenConfiguration {
3893
4297
  /**
4298
+ * @public
3894
4299
  * Nielsen has discontinued the use of breakout code functionality. If you must include this property, set the value to zero.
3895
4300
  */
3896
4301
  BreakoutCode?: number;
3897
4302
  /**
4303
+ * @public
3898
4304
  * Use Distributor ID to specify the distributor ID that is assigned to your organization by Nielsen.
3899
4305
  */
3900
4306
  DistributorId?: string;
@@ -3942,46 +4348,57 @@ export type NielsenUniqueTicPerAudioTrackType = (typeof NielsenUniqueTicPerAudio
3942
4348
  */
3943
4349
  export interface NielsenNonLinearWatermarkSettings {
3944
4350
  /**
4351
+ * @public
3945
4352
  * Choose the type of Nielsen watermarks that you want in your outputs. When you choose NAES 2 and NW, you must provide a value for the setting SID. When you choose CBET, you must provide a value for the setting CSID. When you choose NAES 2, NW, and CBET, you must provide values for both of these settings.
3946
4353
  */
3947
4354
  ActiveWatermarkProcess?: NielsenActiveWatermarkProcessType | string;
3948
4355
  /**
4356
+ * @public
3949
4357
  * Optional. Use this setting when you want the service to include an ADI file in the Nielsen metadata .zip file. To provide an ADI file, store it in Amazon S3 and provide a URL to it here. The URL should be in the following format: S3://bucket/path/ADI-file. For more information about the metadata .zip file, see the setting Metadata destination.
3950
4358
  */
3951
4359
  AdiFilename?: string;
3952
4360
  /**
4361
+ * @public
3953
4362
  * Use the asset ID that you provide to Nielsen to uniquely identify this asset. Required for all Nielsen non-linear watermarking.
3954
4363
  */
3955
4364
  AssetId?: string;
3956
4365
  /**
4366
+ * @public
3957
4367
  * Use the asset name that you provide to Nielsen for this asset. Required for all Nielsen non-linear watermarking.
3958
4368
  */
3959
4369
  AssetName?: string;
3960
4370
  /**
4371
+ * @public
3961
4372
  * Use the CSID that Nielsen provides to you. This CBET source ID should be unique to your Nielsen account but common to all of your output assets that have CBET watermarking. Required when you choose a value for the setting Watermark types that includes CBET.
3962
4373
  */
3963
4374
  CbetSourceId?: string;
3964
4375
  /**
4376
+ * @public
3965
4377
  * Optional. If this asset uses an episode ID with Nielsen, provide it here.
3966
4378
  */
3967
4379
  EpisodeId?: string;
3968
4380
  /**
4381
+ * @public
3969
4382
  * Specify the Amazon S3 location where you want MediaConvert to save your Nielsen non-linear metadata .zip file. This Amazon S3 bucket must be in the same Region as the one where you do your MediaConvert transcoding. If you want to include an ADI file in this .zip file, use the setting ADI file to specify it. MediaConvert delivers the Nielsen metadata .zip files only to your metadata destination Amazon S3 bucket. It doesn't deliver the .zip files to Nielsen. You are responsible for delivering the metadata .zip files to Nielsen.
3970
4383
  */
3971
4384
  MetadataDestination?: string;
3972
4385
  /**
4386
+ * @public
3973
4387
  * Use the SID that Nielsen provides to you. This source ID should be unique to your Nielsen account but common to all of your output assets. Required for all Nielsen non-linear watermarking. This ID should be unique to your Nielsen account but common to all of your output assets. Required for all Nielsen non-linear watermarking.
3974
4388
  */
3975
4389
  SourceId?: number;
3976
4390
  /**
4391
+ * @public
3977
4392
  * Required. Specify whether your source content already contains Nielsen non-linear watermarks. When you set this value to Watermarked, the service fails the job. Nielsen requires that you add non-linear watermarking to only clean content that doesn't already have non-linear Nielsen watermarks.
3978
4393
  */
3979
4394
  SourceWatermarkStatus?: NielsenSourceWatermarkStatusType | string;
3980
4395
  /**
4396
+ * @public
3981
4397
  * Specify the endpoint for the TIC server that you have deployed and configured in the AWS Cloud. Required for all Nielsen non-linear watermarking. MediaConvert can't connect directly to a TIC server. Instead, you must use API Gateway to provide a RESTful interface between MediaConvert and a TIC server that you deploy in your AWS account. For more information on deploying a TIC server in your AWS account and the required API Gateway, contact Nielsen support.
3982
4398
  */
3983
4399
  TicServerUrl?: string;
3984
4400
  /**
4401
+ * @public
3985
4402
  * To create assets that have the same TIC values in each audio track, keep the default value Share TICs. To create assets that have unique TIC values for each audio track, choose Use unique TICs.
3986
4403
  */
3987
4404
  UniqueTicPerAudioTrack?: NielsenUniqueTicPerAudioTrackType | string;
@@ -3992,18 +4409,22 @@ export interface NielsenNonLinearWatermarkSettings {
3992
4409
  */
3993
4410
  export interface AutomatedAbrSettings {
3994
4411
  /**
4412
+ * @public
3995
4413
  * Optional. The maximum target bit rate used in your automated ABR stack. Use this value to set an upper limit on the bandwidth consumed by the highest-quality rendition. This is the rendition that is delivered to viewers with the fastest internet connections. If you don't specify a value, MediaConvert uses 8,000,000 (8 mb/s) by default.
3996
4414
  */
3997
4415
  MaxAbrBitrate?: number;
3998
4416
  /**
4417
+ * @public
3999
4418
  * Optional. The maximum number of renditions that MediaConvert will create in your automated ABR stack. The number of renditions is determined automatically, based on analysis of each job, but will never exceed this limit. When you set this to Auto in the console, which is equivalent to excluding it from your JSON job specification, MediaConvert defaults to a limit of 15.
4000
4419
  */
4001
4420
  MaxRenditions?: number;
4002
4421
  /**
4422
+ * @public
4003
4423
  * Optional. The minimum target bitrate used in your automated ABR stack. Use this value to set a lower limit on the bitrate of video delivered to viewers with slow internet connections. If you don't specify a value, MediaConvert uses 600,000 (600 kb/s) by default.
4004
4424
  */
4005
4425
  MinAbrBitrate?: number;
4006
4426
  /**
4427
+ * @public
4007
4428
  * Optional. Use Automated ABR rules to specify restrictions for the rendition sizes MediaConvert will create in your ABR stack. You can use these rules if your ABR workflow has specific rendition size requirements, but you still want MediaConvert to optimize for video quality and overall file size.
4008
4429
  */
4009
4430
  Rules?: AutomatedAbrRule[];
@@ -4014,6 +4435,7 @@ export interface AutomatedAbrSettings {
4014
4435
  */
4015
4436
  export interface AutomatedEncodingSettings {
4016
4437
  /**
4438
+ * @public
4017
4439
  * Use automated ABR to have MediaConvert set up the renditions in your ABR package for you automatically, based on characteristics of your input video. This feature optimizes video quality while minimizing the overall size of your ABR package.
4018
4440
  */
4019
4441
  AbrSettings?: AutomatedAbrSettings;
@@ -4075,6 +4497,7 @@ export type S3ObjectCannedAcl = (typeof S3ObjectCannedAcl)[keyof typeof S3Object
4075
4497
  */
4076
4498
  export interface S3DestinationAccessControl {
4077
4499
  /**
4500
+ * @public
4078
4501
  * Choose an Amazon S3 canned ACL for MediaConvert to apply to this output.
4079
4502
  */
4080
4503
  CannedAcl?: S3ObjectCannedAcl | string;
@@ -4097,14 +4520,17 @@ export type S3ServerSideEncryptionType = (typeof S3ServerSideEncryptionType)[key
4097
4520
  */
4098
4521
  export interface S3EncryptionSettings {
4099
4522
  /**
4523
+ * @public
4100
4524
  * Specify how you want your data keys managed. AWS uses data keys to encrypt your content. AWS also encrypts the data keys themselves, using a customer master key (CMK), and then stores the encrypted data keys alongside your encrypted content. Use this setting to specify which AWS service manages the CMK. For simplest set up, choose Amazon S3. If you want your master key to be managed by AWS Key Management Service (KMS), choose AWS KMS. By default, when you choose AWS KMS, KMS uses the AWS managed customer master key (CMK) associated with Amazon S3 to encrypt your data keys. You can optionally choose to specify a different, customer managed CMK. Do so by specifying the Amazon Resource Name (ARN) of the key for the setting KMS ARN.
4101
4525
  */
4102
4526
  EncryptionType?: S3ServerSideEncryptionType | string;
4103
4527
  /**
4528
+ * @public
4104
4529
  * Optionally, specify the encryption context that you want to use alongside your KMS key. AWS KMS uses this encryption context as additional authenticated data (AAD) to support authenticated encryption. This value must be a base64-encoded UTF-8 string holding JSON which represents a string-string map. To use this setting, you must also set Server-side encryption to AWS KMS. For more information about encryption context, see: https://docs.aws.amazon.com/kms/latest/developerguide/concepts.html#encrypt_context.
4105
4530
  */
4106
4531
  KmsEncryptionContext?: string;
4107
4532
  /**
4533
+ * @public
4108
4534
  * Optionally, specify the customer master key (CMK) that you want to use to encrypt the data key that AWS uses to encrypt your output content. Enter the Amazon Resource Name (ARN) of the CMK. To use this setting, you must also set Server-side encryption to AWS KMS. If you set Server-side encryption to AWS KMS but don't specify a CMK here, AWS uses the AWS managed CMK associated with Amazon S3.
4109
4535
  */
4110
4536
  KmsKeyArn?: string;
@@ -4115,10 +4541,12 @@ export interface S3EncryptionSettings {
4115
4541
  */
4116
4542
  export interface S3DestinationSettings {
4117
4543
  /**
4544
+ * @public
4118
4545
  * Optional. Have MediaConvert automatically apply Amazon S3 access control for the outputs in this output group. When you don't use this setting, S3 automatically applies the default access control list PRIVATE.
4119
4546
  */
4120
4547
  AccessControl?: S3DestinationAccessControl;
4121
4548
  /**
4549
+ * @public
4122
4550
  * Settings for how your job outputs are encrypted as they are uploaded to Amazon S3.
4123
4551
  */
4124
4552
  Encryption?: S3EncryptionSettings;
@@ -4129,6 +4557,7 @@ export interface S3DestinationSettings {
4129
4557
  */
4130
4558
  export interface DestinationSettings {
4131
4559
  /**
4560
+ * @public
4132
4561
  * Settings associated with S3 destination
4133
4562
  */
4134
4563
  S3Settings?: S3DestinationSettings;
@@ -4163,22 +4592,27 @@ export type CmafInitializationVectorInManifest = (typeof CmafInitializationVecto
4163
4592
  */
4164
4593
  export interface SpekeKeyProviderCmaf {
4165
4594
  /**
4595
+ * @public
4166
4596
  * If you want your key provider to encrypt the content keys that it provides to MediaConvert, set up a certificate with a master key using AWS Certificate Manager. Specify the certificate's Amazon Resource Name (ARN) here.
4167
4597
  */
4168
4598
  CertificateArn?: string;
4169
4599
  /**
4600
+ * @public
4170
4601
  * Specify the DRM system IDs that you want signaled in the DASH manifest that MediaConvert creates as part of this CMAF package. The DASH manifest can currently signal up to three system IDs. For more information, see https://dashif.org/identifiers/content_protection/.
4171
4602
  */
4172
4603
  DashSignaledSystemIds?: string[];
4173
4604
  /**
4605
+ * @public
4174
4606
  * Specify the DRM system ID that you want signaled in the HLS manifest that MediaConvert creates as part of this CMAF package. The HLS manifest can currently signal only one system ID. For more information, see https://dashif.org/identifiers/content_protection/.
4175
4607
  */
4176
4608
  HlsSignaledSystemIds?: string[];
4177
4609
  /**
4610
+ * @public
4178
4611
  * Specify the resource ID that your SPEKE-compliant key provider uses to identify this content.
4179
4612
  */
4180
4613
  ResourceId?: string;
4181
4614
  /**
4615
+ * @public
4182
4616
  * Specify the URL to the key server that your SPEKE-compliant DRM key provider uses to provide keys for encrypting your content.
4183
4617
  */
4184
4618
  Url?: string;
@@ -4189,18 +4623,22 @@ export interface SpekeKeyProviderCmaf {
4189
4623
  */
4190
4624
  export interface StaticKeyProvider {
4191
4625
  /**
4626
+ * @public
4192
4627
  * Relates to DRM implementation. Sets the value of the KEYFORMAT attribute. Must be 'identity' or a reverse DNS string. May be omitted to indicate an implicit value of 'identity'.
4193
4628
  */
4194
4629
  KeyFormat?: string;
4195
4630
  /**
4631
+ * @public
4196
4632
  * Relates to DRM implementation. Either a single positive integer version value or a slash delimited list of version values (1/2/3).
4197
4633
  */
4198
4634
  KeyFormatVersions?: string;
4199
4635
  /**
4636
+ * @public
4200
4637
  * Relates to DRM implementation. Use a 32-character hexidecimal string to specify Key Value.
4201
4638
  */
4202
4639
  StaticKeyValue?: string;
4203
4640
  /**
4641
+ * @public
4204
4642
  * Relates to DRM implementation. The location of the license server used for protecting content.
4205
4643
  */
4206
4644
  Url?: string;
@@ -4223,26 +4661,32 @@ export type CmafKeyProviderType = (typeof CmafKeyProviderType)[keyof typeof Cmaf
4223
4661
  */
4224
4662
  export interface CmafEncryptionSettings {
4225
4663
  /**
4664
+ * @public
4226
4665
  * This is a 128-bit, 16-byte hex value represented by a 32-character text string. If this parameter is not set then the Initialization Vector will follow the segment number by default.
4227
4666
  */
4228
4667
  ConstantInitializationVector?: string;
4229
4668
  /**
4669
+ * @public
4230
4670
  * Specify the encryption scheme that you want the service to use when encrypting your CMAF segments. Choose AES-CBC subsample or AES_CTR.
4231
4671
  */
4232
4672
  EncryptionMethod?: CmafEncryptionType | string;
4233
4673
  /**
4674
+ * @public
4234
4675
  * When you use DRM with CMAF outputs, choose whether the service writes the 128-bit encryption initialization vector in the HLS and DASH manifests.
4235
4676
  */
4236
4677
  InitializationVectorInManifest?: CmafInitializationVectorInManifest | string;
4237
4678
  /**
4679
+ * @public
4238
4680
  * If your output group type is CMAF, use these settings when doing DRM encryption with a SPEKE-compliant key provider. If your output group type is HLS, DASH, or Microsoft Smooth, use the SpekeKeyProvider settings instead.
4239
4681
  */
4240
4682
  SpekeKeyProvider?: SpekeKeyProviderCmaf;
4241
4683
  /**
4684
+ * @public
4242
4685
  * Use these settings to set up encryption with a static key provider.
4243
4686
  */
4244
4687
  StaticKeyProvider?: StaticKeyProvider;
4245
4688
  /**
4689
+ * @public
4246
4690
  * Specify whether your DRM encryption key is static or from a key provider that follows the SPEKE standard. For more information about SPEKE, see https://docs.aws.amazon.com/speke/latest/documentation/what-is-speke.html.
4247
4691
  */
4248
4692
  Type?: CmafKeyProviderType | string;
@@ -4279,26 +4723,32 @@ export type CmafIntervalCadence = (typeof CmafIntervalCadence)[keyof typeof Cmaf
4279
4723
  */
4280
4724
  export interface CmafImageBasedTrickPlaySettings {
4281
4725
  /**
4726
+ * @public
4282
4727
  * The cadence MediaConvert follows for generating thumbnails. If set to FOLLOW_IFRAME, MediaConvert generates thumbnails for each IDR frame in the output (matching the GOP cadence). If set to FOLLOW_CUSTOM, MediaConvert generates thumbnails according to the interval you specify in thumbnailInterval.
4283
4728
  */
4284
4729
  IntervalCadence?: CmafIntervalCadence | string;
4285
4730
  /**
4731
+ * @public
4286
4732
  * Height of each thumbnail within each tile image, in pixels. Leave blank to maintain aspect ratio with thumbnail width. If following the aspect ratio would lead to a total tile height greater than 4096, then the job will be rejected. Must be divisible by 2.
4287
4733
  */
4288
4734
  ThumbnailHeight?: number;
4289
4735
  /**
4736
+ * @public
4290
4737
  * Enter the interval, in seconds, that MediaConvert uses to generate thumbnails. If the interval you enter doesn't align with the output frame rate, MediaConvert automatically rounds the interval to align with the output frame rate. For example, if the output frame rate is 29.97 frames per second and you enter 5, MediaConvert uses a 150 frame interval to generate thumbnails.
4291
4738
  */
4292
4739
  ThumbnailInterval?: number;
4293
4740
  /**
4741
+ * @public
4294
4742
  * Width of each thumbnail within each tile image, in pixels. Default is 312. Must be divisible by 8.
4295
4743
  */
4296
4744
  ThumbnailWidth?: number;
4297
4745
  /**
4746
+ * @public
4298
4747
  * Number of thumbnails in each column of a tile image. Set a value between 2 and 2048. Must be divisible by 2.
4299
4748
  */
4300
4749
  TileHeight?: number;
4301
4750
  /**
4751
+ * @public
4302
4752
  * Number of thumbnails in each row of a tile image. Set a value between 1 and 512.
4303
4753
  */
4304
4754
  TileWidth?: number;
@@ -4465,110 +4915,137 @@ export type CmafWriteSegmentTimelineInRepresentation = (typeof CmafWriteSegmentT
4465
4915
  */
4466
4916
  export interface CmafGroupSettings {
4467
4917
  /**
4918
+ * @public
4468
4919
  * By default, the service creates one top-level .m3u8 HLS manifest and one top -level .mpd DASH manifest for each CMAF output group in your job. These default manifests reference every output in the output group. To create additional top-level manifests that reference a subset of the outputs in the output group, specify a list of them here. For each additional manifest that you specify, the service creates one HLS manifest and one DASH manifest.
4469
4920
  */
4470
4921
  AdditionalManifests?: CmafAdditionalManifest[];
4471
4922
  /**
4923
+ * @public
4472
4924
  * A partial URI prefix that will be put in the manifest file at the top level BaseURL element. Can be used if streams are delivered from a different URL than the manifest file.
4473
4925
  */
4474
4926
  BaseUrl?: string;
4475
4927
  /**
4928
+ * @public
4476
4929
  * Disable this setting only when your workflow requires the #EXT-X-ALLOW-CACHE:no tag. Otherwise, keep the default value Enabled and control caching in your video distribution set up. For example, use the Cache-Control http header.
4477
4930
  */
4478
4931
  ClientCache?: CmafClientCache | string;
4479
4932
  /**
4933
+ * @public
4480
4934
  * Specification to use (RFC-6381 or the default RFC-4281) during m3u8 playlist generation.
4481
4935
  */
4482
4936
  CodecSpecification?: CmafCodecSpecification | string;
4483
4937
  /**
4938
+ * @public
4484
4939
  * Specify how MediaConvert writes SegmentTimeline in your output DASH manifest. To write a SegmentTimeline in each video Representation: Keep the default value, Basic. To write a common SegmentTimeline in the video AdaptationSet: Choose Compact. Note that MediaConvert will still write a SegmentTimeline in any Representation that does not share a common timeline. To write a video AdaptationSet for each different output framerate, and a common SegmentTimeline in each AdaptationSet: Choose Distinct.
4485
4940
  */
4486
4941
  DashManifestStyle?: DashManifestStyle | string;
4487
4942
  /**
4943
+ * @public
4488
4944
  * Use Destination to specify the S3 output location and the output filename base. Destination accepts format identifiers. If you do not specify the base filename in the URI, the service will use the filename of the input file. If your job has multiple inputs, the service uses the filename of the first input file.
4489
4945
  */
4490
4946
  Destination?: string;
4491
4947
  /**
4948
+ * @public
4492
4949
  * Settings associated with the destination. Will vary based on the type of destination
4493
4950
  */
4494
4951
  DestinationSettings?: DestinationSettings;
4495
4952
  /**
4953
+ * @public
4496
4954
  * DRM settings.
4497
4955
  */
4498
4956
  Encryption?: CmafEncryptionSettings;
4499
4957
  /**
4958
+ * @public
4500
4959
  * Specify the length, in whole seconds, of the mp4 fragments. When you don't specify a value, MediaConvert defaults to 2. Related setting: Use Fragment length control to specify whether the encoder enforces this value strictly.
4501
4960
  */
4502
4961
  FragmentLength?: number;
4503
4962
  /**
4963
+ * @public
4504
4964
  * Specify whether MediaConvert generates images for trick play. Keep the default value, None, to not generate any images. Choose Thumbnail to generate tiled thumbnails. Choose Thumbnail and full frame to generate tiled thumbnails and full-resolution images of single frames. When you enable Write HLS manifest, MediaConvert creates a child manifest for each set of images that you generate and adds corresponding entries to the parent manifest. When you enable Write DASH manifest, MediaConvert adds an entry in the .mpd manifest for each set of images that you generate. A common application for these images is Roku trick mode. The thumbnails and full-frame images that MediaConvert creates with this feature are compatible with this Roku specification: https://developer.roku.com/docs/developer-program/media-playback/trick-mode/hls-and-dash.md
4505
4965
  */
4506
4966
  ImageBasedTrickPlay?: CmafImageBasedTrickPlay | string;
4507
4967
  /**
4968
+ * @public
4508
4969
  * Tile and thumbnail settings applicable when imageBasedTrickPlay is ADVANCED
4509
4970
  */
4510
4971
  ImageBasedTrickPlaySettings?: CmafImageBasedTrickPlaySettings;
4511
4972
  /**
4973
+ * @public
4512
4974
  * When set to GZIP, compresses HLS playlist.
4513
4975
  */
4514
4976
  ManifestCompression?: CmafManifestCompression | string;
4515
4977
  /**
4978
+ * @public
4516
4979
  * Indicates whether the output manifest should use floating point values for segment duration.
4517
4980
  */
4518
4981
  ManifestDurationFormat?: CmafManifestDurationFormat | string;
4519
4982
  /**
4983
+ * @public
4520
4984
  * Minimum time of initially buffered media that is needed to ensure smooth playout.
4521
4985
  */
4522
4986
  MinBufferTime?: number;
4523
4987
  /**
4988
+ * @public
4524
4989
  * Keep this setting at the default value of 0, unless you are troubleshooting a problem with how devices play back the end of your video asset. If you know that player devices are hanging on the final segment of your video because the length of your final segment is too short, use this setting to specify a minimum final segment length, in seconds. Choose a value that is greater than or equal to 1 and less than your segment length. When you specify a value for this setting, the encoder will combine any final segment that is shorter than the length that you specify with the previous segment. For example, your segment length is 3 seconds and your final segment is .5 seconds without a minimum final segment length; when you set the minimum final segment length to 1, your final segment is 3.5 seconds.
4525
4990
  */
4526
4991
  MinFinalSegmentLength?: number;
4527
4992
  /**
4993
+ * @public
4528
4994
  * Specify how the value for bandwidth is determined for each video Representation in your output MPD manifest. We recommend that you choose a MPD manifest bandwidth type that is compatible with your downstream player configuration. Max: Use the same value that you specify for Max bitrate in the video output, in bits per second. Average: Use the calculated average bitrate of the encoded video output, in bits per second.
4529
4995
  */
4530
4996
  MpdManifestBandwidthType?: CmafMpdManifestBandwidthType | string;
4531
4997
  /**
4998
+ * @public
4532
4999
  * Specify whether your DASH profile is on-demand or main. When you choose Main profile, the service signals urn:mpeg:dash:profile:isoff-main:2011 in your .mpd DASH manifest. When you choose On-demand, the service signals urn:mpeg:dash:profile:isoff-on-demand:2011 in your .mpd. When you choose On-demand, you must also set the output group setting Segment control to Single file.
4533
5000
  */
4534
5001
  MpdProfile?: CmafMpdProfile | string;
4535
5002
  /**
5003
+ * @public
4536
5004
  * Use this setting only when your output video stream has B-frames, which causes the initial presentation time stamp (PTS) to be offset from the initial decode time stamp (DTS). Specify how MediaConvert handles PTS when writing time stamps in output DASH manifests. Choose Match initial PTS when you want MediaConvert to use the initial PTS as the first time stamp in the manifest. Choose Zero-based to have MediaConvert ignore the initial PTS in the video stream and instead write the initial time stamp as zero in the manifest. For outputs that don't have B-frames, the time stamps in your DASH manifests start at zero regardless of your choice here.
4537
5005
  */
4538
5006
  PtsOffsetHandlingForBFrames?: CmafPtsOffsetHandlingForBFrames | string;
4539
5007
  /**
5008
+ * @public
4540
5009
  * When set to SINGLE_FILE, a single output file is generated, which is internally segmented using the Fragment Length and Segment Length. When set to SEGMENTED_FILES, separate segment files will be created.
4541
5010
  */
4542
5011
  SegmentControl?: CmafSegmentControl | string;
4543
5012
  /**
5013
+ * @public
4544
5014
  * Specify the length, in whole seconds, of each segment. When you don't specify a value, MediaConvert defaults to 10. Related settings: Use Segment length control to specify whether the encoder enforces this value strictly. Use Segment control to specify whether MediaConvert creates separate segment files or one content file that has metadata to mark the segment boundaries.
4545
5015
  */
4546
5016
  SegmentLength?: number;
4547
5017
  /**
5018
+ * @public
4548
5019
  * Specify how you want MediaConvert to determine the segment length. Choose Exact to have the encoder use the exact length that you specify with the setting Segment length. This might result in extra I-frames. Choose Multiple of GOP to have the encoder round up the segment lengths to match the next GOP boundary.
4549
5020
  */
4550
5021
  SegmentLengthControl?: CmafSegmentLengthControl | string;
4551
5022
  /**
5023
+ * @public
4552
5024
  * Include or exclude RESOLUTION attribute for video in EXT-X-STREAM-INF tag of variant manifest.
4553
5025
  */
4554
5026
  StreamInfResolution?: CmafStreamInfResolution | string;
4555
5027
  /**
5028
+ * @public
4556
5029
  * When set to LEGACY, the segment target duration is always rounded up to the nearest integer value above its current value in seconds. When set to SPEC\\_COMPLIANT, the segment target duration is rounded up to the nearest integer value if fraction seconds are greater than or equal to 0.5 (>= 0.5) and rounded down if less than 0.5 (< 0.5). You may need to use LEGACY if your client needs to ensure that the target duration is always longer than the actual duration of the segment. Some older players may experience interrupted playback when the actual duration of a track in a segment is longer than the target duration.
4557
5030
  */
4558
5031
  TargetDurationCompatibilityMode?: CmafTargetDurationCompatibilityMode | string;
4559
5032
  /**
5033
+ * @public
4560
5034
  * Specify the video sample composition time offset mode in the output fMP4 TRUN box. For wider player compatibility, set Video composition offsets to Unsigned or leave blank. The earliest presentation time may be greater than zero, and sample composition time offsets will increment using unsigned integers. For strict fMP4 video and audio timing, set Video composition offsets to Signed. The earliest presentation time will be equal to zero, and sample composition time offsets will increment using signed integers.
4561
5035
  */
4562
5036
  VideoCompositionOffsets?: CmafVideoCompositionOffsets | string;
4563
5037
  /**
5038
+ * @public
4564
5039
  * When set to ENABLED, a DASH MPD manifest will be generated for this output.
4565
5040
  */
4566
5041
  WriteDashManifest?: CmafWriteDASHManifest | string;
4567
5042
  /**
5043
+ * @public
4568
5044
  * When set to ENABLED, an Apple HLS manifest will be generated for this output.
4569
5045
  */
4570
5046
  WriteHlsManifest?: CmafWriteHLSManifest | string;
4571
5047
  /**
5048
+ * @public
4572
5049
  * When you enable Precise segment duration in DASH manifests, your DASH manifest shows precise segment durations. The segment duration information appears inside the SegmentTimeline element, inside SegmentTemplate at the Representation level. When this feature isn't enabled, the segment durations in your DASH manifest are approximate. The segment duration information appears in the duration attribute of the SegmentTemplate element.
4573
5050
  */
4574
5051
  WriteSegmentTimelineInRepresentation?: CmafWriteSegmentTimelineInRepresentation | string;
@@ -4603,19 +5080,23 @@ export type DashIsoPlaybackDeviceCompatibility = (typeof DashIsoPlaybackDeviceCo
4603
5080
  */
4604
5081
  export interface SpekeKeyProvider {
4605
5082
  /**
5083
+ * @public
4606
5084
  * If you want your key provider to encrypt the content keys that it provides to MediaConvert, set up a certificate with a master key using AWS Certificate Manager. Specify the certificate's Amazon Resource Name (ARN) here.
4607
5085
  */
4608
5086
  CertificateArn?: string;
4609
5087
  /**
5088
+ * @public
4610
5089
  * Specify the resource ID that your SPEKE-compliant key provider uses to identify this content.
4611
5090
  */
4612
5091
  ResourceId?: string;
4613
5092
  /**
5093
+ * @public
4614
5094
  * Relates to SPEKE implementation. DRM system identifiers. DASH output groups support a max of two system ids. Other group types support one system id. See
4615
5095
  * https://dashif.org/identifiers/content_protection/ for more details.
4616
5096
  */
4617
5097
  SystemIds?: string[];
4618
5098
  /**
5099
+ * @public
4619
5100
  * Specify the URL to the key server that your SPEKE-compliant DRM key provider uses to provide keys for encrypting your content.
4620
5101
  */
4621
5102
  Url?: string;
@@ -4626,10 +5107,12 @@ export interface SpekeKeyProvider {
4626
5107
  */
4627
5108
  export interface DashIsoEncryptionSettings {
4628
5109
  /**
5110
+ * @public
4629
5111
  * This setting can improve the compatibility of your output with video players on obsolete devices. It applies only to DASH H.264 outputs with DRM encryption. Choose Unencrypted SEI only to correct problems with playback on older devices. Otherwise, keep the default setting CENC v1. If you choose Unencrypted SEI, for that output, the service will exclude the access unit delimiter and will leave the SEI NAL units unencrypted.
4630
5112
  */
4631
5113
  PlaybackDeviceCompatibility?: DashIsoPlaybackDeviceCompatibility | string;
4632
5114
  /**
5115
+ * @public
4633
5116
  * If your output group type is HLS, DASH, or Microsoft Smooth, use these settings when doing DRM encryption with a SPEKE-compliant key provider. If your output group type is CMAF, use the SpekeKeyProviderCmaf settings instead.
4634
5117
  */
4635
5118
  SpekeKeyProvider?: SpekeKeyProvider;
@@ -4678,26 +5161,32 @@ export type DashIsoIntervalCadence = (typeof DashIsoIntervalCadence)[keyof typeo
4678
5161
  */
4679
5162
  export interface DashIsoImageBasedTrickPlaySettings {
4680
5163
  /**
5164
+ * @public
4681
5165
  * The cadence MediaConvert follows for generating thumbnails. If set to FOLLOW_IFRAME, MediaConvert generates thumbnails for each IDR frame in the output (matching the GOP cadence). If set to FOLLOW_CUSTOM, MediaConvert generates thumbnails according to the interval you specify in thumbnailInterval.
4682
5166
  */
4683
5167
  IntervalCadence?: DashIsoIntervalCadence | string;
4684
5168
  /**
5169
+ * @public
4685
5170
  * Height of each thumbnail within each tile image, in pixels. Leave blank to maintain aspect ratio with thumbnail width. If following the aspect ratio would lead to a total tile height greater than 4096, then the job will be rejected. Must be divisible by 2.
4686
5171
  */
4687
5172
  ThumbnailHeight?: number;
4688
5173
  /**
5174
+ * @public
4689
5175
  * Enter the interval, in seconds, that MediaConvert uses to generate thumbnails. If the interval you enter doesn't align with the output frame rate, MediaConvert automatically rounds the interval to align with the output frame rate. For example, if the output frame rate is 29.97 frames per second and you enter 5, MediaConvert uses a 150 frame interval to generate thumbnails.
4690
5176
  */
4691
5177
  ThumbnailInterval?: number;
4692
5178
  /**
5179
+ * @public
4693
5180
  * Width of each thumbnail within each tile image, in pixels. Default is 312. Must be divisible by 8.
4694
5181
  */
4695
5182
  ThumbnailWidth?: number;
4696
5183
  /**
5184
+ * @public
4697
5185
  * Number of thumbnails in each column of a tile image. Set a value between 2 and 2048. Must be divisible by 2.
4698
5186
  */
4699
5187
  TileHeight?: number;
4700
5188
  /**
5189
+ * @public
4701
5190
  * Number of thumbnails in each row of a tile image. Set a value between 1 and 512.
4702
5191
  */
4703
5192
  TileWidth?: number;
@@ -4792,86 +5281,107 @@ export type DashIsoWriteSegmentTimelineInRepresentation = (typeof DashIsoWriteSe
4792
5281
  */
4793
5282
  export interface DashIsoGroupSettings {
4794
5283
  /**
5284
+ * @public
4795
5285
  * By default, the service creates one .mpd DASH manifest for each DASH ISO output group in your job. This default manifest references every output in the output group. To create additional DASH manifests that reference a subset of the outputs in the output group, specify a list of them here.
4796
5286
  */
4797
5287
  AdditionalManifests?: DashAdditionalManifest[];
4798
5288
  /**
5289
+ * @public
4799
5290
  * Use this setting only when your audio codec is a Dolby one (AC3, EAC3, or Atmos) and your downstream workflow requires that your DASH manifest use the Dolby channel configuration tag, rather than the MPEG one. For example, you might need to use this to make dynamic ad insertion work. Specify which audio channel configuration scheme ID URI MediaConvert writes in your DASH manifest. Keep the default value, MPEG channel configuration, to have MediaConvert write this: urn:mpeg:mpegB:cicp:ChannelConfiguration. Choose Dolby channel configuration to have MediaConvert write this instead: tag:dolby.com,2014:dash:audio_channel_configuration:2011.
4800
5291
  */
4801
5292
  AudioChannelConfigSchemeIdUri?: DashIsoGroupAudioChannelConfigSchemeIdUri | string;
4802
5293
  /**
5294
+ * @public
4803
5295
  * A partial URI prefix that will be put in the manifest (.mpd) file at the top level BaseURL element. Can be used if streams are delivered from a different URL than the manifest file.
4804
5296
  */
4805
5297
  BaseUrl?: string;
4806
5298
  /**
5299
+ * @public
4807
5300
  * Specify how MediaConvert writes SegmentTimeline in your output DASH manifest. To write a SegmentTimeline in each video Representation: Keep the default value, Basic. To write a common SegmentTimeline in the video AdaptationSet: Choose Compact. Note that MediaConvert will still write a SegmentTimeline in any Representation that does not share a common timeline. To write a video AdaptationSet for each different output framerate, and a common SegmentTimeline in each AdaptationSet: Choose Distinct.
4808
5301
  */
4809
5302
  DashManifestStyle?: DashManifestStyle | string;
4810
5303
  /**
5304
+ * @public
4811
5305
  * Use Destination to specify the S3 output location and the output filename base. Destination accepts format identifiers. If you do not specify the base filename in the URI, the service will use the filename of the input file. If your job has multiple inputs, the service uses the filename of the first input file.
4812
5306
  */
4813
5307
  Destination?: string;
4814
5308
  /**
5309
+ * @public
4815
5310
  * Settings associated with the destination. Will vary based on the type of destination
4816
5311
  */
4817
5312
  DestinationSettings?: DestinationSettings;
4818
5313
  /**
5314
+ * @public
4819
5315
  * DRM settings.
4820
5316
  */
4821
5317
  Encryption?: DashIsoEncryptionSettings;
4822
5318
  /**
5319
+ * @public
4823
5320
  * Length of fragments to generate (in seconds). Fragment length must be compatible with GOP size and Framerate. Note that fragments will end on the next keyframe after this number of seconds, so actual fragment length may be longer. When Emit Single File is checked, the fragmentation is internal to a single output file and it does not cause the creation of many output files as in other output types.
4824
5321
  */
4825
5322
  FragmentLength?: number;
4826
5323
  /**
5324
+ * @public
4827
5325
  * Supports HbbTV specification as indicated
4828
5326
  */
4829
5327
  HbbtvCompliance?: DashIsoHbbtvCompliance | string;
4830
5328
  /**
5329
+ * @public
4831
5330
  * Specify whether MediaConvert generates images for trick play. Keep the default value, None, to not generate any images. Choose Thumbnail to generate tiled thumbnails. Choose Thumbnail and full frame to generate tiled thumbnails and full-resolution images of single frames. MediaConvert adds an entry in the .mpd manifest for each set of images that you generate. A common application for these images is Roku trick mode. The thumbnails and full-frame images that MediaConvert creates with this feature are compatible with this Roku specification: https://developer.roku.com/docs/developer-program/media-playback/trick-mode/hls-and-dash.md
4832
5331
  */
4833
5332
  ImageBasedTrickPlay?: DashIsoImageBasedTrickPlay | string;
4834
5333
  /**
5334
+ * @public
4835
5335
  * Tile and thumbnail settings applicable when imageBasedTrickPlay is ADVANCED
4836
5336
  */
4837
5337
  ImageBasedTrickPlaySettings?: DashIsoImageBasedTrickPlaySettings;
4838
5338
  /**
5339
+ * @public
4839
5340
  * Minimum time of initially buffered media that is needed to ensure smooth playout.
4840
5341
  */
4841
5342
  MinBufferTime?: number;
4842
5343
  /**
5344
+ * @public
4843
5345
  * Keep this setting at the default value of 0, unless you are troubleshooting a problem with how devices play back the end of your video asset. If you know that player devices are hanging on the final segment of your video because the length of your final segment is too short, use this setting to specify a minimum final segment length, in seconds. Choose a value that is greater than or equal to 1 and less than your segment length. When you specify a value for this setting, the encoder will combine any final segment that is shorter than the length that you specify with the previous segment. For example, your segment length is 3 seconds and your final segment is .5 seconds without a minimum final segment length; when you set the minimum final segment length to 1, your final segment is 3.5 seconds.
4844
5346
  */
4845
5347
  MinFinalSegmentLength?: number;
4846
5348
  /**
5349
+ * @public
4847
5350
  * Specify how the value for bandwidth is determined for each video Representation in your output MPD manifest. We recommend that you choose a MPD manifest bandwidth type that is compatible with your downstream player configuration. Max: Use the same value that you specify for Max bitrate in the video output, in bits per second. Average: Use the calculated average bitrate of the encoded video output, in bits per second.
4848
5351
  */
4849
5352
  MpdManifestBandwidthType?: DashIsoMpdManifestBandwidthType | string;
4850
5353
  /**
5354
+ * @public
4851
5355
  * Specify whether your DASH profile is on-demand or main. When you choose Main profile, the service signals urn:mpeg:dash:profile:isoff-main:2011 in your .mpd DASH manifest. When you choose On-demand, the service signals urn:mpeg:dash:profile:isoff-on-demand:2011 in your .mpd. When you choose On-demand, you must also set the output group setting Segment control to Single file.
4852
5356
  */
4853
5357
  MpdProfile?: DashIsoMpdProfile | string;
4854
5358
  /**
5359
+ * @public
4855
5360
  * Use this setting only when your output video stream has B-frames, which causes the initial presentation time stamp (PTS) to be offset from the initial decode time stamp (DTS). Specify how MediaConvert handles PTS when writing time stamps in output DASH manifests. Choose Match initial PTS when you want MediaConvert to use the initial PTS as the first time stamp in the manifest. Choose Zero-based to have MediaConvert ignore the initial PTS in the video stream and instead write the initial time stamp as zero in the manifest. For outputs that don't have B-frames, the time stamps in your DASH manifests start at zero regardless of your choice here.
4856
5361
  */
4857
5362
  PtsOffsetHandlingForBFrames?: DashIsoPtsOffsetHandlingForBFrames | string;
4858
5363
  /**
5364
+ * @public
4859
5365
  * When set to SINGLE_FILE, a single output file is generated, which is internally segmented using the Fragment Length and Segment Length. When set to SEGMENTED_FILES, separate segment files will be created.
4860
5366
  */
4861
5367
  SegmentControl?: DashIsoSegmentControl | string;
4862
5368
  /**
5369
+ * @public
4863
5370
  * Specify the length, in whole seconds, of each segment. When you don't specify a value, MediaConvert defaults to 30. Related settings: Use Segment length control to specify whether the encoder enforces this value strictly. Use Segment control to specify whether MediaConvert creates separate segment files or one content file that has metadata to mark the segment boundaries.
4864
5371
  */
4865
5372
  SegmentLength?: number;
4866
5373
  /**
5374
+ * @public
4867
5375
  * Specify how you want MediaConvert to determine the segment length. Choose Exact to have the encoder use the exact length that you specify with the setting Segment length. This might result in extra I-frames. Choose Multiple of GOP to have the encoder round up the segment lengths to match the next GOP boundary.
4868
5376
  */
4869
5377
  SegmentLengthControl?: DashIsoSegmentLengthControl | string;
4870
5378
  /**
5379
+ * @public
4871
5380
  * Specify the video sample composition time offset mode in the output fMP4 TRUN box. For wider player compatibility, set Video composition offsets to Unsigned or leave blank. The earliest presentation time may be greater than zero, and sample composition time offsets will increment using unsigned integers. For strict fMP4 video and audio timing, set Video composition offsets to Signed. The earliest presentation time will be equal to zero, and sample composition time offsets will increment using signed integers.
4872
5381
  */
4873
5382
  VideoCompositionOffsets?: DashIsoVideoCompositionOffsets | string;
4874
5383
  /**
5384
+ * @public
4875
5385
  * If you get an HTTP error in the 400 range when you play back your DASH output, enable this setting and run your transcoding job again. When you enable this setting, the service writes precise segment durations in the DASH manifest. The segment duration information appears inside the SegmentTimeline element, inside SegmentTemplate at the Representation level. When you don't enable this setting, the service writes approximate segment durations in your DASH manifest.
4876
5386
  */
4877
5387
  WriteSegmentTimelineInRepresentation?: DashIsoWriteSegmentTimelineInRepresentation | string;
@@ -4882,10 +5392,12 @@ export interface DashIsoGroupSettings {
4882
5392
  */
4883
5393
  export interface FileGroupSettings {
4884
5394
  /**
5395
+ * @public
4885
5396
  * Use Destination to specify the S3 output location and the output filename base. Destination accepts format identifiers. If you do not specify the base filename in the URI, the service will use the filename of the input file. If your job has multiple inputs, the service uses the filename of the first input file.
4886
5397
  */
4887
5398
  Destination?: string;
4888
5399
  /**
5400
+ * @public
4889
5401
  * Settings associated with the destination. Will vary based on the type of destination
4890
5402
  */
4891
5403
  DestinationSettings?: DestinationSettings;
@@ -5017,30 +5529,37 @@ export type HlsKeyProviderType = (typeof HlsKeyProviderType)[keyof typeof HlsKey
5017
5529
  */
5018
5530
  export interface HlsEncryptionSettings {
5019
5531
  /**
5532
+ * @public
5020
5533
  * This is a 128-bit, 16-byte hex value represented by a 32-character text string. If this parameter is not set then the Initialization Vector will follow the segment number by default.
5021
5534
  */
5022
5535
  ConstantInitializationVector?: string;
5023
5536
  /**
5537
+ * @public
5024
5538
  * Encrypts the segments with the given encryption scheme. Leave blank to disable. Selecting 'Disabled' in the web interface also disables encryption.
5025
5539
  */
5026
5540
  EncryptionMethod?: HlsEncryptionType | string;
5027
5541
  /**
5542
+ * @public
5028
5543
  * The Initialization Vector is a 128-bit number used in conjunction with the key for encrypting blocks. If set to INCLUDE, Initialization Vector is listed in the manifest. Otherwise Initialization Vector is not in the manifest.
5029
5544
  */
5030
5545
  InitializationVectorInManifest?: HlsInitializationVectorInManifest | string;
5031
5546
  /**
5547
+ * @public
5032
5548
  * Enable this setting to insert the EXT-X-SESSION-KEY element into the master playlist. This allows for offline Apple HLS FairPlay content protection.
5033
5549
  */
5034
5550
  OfflineEncrypted?: HlsOfflineEncrypted | string;
5035
5551
  /**
5552
+ * @public
5036
5553
  * If your output group type is HLS, DASH, or Microsoft Smooth, use these settings when doing DRM encryption with a SPEKE-compliant key provider. If your output group type is CMAF, use the SpekeKeyProviderCmaf settings instead.
5037
5554
  */
5038
5555
  SpekeKeyProvider?: SpekeKeyProvider;
5039
5556
  /**
5557
+ * @public
5040
5558
  * Use these settings to set up encryption with a static key provider.
5041
5559
  */
5042
5560
  StaticKeyProvider?: StaticKeyProvider;
5043
5561
  /**
5562
+ * @public
5044
5563
  * Specify whether your DRM encryption key is static or from a key provider that follows the SPEKE standard. For more information about SPEKE, see https://docs.aws.amazon.com/speke/latest/documentation/what-is-speke.html.
5045
5564
  */
5046
5565
  Type?: HlsKeyProviderType | string;
@@ -5077,26 +5596,32 @@ export type HlsIntervalCadence = (typeof HlsIntervalCadence)[keyof typeof HlsInt
5077
5596
  */
5078
5597
  export interface HlsImageBasedTrickPlaySettings {
5079
5598
  /**
5599
+ * @public
5080
5600
  * The cadence MediaConvert follows for generating thumbnails. If set to FOLLOW_IFRAME, MediaConvert generates thumbnails for each IDR frame in the output (matching the GOP cadence). If set to FOLLOW_CUSTOM, MediaConvert generates thumbnails according to the interval you specify in thumbnailInterval.
5081
5601
  */
5082
5602
  IntervalCadence?: HlsIntervalCadence | string;
5083
5603
  /**
5604
+ * @public
5084
5605
  * Height of each thumbnail within each tile image, in pixels. Leave blank to maintain aspect ratio with thumbnail width. If following the aspect ratio would lead to a total tile height greater than 4096, then the job will be rejected. Must be divisible by 2.
5085
5606
  */
5086
5607
  ThumbnailHeight?: number;
5087
5608
  /**
5609
+ * @public
5088
5610
  * Enter the interval, in seconds, that MediaConvert uses to generate thumbnails. If the interval you enter doesn't align with the output frame rate, MediaConvert automatically rounds the interval to align with the output frame rate. For example, if the output frame rate is 29.97 frames per second and you enter 5, MediaConvert uses a 150 frame interval to generate thumbnails.
5089
5611
  */
5090
5612
  ThumbnailInterval?: number;
5091
5613
  /**
5614
+ * @public
5092
5615
  * Width of each thumbnail within each tile image, in pixels. Default is 312. Must be divisible by 8.
5093
5616
  */
5094
5617
  ThumbnailWidth?: number;
5095
5618
  /**
5619
+ * @public
5096
5620
  * Number of thumbnails in each column of a tile image. Set a value between 2 and 2048. Must be divisible by 2.
5097
5621
  */
5098
5622
  TileHeight?: number;
5099
5623
  /**
5624
+ * @public
5100
5625
  * Number of thumbnails in each row of a tile image. Set a value between 1 and 512.
5101
5626
  */
5102
5627
  TileWidth?: number;
@@ -5228,130 +5753,162 @@ export type HlsTimedMetadataId3Frame = (typeof HlsTimedMetadataId3Frame)[keyof t
5228
5753
  */
5229
5754
  export interface HlsGroupSettings {
5230
5755
  /**
5756
+ * @public
5231
5757
  * Choose one or more ad marker types to decorate your Apple HLS manifest. This setting does not determine whether SCTE-35 markers appear in the outputs themselves.
5232
5758
  */
5233
5759
  AdMarkers?: (HlsAdMarkers | string)[];
5234
5760
  /**
5761
+ * @public
5235
5762
  * By default, the service creates one top-level .m3u8 HLS manifest for each HLS output group in your job. This default manifest references every output in the output group. To create additional top-level manifests that reference a subset of the outputs in the output group, specify a list of them here.
5236
5763
  */
5237
5764
  AdditionalManifests?: HlsAdditionalManifest[];
5238
5765
  /**
5766
+ * @public
5239
5767
  * Ignore this setting unless you are using FairPlay DRM with Verimatrix and you encounter playback issues. Keep the default value, Include, to output audio-only headers. Choose Exclude to remove the audio-only headers from your audio segments.
5240
5768
  */
5241
5769
  AudioOnlyHeader?: HlsAudioOnlyHeader | string;
5242
5770
  /**
5771
+ * @public
5243
5772
  * A partial URI prefix that will be prepended to each output in the media .m3u8 file. Can be used if base manifest is delivered from a different URL than the main .m3u8 file.
5244
5773
  */
5245
5774
  BaseUrl?: string;
5246
5775
  /**
5776
+ * @public
5247
5777
  * Language to be used on Caption outputs
5248
5778
  */
5249
5779
  CaptionLanguageMappings?: HlsCaptionLanguageMapping[];
5250
5780
  /**
5781
+ * @public
5251
5782
  * Applies only to 608 Embedded output captions. Insert: Include CLOSED-CAPTIONS lines in the manifest. Specify at least one language in the CC1 Language Code field. One CLOSED-CAPTION line is added for each Language Code you specify. Make sure to specify the languages in the order in which they appear in the original source (if the source is embedded format) or the order of the caption selectors (if the source is other than embedded). Otherwise, languages in the manifest will not match up properly with the output captions. None: Include CLOSED-CAPTIONS=NONE line in the manifest. Omit: Omit any CLOSED-CAPTIONS line from the manifest.
5252
5783
  */
5253
5784
  CaptionLanguageSetting?: HlsCaptionLanguageSetting | string;
5254
5785
  /**
5786
+ * @public
5255
5787
  * Set Caption segment length control to Match video to create caption segments that align with the video segments from the first video output in this output group. For example, if the video segments are 2 seconds long, your WebVTT segments will also be 2 seconds long. Keep the default setting, Large segments to create caption segments that are 300 seconds long.
5256
5788
  */
5257
5789
  CaptionSegmentLengthControl?: HlsCaptionSegmentLengthControl | string;
5258
5790
  /**
5791
+ * @public
5259
5792
  * Disable this setting only when your workflow requires the #EXT-X-ALLOW-CACHE:no tag. Otherwise, keep the default value Enabled and control caching in your video distribution set up. For example, use the Cache-Control http header.
5260
5793
  */
5261
5794
  ClientCache?: HlsClientCache | string;
5262
5795
  /**
5796
+ * @public
5263
5797
  * Specification to use (RFC-6381 or the default RFC-4281) during m3u8 playlist generation.
5264
5798
  */
5265
5799
  CodecSpecification?: HlsCodecSpecification | string;
5266
5800
  /**
5801
+ * @public
5267
5802
  * Use Destination to specify the S3 output location and the output filename base. Destination accepts format identifiers. If you do not specify the base filename in the URI, the service will use the filename of the input file. If your job has multiple inputs, the service uses the filename of the first input file.
5268
5803
  */
5269
5804
  Destination?: string;
5270
5805
  /**
5806
+ * @public
5271
5807
  * Settings associated with the destination. Will vary based on the type of destination
5272
5808
  */
5273
5809
  DestinationSettings?: DestinationSettings;
5274
5810
  /**
5811
+ * @public
5275
5812
  * Indicates whether segments should be placed in subdirectories.
5276
5813
  */
5277
5814
  DirectoryStructure?: HlsDirectoryStructure | string;
5278
5815
  /**
5816
+ * @public
5279
5817
  * DRM settings.
5280
5818
  */
5281
5819
  Encryption?: HlsEncryptionSettings;
5282
5820
  /**
5821
+ * @public
5283
5822
  * Specify whether MediaConvert generates images for trick play. Keep the default value, None, to not generate any images. Choose Thumbnail to generate tiled thumbnails. Choose Thumbnail and full frame to generate tiled thumbnails and full-resolution images of single frames. MediaConvert creates a child manifest for each set of images that you generate and adds corresponding entries to the parent manifest. A common application for these images is Roku trick mode. The thumbnails and full-frame images that MediaConvert creates with this feature are compatible with this Roku specification: https://developer.roku.com/docs/developer-program/media-playback/trick-mode/hls-and-dash.md
5284
5823
  */
5285
5824
  ImageBasedTrickPlay?: HlsImageBasedTrickPlay | string;
5286
5825
  /**
5826
+ * @public
5287
5827
  * Tile and thumbnail settings applicable when imageBasedTrickPlay is ADVANCED
5288
5828
  */
5289
5829
  ImageBasedTrickPlaySettings?: HlsImageBasedTrickPlaySettings;
5290
5830
  /**
5831
+ * @public
5291
5832
  * When set to GZIP, compresses HLS playlist.
5292
5833
  */
5293
5834
  ManifestCompression?: HlsManifestCompression | string;
5294
5835
  /**
5836
+ * @public
5295
5837
  * Indicates whether the output manifest should use floating point values for segment duration.
5296
5838
  */
5297
5839
  ManifestDurationFormat?: HlsManifestDurationFormat | string;
5298
5840
  /**
5841
+ * @public
5299
5842
  * Keep this setting at the default value of 0, unless you are troubleshooting a problem with how devices play back the end of your video asset. If you know that player devices are hanging on the final segment of your video because the length of your final segment is too short, use this setting to specify a minimum final segment length, in seconds. Choose a value that is greater than or equal to 1 and less than your segment length. When you specify a value for this setting, the encoder will combine any final segment that is shorter than the length that you specify with the previous segment. For example, your segment length is 3 seconds and your final segment is .5 seconds without a minimum final segment length; when you set the minimum final segment length to 1, your final segment is 3.5 seconds.
5300
5843
  */
5301
5844
  MinFinalSegmentLength?: number;
5302
5845
  /**
5846
+ * @public
5303
5847
  * When set, Minimum Segment Size is enforced by looking ahead and back within the specified range for a nearby avail and extending the segment size if needed.
5304
5848
  */
5305
5849
  MinSegmentLength?: number;
5306
5850
  /**
5851
+ * @public
5307
5852
  * Indicates whether the .m3u8 manifest file should be generated for this HLS output group.
5308
5853
  */
5309
5854
  OutputSelection?: HlsOutputSelection | string;
5310
5855
  /**
5856
+ * @public
5311
5857
  * Includes or excludes EXT-X-PROGRAM-DATE-TIME tag in .m3u8 manifest files. The value is calculated as follows: either the program date and time are initialized using the input timecode source, or the time is initialized using the input timecode source and the date is initialized using the timestamp_offset.
5312
5858
  */
5313
5859
  ProgramDateTime?: HlsProgramDateTime | string;
5314
5860
  /**
5861
+ * @public
5315
5862
  * Period of insertion of EXT-X-PROGRAM-DATE-TIME entry, in seconds.
5316
5863
  */
5317
5864
  ProgramDateTimePeriod?: number;
5318
5865
  /**
5866
+ * @public
5319
5867
  * Specify whether MediaConvert generates HLS manifests while your job is running or when your job is complete. To generate HLS manifests while your job is running: Choose Enabled. Use if you want to play back your content as soon as it's available. MediaConvert writes the parent and child manifests after the first three media segments are written to your destination S3 bucket. It then writes new updated manifests after each additional segment is written. The parent manifest includes the latest BANDWIDTH and AVERAGE-BANDWIDTH attributes, and child manifests include the latest available media segment. When your job completes, the final child playlists include an EXT-X-ENDLIST tag. To generate HLS manifests only when your job completes: Choose Disabled.
5320
5868
  */
5321
5869
  ProgressiveWriteHlsManifest?: HlsProgressiveWriteHlsManifest | string;
5322
5870
  /**
5871
+ * @public
5323
5872
  * When set to SINGLE_FILE, emits program as a single media resource (.ts) file, uses #EXT-X-BYTERANGE tags to index segment for playback.
5324
5873
  */
5325
5874
  SegmentControl?: HlsSegmentControl | string;
5326
5875
  /**
5876
+ * @public
5327
5877
  * Specify the length, in whole seconds, of each segment. When you don't specify a value, MediaConvert defaults to 10. Related settings: Use Segment length control to specify whether the encoder enforces this value strictly. Use Segment control to specify whether MediaConvert creates separate segment files or one content file that has metadata to mark the segment boundaries.
5328
5878
  */
5329
5879
  SegmentLength?: number;
5330
5880
  /**
5881
+ * @public
5331
5882
  * Specify how you want MediaConvert to determine the segment length. Choose Exact to have the encoder use the exact length that you specify with the setting Segment length. This might result in extra I-frames. Choose Multiple of GOP to have the encoder round up the segment lengths to match the next GOP boundary.
5332
5883
  */
5333
5884
  SegmentLengthControl?: HlsSegmentLengthControl | string;
5334
5885
  /**
5886
+ * @public
5335
5887
  * Specify the number of segments to write to a subdirectory before starting a new one. You must also set Directory structure to Subdirectory per stream for this setting to have an effect.
5336
5888
  */
5337
5889
  SegmentsPerSubdirectory?: number;
5338
5890
  /**
5891
+ * @public
5339
5892
  * Include or exclude RESOLUTION attribute for video in EXT-X-STREAM-INF tag of variant manifest.
5340
5893
  */
5341
5894
  StreamInfResolution?: HlsStreamInfResolution | string;
5342
5895
  /**
5896
+ * @public
5343
5897
  * When set to LEGACY, the segment target duration is always rounded up to the nearest integer value above its current value in seconds. When set to SPEC\\_COMPLIANT, the segment target duration is rounded up to the nearest integer value if fraction seconds are greater than or equal to 0.5 (>= 0.5) and rounded down if less than 0.5 (< 0.5). You may need to use LEGACY if your client needs to ensure that the target duration is always longer than the actual duration of the segment. Some older players may experience interrupted playback when the actual duration of a track in a segment is longer than the target duration.
5344
5898
  */
5345
5899
  TargetDurationCompatibilityMode?: HlsTargetDurationCompatibilityMode | string;
5346
5900
  /**
5901
+ * @public
5347
5902
  * Specify the type of the ID3 frame to use for ID3 timestamps in your output. To include ID3 timestamps: Specify PRIV or TDRL and set ID3 metadata to Passthrough. To exclude ID3 timestamps: Set ID3 timestamp frame type to None.
5348
5903
  */
5349
5904
  TimedMetadataId3Frame?: HlsTimedMetadataId3Frame | string;
5350
5905
  /**
5906
+ * @public
5351
5907
  * Specify the interval in seconds to write ID3 timestamps in your output. The first timestamp starts at the output timecode and date, and increases incrementally with each ID3 timestamp. To use the default interval of 10 seconds: Leave blank. To include this metadata in your output: Set ID3 timestamp frame type to PRIV or TDRL, and set ID3 metadata to Passthrough.
5352
5908
  */
5353
5909
  TimedMetadataId3Period?: number;
5354
5910
  /**
5911
+ * @public
5355
5912
  * Provides an extra millisecond delta offset to fine tune the timestamps.
5356
5913
  */
5357
5914
  TimestampDeltaMilliseconds?: number;
@@ -5362,10 +5919,12 @@ export interface HlsGroupSettings {
5362
5919
  */
5363
5920
  export interface MsSmoothAdditionalManifest {
5364
5921
  /**
5922
+ * @public
5365
5923
  * Specify a name modifier that the service adds to the name of this manifest to make it different from the file names of the other main manifests in the output group. For example, say that the default main manifest for your Microsoft Smooth group is film-name.ismv. If you enter "-no-premium" for this setting, then the file name the service generates for this top-level manifest is film-name-no-premium.ismv.
5366
5924
  */
5367
5925
  ManifestNameModifier?: string;
5368
5926
  /**
5927
+ * @public
5369
5928
  * Specify the outputs that you want this additional top-level manifest to reference.
5370
5929
  */
5371
5930
  SelectedOutputs?: string[];
@@ -5388,6 +5947,7 @@ export type MsSmoothAudioDeduplication = (typeof MsSmoothAudioDeduplication)[key
5388
5947
  */
5389
5948
  export interface MsSmoothEncryptionSettings {
5390
5949
  /**
5950
+ * @public
5391
5951
  * If your output group type is HLS, DASH, or Microsoft Smooth, use these settings when doing DRM encryption with a SPEKE-compliant key provider. If your output group type is CMAF, use the SpekeKeyProviderCmaf settings instead.
5392
5952
  */
5393
5953
  SpekeKeyProvider?: SpekeKeyProvider;
@@ -5422,34 +5982,42 @@ export type MsSmoothManifestEncoding = (typeof MsSmoothManifestEncoding)[keyof t
5422
5982
  */
5423
5983
  export interface MsSmoothGroupSettings {
5424
5984
  /**
5985
+ * @public
5425
5986
  * By default, the service creates one .ism Microsoft Smooth Streaming manifest for each Microsoft Smooth Streaming output group in your job. This default manifest references every output in the output group. To create additional manifests that reference a subset of the outputs in the output group, specify a list of them here.
5426
5987
  */
5427
5988
  AdditionalManifests?: MsSmoothAdditionalManifest[];
5428
5989
  /**
5990
+ * @public
5429
5991
  * COMBINE_DUPLICATE_STREAMS combines identical audio encoding settings across a Microsoft Smooth output group into a single audio stream.
5430
5992
  */
5431
5993
  AudioDeduplication?: MsSmoothAudioDeduplication | string;
5432
5994
  /**
5995
+ * @public
5433
5996
  * Use Destination to specify the S3 output location and the output filename base. Destination accepts format identifiers. If you do not specify the base filename in the URI, the service will use the filename of the input file. If your job has multiple inputs, the service uses the filename of the first input file.
5434
5997
  */
5435
5998
  Destination?: string;
5436
5999
  /**
6000
+ * @public
5437
6001
  * Settings associated with the destination. Will vary based on the type of destination
5438
6002
  */
5439
6003
  DestinationSettings?: DestinationSettings;
5440
6004
  /**
6005
+ * @public
5441
6006
  * If you are using DRM, set DRM System to specify the value SpekeKeyProvider.
5442
6007
  */
5443
6008
  Encryption?: MsSmoothEncryptionSettings;
5444
6009
  /**
6010
+ * @public
5445
6011
  * Specify how you want MediaConvert to determine the fragment length. Choose Exact to have the encoder use the exact length that you specify with the setting Fragment length. This might result in extra I-frames. Choose Multiple of GOP to have the encoder round up the segment lengths to match the next GOP boundary.
5446
6012
  */
5447
6013
  FragmentLength?: number;
5448
6014
  /**
6015
+ * @public
5449
6016
  * Specify how you want MediaConvert to determine the fragment length. Choose Exact to have the encoder use the exact length that you specify with the setting Fragment length. This might result in extra I-frames. Choose Multiple of GOP to have the encoder round up the segment lengths to match the next GOP boundary.
5450
6017
  */
5451
6018
  FragmentLengthControl?: MsSmoothFragmentLengthControl | string;
5452
6019
  /**
6020
+ * @public
5453
6021
  * Use Manifest encoding to specify the encoding format for the server and client manifest. Valid options are utf8 and utf16.
5454
6022
  */
5455
6023
  ManifestEncoding?: MsSmoothManifestEncoding | string;
@@ -5475,26 +6043,32 @@ export type OutputGroupType = (typeof OutputGroupType)[keyof typeof OutputGroupT
5475
6043
  */
5476
6044
  export interface OutputGroupSettings {
5477
6045
  /**
6046
+ * @public
5478
6047
  * Settings related to your CMAF output package. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/outputs-file-ABR.html.
5479
6048
  */
5480
6049
  CmafGroupSettings?: CmafGroupSettings;
5481
6050
  /**
6051
+ * @public
5482
6052
  * Settings related to your DASH output package. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/outputs-file-ABR.html.
5483
6053
  */
5484
6054
  DashIsoGroupSettings?: DashIsoGroupSettings;
5485
6055
  /**
6056
+ * @public
5486
6057
  * Settings related to your File output group. MediaConvert uses this group of settings to generate a single standalone file, rather than a streaming package.
5487
6058
  */
5488
6059
  FileGroupSettings?: FileGroupSettings;
5489
6060
  /**
6061
+ * @public
5490
6062
  * Settings related to your HLS output package. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/outputs-file-ABR.html.
5491
6063
  */
5492
6064
  HlsGroupSettings?: HlsGroupSettings;
5493
6065
  /**
6066
+ * @public
5494
6067
  * Settings related to your Microsoft Smooth Streaming output package. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/outputs-file-ABR.html.
5495
6068
  */
5496
6069
  MsSmoothGroupSettings?: MsSmoothGroupSettings;
5497
6070
  /**
6071
+ * @public
5498
6072
  * Type of output group (File group, Apple HLS, DASH ISO, Microsoft Smooth Streaming, CMAF)
5499
6073
  */
5500
6074
  Type?: OutputGroupType | string;
@@ -5626,50 +6200,62 @@ export type CmfcTimedMetadataBoxVersion = (typeof CmfcTimedMetadataBoxVersion)[k
5626
6200
  */
5627
6201
  export interface CmfcSettings {
5628
6202
  /**
6203
+ * @public
5629
6204
  * Specify this setting only when your output will be consumed by a downstream repackaging workflow that is sensitive to very small duration differences between video and audio. For this situation, choose Match video duration. In all other cases, keep the default value, Default codec duration. When you choose Match video duration, MediaConvert pads the output audio streams with silence or trims them to ensure that the total duration of each audio stream is at least as long as the total duration of the video stream. After padding or trimming, the audio stream duration is no more than one frame longer than the video stream. MediaConvert applies audio padding or trimming only to the end of the last segment of the output. For unsegmented outputs, MediaConvert adds padding only to the end of the file. When you keep the default value, any minor discrepancies between audio and video duration will depend on your output audio codec.
5630
6205
  */
5631
6206
  AudioDuration?: CmfcAudioDuration | string;
5632
6207
  /**
6208
+ * @public
5633
6209
  * Specify the audio rendition group for this audio rendition. Specify up to one value for each audio output in your output group. This value appears in your HLS parent manifest in the EXT-X-MEDIA tag of TYPE=AUDIO, as the value for the GROUP-ID attribute. For example, if you specify "audio_aac_1" for Audio group ID, it appears in your manifest like this: #EXT-X-MEDIA:TYPE=AUDIO,GROUP-ID="audio_aac_1". Related setting: To associate the rendition group that this audio track belongs to with a video rendition, include the same value that you provide here for that video output's setting Audio rendition sets.
5634
6210
  */
5635
6211
  AudioGroupId?: string;
5636
6212
  /**
6213
+ * @public
5637
6214
  * List the audio rendition groups that you want included with this video rendition. Use a comma-separated list. For example, say you want to include the audio rendition groups that have the audio group IDs "audio_aac_1" and "audio_dolby". Then you would specify this value: "audio_aac_1,audio_dolby". Related setting: The rendition groups that you include in your comma-separated list should all match values that you specify in the setting Audio group ID for audio renditions in the same output group as this video rendition. Default behavior: If you don't specify anything here and for Audio group ID, MediaConvert puts each audio variant in its own audio rendition group and associates it with every video variant. Each value in your list appears in your HLS parent manifest in the EXT-X-STREAM-INF tag as the value for the AUDIO attribute. To continue the previous example, say that the file name for the child manifest for your video rendition is "amazing_video_1.m3u8". Then, in your parent manifest, each value will appear on separate lines, like this: #EXT-X-STREAM-INF:AUDIO="audio_aac_1"... amazing_video_1.m3u8 #EXT-X-STREAM-INF:AUDIO="audio_dolby"... amazing_video_1.m3u8
5638
6215
  */
5639
6216
  AudioRenditionSets?: string;
5640
6217
  /**
6218
+ * @public
5641
6219
  * Use this setting to control the values that MediaConvert puts in your HLS parent playlist to control how the client player selects which audio track to play. The other options for this setting determine the values that MediaConvert writes for the DEFAULT and AUTOSELECT attributes of the EXT-X-MEDIA entry for the audio variant. For more information about these attributes, see the Apple documentation article https://developer.apple.com/documentation/http_live_streaming/example_playlists_for_http_live_streaming/adding_alternate_media_to_a_playlist. Choose Alternate audio, auto select, default to set DEFAULT=YES and AUTOSELECT=YES. Choose this value for only one variant in your output group. Choose Alternate audio, auto select, not default to set DEFAULT=NO and AUTOSELECT=YES. Choose Alternate Audio, Not Auto Select to set DEFAULT=NO and AUTOSELECT=NO. When you don't specify a value for this setting, MediaConvert defaults to Alternate audio, auto select, default. When there is more than one variant in your output group, you must explicitly choose a value for this setting.
5642
6220
  */
5643
6221
  AudioTrackType?: CmfcAudioTrackType | string;
5644
6222
  /**
6223
+ * @public
5645
6224
  * Specify whether to flag this audio track as descriptive video service (DVS) in your HLS parent manifest. When you choose Flag, MediaConvert includes the parameter CHARACTERISTICS="public.accessibility.describes-video" in the EXT-X-MEDIA entry for this track. When you keep the default choice, Don't flag, MediaConvert leaves this parameter out. The DVS flag can help with accessibility on Apple devices. For more information, see the Apple documentation.
5646
6225
  */
5647
6226
  DescriptiveVideoServiceFlag?: CmfcDescriptiveVideoServiceFlag | string;
5648
6227
  /**
6228
+ * @public
5649
6229
  * Choose Include to have MediaConvert generate an HLS child manifest that lists only the I-frames for this rendition, in addition to your regular manifest for this rendition. You might use this manifest as part of a workflow that creates preview functions for your video. MediaConvert adds both the I-frame only child manifest and the regular child manifest to the parent manifest. When you don't need the I-frame only child manifest, keep the default value Exclude.
5650
6230
  */
5651
6231
  IFrameOnlyManifest?: CmfcIFrameOnlyManifest | string;
5652
6232
  /**
6233
+ * @public
5653
6234
  * To include key-length-value metadata in this output: Set KLV metadata insertion to Passthrough. MediaConvert reads KLV metadata present in your input and writes each instance to a separate event message box in the output, according to MISB ST1910.1. To exclude this KLV metadata: Set KLV metadata insertion to None or leave blank.
5654
6235
  */
5655
6236
  KlvMetadata?: CmfcKlvMetadata | string;
5656
6237
  /**
6238
+ * @public
5657
6239
  * To add an InbandEventStream element in your output MPD manifest for each type of event message, set Manifest metadata signaling to Enabled. For ID3 event messages, the InbandEventStream element schemeIdUri will be same value that you specify for ID3 metadata scheme ID URI. For SCTE35 event messages, the InbandEventStream element schemeIdUri will be "urn:scte:scte35:2013:bin". To leave these elements out of your output MPD manifest, set Manifest metadata signaling to Disabled. To enable Manifest metadata signaling, you must also set SCTE-35 source to Passthrough, ESAM SCTE-35 to insert, or ID3 metadata to Passthrough.
5658
6240
  */
5659
6241
  ManifestMetadataSignaling?: CmfcManifestMetadataSignaling | string;
5660
6242
  /**
6243
+ * @public
5661
6244
  * Use this setting only when you specify SCTE-35 markers from ESAM. Choose INSERT to put SCTE-35 markers in this output at the insertion points that you specify in an ESAM XML document. Provide the document in the setting SCC XML.
5662
6245
  */
5663
6246
  Scte35Esam?: CmfcScte35Esam | string;
5664
6247
  /**
6248
+ * @public
5665
6249
  * Ignore this setting unless you have SCTE-35 markers in your input video file. Choose Passthrough if you want SCTE-35 markers that appear in your input to also appear in this output. Choose None if you don't want those SCTE-35 markers in this output.
5666
6250
  */
5667
6251
  Scte35Source?: CmfcScte35Source | string;
5668
6252
  /**
6253
+ * @public
5669
6254
  * To include ID3 metadata in this output: Set ID3 metadata to Passthrough. Specify this ID3 metadata in Custom ID3 metadata inserter. MediaConvert writes each instance of ID3 metadata in a separate Event Message (eMSG) box. To exclude this ID3 metadata: Set ID3 metadata to None or leave blank.
5670
6255
  */
5671
6256
  TimedMetadata?: CmfcTimedMetadata | string;
5672
6257
  /**
6258
+ * @public
5673
6259
  * Specify the event message box (eMSG) version for ID3 timed metadata in your output.
5674
6260
  * For more information, see ISO/IEC 23009-1:2022 section 5.10.3.3.3 Syntax.
5675
6261
  * Leave blank to use the default value Version 0.
@@ -5677,10 +6263,12 @@ export interface CmfcSettings {
5677
6263
  */
5678
6264
  TimedMetadataBoxVersion?: CmfcTimedMetadataBoxVersion | string;
5679
6265
  /**
6266
+ * @public
5680
6267
  * Specify the event message box (eMSG) scheme ID URI for ID3 timed metadata in your output. For more information, see ISO/IEC 23009-1:2022 section 5.10.3.3.4 Semantics. Leave blank to use the default value: https://aomedia.org/emsg/ID3 When you specify a value for ID3 metadata scheme ID URI, you must also set ID3 metadata to Passthrough.
5681
6268
  */
5682
6269
  TimedMetadataSchemeIdUri?: string;
5683
6270
  /**
6271
+ * @public
5684
6272
  * Specify the event message box (eMSG) value for ID3 timed metadata in your output. For more information, see ISO/IEC 23009-1:2022 section 5.10.3.3.4 Semantics. When you specify a value for ID3 Metadata Value, you must also set ID3 metadata to Passthrough.
5685
6273
  */
5686
6274
  TimedMetadataValue?: string;