@aws-sdk/client-mediaconvert 3.533.0 → 3.536.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (63) hide show
  1. package/dist-types/MediaConvert.d.ts +8 -1
  2. package/dist-types/MediaConvertClient.d.ts +1 -1
  3. package/dist-types/commands/AssociateCertificateCommand.d.ts +2 -1
  4. package/dist-types/commands/CancelJobCommand.d.ts +2 -1
  5. package/dist-types/commands/CreateJobCommand.d.ts +2 -1
  6. package/dist-types/commands/CreateJobTemplateCommand.d.ts +2 -1
  7. package/dist-types/commands/CreatePresetCommand.d.ts +2 -1
  8. package/dist-types/commands/CreateQueueCommand.d.ts +2 -1
  9. package/dist-types/commands/DeleteJobTemplateCommand.d.ts +2 -1
  10. package/dist-types/commands/DeletePolicyCommand.d.ts +2 -1
  11. package/dist-types/commands/DeletePresetCommand.d.ts +2 -1
  12. package/dist-types/commands/DeleteQueueCommand.d.ts +2 -1
  13. package/dist-types/commands/DescribeEndpointsCommand.d.ts +4 -3
  14. package/dist-types/commands/DisassociateCertificateCommand.d.ts +2 -1
  15. package/dist-types/commands/GetJobCommand.d.ts +2 -1
  16. package/dist-types/commands/GetJobTemplateCommand.d.ts +2 -1
  17. package/dist-types/commands/GetPolicyCommand.d.ts +2 -1
  18. package/dist-types/commands/GetPresetCommand.d.ts +2 -1
  19. package/dist-types/commands/GetQueueCommand.d.ts +2 -1
  20. package/dist-types/commands/ListJobTemplatesCommand.d.ts +2 -1
  21. package/dist-types/commands/ListJobsCommand.d.ts +2 -1
  22. package/dist-types/commands/ListPresetsCommand.d.ts +2 -1
  23. package/dist-types/commands/ListQueuesCommand.d.ts +2 -1
  24. package/dist-types/commands/ListTagsForResourceCommand.d.ts +2 -1
  25. package/dist-types/commands/PutPolicyCommand.d.ts +2 -1
  26. package/dist-types/commands/TagResourceCommand.d.ts +2 -1
  27. package/dist-types/commands/UntagResourceCommand.d.ts +2 -1
  28. package/dist-types/commands/UpdateJobTemplateCommand.d.ts +2 -1
  29. package/dist-types/commands/UpdatePresetCommand.d.ts +2 -1
  30. package/dist-types/commands/UpdateQueueCommand.d.ts +2 -1
  31. package/dist-types/models/models_0.d.ts +709 -709
  32. package/dist-types/models/models_1.d.ts +659 -659
  33. package/dist-types/models/models_2.d.ts +126 -126
  34. package/dist-types/ts3.4/MediaConvert.d.ts +7 -0
  35. package/dist-types/ts3.4/commands/AssociateCertificateCommand.d.ts +9 -0
  36. package/dist-types/ts3.4/commands/CancelJobCommand.d.ts +9 -0
  37. package/dist-types/ts3.4/commands/CreateJobCommand.d.ts +9 -0
  38. package/dist-types/ts3.4/commands/CreateJobTemplateCommand.d.ts +9 -0
  39. package/dist-types/ts3.4/commands/CreatePresetCommand.d.ts +9 -0
  40. package/dist-types/ts3.4/commands/CreateQueueCommand.d.ts +9 -0
  41. package/dist-types/ts3.4/commands/DeleteJobTemplateCommand.d.ts +9 -0
  42. package/dist-types/ts3.4/commands/DeletePolicyCommand.d.ts +9 -0
  43. package/dist-types/ts3.4/commands/DeletePresetCommand.d.ts +9 -0
  44. package/dist-types/ts3.4/commands/DeleteQueueCommand.d.ts +9 -0
  45. package/dist-types/ts3.4/commands/DescribeEndpointsCommand.d.ts +9 -0
  46. package/dist-types/ts3.4/commands/DisassociateCertificateCommand.d.ts +9 -0
  47. package/dist-types/ts3.4/commands/GetJobCommand.d.ts +7 -0
  48. package/dist-types/ts3.4/commands/GetJobTemplateCommand.d.ts +9 -0
  49. package/dist-types/ts3.4/commands/GetPolicyCommand.d.ts +9 -0
  50. package/dist-types/ts3.4/commands/GetPresetCommand.d.ts +9 -0
  51. package/dist-types/ts3.4/commands/GetQueueCommand.d.ts +9 -0
  52. package/dist-types/ts3.4/commands/ListJobTemplatesCommand.d.ts +9 -0
  53. package/dist-types/ts3.4/commands/ListJobsCommand.d.ts +9 -0
  54. package/dist-types/ts3.4/commands/ListPresetsCommand.d.ts +9 -0
  55. package/dist-types/ts3.4/commands/ListQueuesCommand.d.ts +9 -0
  56. package/dist-types/ts3.4/commands/ListTagsForResourceCommand.d.ts +9 -0
  57. package/dist-types/ts3.4/commands/PutPolicyCommand.d.ts +9 -0
  58. package/dist-types/ts3.4/commands/TagResourceCommand.d.ts +9 -0
  59. package/dist-types/ts3.4/commands/UntagResourceCommand.d.ts +9 -0
  60. package/dist-types/ts3.4/commands/UpdateJobTemplateCommand.d.ts +9 -0
  61. package/dist-types/ts3.4/commands/UpdatePresetCommand.d.ts +9 -0
  62. package/dist-types/ts3.4/commands/UpdateQueueCommand.d.ts +9 -0
  63. package/package.json +40 -40
@@ -38,81 +38,81 @@ export declare const CmfcTimedMetadataBoxVersion: {
38
38
  */
39
39
  export type CmfcTimedMetadataBoxVersion = (typeof CmfcTimedMetadataBoxVersion)[keyof typeof CmfcTimedMetadataBoxVersion];
40
40
  /**
41
- * @public
42
41
  * These settings relate to the fragmented MP4 container for the segments in your CMAF outputs.
42
+ * @public
43
43
  */
44
44
  export interface CmfcSettings {
45
45
  /**
46
- * @public
47
46
  * Specify this setting only when your output will be consumed by a downstream repackaging workflow that is sensitive to very small duration differences between video and audio. For this situation, choose Match video duration. In all other cases, keep the default value, Default codec duration. When you choose Match video duration, MediaConvert pads the output audio streams with silence or trims them to ensure that the total duration of each audio stream is at least as long as the total duration of the video stream. After padding or trimming, the audio stream duration is no more than one frame longer than the video stream. MediaConvert applies audio padding or trimming only to the end of the last segment of the output. For unsegmented outputs, MediaConvert adds padding only to the end of the file. When you keep the default value, any minor discrepancies between audio and video duration will depend on your output audio codec.
47
+ * @public
48
48
  */
49
49
  AudioDuration?: CmfcAudioDuration;
50
50
  /**
51
- * @public
52
51
  * Specify the audio rendition group for this audio rendition. Specify up to one value for each audio output in your output group. This value appears in your HLS parent manifest in the EXT-X-MEDIA tag of TYPE=AUDIO, as the value for the GROUP-ID attribute. For example, if you specify "audio_aac_1" for Audio group ID, it appears in your manifest like this: #EXT-X-MEDIA:TYPE=AUDIO,GROUP-ID="audio_aac_1". Related setting: To associate the rendition group that this audio track belongs to with a video rendition, include the same value that you provide here for that video output's setting Audio rendition sets.
52
+ * @public
53
53
  */
54
54
  AudioGroupId?: string;
55
55
  /**
56
- * @public
57
56
  * List the audio rendition groups that you want included with this video rendition. Use a comma-separated list. For example, say you want to include the audio rendition groups that have the audio group IDs "audio_aac_1" and "audio_dolby". Then you would specify this value: "audio_aac_1,audio_dolby". Related setting: The rendition groups that you include in your comma-separated list should all match values that you specify in the setting Audio group ID for audio renditions in the same output group as this video rendition. Default behavior: If you don't specify anything here and for Audio group ID, MediaConvert puts each audio variant in its own audio rendition group and associates it with every video variant. Each value in your list appears in your HLS parent manifest in the EXT-X-STREAM-INF tag as the value for the AUDIO attribute. To continue the previous example, say that the file name for the child manifest for your video rendition is "amazing_video_1.m3u8". Then, in your parent manifest, each value will appear on separate lines, like this: #EXT-X-STREAM-INF:AUDIO="audio_aac_1"... amazing_video_1.m3u8 #EXT-X-STREAM-INF:AUDIO="audio_dolby"... amazing_video_1.m3u8
57
+ * @public
58
58
  */
59
59
  AudioRenditionSets?: string;
60
60
  /**
61
- * @public
62
61
  * Use this setting to control the values that MediaConvert puts in your HLS parent playlist to control how the client player selects which audio track to play. Choose Audio-only variant stream (AUDIO_ONLY_VARIANT_STREAM) for any variant that you want to prohibit the client from playing with video. This causes MediaConvert to represent the variant as an EXT-X-STREAM-INF in the HLS manifest. The other options for this setting determine the values that MediaConvert writes for the DEFAULT and AUTOSELECT attributes of the EXT-X-MEDIA entry for the audio variant. For more information about these attributes, see the Apple documentation article https://developer.apple.com/documentation/http_live_streaming/example_playlists_for_http_live_streaming/adding_alternate_media_to_a_playlist. Choose Alternate audio, auto select, default to set DEFAULT=YES and AUTOSELECT=YES. Choose this value for only one variant in your output group. Choose Alternate audio, auto select, not default to set DEFAULT=NO and AUTOSELECT=YES. Choose Alternate Audio, Not Auto Select to set DEFAULT=NO and AUTOSELECT=NO. When you don't specify a value for this setting, MediaConvert defaults to Alternate audio, auto select, default. When there is more than one variant in your output group, you must explicitly choose a value for this setting.
62
+ * @public
63
63
  */
64
64
  AudioTrackType?: CmfcAudioTrackType;
65
65
  /**
66
- * @public
67
66
  * Specify whether to flag this audio track as descriptive video service (DVS) in your HLS parent manifest. When you choose Flag, MediaConvert includes the parameter CHARACTERISTICS="public.accessibility.describes-video" in the EXT-X-MEDIA entry for this track. When you keep the default choice, Don't flag, MediaConvert leaves this parameter out. The DVS flag can help with accessibility on Apple devices. For more information, see the Apple documentation.
67
+ * @public
68
68
  */
69
69
  DescriptiveVideoServiceFlag?: CmfcDescriptiveVideoServiceFlag;
70
70
  /**
71
- * @public
72
71
  * Choose Include to have MediaConvert generate an HLS child manifest that lists only the I-frames for this rendition, in addition to your regular manifest for this rendition. You might use this manifest as part of a workflow that creates preview functions for your video. MediaConvert adds both the I-frame only child manifest and the regular child manifest to the parent manifest. When you don't need the I-frame only child manifest, keep the default value Exclude.
72
+ * @public
73
73
  */
74
74
  IFrameOnlyManifest?: CmfcIFrameOnlyManifest;
75
75
  /**
76
- * @public
77
76
  * To include key-length-value metadata in this output: Set KLV metadata insertion to Passthrough. MediaConvert reads KLV metadata present in your input and writes each instance to a separate event message box in the output, according to MISB ST1910.1. To exclude this KLV metadata: Set KLV metadata insertion to None or leave blank.
77
+ * @public
78
78
  */
79
79
  KlvMetadata?: CmfcKlvMetadata;
80
80
  /**
81
- * @public
82
81
  * To add an InbandEventStream element in your output MPD manifest for each type of event message, set Manifest metadata signaling to Enabled. For ID3 event messages, the InbandEventStream element schemeIdUri will be same value that you specify for ID3 metadata scheme ID URI. For SCTE35 event messages, the InbandEventStream element schemeIdUri will be "urn:scte:scte35:2013:bin". To leave these elements out of your output MPD manifest, set Manifest metadata signaling to Disabled. To enable Manifest metadata signaling, you must also set SCTE-35 source to Passthrough, ESAM SCTE-35 to insert, or ID3 metadata to Passthrough.
82
+ * @public
83
83
  */
84
84
  ManifestMetadataSignaling?: CmfcManifestMetadataSignaling;
85
85
  /**
86
- * @public
87
86
  * Use this setting only when you specify SCTE-35 markers from ESAM. Choose INSERT to put SCTE-35 markers in this output at the insertion points that you specify in an ESAM XML document. Provide the document in the setting SCC XML.
87
+ * @public
88
88
  */
89
89
  Scte35Esam?: CmfcScte35Esam;
90
90
  /**
91
- * @public
92
91
  * Ignore this setting unless you have SCTE-35 markers in your input video file. Choose Passthrough if you want SCTE-35 markers that appear in your input to also appear in this output. Choose None if you don't want those SCTE-35 markers in this output.
92
+ * @public
93
93
  */
94
94
  Scte35Source?: CmfcScte35Source;
95
95
  /**
96
- * @public
97
96
  * To include ID3 metadata in this output: Set ID3 metadata to Passthrough. Specify this ID3 metadata in Custom ID3 metadata inserter. MediaConvert writes each instance of ID3 metadata in a separate Event Message (eMSG) box. To exclude this ID3 metadata: Set ID3 metadata to None or leave blank.
97
+ * @public
98
98
  */
99
99
  TimedMetadata?: CmfcTimedMetadata;
100
100
  /**
101
- * @public
102
101
  * Specify the event message box (eMSG) version for ID3 timed metadata in your output.
103
102
  * For more information, see ISO/IEC 23009-1:2022 section 5.10.3.3.3 Syntax.
104
103
  * Leave blank to use the default value Version 0.
105
104
  * When you specify Version 1, you must also set ID3 metadata to Passthrough.
105
+ * @public
106
106
  */
107
107
  TimedMetadataBoxVersion?: CmfcTimedMetadataBoxVersion;
108
108
  /**
109
- * @public
110
109
  * Specify the event message box (eMSG) scheme ID URI for ID3 timed metadata in your output. For more information, see ISO/IEC 23009-1:2022 section 5.10.3.3.4 Semantics. Leave blank to use the default value: https://aomedia.org/emsg/ID3 When you specify a value for ID3 metadata scheme ID URI, you must also set ID3 metadata to Passthrough.
110
+ * @public
111
111
  */
112
112
  TimedMetadataSchemeIdUri?: string;
113
113
  /**
114
- * @public
115
114
  * Specify the event message box (eMSG) value for ID3 timed metadata in your output. For more information, see ISO/IEC 23009-1:2022 section 5.10.3.3.4 Semantics. When you specify a value for ID3 Metadata Value, you must also set ID3 metadata to Passthrough.
115
+ * @public
116
116
  */
117
117
  TimedMetadataValue?: string;
118
118
  }
@@ -151,13 +151,13 @@ export declare const F4vMoovPlacement: {
151
151
  */
152
152
  export type F4vMoovPlacement = (typeof F4vMoovPlacement)[keyof typeof F4vMoovPlacement];
153
153
  /**
154
- * @public
155
154
  * Settings for F4v container
155
+ * @public
156
156
  */
157
157
  export interface F4vSettings {
158
158
  /**
159
- * @public
160
159
  * To place the MOOV atom at the beginning of your output, which is useful for progressive downloading: Leave blank or choose Progressive download. To place the MOOV at the end of your output: Choose Normal.
160
+ * @public
161
161
  */
162
162
  MoovPlacement?: F4vMoovPlacement;
163
163
  }
@@ -210,23 +210,23 @@ export declare const M2tsDataPtsControl: {
210
210
  */
211
211
  export type M2tsDataPtsControl = (typeof M2tsDataPtsControl)[keyof typeof M2tsDataPtsControl];
212
212
  /**
213
- * @public
214
213
  * Use these settings to insert a DVB Network Information Table (NIT) in the transport stream of this output.
214
+ * @public
215
215
  */
216
216
  export interface DvbNitSettings {
217
217
  /**
218
- * @public
219
218
  * The numeric value placed in the Network Information Table (NIT).
219
+ * @public
220
220
  */
221
221
  NetworkId?: number;
222
222
  /**
223
- * @public
224
223
  * The network name text placed in the network_name_descriptor inside the Network Information Table. Maximum length is 256 characters.
224
+ * @public
225
225
  */
226
226
  NetworkName?: string;
227
227
  /**
228
- * @public
229
228
  * The number of milliseconds between instances of this table in the output transport stream.
229
+ * @public
230
230
  */
231
231
  NitInterval?: number;
232
232
  }
@@ -245,39 +245,39 @@ export declare const OutputSdt: {
245
245
  */
246
246
  export type OutputSdt = (typeof OutputSdt)[keyof typeof OutputSdt];
247
247
  /**
248
- * @public
249
248
  * Use these settings to insert a DVB Service Description Table (SDT) in the transport stream of this output.
249
+ * @public
250
250
  */
251
251
  export interface DvbSdtSettings {
252
252
  /**
253
- * @public
254
253
  * Selects method of inserting SDT information into output stream. "Follow input SDT" copies SDT information from input stream to output stream. "Follow input SDT if present" copies SDT information from input stream to output stream if SDT information is present in the input, otherwise it will fall back on the user-defined values. Enter "SDT Manually" means user will enter the SDT information. "No SDT" means output stream will not contain SDT information.
254
+ * @public
255
255
  */
256
256
  OutputSdt?: OutputSdt;
257
257
  /**
258
- * @public
259
258
  * The number of milliseconds between instances of this table in the output transport stream.
259
+ * @public
260
260
  */
261
261
  SdtInterval?: number;
262
262
  /**
263
- * @public
264
263
  * The service name placed in the service_descriptor in the Service Description Table. Maximum length is 256 characters.
264
+ * @public
265
265
  */
266
266
  ServiceName?: string;
267
267
  /**
268
- * @public
269
268
  * The service provider name placed in the service_descriptor in the Service Description Table. Maximum length is 256 characters.
269
+ * @public
270
270
  */
271
271
  ServiceProviderName?: string;
272
272
  }
273
273
  /**
274
- * @public
275
274
  * Use these settings to insert a DVB Time and Date Table (TDT) in the transport stream of this output.
275
+ * @public
276
276
  */
277
277
  export interface DvbTdtSettings {
278
278
  /**
279
- * @public
280
279
  * The number of milliseconds between instances of this table in the output transport stream.
280
+ * @public
281
281
  */
282
282
  TdtInterval?: number;
283
283
  }
@@ -390,13 +390,13 @@ export declare const M2tsRateMode: {
390
390
  */
391
391
  export type M2tsRateMode = (typeof M2tsRateMode)[keyof typeof M2tsRateMode];
392
392
  /**
393
- * @public
394
393
  * Settings for SCTE-35 signals from ESAM. Include this in your job settings to put SCTE-35 markers in your HLS and transport stream outputs at the insertion points that you specify in an ESAM XML document. Provide the document in the setting SCC XML.
394
+ * @public
395
395
  */
396
396
  export interface M2tsScte35Esam {
397
397
  /**
398
- * @public
399
398
  * Packet Identifier (PID) of the SCTE-35 stream in the transport stream generated by ESAM.
399
+ * @public
400
400
  */
401
401
  Scte35EsamPid?: number;
402
402
  }
@@ -441,213 +441,213 @@ export declare const M2tsSegmentationStyle: {
441
441
  */
442
442
  export type M2tsSegmentationStyle = (typeof M2tsSegmentationStyle)[keyof typeof M2tsSegmentationStyle];
443
443
  /**
444
- * @public
445
444
  * MPEG-2 TS container settings. These apply to outputs in a File output group when the output's container is MPEG-2 Transport Stream (M2TS). In these assets, data is organized by the program map table (PMT). Each transport stream program contains subsets of data, including audio, video, and metadata. Each of these subsets of data has a numerical label called a packet identifier (PID). Each transport stream program corresponds to one MediaConvert output. The PMT lists the types of data in a program along with their PID. Downstream systems and players use the program map table to look up the PID for each type of data it accesses and then uses the PIDs to locate specific data within the asset.
445
+ * @public
446
446
  */
447
447
  export interface M2tsSettings {
448
448
  /**
449
- * @public
450
449
  * Selects between the DVB and ATSC buffer models for Dolby Digital audio.
450
+ * @public
451
451
  */
452
452
  AudioBufferModel?: M2tsAudioBufferModel;
453
453
  /**
454
- * @public
455
454
  * Specify this setting only when your output will be consumed by a downstream repackaging workflow that is sensitive to very small duration differences between video and audio. For this situation, choose Match video duration. In all other cases, keep the default value, Default codec duration. When you choose Match video duration, MediaConvert pads the output audio streams with silence or trims them to ensure that the total duration of each audio stream is at least as long as the total duration of the video stream. After padding or trimming, the audio stream duration is no more than one frame longer than the video stream. MediaConvert applies audio padding or trimming only to the end of the last segment of the output. For unsegmented outputs, MediaConvert adds padding only to the end of the file. When you keep the default value, any minor discrepancies between audio and video duration will depend on your output audio codec.
455
+ * @public
456
456
  */
457
457
  AudioDuration?: M2tsAudioDuration;
458
458
  /**
459
- * @public
460
459
  * The number of audio frames to insert for each PES packet.
460
+ * @public
461
461
  */
462
462
  AudioFramesPerPes?: number;
463
463
  /**
464
- * @public
465
464
  * Specify the packet identifiers (PIDs) for any elementary audio streams you include in this output. Specify multiple PIDs as a JSON array. Default is the range 482-492.
465
+ * @public
466
466
  */
467
467
  AudioPids?: number[];
468
468
  /**
469
- * @public
470
469
  * Specify the output bitrate of the transport stream in bits per second. Setting to 0 lets the muxer automatically determine the appropriate bitrate. Other common values are 3750000, 7500000, and 15000000.
470
+ * @public
471
471
  */
472
472
  Bitrate?: number;
473
473
  /**
474
- * @public
475
474
  * Controls what buffer model to use for accurate interleaving. If set to MULTIPLEX, use multiplex buffer model. If set to NONE, this can lead to lower latency, but low-memory devices may not be able to play back the stream without interruptions.
475
+ * @public
476
476
  */
477
477
  BufferModel?: M2tsBufferModel;
478
478
  /**
479
- * @public
480
479
  * If you select ALIGN_TO_VIDEO, MediaConvert writes captions and data packets with Presentation Timestamp (PTS) values greater than or equal to the first video packet PTS (MediaConvert drops captions and data packets with lesser PTS values). Keep the default value to allow all PTS values.
480
+ * @public
481
481
  */
482
482
  DataPTSControl?: M2tsDataPtsControl;
483
483
  /**
484
- * @public
485
484
  * Use these settings to insert a DVB Network Information Table (NIT) in the transport stream of this output.
485
+ * @public
486
486
  */
487
487
  DvbNitSettings?: DvbNitSettings;
488
488
  /**
489
- * @public
490
489
  * Use these settings to insert a DVB Service Description Table (SDT) in the transport stream of this output.
490
+ * @public
491
491
  */
492
492
  DvbSdtSettings?: DvbSdtSettings;
493
493
  /**
494
- * @public
495
494
  * Specify the packet identifiers (PIDs) for DVB subtitle data included in this output. Specify multiple PIDs as a JSON array. Default is the range 460-479.
495
+ * @public
496
496
  */
497
497
  DvbSubPids?: number[];
498
498
  /**
499
- * @public
500
499
  * Use these settings to insert a DVB Time and Date Table (TDT) in the transport stream of this output.
500
+ * @public
501
501
  */
502
502
  DvbTdtSettings?: DvbTdtSettings;
503
503
  /**
504
- * @public
505
504
  * Specify the packet identifier (PID) for DVB teletext data you include in this output. Default is 499.
505
+ * @public
506
506
  */
507
507
  DvbTeletextPid?: number;
508
508
  /**
509
- * @public
510
509
  * When set to VIDEO_AND_FIXED_INTERVALS, audio EBP markers will be added to partitions 3 and 4. The interval between these additional markers will be fixed, and will be slightly shorter than the video EBP marker interval. When set to VIDEO_INTERVAL, these additional markers will not be inserted. Only applicable when EBP segmentation markers are is selected (segmentationMarkers is EBP or EBP_LEGACY).
510
+ * @public
511
511
  */
512
512
  EbpAudioInterval?: M2tsEbpAudioInterval;
513
513
  /**
514
- * @public
515
514
  * Selects which PIDs to place EBP markers on. They can either be placed only on the video PID, or on both the video PID and all audio PIDs. Only applicable when EBP segmentation markers are is selected (segmentationMarkers is EBP or EBP_LEGACY).
515
+ * @public
516
516
  */
517
517
  EbpPlacement?: M2tsEbpPlacement;
518
518
  /**
519
- * @public
520
519
  * Controls whether to include the ES Rate field in the PES header.
520
+ * @public
521
521
  */
522
522
  EsRateInPes?: M2tsEsRateInPes;
523
523
  /**
524
- * @public
525
524
  * Keep the default value unless you know that your audio EBP markers are incorrectly appearing before your video EBP markers. To correct this problem, set this value to Force.
525
+ * @public
526
526
  */
527
527
  ForceTsVideoEbpOrder?: M2tsForceTsVideoEbpOrder;
528
528
  /**
529
- * @public
530
529
  * The length, in seconds, of each fragment. Only used with EBP markers.
530
+ * @public
531
531
  */
532
532
  FragmentTime?: number;
533
533
  /**
534
- * @public
535
534
  * To include key-length-value metadata in this output: Set KLV metadata insertion to Passthrough. MediaConvert reads KLV metadata present in your input and passes it through to the output transport stream. To exclude this KLV metadata: Set KLV metadata insertion to None or leave blank.
535
+ * @public
536
536
  */
537
537
  KlvMetadata?: M2tsKlvMetadata;
538
538
  /**
539
- * @public
540
539
  * Specify the maximum time, in milliseconds, between Program Clock References (PCRs) inserted into the transport stream.
540
+ * @public
541
541
  */
542
542
  MaxPcrInterval?: number;
543
543
  /**
544
- * @public
545
544
  * When set, enforces that Encoder Boundary Points do not come within the specified time interval of each other by looking ahead at input video. If another EBP is going to come in within the specified time interval, the current EBP is not emitted, and the segment is "stretched" to the next marker. The lookahead value does not add latency to the system. The Live Event must be configured elsewhere to create sufficient latency to make the lookahead accurate.
545
+ * @public
546
546
  */
547
547
  MinEbpInterval?: number;
548
548
  /**
549
- * @public
550
549
  * If INSERT, Nielsen inaudible tones for media tracking will be detected in the input audio and an equivalent ID3 tag will be inserted in the output.
550
+ * @public
551
551
  */
552
552
  NielsenId3?: M2tsNielsenId3;
553
553
  /**
554
- * @public
555
554
  * Value in bits per second of extra null packets to insert into the transport stream. This can be used if a downstream encryption system requires periodic null packets.
555
+ * @public
556
556
  */
557
557
  NullPacketBitrate?: number;
558
558
  /**
559
- * @public
560
559
  * The number of milliseconds between instances of this table in the output transport stream.
560
+ * @public
561
561
  */
562
562
  PatInterval?: number;
563
563
  /**
564
- * @public
565
564
  * When set to PCR_EVERY_PES_PACKET, a Program Clock Reference value is inserted for every Packetized Elementary Stream (PES) header. This is effective only when the PCR PID is the same as the video or audio elementary stream.
565
+ * @public
566
566
  */
567
567
  PcrControl?: M2tsPcrControl;
568
568
  /**
569
- * @public
570
569
  * Specify the packet identifier (PID) for the program clock reference (PCR) in this output. If you do not specify a value, the service will use the value for Video PID.
570
+ * @public
571
571
  */
572
572
  PcrPid?: number;
573
573
  /**
574
- * @public
575
574
  * Specify the number of milliseconds between instances of the program map table (PMT) in the output transport stream.
575
+ * @public
576
576
  */
577
577
  PmtInterval?: number;
578
578
  /**
579
- * @public
580
579
  * Specify the packet identifier (PID) for the program map table (PMT) itself. Default is 480.
580
+ * @public
581
581
  */
582
582
  PmtPid?: number;
583
583
  /**
584
- * @public
585
584
  * Specify the packet identifier (PID) of the private metadata stream. Default is 503.
585
+ * @public
586
586
  */
587
587
  PrivateMetadataPid?: number;
588
588
  /**
589
- * @public
590
589
  * Use Program number to specify the program number used in the program map table (PMT) for this output. Default is 1. Program numbers and program map tables are parts of MPEG-2 transport stream containers, used for organizing data.
590
+ * @public
591
591
  */
592
592
  ProgramNumber?: number;
593
593
  /**
594
- * @public
595
594
  * Manually specify the initial PTS offset, in seconds, when you set PTS offset to Seconds. Enter an integer from 0 to 3600. Leave blank to keep the default value 2.
595
+ * @public
596
596
  */
597
597
  PtsOffset?: number;
598
598
  /**
599
- * @public
600
599
  * Specify the initial presentation timestamp (PTS) offset for your transport stream output. To let MediaConvert automatically determine the initial PTS offset: Keep the default value, Auto. We recommend that you choose Auto for the widest player compatibility. The initial PTS will be at least two seconds and vary depending on your output's bitrate, HRD buffer size and HRD buffer initial fill percentage. To manually specify an initial PTS offset: Choose Seconds. Then specify the number of seconds with PTS offset.
600
+ * @public
601
601
  */
602
602
  PtsOffsetMode?: TsPtsOffset;
603
603
  /**
604
- * @public
605
604
  * When set to CBR, inserts null packets into transport stream to fill specified bitrate. When set to VBR, the bitrate setting acts as the maximum bitrate, but the output will not be padded up to that bitrate.
605
+ * @public
606
606
  */
607
607
  RateMode?: M2tsRateMode;
608
608
  /**
609
- * @public
610
609
  * Include this in your job settings to put SCTE-35 markers in your HLS and transport stream outputs at the insertion points that you specify in an ESAM XML document. Provide the document in the setting SCC XML.
610
+ * @public
611
611
  */
612
612
  Scte35Esam?: M2tsScte35Esam;
613
613
  /**
614
- * @public
615
614
  * Specify the packet identifier (PID) of the SCTE-35 stream in the transport stream.
615
+ * @public
616
616
  */
617
617
  Scte35Pid?: number;
618
618
  /**
619
- * @public
620
619
  * For SCTE-35 markers from your input-- Choose Passthrough if you want SCTE-35 markers that appear in your input to also appear in this output. Choose None if you don't want SCTE-35 markers in this output. For SCTE-35 markers from an ESAM XML document-- Choose None. Also provide the ESAM XML as a string in the setting Signal processing notification XML. Also enable ESAM SCTE-35 (include the property scte35Esam).
620
+ * @public
621
621
  */
622
622
  Scte35Source?: M2tsScte35Source;
623
623
  /**
624
- * @public
625
624
  * Inserts segmentation markers at each segmentation_time period. rai_segstart sets the Random Access Indicator bit in the adaptation field. rai_adapt sets the RAI bit and adds the current timecode in the private data bytes. psi_segstart inserts PAT and PMT tables at the start of segments. ebp adds Encoder Boundary Point information to the adaptation field as per OpenCable specification OC-SP-EBP-I01-130118. ebp_legacy adds Encoder Boundary Point information to the adaptation field using a legacy proprietary format.
625
+ * @public
626
626
  */
627
627
  SegmentationMarkers?: M2tsSegmentationMarkers;
628
628
  /**
629
- * @public
630
629
  * The segmentation style parameter controls how segmentation markers are inserted into the transport stream. With avails, it is possible that segments may be truncated, which can influence where future segmentation markers are inserted. When a segmentation style of "reset_cadence" is selected and a segment is truncated due to an avail, we will reset the segmentation cadence. This means the subsequent segment will have a duration of of $segmentation_time seconds. When a segmentation style of "maintain_cadence" is selected and a segment is truncated due to an avail, we will not reset the segmentation cadence. This means the subsequent segment will likely be truncated as well. However, all segments after that will have a duration of $segmentation_time seconds. Note that EBP lookahead is a slight exception to this rule.
630
+ * @public
631
631
  */
632
632
  SegmentationStyle?: M2tsSegmentationStyle;
633
633
  /**
634
- * @public
635
634
  * Specify the length, in seconds, of each segment. Required unless markers is set to _none_.
635
+ * @public
636
636
  */
637
637
  SegmentationTime?: number;
638
638
  /**
639
- * @public
640
639
  * Packet Identifier (PID) of the ID3 metadata stream in the transport stream.
640
+ * @public
641
641
  */
642
642
  TimedMetadataPid?: number;
643
643
  /**
644
- * @public
645
644
  * Specify the ID for the transport stream itself in the program map table for this output. Transport stream IDs and program map tables are parts of MPEG-2 transport stream containers, used for organizing data.
645
+ * @public
646
646
  */
647
647
  TransportStreamId?: number;
648
648
  /**
649
- * @public
650
649
  * Specify the packet identifier (PID) of the elementary video stream in the transport stream.
650
+ * @public
651
651
  */
652
652
  VideoPid?: number;
653
653
  }
@@ -724,113 +724,113 @@ export declare const TimedMetadata: {
724
724
  */
725
725
  export type TimedMetadata = (typeof TimedMetadata)[keyof typeof TimedMetadata];
726
726
  /**
727
- * @public
728
727
  * These settings relate to the MPEG-2 transport stream (MPEG2-TS) container for the MPEG2-TS segments in your HLS outputs.
728
+ * @public
729
729
  */
730
730
  export interface M3u8Settings {
731
731
  /**
732
- * @public
733
732
  * Specify this setting only when your output will be consumed by a downstream repackaging workflow that is sensitive to very small duration differences between video and audio. For this situation, choose Match video duration. In all other cases, keep the default value, Default codec duration. When you choose Match video duration, MediaConvert pads the output audio streams with silence or trims them to ensure that the total duration of each audio stream is at least as long as the total duration of the video stream. After padding or trimming, the audio stream duration is no more than one frame longer than the video stream. MediaConvert applies audio padding or trimming only to the end of the last segment of the output. For unsegmented outputs, MediaConvert adds padding only to the end of the file. When you keep the default value, any minor discrepancies between audio and video duration will depend on your output audio codec.
733
+ * @public
734
734
  */
735
735
  AudioDuration?: M3u8AudioDuration;
736
736
  /**
737
- * @public
738
737
  * The number of audio frames to insert for each PES packet.
738
+ * @public
739
739
  */
740
740
  AudioFramesPerPes?: number;
741
741
  /**
742
- * @public
743
742
  * Packet Identifier (PID) of the elementary audio stream(s) in the transport stream. Multiple values are accepted, and can be entered in ranges and/or by comma separation.
743
+ * @public
744
744
  */
745
745
  AudioPids?: number[];
746
746
  /**
747
- * @public
748
747
  * If you select ALIGN_TO_VIDEO, MediaConvert writes captions and data packets with Presentation Timestamp (PTS) values greater than or equal to the first video packet PTS (MediaConvert drops captions and data packets with lesser PTS values). Keep the default value AUTO to allow all PTS values.
748
+ * @public
749
749
  */
750
750
  DataPTSControl?: M3u8DataPtsControl;
751
751
  /**
752
- * @public
753
752
  * Specify the maximum time, in milliseconds, between Program Clock References (PCRs) inserted into the transport stream.
753
+ * @public
754
754
  */
755
755
  MaxPcrInterval?: number;
756
756
  /**
757
- * @public
758
757
  * If INSERT, Nielsen inaudible tones for media tracking will be detected in the input audio and an equivalent ID3 tag will be inserted in the output.
758
+ * @public
759
759
  */
760
760
  NielsenId3?: M3u8NielsenId3;
761
761
  /**
762
- * @public
763
762
  * The number of milliseconds between instances of this table in the output transport stream.
763
+ * @public
764
764
  */
765
765
  PatInterval?: number;
766
766
  /**
767
- * @public
768
767
  * When set to PCR_EVERY_PES_PACKET a Program Clock Reference value is inserted for every Packetized Elementary Stream (PES) header. This parameter is effective only when the PCR PID is the same as the video or audio elementary stream.
768
+ * @public
769
769
  */
770
770
  PcrControl?: M3u8PcrControl;
771
771
  /**
772
- * @public
773
772
  * Packet Identifier (PID) of the Program Clock Reference (PCR) in the transport stream. When no value is given, the encoder will assign the same value as the Video PID.
773
+ * @public
774
774
  */
775
775
  PcrPid?: number;
776
776
  /**
777
- * @public
778
777
  * The number of milliseconds between instances of this table in the output transport stream.
778
+ * @public
779
779
  */
780
780
  PmtInterval?: number;
781
781
  /**
782
- * @public
783
782
  * Packet Identifier (PID) for the Program Map Table (PMT) in the transport stream.
783
+ * @public
784
784
  */
785
785
  PmtPid?: number;
786
786
  /**
787
- * @public
788
787
  * Packet Identifier (PID) of the private metadata stream in the transport stream.
788
+ * @public
789
789
  */
790
790
  PrivateMetadataPid?: number;
791
791
  /**
792
- * @public
793
792
  * The value of the program number field in the Program Map Table.
793
+ * @public
794
794
  */
795
795
  ProgramNumber?: number;
796
796
  /**
797
- * @public
798
797
  * Manually specify the initial PTS offset, in seconds, when you set PTS offset to Seconds. Enter an integer from 0 to 3600. Leave blank to keep the default value 2.
798
+ * @public
799
799
  */
800
800
  PtsOffset?: number;
801
801
  /**
802
- * @public
803
802
  * Specify the initial presentation timestamp (PTS) offset for your transport stream output. To let MediaConvert automatically determine the initial PTS offset: Keep the default value, Auto. We recommend that you choose Auto for the widest player compatibility. The initial PTS will be at least two seconds and vary depending on your output's bitrate, HRD buffer size and HRD buffer initial fill percentage. To manually specify an initial PTS offset: Choose Seconds. Then specify the number of seconds with PTS offset.
803
+ * @public
804
804
  */
805
805
  PtsOffsetMode?: TsPtsOffset;
806
806
  /**
807
- * @public
808
807
  * Packet Identifier (PID) of the SCTE-35 stream in the transport stream.
808
+ * @public
809
809
  */
810
810
  Scte35Pid?: number;
811
811
  /**
812
- * @public
813
812
  * For SCTE-35 markers from your input-- Choose Passthrough if you want SCTE-35 markers that appear in your input to also appear in this output. Choose None if you don't want SCTE-35 markers in this output. For SCTE-35 markers from an ESAM XML document-- Choose None if you don't want manifest conditioning. Choose Passthrough and choose Ad markers if you do want manifest conditioning. In both cases, also provide the ESAM XML as a string in the setting Signal processing notification XML.
813
+ * @public
814
814
  */
815
815
  Scte35Source?: M3u8Scte35Source;
816
816
  /**
817
- * @public
818
817
  * Set ID3 metadata to Passthrough to include ID3 metadata in this output. This includes ID3 metadata from the following features: ID3 timestamp period, and Custom ID3 metadata inserter. To exclude this ID3 metadata in this output: set ID3 metadata to None or leave blank.
818
+ * @public
819
819
  */
820
820
  TimedMetadata?: TimedMetadata;
821
821
  /**
822
- * @public
823
822
  * Packet Identifier (PID) of the ID3 metadata stream in the transport stream.
823
+ * @public
824
824
  */
825
825
  TimedMetadataPid?: number;
826
826
  /**
827
- * @public
828
827
  * The value of the transport stream ID field in the Program Map Table.
828
+ * @public
829
829
  */
830
830
  TransportStreamId?: number;
831
831
  /**
832
- * @public
833
832
  * Packet Identifier (PID) of the elementary video stream in the transport stream.
833
+ * @public
834
834
  */
835
835
  VideoPid?: number;
836
836
  }
@@ -895,33 +895,33 @@ export declare const MovReference: {
895
895
  */
896
896
  export type MovReference = (typeof MovReference)[keyof typeof MovReference];
897
897
  /**
898
- * @public
899
898
  * These settings relate to your QuickTime MOV output container.
899
+ * @public
900
900
  */
901
901
  export interface MovSettings {
902
902
  /**
903
- * @public
904
903
  * When enabled, include 'clap' atom if appropriate for the video output settings.
904
+ * @public
905
905
  */
906
906
  ClapAtom?: MovClapAtom;
907
907
  /**
908
- * @public
909
908
  * When enabled, file composition times will start at zero, composition times in the 'ctts' (composition time to sample) box for B-frames will be negative, and a 'cslg' (composition shift least greatest) box will be included per 14496-1 amendment 1. This improves compatibility with Apple players and tools.
909
+ * @public
910
910
  */
911
911
  CslgAtom?: MovCslgAtom;
912
912
  /**
913
- * @public
914
913
  * When set to XDCAM, writes MPEG2 video streams into the QuickTime file using XDCAM fourcc codes. This increases compatibility with Apple editors and players, but may decrease compatibility with other players. Only applicable when the video codec is MPEG2.
914
+ * @public
915
915
  */
916
916
  Mpeg2FourCCControl?: MovMpeg2FourCCControl;
917
917
  /**
918
- * @public
919
918
  * Unless you need Omneon compatibility: Keep the default value, None. To make this output compatible with Omneon: Choose Omneon. When you do, MediaConvert increases the length of the 'elst' edit list atom. Note that this might cause file rejections when a recipient of the output file doesn't expect this extra padding.
919
+ * @public
920
920
  */
921
921
  PaddingControl?: MovPaddingControl;
922
922
  /**
923
- * @public
924
923
  * Always keep the default value (SELF_CONTAINED) for this setting.
924
+ * @public
925
925
  */
926
926
  Reference?: MovReference;
927
927
  }
@@ -962,38 +962,38 @@ export declare const Mp4MoovPlacement: {
962
962
  */
963
963
  export type Mp4MoovPlacement = (typeof Mp4MoovPlacement)[keyof typeof Mp4MoovPlacement];
964
964
  /**
965
- * @public
966
965
  * These settings relate to your MP4 output container. You can create audio only outputs with this container. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/supported-codecs-containers-audio-only.html#output-codecs-and-containers-supported-for-audio-only.
966
+ * @public
967
967
  */
968
968
  export interface Mp4Settings {
969
969
  /**
970
- * @public
971
970
  * Specify this setting only when your output will be consumed by a downstream repackaging workflow that is sensitive to very small duration differences between video and audio. For this situation, choose Match video duration. In all other cases, keep the default value, Default codec duration. When you choose Match video duration, MediaConvert pads the output audio streams with silence or trims them to ensure that the total duration of each audio stream is at least as long as the total duration of the video stream. After padding or trimming, the audio stream duration is no more than one frame longer than the video stream. MediaConvert applies audio padding or trimming only to the end of the last segment of the output. For unsegmented outputs, MediaConvert adds padding only to the end of the file. When you keep the default value, any minor discrepancies between audio and video duration will depend on your output audio codec.
971
+ * @public
972
972
  */
973
973
  AudioDuration?: CmfcAudioDuration;
974
974
  /**
975
- * @public
976
975
  * When enabled, file composition times will start at zero, composition times in the 'ctts' (composition time to sample) box for B-frames will be negative, and a 'cslg' (composition shift least greatest) box will be included per 14496-1 amendment 1. This improves compatibility with Apple players and tools.
976
+ * @public
977
977
  */
978
978
  CslgAtom?: Mp4CslgAtom;
979
979
  /**
980
- * @public
981
980
  * Ignore this setting unless compliance to the CTTS box version specification matters in your workflow. Specify a value of 1 to set your CTTS box version to 1 and make your output compliant with the specification. When you specify a value of 1, you must also set CSLG atom to the value INCLUDE. Keep the default value 0 to set your CTTS box version to 0. This can provide backward compatibility for some players and packagers.
981
+ * @public
982
982
  */
983
983
  CttsVersion?: number;
984
984
  /**
985
- * @public
986
985
  * Inserts a free-space box immediately after the moov box.
986
+ * @public
987
987
  */
988
988
  FreeSpaceBox?: Mp4FreeSpaceBox;
989
989
  /**
990
- * @public
991
990
  * To place the MOOV atom at the beginning of your output, which is useful for progressive downloading: Leave blank or choose Progressive download. To place the MOOV at the end of your output: Choose Normal.
991
+ * @public
992
992
  */
993
993
  MoovPlacement?: Mp4MoovPlacement;
994
994
  /**
995
- * @public
996
995
  * Overrides the "Major Brand" field in the output file. Usually not necessary to specify.
996
+ * @public
997
997
  */
998
998
  Mp4MajorBrand?: string;
999
999
  }
@@ -1106,66 +1106,66 @@ export declare const MpdTimedMetadataBoxVersion: {
1106
1106
  */
1107
1107
  export type MpdTimedMetadataBoxVersion = (typeof MpdTimedMetadataBoxVersion)[keyof typeof MpdTimedMetadataBoxVersion];
1108
1108
  /**
1109
- * @public
1110
1109
  * These settings relate to the fragmented MP4 container for the segments in your DASH outputs.
1110
+ * @public
1111
1111
  */
1112
1112
  export interface MpdSettings {
1113
1113
  /**
1114
- * @public
1115
1114
  * Optional. Choose Include to have MediaConvert mark up your DASH manifest with <Accessibility> elements for embedded 608 captions. This markup isn't generally required, but some video players require it to discover and play embedded 608 captions. Keep the default value, Exclude, to leave these elements out. When you enable this setting, this is the markup that MediaConvert includes in your manifest: <Accessibility schemeIdUri="urn:scte:dash:cc:cea-608:2015" value="CC1=eng"/>
1115
+ * @public
1116
1116
  */
1117
1117
  AccessibilityCaptionHints?: MpdAccessibilityCaptionHints;
1118
1118
  /**
1119
- * @public
1120
1119
  * Specify this setting only when your output will be consumed by a downstream repackaging workflow that is sensitive to very small duration differences between video and audio. For this situation, choose Match video duration. In all other cases, keep the default value, Default codec duration. When you choose Match video duration, MediaConvert pads the output audio streams with silence or trims them to ensure that the total duration of each audio stream is at least as long as the total duration of the video stream. After padding or trimming, the audio stream duration is no more than one frame longer than the video stream. MediaConvert applies audio padding or trimming only to the end of the last segment of the output. For unsegmented outputs, MediaConvert adds padding only to the end of the file. When you keep the default value, any minor discrepancies between audio and video duration will depend on your output audio codec.
1120
+ * @public
1121
1121
  */
1122
1122
  AudioDuration?: MpdAudioDuration;
1123
1123
  /**
1124
- * @public
1125
1124
  * Use this setting only in DASH output groups that include sidecar TTML or IMSC captions. You specify sidecar captions in a separate output from your audio and video. Choose Raw for captions in a single XML file in a raw container. Choose Fragmented MPEG-4 for captions in XML format contained within fragmented MP4 files. This set of fragmented MP4 files is separate from your video and audio fragmented MP4 files.
1125
+ * @public
1126
1126
  */
1127
1127
  CaptionContainerType?: MpdCaptionContainerType;
1128
1128
  /**
1129
- * @public
1130
1129
  * To include key-length-value metadata in this output: Set KLV metadata insertion to Passthrough. MediaConvert reads KLV metadata present in your input and writes each instance to a separate event message box in the output, according to MISB ST1910.1. To exclude this KLV metadata: Set KLV metadata insertion to None or leave blank.
1130
+ * @public
1131
1131
  */
1132
1132
  KlvMetadata?: MpdKlvMetadata;
1133
1133
  /**
1134
- * @public
1135
1134
  * To add an InbandEventStream element in your output MPD manifest for each type of event message, set Manifest metadata signaling to Enabled. For ID3 event messages, the InbandEventStream element schemeIdUri will be same value that you specify for ID3 metadata scheme ID URI. For SCTE35 event messages, the InbandEventStream element schemeIdUri will be "urn:scte:scte35:2013:bin". To leave these elements out of your output MPD manifest, set Manifest metadata signaling to Disabled. To enable Manifest metadata signaling, you must also set SCTE-35 source to Passthrough, ESAM SCTE-35 to insert, or ID3 metadata to Passthrough.
1135
+ * @public
1136
1136
  */
1137
1137
  ManifestMetadataSignaling?: MpdManifestMetadataSignaling;
1138
1138
  /**
1139
- * @public
1140
1139
  * Use this setting only when you specify SCTE-35 markers from ESAM. Choose INSERT to put SCTE-35 markers in this output at the insertion points that you specify in an ESAM XML document. Provide the document in the setting SCC XML.
1140
+ * @public
1141
1141
  */
1142
1142
  Scte35Esam?: MpdScte35Esam;
1143
1143
  /**
1144
- * @public
1145
1144
  * Ignore this setting unless you have SCTE-35 markers in your input video file. Choose Passthrough if you want SCTE-35 markers that appear in your input to also appear in this output. Choose None if you don't want those SCTE-35 markers in this output.
1145
+ * @public
1146
1146
  */
1147
1147
  Scte35Source?: MpdScte35Source;
1148
1148
  /**
1149
- * @public
1150
1149
  * To include ID3 metadata in this output: Set ID3 metadata to Passthrough. Specify this ID3 metadata in Custom ID3 metadata inserter. MediaConvert writes each instance of ID3 metadata in a separate Event Message (eMSG) box. To exclude this ID3 metadata: Set ID3 metadata to None or leave blank.
1150
+ * @public
1151
1151
  */
1152
1152
  TimedMetadata?: MpdTimedMetadata;
1153
1153
  /**
1154
- * @public
1155
1154
  * Specify the event message box (eMSG) version for ID3 timed metadata in your output.
1156
1155
  * For more information, see ISO/IEC 23009-1:2022 section 5.10.3.3.3 Syntax.
1157
1156
  * Leave blank to use the default value Version 0.
1158
1157
  * When you specify Version 1, you must also set ID3 metadata to Passthrough.
1158
+ * @public
1159
1159
  */
1160
1160
  TimedMetadataBoxVersion?: MpdTimedMetadataBoxVersion;
1161
1161
  /**
1162
- * @public
1163
1162
  * Specify the event message box (eMSG) scheme ID URI for ID3 timed metadata in your output. For more information, see ISO/IEC 23009-1:2022 section 5.10.3.3.4 Semantics. Leave blank to use the default value: https://aomedia.org/emsg/ID3 When you specify a value for ID3 metadata scheme ID URI, you must also set ID3 metadata to Passthrough.
1163
+ * @public
1164
1164
  */
1165
1165
  TimedMetadataSchemeIdUri?: string;
1166
1166
  /**
1167
- * @public
1168
1167
  * Specify the event message box (eMSG) value for ID3 timed metadata in your output. For more information, see ISO/IEC 23009-1:2022 section 5.10.3.3.4 Semantics. When you specify a value for ID3 Metadata Value, you must also set ID3 metadata to Passthrough.
1168
+ * @public
1169
1169
  */
1170
1170
  TimedMetadataValue?: string;
1171
1171
  }
@@ -1209,90 +1209,90 @@ export declare const MxfXavcDurationMode: {
1209
1209
  */
1210
1210
  export type MxfXavcDurationMode = (typeof MxfXavcDurationMode)[keyof typeof MxfXavcDurationMode];
1211
1211
  /**
1212
- * @public
1213
1212
  * Specify the XAVC profile settings for MXF outputs when you set your MXF profile to XAVC.
1213
+ * @public
1214
1214
  */
1215
1215
  export interface MxfXavcProfileSettings {
1216
1216
  /**
1217
- * @public
1218
1217
  * To create an output that complies with the XAVC file format guidelines for interoperability, keep the default value, Drop frames for compliance. To include all frames from your input in this output, keep the default setting, Allow any duration. The number of frames that MediaConvert excludes when you set this to Drop frames for compliance depends on the output frame rate and duration.
1218
+ * @public
1219
1219
  */
1220
1220
  DurationMode?: MxfXavcDurationMode;
1221
1221
  /**
1222
- * @public
1223
1222
  * Specify a value for this setting only for outputs that you set up with one of these two XAVC profiles: XAVC HD Intra CBG or XAVC 4K Intra CBG. Specify the amount of space in each frame that the service reserves for ancillary data, such as teletext captions. The default value for this setting is 1492 bytes per frame. This should be sufficient to prevent overflow unless you have multiple pages of teletext captions data. If you have a large amount of teletext data, specify a larger number.
1223
+ * @public
1224
1224
  */
1225
1225
  MaxAncDataSize?: number;
1226
1226
  }
1227
1227
  /**
1228
- * @public
1229
1228
  * These settings relate to your MXF output container.
1229
+ * @public
1230
1230
  */
1231
1231
  export interface MxfSettings {
1232
1232
  /**
1233
- * @public
1234
1233
  * Optional. When you have AFD signaling set up in your output video stream, use this setting to choose whether to also include it in the MXF wrapper. Choose Don't copy to exclude AFD signaling from the MXF wrapper. Choose Copy from video stream to copy the AFD values from the video stream for this output to the MXF wrapper. Regardless of which option you choose, the AFD values remain in the video stream. Related settings: To set up your output to include or exclude AFD values, see AfdSignaling, under VideoDescription. On the console, find AFD signaling under the output's video encoding settings.
1234
+ * @public
1235
1235
  */
1236
1236
  AfdSignaling?: MxfAfdSignaling;
1237
1237
  /**
1238
- * @public
1239
1238
  * Specify the MXF profile, also called shim, for this output. To automatically select a profile according to your output video codec and resolution, leave blank. For a list of codecs supported with each MXF profile, see https://docs.aws.amazon.com/mediaconvert/latest/ug/codecs-supported-with-each-mxf-profile.html. For more information about the automatic selection behavior, see https://docs.aws.amazon.com/mediaconvert/latest/ug/default-automatic-selection-of-mxf-profiles.html.
1239
+ * @public
1240
1240
  */
1241
1241
  Profile?: MxfProfile;
1242
1242
  /**
1243
- * @public
1244
1243
  * Specify the XAVC profile settings for MXF outputs when you set your MXF profile to XAVC.
1244
+ * @public
1245
1245
  */
1246
1246
  XavcProfileSettings?: MxfXavcProfileSettings;
1247
1247
  }
1248
1248
  /**
1249
- * @public
1250
1249
  * Container specific settings.
1250
+ * @public
1251
1251
  */
1252
1252
  export interface ContainerSettings {
1253
1253
  /**
1254
- * @public
1255
1254
  * These settings relate to the fragmented MP4 container for the segments in your CMAF outputs.
1255
+ * @public
1256
1256
  */
1257
1257
  CmfcSettings?: CmfcSettings;
1258
1258
  /**
1259
- * @public
1260
1259
  * Container for this output. Some containers require a container settings object. If not specified, the default object will be created.
1260
+ * @public
1261
1261
  */
1262
1262
  Container?: ContainerType;
1263
1263
  /**
1264
- * @public
1265
1264
  * Settings for F4v container
1265
+ * @public
1266
1266
  */
1267
1267
  F4vSettings?: F4vSettings;
1268
1268
  /**
1269
- * @public
1270
1269
  * MPEG-2 TS container settings. These apply to outputs in a File output group when the output's container is MPEG-2 Transport Stream (M2TS). In these assets, data is organized by the program map table (PMT). Each transport stream program contains subsets of data, including audio, video, and metadata. Each of these subsets of data has a numerical label called a packet identifier (PID). Each transport stream program corresponds to one MediaConvert output. The PMT lists the types of data in a program along with their PID. Downstream systems and players use the program map table to look up the PID for each type of data it accesses and then uses the PIDs to locate specific data within the asset.
1270
+ * @public
1271
1271
  */
1272
1272
  M2tsSettings?: M2tsSettings;
1273
1273
  /**
1274
- * @public
1275
1274
  * These settings relate to the MPEG-2 transport stream (MPEG2-TS) container for the MPEG2-TS segments in your HLS outputs.
1275
+ * @public
1276
1276
  */
1277
1277
  M3u8Settings?: M3u8Settings;
1278
1278
  /**
1279
- * @public
1280
1279
  * These settings relate to your QuickTime MOV output container.
1280
+ * @public
1281
1281
  */
1282
1282
  MovSettings?: MovSettings;
1283
1283
  /**
1284
- * @public
1285
1284
  * These settings relate to your MP4 output container. You can create audio only outputs with this container. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/supported-codecs-containers-audio-only.html#output-codecs-and-containers-supported-for-audio-only.
1285
+ * @public
1286
1286
  */
1287
1287
  Mp4Settings?: Mp4Settings;
1288
1288
  /**
1289
- * @public
1290
1289
  * These settings relate to the fragmented MP4 container for the segments in your DASH outputs.
1290
+ * @public
1291
1291
  */
1292
1292
  MpdSettings?: MpdSettings;
1293
1293
  /**
1294
- * @public
1295
1294
  * These settings relate to your MXF output container.
1295
+ * @public
1296
1296
  */
1297
1297
  MxfSettings?: MxfSettings;
1298
1298
  }
@@ -1347,54 +1347,54 @@ export declare const HlsIFrameOnlyManifest: {
1347
1347
  */
1348
1348
  export type HlsIFrameOnlyManifest = (typeof HlsIFrameOnlyManifest)[keyof typeof HlsIFrameOnlyManifest];
1349
1349
  /**
1350
- * @public
1351
1350
  * Settings for HLS output groups
1351
+ * @public
1352
1352
  */
1353
1353
  export interface HlsSettings {
1354
1354
  /**
1355
- * @public
1356
1355
  * Specifies the group to which the audio rendition belongs.
1356
+ * @public
1357
1357
  */
1358
1358
  AudioGroupId?: string;
1359
1359
  /**
1360
- * @public
1361
1360
  * Use this setting only in audio-only outputs. Choose MPEG-2 Transport Stream (M2TS) to create a file in an MPEG2-TS container. Keep the default value Automatic to create an audio-only file in a raw container. Regardless of the value that you specify here, if this output has video, the service will place the output into an MPEG2-TS container.
1361
+ * @public
1362
1362
  */
1363
1363
  AudioOnlyContainer?: HlsAudioOnlyContainer;
1364
1364
  /**
1365
- * @public
1366
1365
  * List all the audio groups that are used with the video output stream. Input all the audio GROUP-IDs that are associated to the video, separate by ','.
1366
+ * @public
1367
1367
  */
1368
1368
  AudioRenditionSets?: string;
1369
1369
  /**
1370
- * @public
1371
1370
  * Four types of audio-only tracks are supported: Audio-Only Variant Stream The client can play back this audio-only stream instead of video in low-bandwidth scenarios. Represented as an EXT-X-STREAM-INF in the HLS manifest. Alternate Audio, Auto Select, Default Alternate rendition that the client should try to play back by default. Represented as an EXT-X-MEDIA in the HLS manifest with DEFAULT=YES, AUTOSELECT=YES Alternate Audio, Auto Select, Not Default Alternate rendition that the client may try to play back by default. Represented as an EXT-X-MEDIA in the HLS manifest with DEFAULT=NO, AUTOSELECT=YES Alternate Audio, not Auto Select Alternate rendition that the client will not try to play back by default. Represented as an EXT-X-MEDIA in the HLS manifest with DEFAULT=NO, AUTOSELECT=NO
1371
+ * @public
1372
1372
  */
1373
1373
  AudioTrackType?: HlsAudioTrackType;
1374
1374
  /**
1375
- * @public
1376
1375
  * Specify whether to flag this audio track as descriptive video service (DVS) in your HLS parent manifest. When you choose Flag, MediaConvert includes the parameter CHARACTERISTICS="public.accessibility.describes-video" in the EXT-X-MEDIA entry for this track. When you keep the default choice, Don't flag, MediaConvert leaves this parameter out. The DVS flag can help with accessibility on Apple devices. For more information, see the Apple documentation.
1376
+ * @public
1377
1377
  */
1378
1378
  DescriptiveVideoServiceFlag?: HlsDescriptiveVideoServiceFlag;
1379
1379
  /**
1380
- * @public
1381
1380
  * Choose Include to have MediaConvert generate a child manifest that lists only the I-frames for this rendition, in addition to your regular manifest for this rendition. You might use this manifest as part of a workflow that creates preview functions for your video. MediaConvert adds both the I-frame only child manifest and the regular child manifest to the parent manifest. When you don't need the I-frame only child manifest, keep the default value Exclude.
1381
+ * @public
1382
1382
  */
1383
1383
  IFrameOnlyManifest?: HlsIFrameOnlyManifest;
1384
1384
  /**
1385
- * @public
1386
1385
  * Use this setting to add an identifying string to the filename of each segment. The service adds this string between the name modifier and segment index number. You can use format identifiers in the string. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/using-variables-in-your-job-settings.html
1386
+ * @public
1387
1387
  */
1388
1388
  SegmentModifier?: string;
1389
1389
  }
1390
1390
  /**
1391
- * @public
1392
1391
  * Specific settings for this type of output.
1392
+ * @public
1393
1393
  */
1394
1394
  export interface OutputSettings {
1395
1395
  /**
1396
- * @public
1397
1396
  * Settings for HLS output groups
1397
+ * @public
1398
1398
  */
1399
1399
  HlsSettings?: HlsSettings;
1400
1400
  }
@@ -1489,18 +1489,18 @@ export declare const Av1FramerateConversionAlgorithm: {
1489
1489
  */
1490
1490
  export type Av1FramerateConversionAlgorithm = (typeof Av1FramerateConversionAlgorithm)[keyof typeof Av1FramerateConversionAlgorithm];
1491
1491
  /**
1492
- * @public
1493
1492
  * Settings for quality-defined variable bitrate encoding with the AV1 codec. Use these settings only when you set QVBR for Rate control mode.
1493
+ * @public
1494
1494
  */
1495
1495
  export interface Av1QvbrSettings {
1496
1496
  /**
1497
- * @public
1498
1497
  * Use this setting only when you set Rate control mode to QVBR. Specify the target quality level for this output. MediaConvert determines the right number of bits to use for each part of the video to maintain the video quality that you specify. When you keep the default value, AUTO, MediaConvert picks a quality level for you, based on characteristics of your input video. If you prefer to specify a quality level, specify a number from 1 through 10. Use higher numbers for greater quality. Level 10 results in nearly lossless compression. The quality level for most broadcast-quality transcodes is between 6 and 9. Optionally, to specify a value between whole numbers, also provide a value for the setting qvbrQualityLevelFineTune. For example, if you want your QVBR quality level to be 7.33, set qvbrQualityLevel to 7 and set qvbrQualityLevelFineTune to .33.
1498
+ * @public
1499
1499
  */
1500
1500
  QvbrQualityLevel?: number;
1501
1501
  /**
1502
- * @public
1503
1502
  * Optional. Specify a value here to set the QVBR quality to a level that is between whole numbers. For example, if you want your QVBR quality level to be 7.33, set qvbrQualityLevel to 7 and set qvbrQualityLevelFineTune to .33. MediaConvert rounds your QVBR quality level to the nearest third of a whole number. For example, if you set qvbrQualityLevel to 7 and you set qvbrQualityLevelFineTune to .25, your actual QVBR quality level is 7.33.
1503
+ * @public
1504
1504
  */
1505
1505
  QvbrQualityLevelFineTune?: number;
1506
1506
  }
@@ -1528,78 +1528,78 @@ export declare const Av1SpatialAdaptiveQuantization: {
1528
1528
  */
1529
1529
  export type Av1SpatialAdaptiveQuantization = (typeof Av1SpatialAdaptiveQuantization)[keyof typeof Av1SpatialAdaptiveQuantization];
1530
1530
  /**
1531
- * @public
1532
1531
  * Required when you set Codec, under VideoDescription>CodecSettings to the value AV1.
1532
+ * @public
1533
1533
  */
1534
1534
  export interface Av1Settings {
1535
1535
  /**
1536
- * @public
1537
1536
  * Specify the strength of any adaptive quantization filters that you enable. The value that you choose here applies to Spatial adaptive quantization.
1537
+ * @public
1538
1538
  */
1539
1539
  AdaptiveQuantization?: Av1AdaptiveQuantization;
1540
1540
  /**
1541
- * @public
1542
1541
  * Specify the Bit depth. You can choose 8-bit or 10-bit.
1542
+ * @public
1543
1543
  */
1544
1544
  BitDepth?: Av1BitDepth;
1545
1545
  /**
1546
- * @public
1547
1546
  * Film grain synthesis replaces film grain present in your content with similar quality synthesized AV1 film grain. We recommend that you choose Enabled to reduce the bandwidth of your QVBR quality level 5, 6, 7, or 8 outputs. For QVBR quality level 9 or 10 outputs we recommend that you keep the default value, Disabled. When you include Film grain synthesis, you cannot include the Noise reducer preprocessor.
1547
+ * @public
1548
1548
  */
1549
1549
  FilmGrainSynthesis?: Av1FilmGrainSynthesis;
1550
1550
  /**
1551
- * @public
1552
1551
  * Use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.
1552
+ * @public
1553
1553
  */
1554
1554
  FramerateControl?: Av1FramerateControl;
1555
1555
  /**
1556
- * @public
1557
1556
  * Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate. For numerically simple conversions, such as 60 fps to 30 fps: We recommend that you keep the default value, Drop duplicate. For numerically complex conversions, to avoid stutter: Choose Interpolate. This results in a smooth picture, but might introduce undesirable video artifacts. For complex frame rate conversions, especially if your source video has already been converted from its original cadence: Choose FrameFormer to do motion-compensated interpolation. FrameFormer uses the best conversion method frame by frame. Note that using FrameFormer increases the transcoding time and incurs a significant add-on cost. When you choose FrameFormer, your input video resolution must be at least 128x96.
1557
+ * @public
1558
1558
  */
1559
1559
  FramerateConversionAlgorithm?: Av1FramerateConversionAlgorithm;
1560
1560
  /**
1561
- * @public
1562
1561
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateDenominator to specify the denominator of this fraction. In this example, use 1001 for the value of FramerateDenominator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
1562
+ * @public
1563
1563
  */
1564
1564
  FramerateDenominator?: number;
1565
1565
  /**
1566
- * @public
1567
1566
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateNumerator to specify the numerator of this fraction. In this example, use 24000 for the value of FramerateNumerator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
1567
+ * @public
1568
1568
  */
1569
1569
  FramerateNumerator?: number;
1570
1570
  /**
1571
- * @public
1572
1571
  * Specify the GOP length (keyframe interval) in frames. With AV1, MediaConvert doesn't support GOP length in seconds. This value must be greater than zero and preferably equal to 1 + ((numberBFrames + 1) * x), where x is an integer value.
1572
+ * @public
1573
1573
  */
1574
1574
  GopSize?: number;
1575
1575
  /**
1576
- * @public
1577
1576
  * Maximum bitrate in bits/second. For example, enter five megabits per second as 5000000. Required when Rate control mode is QVBR.
1577
+ * @public
1578
1578
  */
1579
1579
  MaxBitrate?: number;
1580
1580
  /**
1581
- * @public
1582
1581
  * Specify from the number of B-frames, in the range of 0-15. For AV1 encoding, we recommend using 7 or 15. Choose a larger number for a lower bitrate and smaller file size; choose a smaller number for better video quality.
1582
+ * @public
1583
1583
  */
1584
1584
  NumberBFramesBetweenReferenceFrames?: number;
1585
1585
  /**
1586
- * @public
1587
1586
  * Settings for quality-defined variable bitrate encoding with the H.265 codec. Use these settings only when you set QVBR for Rate control mode.
1587
+ * @public
1588
1588
  */
1589
1589
  QvbrSettings?: Av1QvbrSettings;
1590
1590
  /**
1591
- * @public
1592
1591
  * 'With AV1 outputs, for rate control mode, MediaConvert supports only quality-defined variable bitrate (QVBR). You can''t use CBR or VBR.'
1592
+ * @public
1593
1593
  */
1594
1594
  RateControlMode?: Av1RateControlMode;
1595
1595
  /**
1596
- * @public
1597
1596
  * Specify the number of slices per picture. This value must be 1, 2, 4, 8, 16, or 32. For progressive pictures, this value must be less than or equal to the number of macroblock rows. For interlaced pictures, this value must be less than or equal to half the number of macroblock rows.
1597
+ * @public
1598
1598
  */
1599
1599
  Slices?: number;
1600
1600
  /**
1601
- * @public
1602
1601
  * Keep the default value, Enabled, to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher.
1602
+ * @public
1603
1603
  */
1604
1604
  SpatialAdaptiveQuantization?: Av1SpatialAdaptiveQuantization;
1605
1605
  }
@@ -1630,13 +1630,13 @@ export declare const AvcIntraUhdQualityTuningLevel: {
1630
1630
  */
1631
1631
  export type AvcIntraUhdQualityTuningLevel = (typeof AvcIntraUhdQualityTuningLevel)[keyof typeof AvcIntraUhdQualityTuningLevel];
1632
1632
  /**
1633
- * @public
1634
1633
  * Optional when you set AVC-Intra class to Class 4K/2K. When you set AVC-Intra class to a different value, this object isn't allowed.
1634
+ * @public
1635
1635
  */
1636
1636
  export interface AvcIntraUhdSettings {
1637
1637
  /**
1638
- * @public
1639
1638
  * Optional. Use Quality tuning level to choose how many transcoding passes MediaConvert does with your video. When you choose Multi-pass, your video quality is better and your output bitrate is more accurate. That is, the actual bitrate of your output is closer to the target bitrate defined in the specification. When you choose Single-pass, your encoding time is faster. The default behavior is Single-pass.
1639
+ * @public
1640
1640
  */
1641
1641
  QualityTuningLevel?: AvcIntraUhdQualityTuningLevel;
1642
1642
  }
@@ -1717,58 +1717,58 @@ export declare const AvcIntraTelecine: {
1717
1717
  */
1718
1718
  export type AvcIntraTelecine = (typeof AvcIntraTelecine)[keyof typeof AvcIntraTelecine];
1719
1719
  /**
1720
- * @public
1721
1720
  * Required when you choose AVC-Intra for your output video codec. For more information about the AVC-Intra settings, see the relevant specification. For detailed information about SD and HD in AVC-Intra, see https://ieeexplore.ieee.org/document/7290936. For information about 4K/2K in AVC-Intra, see https://pro-av.panasonic.net/en/avc-ultra/AVC-ULTRAoverview.pdf.
1721
+ * @public
1722
1722
  */
1723
1723
  export interface AvcIntraSettings {
1724
1724
  /**
1725
- * @public
1726
1725
  * Specify the AVC-Intra class of your output. The AVC-Intra class selection determines the output video bit rate depending on the frame rate of the output. Outputs with higher class values have higher bitrates and improved image quality. Note that for Class 4K/2K, MediaConvert supports only 4:2:2 chroma subsampling.
1726
+ * @public
1727
1727
  */
1728
1728
  AvcIntraClass?: AvcIntraClass;
1729
1729
  /**
1730
- * @public
1731
1730
  * Optional when you set AVC-Intra class to Class 4K/2K. When you set AVC-Intra class to a different value, this object isn't allowed.
1731
+ * @public
1732
1732
  */
1733
1733
  AvcIntraUhdSettings?: AvcIntraUhdSettings;
1734
1734
  /**
1735
- * @public
1736
1735
  * If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.
1736
+ * @public
1737
1737
  */
1738
1738
  FramerateControl?: AvcIntraFramerateControl;
1739
1739
  /**
1740
- * @public
1741
1740
  * Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate. For numerically simple conversions, such as 60 fps to 30 fps: We recommend that you keep the default value, Drop duplicate. For numerically complex conversions, to avoid stutter: Choose Interpolate. This results in a smooth picture, but might introduce undesirable video artifacts. For complex frame rate conversions, especially if your source video has already been converted from its original cadence: Choose FrameFormer to do motion-compensated interpolation. FrameFormer uses the best conversion method frame by frame. Note that using FrameFormer increases the transcoding time and incurs a significant add-on cost. When you choose FrameFormer, your input video resolution must be at least 128x96.
1741
+ * @public
1742
1742
  */
1743
1743
  FramerateConversionAlgorithm?: AvcIntraFramerateConversionAlgorithm;
1744
1744
  /**
1745
- * @public
1746
1745
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateDenominator to specify the denominator of this fraction. In this example, use 1001 for the value of FramerateDenominator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
1746
+ * @public
1747
1747
  */
1748
1748
  FramerateDenominator?: number;
1749
1749
  /**
1750
- * @public
1751
1750
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateNumerator to specify the numerator of this fraction. In this example, use 24000 for the value of FramerateNumerator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
1751
+ * @public
1752
1752
  */
1753
1753
  FramerateNumerator?: number;
1754
1754
  /**
1755
- * @public
1756
1755
  * Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.
1756
+ * @public
1757
1757
  */
1758
1758
  InterlaceMode?: AvcIntraInterlaceMode;
1759
1759
  /**
1760
- * @public
1761
1760
  * Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive.
1761
+ * @public
1762
1762
  */
1763
1763
  ScanTypeConversionMode?: AvcIntraScanTypeConversionMode;
1764
1764
  /**
1765
- * @public
1766
1765
  * Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25.
1766
+ * @public
1767
1767
  */
1768
1768
  SlowPal?: AvcIntraSlowPal;
1769
1769
  /**
1770
- * @public
1771
1770
  * When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard telecine to create a smoother picture. When you keep the default value, None, MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture.
1771
+ * @public
1772
1772
  */
1773
1773
  Telecine?: AvcIntraTelecine;
1774
1774
  }
@@ -1796,28 +1796,28 @@ export declare const VideoCodec: {
1796
1796
  */
1797
1797
  export type VideoCodec = (typeof VideoCodec)[keyof typeof VideoCodec];
1798
1798
  /**
1799
- * @public
1800
1799
  * Required when you set Codec to the value FRAME_CAPTURE.
1800
+ * @public
1801
1801
  */
1802
1802
  export interface FrameCaptureSettings {
1803
1803
  /**
1804
- * @public
1805
1804
  * Frame capture will encode the first frame of the output stream, then one frame every framerateDenominator/framerateNumerator seconds. For example, settings of framerateNumerator = 1 and framerateDenominator = 3 (a rate of 1/3 frame per second) will capture the first frame, then 1 frame every 3s. Files will be named as filename.n.jpg where n is the 0-based sequence number of each Capture.
1805
+ * @public
1806
1806
  */
1807
1807
  FramerateDenominator?: number;
1808
1808
  /**
1809
- * @public
1810
1809
  * Frame capture will encode the first frame of the output stream, then one frame every framerateDenominator/framerateNumerator seconds. For example, settings of framerateNumerator = 1 and framerateDenominator = 3 (a rate of 1/3 frame per second) will capture the first frame, then 1 frame every 3s. Files will be named as filename.NNNNNNN.jpg where N is the 0-based frame sequence number zero padded to 7 decimal places.
1810
+ * @public
1811
1811
  */
1812
1812
  FramerateNumerator?: number;
1813
1813
  /**
1814
- * @public
1815
1814
  * Maximum number of captures (encoded jpg output files).
1815
+ * @public
1816
1816
  */
1817
1817
  MaxCaptures?: number;
1818
1818
  /**
1819
- * @public
1820
1819
  * JPEG Quality - a higher value equals higher quality.
1820
+ * @public
1821
1821
  */
1822
1822
  Quality?: number;
1823
1823
  }
@@ -1868,18 +1868,18 @@ export declare const BandwidthReductionFilterStrength: {
1868
1868
  */
1869
1869
  export type BandwidthReductionFilterStrength = (typeof BandwidthReductionFilterStrength)[keyof typeof BandwidthReductionFilterStrength];
1870
1870
  /**
1871
- * @public
1872
1871
  * The Bandwidth reduction filter increases the video quality of your output relative to its bitrate. Use to lower the bitrate of your constant quality QVBR output, with little or no perceptual decrease in quality. Or, use to increase the video quality of outputs with other rate control modes relative to the bitrate that you specify. Bandwidth reduction increases further when your input is low quality or noisy. Outputs that use this feature incur pro-tier pricing. When you include Bandwidth reduction filter, you cannot include the Noise reducer preprocessor.
1872
+ * @public
1873
1873
  */
1874
1874
  export interface BandwidthReductionFilter {
1875
1875
  /**
1876
- * @public
1877
1876
  * Optionally specify the level of sharpening to apply when you use the Bandwidth reduction filter. Sharpening adds contrast to the edges of your video content and can reduce softness. Keep the default value Off to apply no sharpening. Set Sharpening strength to Low to apply a minimal amount of sharpening, or High to apply a maximum amount of sharpening.
1877
+ * @public
1878
1878
  */
1879
1879
  Sharpening?: BandwidthReductionFilterSharpening;
1880
1880
  /**
1881
- * @public
1882
1881
  * Specify the strength of the Bandwidth reduction filter. For most workflows, we recommend that you choose Auto to reduce the bandwidth of your output with little to no perceptual decrease in video quality. For high quality and high bitrate outputs, choose Low. For the most bandwidth reduction, choose High. We recommend that you choose High for low bitrate outputs. Note that High may incur a slight increase in the softness of your output.
1882
+ * @public
1883
1883
  */
1884
1884
  Strength?: BandwidthReductionFilterStrength;
1885
1885
  }
@@ -2078,23 +2078,23 @@ export declare const H264QualityTuningLevel: {
2078
2078
  */
2079
2079
  export type H264QualityTuningLevel = (typeof H264QualityTuningLevel)[keyof typeof H264QualityTuningLevel];
2080
2080
  /**
2081
- * @public
2082
2081
  * Settings for quality-defined variable bitrate encoding with the H.264 codec. Use these settings only when you set QVBR for Rate control mode.
2082
+ * @public
2083
2083
  */
2084
2084
  export interface H264QvbrSettings {
2085
2085
  /**
2086
- * @public
2087
2086
  * Use this setting only when Rate control mode is QVBR and Quality tuning level is Multi-pass HQ. For Max average bitrate values suited to the complexity of your input video, the service limits the average bitrate of the video part of this output to the value that you choose. That is, the total size of the video element is less than or equal to the value you set multiplied by the number of seconds of encoded output.
2087
+ * @public
2088
2088
  */
2089
2089
  MaxAverageBitrate?: number;
2090
2090
  /**
2091
- * @public
2092
2091
  * Use this setting only when you set Rate control mode to QVBR. Specify the target quality level for this output. MediaConvert determines the right number of bits to use for each part of the video to maintain the video quality that you specify. When you keep the default value, AUTO, MediaConvert picks a quality level for you, based on characteristics of your input video. If you prefer to specify a quality level, specify a number from 1 through 10. Use higher numbers for greater quality. Level 10 results in nearly lossless compression. The quality level for most broadcast-quality transcodes is between 6 and 9. Optionally, to specify a value between whole numbers, also provide a value for the setting qvbrQualityLevelFineTune. For example, if you want your QVBR quality level to be 7.33, set qvbrQualityLevel to 7 and set qvbrQualityLevelFineTune to .33.
2092
+ * @public
2093
2093
  */
2094
2094
  QvbrQualityLevel?: number;
2095
2095
  /**
2096
- * @public
2097
2096
  * Optional. Specify a value here to set the QVBR quality to a level that is between whole numbers. For example, if you want your QVBR quality level to be 7.33, set qvbrQualityLevel to 7 and set qvbrQualityLevelFineTune to .33. MediaConvert rounds your QVBR quality level to the nearest third of a whole number. For example, if you set qvbrQualityLevel to 7 and you set qvbrQualityLevelFineTune to .25, your actual QVBR quality level is 7.33.
2097
+ * @public
2098
2098
  */
2099
2099
  QvbrQualityLevelFineTune?: number;
2100
2100
  }
@@ -2222,223 +2222,223 @@ export declare const H264UnregisteredSeiTimecode: {
2222
2222
  */
2223
2223
  export type H264UnregisteredSeiTimecode = (typeof H264UnregisteredSeiTimecode)[keyof typeof H264UnregisteredSeiTimecode];
2224
2224
  /**
2225
- * @public
2226
2225
  * Required when you set Codec to the value H_264.
2226
+ * @public
2227
2227
  */
2228
2228
  export interface H264Settings {
2229
2229
  /**
2230
- * @public
2231
2230
  * Keep the default value, Auto, for this setting to have MediaConvert automatically apply the best types of quantization for your video content. When you want to apply your quantization settings manually, you must set H264AdaptiveQuantization to a value other than Auto. Use this setting to specify the strength of any adaptive quantization filters that you enable. If you don't want MediaConvert to do any adaptive quantization in this transcode, set Adaptive quantization to Off. Related settings: The value that you choose here applies to the following settings: H264FlickerAdaptiveQuantization, H264SpatialAdaptiveQuantization, and H264TemporalAdaptiveQuantization.
2231
+ * @public
2232
2232
  */
2233
2233
  AdaptiveQuantization?: H264AdaptiveQuantization;
2234
2234
  /**
2235
- * @public
2236
2235
  * The Bandwidth reduction filter increases the video quality of your output relative to its bitrate. Use to lower the bitrate of your constant quality QVBR output, with little or no perceptual decrease in quality. Or, use to increase the video quality of outputs with other rate control modes relative to the bitrate that you specify. Bandwidth reduction increases further when your input is low quality or noisy. Outputs that use this feature incur pro-tier pricing. When you include Bandwidth reduction filter, you cannot include the Noise reducer preprocessor.
2236
+ * @public
2237
2237
  */
2238
2238
  BandwidthReductionFilter?: BandwidthReductionFilter;
2239
2239
  /**
2240
- * @public
2241
2240
  * Specify the average bitrate in bits per second. Required for VBR and CBR. For MS Smooth outputs, bitrates must be unique when rounded down to the nearest multiple of 1000.
2241
+ * @public
2242
2242
  */
2243
2243
  Bitrate?: number;
2244
2244
  /**
2245
- * @public
2246
2245
  * Specify an H.264 level that is consistent with your output video settings. If you aren't sure what level to specify, choose Auto.
2246
+ * @public
2247
2247
  */
2248
2248
  CodecLevel?: H264CodecLevel;
2249
2249
  /**
2250
- * @public
2251
2250
  * H.264 Profile. High 4:2:2 and 10-bit profiles are only available with the AVC-I License.
2251
+ * @public
2252
2252
  */
2253
2253
  CodecProfile?: H264CodecProfile;
2254
2254
  /**
2255
- * @public
2256
2255
  * Specify whether to allow the number of B-frames in your output GOP structure to vary or not depending on your input video content. To improve the subjective video quality of your output that has high-motion content: Leave blank or keep the default value Adaptive. MediaConvert will use fewer B-frames for high-motion video content than low-motion content. The maximum number of B- frames is limited by the value that you choose for B-frames between reference frames. To use the same number B-frames for all types of content: Choose Static.
2256
+ * @public
2257
2257
  */
2258
2258
  DynamicSubGop?: H264DynamicSubGop;
2259
2259
  /**
2260
- * @public
2261
2260
  * Optionally include or suppress markers at the end of your output that signal the end of the video stream. To include end of stream markers: Leave blank or keep the default value, Include. To not include end of stream markers: Choose Suppress. This is useful when your output will be inserted into another stream.
2261
+ * @public
2262
2262
  */
2263
2263
  EndOfStreamMarkers?: H264EndOfStreamMarkers;
2264
2264
  /**
2265
- * @public
2266
2265
  * Entropy encoding mode. Use CABAC (must be in Main or High profile) or CAVLC.
2266
+ * @public
2267
2267
  */
2268
2268
  EntropyEncoding?: H264EntropyEncoding;
2269
2269
  /**
2270
- * @public
2271
2270
  * The video encoding method for your MPEG-4 AVC output. Keep the default value, PAFF, to have MediaConvert use PAFF encoding for interlaced outputs. Choose Force field to disable PAFF encoding and create separate interlaced fields. Choose MBAFF to disable PAFF and have MediaConvert use MBAFF encoding for interlaced outputs.
2271
+ * @public
2272
2272
  */
2273
2273
  FieldEncoding?: H264FieldEncoding;
2274
2274
  /**
2275
- * @public
2276
2275
  * Only use this setting when you change the default value, AUTO, for the setting H264AdaptiveQuantization. When you keep all defaults, excluding H264AdaptiveQuantization and all other adaptive quantization from your JSON job specification, MediaConvert automatically applies the best types of quantization for your video content. When you set H264AdaptiveQuantization to a value other than AUTO, the default value for H264FlickerAdaptiveQuantization is Disabled. Change this value to Enabled to reduce I-frame pop. I-frame pop appears as a visual flicker that can arise when the encoder saves bits by copying some macroblocks many times from frame to frame, and then refreshes them at the I-frame. When you enable this setting, the encoder updates these macroblocks slightly more often to smooth out the flicker. To manually enable or disable H264FlickerAdaptiveQuantization, you must set Adaptive quantization to a value other than AUTO.
2276
+ * @public
2277
2277
  */
2278
2278
  FlickerAdaptiveQuantization?: H264FlickerAdaptiveQuantization;
2279
2279
  /**
2280
- * @public
2281
2280
  * If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.
2281
+ * @public
2282
2282
  */
2283
2283
  FramerateControl?: H264FramerateControl;
2284
2284
  /**
2285
- * @public
2286
2285
  * Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate. For numerically simple conversions, such as 60 fps to 30 fps: We recommend that you keep the default value, Drop duplicate. For numerically complex conversions, to avoid stutter: Choose Interpolate. This results in a smooth picture, but might introduce undesirable video artifacts. For complex frame rate conversions, especially if your source video has already been converted from its original cadence: Choose FrameFormer to do motion-compensated interpolation. FrameFormer uses the best conversion method frame by frame. Note that using FrameFormer increases the transcoding time and incurs a significant add-on cost. When you choose FrameFormer, your input video resolution must be at least 128x96.
2286
+ * @public
2287
2287
  */
2288
2288
  FramerateConversionAlgorithm?: H264FramerateConversionAlgorithm;
2289
2289
  /**
2290
- * @public
2291
2290
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateDenominator to specify the denominator of this fraction. In this example, use 1001 for the value of FramerateDenominator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
2291
+ * @public
2292
2292
  */
2293
2293
  FramerateDenominator?: number;
2294
2294
  /**
2295
- * @public
2296
2295
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateNumerator to specify the numerator of this fraction. In this example, use 24000 for the value of FramerateNumerator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
2296
+ * @public
2297
2297
  */
2298
2298
  FramerateNumerator?: number;
2299
2299
  /**
2300
- * @public
2301
2300
  * Specify whether to allow B-frames to be referenced by other frame types. To use reference B-frames when your GOP structure has 1 or more B-frames: Leave blank or keep the default value Enabled. We recommend that you choose Enabled to help improve the video quality of your output relative to its bitrate. To not use reference B-frames: Choose Disabled.
2301
+ * @public
2302
2302
  */
2303
2303
  GopBReference?: H264GopBReference;
2304
2304
  /**
2305
- * @public
2306
2305
  * Specify the relative frequency of open to closed GOPs in this output. For example, if you want to allow four open GOPs and then require a closed GOP, set this value to 5. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. In the console, do this by keeping the default empty value. If you do explicitly specify a value, for segmented outputs, don't set this value to 0.
2306
+ * @public
2307
2307
  */
2308
2308
  GopClosedCadence?: number;
2309
2309
  /**
2310
- * @public
2311
2310
  * Use this setting only when you set GOP mode control to Specified, frames or Specified, seconds. Specify the GOP length using a whole number of frames or a decimal value of seconds. MediaConvert will interpret this value as frames or seconds depending on the value you choose for GOP mode control. If you want to allow MediaConvert to automatically determine GOP size, leave GOP size blank and set GOP mode control to Auto. If your output group specifies HLS, DASH, or CMAF, leave GOP size blank and set GOP mode control to Auto in each output in your output group.
2311
+ * @public
2312
2312
  */
2313
2313
  GopSize?: number;
2314
2314
  /**
2315
- * @public
2316
2315
  * Specify how the transcoder determines GOP size for this output. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, choose Auto and and leave GOP size blank. By default, if you don't specify GOP mode control, MediaConvert will use automatic behavior. If your output group specifies HLS, DASH, or CMAF, set GOP mode control to Auto and leave GOP size blank in each output in your output group. To explicitly specify the GOP length, choose Specified, frames or Specified, seconds and then provide the GOP length in the related setting GOP size.
2316
+ * @public
2317
2317
  */
2318
2318
  GopSizeUnits?: H264GopSizeUnits;
2319
2319
  /**
2320
- * @public
2321
2320
  * If your downstream systems have strict buffer requirements: Specify the minimum percentage of the HRD buffer that's available at the end of each encoded video segment. For the best video quality: Set to 0 or leave blank to automatically determine the final buffer fill percentage.
2321
+ * @public
2322
2322
  */
2323
2323
  HrdBufferFinalFillPercentage?: number;
2324
2324
  /**
2325
- * @public
2326
2325
  * Percentage of the buffer that should initially be filled (HRD buffer model).
2326
+ * @public
2327
2327
  */
2328
2328
  HrdBufferInitialFillPercentage?: number;
2329
2329
  /**
2330
- * @public
2331
2330
  * Size of buffer (HRD buffer model) in bits. For example, enter five megabits as 5000000.
2331
+ * @public
2332
2332
  */
2333
2333
  HrdBufferSize?: number;
2334
2334
  /**
2335
- * @public
2336
2335
  * Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.
2336
+ * @public
2337
2337
  */
2338
2338
  InterlaceMode?: H264InterlaceMode;
2339
2339
  /**
2340
- * @public
2341
2340
  * Maximum bitrate in bits/second. For example, enter five megabits per second as 5000000. Required when Rate control mode is QVBR.
2341
+ * @public
2342
2342
  */
2343
2343
  MaxBitrate?: number;
2344
2344
  /**
2345
- * @public
2346
2345
  * Use this setting only when you also enable Scene change detection. This setting determines how the encoder manages the spacing between I-frames that it inserts as part of the I-frame cadence and the I-frames that it inserts for Scene change detection. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, do this by keeping the default empty value. When you explicitly specify a value for this setting, the encoder determines whether to skip a cadence-driven I-frame by the value you set. For example, if you set Min I interval to 5 and a cadence-driven I-frame would fall within 5 frames of a scene-change I-frame, then the encoder skips the cadence-driven I-frame. In this way, one GOP is shrunk slightly and one GOP is stretched slightly. When the cadence-driven I-frames are farther from the scene-change I-frame than the value you set, then the encoder leaves all I-frames in place and the GOPs surrounding the scene change are smaller than the usual cadence GOPs.
2346
+ * @public
2347
2347
  */
2348
2348
  MinIInterval?: number;
2349
2349
  /**
2350
- * @public
2351
2350
  * Specify the number of B-frames between reference frames in this output. For the best video quality: Leave blank. MediaConvert automatically determines the number of B-frames to use based on the characteristics of your input video. To manually specify the number of B-frames between reference frames: Enter an integer from 0 to 7.
2351
+ * @public
2352
2352
  */
2353
2353
  NumberBFramesBetweenReferenceFrames?: number;
2354
2354
  /**
2355
- * @public
2356
2355
  * Number of reference frames to use. The encoder may use more than requested if using B-frames and/or interlaced encoding.
2356
+ * @public
2357
2357
  */
2358
2358
  NumberReferenceFrames?: number;
2359
2359
  /**
2360
- * @public
2361
2360
  * Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source, uses the PAR from your input video for your output. To specify a different PAR in the console, choose any value other than Follow source. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings.
2361
+ * @public
2362
2362
  */
2363
2363
  ParControl?: H264ParControl;
2364
2364
  /**
2365
- * @public
2366
2365
  * Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parDenominator is 33.
2366
+ * @public
2367
2367
  */
2368
2368
  ParDenominator?: number;
2369
2369
  /**
2370
- * @public
2371
2370
  * Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parNumerator is 40.
2371
+ * @public
2372
2372
  */
2373
2373
  ParNumerator?: number;
2374
2374
  /**
2375
- * @public
2376
2375
  * The Quality tuning level you choose represents a trade-off between the encoding speed of your job and the output video quality. For the fastest encoding speed at the cost of video quality: Choose Single pass. For a good balance between encoding speed and video quality: Leave blank or keep the default value Single pass HQ. For the best video quality, at the cost of encoding speed: Choose Multi pass HQ. MediaConvert performs an analysis pass on your input followed by an encoding pass. Outputs that use this feature incur pro-tier pricing.
2376
+ * @public
2377
2377
  */
2378
2378
  QualityTuningLevel?: H264QualityTuningLevel;
2379
2379
  /**
2380
- * @public
2381
2380
  * Settings for quality-defined variable bitrate encoding with the H.265 codec. Use these settings only when you set QVBR for Rate control mode.
2381
+ * @public
2382
2382
  */
2383
2383
  QvbrSettings?: H264QvbrSettings;
2384
2384
  /**
2385
- * @public
2386
2385
  * Use this setting to specify whether this output has a variable bitrate (VBR), constant bitrate (CBR) or quality-defined variable bitrate (QVBR).
2386
+ * @public
2387
2387
  */
2388
2388
  RateControlMode?: H264RateControlMode;
2389
2389
  /**
2390
- * @public
2391
2390
  * Places a PPS header on each encoded picture, even if repeated.
2391
+ * @public
2392
2392
  */
2393
2393
  RepeatPps?: H264RepeatPps;
2394
2394
  /**
2395
- * @public
2396
2395
  * Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive.
2396
+ * @public
2397
2397
  */
2398
2398
  ScanTypeConversionMode?: H264ScanTypeConversionMode;
2399
2399
  /**
2400
- * @public
2401
2400
  * Enable this setting to insert I-frames at scene changes that the service automatically detects. This improves video quality and is enabled by default. If this output uses QVBR, choose Transition detection for further video quality improvement. For more information about QVBR, see https://docs.aws.amazon.com/console/mediaconvert/cbr-vbr-qvbr.
2401
+ * @public
2402
2402
  */
2403
2403
  SceneChangeDetect?: H264SceneChangeDetect;
2404
2404
  /**
2405
- * @public
2406
2405
  * Number of slices per picture. Must be less than or equal to the number of macroblock rows for progressive pictures, and less than or equal to half the number of macroblock rows for interlaced pictures.
2406
+ * @public
2407
2407
  */
2408
2408
  Slices?: number;
2409
2409
  /**
2410
- * @public
2411
2410
  * Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25.
2411
+ * @public
2412
2412
  */
2413
2413
  SlowPal?: H264SlowPal;
2414
2414
  /**
2415
- * @public
2416
2415
  * Ignore this setting unless you need to comply with a specification that requires a specific value. If you don't have a specification requirement, we recommend that you adjust the softness of your output by using a lower value for the setting Sharpness or by enabling a noise reducer filter. The Softness setting specifies the quantization matrices that the encoder uses. Keep the default value, 0, for flat quantization. Choose the value 1 or 16 to use the default JVT softening quantization matricies from the H.264 specification. Choose a value from 17 to 128 to use planar interpolation. Increasing values from 17 to 128 result in increasing reduction of high-frequency data. The value 128 results in the softest video.
2416
+ * @public
2417
2417
  */
2418
2418
  Softness?: number;
2419
2419
  /**
2420
- * @public
2421
2420
  * Only use this setting when you change the default value, Auto, for the setting H264AdaptiveQuantization. When you keep all defaults, excluding H264AdaptiveQuantization and all other adaptive quantization from your JSON job specification, MediaConvert automatically applies the best types of quantization for your video content. When you set H264AdaptiveQuantization to a value other than AUTO, the default value for H264SpatialAdaptiveQuantization is Enabled. Keep this default value to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to set H264SpatialAdaptiveQuantization to Disabled. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher. To manually enable or disable H264SpatialAdaptiveQuantization, you must set Adaptive quantization to a value other than AUTO.
2421
+ * @public
2422
2422
  */
2423
2423
  SpatialAdaptiveQuantization?: H264SpatialAdaptiveQuantization;
2424
2424
  /**
2425
- * @public
2426
2425
  * Produces a bitstream compliant with SMPTE RP-2027.
2426
+ * @public
2427
2427
  */
2428
2428
  Syntax?: H264Syntax;
2429
2429
  /**
2430
- * @public
2431
2430
  * When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard or soft telecine to create a smoother picture. Hard telecine produces a 29.97i output. Soft telecine produces an output with a 23.976 output that signals to the video player device to do the conversion during play back. When you keep the default value, None, MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture.
2431
+ * @public
2432
2432
  */
2433
2433
  Telecine?: H264Telecine;
2434
2434
  /**
2435
- * @public
2436
2435
  * Only use this setting when you change the default value, AUTO, for the setting H264AdaptiveQuantization. When you keep all defaults, excluding H264AdaptiveQuantization and all other adaptive quantization from your JSON job specification, MediaConvert automatically applies the best types of quantization for your video content. When you set H264AdaptiveQuantization to a value other than AUTO, the default value for H264TemporalAdaptiveQuantization is Enabled. Keep this default value to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to set H264TemporalAdaptiveQuantization to Disabled. Related setting: When you enable temporal quantization, adjust the strength of the filter with the setting Adaptive quantization. To manually enable or disable H264TemporalAdaptiveQuantization, you must set Adaptive quantization to a value other than AUTO.
2436
+ * @public
2437
2437
  */
2438
2438
  TemporalAdaptiveQuantization?: H264TemporalAdaptiveQuantization;
2439
2439
  /**
2440
- * @public
2441
2440
  * Inserts timecode for each frame as 4 bytes of an unregistered SEI message.
2441
+ * @public
2442
2442
  */
2443
2443
  UnregisteredSeiTimecode?: H264UnregisteredSeiTimecode;
2444
2444
  }
@@ -2640,23 +2640,23 @@ export declare const H265QualityTuningLevel: {
2640
2640
  */
2641
2641
  export type H265QualityTuningLevel = (typeof H265QualityTuningLevel)[keyof typeof H265QualityTuningLevel];
2642
2642
  /**
2643
- * @public
2644
2643
  * Settings for quality-defined variable bitrate encoding with the H.265 codec. Use these settings only when you set QVBR for Rate control mode.
2644
+ * @public
2645
2645
  */
2646
2646
  export interface H265QvbrSettings {
2647
2647
  /**
2648
- * @public
2649
2648
  * Use this setting only when Rate control mode is QVBR and Quality tuning level is Multi-pass HQ. For Max average bitrate values suited to the complexity of your input video, the service limits the average bitrate of the video part of this output to the value that you choose. That is, the total size of the video element is less than or equal to the value you set multiplied by the number of seconds of encoded output.
2649
+ * @public
2650
2650
  */
2651
2651
  MaxAverageBitrate?: number;
2652
2652
  /**
2653
- * @public
2654
2653
  * Use this setting only when you set Rate control mode to QVBR. Specify the target quality level for this output. MediaConvert determines the right number of bits to use for each part of the video to maintain the video quality that you specify. When you keep the default value, AUTO, MediaConvert picks a quality level for you, based on characteristics of your input video. If you prefer to specify a quality level, specify a number from 1 through 10. Use higher numbers for greater quality. Level 10 results in nearly lossless compression. The quality level for most broadcast-quality transcodes is between 6 and 9. Optionally, to specify a value between whole numbers, also provide a value for the setting qvbrQualityLevelFineTune. For example, if you want your QVBR quality level to be 7.33, set qvbrQualityLevel to 7 and set qvbrQualityLevelFineTune to .33.
2654
+ * @public
2655
2655
  */
2656
2656
  QvbrQualityLevel?: number;
2657
2657
  /**
2658
- * @public
2659
2658
  * Optional. Specify a value here to set the QVBR quality to a level that is between whole numbers. For example, if you want your QVBR quality level to be 7.33, set qvbrQualityLevel to 7 and set qvbrQualityLevelFineTune to .33. MediaConvert rounds your QVBR quality level to the nearest third of a whole number. For example, if you set qvbrQualityLevel to 7 and you set qvbrQualityLevelFineTune to .25, your actual QVBR quality level is 7.33.
2659
+ * @public
2660
2660
  */
2661
2661
  QvbrQualityLevelFineTune?: number;
2662
2662
  }
@@ -2809,223 +2809,223 @@ export declare const H265WriteMp4PackagingType: {
2809
2809
  */
2810
2810
  export type H265WriteMp4PackagingType = (typeof H265WriteMp4PackagingType)[keyof typeof H265WriteMp4PackagingType];
2811
2811
  /**
2812
- * @public
2813
2812
  * Settings for H265 codec
2813
+ * @public
2814
2814
  */
2815
2815
  export interface H265Settings {
2816
2816
  /**
2817
- * @public
2818
2817
  * When you set Adaptive Quantization to Auto, or leave blank, MediaConvert automatically applies quantization to improve the video quality of your output. Set Adaptive Quantization to Low, Medium, High, Higher, or Max to manually control the strength of the quantization filter. When you do, you can specify a value for Spatial Adaptive Quantization, Temporal Adaptive Quantization, and Flicker Adaptive Quantization, to further control the quantization filter. Set Adaptive Quantization to Off to apply no quantization to your output.
2818
+ * @public
2819
2819
  */
2820
2820
  AdaptiveQuantization?: H265AdaptiveQuantization;
2821
2821
  /**
2822
- * @public
2823
2822
  * Enables Alternate Transfer Function SEI message for outputs using Hybrid Log Gamma (HLG) Electro-Optical Transfer Function (EOTF).
2823
+ * @public
2824
2824
  */
2825
2825
  AlternateTransferFunctionSei?: H265AlternateTransferFunctionSei;
2826
2826
  /**
2827
- * @public
2828
2827
  * The Bandwidth reduction filter increases the video quality of your output relative to its bitrate. Use to lower the bitrate of your constant quality QVBR output, with little or no perceptual decrease in quality. Or, use to increase the video quality of outputs with other rate control modes relative to the bitrate that you specify. Bandwidth reduction increases further when your input is low quality or noisy. Outputs that use this feature incur pro-tier pricing. When you include Bandwidth reduction filter, you cannot include the Noise reducer preprocessor.
2828
+ * @public
2829
2829
  */
2830
2830
  BandwidthReductionFilter?: BandwidthReductionFilter;
2831
2831
  /**
2832
- * @public
2833
2832
  * Specify the average bitrate in bits per second. Required for VBR and CBR. For MS Smooth outputs, bitrates must be unique when rounded down to the nearest multiple of 1000.
2833
+ * @public
2834
2834
  */
2835
2835
  Bitrate?: number;
2836
2836
  /**
2837
- * @public
2838
2837
  * H.265 Level.
2838
+ * @public
2839
2839
  */
2840
2840
  CodecLevel?: H265CodecLevel;
2841
2841
  /**
2842
- * @public
2843
2842
  * Represents the Profile and Tier, per the HEVC (H.265) specification. Selections are grouped as [Profile] / [Tier], so "Main/High" represents Main Profile with High Tier. 4:2:2 profiles are only available with the HEVC 4:2:2 License.
2843
+ * @public
2844
2844
  */
2845
2845
  CodecProfile?: H265CodecProfile;
2846
2846
  /**
2847
- * @public
2848
2847
  * Specify whether to allow the number of B-frames in your output GOP structure to vary or not depending on your input video content. To improve the subjective video quality of your output that has high-motion content: Leave blank or keep the default value Adaptive. MediaConvert will use fewer B-frames for high-motion video content than low-motion content. The maximum number of B- frames is limited by the value that you choose for B-frames between reference frames. To use the same number B-frames for all types of content: Choose Static.
2848
+ * @public
2849
2849
  */
2850
2850
  DynamicSubGop?: H265DynamicSubGop;
2851
2851
  /**
2852
- * @public
2853
2852
  * Optionally include or suppress markers at the end of your output that signal the end of the video stream. To include end of stream markers: Leave blank or keep the default value, Include. To not include end of stream markers: Choose Suppress. This is useful when your output will be inserted into another stream.
2853
+ * @public
2854
2854
  */
2855
2855
  EndOfStreamMarkers?: H265EndOfStreamMarkers;
2856
2856
  /**
2857
- * @public
2858
2857
  * Enable this setting to have the encoder reduce I-frame pop. I-frame pop appears as a visual flicker that can arise when the encoder saves bits by copying some macroblocks many times from frame to frame, and then refreshes them at the I-frame. When you enable this setting, the encoder updates these macroblocks slightly more often to smooth out the flicker. This setting is disabled by default. Related setting: In addition to enabling this setting, you must also set adaptiveQuantization to a value other than Off.
2858
+ * @public
2859
2859
  */
2860
2860
  FlickerAdaptiveQuantization?: H265FlickerAdaptiveQuantization;
2861
2861
  /**
2862
- * @public
2863
2862
  * Use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.
2863
+ * @public
2864
2864
  */
2865
2865
  FramerateControl?: H265FramerateControl;
2866
2866
  /**
2867
- * @public
2868
2867
  * Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate. For numerically simple conversions, such as 60 fps to 30 fps: We recommend that you keep the default value, Drop duplicate. For numerically complex conversions, to avoid stutter: Choose Interpolate. This results in a smooth picture, but might introduce undesirable video artifacts. For complex frame rate conversions, especially if your source video has already been converted from its original cadence: Choose FrameFormer to do motion-compensated interpolation. FrameFormer uses the best conversion method frame by frame. Note that using FrameFormer increases the transcoding time and incurs a significant add-on cost. When you choose FrameFormer, your input video resolution must be at least 128x96.
2868
+ * @public
2869
2869
  */
2870
2870
  FramerateConversionAlgorithm?: H265FramerateConversionAlgorithm;
2871
2871
  /**
2872
- * @public
2873
2872
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateDenominator to specify the denominator of this fraction. In this example, use 1001 for the value of FramerateDenominator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
2873
+ * @public
2874
2874
  */
2875
2875
  FramerateDenominator?: number;
2876
2876
  /**
2877
- * @public
2878
2877
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateNumerator to specify the numerator of this fraction. In this example, use 24000 for the value of FramerateNumerator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
2878
+ * @public
2879
2879
  */
2880
2880
  FramerateNumerator?: number;
2881
2881
  /**
2882
- * @public
2883
2882
  * Specify whether to allow B-frames to be referenced by other frame types. To use reference B-frames when your GOP structure has 1 or more B-frames: Leave blank or keep the default value Enabled. We recommend that you choose Enabled to help improve the video quality of your output relative to its bitrate. To not use reference B-frames: Choose Disabled.
2883
+ * @public
2884
2884
  */
2885
2885
  GopBReference?: H265GopBReference;
2886
2886
  /**
2887
- * @public
2888
2887
  * Specify the relative frequency of open to closed GOPs in this output. For example, if you want to allow four open GOPs and then require a closed GOP, set this value to 5. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, do this by keeping the default empty value. If you do explicitly specify a value, for segmented outputs, don't set this value to 0.
2888
+ * @public
2889
2889
  */
2890
2890
  GopClosedCadence?: number;
2891
2891
  /**
2892
- * @public
2893
2892
  * Use this setting only when you set GOP mode control to Specified, frames or Specified, seconds. Specify the GOP length using a whole number of frames or a decimal value of seconds. MediaConvert will interpret this value as frames or seconds depending on the value you choose for GOP mode control. If you want to allow MediaConvert to automatically determine GOP size, leave GOP size blank and set GOP mode control to Auto. If your output group specifies HLS, DASH, or CMAF, leave GOP size blank and set GOP mode control to Auto in each output in your output group.
2893
+ * @public
2894
2894
  */
2895
2895
  GopSize?: number;
2896
2896
  /**
2897
- * @public
2898
2897
  * Specify how the transcoder determines GOP size for this output. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, choose Auto and and leave GOP size blank. By default, if you don't specify GOP mode control, MediaConvert will use automatic behavior. If your output group specifies HLS, DASH, or CMAF, set GOP mode control to Auto and leave GOP size blank in each output in your output group. To explicitly specify the GOP length, choose Specified, frames or Specified, seconds and then provide the GOP length in the related setting GOP size.
2898
+ * @public
2899
2899
  */
2900
2900
  GopSizeUnits?: H265GopSizeUnits;
2901
2901
  /**
2902
- * @public
2903
2902
  * If your downstream systems have strict buffer requirements: Specify the minimum percentage of the HRD buffer that's available at the end of each encoded video segment. For the best video quality: Set to 0 or leave blank to automatically determine the final buffer fill percentage.
2903
+ * @public
2904
2904
  */
2905
2905
  HrdBufferFinalFillPercentage?: number;
2906
2906
  /**
2907
- * @public
2908
2907
  * Percentage of the buffer that should initially be filled (HRD buffer model).
2908
+ * @public
2909
2909
  */
2910
2910
  HrdBufferInitialFillPercentage?: number;
2911
2911
  /**
2912
- * @public
2913
2912
  * Size of buffer (HRD buffer model) in bits. For example, enter five megabits as 5000000.
2913
+ * @public
2914
2914
  */
2915
2915
  HrdBufferSize?: number;
2916
2916
  /**
2917
- * @public
2918
2917
  * Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.
2918
+ * @public
2919
2919
  */
2920
2920
  InterlaceMode?: H265InterlaceMode;
2921
2921
  /**
2922
- * @public
2923
2922
  * Maximum bitrate in bits/second. For example, enter five megabits per second as 5000000. Required when Rate control mode is QVBR.
2923
+ * @public
2924
2924
  */
2925
2925
  MaxBitrate?: number;
2926
2926
  /**
2927
- * @public
2928
2927
  * Use this setting only when you also enable Scene change detection. This setting determines how the encoder manages the spacing between I-frames that it inserts as part of the I-frame cadence and the I-frames that it inserts for Scene change detection. We recommend that you have the transcoder automatically choose this value for you based on characteristics of your input video. To enable this automatic behavior, do this by keeping the default empty value. When you explicitly specify a value for this setting, the encoder determines whether to skip a cadence-driven I-frame by the value you set. For example, if you set Min I interval to 5 and a cadence-driven I-frame would fall within 5 frames of a scene-change I-frame, then the encoder skips the cadence-driven I-frame. In this way, one GOP is shrunk slightly and one GOP is stretched slightly. When the cadence-driven I-frames are farther from the scene-change I-frame than the value you set, then the encoder leaves all I-frames in place and the GOPs surrounding the scene change are smaller than the usual cadence GOPs.
2928
+ * @public
2929
2929
  */
2930
2930
  MinIInterval?: number;
2931
2931
  /**
2932
- * @public
2933
2932
  * Specify the number of B-frames between reference frames in this output. For the best video quality: Leave blank. MediaConvert automatically determines the number of B-frames to use based on the characteristics of your input video. To manually specify the number of B-frames between reference frames: Enter an integer from 0 to 7.
2933
+ * @public
2934
2934
  */
2935
2935
  NumberBFramesBetweenReferenceFrames?: number;
2936
2936
  /**
2937
- * @public
2938
2937
  * Number of reference frames to use. The encoder may use more than requested if using B-frames and/or interlaced encoding.
2938
+ * @public
2939
2939
  */
2940
2940
  NumberReferenceFrames?: number;
2941
2941
  /**
2942
- * @public
2943
2942
  * Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source, uses the PAR from your input video for your output. To specify a different PAR, choose any value other than Follow source. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings.
2943
+ * @public
2944
2944
  */
2945
2945
  ParControl?: H265ParControl;
2946
2946
  /**
2947
- * @public
2948
2947
  * Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parDenominator is 33.
2948
+ * @public
2949
2949
  */
2950
2950
  ParDenominator?: number;
2951
2951
  /**
2952
- * @public
2953
2952
  * Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parNumerator is 40.
2953
+ * @public
2954
2954
  */
2955
2955
  ParNumerator?: number;
2956
2956
  /**
2957
- * @public
2958
2957
  * Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding.
2958
+ * @public
2959
2959
  */
2960
2960
  QualityTuningLevel?: H265QualityTuningLevel;
2961
2961
  /**
2962
- * @public
2963
2962
  * Settings for quality-defined variable bitrate encoding with the H.265 codec. Use these settings only when you set QVBR for Rate control mode.
2963
+ * @public
2964
2964
  */
2965
2965
  QvbrSettings?: H265QvbrSettings;
2966
2966
  /**
2967
- * @public
2968
2967
  * Use this setting to specify whether this output has a variable bitrate (VBR), constant bitrate (CBR) or quality-defined variable bitrate (QVBR).
2968
+ * @public
2969
2969
  */
2970
2970
  RateControlMode?: H265RateControlMode;
2971
2971
  /**
2972
- * @public
2973
2972
  * Specify Sample Adaptive Offset (SAO) filter strength. Adaptive mode dynamically selects best strength based on content
2973
+ * @public
2974
2974
  */
2975
2975
  SampleAdaptiveOffsetFilterMode?: H265SampleAdaptiveOffsetFilterMode;
2976
2976
  /**
2977
- * @public
2978
2977
  * Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive.
2978
+ * @public
2979
2979
  */
2980
2980
  ScanTypeConversionMode?: H265ScanTypeConversionMode;
2981
2981
  /**
2982
- * @public
2983
2982
  * Enable this setting to insert I-frames at scene changes that the service automatically detects. This improves video quality and is enabled by default. If this output uses QVBR, choose Transition detection for further video quality improvement. For more information about QVBR, see https://docs.aws.amazon.com/console/mediaconvert/cbr-vbr-qvbr.
2983
+ * @public
2984
2984
  */
2985
2985
  SceneChangeDetect?: H265SceneChangeDetect;
2986
2986
  /**
2987
- * @public
2988
2987
  * Number of slices per picture. Must be less than or equal to the number of macroblock rows for progressive pictures, and less than or equal to half the number of macroblock rows for interlaced pictures.
2988
+ * @public
2989
2989
  */
2990
2990
  Slices?: number;
2991
2991
  /**
2992
- * @public
2993
2992
  * Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25.
2993
+ * @public
2994
2994
  */
2995
2995
  SlowPal?: H265SlowPal;
2996
2996
  /**
2997
- * @public
2998
2997
  * Keep the default value, Enabled, to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher.
2998
+ * @public
2999
2999
  */
3000
3000
  SpatialAdaptiveQuantization?: H265SpatialAdaptiveQuantization;
3001
3001
  /**
3002
- * @public
3003
3002
  * This field applies only if the Streams > Advanced > Framerate field is set to 29.970. This field works with the Streams > Advanced > Preprocessors > Deinterlacer field and the Streams > Advanced > Interlaced Mode field to identify the scan type for the output: Progressive, Interlaced, Hard Telecine or Soft Telecine. - Hard: produces 29.97i output from 23.976 input. - Soft: produces 23.976; the player converts this output to 29.97i.
3003
+ * @public
3004
3004
  */
3005
3005
  Telecine?: H265Telecine;
3006
3006
  /**
3007
- * @public
3008
3007
  * Keep the default value, Enabled, to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to disable this feature. Related setting: When you enable temporal quantization, adjust the strength of the filter with the setting Adaptive quantization.
3008
+ * @public
3009
3009
  */
3010
3010
  TemporalAdaptiveQuantization?: H265TemporalAdaptiveQuantization;
3011
3011
  /**
3012
- * @public
3013
3012
  * Enables temporal layer identifiers in the encoded bitstream. Up to 3 layers are supported depending on GOP structure: I- and P-frames form one layer, reference B-frames can form a second layer and non-reference b-frames can form a third layer. Decoders can optionally decode only the lower temporal layers to generate a lower frame rate output. For example, given a bitstream with temporal IDs and with b-frames = 1 (i.e. IbPbPb display order), a decoder could decode all the frames for full frame rate output or only the I and P frames (lowest temporal layer) for a half frame rate output.
3013
+ * @public
3014
3014
  */
3015
3015
  TemporalIds?: H265TemporalIds;
3016
3016
  /**
3017
- * @public
3018
3017
  * Enable use of tiles, allowing horizontal as well as vertical subdivision of the encoded pictures.
3018
+ * @public
3019
3019
  */
3020
3020
  Tiles?: H265Tiles;
3021
3021
  /**
3022
- * @public
3023
3022
  * Inserts timecode for each frame as 4 bytes of an unregistered SEI message.
3023
+ * @public
3024
3024
  */
3025
3025
  UnregisteredSeiTimecode?: H265UnregisteredSeiTimecode;
3026
3026
  /**
3027
- * @public
3028
3027
  * If the location of parameter set NAL units doesn't matter in your workflow, ignore this setting. Use this setting only with CMAF or DASH outputs, or with standalone file outputs in an MPEG-4 container (MP4 outputs). Choose HVC1 to mark your output as HVC1. This makes your output compliant with the following specification: ISO IECJTC1 SC29 N13798 Text ISO/IEC FDIS 14496-15 3rd Edition. For these outputs, the service stores parameter set NAL units in the sample headers but not in the samples directly. For MP4 outputs, when you choose HVC1, your output video might not work properly with some downstream systems and video players. The service defaults to marking your output as HEV1. For these outputs, the service writes parameter set NAL units directly into the samples.
3028
+ * @public
3029
3029
  */
3030
3030
  WriteMp4PackagingType?: H265WriteMp4PackagingType;
3031
3031
  }
@@ -3271,173 +3271,173 @@ export declare const Mpeg2TemporalAdaptiveQuantization: {
3271
3271
  */
3272
3272
  export type Mpeg2TemporalAdaptiveQuantization = (typeof Mpeg2TemporalAdaptiveQuantization)[keyof typeof Mpeg2TemporalAdaptiveQuantization];
3273
3273
  /**
3274
- * @public
3275
3274
  * Required when you set Codec to the value MPEG2.
3275
+ * @public
3276
3276
  */
3277
3277
  export interface Mpeg2Settings {
3278
3278
  /**
3279
- * @public
3280
3279
  * Specify the strength of any adaptive quantization filters that you enable. The value that you choose here applies to the following settings: Spatial adaptive quantization, and Temporal adaptive quantization.
3280
+ * @public
3281
3281
  */
3282
3282
  AdaptiveQuantization?: Mpeg2AdaptiveQuantization;
3283
3283
  /**
3284
- * @public
3285
3284
  * Specify the average bitrate in bits per second. Required for VBR and CBR. For MS Smooth outputs, bitrates must be unique when rounded down to the nearest multiple of 1000.
3285
+ * @public
3286
3286
  */
3287
3287
  Bitrate?: number;
3288
3288
  /**
3289
- * @public
3290
3289
  * Use Level to set the MPEG-2 level for the video output.
3290
+ * @public
3291
3291
  */
3292
3292
  CodecLevel?: Mpeg2CodecLevel;
3293
3293
  /**
3294
- * @public
3295
3294
  * Use Profile to set the MPEG-2 profile for the video output.
3295
+ * @public
3296
3296
  */
3297
3297
  CodecProfile?: Mpeg2CodecProfile;
3298
3298
  /**
3299
- * @public
3300
3299
  * Choose Adaptive to improve subjective video quality for high-motion content. This will cause the service to use fewer B-frames (which infer information based on other frames) for high-motion portions of the video and more B-frames for low-motion portions. The maximum number of B-frames is limited by the value you provide for the setting B frames between reference frames.
3300
+ * @public
3301
3301
  */
3302
3302
  DynamicSubGop?: Mpeg2DynamicSubGop;
3303
3303
  /**
3304
- * @public
3305
3304
  * If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.
3305
+ * @public
3306
3306
  */
3307
3307
  FramerateControl?: Mpeg2FramerateControl;
3308
3308
  /**
3309
- * @public
3310
3309
  * Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate. For numerically simple conversions, such as 60 fps to 30 fps: We recommend that you keep the default value, Drop duplicate. For numerically complex conversions, to avoid stutter: Choose Interpolate. This results in a smooth picture, but might introduce undesirable video artifacts. For complex frame rate conversions, especially if your source video has already been converted from its original cadence: Choose FrameFormer to do motion-compensated interpolation. FrameFormer uses the best conversion method frame by frame. Note that using FrameFormer increases the transcoding time and incurs a significant add-on cost. When you choose FrameFormer, your input video resolution must be at least 128x96.
3310
+ * @public
3311
3311
  */
3312
3312
  FramerateConversionAlgorithm?: Mpeg2FramerateConversionAlgorithm;
3313
3313
  /**
3314
- * @public
3315
3314
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateDenominator to specify the denominator of this fraction. In this example, use 1001 for the value of FramerateDenominator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
3315
+ * @public
3316
3316
  */
3317
3317
  FramerateDenominator?: number;
3318
3318
  /**
3319
- * @public
3320
3319
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateNumerator to specify the numerator of this fraction. In this example, use 24000 for the value of FramerateNumerator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
3320
+ * @public
3321
3321
  */
3322
3322
  FramerateNumerator?: number;
3323
3323
  /**
3324
- * @public
3325
3324
  * Specify the relative frequency of open to closed GOPs in this output. For example, if you want to allow four open GOPs and then require a closed GOP, set this value to 5. When you create a streaming output, we recommend that you keep the default value, 1, so that players starting mid-stream receive an IDR frame as quickly as possible. Don't set this value to 0; that would break output segmenting.
3325
+ * @public
3326
3326
  */
3327
3327
  GopClosedCadence?: number;
3328
3328
  /**
3329
- * @public
3330
3329
  * Specify the interval between keyframes, in seconds or frames, for this output. Default: 12 Related settings: When you specify the GOP size in seconds, set GOP mode control to Specified, seconds. The default value for GOP mode control is Frames.
3330
+ * @public
3331
3331
  */
3332
3332
  GopSize?: number;
3333
3333
  /**
3334
- * @public
3335
3334
  * Specify the units for GOP size. If you don't specify a value here, by default the encoder measures GOP size in frames.
3335
+ * @public
3336
3336
  */
3337
3337
  GopSizeUnits?: Mpeg2GopSizeUnits;
3338
3338
  /**
3339
- * @public
3340
3339
  * If your downstream systems have strict buffer requirements: Specify the minimum percentage of the HRD buffer that's available at the end of each encoded video segment. For the best video quality: Set to 0 or leave blank to automatically determine the final buffer fill percentage.
3340
+ * @public
3341
3341
  */
3342
3342
  HrdBufferFinalFillPercentage?: number;
3343
3343
  /**
3344
- * @public
3345
3344
  * Percentage of the buffer that should initially be filled (HRD buffer model).
3345
+ * @public
3346
3346
  */
3347
3347
  HrdBufferInitialFillPercentage?: number;
3348
3348
  /**
3349
- * @public
3350
3349
  * Size of buffer (HRD buffer model) in bits. For example, enter five megabits as 5000000.
3350
+ * @public
3351
3351
  */
3352
3352
  HrdBufferSize?: number;
3353
3353
  /**
3354
- * @public
3355
3354
  * Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.
3355
+ * @public
3356
3356
  */
3357
3357
  InterlaceMode?: Mpeg2InterlaceMode;
3358
3358
  /**
3359
- * @public
3360
3359
  * Use Intra DC precision to set quantization precision for intra-block DC coefficients. If you choose the value auto, the service will automatically select the precision based on the per-frame compression ratio.
3360
+ * @public
3361
3361
  */
3362
3362
  IntraDcPrecision?: Mpeg2IntraDcPrecision;
3363
3363
  /**
3364
- * @public
3365
3364
  * Maximum bitrate in bits/second. For example, enter five megabits per second as 5000000.
3365
+ * @public
3366
3366
  */
3367
3367
  MaxBitrate?: number;
3368
3368
  /**
3369
- * @public
3370
3369
  * Use this setting only when you also enable Scene change detection. This setting determines how the encoder manages the spacing between I-frames that it inserts as part of the I-frame cadence and the I-frames that it inserts for Scene change detection. When you specify a value for this setting, the encoder determines whether to skip a cadence-driven I-frame by the value you set. For example, if you set Min I interval to 5 and a cadence-driven I-frame would fall within 5 frames of a scene-change I-frame, then the encoder skips the cadence-driven I-frame. In this way, one GOP is shrunk slightly and one GOP is stretched slightly. When the cadence-driven I-frames are farther from the scene-change I-frame than the value you set, then the encoder leaves all I-frames in place and the GOPs surrounding the scene change are smaller than the usual cadence GOPs.
3370
+ * @public
3371
3371
  */
3372
3372
  MinIInterval?: number;
3373
3373
  /**
3374
- * @public
3375
3374
  * Specify the number of B-frames that MediaConvert puts between reference frames in this output. Valid values are whole numbers from 0 through 7. When you don't specify a value, MediaConvert defaults to 2.
3375
+ * @public
3376
3376
  */
3377
3377
  NumberBFramesBetweenReferenceFrames?: number;
3378
3378
  /**
3379
- * @public
3380
3379
  * Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source, uses the PAR from your input video for your output. To specify a different PAR in the console, choose any value other than Follow source. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings.
3380
+ * @public
3381
3381
  */
3382
3382
  ParControl?: Mpeg2ParControl;
3383
3383
  /**
3384
- * @public
3385
3384
  * Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parDenominator is 33.
3385
+ * @public
3386
3386
  */
3387
3387
  ParDenominator?: number;
3388
3388
  /**
3389
- * @public
3390
3389
  * Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parNumerator is 40.
3390
+ * @public
3391
3391
  */
3392
3392
  ParNumerator?: number;
3393
3393
  /**
3394
- * @public
3395
3394
  * Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding.
3395
+ * @public
3396
3396
  */
3397
3397
  QualityTuningLevel?: Mpeg2QualityTuningLevel;
3398
3398
  /**
3399
- * @public
3400
3399
  * Use Rate control mode to specify whether the bitrate is variable (vbr) or constant (cbr).
3400
+ * @public
3401
3401
  */
3402
3402
  RateControlMode?: Mpeg2RateControlMode;
3403
3403
  /**
3404
- * @public
3405
3404
  * Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive.
3405
+ * @public
3406
3406
  */
3407
3407
  ScanTypeConversionMode?: Mpeg2ScanTypeConversionMode;
3408
3408
  /**
3409
- * @public
3410
3409
  * Enable this setting to insert I-frames at scene changes that the service automatically detects. This improves video quality and is enabled by default.
3410
+ * @public
3411
3411
  */
3412
3412
  SceneChangeDetect?: Mpeg2SceneChangeDetect;
3413
3413
  /**
3414
- * @public
3415
3414
  * Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25.
3415
+ * @public
3416
3416
  */
3417
3417
  SlowPal?: Mpeg2SlowPal;
3418
3418
  /**
3419
- * @public
3420
3419
  * Ignore this setting unless you need to comply with a specification that requires a specific value. If you don't have a specification requirement, we recommend that you adjust the softness of your output by using a lower value for the setting Sharpness or by enabling a noise reducer filter. The Softness setting specifies the quantization matrices that the encoder uses. Keep the default value, 0, to use the AWS Elemental default matrices. Choose a value from 17 to 128 to use planar interpolation. Increasing values from 17 to 128 result in increasing reduction of high-frequency data. The value 128 results in the softest video.
3420
+ * @public
3421
3421
  */
3422
3422
  Softness?: number;
3423
3423
  /**
3424
- * @public
3425
3424
  * Keep the default value, Enabled, to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher.
3425
+ * @public
3426
3426
  */
3427
3427
  SpatialAdaptiveQuantization?: Mpeg2SpatialAdaptiveQuantization;
3428
3428
  /**
3429
- * @public
3430
3429
  * Specify whether this output's video uses the D10 syntax. Keep the default value to not use the syntax. Related settings: When you choose D10 for your MXF profile, you must also set this value to D10.
3430
+ * @public
3431
3431
  */
3432
3432
  Syntax?: Mpeg2Syntax;
3433
3433
  /**
3434
- * @public
3435
3434
  * When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard or soft telecine to create a smoother picture. Hard telecine produces a 29.97i output. Soft telecine produces an output with a 23.976 output that signals to the video player device to do the conversion during play back. When you keep the default value, None, MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture.
3435
+ * @public
3436
3436
  */
3437
3437
  Telecine?: Mpeg2Telecine;
3438
3438
  /**
3439
- * @public
3440
3439
  * Keep the default value, Enabled, to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to disable this feature. Related setting: When you enable temporal quantization, adjust the strength of the filter with the setting Adaptive quantization.
3440
+ * @public
3441
3441
  */
3442
3442
  TemporalAdaptiveQuantization?: Mpeg2TemporalAdaptiveQuantization;
3443
3443
  }
@@ -3558,73 +3558,73 @@ export declare const ProresTelecine: {
3558
3558
  */
3559
3559
  export type ProresTelecine = (typeof ProresTelecine)[keyof typeof ProresTelecine];
3560
3560
  /**
3561
- * @public
3562
3561
  * Required when you set Codec to the value PRORES.
3562
+ * @public
3563
3563
  */
3564
3564
  export interface ProresSettings {
3565
3565
  /**
3566
- * @public
3567
3566
  * This setting applies only to ProRes 4444 and ProRes 4444 XQ outputs that you create from inputs that use 4:4:4 chroma sampling. Set Preserve 4:4:4 sampling to allow outputs to also use 4:4:4 chroma sampling. You must specify a value for this setting when your output codec profile supports 4:4:4 chroma sampling. Related Settings: For Apple ProRes outputs with 4:4:4 chroma sampling: Choose Preserve 4:4:4 sampling. Use when your input has 4:4:4 chroma sampling and your output codec Profile is Apple ProRes 4444 or 4444 XQ. Note that when you choose Preserve 4:4:4 sampling, you cannot include any of the following Preprocessors: Dolby Vision, HDR10+, or Noise reducer.
3567
+ * @public
3568
3568
  */
3569
3569
  ChromaSampling?: ProresChromaSampling;
3570
3570
  /**
3571
- * @public
3572
3571
  * Use Profile to specify the type of Apple ProRes codec to use for this output.
3572
+ * @public
3573
3573
  */
3574
3574
  CodecProfile?: ProresCodecProfile;
3575
3575
  /**
3576
- * @public
3577
3576
  * If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.
3577
+ * @public
3578
3578
  */
3579
3579
  FramerateControl?: ProresFramerateControl;
3580
3580
  /**
3581
- * @public
3582
3581
  * Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate. For numerically simple conversions, such as 60 fps to 30 fps: We recommend that you keep the default value, Drop duplicate. For numerically complex conversions, to avoid stutter: Choose Interpolate. This results in a smooth picture, but might introduce undesirable video artifacts. For complex frame rate conversions, especially if your source video has already been converted from its original cadence: Choose FrameFormer to do motion-compensated interpolation. FrameFormer uses the best conversion method frame by frame. Note that using FrameFormer increases the transcoding time and incurs a significant add-on cost. When you choose FrameFormer, your input video resolution must be at least 128x96.
3582
+ * @public
3583
3583
  */
3584
3584
  FramerateConversionAlgorithm?: ProresFramerateConversionAlgorithm;
3585
3585
  /**
3586
- * @public
3587
3586
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateDenominator to specify the denominator of this fraction. In this example, use 1001 for the value of FramerateDenominator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
3587
+ * @public
3588
3588
  */
3589
3589
  FramerateDenominator?: number;
3590
3590
  /**
3591
- * @public
3592
3591
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateNumerator to specify the numerator of this fraction. In this example, use 24000 for the value of FramerateNumerator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
3592
+ * @public
3593
3593
  */
3594
3594
  FramerateNumerator?: number;
3595
3595
  /**
3596
- * @public
3597
3596
  * Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.
3597
+ * @public
3598
3598
  */
3599
3599
  InterlaceMode?: ProresInterlaceMode;
3600
3600
  /**
3601
- * @public
3602
3601
  * Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source, uses the PAR from your input video for your output. To specify a different PAR, choose any value other than Follow source. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings.
3602
+ * @public
3603
3603
  */
3604
3604
  ParControl?: ProresParControl;
3605
3605
  /**
3606
- * @public
3607
3606
  * Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parDenominator is 33.
3607
+ * @public
3608
3608
  */
3609
3609
  ParDenominator?: number;
3610
3610
  /**
3611
- * @public
3612
3611
  * Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parNumerator is 40.
3612
+ * @public
3613
3613
  */
3614
3614
  ParNumerator?: number;
3615
3615
  /**
3616
- * @public
3617
3616
  * Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive.
3617
+ * @public
3618
3618
  */
3619
3619
  ScanTypeConversionMode?: ProresScanTypeConversionMode;
3620
3620
  /**
3621
- * @public
3622
3621
  * Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output. When you enable slow PAL, MediaConvert relabels the video frames to 25 fps and resamples your audio to keep it synchronized with the video. Note that enabling this setting will slightly reduce the duration of your video. Required settings: You must also set Framerate to 25.
3622
+ * @public
3623
3623
  */
3624
3624
  SlowPal?: ProresSlowPal;
3625
3625
  /**
3626
- * @public
3627
3626
  * When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard telecine to create a smoother picture. When you keep the default value, None, MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture.
3627
+ * @public
3628
3628
  */
3629
3629
  Telecine?: ProresTelecine;
3630
3630
  }
@@ -3715,53 +3715,53 @@ export declare const UncompressedTelecine: {
3715
3715
  */
3716
3716
  export type UncompressedTelecine = (typeof UncompressedTelecine)[keyof typeof UncompressedTelecine];
3717
3717
  /**
3718
- * @public
3719
3718
  * Required when you set Codec, under VideoDescription>CodecSettings to the value UNCOMPRESSED.
3719
+ * @public
3720
3720
  */
3721
3721
  export interface UncompressedSettings {
3722
3722
  /**
3723
- * @public
3724
3723
  * The four character code for the uncompressed video.
3724
+ * @public
3725
3725
  */
3726
3726
  Fourcc?: UncompressedFourcc;
3727
3727
  /**
3728
- * @public
3729
3728
  * Use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.
3729
+ * @public
3730
3730
  */
3731
3731
  FramerateControl?: UncompressedFramerateControl;
3732
3732
  /**
3733
- * @public
3734
3733
  * Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate. For numerically simple conversions, such as 60 fps to 30 fps: We recommend that you keep the default value, Drop duplicate. For numerically complex conversions, to avoid stutter: Choose Interpolate. This results in a smooth picture, but might introduce undesirable video artifacts. For complex frame rate conversions, especially if your source video has already been converted from its original cadence: Choose FrameFormer to do motion-compensated interpolation. FrameFormer uses the best conversion method frame by frame. Note that using FrameFormer increases the transcoding time and incurs a significant add-on cost. When you choose FrameFormer, your input video resolution must be at least 128x96.
3734
+ * @public
3735
3735
  */
3736
3736
  FramerateConversionAlgorithm?: UncompressedFramerateConversionAlgorithm;
3737
3737
  /**
3738
- * @public
3739
3738
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateDenominator to specify the denominator of this fraction. In this example, use 1001 for the value of FramerateDenominator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
3739
+ * @public
3740
3740
  */
3741
3741
  FramerateDenominator?: number;
3742
3742
  /**
3743
- * @public
3744
3743
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateNumerator to specify the numerator of this fraction. In this example, use 24000 for the value of FramerateNumerator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
3744
+ * @public
3745
3745
  */
3746
3746
  FramerateNumerator?: number;
3747
3747
  /**
3748
- * @public
3749
3748
  * Optional. Choose the scan line type for this output. If you don't specify a value, MediaConvert will create a progressive output.
3749
+ * @public
3750
3750
  */
3751
3751
  InterlaceMode?: UncompressedInterlaceMode;
3752
3752
  /**
3753
- * @public
3754
3753
  * Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive.
3754
+ * @public
3755
3755
  */
3756
3756
  ScanTypeConversionMode?: UncompressedScanTypeConversionMode;
3757
3757
  /**
3758
- * @public
3759
3758
  * Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output by relabeling the video frames and resampling your audio. Note that enabling this setting will slightly reduce the duration of your video. Related settings: You must also set Framerate to 25.
3759
+ * @public
3760
3760
  */
3761
3761
  SlowPal?: UncompressedSlowPal;
3762
3762
  /**
3763
- * @public
3764
3763
  * When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard telecine to create a smoother picture. When you keep the default value, None, MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture.
3764
+ * @public
3765
3765
  */
3766
3766
  Telecine?: UncompressedTelecine;
3767
3767
  }
@@ -3852,53 +3852,53 @@ export declare const Vc3Class: {
3852
3852
  */
3853
3853
  export type Vc3Class = (typeof Vc3Class)[keyof typeof Vc3Class];
3854
3854
  /**
3855
- * @public
3856
3855
  * Required when you set Codec to the value VC3
3856
+ * @public
3857
3857
  */
3858
3858
  export interface Vc3Settings {
3859
3859
  /**
3860
- * @public
3861
3860
  * If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.
3861
+ * @public
3862
3862
  */
3863
3863
  FramerateControl?: Vc3FramerateControl;
3864
3864
  /**
3865
- * @public
3866
3865
  * Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate. For numerically simple conversions, such as 60 fps to 30 fps: We recommend that you keep the default value, Drop duplicate. For numerically complex conversions, to avoid stutter: Choose Interpolate. This results in a smooth picture, but might introduce undesirable video artifacts. For complex frame rate conversions, especially if your source video has already been converted from its original cadence: Choose FrameFormer to do motion-compensated interpolation. FrameFormer uses the best conversion method frame by frame. Note that using FrameFormer increases the transcoding time and incurs a significant add-on cost. When you choose FrameFormer, your input video resolution must be at least 128x96.
3866
+ * @public
3867
3867
  */
3868
3868
  FramerateConversionAlgorithm?: Vc3FramerateConversionAlgorithm;
3869
3869
  /**
3870
- * @public
3871
3870
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateDenominator to specify the denominator of this fraction. In this example, use 1001 for the value of FramerateDenominator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
3871
+ * @public
3872
3872
  */
3873
3873
  FramerateDenominator?: number;
3874
3874
  /**
3875
- * @public
3876
3875
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateNumerator to specify the numerator of this fraction. In this example, use 24000 for the value of FramerateNumerator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
3876
+ * @public
3877
3877
  */
3878
3878
  FramerateNumerator?: number;
3879
3879
  /**
3880
- * @public
3881
3880
  * Optional. Choose the scan line type for this output. If you don't specify a value, MediaConvert will create a progressive output.
3881
+ * @public
3882
3882
  */
3883
3883
  InterlaceMode?: Vc3InterlaceMode;
3884
3884
  /**
3885
- * @public
3886
3885
  * Use this setting for interlaced outputs, when your output frame rate is half of your input frame rate. In this situation, choose Optimized interlacing to create a better quality interlaced output. In this case, each progressive frame from the input corresponds to an interlaced field in the output. Keep the default value, Basic interlacing, for all other output frame rates. With basic interlacing, MediaConvert performs any frame rate conversion first and then interlaces the frames. When you choose Optimized interlacing and you set your output frame rate to a value that isn't suitable for optimized interlacing, MediaConvert automatically falls back to basic interlacing. Required settings: To use optimized interlacing, you must set Telecine to None or Soft. You can't use optimized interlacing for hard telecine outputs. You must also set Interlace mode to a value other than Progressive.
3886
+ * @public
3887
3887
  */
3888
3888
  ScanTypeConversionMode?: Vc3ScanTypeConversionMode;
3889
3889
  /**
3890
- * @public
3891
3890
  * Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output by relabeling the video frames and resampling your audio. Note that enabling this setting will slightly reduce the duration of your video. Related settings: You must also set Framerate to 25.
3891
+ * @public
3892
3892
  */
3893
3893
  SlowPal?: Vc3SlowPal;
3894
3894
  /**
3895
- * @public
3896
3895
  * When you do frame rate conversion from 23.976 frames per second (fps) to 29.97 fps, and your output scan type is interlaced, you can optionally enable hard telecine to create a smoother picture. When you keep the default value, None, MediaConvert does a standard frame rate conversion to 29.97 without doing anything with the field polarity to create a smoother picture.
3896
+ * @public
3897
3897
  */
3898
3898
  Telecine?: Vc3Telecine;
3899
3899
  /**
3900
- * @public
3901
3900
  * Specify the VC3 class to choose the quality characteristics for this output. VC3 class, together with the settings Framerate (framerateNumerator and framerateDenominator) and Resolution (height and width), determine your output bitrate. For example, say that your video resolution is 1920x1080 and your framerate is 29.97. Then Class 145 gives you an output with a bitrate of approximately 145 Mbps and Class 220 gives you and output with a bitrate of approximately 220 Mbps. VC3 class also specifies the color bit depth of your output.
3901
+ * @public
3902
3902
  */
3903
3903
  Vc3Class?: Vc3Class;
3904
3904
  }
@@ -3963,73 +3963,73 @@ export declare const Vp8RateControlMode: {
3963
3963
  */
3964
3964
  export type Vp8RateControlMode = (typeof Vp8RateControlMode)[keyof typeof Vp8RateControlMode];
3965
3965
  /**
3966
- * @public
3967
3966
  * Required when you set Codec to the value VP8.
3967
+ * @public
3968
3968
  */
3969
3969
  export interface Vp8Settings {
3970
3970
  /**
3971
- * @public
3972
3971
  * Target bitrate in bits/second. For example, enter five megabits per second as 5000000.
3972
+ * @public
3973
3973
  */
3974
3974
  Bitrate?: number;
3975
3975
  /**
3976
- * @public
3977
3976
  * If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.
3977
+ * @public
3978
3978
  */
3979
3979
  FramerateControl?: Vp8FramerateControl;
3980
3980
  /**
3981
- * @public
3982
3981
  * Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate. For numerically simple conversions, such as 60 fps to 30 fps: We recommend that you keep the default value, Drop duplicate. For numerically complex conversions, to avoid stutter: Choose Interpolate. This results in a smooth picture, but might introduce undesirable video artifacts. For complex frame rate conversions, especially if your source video has already been converted from its original cadence: Choose FrameFormer to do motion-compensated interpolation. FrameFormer uses the best conversion method frame by frame. Note that using FrameFormer increases the transcoding time and incurs a significant add-on cost. When you choose FrameFormer, your input video resolution must be at least 128x96.
3982
+ * @public
3983
3983
  */
3984
3984
  FramerateConversionAlgorithm?: Vp8FramerateConversionAlgorithm;
3985
3985
  /**
3986
- * @public
3987
3986
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateDenominator to specify the denominator of this fraction. In this example, use 1001 for the value of FramerateDenominator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
3987
+ * @public
3988
3988
  */
3989
3989
  FramerateDenominator?: number;
3990
3990
  /**
3991
- * @public
3992
3991
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateNumerator to specify the numerator of this fraction. In this example, use 24000 for the value of FramerateNumerator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
3992
+ * @public
3993
3993
  */
3994
3994
  FramerateNumerator?: number;
3995
3995
  /**
3996
- * @public
3997
3996
  * GOP Length (keyframe interval) in frames. Must be greater than zero.
3997
+ * @public
3998
3998
  */
3999
3999
  GopSize?: number;
4000
4000
  /**
4001
- * @public
4002
4001
  * Optional. Size of buffer (HRD buffer model) in bits. For example, enter five megabits as 5000000.
4002
+ * @public
4003
4003
  */
4004
4004
  HrdBufferSize?: number;
4005
4005
  /**
4006
- * @public
4007
4006
  * Ignore this setting unless you set qualityTuningLevel to MULTI_PASS. Optional. Specify the maximum bitrate in bits/second. For example, enter five megabits per second as 5000000. The default behavior uses twice the target bitrate as the maximum bitrate.
4007
+ * @public
4008
4008
  */
4009
4009
  MaxBitrate?: number;
4010
4010
  /**
4011
- * @public
4012
4011
  * Optional. Specify how the service determines the pixel aspect ratio (PAR) for this output. The default behavior, Follow source, uses the PAR from your input video for your output. To specify a different PAR in the console, choose any value other than Follow source. When you choose SPECIFIED for this setting, you must also specify values for the parNumerator and parDenominator settings.
4012
+ * @public
4013
4013
  */
4014
4014
  ParControl?: Vp8ParControl;
4015
4015
  /**
4016
- * @public
4017
4016
  * Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parDenominator is 33.
4017
+ * @public
4018
4018
  */
4019
4019
  ParDenominator?: number;
4020
4020
  /**
4021
- * @public
4022
4021
  * Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parNumerator is 40.
4022
+ * @public
4023
4023
  */
4024
4024
  ParNumerator?: number;
4025
4025
  /**
4026
- * @public
4027
4026
  * Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, multi-pass encoding.
4027
+ * @public
4028
4028
  */
4029
4029
  QualityTuningLevel?: Vp8QualityTuningLevel;
4030
4030
  /**
4031
- * @public
4032
4031
  * With the VP8 codec, you can use only the variable bitrate (VBR) rate control mode.
4032
+ * @public
4033
4033
  */
4034
4034
  RateControlMode?: Vp8RateControlMode;
4035
4035
  }
@@ -4094,73 +4094,73 @@ export declare const Vp9RateControlMode: {
4094
4094
  */
4095
4095
  export type Vp9RateControlMode = (typeof Vp9RateControlMode)[keyof typeof Vp9RateControlMode];
4096
4096
  /**
4097
- * @public
4098
4097
  * Required when you set Codec to the value VP9.
4098
+ * @public
4099
4099
  */
4100
4100
  export interface Vp9Settings {
4101
4101
  /**
4102
- * @public
4103
4102
  * Target bitrate in bits/second. For example, enter five megabits per second as 5000000.
4103
+ * @public
4104
4104
  */
4105
4105
  Bitrate?: number;
4106
4106
  /**
4107
- * @public
4108
4107
  * If you are using the console, use the Framerate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list or choose Custom. The framerates shown in the dropdown list are decimal approximations of fractions. If you choose Custom, specify your frame rate as a fraction.
4108
+ * @public
4109
4109
  */
4110
4110
  FramerateControl?: Vp9FramerateControl;
4111
4111
  /**
4112
- * @public
4113
4112
  * Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate. For numerically simple conversions, such as 60 fps to 30 fps: We recommend that you keep the default value, Drop duplicate. For numerically complex conversions, to avoid stutter: Choose Interpolate. This results in a smooth picture, but might introduce undesirable video artifacts. For complex frame rate conversions, especially if your source video has already been converted from its original cadence: Choose FrameFormer to do motion-compensated interpolation. FrameFormer uses the best conversion method frame by frame. Note that using FrameFormer increases the transcoding time and incurs a significant add-on cost. When you choose FrameFormer, your input video resolution must be at least 128x96.
4113
+ * @public
4114
4114
  */
4115
4115
  FramerateConversionAlgorithm?: Vp9FramerateConversionAlgorithm;
4116
4116
  /**
4117
- * @public
4118
4117
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateDenominator to specify the denominator of this fraction. In this example, use 1001 for the value of FramerateDenominator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
4118
+ * @public
4119
4119
  */
4120
4120
  FramerateDenominator?: number;
4121
4121
  /**
4122
- * @public
4123
4122
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateNumerator to specify the numerator of this fraction. In this example, use 24000 for the value of FramerateNumerator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
4123
+ * @public
4124
4124
  */
4125
4125
  FramerateNumerator?: number;
4126
4126
  /**
4127
- * @public
4128
4127
  * GOP Length (keyframe interval) in frames. Must be greater than zero.
4128
+ * @public
4129
4129
  */
4130
4130
  GopSize?: number;
4131
4131
  /**
4132
- * @public
4133
4132
  * Size of buffer (HRD buffer model) in bits. For example, enter five megabits as 5000000.
4133
+ * @public
4134
4134
  */
4135
4135
  HrdBufferSize?: number;
4136
4136
  /**
4137
- * @public
4138
4137
  * Ignore this setting unless you set qualityTuningLevel to MULTI_PASS. Optional. Specify the maximum bitrate in bits/second. For example, enter five megabits per second as 5000000. The default behavior uses twice the target bitrate as the maximum bitrate.
4138
+ * @public
4139
4139
  */
4140
4140
  MaxBitrate?: number;
4141
4141
  /**
4142
- * @public
4143
4142
  * Optional. Specify how the service determines the pixel aspect ratio for this output. The default behavior is to use the same pixel aspect ratio as your input video.
4143
+ * @public
4144
4144
  */
4145
4145
  ParControl?: Vp9ParControl;
4146
4146
  /**
4147
- * @public
4148
4147
  * Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parDenominator is 33.
4148
+ * @public
4149
4149
  */
4150
4150
  ParDenominator?: number;
4151
4151
  /**
4152
- * @public
4153
4152
  * Required when you set Pixel aspect ratio to SPECIFIED. On the console, this corresponds to any value other than Follow source. When you specify an output pixel aspect ratio (PAR) that is different from your input video PAR, provide your output PAR as a ratio. For example, for D1/DV NTSC widescreen, you would specify the ratio 40:33. In this example, the value for parNumerator is 40.
4153
+ * @public
4154
4154
  */
4155
4155
  ParNumerator?: number;
4156
4156
  /**
4157
- * @public
4158
4157
  * Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, multi-pass encoding.
4158
+ * @public
4159
4159
  */
4160
4160
  QualityTuningLevel?: Vp9QualityTuningLevel;
4161
4161
  /**
4162
- * @public
4163
4162
  * With the VP9 codec, you can use only the variable bitrate (VBR) rate control mode.
4163
+ * @public
4164
4164
  */
4165
4165
  RateControlMode?: Vp9RateControlMode;
4166
4166
  }
@@ -4284,13 +4284,13 @@ export declare const Xavc4kIntraCbgProfileClass: {
4284
4284
  */
4285
4285
  export type Xavc4kIntraCbgProfileClass = (typeof Xavc4kIntraCbgProfileClass)[keyof typeof Xavc4kIntraCbgProfileClass];
4286
4286
  /**
4287
- * @public
4288
4287
  * Required when you set Profile to the value XAVC_4K_INTRA_CBG.
4288
+ * @public
4289
4289
  */
4290
4290
  export interface Xavc4kIntraCbgProfileSettings {
4291
4291
  /**
4292
- * @public
4293
4292
  * Specify the XAVC Intra 4k (CBG) Class to set the bitrate of your output. Outputs of the same class have similar image quality over the operating points that are valid for that class.
4293
+ * @public
4294
4294
  */
4295
4295
  XavcClass?: Xavc4kIntraCbgProfileClass;
4296
4296
  }
@@ -4308,13 +4308,13 @@ export declare const Xavc4kIntraVbrProfileClass: {
4308
4308
  */
4309
4309
  export type Xavc4kIntraVbrProfileClass = (typeof Xavc4kIntraVbrProfileClass)[keyof typeof Xavc4kIntraVbrProfileClass];
4310
4310
  /**
4311
- * @public
4312
4311
  * Required when you set Profile to the value XAVC_4K_INTRA_VBR.
4312
+ * @public
4313
4313
  */
4314
4314
  export interface Xavc4kIntraVbrProfileSettings {
4315
4315
  /**
4316
- * @public
4317
4316
  * Specify the XAVC Intra 4k (VBR) Class to set the bitrate of your output. Outputs of the same class have similar image quality over the operating points that are valid for that class.
4317
+ * @public
4318
4318
  */
4319
4319
  XavcClass?: Xavc4kIntraVbrProfileClass;
4320
4320
  }
@@ -4381,48 +4381,48 @@ export declare const Xavc4kProfileQualityTuningLevel: {
4381
4381
  */
4382
4382
  export type Xavc4kProfileQualityTuningLevel = (typeof Xavc4kProfileQualityTuningLevel)[keyof typeof Xavc4kProfileQualityTuningLevel];
4383
4383
  /**
4384
- * @public
4385
4384
  * Required when you set Profile to the value XAVC_4K.
4385
+ * @public
4386
4386
  */
4387
4387
  export interface Xavc4kProfileSettings {
4388
4388
  /**
4389
- * @public
4390
4389
  * Specify the XAVC 4k (Long GOP) Bitrate Class to set the bitrate of your output. Outputs of the same class have similar image quality over the operating points that are valid for that class.
4390
+ * @public
4391
4391
  */
4392
4392
  BitrateClass?: Xavc4kProfileBitrateClass;
4393
4393
  /**
4394
- * @public
4395
4394
  * Specify the codec profile for this output. Choose High, 8-bit, 4:2:0 (HIGH) or High, 10-bit, 4:2:2 (HIGH_422). These profiles are specified in ITU-T H.264.
4395
+ * @public
4396
4396
  */
4397
4397
  CodecProfile?: Xavc4kProfileCodecProfile;
4398
4398
  /**
4399
- * @public
4400
4399
  * The best way to set up adaptive quantization is to keep the default value, Auto, for the setting Adaptive quantization. When you do so, MediaConvert automatically applies the best types of quantization for your video content. Include this setting in your JSON job specification only when you choose to change the default value for Adaptive quantization. Enable this setting to have the encoder reduce I-frame pop. I-frame pop appears as a visual flicker that can arise when the encoder saves bits by copying some macroblocks many times from frame to frame, and then refreshes them at the I-frame. When you enable this setting, the encoder updates these macroblocks slightly more often to smooth out the flicker. This setting is disabled by default. Related setting: In addition to enabling this setting, you must also set Adaptive quantization to a value other than Off or Auto. Use Adaptive quantization to adjust the degree of smoothing that Flicker adaptive quantization provides.
4400
+ * @public
4401
4401
  */
4402
4402
  FlickerAdaptiveQuantization?: XavcFlickerAdaptiveQuantization;
4403
4403
  /**
4404
- * @public
4405
4404
  * Specify whether the encoder uses B-frames as reference frames for other pictures in the same GOP. Choose Allow to allow the encoder to use B-frames as reference frames. Choose Don't allow to prevent the encoder from using B-frames as reference frames.
4405
+ * @public
4406
4406
  */
4407
4407
  GopBReference?: XavcGopBReference;
4408
4408
  /**
4409
- * @public
4410
4409
  * Frequency of closed GOPs. In streaming applications, it is recommended that this be set to 1 so a decoder joining mid-stream will receive an IDR frame as quickly as possible. Setting this value to 0 will break output segmenting.
4410
+ * @public
4411
4411
  */
4412
4412
  GopClosedCadence?: number;
4413
4413
  /**
4414
- * @public
4415
4414
  * Specify the size of the buffer that MediaConvert uses in the HRD buffer model for this output. Specify this value in bits; for example, enter five megabits as 5000000. When you don't set this value, or you set it to zero, MediaConvert calculates the default by doubling the bitrate of this output point.
4415
+ * @public
4416
4416
  */
4417
4417
  HrdBufferSize?: number;
4418
4418
  /**
4419
- * @public
4420
4419
  * Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding.
4420
+ * @public
4421
4421
  */
4422
4422
  QualityTuningLevel?: Xavc4kProfileQualityTuningLevel;
4423
4423
  /**
4424
- * @public
4425
4424
  * Number of slices per picture. Must be less than or equal to the number of macroblock rows for progressive pictures, and less than or equal to half the number of macroblock rows for interlaced pictures.
4425
+ * @public
4426
4426
  */
4427
4427
  Slices?: number;
4428
4428
  }
@@ -4440,13 +4440,13 @@ export declare const XavcHdIntraCbgProfileClass: {
4440
4440
  */
4441
4441
  export type XavcHdIntraCbgProfileClass = (typeof XavcHdIntraCbgProfileClass)[keyof typeof XavcHdIntraCbgProfileClass];
4442
4442
  /**
4443
- * @public
4444
4443
  * Required when you set Profile to the value XAVC_HD_INTRA_CBG.
4444
+ * @public
4445
4445
  */
4446
4446
  export interface XavcHdIntraCbgProfileSettings {
4447
4447
  /**
4448
- * @public
4449
4448
  * Specify the XAVC Intra HD (CBG) Class to set the bitrate of your output. Outputs of the same class have similar image quality over the operating points that are valid for that class.
4449
+ * @public
4450
4450
  */
4451
4451
  XavcClass?: XavcHdIntraCbgProfileClass;
4452
4452
  }
@@ -4504,210 +4504,210 @@ export declare const XavcHdProfileTelecine: {
4504
4504
  */
4505
4505
  export type XavcHdProfileTelecine = (typeof XavcHdProfileTelecine)[keyof typeof XavcHdProfileTelecine];
4506
4506
  /**
4507
- * @public
4508
4507
  * Required when you set Profile to the value XAVC_HD.
4508
+ * @public
4509
4509
  */
4510
4510
  export interface XavcHdProfileSettings {
4511
4511
  /**
4512
- * @public
4513
4512
  * Specify the XAVC HD (Long GOP) Bitrate Class to set the bitrate of your output. Outputs of the same class have similar image quality over the operating points that are valid for that class.
4513
+ * @public
4514
4514
  */
4515
4515
  BitrateClass?: XavcHdProfileBitrateClass;
4516
4516
  /**
4517
- * @public
4518
4517
  * The best way to set up adaptive quantization is to keep the default value, Auto, for the setting Adaptive quantization. When you do so, MediaConvert automatically applies the best types of quantization for your video content. Include this setting in your JSON job specification only when you choose to change the default value for Adaptive quantization. Enable this setting to have the encoder reduce I-frame pop. I-frame pop appears as a visual flicker that can arise when the encoder saves bits by copying some macroblocks many times from frame to frame, and then refreshes them at the I-frame. When you enable this setting, the encoder updates these macroblocks slightly more often to smooth out the flicker. This setting is disabled by default. Related setting: In addition to enabling this setting, you must also set Adaptive quantization to a value other than Off or Auto. Use Adaptive quantization to adjust the degree of smoothing that Flicker adaptive quantization provides.
4518
+ * @public
4519
4519
  */
4520
4520
  FlickerAdaptiveQuantization?: XavcFlickerAdaptiveQuantization;
4521
4521
  /**
4522
- * @public
4523
4522
  * Specify whether the encoder uses B-frames as reference frames for other pictures in the same GOP. Choose Allow to allow the encoder to use B-frames as reference frames. Choose Don't allow to prevent the encoder from using B-frames as reference frames.
4523
+ * @public
4524
4524
  */
4525
4525
  GopBReference?: XavcGopBReference;
4526
4526
  /**
4527
- * @public
4528
4527
  * Frequency of closed GOPs. In streaming applications, it is recommended that this be set to 1 so a decoder joining mid-stream will receive an IDR frame as quickly as possible. Setting this value to 0 will break output segmenting.
4528
+ * @public
4529
4529
  */
4530
4530
  GopClosedCadence?: number;
4531
4531
  /**
4532
- * @public
4533
4532
  * Specify the size of the buffer that MediaConvert uses in the HRD buffer model for this output. Specify this value in bits; for example, enter five megabits as 5000000. When you don't set this value, or you set it to zero, MediaConvert calculates the default by doubling the bitrate of this output point.
4533
+ * @public
4534
4534
  */
4535
4535
  HrdBufferSize?: number;
4536
4536
  /**
4537
- * @public
4538
4537
  * Choose the scan line type for the output. Keep the default value, Progressive to create a progressive output, regardless of the scan type of your input. Use Top field first or Bottom field first to create an output that's interlaced with the same field polarity throughout. Use Follow, default top or Follow, default bottom to produce outputs with the same field polarity as the source. For jobs that have multiple inputs, the output field polarity might change over the course of the output. Follow behavior depends on the input scan type. If the source is interlaced, the output will be interlaced with the same polarity as the source. If the source is progressive, the output will be interlaced with top field bottom field first, depending on which of the Follow options you choose.
4538
+ * @public
4539
4539
  */
4540
4540
  InterlaceMode?: XavcInterlaceMode;
4541
4541
  /**
4542
- * @public
4543
4542
  * Optional. Use Quality tuning level to choose how you want to trade off encoding speed for output video quality. The default behavior is faster, lower quality, single-pass encoding.
4543
+ * @public
4544
4544
  */
4545
4545
  QualityTuningLevel?: XavcHdProfileQualityTuningLevel;
4546
4546
  /**
4547
- * @public
4548
4547
  * Number of slices per picture. Must be less than or equal to the number of macroblock rows for progressive pictures, and less than or equal to half the number of macroblock rows for interlaced pictures.
4548
+ * @public
4549
4549
  */
4550
4550
  Slices?: number;
4551
4551
  /**
4552
- * @public
4553
4552
  * Ignore this setting unless you set Frame rate (framerateNumerator divided by framerateDenominator) to 29.970. If your input framerate is 23.976, choose Hard. Otherwise, keep the default value None. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/working-with-telecine-and-inverse-telecine.html.
4553
+ * @public
4554
4554
  */
4555
4555
  Telecine?: XavcHdProfileTelecine;
4556
4556
  }
4557
4557
  /**
4558
- * @public
4559
4558
  * Required when you set Codec to the value XAVC.
4559
+ * @public
4560
4560
  */
4561
4561
  export interface XavcSettings {
4562
4562
  /**
4563
- * @public
4564
4563
  * Keep the default value, Auto, for this setting to have MediaConvert automatically apply the best types of quantization for your video content. When you want to apply your quantization settings manually, you must set Adaptive quantization to a value other than Auto. Use this setting to specify the strength of any adaptive quantization filters that you enable. If you don't want MediaConvert to do any adaptive quantization in this transcode, set Adaptive quantization to Off. Related settings: The value that you choose here applies to the following settings: Flicker adaptive quantization (flickerAdaptiveQuantization), Spatial adaptive quantization, and Temporal adaptive quantization.
4564
+ * @public
4565
4565
  */
4566
4566
  AdaptiveQuantization?: XavcAdaptiveQuantization;
4567
4567
  /**
4568
- * @public
4569
4568
  * Optional. Choose a specific entropy encoding mode only when you want to override XAVC recommendations. If you choose the value auto, MediaConvert uses the mode that the XAVC file format specifies given this output's operating point.
4569
+ * @public
4570
4570
  */
4571
4571
  EntropyEncoding?: XavcEntropyEncoding;
4572
4572
  /**
4573
- * @public
4574
4573
  * If you are using the console, use the Frame rate setting to specify the frame rate for this output. If you want to keep the same frame rate as the input video, choose Follow source. If you want to do frame rate conversion, choose a frame rate from the dropdown list. The framerates shown in the dropdown list are decimal approximations of fractions.
4574
+ * @public
4575
4575
  */
4576
4576
  FramerateControl?: XavcFramerateControl;
4577
4577
  /**
4578
- * @public
4579
4578
  * Choose the method that you want MediaConvert to use when increasing or decreasing the frame rate. For numerically simple conversions, such as 60 fps to 30 fps: We recommend that you keep the default value, Drop duplicate. For numerically complex conversions, to avoid stutter: Choose Interpolate. This results in a smooth picture, but might introduce undesirable video artifacts. For complex frame rate conversions, especially if your source video has already been converted from its original cadence: Choose FrameFormer to do motion-compensated interpolation. FrameFormer uses the best conversion method frame by frame. Note that using FrameFormer increases the transcoding time and incurs a significant add-on cost. When you choose FrameFormer, your input video resolution must be at least 128x96.
4579
+ * @public
4580
4580
  */
4581
4581
  FramerateConversionAlgorithm?: XavcFramerateConversionAlgorithm;
4582
4582
  /**
4583
- * @public
4584
4583
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateDenominator to specify the denominator of this fraction. In this example, use 1001 for the value of FramerateDenominator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Frame rate. In this example, specify 23.976.
4584
+ * @public
4585
4585
  */
4586
4586
  FramerateDenominator?: number;
4587
4587
  /**
4588
- * @public
4589
4588
  * When you use the API for transcode jobs that use frame rate conversion, specify the frame rate as a fraction. For example, 24000 / 1001 = 23.976 fps. Use FramerateNumerator to specify the numerator of this fraction. In this example, use 24000 for the value of FramerateNumerator. When you use the console for transcode jobs that use frame rate conversion, provide the value as a decimal number for Framerate. In this example, specify 23.976.
4589
+ * @public
4590
4590
  */
4591
4591
  FramerateNumerator?: number;
4592
4592
  /**
4593
- * @public
4594
4593
  * Specify the XAVC profile for this output. For more information, see the Sony documentation at https://www.xavc-info.org/. Note that MediaConvert doesn't support the interlaced video XAVC operating points for XAVC_HD_INTRA_CBG. To create an interlaced XAVC output, choose the profile XAVC_HD.
4594
+ * @public
4595
4595
  */
4596
4596
  Profile?: XavcProfile;
4597
4597
  /**
4598
- * @public
4599
4598
  * Ignore this setting unless your input frame rate is 23.976 or 24 frames per second (fps). Enable slow PAL to create a 25 fps output by relabeling the video frames and resampling your audio. Note that enabling this setting will slightly reduce the duration of your video. Related settings: You must also set Frame rate to 25.
4599
+ * @public
4600
4600
  */
4601
4601
  SlowPal?: XavcSlowPal;
4602
4602
  /**
4603
- * @public
4604
4603
  * Ignore this setting unless your downstream workflow requires that you specify it explicitly. Otherwise, we recommend that you adjust the softness of your output by using a lower value for the setting Sharpness or by enabling a noise reducer filter. The Softness setting specifies the quantization matrices that the encoder uses. Keep the default value, 0, for flat quantization. Choose the value 1 or 16 to use the default JVT softening quantization matricies from the H.264 specification. Choose a value from 17 to 128 to use planar interpolation. Increasing values from 17 to 128 result in increasing reduction of high-frequency data. The value 128 results in the softest video.
4604
+ * @public
4605
4605
  */
4606
4606
  Softness?: number;
4607
4607
  /**
4608
- * @public
4609
4608
  * The best way to set up adaptive quantization is to keep the default value, Auto, for the setting Adaptive quantization. When you do so, MediaConvert automatically applies the best types of quantization for your video content. Include this setting in your JSON job specification only when you choose to change the default value for Adaptive quantization. For this setting, keep the default value, Enabled, to adjust quantization within each frame based on spatial variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas that can sustain more distortion with no noticeable visual degradation and uses more bits on areas where any small distortion will be noticeable. For example, complex textured blocks are encoded with fewer bits and smooth textured blocks are encoded with more bits. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen with a lot of complex texture, you might choose to disable this feature. Related setting: When you enable spatial adaptive quantization, set the value for Adaptive quantization depending on your content. For homogeneous content, such as cartoons and video games, set it to Low. For content with a wider variety of textures, set it to High or Higher.
4609
+ * @public
4610
4610
  */
4611
4611
  SpatialAdaptiveQuantization?: XavcSpatialAdaptiveQuantization;
4612
4612
  /**
4613
- * @public
4614
4613
  * The best way to set up adaptive quantization is to keep the default value, Auto, for the setting Adaptive quantization. When you do so, MediaConvert automatically applies the best types of quantization for your video content. Include this setting in your JSON job specification only when you choose to change the default value for Adaptive quantization. For this setting, keep the default value, Enabled, to adjust quantization within each frame based on temporal variation of content complexity. When you enable this feature, the encoder uses fewer bits on areas of the frame that aren't moving and uses more bits on complex objects with sharp edges that move a lot. For example, this feature improves the readability of text tickers on newscasts and scoreboards on sports matches. Enabling this feature will almost always improve your video quality. Note, though, that this feature doesn't take into account where the viewer's attention is likely to be. If viewers are likely to be focusing their attention on a part of the screen that doesn't have moving objects with sharp edges, such as sports athletes' faces, you might choose to disable this feature. Related setting: When you enable temporal adaptive quantization, adjust the strength of the filter with the setting Adaptive quantization.
4614
+ * @public
4615
4615
  */
4616
4616
  TemporalAdaptiveQuantization?: XavcTemporalAdaptiveQuantization;
4617
4617
  /**
4618
- * @public
4619
4618
  * Required when you set Profile to the value XAVC_4K_INTRA_CBG.
4619
+ * @public
4620
4620
  */
4621
4621
  Xavc4kIntraCbgProfileSettings?: Xavc4kIntraCbgProfileSettings;
4622
4622
  /**
4623
- * @public
4624
4623
  * Required when you set Profile to the value XAVC_4K_INTRA_VBR.
4624
+ * @public
4625
4625
  */
4626
4626
  Xavc4kIntraVbrProfileSettings?: Xavc4kIntraVbrProfileSettings;
4627
4627
  /**
4628
- * @public
4629
4628
  * Required when you set Profile to the value XAVC_4K.
4629
+ * @public
4630
4630
  */
4631
4631
  Xavc4kProfileSettings?: Xavc4kProfileSettings;
4632
4632
  /**
4633
- * @public
4634
4633
  * Required when you set Profile to the value XAVC_HD_INTRA_CBG.
4634
+ * @public
4635
4635
  */
4636
4636
  XavcHdIntraCbgProfileSettings?: XavcHdIntraCbgProfileSettings;
4637
4637
  /**
4638
- * @public
4639
4638
  * Required when you set Profile to the value XAVC_HD.
4639
+ * @public
4640
4640
  */
4641
4641
  XavcHdProfileSettings?: XavcHdProfileSettings;
4642
4642
  }
4643
4643
  /**
4644
- * @public
4645
4644
  * Video codec settings contains the group of settings related to video encoding. The settings in this group vary depending on the value that you choose for Video codec. For each codec enum that you choose, define the corresponding settings object. The following lists the codec enum, settings object pairs. * AV1, Av1Settings * AVC_INTRA, AvcIntraSettings * FRAME_CAPTURE, FrameCaptureSettings * H_264, H264Settings * H_265, H265Settings * MPEG2, Mpeg2Settings * PRORES, ProresSettings * UNCOMPRESSED, UncompressedSettings * VC3, Vc3Settings * VP8, Vp8Settings * VP9, Vp9Settings * XAVC, XavcSettings
4645
+ * @public
4646
4646
  */
4647
4647
  export interface VideoCodecSettings {
4648
4648
  /**
4649
- * @public
4650
4649
  * Required when you set Codec, under VideoDescription>CodecSettings to the value AV1.
4650
+ * @public
4651
4651
  */
4652
4652
  Av1Settings?: Av1Settings;
4653
4653
  /**
4654
- * @public
4655
4654
  * Required when you choose AVC-Intra for your output video codec. For more information about the AVC-Intra settings, see the relevant specification. For detailed information about SD and HD in AVC-Intra, see https://ieeexplore.ieee.org/document/7290936. For information about 4K/2K in AVC-Intra, see https://pro-av.panasonic.net/en/avc-ultra/AVC-ULTRAoverview.pdf.
4655
+ * @public
4656
4656
  */
4657
4657
  AvcIntraSettings?: AvcIntraSettings;
4658
4658
  /**
4659
- * @public
4660
4659
  * Specifies the video codec. This must be equal to one of the enum values defined by the object VideoCodec. To passthrough the video stream of your input JPEG2000, VC-3, AVC-INTRA or Apple ProRes video without any video encoding: Choose Passthrough. If you have multiple input videos, note that they must have identical encoding attributes. When you choose Passthrough, your output container must be MXF or QuickTime MOV.
4660
+ * @public
4661
4661
  */
4662
4662
  Codec?: VideoCodec;
4663
4663
  /**
4664
- * @public
4665
4664
  * Required when you set Codec to the value FRAME_CAPTURE.
4665
+ * @public
4666
4666
  */
4667
4667
  FrameCaptureSettings?: FrameCaptureSettings;
4668
4668
  /**
4669
- * @public
4670
4669
  * Required when you set Codec to the value H_264.
4670
+ * @public
4671
4671
  */
4672
4672
  H264Settings?: H264Settings;
4673
4673
  /**
4674
- * @public
4675
4674
  * Settings for H265 codec
4675
+ * @public
4676
4676
  */
4677
4677
  H265Settings?: H265Settings;
4678
4678
  /**
4679
- * @public
4680
4679
  * Required when you set Codec to the value MPEG2.
4680
+ * @public
4681
4681
  */
4682
4682
  Mpeg2Settings?: Mpeg2Settings;
4683
4683
  /**
4684
- * @public
4685
4684
  * Required when you set Codec to the value PRORES.
4685
+ * @public
4686
4686
  */
4687
4687
  ProresSettings?: ProresSettings;
4688
4688
  /**
4689
- * @public
4690
4689
  * Required when you set Codec, under VideoDescription>CodecSettings to the value UNCOMPRESSED.
4690
+ * @public
4691
4691
  */
4692
4692
  UncompressedSettings?: UncompressedSettings;
4693
4693
  /**
4694
- * @public
4695
4694
  * Required when you set Codec to the value VC3
4695
+ * @public
4696
4696
  */
4697
4697
  Vc3Settings?: Vc3Settings;
4698
4698
  /**
4699
- * @public
4700
4699
  * Required when you set Codec to the value VP8.
4700
+ * @public
4701
4701
  */
4702
4702
  Vp8Settings?: Vp8Settings;
4703
4703
  /**
4704
- * @public
4705
4704
  * Required when you set Codec to the value VP9.
4705
+ * @public
4706
4706
  */
4707
4707
  Vp9Settings?: Vp9Settings;
4708
4708
  /**
4709
- * @public
4710
4709
  * Required when you set Codec to the value XAVC.
4710
+ * @public
4711
4711
  */
4712
4712
  XavcSettings?: XavcSettings;
4713
4713
  }
@@ -4776,28 +4776,28 @@ export declare const VideoTimecodeInsertion: {
4776
4776
  */
4777
4777
  export type VideoTimecodeInsertion = (typeof VideoTimecodeInsertion)[keyof typeof VideoTimecodeInsertion];
4778
4778
  /**
4779
- * @public
4780
4779
  * Specify YUV limits and RGB tolerances when you set Sample range conversion to Limited range clip.
4780
+ * @public
4781
4781
  */
4782
4782
  export interface ClipLimits {
4783
4783
  /**
4784
- * @public
4785
4784
  * Specify the Maximum RGB color sample range tolerance for your output. MediaConvert corrects any YUV values that, when converted to RGB, would be outside the upper tolerance that you specify. Enter an integer from 90 to 105 as an offset percentage to the maximum possible value. Leave blank to use the default value 100. When you specify a value for Maximum RGB tolerance, you must set Sample range conversion to Limited range clip.
4785
+ * @public
4786
4786
  */
4787
4787
  MaximumRGBTolerance?: number;
4788
4788
  /**
4789
- * @public
4790
4789
  * Specify the Maximum YUV color sample limit. MediaConvert conforms any pixels in your input above the value that you specify to typical limited range bounds. Enter an integer from 920 to 1023. Leave blank to use the default value 940. The value that you enter applies to 10-bit ranges. For 8-bit ranges, MediaConvert automatically scales this value down. When you specify a value for Maximum YUV, you must set Sample range conversion to Limited range clip.
4790
+ * @public
4791
4791
  */
4792
4792
  MaximumYUV?: number;
4793
4793
  /**
4794
- * @public
4795
4794
  * Specify the Minimum RGB color sample range tolerance for your output. MediaConvert corrects any YUV values that, when converted to RGB, would be outside the lower tolerance that you specify. Enter an integer from -5 to 10 as an offset percentage to the minimum possible value. Leave blank to use the default value 0. When you specify a value for Minimum RGB tolerance, you must set Sample range conversion to Limited range clip.
4795
+ * @public
4796
4796
  */
4797
4797
  MinimumRGBTolerance?: number;
4798
4798
  /**
4799
- * @public
4800
4799
  * Specify the Minimum YUV color sample limit. MediaConvert conforms any pixels in your input below the value that you specify to typical limited range bounds. Enter an integer from 0 to 128. Leave blank to use the default value 64. The value that you enter applies to 10-bit ranges. For 8-bit ranges, MediaConvert automatically scales this value down. When you specify a value for Minumum YUV, you must set Sample range conversion to Limited range clip.
4800
+ * @public
4801
4801
  */
4802
4802
  MinimumYUV?: number;
4803
4803
  }
@@ -4845,68 +4845,68 @@ export declare const SampleRangeConversion: {
4845
4845
  */
4846
4846
  export type SampleRangeConversion = (typeof SampleRangeConversion)[keyof typeof SampleRangeConversion];
4847
4847
  /**
4848
- * @public
4849
4848
  * Settings for color correction.
4849
+ * @public
4850
4850
  */
4851
4851
  export interface ColorCorrector {
4852
4852
  /**
4853
- * @public
4854
4853
  * Brightness level.
4854
+ * @public
4855
4855
  */
4856
4856
  Brightness?: number;
4857
4857
  /**
4858
- * @public
4859
4858
  * Specify YUV limits and RGB tolerances when you set Sample range conversion to Limited range clip.
4859
+ * @public
4860
4860
  */
4861
4861
  ClipLimits?: ClipLimits;
4862
4862
  /**
4863
- * @public
4864
4863
  * Specify the color space you want for this output. The service supports conversion between HDR formats, between SDR formats, from SDR to HDR, and from HDR to SDR. SDR to HDR conversion doesn't upgrade the dynamic range. The converted video has an HDR format, but visually appears the same as an unconverted output. HDR to SDR conversion uses tone mapping to approximate the outcome of manually regrading from HDR to SDR. When you specify an output color space, MediaConvert uses the following color space metadata, which includes color primaries, transfer characteristics, and matrix coefficients:
4865
4864
  * * HDR 10: BT.2020, PQ, BT.2020 non-constant
4866
4865
  * * HLG 2020: BT.2020, HLG, BT.2020 non-constant
4867
4866
  * * P3DCI (Theater): DCIP3, SMPTE 428M, BT.709
4868
4867
  * * P3D65 (SDR): Display P3, sRGB, BT.709
4869
4868
  * * P3D65 (HDR): Display P3, PQ, BT.709
4869
+ * @public
4870
4870
  */
4871
4871
  ColorSpaceConversion?: ColorSpaceConversion;
4872
4872
  /**
4873
- * @public
4874
4873
  * Contrast level.
4874
+ * @public
4875
4875
  */
4876
4876
  Contrast?: number;
4877
4877
  /**
4878
- * @public
4879
4878
  * Use these settings when you convert to the HDR 10 color space. Specify the SMPTE ST 2086 Mastering Display Color Volume static metadata that you want signaled in the output. These values don't affect the pixel values that are encoded in the video stream. They are intended to help the downstream video player display content in a way that reflects the intentions of the the content creator. When you set Color space conversion to HDR 10, these settings are required. You must set values for Max frame average light level and Max content light level; these settings don't have a default value. The default values for the other HDR 10 metadata settings are defined by the P3D65 color space. For more information about MediaConvert HDR jobs, see https://docs.aws.amazon.com/console/mediaconvert/hdr.
4879
+ * @public
4880
4880
  */
4881
4881
  Hdr10Metadata?: Hdr10Metadata;
4882
4882
  /**
4883
- * @public
4884
4883
  * Specify how MediaConvert maps brightness and colors from your HDR input to your SDR output. The mode that you select represents a creative choice, with different tradeoffs in the details and tones of your output. To maintain details in bright or saturated areas of your output: Choose Preserve details. For some sources, your SDR output may look less bright and less saturated when compared to your HDR source. MediaConvert automatically applies this mode for HLG sources, regardless of your choice. For a bright and saturated output: Choose Vibrant. We recommend that you choose this mode when any of your source content is HDR10, and for the best results when it is mastered for 1000 nits. You may notice loss of details in bright or saturated areas of your output. HDR to SDR tone mapping has no effect when your input is SDR.
4884
+ * @public
4885
4885
  */
4886
4886
  HdrToSdrToneMapper?: HDRToSDRToneMapper;
4887
4887
  /**
4888
- * @public
4889
4888
  * Hue in degrees.
4889
+ * @public
4890
4890
  */
4891
4891
  Hue?: number;
4892
4892
  /**
4893
- * @public
4894
4893
  * Specify the maximum mastering display luminance. Enter an integer from 0 to 2147483647, in units of 0.0001 nits. For example, enter 10000000 for 1000 nits.
4894
+ * @public
4895
4895
  */
4896
4896
  MaxLuminance?: number;
4897
4897
  /**
4898
- * @public
4899
4898
  * Specify how MediaConvert limits the color sample range for this output. To create a limited range output from a full range input: Choose Limited range squeeze. For full range inputs, MediaConvert performs a linear offset to color samples equally across all pixels and frames. Color samples in 10-bit outputs are limited to 64 through 940, and 8-bit outputs are limited to 16 through 235. Note: For limited range inputs, values for color samples are passed through to your output unchanged. MediaConvert does not limit the sample range. To correct pixels in your input that are out of range or out of gamut: Choose Limited range clip. Use for broadcast applications. MediaConvert conforms any pixels outside of the values that you specify under Minimum YUV and Maximum YUV to limited range bounds. MediaConvert also corrects any YUV values that, when converted to RGB, would be outside the bounds you specify under Minimum RGB tolerance and Maximum RGB tolerance. With either limited range conversion, MediaConvert writes the sample range metadata in the output.
4899
+ * @public
4900
4900
  */
4901
4901
  SampleRangeConversion?: SampleRangeConversion;
4902
4902
  /**
4903
- * @public
4904
4903
  * Saturation level.
4904
+ * @public
4905
4905
  */
4906
4906
  Saturation?: number;
4907
4907
  /**
4908
- * @public
4909
4908
  * Specify the reference white level, in nits, for all of your SDR inputs. Use to correct brightness levels within HDR10 outputs. The following color metadata must be present in your SDR input: color primaries, transfer characteristics, and matrix coefficients. If your SDR input has missing color metadata, or if you want to correct input color metadata, manually specify a color space in the input video selector. For 1,000 nit peak brightness displays, we recommend that you set SDR reference white level to 203 (according to ITU-R BT.2408). Leave blank to use the default value of 100, or specify an integer from 100 to 1000.
4909
+ * @public
4910
4910
  */
4911
4911
  SdrReferenceWhiteLevel?: number;
4912
4912
  }
@@ -4951,42 +4951,42 @@ export declare const DeinterlacerMode: {
4951
4951
  */
4952
4952
  export type DeinterlacerMode = (typeof DeinterlacerMode)[keyof typeof DeinterlacerMode];
4953
4953
  /**
4954
- * @public
4955
4954
  * Settings for deinterlacer
4955
+ * @public
4956
4956
  */
4957
4957
  export interface Deinterlacer {
4958
4958
  /**
4959
- * @public
4960
4959
  * Only applies when you set Deinterlace mode to Deinterlace or Adaptive. Interpolate produces sharper pictures, while blend produces smoother motion. If your source file includes a ticker, such as a scrolling headline at the bottom of the frame: Choose Interpolate ticker or Blend ticker. To apply field doubling: Choose Linear interpolation. Note that Linear interpolation may introduce video artifacts into your output.
4960
+ * @public
4961
4961
  */
4962
4962
  Algorithm?: DeinterlaceAlgorithm;
4963
4963
  /**
4964
- * @public
4965
4964
  * - When set to NORMAL (default), the deinterlacer does not convert frames that are tagged in metadata as progressive. It will only convert those that are tagged as some other type. - When set to FORCE_ALL_FRAMES, the deinterlacer converts every frame to progressive - even those that are already tagged as progressive. Turn Force mode on only if there is a good chance that the metadata has tagged frames as progressive when they are not progressive. Do not turn on otherwise; processing frames that are already progressive into progressive will probably result in lower quality video.
4965
+ * @public
4966
4966
  */
4967
4967
  Control?: DeinterlacerControl;
4968
4968
  /**
4969
- * @public
4970
4969
  * Use Deinterlacer to choose how the service will do deinterlacing. Default is Deinterlace.
4971
4970
  * - Deinterlace converts interlaced to progressive.
4972
4971
  * - Inverse telecine converts Hard Telecine 29.97i to progressive 23.976p.
4973
4972
  * - Adaptive auto-detects and converts to progressive.
4973
+ * @public
4974
4974
  */
4975
4975
  Mode?: DeinterlacerMode;
4976
4976
  }
4977
4977
  /**
4978
- * @public
4979
4978
  * Use these settings when you set DolbyVisionLevel6Mode to SPECIFY to override the MaxCLL and MaxFALL values in your input with new values.
4979
+ * @public
4980
4980
  */
4981
4981
  export interface DolbyVisionLevel6Metadata {
4982
4982
  /**
4983
- * @public
4984
4983
  * Maximum Content Light Level. Static HDR metadata that corresponds to the brightest pixel in the entire stream. Measured in nits.
4984
+ * @public
4985
4985
  */
4986
4986
  MaxCll?: number;
4987
4987
  /**
4988
- * @public
4989
4988
  * Maximum Frame-Average Light Level. Static HDR metadata that corresponds to the highest frame-average brightness in the entire stream. Measured in nits.
4989
+ * @public
4990
4990
  */
4991
4991
  MaxFall?: number;
4992
4992
  }
@@ -5028,44 +5028,44 @@ export declare const DolbyVisionProfile: {
5028
5028
  */
5029
5029
  export type DolbyVisionProfile = (typeof DolbyVisionProfile)[keyof typeof DolbyVisionProfile];
5030
5030
  /**
5031
- * @public
5032
5031
  * Create Dolby Vision Profile 5 or Profile 8.1 compatible video output.
5032
+ * @public
5033
5033
  */
5034
5034
  export interface DolbyVision {
5035
5035
  /**
5036
- * @public
5037
5036
  * Use these settings when you set DolbyVisionLevel6Mode to SPECIFY to override the MaxCLL and MaxFALL values in your input with new values.
5037
+ * @public
5038
5038
  */
5039
5039
  L6Metadata?: DolbyVisionLevel6Metadata;
5040
5040
  /**
5041
- * @public
5042
5041
  * Use Dolby Vision Mode to choose how the service will handle Dolby Vision MaxCLL and MaxFALL properies.
5042
+ * @public
5043
5043
  */
5044
5044
  L6Mode?: DolbyVisionLevel6Mode;
5045
5045
  /**
5046
- * @public
5047
5046
  * Required when you set Dolby Vision Profile to Profile 8.1. When you set Content mapping to None, content mapping is not applied to the HDR10-compatible signal. Depending on the source peak nit level, clipping might occur on HDR devices without Dolby Vision. When you set Content mapping to HDR10 1000, the transcoder creates a 1,000 nits peak HDR10-compatible signal by applying static content mapping to the source. This mode is speed-optimized for PQ10 sources with metadata that is created from analysis. For graded Dolby Vision content, be aware that creative intent might not be guaranteed with extreme 1,000 nits trims.
5047
+ * @public
5048
5048
  */
5049
5049
  Mapping?: DolbyVisionMapping;
5050
5050
  /**
5051
- * @public
5052
5051
  * Required when you enable Dolby Vision. Use Profile 5 to include frame-interleaved Dolby Vision metadata in your output. Your input must include Dolby Vision metadata or an HDR10 YUV color space. Use Profile 8.1 to include frame-interleaved Dolby Vision metadata and HDR10 metadata in your output. Your input must include Dolby Vision metadata.
5052
+ * @public
5053
5053
  */
5054
5054
  Profile?: DolbyVisionProfile;
5055
5055
  }
5056
5056
  /**
5057
- * @public
5058
5057
  * Setting for HDR10+ metadata insertion
5058
+ * @public
5059
5059
  */
5060
5060
  export interface Hdr10Plus {
5061
5061
  /**
5062
- * @public
5063
5062
  * Specify the HDR10+ mastering display normalized peak luminance, in nits. This is the normalized actual peak luminance of the mastering display, as defined by ST 2094-40.
5063
+ * @public
5064
5064
  */
5065
5065
  MasteringMonitorNits?: number;
5066
5066
  /**
5067
- * @public
5068
5067
  * Specify the HDR10+ target display nominal peak luminance, in nits. This is the nominal maximum luminance of the target display as defined by ST 2094-40.
5068
+ * @public
5069
5069
  */
5070
5070
  TargetMonitorNits?: number;
5071
5071
  }
@@ -5088,34 +5088,34 @@ export declare const NoiseReducerFilter: {
5088
5088
  */
5089
5089
  export type NoiseReducerFilter = (typeof NoiseReducerFilter)[keyof typeof NoiseReducerFilter];
5090
5090
  /**
5091
- * @public
5092
5091
  * Settings for a noise reducer filter
5092
+ * @public
5093
5093
  */
5094
5094
  export interface NoiseReducerFilterSettings {
5095
5095
  /**
5096
- * @public
5097
5096
  * Relative strength of noise reducing filter. Higher values produce stronger filtering.
5097
+ * @public
5098
5098
  */
5099
5099
  Strength?: number;
5100
5100
  }
5101
5101
  /**
5102
- * @public
5103
5102
  * Noise reducer filter settings for spatial filter.
5103
+ * @public
5104
5104
  */
5105
5105
  export interface NoiseReducerSpatialFilterSettings {
5106
5106
  /**
5107
- * @public
5108
5107
  * Specify strength of post noise reduction sharpening filter, with 0 disabling the filter and 3 enabling it at maximum strength.
5108
+ * @public
5109
5109
  */
5110
5110
  PostFilterSharpenStrength?: number;
5111
5111
  /**
5112
- * @public
5113
5112
  * The speed of the filter, from -2 (lower speed) to 3 (higher speed), with 0 being the nominal value.
5113
+ * @public
5114
5114
  */
5115
5115
  Speed?: number;
5116
5116
  /**
5117
- * @public
5118
5117
  * Relative strength of noise reducing filter. Higher values produce stronger filtering.
5118
+ * @public
5119
5119
  */
5120
5120
  Strength?: number;
5121
5121
  }
@@ -5146,59 +5146,59 @@ export declare const NoiseFilterPostTemporalSharpeningStrength: {
5146
5146
  */
5147
5147
  export type NoiseFilterPostTemporalSharpeningStrength = (typeof NoiseFilterPostTemporalSharpeningStrength)[keyof typeof NoiseFilterPostTemporalSharpeningStrength];
5148
5148
  /**
5149
- * @public
5150
5149
  * Noise reducer filter settings for temporal filter.
5150
+ * @public
5151
5151
  */
5152
5152
  export interface NoiseReducerTemporalFilterSettings {
5153
5153
  /**
5154
- * @public
5155
5154
  * Use Aggressive mode for content that has complex motion. Higher values produce stronger temporal filtering. This filters highly complex scenes more aggressively and creates better VQ for low bitrate outputs.
5155
+ * @public
5156
5156
  */
5157
5157
  AggressiveMode?: number;
5158
5158
  /**
5159
- * @public
5160
5159
  * When you set Noise reducer to Temporal, the bandwidth and sharpness of your output is reduced. You can optionally use Post temporal sharpening to apply sharpening to the edges of your output. Note that Post temporal sharpening will also make the bandwidth reduction from the Noise reducer smaller. The default behavior, Auto, allows the transcoder to determine whether to apply sharpening, depending on your input type and quality. When you set Post temporal sharpening to Enabled, specify how much sharpening is applied using Post temporal sharpening strength. Set Post temporal sharpening to Disabled to not apply sharpening.
5160
+ * @public
5161
5161
  */
5162
5162
  PostTemporalSharpening?: NoiseFilterPostTemporalSharpening;
5163
5163
  /**
5164
- * @public
5165
5164
  * Use Post temporal sharpening strength to define the amount of sharpening the transcoder applies to your output. Set Post temporal sharpening strength to Low, Medium, or High to indicate the amount of sharpening.
5165
+ * @public
5166
5166
  */
5167
5167
  PostTemporalSharpeningStrength?: NoiseFilterPostTemporalSharpeningStrength;
5168
5168
  /**
5169
- * @public
5170
5169
  * The speed of the filter (higher number is faster). Low setting reduces bit rate at the cost of transcode time, high setting improves transcode time at the cost of bit rate.
5170
+ * @public
5171
5171
  */
5172
5172
  Speed?: number;
5173
5173
  /**
5174
- * @public
5175
5174
  * Specify the strength of the noise reducing filter on this output. Higher values produce stronger filtering. We recommend the following value ranges, depending on the result that you want: * 0-2 for complexity reduction with minimal sharpness loss * 2-8 for complexity reduction with image preservation * 8-16 for a high level of complexity reduction
5175
+ * @public
5176
5176
  */
5177
5177
  Strength?: number;
5178
5178
  }
5179
5179
  /**
5180
- * @public
5181
5180
  * Enable the Noise reducer feature to remove noise from your video output if necessary. Enable or disable this feature for each output individually. This setting is disabled by default. When you enable Noise reducer, you must also select a value for Noise reducer filter. For AVC outputs, when you include Noise reducer, you cannot include the Bandwidth reduction filter.
5181
+ * @public
5182
5182
  */
5183
5183
  export interface NoiseReducer {
5184
5184
  /**
5185
- * @public
5186
5185
  * Use Noise reducer filter to select one of the following spatial image filtering functions. To use this setting, you must also enable Noise reducer. * Bilateral preserves edges while reducing noise. * Mean (softest), Gaussian, Lanczos, and Sharpen (sharpest) do convolution filtering. * Conserve does min/max noise reduction. * Spatial does frequency-domain filtering based on JND principles. * Temporal optimizes video quality for complex motion.
5186
+ * @public
5187
5187
  */
5188
5188
  Filter?: NoiseReducerFilter;
5189
5189
  /**
5190
- * @public
5191
5190
  * Settings for a noise reducer filter
5191
+ * @public
5192
5192
  */
5193
5193
  FilterSettings?: NoiseReducerFilterSettings;
5194
5194
  /**
5195
- * @public
5196
5195
  * Noise reducer filter settings for spatial filter.
5196
+ * @public
5197
5197
  */
5198
5198
  SpatialFilterSettings?: NoiseReducerSpatialFilterSettings;
5199
5199
  /**
5200
- * @public
5201
5200
  * Noise reducer filter settings for temporal filter.
5201
+ * @public
5202
5202
  */
5203
5203
  TemporalFilterSettings?: NoiseReducerTemporalFilterSettings;
5204
5204
  }
@@ -5218,39 +5218,39 @@ export declare const WatermarkingStrength: {
5218
5218
  */
5219
5219
  export type WatermarkingStrength = (typeof WatermarkingStrength)[keyof typeof WatermarkingStrength];
5220
5220
  /**
5221
- * @public
5222
5221
  * For forensic video watermarking, MediaConvert supports Nagra NexGuard File Marker watermarking. MediaConvert supports both PreRelease Content (NGPR/G2) and OTT Streaming workflows.
5222
+ * @public
5223
5223
  */
5224
5224
  export interface NexGuardFileMarkerSettings {
5225
5225
  /**
5226
- * @public
5227
5226
  * Use the base64 license string that Nagra provides you. Enter it directly in your JSON job specification or in the console. Required when you include Nagra NexGuard File Marker watermarking in your job.
5227
+ * @public
5228
5228
  */
5229
5229
  License?: string;
5230
5230
  /**
5231
- * @public
5232
5231
  * Specify the payload ID that you want associated with this output. Valid values vary depending on your Nagra NexGuard forensic watermarking workflow. Required when you include Nagra NexGuard File Marker watermarking in your job. For PreRelease Content (NGPR/G2), specify an integer from 1 through 4,194,303. You must generate a unique ID for each asset you watermark, and keep a record of which ID you have assigned to each asset. Neither Nagra nor MediaConvert keep track of the relationship between output files and your IDs. For OTT Streaming, create two adaptive bitrate (ABR) stacks for each asset. Do this by setting up two output groups. For one output group, set the value of Payload ID to 0 in every output. For the other output group, set Payload ID to 1 in every output.
5232
+ * @public
5233
5233
  */
5234
5234
  Payload?: number;
5235
5235
  /**
5236
- * @public
5237
5236
  * Enter one of the watermarking preset strings that Nagra provides you. Required when you include Nagra NexGuard File Marker watermarking in your job.
5237
+ * @public
5238
5238
  */
5239
5239
  Preset?: string;
5240
5240
  /**
5241
- * @public
5242
5241
  * Optional. Ignore this setting unless Nagra support directs you to specify a value. When you don't specify a value here, the Nagra NexGuard library uses its default value.
5242
+ * @public
5243
5243
  */
5244
5244
  Strength?: WatermarkingStrength;
5245
5245
  }
5246
5246
  /**
5247
- * @public
5248
5247
  * If you work with a third party video watermarking partner, use the group of settings that correspond with your watermarking partner to include watermarks in your output.
5248
+ * @public
5249
5249
  */
5250
5250
  export interface PartnerWatermarking {
5251
5251
  /**
5252
- * @public
5253
5252
  * For forensic video watermarking, MediaConvert supports Nagra NexGuard File Marker watermarking. MediaConvert supports both PreRelease Content (NGPR/G2) and OTT Streaming workflows.
5253
+ * @public
5254
5254
  */
5255
5255
  NexguardFileMarkerSettings?: NexGuardFileMarkerSettings;
5256
5256
  }
@@ -5274,227 +5274,227 @@ export declare const TimecodeBurninPosition: {
5274
5274
  */
5275
5275
  export type TimecodeBurninPosition = (typeof TimecodeBurninPosition)[keyof typeof TimecodeBurninPosition];
5276
5276
  /**
5277
- * @public
5278
5277
  * Settings for burning the output timecode and specified prefix into the output.
5278
+ * @public
5279
5279
  */
5280
5280
  export interface TimecodeBurnin {
5281
5281
  /**
5282
- * @public
5283
5282
  * Use Font size to set the font size of any burned-in timecode. Valid values are 10, 16, 32, 48.
5283
+ * @public
5284
5284
  */
5285
5285
  FontSize?: number;
5286
5286
  /**
5287
- * @public
5288
5287
  * Use Position under Timecode burn-in to specify the location the burned-in timecode on output video.
5288
+ * @public
5289
5289
  */
5290
5290
  Position?: TimecodeBurninPosition;
5291
5291
  /**
5292
- * @public
5293
5292
  * Use Prefix to place ASCII characters before any burned-in timecode. For example, a prefix of "EZ-" will result in the timecode "EZ-00:00:00:00". Provide either the characters themselves or the ASCII code equivalents. The supported range of characters is 0x20 through 0x7e. This includes letters, numbers, and all special characters represented on a standard English keyboard.
5293
+ * @public
5294
5294
  */
5295
5295
  Prefix?: string;
5296
5296
  }
5297
5297
  /**
5298
- * @public
5299
5298
  * Find additional transcoding features under Preprocessors. Enable the features at each output individually. These features are disabled by default.
5299
+ * @public
5300
5300
  */
5301
5301
  export interface VideoPreprocessor {
5302
5302
  /**
5303
- * @public
5304
5303
  * Use these settings to convert the color space or to modify properties such as hue and contrast for this output. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/converting-the-color-space.html.
5304
+ * @public
5305
5305
  */
5306
5306
  ColorCorrector?: ColorCorrector;
5307
5307
  /**
5308
- * @public
5309
5308
  * Use the deinterlacer to produce smoother motion and a clearer picture. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/working-with-scan-type.html.
5309
+ * @public
5310
5310
  */
5311
5311
  Deinterlacer?: Deinterlacer;
5312
5312
  /**
5313
- * @public
5314
5313
  * Enable Dolby Vision feature to produce Dolby Vision compatible video output.
5314
+ * @public
5315
5315
  */
5316
5316
  DolbyVision?: DolbyVision;
5317
5317
  /**
5318
- * @public
5319
5318
  * Enable HDR10+ analysis and metadata injection. Compatible with HEVC only.
5319
+ * @public
5320
5320
  */
5321
5321
  Hdr10Plus?: Hdr10Plus;
5322
5322
  /**
5323
- * @public
5324
5323
  * Enable the Image inserter feature to include a graphic overlay on your video. Enable or disable this feature for each output individually. This setting is disabled by default.
5324
+ * @public
5325
5325
  */
5326
5326
  ImageInserter?: ImageInserter;
5327
5327
  /**
5328
- * @public
5329
5328
  * Enable the Noise reducer feature to remove noise from your video output if necessary. Enable or disable this feature for each output individually. This setting is disabled by default. When you enable Noise reducer, you must also select a value for Noise reducer filter. For AVC outputs, when you include Noise reducer, you cannot include the Bandwidth reduction filter.
5329
+ * @public
5330
5330
  */
5331
5331
  NoiseReducer?: NoiseReducer;
5332
5332
  /**
5333
- * @public
5334
5333
  * If you work with a third party video watermarking partner, use the group of settings that correspond with your watermarking partner to include watermarks in your output.
5334
+ * @public
5335
5335
  */
5336
5336
  PartnerWatermarking?: PartnerWatermarking;
5337
5337
  /**
5338
- * @public
5339
5338
  * Settings for burning the output timecode and specified prefix into the output.
5339
+ * @public
5340
5340
  */
5341
5341
  TimecodeBurnin?: TimecodeBurnin;
5342
5342
  }
5343
5343
  /**
5344
- * @public
5345
5344
  * Settings related to video encoding of your output. The specific video settings depend on the video codec that you choose.
5345
+ * @public
5346
5346
  */
5347
5347
  export interface VideoDescription {
5348
5348
  /**
5349
- * @public
5350
5349
  * This setting only applies to H.264, H.265, and MPEG2 outputs. Use Insert AFD signaling to specify whether the service includes AFD values in the output video data and what those values are. * Choose None to remove all AFD values from this output. * Choose Fixed to ignore input AFD values and instead encode the value specified in the job. * Choose Auto to calculate output AFD values based on the input AFD scaler data.
5350
+ * @public
5351
5351
  */
5352
5352
  AfdSignaling?: AfdSignaling;
5353
5353
  /**
5354
- * @public
5355
5354
  * The anti-alias filter is automatically applied to all outputs. The service no longer accepts the value DISABLED for AntiAlias. If you specify that in your job, the service will ignore the setting.
5355
+ * @public
5356
5356
  */
5357
5357
  AntiAlias?: AntiAlias;
5358
5358
  /**
5359
- * @public
5360
5359
  * Video codec settings contains the group of settings related to video encoding. The settings in this group vary depending on the value that you choose for Video codec. For each codec enum that you choose, define the corresponding settings object. The following lists the codec enum, settings object pairs. * AV1, Av1Settings * AVC_INTRA, AvcIntraSettings * FRAME_CAPTURE, FrameCaptureSettings * H_264, H264Settings * H_265, H265Settings * MPEG2, Mpeg2Settings * PRORES, ProresSettings * UNCOMPRESSED, UncompressedSettings * VC3, Vc3Settings * VP8, Vp8Settings * VP9, Vp9Settings * XAVC, XavcSettings
5360
+ * @public
5361
5361
  */
5362
5362
  CodecSettings?: VideoCodecSettings;
5363
5363
  /**
5364
- * @public
5365
5364
  * Choose Insert for this setting to include color metadata in this output. Choose Ignore to exclude color metadata from this output. If you don't specify a value, the service sets this to Insert by default.
5365
+ * @public
5366
5366
  */
5367
5367
  ColorMetadata?: ColorMetadata;
5368
5368
  /**
5369
- * @public
5370
5369
  * Use Cropping selection to specify the video area that the service will include in the output video frame.
5370
+ * @public
5371
5371
  */
5372
5372
  Crop?: Rectangle;
5373
5373
  /**
5374
- * @public
5375
5374
  * Applies only to 29.97 fps outputs. When this feature is enabled, the service will use drop-frame timecode on outputs. If it is not possible to use drop-frame timecode, the system will fall back to non-drop-frame. This setting is enabled by default when Timecode insertion is enabled.
5375
+ * @public
5376
5376
  */
5377
5377
  DropFrameTimecode?: DropFrameTimecode;
5378
5378
  /**
5379
- * @public
5380
5379
  * Applies only if you set AFD Signaling to Fixed. Use Fixed to specify a four-bit AFD value which the service will write on all frames of this video output.
5380
+ * @public
5381
5381
  */
5382
5382
  FixedAfd?: number;
5383
5383
  /**
5384
- * @public
5385
5384
  * Use Height to define the video resolution height, in pixels, for this output. To use the same resolution as your input: Leave both Width and Height blank. To evenly scale from your input resolution: Leave Height blank and enter a value for Width. For example, if your input is 1920x1080 and you set Width to 1280, your output will be 1280x720.
5385
+ * @public
5386
5386
  */
5387
5387
  Height?: number;
5388
5388
  /**
5389
- * @public
5390
5389
  * Use Selection placement to define the video area in your output frame. The area outside of the rectangle that you specify here is black.
5390
+ * @public
5391
5391
  */
5392
5392
  Position?: Rectangle;
5393
5393
  /**
5394
- * @public
5395
5394
  * Use Respond to AFD to specify how the service changes the video itself in response to AFD values in the input. * Choose Respond to clip the input video frame according to the AFD value, input display aspect ratio, and output display aspect ratio. * Choose Passthrough to include the input AFD values. Do not choose this when AfdSignaling is set to NONE. A preferred implementation of this workflow is to set RespondToAfd to and set AfdSignaling to AUTO. * Choose None to remove all input AFD values from this output.
5395
+ * @public
5396
5396
  */
5397
5397
  RespondToAfd?: RespondToAfd;
5398
5398
  /**
5399
- * @public
5400
5399
  * Specify the video Scaling behavior when your output has a different resolution than your input. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/video-scaling.html
5400
+ * @public
5401
5401
  */
5402
5402
  ScalingBehavior?: ScalingBehavior;
5403
5403
  /**
5404
- * @public
5405
5404
  * Use Sharpness setting to specify the strength of anti-aliasing. This setting changes the width of the anti-alias filter kernel used for scaling. Sharpness only applies if your output resolution is different from your input resolution. 0 is the softest setting, 100 the sharpest, and 50 recommended for most content.
5405
+ * @public
5406
5406
  */
5407
5407
  Sharpness?: number;
5408
5408
  /**
5409
- * @public
5410
5409
  * Applies only to H.264, H.265, MPEG2, and ProRes outputs. Only enable Timecode insertion when the input frame rate is identical to the output frame rate. To include timecodes in this output, set Timecode insertion to PIC_TIMING_SEI. To leave them out, set it to DISABLED. Default is DISABLED. When the service inserts timecodes in an output, by default, it uses any embedded timecodes from the input. If none are present, the service will set the timecode for the first output frame to zero. To change this default behavior, adjust the settings under Timecode configuration. In the console, these settings are located under Job > Job settings > Timecode configuration. Note - Timecode source under input settings does not affect the timecodes that are inserted in the output. Source under Job settings > Timecode configuration does.
5410
+ * @public
5411
5411
  */
5412
5412
  TimecodeInsertion?: VideoTimecodeInsertion;
5413
5413
  /**
5414
- * @public
5415
5414
  * Find additional transcoding features under Preprocessors. Enable the features at each output individually. These features are disabled by default.
5415
+ * @public
5416
5416
  */
5417
5417
  VideoPreprocessors?: VideoPreprocessor;
5418
5418
  /**
5419
- * @public
5420
5419
  * Use Width to define the video resolution width, in pixels, for this output. To use the same resolution as your input: Leave both Width and Height blank. To evenly scale from your input resolution: Leave Width blank and enter a value for Height. For example, if your input is 1920x1080 and you set Height to 720, your output will be 1280x720.
5420
+ * @public
5421
5421
  */
5422
5422
  Width?: number;
5423
5423
  }
5424
5424
  /**
5425
- * @public
5426
5425
  * Each output in your job is a collection of settings that describes how you want MediaConvert to encode a single output file or stream. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/create-outputs.html.
5426
+ * @public
5427
5427
  */
5428
5428
  export interface Output {
5429
5429
  /**
5430
- * @public
5431
5430
  * Contains groups of audio encoding settings organized by audio codec. Include one instance of per output. Can contain multiple groups of encoding settings.
5431
+ * @public
5432
5432
  */
5433
5433
  AudioDescriptions?: AudioDescription[];
5434
5434
  /**
5435
- * @public
5436
5435
  * Contains groups of captions settings. For each output that has captions, include one instance of CaptionDescriptions. Can contain multiple groups of captions settings.
5436
+ * @public
5437
5437
  */
5438
5438
  CaptionDescriptions?: CaptionDescription[];
5439
5439
  /**
5440
- * @public
5441
5440
  * Container specific settings.
5441
+ * @public
5442
5442
  */
5443
5443
  ContainerSettings?: ContainerSettings;
5444
5444
  /**
5445
- * @public
5446
5445
  * Use Extension to specify the file extension for outputs in File output groups. If you do not specify a value, the service will use default extensions by container type as follows * MPEG-2 transport stream, m2ts * Quicktime, mov * MXF container, mxf * MPEG-4 container, mp4 * WebM container, webm * No Container, the service will use codec extensions (e.g. AAC, H265, H265, AC3)
5446
+ * @public
5447
5447
  */
5448
5448
  Extension?: string;
5449
5449
  /**
5450
- * @public
5451
5450
  * Use Name modifier to have the service add a string to the end of each output filename. You specify the base filename as part of your destination URI. When you create multiple outputs in the same output group, Name modifier is required. Name modifier also accepts format identifiers. For DASH ISO outputs, if you use the format identifiers $Number$ or $Time$ in one output, you must use them in the same way in all outputs of the output group.
5451
+ * @public
5452
5452
  */
5453
5453
  NameModifier?: string;
5454
5454
  /**
5455
- * @public
5456
5455
  * Specific settings for this type of output.
5456
+ * @public
5457
5457
  */
5458
5458
  OutputSettings?: OutputSettings;
5459
5459
  /**
5460
- * @public
5461
5460
  * Use Preset to specify a preset for your transcoding settings. Provide the system or custom preset name. You can specify either Preset or Container settings, but not both.
5461
+ * @public
5462
5462
  */
5463
5463
  Preset?: string;
5464
5464
  /**
5465
- * @public
5466
5465
  * VideoDescription contains a group of video encoding settings. The specific video settings depend on the video codec that you choose for the property codec. Include one instance of VideoDescription per output.
5466
+ * @public
5467
5467
  */
5468
5468
  VideoDescription?: VideoDescription;
5469
5469
  }
5470
5470
  /**
5471
- * @public
5472
5471
  * Group of outputs
5472
+ * @public
5473
5473
  */
5474
5474
  export interface OutputGroup {
5475
5475
  /**
5476
- * @public
5477
5476
  * Use automated encoding to have MediaConvert choose your encoding settings for you, based on characteristics of your input video.
5477
+ * @public
5478
5478
  */
5479
5479
  AutomatedEncodingSettings?: AutomatedEncodingSettings;
5480
5480
  /**
5481
- * @public
5482
5481
  * Use Custom Group Name to specify a name for the output group. This value is displayed on the console and can make your job settings JSON more human-readable. It does not affect your outputs. Use up to twelve characters that are either letters, numbers, spaces, or underscores.
5482
+ * @public
5483
5483
  */
5484
5484
  CustomName?: string;
5485
5485
  /**
5486
- * @public
5487
5486
  * Name of the output group
5487
+ * @public
5488
5488
  */
5489
5489
  Name?: string;
5490
5490
  /**
5491
- * @public
5492
5491
  * Output Group settings, including type
5492
+ * @public
5493
5493
  */
5494
5494
  OutputGroupSettings?: OutputGroupSettings;
5495
5495
  /**
5496
- * @public
5497
5496
  * This object holds groups of encoding settings, one group of settings per output.
5497
+ * @public
5498
5498
  */
5499
5499
  Outputs?: Output[];
5500
5500
  }
@@ -5512,115 +5512,115 @@ export declare const TimecodeSource: {
5512
5512
  */
5513
5513
  export type TimecodeSource = (typeof TimecodeSource)[keyof typeof TimecodeSource];
5514
5514
  /**
5515
- * @public
5516
5515
  * These settings control how the service handles timecodes throughout the job. These settings don't affect input clipping.
5516
+ * @public
5517
5517
  */
5518
5518
  export interface TimecodeConfig {
5519
5519
  /**
5520
- * @public
5521
5520
  * If you use an editing platform that relies on an anchor timecode, use Anchor Timecode to specify a timecode that will match the input video frame to the output video frame. Use 24-hour format with frame number, (HH:MM:SS:FF) or (HH:MM:SS;FF). This setting ignores frame rate conversion. System behavior for Anchor Timecode varies depending on your setting for Source. * If Source is set to Specified Start, the first input frame is the specified value in Start Timecode. Anchor Timecode and Start Timecode are used calculate output timecode. * If Source is set to Start at 0 the first frame is 00:00:00:00. * If Source is set to Embedded, the first frame is the timecode value on the first input frame of the input.
5521
+ * @public
5522
5522
  */
5523
5523
  Anchor?: string;
5524
5524
  /**
5525
- * @public
5526
5525
  * Use Source to set how timecodes are handled within this job. To make sure that your video, audio, captions, and markers are synchronized and that time-based features, such as image inserter, work correctly, choose the Timecode source option that matches your assets. All timecodes are in a 24-hour format with frame number (HH:MM:SS:FF). * Embedded - Use the timecode that is in the input video. If no embedded timecode is in the source, the service will use Start at 0 instead. * Start at 0 - Set the timecode of the initial frame to 00:00:00:00. * Specified Start - Set the timecode of the initial frame to a value other than zero. You use Start timecode to provide this value.
5526
+ * @public
5527
5527
  */
5528
5528
  Source?: TimecodeSource;
5529
5529
  /**
5530
- * @public
5531
5530
  * Only use when you set Source to Specified start. Use Start timecode to specify the timecode for the initial frame. Use 24-hour format with frame number, (HH:MM:SS:FF) or (HH:MM:SS;FF).
5531
+ * @public
5532
5532
  */
5533
5533
  Start?: string;
5534
5534
  /**
5535
- * @public
5536
5535
  * Only applies to outputs that support program-date-time stamp. Use Timestamp offset to overwrite the timecode date without affecting the time and frame number. Provide the new date as a string in the format "yyyy-mm-dd". To use Timestamp offset, you must also enable Insert program-date-time in the output settings. For example, if the date part of your timecodes is 2002-1-25 and you want to change it to one year later, set Timestamp offset to 2003-1-25.
5536
+ * @public
5537
5537
  */
5538
5538
  TimestampOffset?: string;
5539
5539
  }
5540
5540
  /**
5541
- * @public
5542
5541
  * Insert user-defined custom ID3 metadata at timecodes that you specify. In each output that you want to include this metadata, you must set ID3 metadata to Passthrough.
5542
+ * @public
5543
5543
  */
5544
5544
  export interface TimedMetadataInsertion {
5545
5545
  /**
5546
- * @public
5547
5546
  * Id3Insertions contains the array of Id3Insertion instances.
5547
+ * @public
5548
5548
  */
5549
5549
  Id3Insertions?: Id3Insertion[];
5550
5550
  }
5551
5551
  /**
5552
- * @public
5553
5552
  * JobSettings contains all the transcode settings for a job.
5553
+ * @public
5554
5554
  */
5555
5555
  export interface JobSettings {
5556
5556
  /**
5557
- * @public
5558
5557
  * When specified, this offset (in milliseconds) is added to the input Ad Avail PTS time.
5558
+ * @public
5559
5559
  */
5560
5560
  AdAvailOffset?: number;
5561
5561
  /**
5562
- * @public
5563
5562
  * Settings for ad avail blanking. Video can be blanked or overlaid with an image, and audio muted during SCTE-35 triggered ad avails.
5563
+ * @public
5564
5564
  */
5565
5565
  AvailBlanking?: AvailBlanking;
5566
5566
  /**
5567
- * @public
5568
5567
  * Use 3D LUTs to specify custom color mapping behavior when you convert from one color space into another. You can include up to 8 different 3D LUTs.
5568
+ * @public
5569
5569
  */
5570
5570
  ColorConversion3DLUTSettings?: ColorConversion3DLUTSetting[];
5571
5571
  /**
5572
- * @public
5573
5572
  * Settings for Event Signaling And Messaging (ESAM). If you don't do ad insertion, you can ignore these settings.
5573
+ * @public
5574
5574
  */
5575
5575
  Esam?: EsamSettings;
5576
5576
  /**
5577
- * @public
5578
5577
  * If your source content has EIA-608 Line 21 Data Services, enable this feature to specify what MediaConvert does with the Extended Data Services (XDS) packets. You can choose to pass through XDS packets, or remove them from the output. For more information about XDS, see EIA-608 Line Data Services, section 9.5.1.5 05h Content Advisory.
5578
+ * @public
5579
5579
  */
5580
5580
  ExtendedDataServices?: ExtendedDataServices;
5581
5581
  /**
5582
- * @public
5583
5582
  * Specify the input that MediaConvert references for your default output settings. MediaConvert uses this input's Resolution, Frame rate, and Pixel aspect ratio for all outputs that you don't manually specify different output settings for. Enabling this setting will disable "Follow source" for all other inputs. If MediaConvert cannot follow your source, for example if you specify an audio-only input, MediaConvert uses the first followable input instead. In your JSON job specification, enter an integer from 1 to 150 corresponding to the order of your inputs.
5583
+ * @public
5584
5584
  */
5585
5585
  FollowSource?: number;
5586
5586
  /**
5587
- * @public
5588
5587
  * Use Inputs to define source file used in the transcode job. There can be multiple inputs add in a job. These inputs will be concantenated together to create the output.
5588
+ * @public
5589
5589
  */
5590
5590
  Inputs?: Input[];
5591
5591
  /**
5592
- * @public
5593
5592
  * Use these settings only when you use Kantar watermarking. Specify the values that MediaConvert uses to generate and place Kantar watermarks in your output audio. These settings apply to every output in your job. In addition to specifying these values, you also need to store your Kantar credentials in AWS Secrets Manager. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/kantar-watermarking.html.
5593
+ * @public
5594
5594
  */
5595
5595
  KantarWatermark?: KantarWatermarkSettings;
5596
5596
  /**
5597
- * @public
5598
5597
  * Overlay motion graphics on top of your video. The motion graphics that you specify here appear on all outputs in all output groups. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/motion-graphic-overlay.html.
5598
+ * @public
5599
5599
  */
5600
5600
  MotionImageInserter?: MotionImageInserter;
5601
5601
  /**
5602
- * @public
5603
5602
  * Settings for your Nielsen configuration. If you don't do Nielsen measurement and analytics, ignore these settings. When you enable Nielsen configuration, MediaConvert enables PCM to ID3 tagging for all outputs in the job.
5603
+ * @public
5604
5604
  */
5605
5605
  NielsenConfiguration?: NielsenConfiguration;
5606
5606
  /**
5607
- * @public
5608
5607
  * Ignore these settings unless you are using Nielsen non-linear watermarking. Specify the values that MediaConvert uses to generate and place Nielsen watermarks in your output audio. In addition to specifying these values, you also need to set up your cloud TIC server. These settings apply to every output in your job. The MediaConvert implementation is currently with the following Nielsen versions: Nielsen Watermark SDK Version 5.2.1 Nielsen NLM Watermark Engine Version 1.2.7 Nielsen Watermark Authenticator [SID_TIC] Version [5.0.0]
5608
+ * @public
5609
5609
  */
5610
5610
  NielsenNonLinearWatermark?: NielsenNonLinearWatermarkSettings;
5611
5611
  /**
5612
- * @public
5613
5612
  * Contains one group of settings for each set of outputs that share a common package type. All unpackaged files (MPEG-4, MPEG-2 TS, Quicktime, MXF, and no container) are grouped in a single output group as well. Required in is a group of settings that apply to the whole group. This required object depends on the value you set for Type. Type, settings object pairs are as follows. * FILE_GROUP_SETTINGS, FileGroupSettings * HLS_GROUP_SETTINGS, HlsGroupSettings * DASH_ISO_GROUP_SETTINGS, DashIsoGroupSettings * MS_SMOOTH_GROUP_SETTINGS, MsSmoothGroupSettings * CMAF_GROUP_SETTINGS, CmafGroupSettings
5613
+ * @public
5614
5614
  */
5615
5615
  OutputGroups?: OutputGroup[];
5616
5616
  /**
5617
- * @public
5618
5617
  * These settings control how the service handles timecodes throughout the job. These settings don't affect input clipping.
5618
+ * @public
5619
5619
  */
5620
5620
  TimecodeConfig?: TimecodeConfig;
5621
5621
  /**
5622
- * @public
5623
5622
  * Insert user-defined custom ID3 metadata at timecodes that you specify. In each output that you want to include this metadata, you must set ID3 metadata to Passthrough.
5623
+ * @public
5624
5624
  */
5625
5625
  TimedMetadataInsertion?: TimedMetadataInsertion;
5626
5626
  }
@@ -5677,256 +5677,256 @@ export declare const StatusUpdateInterval: {
5677
5677
  */
5678
5678
  export type StatusUpdateInterval = (typeof StatusUpdateInterval)[keyof typeof StatusUpdateInterval];
5679
5679
  /**
5680
- * @public
5681
5680
  * Information about when jobs are submitted, started, and finished is specified in Unix epoch format in seconds.
5681
+ * @public
5682
5682
  */
5683
5683
  export interface Timing {
5684
5684
  /**
5685
- * @public
5686
5685
  * The time, in Unix epoch format, that the transcoding job finished
5686
+ * @public
5687
5687
  */
5688
5688
  FinishTime?: Date;
5689
5689
  /**
5690
- * @public
5691
5690
  * The time, in Unix epoch format, that transcoding for the job began.
5691
+ * @public
5692
5692
  */
5693
5693
  StartTime?: Date;
5694
5694
  /**
5695
- * @public
5696
5695
  * The time, in Unix epoch format, that you submitted the job.
5696
+ * @public
5697
5697
  */
5698
5698
  SubmitTime?: Date;
5699
5699
  }
5700
5700
  /**
5701
- * @public
5702
5701
  * Contains any warning codes and their count for the job.
5702
+ * @public
5703
5703
  */
5704
5704
  export interface WarningGroup {
5705
5705
  /**
5706
- * @public
5707
5706
  * Warning code that identifies a specific warning in the job. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/warning_codes.html
5707
+ * @public
5708
5708
  */
5709
5709
  Code: number | undefined;
5710
5710
  /**
5711
- * @public
5712
5711
  * The number of times this warning occurred in the job.
5712
+ * @public
5713
5713
  */
5714
5714
  Count: number | undefined;
5715
5715
  }
5716
5716
  /**
5717
- * @public
5718
5717
  * Each job converts an input file into an output file or files. For more information, see the User Guide at https://docs.aws.amazon.com/mediaconvert/latest/ug/what-is.html
5718
+ * @public
5719
5719
  */
5720
5720
  export interface Job {
5721
5721
  /**
5722
- * @public
5723
5722
  * Accelerated transcoding can significantly speed up jobs with long, visually complex content.
5723
+ * @public
5724
5724
  */
5725
5725
  AccelerationSettings?: AccelerationSettings;
5726
5726
  /**
5727
- * @public
5728
5727
  * Describes whether the current job is running with accelerated transcoding. For jobs that have Acceleration (AccelerationMode) set to DISABLED, AccelerationStatus is always NOT_APPLICABLE. For jobs that have Acceleration (AccelerationMode) set to ENABLED or PREFERRED, AccelerationStatus is one of the other states. AccelerationStatus is IN_PROGRESS initially, while the service determines whether the input files and job settings are compatible with accelerated transcoding. If they are, AcclerationStatus is ACCELERATED. If your input files and job settings aren't compatible with accelerated transcoding, the service either fails your job or runs it without accelerated transcoding, depending on how you set Acceleration (AccelerationMode). When the service runs your job without accelerated transcoding, AccelerationStatus is NOT_ACCELERATED.
5728
+ * @public
5729
5729
  */
5730
5730
  AccelerationStatus?: AccelerationStatus;
5731
5731
  /**
5732
- * @public
5733
5732
  * An identifier for this resource that is unique within all of AWS.
5733
+ * @public
5734
5734
  */
5735
5735
  Arn?: string;
5736
5736
  /**
5737
- * @public
5738
5737
  * The tag type that AWS Billing and Cost Management will use to sort your AWS Elemental MediaConvert costs on any billing report that you set up.
5738
+ * @public
5739
5739
  */
5740
5740
  BillingTagsSource?: BillingTagsSource;
5741
5741
  /**
5742
- * @public
5743
5742
  * Prevent duplicate jobs from being created and ensure idempotency for your requests. A client request token can be any string that includes up to 64 ASCII characters. If you reuse a client request token within one minute of a successful request, the API returns the job details of the original request instead. For more information see https://docs.aws.amazon.com/mediaconvert/latest/apireference/idempotency.html.
5743
+ * @public
5744
5744
  */
5745
5745
  ClientRequestToken?: string;
5746
5746
  /**
5747
- * @public
5748
5747
  * The time, in Unix epoch format in seconds, when the job got created.
5748
+ * @public
5749
5749
  */
5750
5750
  CreatedAt?: Date;
5751
5751
  /**
5752
- * @public
5753
5752
  * A job's phase can be PROBING, TRANSCODING OR UPLOADING
5753
+ * @public
5754
5754
  */
5755
5755
  CurrentPhase?: JobPhase;
5756
5756
  /**
5757
- * @public
5758
5757
  * Error code for the job
5758
+ * @public
5759
5759
  */
5760
5760
  ErrorCode?: number;
5761
5761
  /**
5762
- * @public
5763
5762
  * Error message of Job
5763
+ * @public
5764
5764
  */
5765
5765
  ErrorMessage?: string;
5766
5766
  /**
5767
- * @public
5768
5767
  * Optional list of hop destinations.
5768
+ * @public
5769
5769
  */
5770
5770
  HopDestinations?: HopDestination[];
5771
5771
  /**
5772
- * @public
5773
5772
  * A portion of the job's ARN, unique within your AWS Elemental MediaConvert resources
5773
+ * @public
5774
5774
  */
5775
5775
  Id?: string;
5776
5776
  /**
5777
- * @public
5778
5777
  * An estimate of how far your job has progressed. This estimate is shown as a percentage of the total time from when your job leaves its queue to when your output files appear in your output Amazon S3 bucket. AWS Elemental MediaConvert provides jobPercentComplete in CloudWatch STATUS_UPDATE events and in the response to GetJob and ListJobs requests. The jobPercentComplete estimate is reliable for the following input containers: Quicktime, Transport Stream, MP4, and MXF. For some jobs, the service can't provide information about job progress. In those cases, jobPercentComplete returns a null value.
5778
+ * @public
5779
5779
  */
5780
5780
  JobPercentComplete?: number;
5781
5781
  /**
5782
- * @public
5783
5782
  * The job template that the job is created from, if it is created from a job template.
5783
+ * @public
5784
5784
  */
5785
5785
  JobTemplate?: string;
5786
5786
  /**
5787
- * @public
5788
5787
  * Provides messages from the service about jobs that you have already successfully submitted.
5788
+ * @public
5789
5789
  */
5790
5790
  Messages?: JobMessages;
5791
5791
  /**
5792
- * @public
5793
5792
  * List of output group details
5793
+ * @public
5794
5794
  */
5795
5795
  OutputGroupDetails?: OutputGroupDetail[];
5796
5796
  /**
5797
- * @public
5798
5797
  * Relative priority on the job.
5798
+ * @public
5799
5799
  */
5800
5800
  Priority?: number;
5801
5801
  /**
5802
- * @public
5803
5802
  * When you create a job, you can specify a queue to send it to. If you don't specify, the job will go to the default queue. For more about queues, see the User Guide topic at https://docs.aws.amazon.com/mediaconvert/latest/ug/what-is.html
5803
+ * @public
5804
5804
  */
5805
5805
  Queue?: string;
5806
5806
  /**
5807
- * @public
5808
5807
  * The job's queue hopping history.
5808
+ * @public
5809
5809
  */
5810
5810
  QueueTransitions?: QueueTransition[];
5811
5811
  /**
5812
- * @public
5813
5812
  * The number of times that the service automatically attempted to process your job after encountering an error.
5813
+ * @public
5814
5814
  */
5815
5815
  RetryCount?: number;
5816
5816
  /**
5817
- * @public
5818
5817
  * The IAM role you use for creating this job. For details about permissions, see the User Guide topic at the User Guide at https://docs.aws.amazon.com/mediaconvert/latest/ug/iam-role.html
5818
+ * @public
5819
5819
  */
5820
5820
  Role: string | undefined;
5821
5821
  /**
5822
- * @public
5823
5822
  * JobSettings contains all the transcode settings for a job.
5823
+ * @public
5824
5824
  */
5825
5825
  Settings: JobSettings | undefined;
5826
5826
  /**
5827
- * @public
5828
5827
  * Enable this setting when you run a test job to estimate how many reserved transcoding slots (RTS) you need. When this is enabled, MediaConvert runs your job from an on-demand queue with similar performance to what you will see with one RTS in a reserved queue. This setting is disabled by default.
5828
+ * @public
5829
5829
  */
5830
5830
  SimulateReservedQueue?: SimulateReservedQueue;
5831
5831
  /**
5832
- * @public
5833
5832
  * A job's status can be SUBMITTED, PROGRESSING, COMPLETE, CANCELED, or ERROR.
5833
+ * @public
5834
5834
  */
5835
5835
  Status?: JobStatus;
5836
5836
  /**
5837
- * @public
5838
5837
  * Specify how often MediaConvert sends STATUS_UPDATE events to Amazon CloudWatch Events. Set the interval, in seconds, between status updates. MediaConvert sends an update at this interval from the time the service begins processing your job to the time it completes the transcode or encounters an error.
5838
+ * @public
5839
5839
  */
5840
5840
  StatusUpdateInterval?: StatusUpdateInterval;
5841
5841
  /**
5842
- * @public
5843
5842
  * Information about when jobs are submitted, started, and finished is specified in Unix epoch format in seconds.
5843
+ * @public
5844
5844
  */
5845
5845
  Timing?: Timing;
5846
5846
  /**
5847
- * @public
5848
5847
  * User-defined metadata that you want to associate with an MediaConvert job. You specify metadata in key/value pairs.
5848
+ * @public
5849
5849
  */
5850
5850
  UserMetadata?: Record<string, string>;
5851
5851
  /**
5852
- * @public
5853
5852
  * Contains any warning messages for the job. Use to help identify potential issues with your input, output, or job. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/warning_codes.html
5853
+ * @public
5854
5854
  */
5855
5855
  Warnings?: WarningGroup[];
5856
5856
  }
5857
5857
  /**
5858
- * @public
5859
5858
  * JobTemplateSettings contains all the transcode settings saved in the template that will be applied to jobs created from it.
5859
+ * @public
5860
5860
  */
5861
5861
  export interface JobTemplateSettings {
5862
5862
  /**
5863
- * @public
5864
5863
  * When specified, this offset (in milliseconds) is added to the input Ad Avail PTS time.
5864
+ * @public
5865
5865
  */
5866
5866
  AdAvailOffset?: number;
5867
5867
  /**
5868
- * @public
5869
5868
  * Settings for ad avail blanking. Video can be blanked or overlaid with an image, and audio muted during SCTE-35 triggered ad avails.
5869
+ * @public
5870
5870
  */
5871
5871
  AvailBlanking?: AvailBlanking;
5872
5872
  /**
5873
- * @public
5874
5873
  * Use 3D LUTs to specify custom color mapping behavior when you convert from one color space into another. You can include up to 8 different 3D LUTs.
5874
+ * @public
5875
5875
  */
5876
5876
  ColorConversion3DLUTSettings?: ColorConversion3DLUTSetting[];
5877
5877
  /**
5878
- * @public
5879
5878
  * Settings for Event Signaling And Messaging (ESAM). If you don't do ad insertion, you can ignore these settings.
5879
+ * @public
5880
5880
  */
5881
5881
  Esam?: EsamSettings;
5882
5882
  /**
5883
- * @public
5884
5883
  * If your source content has EIA-608 Line 21 Data Services, enable this feature to specify what MediaConvert does with the Extended Data Services (XDS) packets. You can choose to pass through XDS packets, or remove them from the output. For more information about XDS, see EIA-608 Line Data Services, section 9.5.1.5 05h Content Advisory.
5884
+ * @public
5885
5885
  */
5886
5886
  ExtendedDataServices?: ExtendedDataServices;
5887
5887
  /**
5888
- * @public
5889
5888
  * Specify the input that MediaConvert references for your default output settings. MediaConvert uses this input's Resolution, Frame rate, and Pixel aspect ratio for all outputs that you don't manually specify different output settings for. Enabling this setting will disable "Follow source" for all other inputs. If MediaConvert cannot follow your source, for example if you specify an audio-only input, MediaConvert uses the first followable input instead. In your JSON job specification, enter an integer from 1 to 150 corresponding to the order of your inputs.
5889
+ * @public
5890
5890
  */
5891
5891
  FollowSource?: number;
5892
5892
  /**
5893
- * @public
5894
5893
  * Use Inputs to define the source file used in the transcode job. There can only be one input in a job template. Using the API, you can include multiple inputs when referencing a job template.
5894
+ * @public
5895
5895
  */
5896
5896
  Inputs?: InputTemplate[];
5897
5897
  /**
5898
- * @public
5899
5898
  * Use these settings only when you use Kantar watermarking. Specify the values that MediaConvert uses to generate and place Kantar watermarks in your output audio. These settings apply to every output in your job. In addition to specifying these values, you also need to store your Kantar credentials in AWS Secrets Manager. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/kantar-watermarking.html.
5899
+ * @public
5900
5900
  */
5901
5901
  KantarWatermark?: KantarWatermarkSettings;
5902
5902
  /**
5903
- * @public
5904
5903
  * Overlay motion graphics on top of your video. The motion graphics that you specify here appear on all outputs in all output groups. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/motion-graphic-overlay.html.
5904
+ * @public
5905
5905
  */
5906
5906
  MotionImageInserter?: MotionImageInserter;
5907
5907
  /**
5908
- * @public
5909
5908
  * Settings for your Nielsen configuration. If you don't do Nielsen measurement and analytics, ignore these settings. When you enable Nielsen configuration, MediaConvert enables PCM to ID3 tagging for all outputs in the job.
5909
+ * @public
5910
5910
  */
5911
5911
  NielsenConfiguration?: NielsenConfiguration;
5912
5912
  /**
5913
- * @public
5914
5913
  * Ignore these settings unless you are using Nielsen non-linear watermarking. Specify the values that MediaConvert uses to generate and place Nielsen watermarks in your output audio. In addition to specifying these values, you also need to set up your cloud TIC server. These settings apply to every output in your job. The MediaConvert implementation is currently with the following Nielsen versions: Nielsen Watermark SDK Version 5.2.1 Nielsen NLM Watermark Engine Version 1.2.7 Nielsen Watermark Authenticator [SID_TIC] Version [5.0.0]
5914
+ * @public
5915
5915
  */
5916
5916
  NielsenNonLinearWatermark?: NielsenNonLinearWatermarkSettings;
5917
5917
  /**
5918
- * @public
5919
5918
  * Contains one group of settings for each set of outputs that share a common package type. All unpackaged files (MPEG-4, MPEG-2 TS, Quicktime, MXF, and no container) are grouped in a single output group as well. Required in is a group of settings that apply to the whole group. This required object depends on the value you set for Type. Type, settings object pairs are as follows. * FILE_GROUP_SETTINGS, FileGroupSettings * HLS_GROUP_SETTINGS, HlsGroupSettings * DASH_ISO_GROUP_SETTINGS, DashIsoGroupSettings * MS_SMOOTH_GROUP_SETTINGS, MsSmoothGroupSettings * CMAF_GROUP_SETTINGS, CmafGroupSettings
5919
+ * @public
5920
5920
  */
5921
5921
  OutputGroups?: OutputGroup[];
5922
5922
  /**
5923
- * @public
5924
5923
  * These settings control how the service handles timecodes throughout the job. These settings don't affect input clipping.
5924
+ * @public
5925
5925
  */
5926
5926
  TimecodeConfig?: TimecodeConfig;
5927
5927
  /**
5928
- * @public
5929
5928
  * Insert user-defined custom ID3 metadata at timecodes that you specify. In each output that you want to include this metadata, you must set ID3 metadata to Passthrough.
5929
+ * @public
5930
5930
  */
5931
5931
  TimedMetadataInsertion?: TimedMetadataInsertion;
5932
5932
  }
@@ -5943,145 +5943,145 @@ export declare const Type: {
5943
5943
  */
5944
5944
  export type Type = (typeof Type)[keyof typeof Type];
5945
5945
  /**
5946
- * @public
5947
5946
  * A job template is a pre-made set of encoding instructions that you can use to quickly create a job.
5947
+ * @public
5948
5948
  */
5949
5949
  export interface JobTemplate {
5950
5950
  /**
5951
- * @public
5952
5951
  * Accelerated transcoding can significantly speed up jobs with long, visually complex content.
5952
+ * @public
5953
5953
  */
5954
5954
  AccelerationSettings?: AccelerationSettings;
5955
5955
  /**
5956
- * @public
5957
5956
  * An identifier for this resource that is unique within all of AWS.
5957
+ * @public
5958
5958
  */
5959
5959
  Arn?: string;
5960
5960
  /**
5961
- * @public
5962
5961
  * An optional category you create to organize your job templates.
5962
+ * @public
5963
5963
  */
5964
5964
  Category?: string;
5965
5965
  /**
5966
- * @public
5967
5966
  * The timestamp in epoch seconds for Job template creation.
5967
+ * @public
5968
5968
  */
5969
5969
  CreatedAt?: Date;
5970
5970
  /**
5971
- * @public
5972
5971
  * An optional description you create for each job template.
5972
+ * @public
5973
5973
  */
5974
5974
  Description?: string;
5975
5975
  /**
5976
- * @public
5977
5976
  * Optional list of hop destinations.
5977
+ * @public
5978
5978
  */
5979
5979
  HopDestinations?: HopDestination[];
5980
5980
  /**
5981
- * @public
5982
5981
  * The timestamp in epoch seconds when the Job template was last updated.
5982
+ * @public
5983
5983
  */
5984
5984
  LastUpdated?: Date;
5985
5985
  /**
5986
- * @public
5987
5986
  * A name you create for each job template. Each name must be unique within your account.
5987
+ * @public
5988
5988
  */
5989
5989
  Name: string | undefined;
5990
5990
  /**
5991
- * @public
5992
5991
  * Relative priority on the job.
5992
+ * @public
5993
5993
  */
5994
5994
  Priority?: number;
5995
5995
  /**
5996
- * @public
5997
5996
  * Optional. The queue that jobs created from this template are assigned to. If you don't specify this, jobs will go to the default queue.
5997
+ * @public
5998
5998
  */
5999
5999
  Queue?: string;
6000
6000
  /**
6001
- * @public
6002
6001
  * JobTemplateSettings contains all the transcode settings saved in the template that will be applied to jobs created from it.
6002
+ * @public
6003
6003
  */
6004
6004
  Settings: JobTemplateSettings | undefined;
6005
6005
  /**
6006
- * @public
6007
6006
  * Specify how often MediaConvert sends STATUS_UPDATE events to Amazon CloudWatch Events. Set the interval, in seconds, between status updates. MediaConvert sends an update at this interval from the time the service begins processing your job to the time it completes the transcode or encounters an error.
6007
+ * @public
6008
6008
  */
6009
6009
  StatusUpdateInterval?: StatusUpdateInterval;
6010
6010
  /**
6011
- * @public
6012
6011
  * A job template can be of two types: system or custom. System or built-in job templates can't be modified or deleted by the user.
6012
+ * @public
6013
6013
  */
6014
6014
  Type?: Type;
6015
6015
  }
6016
6016
  /**
6017
- * @public
6018
6017
  * Settings for preset
6018
+ * @public
6019
6019
  */
6020
6020
  export interface PresetSettings {
6021
6021
  /**
6022
- * @public
6023
6022
  * Contains groups of audio encoding settings organized by audio codec. Include one instance of per output. Can contain multiple groups of encoding settings.
6023
+ * @public
6024
6024
  */
6025
6025
  AudioDescriptions?: AudioDescription[];
6026
6026
  /**
6027
- * @public
6028
6027
  * This object holds groups of settings related to captions for one output. For each output that has captions, include one instance of CaptionDescriptions.
6028
+ * @public
6029
6029
  */
6030
6030
  CaptionDescriptions?: CaptionDescriptionPreset[];
6031
6031
  /**
6032
- * @public
6033
6032
  * Container specific settings.
6033
+ * @public
6034
6034
  */
6035
6035
  ContainerSettings?: ContainerSettings;
6036
6036
  /**
6037
- * @public
6038
6037
  * VideoDescription contains a group of video encoding settings. The specific video settings depend on the video codec that you choose for the property codec. Include one instance of VideoDescription per output.
6038
+ * @public
6039
6039
  */
6040
6040
  VideoDescription?: VideoDescription;
6041
6041
  }
6042
6042
  /**
6043
- * @public
6044
6043
  * A preset is a collection of preconfigured media conversion settings that you want MediaConvert to apply to the output during the conversion process.
6044
+ * @public
6045
6045
  */
6046
6046
  export interface Preset {
6047
6047
  /**
6048
- * @public
6049
6048
  * An identifier for this resource that is unique within all of AWS.
6049
+ * @public
6050
6050
  */
6051
6051
  Arn?: string;
6052
6052
  /**
6053
- * @public
6054
6053
  * An optional category you create to organize your presets.
6054
+ * @public
6055
6055
  */
6056
6056
  Category?: string;
6057
6057
  /**
6058
- * @public
6059
6058
  * The timestamp in epoch seconds for preset creation.
6059
+ * @public
6060
6060
  */
6061
6061
  CreatedAt?: Date;
6062
6062
  /**
6063
- * @public
6064
6063
  * An optional description you create for each preset.
6064
+ * @public
6065
6065
  */
6066
6066
  Description?: string;
6067
6067
  /**
6068
- * @public
6069
6068
  * The timestamp in epoch seconds when the preset was last updated.
6069
+ * @public
6070
6070
  */
6071
6071
  LastUpdated?: Date;
6072
6072
  /**
6073
- * @public
6074
6073
  * A name you create for each preset. Each name must be unique within your account.
6074
+ * @public
6075
6075
  */
6076
6076
  Name: string | undefined;
6077
6077
  /**
6078
- * @public
6079
6078
  * Settings for preset
6079
+ * @public
6080
6080
  */
6081
6081
  Settings: PresetSettings | undefined;
6082
6082
  /**
6083
- * @public
6084
6083
  * A preset can be of two types: system or custom. System or built-in preset can't be modified or deleted by the user.
6084
+ * @public
6085
6085
  */
6086
6086
  Type?: Type;
6087
6087
  }
@@ -6133,38 +6133,38 @@ export declare const ReservationPlanStatus: {
6133
6133
  */
6134
6134
  export type ReservationPlanStatus = (typeof ReservationPlanStatus)[keyof typeof ReservationPlanStatus];
6135
6135
  /**
6136
- * @public
6137
6136
  * Details about the pricing plan for your reserved queue. Required for reserved queues and not applicable to on-demand queues.
6137
+ * @public
6138
6138
  */
6139
6139
  export interface ReservationPlan {
6140
6140
  /**
6141
- * @public
6142
6141
  * The length of the term of your reserved queue pricing plan commitment.
6142
+ * @public
6143
6143
  */
6144
6144
  Commitment?: Commitment;
6145
6145
  /**
6146
- * @public
6147
6146
  * The timestamp in epoch seconds for when the current pricing plan term for this reserved queue expires.
6147
+ * @public
6148
6148
  */
6149
6149
  ExpiresAt?: Date;
6150
6150
  /**
6151
- * @public
6152
6151
  * The timestamp in epoch seconds for when you set up the current pricing plan for this reserved queue.
6152
+ * @public
6153
6153
  */
6154
6154
  PurchasedAt?: Date;
6155
6155
  /**
6156
- * @public
6157
6156
  * Specifies whether the term of your reserved queue pricing plan is automatically extended (AUTO_RENEW) or expires (EXPIRE) at the end of the term.
6157
+ * @public
6158
6158
  */
6159
6159
  RenewalType?: RenewalType;
6160
6160
  /**
6161
- * @public
6162
6161
  * Specifies the number of reserved transcode slots (RTS) for this queue. The number of RTS determines how many jobs the queue can process in parallel; each RTS can process one job at a time. When you increase this number, you extend your existing commitment with a new 12-month commitment for a larger number of RTS. The new commitment begins when you purchase the additional capacity. You can't decrease the number of RTS in your reserved queue.
6162
+ * @public
6163
6163
  */
6164
6164
  ReservedSlots?: number;
6165
6165
  /**
6166
- * @public
6167
6166
  * Specifies whether the pricing plan for your reserved queue is ACTIVE or EXPIRED.
6167
+ * @public
6168
6168
  */
6169
6169
  Status?: ReservationPlanStatus;
6170
6170
  }
@@ -6181,63 +6181,63 @@ export declare const QueueStatus: {
6181
6181
  */
6182
6182
  export type QueueStatus = (typeof QueueStatus)[keyof typeof QueueStatus];
6183
6183
  /**
6184
- * @public
6185
6184
  * You can use queues to manage the resources that are available to your AWS account for running multiple transcoding jobs at the same time. If you don't specify a queue, the service sends all jobs through the default queue. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/working-with-queues.html.
6185
+ * @public
6186
6186
  */
6187
6187
  export interface Queue {
6188
6188
  /**
6189
- * @public
6190
6189
  * An identifier for this resource that is unique within all of AWS.
6190
+ * @public
6191
6191
  */
6192
6192
  Arn?: string;
6193
6193
  /**
6194
- * @public
6195
6194
  * The timestamp in epoch seconds for when you created the queue.
6195
+ * @public
6196
6196
  */
6197
6197
  CreatedAt?: Date;
6198
6198
  /**
6199
- * @public
6200
6199
  * An optional description that you create for each queue.
6200
+ * @public
6201
6201
  */
6202
6202
  Description?: string;
6203
6203
  /**
6204
- * @public
6205
6204
  * The timestamp in epoch seconds for when you most recently updated the queue.
6205
+ * @public
6206
6206
  */
6207
6207
  LastUpdated?: Date;
6208
6208
  /**
6209
- * @public
6210
6209
  * A name that you create for each queue. Each name must be unique within your account.
6210
+ * @public
6211
6211
  */
6212
6212
  Name: string | undefined;
6213
6213
  /**
6214
- * @public
6215
6214
  * Specifies whether the pricing plan for the queue is on-demand or reserved. For on-demand, you pay per minute, billed in increments of .01 minute. For reserved, you pay for the transcoding capacity of the entire queue, regardless of how much or how little you use it. Reserved pricing requires a 12-month commitment.
6215
+ * @public
6216
6216
  */
6217
6217
  PricingPlan?: PricingPlan;
6218
6218
  /**
6219
- * @public
6220
6219
  * The estimated number of jobs with a PROGRESSING status.
6220
+ * @public
6221
6221
  */
6222
6222
  ProgressingJobsCount?: number;
6223
6223
  /**
6224
- * @public
6225
6224
  * Details about the pricing plan for your reserved queue. Required for reserved queues and not applicable to on-demand queues.
6225
+ * @public
6226
6226
  */
6227
6227
  ReservationPlan?: ReservationPlan;
6228
6228
  /**
6229
- * @public
6230
6229
  * Queues can be ACTIVE or PAUSED. If you pause a queue, the service won't begin processing jobs in that queue. Jobs that are running when you pause the queue continue to run until they finish or result in an error.
6230
+ * @public
6231
6231
  */
6232
6232
  Status?: QueueStatus;
6233
6233
  /**
6234
- * @public
6235
6234
  * The estimated number of jobs with a SUBMITTED status.
6235
+ * @public
6236
6236
  */
6237
6237
  SubmittedJobsCount?: number;
6238
6238
  /**
6239
- * @public
6240
6239
  * Specifies whether this on-demand queue is system or custom. System queues are built in. You can't modify or delete system queues. You can create and modify custom queues.
6240
+ * @public
6241
6241
  */
6242
6242
  Type?: Type;
6243
6243
  }
@@ -6246,8 +6246,8 @@ export interface Queue {
6246
6246
  */
6247
6247
  export interface AssociateCertificateRequest {
6248
6248
  /**
6249
- * @public
6250
6249
  * The ARN of the ACM certificate that you want to associate with your MediaConvert resource.
6250
+ * @public
6251
6251
  */
6252
6252
  Arn: string | undefined;
6253
6253
  }
@@ -6257,8 +6257,8 @@ export interface AssociateCertificateRequest {
6257
6257
  export interface AssociateCertificateResponse {
6258
6258
  }
6259
6259
  /**
6260
- * @public
6261
6260
  * The service can't process your request because of a problem in the request. Please check your request form and syntax.
6261
+ * @public
6262
6262
  */
6263
6263
  export declare class BadRequestException extends __BaseException {
6264
6264
  readonly name: "BadRequestException";
@@ -6270,8 +6270,8 @@ export declare class BadRequestException extends __BaseException {
6270
6270
  constructor(opts: __ExceptionOptionType<BadRequestException, __BaseException>);
6271
6271
  }
6272
6272
  /**
6273
- * @public
6274
6273
  * The service couldn't complete your request because there is a conflict with the current state of the resource.
6274
+ * @public
6275
6275
  */
6276
6276
  export declare class ConflictException extends __BaseException {
6277
6277
  readonly name: "ConflictException";
@@ -6283,8 +6283,8 @@ export declare class ConflictException extends __BaseException {
6283
6283
  constructor(opts: __ExceptionOptionType<ConflictException, __BaseException>);
6284
6284
  }
6285
6285
  /**
6286
- * @public
6287
6286
  * You don't have permissions for this action with the credentials you sent.
6287
+ * @public
6288
6288
  */
6289
6289
  export declare class ForbiddenException extends __BaseException {
6290
6290
  readonly name: "ForbiddenException";
@@ -6296,8 +6296,8 @@ export declare class ForbiddenException extends __BaseException {
6296
6296
  constructor(opts: __ExceptionOptionType<ForbiddenException, __BaseException>);
6297
6297
  }
6298
6298
  /**
6299
- * @public
6300
6299
  * The service encountered an unexpected condition and can't fulfill your request.
6300
+ * @public
6301
6301
  */
6302
6302
  export declare class InternalServerErrorException extends __BaseException {
6303
6303
  readonly name: "InternalServerErrorException";
@@ -6309,8 +6309,8 @@ export declare class InternalServerErrorException extends __BaseException {
6309
6309
  constructor(opts: __ExceptionOptionType<InternalServerErrorException, __BaseException>);
6310
6310
  }
6311
6311
  /**
6312
- * @public
6313
6312
  * The resource you requested doesn't exist.
6313
+ * @public
6314
6314
  */
6315
6315
  export declare class NotFoundException extends __BaseException {
6316
6316
  readonly name: "NotFoundException";
@@ -6322,8 +6322,8 @@ export declare class NotFoundException extends __BaseException {
6322
6322
  constructor(opts: __ExceptionOptionType<NotFoundException, __BaseException>);
6323
6323
  }
6324
6324
  /**
6325
- * @public
6326
6325
  * Too many requests have been sent in too short of a time. The service limits the rate at which it will accept requests.
6326
+ * @public
6327
6327
  */
6328
6328
  export declare class TooManyRequestsException extends __BaseException {
6329
6329
  readonly name: "TooManyRequestsException";