@aws-sdk/client-mediaconvert 3.808.0 → 3.810.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist-cjs/index.js CHANGED
@@ -90,6 +90,7 @@ __export(index_exports, {
90
90
  CaptionSourceByteRateLimit: () => CaptionSourceByteRateLimit,
91
91
  CaptionSourceConvertPaintOnToPopOn: () => CaptionSourceConvertPaintOnToPopOn,
92
92
  CaptionSourceType: () => CaptionSourceType,
93
+ CaptionSourceUpconvertSTLToTeletext: () => CaptionSourceUpconvertSTLToTeletext,
93
94
  ChromaPositionMode: () => ChromaPositionMode,
94
95
  CmafClientCache: () => CmafClientCache,
95
96
  CmafCodecSpecification: () => CmafCodecSpecification,
@@ -2761,7 +2762,8 @@ var se_FileSourceSettings = /* @__PURE__ */ __name((input, context) => {
2761
2762
  framerate: [, (_) => se_CaptionSourceFramerate(_, context), `Framerate`],
2762
2763
  sourceFile: [, , `SourceFile`],
2763
2764
  timeDelta: [, , `TimeDelta`],
2764
- timeDeltaUnits: [, , `TimeDeltaUnits`]
2765
+ timeDeltaUnits: [, , `TimeDeltaUnits`],
2766
+ upconvertSTLToTeletext: [, , `UpconvertSTLToTeletext`]
2765
2767
  });
2766
2768
  }, "se_FileSourceSettings");
2767
2769
  var se_FlacSettings = /* @__PURE__ */ __name((input, context) => {
@@ -3784,6 +3786,7 @@ var se_VideoDescription = /* @__PURE__ */ __name((input, context) => {
3784
3786
  }, "se_VideoDescription");
3785
3787
  var se_VideoOverlay = /* @__PURE__ */ __name((input, context) => {
3786
3788
  return (0, import_smithy_client.take)(input, {
3789
+ crop: [, (_) => se_VideoOverlayCrop(_, context), `Crop`],
3787
3790
  endTimecode: [, , `EndTimecode`],
3788
3791
  initialPosition: [, (_) => se_VideoOverlayPosition(_, context), `InitialPosition`],
3789
3792
  input: [, (_) => se_VideoOverlayInput(_, context), `Input`],
@@ -3792,6 +3795,15 @@ var se_VideoOverlay = /* @__PURE__ */ __name((input, context) => {
3792
3795
  transitions: [, (_) => se___listOfVideoOverlayTransition(_, context), `Transitions`]
3793
3796
  });
3794
3797
  }, "se_VideoOverlay");
3798
+ var se_VideoOverlayCrop = /* @__PURE__ */ __name((input, context) => {
3799
+ return (0, import_smithy_client.take)(input, {
3800
+ height: [, , `Height`],
3801
+ unit: [, , `Unit`],
3802
+ width: [, , `Width`],
3803
+ x: [, , `X`],
3804
+ y: [, , `Y`]
3805
+ });
3806
+ }, "se_VideoOverlayCrop");
3795
3807
  var se_VideoOverlayInput = /* @__PURE__ */ __name((input, context) => {
3796
3808
  return (0, import_smithy_client.take)(input, {
3797
3809
  fileInput: [, , `FileInput`],
@@ -4979,7 +4991,8 @@ var de_FileSourceSettings = /* @__PURE__ */ __name((output, context) => {
4979
4991
  Framerate: [, (_) => de_CaptionSourceFramerate(_, context), `framerate`],
4980
4992
  SourceFile: [, import_smithy_client.expectString, `sourceFile`],
4981
4993
  TimeDelta: [, import_smithy_client.expectInt32, `timeDelta`],
4982
- TimeDeltaUnits: [, import_smithy_client.expectString, `timeDeltaUnits`]
4994
+ TimeDeltaUnits: [, import_smithy_client.expectString, `timeDeltaUnits`],
4995
+ UpconvertSTLToTeletext: [, import_smithy_client.expectString, `upconvertSTLToTeletext`]
4983
4996
  });
4984
4997
  }, "de_FileSourceSettings");
4985
4998
  var de_FlacSettings = /* @__PURE__ */ __name((output, context) => {
@@ -6195,6 +6208,7 @@ var de_VideoDetail = /* @__PURE__ */ __name((output, context) => {
6195
6208
  }, "de_VideoDetail");
6196
6209
  var de_VideoOverlay = /* @__PURE__ */ __name((output, context) => {
6197
6210
  return (0, import_smithy_client.take)(output, {
6211
+ Crop: [, (_) => de_VideoOverlayCrop(_, context), `crop`],
6198
6212
  EndTimecode: [, import_smithy_client.expectString, `endTimecode`],
6199
6213
  InitialPosition: [, (_) => de_VideoOverlayPosition(_, context), `initialPosition`],
6200
6214
  Input: [, (_) => de_VideoOverlayInput(_, context), `input`],
@@ -6203,6 +6217,15 @@ var de_VideoOverlay = /* @__PURE__ */ __name((output, context) => {
6203
6217
  Transitions: [, (_) => de___listOfVideoOverlayTransition(_, context), `transitions`]
6204
6218
  });
6205
6219
  }, "de_VideoOverlay");
6220
+ var de_VideoOverlayCrop = /* @__PURE__ */ __name((output, context) => {
6221
+ return (0, import_smithy_client.take)(output, {
6222
+ Height: [, import_smithy_client.expectInt32, `height`],
6223
+ Unit: [, import_smithy_client.expectString, `unit`],
6224
+ Width: [, import_smithy_client.expectInt32, `width`],
6225
+ X: [, import_smithy_client.expectInt32, `x`],
6226
+ Y: [, import_smithy_client.expectInt32, `y`]
6227
+ });
6228
+ }, "de_VideoOverlayCrop");
6206
6229
  var de_VideoOverlayInput = /* @__PURE__ */ __name((output, context) => {
6207
6230
  return (0, import_smithy_client.take)(output, {
6208
6231
  FileInput: [, import_smithy_client.expectString, `fileInput`],
@@ -7725,6 +7748,10 @@ var FileSourceTimeDeltaUnits = {
7725
7748
  MILLISECONDS: "MILLISECONDS",
7726
7749
  SECONDS: "SECONDS"
7727
7750
  };
7751
+ var CaptionSourceUpconvertSTLToTeletext = {
7752
+ DISABLED: "DISABLED",
7753
+ UPCONVERT: "UPCONVERT"
7754
+ };
7728
7755
  var CaptionSourceType = {
7729
7756
  ANCILLARY: "ANCILLARY",
7730
7757
  DVB_SUB: "DVB_SUB",
@@ -8133,12 +8160,12 @@ var HlsTimedMetadataId3Frame = {
8133
8160
  PRIV: "PRIV",
8134
8161
  TDRL: "TDRL"
8135
8162
  };
8163
+
8164
+ // src/models/models_1.ts
8136
8165
  var MsSmoothAudioDeduplication = {
8137
8166
  COMBINE_DUPLICATE_STREAMS: "COMBINE_DUPLICATE_STREAMS",
8138
8167
  NONE: "NONE"
8139
8168
  };
8140
-
8141
- // src/models/models_1.ts
8142
8169
  var MsSmoothFragmentLengthControl = {
8143
8170
  EXACT: "EXACT",
8144
8171
  GOP_MULTIPLE: "GOP_MULTIPLE"
@@ -9460,6 +9487,7 @@ var StatusUpdateInterval = {
9460
9487
  FileSourceConvert608To708,
9461
9488
  CaptionSourceConvertPaintOnToPopOn,
9462
9489
  FileSourceTimeDeltaUnits,
9490
+ CaptionSourceUpconvertSTLToTeletext,
9463
9491
  CaptionSourceType,
9464
9492
  InputDeblockFilter,
9465
9493
  DecryptionMode,
@@ -755,6 +755,10 @@ export const FileSourceTimeDeltaUnits = {
755
755
  MILLISECONDS: "MILLISECONDS",
756
756
  SECONDS: "SECONDS",
757
757
  };
758
+ export const CaptionSourceUpconvertSTLToTeletext = {
759
+ DISABLED: "DISABLED",
760
+ UPCONVERT: "UPCONVERT",
761
+ };
758
762
  export const CaptionSourceType = {
759
763
  ANCILLARY: "ANCILLARY",
760
764
  DVB_SUB: "DVB_SUB",
@@ -1163,7 +1167,3 @@ export const HlsTimedMetadataId3Frame = {
1163
1167
  PRIV: "PRIV",
1164
1168
  TDRL: "TDRL",
1165
1169
  };
1166
- export const MsSmoothAudioDeduplication = {
1167
- COMBINE_DUPLICATE_STREAMS: "COMBINE_DUPLICATE_STREAMS",
1168
- NONE: "NONE",
1169
- };
@@ -1,3 +1,7 @@
1
+ export const MsSmoothAudioDeduplication = {
2
+ COMBINE_DUPLICATE_STREAMS: "COMBINE_DUPLICATE_STREAMS",
3
+ NONE: "NONE",
4
+ };
1
5
  export const MsSmoothFragmentLengthControl = {
2
6
  EXACT: "EXACT",
3
7
  GOP_MULTIPLE: "GOP_MULTIPLE",
@@ -1842,6 +1842,7 @@ const se_FileSourceSettings = (input, context) => {
1842
1842
  sourceFile: [, , `SourceFile`],
1843
1843
  timeDelta: [, , `TimeDelta`],
1844
1844
  timeDeltaUnits: [, , `TimeDeltaUnits`],
1845
+ upconvertSTLToTeletext: [, , `UpconvertSTLToTeletext`],
1845
1846
  });
1846
1847
  };
1847
1848
  const se_FlacSettings = (input, context) => {
@@ -2864,6 +2865,7 @@ const se_VideoDescription = (input, context) => {
2864
2865
  };
2865
2866
  const se_VideoOverlay = (input, context) => {
2866
2867
  return take(input, {
2868
+ crop: [, (_) => se_VideoOverlayCrop(_, context), `Crop`],
2867
2869
  endTimecode: [, , `EndTimecode`],
2868
2870
  initialPosition: [, (_) => se_VideoOverlayPosition(_, context), `InitialPosition`],
2869
2871
  input: [, (_) => se_VideoOverlayInput(_, context), `Input`],
@@ -2872,6 +2874,15 @@ const se_VideoOverlay = (input, context) => {
2872
2874
  transitions: [, (_) => se___listOfVideoOverlayTransition(_, context), `Transitions`],
2873
2875
  });
2874
2876
  };
2877
+ const se_VideoOverlayCrop = (input, context) => {
2878
+ return take(input, {
2879
+ height: [, , `Height`],
2880
+ unit: [, , `Unit`],
2881
+ width: [, , `Width`],
2882
+ x: [, , `X`],
2883
+ y: [, , `Y`],
2884
+ });
2885
+ };
2875
2886
  const se_VideoOverlayInput = (input, context) => {
2876
2887
  return take(input, {
2877
2888
  fileInput: [, , `FileInput`],
@@ -4138,6 +4149,7 @@ const de_FileSourceSettings = (output, context) => {
4138
4149
  SourceFile: [, __expectString, `sourceFile`],
4139
4150
  TimeDelta: [, __expectInt32, `timeDelta`],
4140
4151
  TimeDeltaUnits: [, __expectString, `timeDeltaUnits`],
4152
+ UpconvertSTLToTeletext: [, __expectString, `upconvertSTLToTeletext`],
4141
4153
  });
4142
4154
  };
4143
4155
  const de_FlacSettings = (output, context) => {
@@ -5353,6 +5365,7 @@ const de_VideoDetail = (output, context) => {
5353
5365
  };
5354
5366
  const de_VideoOverlay = (output, context) => {
5355
5367
  return take(output, {
5368
+ Crop: [, (_) => de_VideoOverlayCrop(_, context), `crop`],
5356
5369
  EndTimecode: [, __expectString, `endTimecode`],
5357
5370
  InitialPosition: [, (_) => de_VideoOverlayPosition(_, context), `initialPosition`],
5358
5371
  Input: [, (_) => de_VideoOverlayInput(_, context), `input`],
@@ -5361,6 +5374,15 @@ const de_VideoOverlay = (output, context) => {
5361
5374
  Transitions: [, (_) => de___listOfVideoOverlayTransition(_, context), `transitions`],
5362
5375
  });
5363
5376
  };
5377
+ const de_VideoOverlayCrop = (output, context) => {
5378
+ return take(output, {
5379
+ Height: [, __expectInt32, `height`],
5380
+ Unit: [, __expectString, `unit`],
5381
+ Width: [, __expectInt32, `width`],
5382
+ X: [, __expectInt32, `x`],
5383
+ Y: [, __expectInt32, `y`],
5384
+ });
5385
+ };
5364
5386
  const de_VideoOverlayInput = (output, context) => {
5365
5387
  return take(output, {
5366
5388
  FileInput: [, __expectString, `fileInput`],
@@ -165,6 +165,7 @@ declare const CreateJobCommand_base: {
165
165
  * SourceFile: "STRING_VALUE",
166
166
  * TimeDelta: Number("int"),
167
167
  * TimeDeltaUnits: "SECONDS" || "MILLISECONDS",
168
+ * UpconvertSTLToTeletext: "UPCONVERT" || "DISABLED",
168
169
  * },
169
170
  * SourceType: "ANCILLARY" || "DVB_SUB" || "EMBEDDED" || "SCTE20" || "SCC" || "TTML" || "STL" || "SRT" || "SMI" || "SMPTE_TT" || "TELETEXT" || "NULL_SOURCE" || "IMSC" || "WEBVTT",
170
171
  * TeletextSourceSettings: { // TeletextSourceSettings
@@ -255,6 +256,13 @@ declare const CreateJobCommand_base: {
255
256
  * },
256
257
  * VideoOverlays: [ // __listOfVideoOverlay
257
258
  * { // VideoOverlay
259
+ * Crop: { // VideoOverlayCrop
260
+ * Height: Number("int"),
261
+ * Unit: "PIXELS" || "PERCENTAGE",
262
+ * Width: Number("int"),
263
+ * X: Number("int"),
264
+ * Y: Number("int"),
265
+ * },
258
266
  * EndTimecode: "STRING_VALUE",
259
267
  * InitialPosition: { // VideoOverlayPosition
260
268
  * Height: Number("int"),
@@ -1759,6 +1767,7 @@ declare const CreateJobCommand_base: {
1759
1767
  * // SourceFile: "STRING_VALUE",
1760
1768
  * // TimeDelta: Number("int"),
1761
1769
  * // TimeDeltaUnits: "SECONDS" || "MILLISECONDS",
1770
+ * // UpconvertSTLToTeletext: "UPCONVERT" || "DISABLED",
1762
1771
  * // },
1763
1772
  * // SourceType: "ANCILLARY" || "DVB_SUB" || "EMBEDDED" || "SCTE20" || "SCC" || "TTML" || "STL" || "SRT" || "SMI" || "SMPTE_TT" || "TELETEXT" || "NULL_SOURCE" || "IMSC" || "WEBVTT",
1764
1773
  * // TeletextSourceSettings: { // TeletextSourceSettings
@@ -1849,6 +1858,13 @@ declare const CreateJobCommand_base: {
1849
1858
  * // },
1850
1859
  * // VideoOverlays: [ // __listOfVideoOverlay
1851
1860
  * // { // VideoOverlay
1861
+ * // Crop: { // VideoOverlayCrop
1862
+ * // Height: Number("int"),
1863
+ * // Unit: "PIXELS" || "PERCENTAGE",
1864
+ * // Width: Number("int"),
1865
+ * // X: Number("int"),
1866
+ * // Y: Number("int"),
1867
+ * // },
1852
1868
  * // EndTimecode: "STRING_VALUE",
1853
1869
  * // InitialPosition: { // VideoOverlayPosition
1854
1870
  * // Height: Number("int"),
@@ -163,6 +163,7 @@ declare const CreateJobTemplateCommand_base: {
163
163
  * SourceFile: "STRING_VALUE",
164
164
  * TimeDelta: Number("int"),
165
165
  * TimeDeltaUnits: "SECONDS" || "MILLISECONDS",
166
+ * UpconvertSTLToTeletext: "UPCONVERT" || "DISABLED",
166
167
  * },
167
168
  * SourceType: "ANCILLARY" || "DVB_SUB" || "EMBEDDED" || "SCTE20" || "SCC" || "TTML" || "STL" || "SRT" || "SMI" || "SMPTE_TT" || "TELETEXT" || "NULL_SOURCE" || "IMSC" || "WEBVTT",
168
169
  * TeletextSourceSettings: { // TeletextSourceSettings
@@ -236,6 +237,13 @@ declare const CreateJobTemplateCommand_base: {
236
237
  * TimecodeStart: "STRING_VALUE",
237
238
  * VideoOverlays: [ // __listOfVideoOverlay
238
239
  * { // VideoOverlay
240
+ * Crop: { // VideoOverlayCrop
241
+ * Height: Number("int"),
242
+ * Unit: "PIXELS" || "PERCENTAGE",
243
+ * Width: Number("int"),
244
+ * X: Number("int"),
245
+ * Y: Number("int"),
246
+ * },
239
247
  * EndTimecode: "STRING_VALUE",
240
248
  * InitialPosition: { // VideoOverlayPosition
241
249
  * Height: Number("int"),
@@ -1699,6 +1707,7 @@ declare const CreateJobTemplateCommand_base: {
1699
1707
  * // SourceFile: "STRING_VALUE",
1700
1708
  * // TimeDelta: Number("int"),
1701
1709
  * // TimeDeltaUnits: "SECONDS" || "MILLISECONDS",
1710
+ * // UpconvertSTLToTeletext: "UPCONVERT" || "DISABLED",
1702
1711
  * // },
1703
1712
  * // SourceType: "ANCILLARY" || "DVB_SUB" || "EMBEDDED" || "SCTE20" || "SCC" || "TTML" || "STL" || "SRT" || "SMI" || "SMPTE_TT" || "TELETEXT" || "NULL_SOURCE" || "IMSC" || "WEBVTT",
1704
1713
  * // TeletextSourceSettings: { // TeletextSourceSettings
@@ -1772,6 +1781,13 @@ declare const CreateJobTemplateCommand_base: {
1772
1781
  * // TimecodeStart: "STRING_VALUE",
1773
1782
  * // VideoOverlays: [ // __listOfVideoOverlay
1774
1783
  * // { // VideoOverlay
1784
+ * // Crop: { // VideoOverlayCrop
1785
+ * // Height: Number("int"),
1786
+ * // Unit: "PIXELS" || "PERCENTAGE",
1787
+ * // Width: Number("int"),
1788
+ * // X: Number("int"),
1789
+ * // Y: Number("int"),
1790
+ * // },
1775
1791
  * // EndTimecode: "STRING_VALUE",
1776
1792
  * // InitialPosition: { // VideoOverlayPosition
1777
1793
  * // Height: Number("int"),
@@ -209,6 +209,7 @@ declare const GetJobCommand_base: {
209
209
  * // SourceFile: "STRING_VALUE",
210
210
  * // TimeDelta: Number("int"),
211
211
  * // TimeDeltaUnits: "SECONDS" || "MILLISECONDS",
212
+ * // UpconvertSTLToTeletext: "UPCONVERT" || "DISABLED",
212
213
  * // },
213
214
  * // SourceType: "ANCILLARY" || "DVB_SUB" || "EMBEDDED" || "SCTE20" || "SCC" || "TTML" || "STL" || "SRT" || "SMI" || "SMPTE_TT" || "TELETEXT" || "NULL_SOURCE" || "IMSC" || "WEBVTT",
214
215
  * // TeletextSourceSettings: { // TeletextSourceSettings
@@ -299,6 +300,13 @@ declare const GetJobCommand_base: {
299
300
  * // },
300
301
  * // VideoOverlays: [ // __listOfVideoOverlay
301
302
  * // { // VideoOverlay
303
+ * // Crop: { // VideoOverlayCrop
304
+ * // Height: Number("int"),
305
+ * // Unit: "PIXELS" || "PERCENTAGE",
306
+ * // Width: Number("int"),
307
+ * // X: Number("int"),
308
+ * // Y: Number("int"),
309
+ * // },
302
310
  * // EndTimecode: "STRING_VALUE",
303
311
  * // InitialPosition: { // VideoOverlayPosition
304
312
  * // Height: Number("int"),
@@ -172,6 +172,7 @@ declare const GetJobTemplateCommand_base: {
172
172
  * // SourceFile: "STRING_VALUE",
173
173
  * // TimeDelta: Number("int"),
174
174
  * // TimeDeltaUnits: "SECONDS" || "MILLISECONDS",
175
+ * // UpconvertSTLToTeletext: "UPCONVERT" || "DISABLED",
175
176
  * // },
176
177
  * // SourceType: "ANCILLARY" || "DVB_SUB" || "EMBEDDED" || "SCTE20" || "SCC" || "TTML" || "STL" || "SRT" || "SMI" || "SMPTE_TT" || "TELETEXT" || "NULL_SOURCE" || "IMSC" || "WEBVTT",
177
178
  * // TeletextSourceSettings: { // TeletextSourceSettings
@@ -245,6 +246,13 @@ declare const GetJobTemplateCommand_base: {
245
246
  * // TimecodeStart: "STRING_VALUE",
246
247
  * // VideoOverlays: [ // __listOfVideoOverlay
247
248
  * // { // VideoOverlay
249
+ * // Crop: { // VideoOverlayCrop
250
+ * // Height: Number("int"),
251
+ * // Unit: "PIXELS" || "PERCENTAGE",
252
+ * // Width: Number("int"),
253
+ * // X: Number("int"),
254
+ * // Y: Number("int"),
255
+ * // },
248
256
  * // EndTimecode: "STRING_VALUE",
249
257
  * // InitialPosition: { // VideoOverlayPosition
250
258
  * // Height: Number("int"),
@@ -177,6 +177,7 @@ declare const ListJobTemplatesCommand_base: {
177
177
  * // SourceFile: "STRING_VALUE",
178
178
  * // TimeDelta: Number("int"),
179
179
  * // TimeDeltaUnits: "SECONDS" || "MILLISECONDS",
180
+ * // UpconvertSTLToTeletext: "UPCONVERT" || "DISABLED",
180
181
  * // },
181
182
  * // SourceType: "ANCILLARY" || "DVB_SUB" || "EMBEDDED" || "SCTE20" || "SCC" || "TTML" || "STL" || "SRT" || "SMI" || "SMPTE_TT" || "TELETEXT" || "NULL_SOURCE" || "IMSC" || "WEBVTT",
182
183
  * // TeletextSourceSettings: { // TeletextSourceSettings
@@ -250,6 +251,13 @@ declare const ListJobTemplatesCommand_base: {
250
251
  * // TimecodeStart: "STRING_VALUE",
251
252
  * // VideoOverlays: [ // __listOfVideoOverlay
252
253
  * // { // VideoOverlay
254
+ * // Crop: { // VideoOverlayCrop
255
+ * // Height: Number("int"),
256
+ * // Unit: "PIXELS" || "PERCENTAGE",
257
+ * // Width: Number("int"),
258
+ * // X: Number("int"),
259
+ * // Y: Number("int"),
260
+ * // },
253
261
  * // EndTimecode: "STRING_VALUE",
254
262
  * // InitialPosition: { // VideoOverlayPosition
255
263
  * // Height: Number("int"),
@@ -214,6 +214,7 @@ declare const ListJobsCommand_base: {
214
214
  * // SourceFile: "STRING_VALUE",
215
215
  * // TimeDelta: Number("int"),
216
216
  * // TimeDeltaUnits: "SECONDS" || "MILLISECONDS",
217
+ * // UpconvertSTLToTeletext: "UPCONVERT" || "DISABLED",
217
218
  * // },
218
219
  * // SourceType: "ANCILLARY" || "DVB_SUB" || "EMBEDDED" || "SCTE20" || "SCC" || "TTML" || "STL" || "SRT" || "SMI" || "SMPTE_TT" || "TELETEXT" || "NULL_SOURCE" || "IMSC" || "WEBVTT",
219
220
  * // TeletextSourceSettings: { // TeletextSourceSettings
@@ -304,6 +305,13 @@ declare const ListJobsCommand_base: {
304
305
  * // },
305
306
  * // VideoOverlays: [ // __listOfVideoOverlay
306
307
  * // { // VideoOverlay
308
+ * // Crop: { // VideoOverlayCrop
309
+ * // Height: Number("int"),
310
+ * // Unit: "PIXELS" || "PERCENTAGE",
311
+ * // Width: Number("int"),
312
+ * // X: Number("int"),
313
+ * // Y: Number("int"),
314
+ * // },
307
315
  * // EndTimecode: "STRING_VALUE",
308
316
  * // InitialPosition: { // VideoOverlayPosition
309
317
  * // Height: Number("int"),
@@ -215,6 +215,7 @@ declare const SearchJobsCommand_base: {
215
215
  * // SourceFile: "STRING_VALUE",
216
216
  * // TimeDelta: Number("int"),
217
217
  * // TimeDeltaUnits: "SECONDS" || "MILLISECONDS",
218
+ * // UpconvertSTLToTeletext: "UPCONVERT" || "DISABLED",
218
219
  * // },
219
220
  * // SourceType: "ANCILLARY" || "DVB_SUB" || "EMBEDDED" || "SCTE20" || "SCC" || "TTML" || "STL" || "SRT" || "SMI" || "SMPTE_TT" || "TELETEXT" || "NULL_SOURCE" || "IMSC" || "WEBVTT",
220
221
  * // TeletextSourceSettings: { // TeletextSourceSettings
@@ -305,6 +306,13 @@ declare const SearchJobsCommand_base: {
305
306
  * // },
306
307
  * // VideoOverlays: [ // __listOfVideoOverlay
307
308
  * // { // VideoOverlay
309
+ * // Crop: { // VideoOverlayCrop
310
+ * // Height: Number("int"),
311
+ * // Unit: "PIXELS" || "PERCENTAGE",
312
+ * // Width: Number("int"),
313
+ * // X: Number("int"),
314
+ * // Y: Number("int"),
315
+ * // },
308
316
  * // EndTimecode: "STRING_VALUE",
309
317
  * // InitialPosition: { // VideoOverlayPosition
310
318
  * // Height: Number("int"),
@@ -163,6 +163,7 @@ declare const UpdateJobTemplateCommand_base: {
163
163
  * SourceFile: "STRING_VALUE",
164
164
  * TimeDelta: Number("int"),
165
165
  * TimeDeltaUnits: "SECONDS" || "MILLISECONDS",
166
+ * UpconvertSTLToTeletext: "UPCONVERT" || "DISABLED",
166
167
  * },
167
168
  * SourceType: "ANCILLARY" || "DVB_SUB" || "EMBEDDED" || "SCTE20" || "SCC" || "TTML" || "STL" || "SRT" || "SMI" || "SMPTE_TT" || "TELETEXT" || "NULL_SOURCE" || "IMSC" || "WEBVTT",
168
169
  * TeletextSourceSettings: { // TeletextSourceSettings
@@ -236,6 +237,13 @@ declare const UpdateJobTemplateCommand_base: {
236
237
  * TimecodeStart: "STRING_VALUE",
237
238
  * VideoOverlays: [ // __listOfVideoOverlay
238
239
  * { // VideoOverlay
240
+ * Crop: { // VideoOverlayCrop
241
+ * Height: Number("int"),
242
+ * Unit: "PIXELS" || "PERCENTAGE",
243
+ * Width: Number("int"),
244
+ * X: Number("int"),
245
+ * Y: Number("int"),
246
+ * },
239
247
  * EndTimecode: "STRING_VALUE",
240
248
  * InitialPosition: { // VideoOverlayPosition
241
249
  * Height: Number("int"),
@@ -1696,6 +1704,7 @@ declare const UpdateJobTemplateCommand_base: {
1696
1704
  * // SourceFile: "STRING_VALUE",
1697
1705
  * // TimeDelta: Number("int"),
1698
1706
  * // TimeDeltaUnits: "SECONDS" || "MILLISECONDS",
1707
+ * // UpconvertSTLToTeletext: "UPCONVERT" || "DISABLED",
1699
1708
  * // },
1700
1709
  * // SourceType: "ANCILLARY" || "DVB_SUB" || "EMBEDDED" || "SCTE20" || "SCC" || "TTML" || "STL" || "SRT" || "SMI" || "SMPTE_TT" || "TELETEXT" || "NULL_SOURCE" || "IMSC" || "WEBVTT",
1701
1710
  * // TeletextSourceSettings: { // TeletextSourceSettings
@@ -1769,6 +1778,13 @@ declare const UpdateJobTemplateCommand_base: {
1769
1778
  * // TimecodeStart: "STRING_VALUE",
1770
1779
  * // VideoOverlays: [ // __listOfVideoOverlay
1771
1780
  * // { // VideoOverlay
1781
+ * // Crop: { // VideoOverlayCrop
1782
+ * // Height: Number("int"),
1783
+ * // Unit: "PIXELS" || "PERCENTAGE",
1784
+ * // Width: Number("int"),
1785
+ * // X: Number("int"),
1786
+ * // Y: Number("int"),
1787
+ * // },
1772
1788
  * // EndTimecode: "STRING_VALUE",
1773
1789
  * // InitialPosition: { // VideoOverlayPosition
1774
1790
  * // Height: Number("int"),
@@ -3216,6 +3216,18 @@ export declare const FileSourceTimeDeltaUnits: {
3216
3216
  * @public
3217
3217
  */
3218
3218
  export type FileSourceTimeDeltaUnits = (typeof FileSourceTimeDeltaUnits)[keyof typeof FileSourceTimeDeltaUnits];
3219
+ /**
3220
+ * @public
3221
+ * @enum
3222
+ */
3223
+ export declare const CaptionSourceUpconvertSTLToTeletext: {
3224
+ readonly DISABLED: "DISABLED";
3225
+ readonly UPCONVERT: "UPCONVERT";
3226
+ };
3227
+ /**
3228
+ * @public
3229
+ */
3230
+ export type CaptionSourceUpconvertSTLToTeletext = (typeof CaptionSourceUpconvertSTLToTeletext)[keyof typeof CaptionSourceUpconvertSTLToTeletext];
3219
3231
  /**
3220
3232
  * If your input captions are SCC, SMI, SRT, STL, TTML, WebVTT, or IMSC 1.1 in an xml file, specify the URI of the input caption source file. If your caption source is IMSC in an IMF package, use TrackSourceSettings instead of FileSoureSettings.
3221
3233
  * @public
@@ -3256,6 +3268,11 @@ export interface FileSourceSettings {
3256
3268
  * @public
3257
3269
  */
3258
3270
  TimeDeltaUnits?: FileSourceTimeDeltaUnits | undefined;
3271
+ /**
3272
+ * Specify whether this set of input captions appears in your outputs in both STL and Teletext format. If you choose Upconvert, MediaConvert includes the captions data in two ways: it passes the STL data through using the Teletext compatibility bytes fields of the Teletext wrapper, and it also translates the STL data into Teletext.
3273
+ * @public
3274
+ */
3275
+ UpconvertSTLToTeletext?: CaptionSourceUpconvertSTLToTeletext | undefined;
3259
3276
  }
3260
3277
  /**
3261
3278
  * @public
@@ -3713,6 +3730,37 @@ export declare const VideoOverlayUnit: {
3713
3730
  * @public
3714
3731
  */
3715
3732
  export type VideoOverlayUnit = (typeof VideoOverlayUnit)[keyof typeof VideoOverlayUnit];
3733
+ /**
3734
+ * Specify a rectangle of content to crop and use from your video overlay's input video. When you do, MediaConvert uses the cropped dimensions that you specify under X offset, Y offset, Width, and Height.
3735
+ * @public
3736
+ */
3737
+ export interface VideoOverlayCrop {
3738
+ /**
3739
+ * Specify the height of the video overlay cropping rectangle. To use the same height as your overlay input video: Keep blank, or enter 0. To specify a different height for the cropping rectangle: Enter an integer representing the Unit type that you choose, either Pixels or Percentage. For example, when you enter 100 and choose Pixels, the cropping rectangle will 100 pixels high. When you enter 10, choose Percentage, and your overlay input video is 1920x1080, the cropping rectangle will be 108 pixels high.
3740
+ * @public
3741
+ */
3742
+ Height?: number | undefined;
3743
+ /**
3744
+ * Specify the Unit type to use when you enter a value for X position, Y position, Width, or Height. You can choose Pixels or Percentage. Leave blank to use the default value, Pixels.
3745
+ * @public
3746
+ */
3747
+ Unit?: VideoOverlayUnit | undefined;
3748
+ /**
3749
+ * Specify the width of the video overlay cropping rectangle. To use the same width as your overlay input video: Keep blank, or enter 0. To specify a different width for the cropping rectangle: Enter an integer representing the Unit type that you choose, either Pixels or Percentage. For example, when you enter 100 and choose Pixels, the cropping rectangle will 100 pixels wide. When you enter 10, choose Percentage, and your overlay input video is 1920x1080, the cropping rectangle will be 192 pixels wide.
3750
+ * @public
3751
+ */
3752
+ Width?: number | undefined;
3753
+ /**
3754
+ * Specify the distance between the cropping rectangle and the left edge of your overlay video's frame. To position the cropping rectangle along the left edge: Keep blank, or enter 0. To position the cropping rectangle to the right, relative to the left edge of your overlay video's frame: Enter an integer representing the Unit type that you choose, either Pixels or Percentage. For example, when you enter 10 and choose Pixels, the cropping rectangle will be positioned 10 pixels from the left edge of the overlay video's frame. When you enter 10, choose Percentage, and your overlay input video is 1920x1080, the cropping rectangle will be positioned 192 pixels from the left edge of the overlay video's frame.
3755
+ * @public
3756
+ */
3757
+ X?: number | undefined;
3758
+ /**
3759
+ * Specify the distance between the cropping rectangle and the top edge of your overlay video's frame. To position the cropping rectangle along the top edge: Keep blank, or enter 0. To position the cropping rectangle down, relative to the top edge of your overlay video's frame: Enter an integer representing the Unit type that you choose, either Pixels or Percentage. For example, when you enter 10 and choose Pixels, the cropping rectangle will be positioned 10 pixels from the top edge of the overlay video's frame. When you enter 10, choose Percentage, and your overlay input video is 1920x1080, the cropping rectangle will be positioned 108 pixels from the top edge of the overlay video's frame.
3760
+ * @public
3761
+ */
3762
+ Y?: number | undefined;
3763
+ }
3716
3764
  /**
3717
3765
  * position of video overlay
3718
3766
  * @public
@@ -3825,6 +3873,11 @@ export interface VideoOverlayTransition {
3825
3873
  * @public
3826
3874
  */
3827
3875
  export interface VideoOverlay {
3876
+ /**
3877
+ * Specify a rectangle of content to crop and use from your video overlay's input video. When you do, MediaConvert uses the cropped dimensions that you specify under X offset, Y offset, Width, and Height.
3878
+ * @public
3879
+ */
3880
+ Crop?: VideoOverlayCrop | undefined;
3828
3881
  /**
3829
3882
  * Enter the end timecode in the base input video for this overlay. Your overlay will be active through this frame. To display your video overlay for the duration of the base input video: Leave blank. Use the format HH:MM:SS:FF or HH:MM:SS;FF, where HH is the hour, MM is the minute, SS isthe second, and FF is the frame number. When entering this value, take into account your choice for the base input video's timecode source. For example, if you have embedded timecodes that start at 01:00:00:00 and you want your overlay to end ten minutes into the video, enter 01:10:00:00.
3830
3883
  * @public
@@ -6462,26 +6515,3 @@ export interface MsSmoothAdditionalManifest {
6462
6515
  */
6463
6516
  SelectedOutputs?: string[] | undefined;
6464
6517
  }
6465
- /**
6466
- * @public
6467
- * @enum
6468
- */
6469
- export declare const MsSmoothAudioDeduplication: {
6470
- readonly COMBINE_DUPLICATE_STREAMS: "COMBINE_DUPLICATE_STREAMS";
6471
- readonly NONE: "NONE";
6472
- };
6473
- /**
6474
- * @public
6475
- */
6476
- export type MsSmoothAudioDeduplication = (typeof MsSmoothAudioDeduplication)[keyof typeof MsSmoothAudioDeduplication];
6477
- /**
6478
- * If you are using DRM, set DRM System to specify the value SpekeKeyProvider.
6479
- * @public
6480
- */
6481
- export interface MsSmoothEncryptionSettings {
6482
- /**
6483
- * If your output group type is HLS, DASH, or Microsoft Smooth, use these settings when doing DRM encryption with a SPEKE-compliant key provider. If your output group type is CMAF, use the SpekeKeyProviderCmaf settings instead.
6484
- * @public
6485
- */
6486
- SpekeKeyProvider?: SpekeKeyProvider | undefined;
6487
- }
@@ -1,4 +1,27 @@
1
- import { AudioDescription, AutomatedEncodingSettings, AvailBlanking, CaptionDescription, CmafGroupSettings, ColorConversion3DLUTSetting, DashIsoGroupSettings, DestinationSettings, EsamSettings, ExtendedDataServices, FileGroupSettings, FrameMetricType, Hdr10Metadata, HlsGroupSettings, Id3Insertion, ImageInserter, Input, KantarWatermarkSettings, MotionImageInserter, MsSmoothAdditionalManifest, MsSmoothAudioDeduplication, MsSmoothEncryptionSettings, NielsenConfiguration, NielsenNonLinearWatermarkSettings, Rectangle } from "./models_0";
1
+ import { AudioDescription, AutomatedEncodingSettings, AvailBlanking, CaptionDescription, CmafGroupSettings, ColorConversion3DLUTSetting, DashIsoGroupSettings, DestinationSettings, EsamSettings, ExtendedDataServices, FileGroupSettings, FrameMetricType, Hdr10Metadata, HlsGroupSettings, Id3Insertion, ImageInserter, Input, KantarWatermarkSettings, MotionImageInserter, MsSmoothAdditionalManifest, NielsenConfiguration, NielsenNonLinearWatermarkSettings, Rectangle, SpekeKeyProvider } from "./models_0";
2
+ /**
3
+ * @public
4
+ * @enum
5
+ */
6
+ export declare const MsSmoothAudioDeduplication: {
7
+ readonly COMBINE_DUPLICATE_STREAMS: "COMBINE_DUPLICATE_STREAMS";
8
+ readonly NONE: "NONE";
9
+ };
10
+ /**
11
+ * @public
12
+ */
13
+ export type MsSmoothAudioDeduplication = (typeof MsSmoothAudioDeduplication)[keyof typeof MsSmoothAudioDeduplication];
14
+ /**
15
+ * If you are using DRM, set DRM System to specify the value SpekeKeyProvider.
16
+ * @public
17
+ */
18
+ export interface MsSmoothEncryptionSettings {
19
+ /**
20
+ * If your output group type is HLS, DASH, or Microsoft Smooth, use these settings when doing DRM encryption with a SPEKE-compliant key provider. If your output group type is CMAF, use the SpekeKeyProviderCmaf settings instead.
21
+ * @public
22
+ */
23
+ SpekeKeyProvider?: SpekeKeyProvider | undefined;
24
+ }
2
25
  /**
3
26
  * @public
4
27
  * @enum
@@ -115,7 +138,7 @@ export interface OutputGroupSettings {
115
138
  */
116
139
  MsSmoothGroupSettings?: MsSmoothGroupSettings | undefined;
117
140
  /**
118
- * Optionally choose one or more per frame metric reports to generate along with your output. You can use these metrics to analyze your video output according to one or more commonly used image quality metrics. You can specify per frame metrics for output groups or for individual outputs. When you do, MediaConvert writes a CSV (Comma-Separated Values) file to your S3 output destination, named after the video, video codec, and metric type. For example: video_h264_PSNR.csv Jobs that generate per frame metrics will take longer to complete, depending on the resolution and complexity of your output. For example, some 4K jobs might take up to twice as long to complete. Note that when analyzing the video quality of your output, or when comparing the video quality of multiple different outputs, we generally also recommend a detailed visual review in a controlled environment. You can choose from the following per frame metrics: * PSNR: Peak Signal-to-Noise Ratio * SSIM: Structural Similarity Index Measure * MS_SSIM: Multi-Scale Similarity Index Measure * PSNR_HVS: Peak Signal-to-Noise Ratio, Human Visual System * VMAF: Video Multi-Method Assessment Fusion * QVBR: Quality-Defined Variable Bitrate. This option is only available when your output uses the QVBR rate control mode.
141
+ * Optionally choose one or more per frame metric reports to generate along with your output. You can use these metrics to analyze your video output according to one or more commonly used image quality metrics. You can specify per frame metrics for output groups or for individual outputs. When you do, MediaConvert writes a CSV (Comma-Separated Values) file to your S3 output destination, named after the output name and metric type. For example: videofile_PSNR.csv Jobs that generate per frame metrics will take longer to complete, depending on the resolution and complexity of your output. For example, some 4K jobs might take up to twice as long to complete. Note that when analyzing the video quality of your output, or when comparing the video quality of multiple different outputs, we generally also recommend a detailed visual review in a controlled environment. You can choose from the following per frame metrics: * PSNR: Peak Signal-to-Noise Ratio * SSIM: Structural Similarity Index Measure * MS_SSIM: Multi-Scale Similarity Index Measure * PSNR_HVS: Peak Signal-to-Noise Ratio, Human Visual System * VMAF: Video Multi-Method Assessment Fusion * QVBR: Quality-Defined Variable Bitrate. This option is only available when your output uses the QVBR rate control mode.
119
142
  * @public
120
143
  */
121
144
  PerFrameMetrics?: FrameMetricType[] | undefined;
@@ -1837,7 +1860,7 @@ export interface Av1Settings {
1837
1860
  */
1838
1861
  NumberBFramesBetweenReferenceFrames?: number | undefined;
1839
1862
  /**
1840
- * Optionally choose one or more per frame metric reports to generate along with your output. You can use these metrics to analyze your video output according to one or more commonly used image quality metrics. You can specify per frame metrics for output groups or for individual outputs. When you do, MediaConvert writes a CSV (Comma-Separated Values) file to your S3 output destination, named after the video, video codec, and metric type. For example: video_h264_PSNR.csv Jobs that generate per frame metrics will take longer to complete, depending on the resolution and complexity of your output. For example, some 4K jobs might take up to twice as long to complete. Note that when analyzing the video quality of your output, or when comparing the video quality of multiple different outputs, we generally also recommend a detailed visual review in a controlled environment. You can choose from the following per frame metrics: * PSNR: Peak Signal-to-Noise Ratio * SSIM: Structural Similarity Index Measure * MS_SSIM: Multi-Scale Similarity Index Measure * PSNR_HVS: Peak Signal-to-Noise Ratio, Human Visual System * VMAF: Video Multi-Method Assessment Fusion * QVBR: Quality-Defined Variable Bitrate. This option is only available when your output uses the QVBR rate control mode.
1863
+ * Optionally choose one or more per frame metric reports to generate along with your output. You can use these metrics to analyze your video output according to one or more commonly used image quality metrics. You can specify per frame metrics for output groups or for individual outputs. When you do, MediaConvert writes a CSV (Comma-Separated Values) file to your S3 output destination, named after the output name and metric type. For example: videofile_PSNR.csv Jobs that generate per frame metrics will take longer to complete, depending on the resolution and complexity of your output. For example, some 4K jobs might take up to twice as long to complete. Note that when analyzing the video quality of your output, or when comparing the video quality of multiple different outputs, we generally also recommend a detailed visual review in a controlled environment. You can choose from the following per frame metrics: * PSNR: Peak Signal-to-Noise Ratio * SSIM: Structural Similarity Index Measure * MS_SSIM: Multi-Scale Similarity Index Measure * PSNR_HVS: Peak Signal-to-Noise Ratio, Human Visual System * VMAF: Video Multi-Method Assessment Fusion * QVBR: Quality-Defined Variable Bitrate. This option is only available when your output uses the QVBR rate control mode.
1841
1864
  * @public
1842
1865
  */
1843
1866
  PerFrameMetrics?: FrameMetricType[] | undefined;
@@ -2017,7 +2040,7 @@ export interface AvcIntraSettings {
2017
2040
  */
2018
2041
  InterlaceMode?: AvcIntraInterlaceMode | undefined;
2019
2042
  /**
2020
- * Optionally choose one or more per frame metric reports to generate along with your output. You can use these metrics to analyze your video output according to one or more commonly used image quality metrics. You can specify per frame metrics for output groups or for individual outputs. When you do, MediaConvert writes a CSV (Comma-Separated Values) file to your S3 output destination, named after the video, video codec, and metric type. For example: video_h264_PSNR.csv Jobs that generate per frame metrics will take longer to complete, depending on the resolution and complexity of your output. For example, some 4K jobs might take up to twice as long to complete. Note that when analyzing the video quality of your output, or when comparing the video quality of multiple different outputs, we generally also recommend a detailed visual review in a controlled environment. You can choose from the following per frame metrics: * PSNR: Peak Signal-to-Noise Ratio * SSIM: Structural Similarity Index Measure * MS_SSIM: Multi-Scale Similarity Index Measure * PSNR_HVS: Peak Signal-to-Noise Ratio, Human Visual System * VMAF: Video Multi-Method Assessment Fusion * QVBR: Quality-Defined Variable Bitrate. This option is only available when your output uses the QVBR rate control mode.
2043
+ * Optionally choose one or more per frame metric reports to generate along with your output. You can use these metrics to analyze your video output according to one or more commonly used image quality metrics. You can specify per frame metrics for output groups or for individual outputs. When you do, MediaConvert writes a CSV (Comma-Separated Values) file to your S3 output destination, named after the output name and metric type. For example: videofile_PSNR.csv Jobs that generate per frame metrics will take longer to complete, depending on the resolution and complexity of your output. For example, some 4K jobs might take up to twice as long to complete. Note that when analyzing the video quality of your output, or when comparing the video quality of multiple different outputs, we generally also recommend a detailed visual review in a controlled environment. You can choose from the following per frame metrics: * PSNR: Peak Signal-to-Noise Ratio * SSIM: Structural Similarity Index Measure * MS_SSIM: Multi-Scale Similarity Index Measure * PSNR_HVS: Peak Signal-to-Noise Ratio, Human Visual System * VMAF: Video Multi-Method Assessment Fusion * QVBR: Quality-Defined Variable Bitrate. This option is only available when your output uses the QVBR rate control mode.
2021
2044
  * @public
2022
2045
  */
2023
2046
  PerFrameMetrics?: FrameMetricType[] | undefined;
@@ -2713,7 +2736,7 @@ export interface H264Settings {
2713
2736
  */
2714
2737
  ParNumerator?: number | undefined;
2715
2738
  /**
2716
- * Optionally choose one or more per frame metric reports to generate along with your output. You can use these metrics to analyze your video output according to one or more commonly used image quality metrics. You can specify per frame metrics for output groups or for individual outputs. When you do, MediaConvert writes a CSV (Comma-Separated Values) file to your S3 output destination, named after the video, video codec, and metric type. For example: video_h264_PSNR.csv Jobs that generate per frame metrics will take longer to complete, depending on the resolution and complexity of your output. For example, some 4K jobs might take up to twice as long to complete. Note that when analyzing the video quality of your output, or when comparing the video quality of multiple different outputs, we generally also recommend a detailed visual review in a controlled environment. You can choose from the following per frame metrics: * PSNR: Peak Signal-to-Noise Ratio * SSIM: Structural Similarity Index Measure * MS_SSIM: Multi-Scale Similarity Index Measure * PSNR_HVS: Peak Signal-to-Noise Ratio, Human Visual System * VMAF: Video Multi-Method Assessment Fusion * QVBR: Quality-Defined Variable Bitrate. This option is only available when your output uses the QVBR rate control mode.
2739
+ * Optionally choose one or more per frame metric reports to generate along with your output. You can use these metrics to analyze your video output according to one or more commonly used image quality metrics. You can specify per frame metrics for output groups or for individual outputs. When you do, MediaConvert writes a CSV (Comma-Separated Values) file to your S3 output destination, named after the output name and metric type. For example: videofile_PSNR.csv Jobs that generate per frame metrics will take longer to complete, depending on the resolution and complexity of your output. For example, some 4K jobs might take up to twice as long to complete. Note that when analyzing the video quality of your output, or when comparing the video quality of multiple different outputs, we generally also recommend a detailed visual review in a controlled environment. You can choose from the following per frame metrics: * PSNR: Peak Signal-to-Noise Ratio * SSIM: Structural Similarity Index Measure * MS_SSIM: Multi-Scale Similarity Index Measure * PSNR_HVS: Peak Signal-to-Noise Ratio, Human Visual System * VMAF: Video Multi-Method Assessment Fusion * QVBR: Quality-Defined Variable Bitrate. This option is only available when your output uses the QVBR rate control mode.
2717
2740
  * @public
2718
2741
  */
2719
2742
  PerFrameMetrics?: FrameMetricType[] | undefined;
@@ -3328,7 +3351,7 @@ export interface H265Settings {
3328
3351
  */
3329
3352
  ParNumerator?: number | undefined;
3330
3353
  /**
3331
- * Optionally choose one or more per frame metric reports to generate along with your output. You can use these metrics to analyze your video output according to one or more commonly used image quality metrics. You can specify per frame metrics for output groups or for individual outputs. When you do, MediaConvert writes a CSV (Comma-Separated Values) file to your S3 output destination, named after the video, video codec, and metric type. For example: video_h264_PSNR.csv Jobs that generate per frame metrics will take longer to complete, depending on the resolution and complexity of your output. For example, some 4K jobs might take up to twice as long to complete. Note that when analyzing the video quality of your output, or when comparing the video quality of multiple different outputs, we generally also recommend a detailed visual review in a controlled environment. You can choose from the following per frame metrics: * PSNR: Peak Signal-to-Noise Ratio * SSIM: Structural Similarity Index Measure * MS_SSIM: Multi-Scale Similarity Index Measure * PSNR_HVS: Peak Signal-to-Noise Ratio, Human Visual System * VMAF: Video Multi-Method Assessment Fusion * QVBR: Quality-Defined Variable Bitrate. This option is only available when your output uses the QVBR rate control mode.
3354
+ * Optionally choose one or more per frame metric reports to generate along with your output. You can use these metrics to analyze your video output according to one or more commonly used image quality metrics. You can specify per frame metrics for output groups or for individual outputs. When you do, MediaConvert writes a CSV (Comma-Separated Values) file to your S3 output destination, named after the output name and metric type. For example: videofile_PSNR.csv Jobs that generate per frame metrics will take longer to complete, depending on the resolution and complexity of your output. For example, some 4K jobs might take up to twice as long to complete. Note that when analyzing the video quality of your output, or when comparing the video quality of multiple different outputs, we generally also recommend a detailed visual review in a controlled environment. You can choose from the following per frame metrics: * PSNR: Peak Signal-to-Noise Ratio * SSIM: Structural Similarity Index Measure * MS_SSIM: Multi-Scale Similarity Index Measure * PSNR_HVS: Peak Signal-to-Noise Ratio, Human Visual System * VMAF: Video Multi-Method Assessment Fusion * QVBR: Quality-Defined Variable Bitrate. This option is only available when your output uses the QVBR rate control mode.
3332
3355
  * @public
3333
3356
  */
3334
3357
  PerFrameMetrics?: FrameMetricType[] | undefined;
@@ -3771,7 +3794,7 @@ export interface Mpeg2Settings {
3771
3794
  */
3772
3795
  ParNumerator?: number | undefined;
3773
3796
  /**
3774
- * Optionally choose one or more per frame metric reports to generate along with your output. You can use these metrics to analyze your video output according to one or more commonly used image quality metrics. You can specify per frame metrics for output groups or for individual outputs. When you do, MediaConvert writes a CSV (Comma-Separated Values) file to your S3 output destination, named after the video, video codec, and metric type. For example: video_h264_PSNR.csv Jobs that generate per frame metrics will take longer to complete, depending on the resolution and complexity of your output. For example, some 4K jobs might take up to twice as long to complete. Note that when analyzing the video quality of your output, or when comparing the video quality of multiple different outputs, we generally also recommend a detailed visual review in a controlled environment. You can choose from the following per frame metrics: * PSNR: Peak Signal-to-Noise Ratio * SSIM: Structural Similarity Index Measure * MS_SSIM: Multi-Scale Similarity Index Measure * PSNR_HVS: Peak Signal-to-Noise Ratio, Human Visual System * VMAF: Video Multi-Method Assessment Fusion * QVBR: Quality-Defined Variable Bitrate. This option is only available when your output uses the QVBR rate control mode.
3797
+ * Optionally choose one or more per frame metric reports to generate along with your output. You can use these metrics to analyze your video output according to one or more commonly used image quality metrics. You can specify per frame metrics for output groups or for individual outputs. When you do, MediaConvert writes a CSV (Comma-Separated Values) file to your S3 output destination, named after the output name and metric type. For example: videofile_PSNR.csv Jobs that generate per frame metrics will take longer to complete, depending on the resolution and complexity of your output. For example, some 4K jobs might take up to twice as long to complete. Note that when analyzing the video quality of your output, or when comparing the video quality of multiple different outputs, we generally also recommend a detailed visual review in a controlled environment. You can choose from the following per frame metrics: * PSNR: Peak Signal-to-Noise Ratio * SSIM: Structural Similarity Index Measure * MS_SSIM: Multi-Scale Similarity Index Measure * PSNR_HVS: Peak Signal-to-Noise Ratio, Human Visual System * VMAF: Video Multi-Method Assessment Fusion * QVBR: Quality-Defined Variable Bitrate. This option is only available when your output uses the QVBR rate control mode.
3775
3798
  * @public
3776
3799
  */
3777
3800
  PerFrameMetrics?: FrameMetricType[] | undefined;
@@ -3999,7 +4022,7 @@ export interface ProresSettings {
3999
4022
  */
4000
4023
  ParNumerator?: number | undefined;
4001
4024
  /**
4002
- * Optionally choose one or more per frame metric reports to generate along with your output. You can use these metrics to analyze your video output according to one or more commonly used image quality metrics. You can specify per frame metrics for output groups or for individual outputs. When you do, MediaConvert writes a CSV (Comma-Separated Values) file to your S3 output destination, named after the video, video codec, and metric type. For example: video_h264_PSNR.csv Jobs that generate per frame metrics will take longer to complete, depending on the resolution and complexity of your output. For example, some 4K jobs might take up to twice as long to complete. Note that when analyzing the video quality of your output, or when comparing the video quality of multiple different outputs, we generally also recommend a detailed visual review in a controlled environment. You can choose from the following per frame metrics: * PSNR: Peak Signal-to-Noise Ratio * SSIM: Structural Similarity Index Measure * MS_SSIM: Multi-Scale Similarity Index Measure * PSNR_HVS: Peak Signal-to-Noise Ratio, Human Visual System * VMAF: Video Multi-Method Assessment Fusion * QVBR: Quality-Defined Variable Bitrate. This option is only available when your output uses the QVBR rate control mode.
4025
+ * Optionally choose one or more per frame metric reports to generate along with your output. You can use these metrics to analyze your video output according to one or more commonly used image quality metrics. You can specify per frame metrics for output groups or for individual outputs. When you do, MediaConvert writes a CSV (Comma-Separated Values) file to your S3 output destination, named after the output name and metric type. For example: videofile_PSNR.csv Jobs that generate per frame metrics will take longer to complete, depending on the resolution and complexity of your output. For example, some 4K jobs might take up to twice as long to complete. Note that when analyzing the video quality of your output, or when comparing the video quality of multiple different outputs, we generally also recommend a detailed visual review in a controlled environment. You can choose from the following per frame metrics: * PSNR: Peak Signal-to-Noise Ratio * SSIM: Structural Similarity Index Measure * MS_SSIM: Multi-Scale Similarity Index Measure * PSNR_HVS: Peak Signal-to-Noise Ratio, Human Visual System * VMAF: Video Multi-Method Assessment Fusion * QVBR: Quality-Defined Variable Bitrate. This option is only available when your output uses the QVBR rate control mode.
4003
4026
  * @public
4004
4027
  */
4005
4028
  PerFrameMetrics?: FrameMetricType[] | undefined;
@@ -4986,7 +5009,7 @@ export interface XavcSettings {
4986
5009
  */
4987
5010
  FramerateNumerator?: number | undefined;
4988
5011
  /**
4989
- * Optionally choose one or more per frame metric reports to generate along with your output. You can use these metrics to analyze your video output according to one or more commonly used image quality metrics. You can specify per frame metrics for output groups or for individual outputs. When you do, MediaConvert writes a CSV (Comma-Separated Values) file to your S3 output destination, named after the video, video codec, and metric type. For example: video_h264_PSNR.csv Jobs that generate per frame metrics will take longer to complete, depending on the resolution and complexity of your output. For example, some 4K jobs might take up to twice as long to complete. Note that when analyzing the video quality of your output, or when comparing the video quality of multiple different outputs, we generally also recommend a detailed visual review in a controlled environment. You can choose from the following per frame metrics: * PSNR: Peak Signal-to-Noise Ratio * SSIM: Structural Similarity Index Measure * MS_SSIM: Multi-Scale Similarity Index Measure * PSNR_HVS: Peak Signal-to-Noise Ratio, Human Visual System * VMAF: Video Multi-Method Assessment Fusion * QVBR: Quality-Defined Variable Bitrate. This option is only available when your output uses the QVBR rate control mode.
5012
+ * Optionally choose one or more per frame metric reports to generate along with your output. You can use these metrics to analyze your video output according to one or more commonly used image quality metrics. You can specify per frame metrics for output groups or for individual outputs. When you do, MediaConvert writes a CSV (Comma-Separated Values) file to your S3 output destination, named after the output name and metric type. For example: videofile_PSNR.csv Jobs that generate per frame metrics will take longer to complete, depending on the resolution and complexity of your output. For example, some 4K jobs might take up to twice as long to complete. Note that when analyzing the video quality of your output, or when comparing the video quality of multiple different outputs, we generally also recommend a detailed visual review in a controlled environment. You can choose from the following per frame metrics: * PSNR: Peak Signal-to-Noise Ratio * SSIM: Structural Similarity Index Measure * MS_SSIM: Multi-Scale Similarity Index Measure * PSNR_HVS: Peak Signal-to-Noise Ratio, Human Visual System * VMAF: Video Multi-Method Assessment Fusion * QVBR: Quality-Defined Variable Bitrate. This option is only available when your output uses the QVBR rate control mode.
4990
5013
  * @public
4991
5014
  */
4992
5015
  PerFrameMetrics?: FrameMetricType[] | undefined;
@@ -6104,40 +6127,3 @@ export declare const StatusUpdateInterval: {
6104
6127
  * @public
6105
6128
  */
6106
6129
  export type StatusUpdateInterval = (typeof StatusUpdateInterval)[keyof typeof StatusUpdateInterval];
6107
- /**
6108
- * Information about when jobs are submitted, started, and finished is specified in Unix epoch format in seconds.
6109
- * @public
6110
- */
6111
- export interface Timing {
6112
- /**
6113
- * The time, in Unix epoch format, that the transcoding job finished
6114
- * @public
6115
- */
6116
- FinishTime?: Date | undefined;
6117
- /**
6118
- * The time, in Unix epoch format, that transcoding for the job began.
6119
- * @public
6120
- */
6121
- StartTime?: Date | undefined;
6122
- /**
6123
- * The time, in Unix epoch format, that you submitted the job.
6124
- * @public
6125
- */
6126
- SubmitTime?: Date | undefined;
6127
- }
6128
- /**
6129
- * Contains any warning codes and their count for the job.
6130
- * @public
6131
- */
6132
- export interface WarningGroup {
6133
- /**
6134
- * Warning code that identifies a specific warning in the job. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/warning_codes.html
6135
- * @public
6136
- */
6137
- Code: number | undefined;
6138
- /**
6139
- * The number of times this warning occurred in the job.
6140
- * @public
6141
- */
6142
- Count: number | undefined;
6143
- }
@@ -1,7 +1,44 @@
1
1
  import { ExceptionOptionType as __ExceptionOptionType } from "@smithy/smithy-client";
2
2
  import { MediaConvertServiceException as __BaseException } from "./MediaConvertServiceException";
3
3
  import { AccelerationSettings, AccelerationStatus, AudioDescription, AvailBlanking, BillingTagsSource, CaptionDescriptionPreset, ColorConversion3DLUTSetting, Endpoint, EsamSettings, ExtendedDataServices, HopDestination, InputTemplate, JobMessages, JobPhase, KantarWatermarkSettings, MotionImageInserter, NielsenConfiguration, NielsenNonLinearWatermarkSettings, OutputGroupDetail, QueueTransition } from "./models_0";
4
- import { ContainerSettings, JobSettings, JobStatus, OutputGroup, SimulateReservedQueue, StatusUpdateInterval, TimecodeConfig, TimedMetadataInsertion, Timing, VideoDescription, WarningGroup } from "./models_1";
4
+ import { ContainerSettings, JobSettings, JobStatus, OutputGroup, SimulateReservedQueue, StatusUpdateInterval, TimecodeConfig, TimedMetadataInsertion, VideoDescription } from "./models_1";
5
+ /**
6
+ * Information about when jobs are submitted, started, and finished is specified in Unix epoch format in seconds.
7
+ * @public
8
+ */
9
+ export interface Timing {
10
+ /**
11
+ * The time, in Unix epoch format, that the transcoding job finished
12
+ * @public
13
+ */
14
+ FinishTime?: Date | undefined;
15
+ /**
16
+ * The time, in Unix epoch format, that transcoding for the job began.
17
+ * @public
18
+ */
19
+ StartTime?: Date | undefined;
20
+ /**
21
+ * The time, in Unix epoch format, that you submitted the job.
22
+ * @public
23
+ */
24
+ SubmitTime?: Date | undefined;
25
+ }
26
+ /**
27
+ * Contains any warning codes and their count for the job.
28
+ * @public
29
+ */
30
+ export interface WarningGroup {
31
+ /**
32
+ * Warning code that identifies a specific warning in the job. For more information, see https://docs.aws.amazon.com/mediaconvert/latest/ug/warning_codes.html
33
+ * @public
34
+ */
35
+ Code: number | undefined;
36
+ /**
37
+ * The number of times this warning occurred in the job.
38
+ * @public
39
+ */
40
+ Count: number | undefined;
41
+ }
5
42
  /**
6
43
  * Each job converts an input file into an output file or files. For more information, see the User Guide at https://docs.aws.amazon.com/mediaconvert/latest/ug/what-is.html
7
44
  * @public
@@ -1317,6 +1317,12 @@ export declare const FileSourceTimeDeltaUnits: {
1317
1317
  };
1318
1318
  export type FileSourceTimeDeltaUnits =
1319
1319
  (typeof FileSourceTimeDeltaUnits)[keyof typeof FileSourceTimeDeltaUnits];
1320
+ export declare const CaptionSourceUpconvertSTLToTeletext: {
1321
+ readonly DISABLED: "DISABLED";
1322
+ readonly UPCONVERT: "UPCONVERT";
1323
+ };
1324
+ export type CaptionSourceUpconvertSTLToTeletext =
1325
+ (typeof CaptionSourceUpconvertSTLToTeletext)[keyof typeof CaptionSourceUpconvertSTLToTeletext];
1320
1326
  export interface FileSourceSettings {
1321
1327
  ByteRateLimit?: CaptionSourceByteRateLimit | undefined;
1322
1328
  Convert608To708?: FileSourceConvert608To708 | undefined;
@@ -1325,6 +1331,7 @@ export interface FileSourceSettings {
1325
1331
  SourceFile?: string | undefined;
1326
1332
  TimeDelta?: number | undefined;
1327
1333
  TimeDeltaUnits?: FileSourceTimeDeltaUnits | undefined;
1334
+ UpconvertSTLToTeletext?: CaptionSourceUpconvertSTLToTeletext | undefined;
1328
1335
  }
1329
1336
  export declare const CaptionSourceType: {
1330
1337
  readonly ANCILLARY: "ANCILLARY";
@@ -1473,6 +1480,13 @@ export declare const VideoOverlayUnit: {
1473
1480
  };
1474
1481
  export type VideoOverlayUnit =
1475
1482
  (typeof VideoOverlayUnit)[keyof typeof VideoOverlayUnit];
1483
+ export interface VideoOverlayCrop {
1484
+ Height?: number | undefined;
1485
+ Unit?: VideoOverlayUnit | undefined;
1486
+ Width?: number | undefined;
1487
+ X?: number | undefined;
1488
+ Y?: number | undefined;
1489
+ }
1476
1490
  export interface VideoOverlayPosition {
1477
1491
  Height?: number | undefined;
1478
1492
  Unit?: VideoOverlayUnit | undefined;
@@ -1502,6 +1516,7 @@ export interface VideoOverlayTransition {
1502
1516
  StartTimecode?: string | undefined;
1503
1517
  }
1504
1518
  export interface VideoOverlay {
1519
+ Crop?: VideoOverlayCrop | undefined;
1505
1520
  EndTimecode?: string | undefined;
1506
1521
  InitialPosition?: VideoOverlayPosition | undefined;
1507
1522
  Input?: VideoOverlayInput | undefined;
@@ -2383,12 +2398,3 @@ export interface MsSmoothAdditionalManifest {
2383
2398
  ManifestNameModifier?: string | undefined;
2384
2399
  SelectedOutputs?: string[] | undefined;
2385
2400
  }
2386
- export declare const MsSmoothAudioDeduplication: {
2387
- readonly COMBINE_DUPLICATE_STREAMS: "COMBINE_DUPLICATE_STREAMS";
2388
- readonly NONE: "NONE";
2389
- };
2390
- export type MsSmoothAudioDeduplication =
2391
- (typeof MsSmoothAudioDeduplication)[keyof typeof MsSmoothAudioDeduplication];
2392
- export interface MsSmoothEncryptionSettings {
2393
- SpekeKeyProvider?: SpekeKeyProvider | undefined;
2394
- }
@@ -19,12 +19,20 @@ import {
19
19
  KantarWatermarkSettings,
20
20
  MotionImageInserter,
21
21
  MsSmoothAdditionalManifest,
22
- MsSmoothAudioDeduplication,
23
- MsSmoothEncryptionSettings,
24
22
  NielsenConfiguration,
25
23
  NielsenNonLinearWatermarkSettings,
26
24
  Rectangle,
25
+ SpekeKeyProvider,
27
26
  } from "./models_0";
27
+ export declare const MsSmoothAudioDeduplication: {
28
+ readonly COMBINE_DUPLICATE_STREAMS: "COMBINE_DUPLICATE_STREAMS";
29
+ readonly NONE: "NONE";
30
+ };
31
+ export type MsSmoothAudioDeduplication =
32
+ (typeof MsSmoothAudioDeduplication)[keyof typeof MsSmoothAudioDeduplication];
33
+ export interface MsSmoothEncryptionSettings {
34
+ SpekeKeyProvider?: SpekeKeyProvider | undefined;
35
+ }
28
36
  export declare const MsSmoothFragmentLengthControl: {
29
37
  readonly EXACT: "EXACT";
30
38
  readonly GOP_MULTIPLE: "GOP_MULTIPLE";
@@ -2298,12 +2306,3 @@ export declare const StatusUpdateInterval: {
2298
2306
  };
2299
2307
  export type StatusUpdateInterval =
2300
2308
  (typeof StatusUpdateInterval)[keyof typeof StatusUpdateInterval];
2301
- export interface Timing {
2302
- FinishTime?: Date | undefined;
2303
- StartTime?: Date | undefined;
2304
- SubmitTime?: Date | undefined;
2305
- }
2306
- export interface WarningGroup {
2307
- Code: number | undefined;
2308
- Count: number | undefined;
2309
- }
@@ -31,10 +31,17 @@ import {
31
31
  StatusUpdateInterval,
32
32
  TimecodeConfig,
33
33
  TimedMetadataInsertion,
34
- Timing,
35
34
  VideoDescription,
36
- WarningGroup,
37
35
  } from "./models_1";
36
+ export interface Timing {
37
+ FinishTime?: Date | undefined;
38
+ StartTime?: Date | undefined;
39
+ SubmitTime?: Date | undefined;
40
+ }
41
+ export interface WarningGroup {
42
+ Code: number | undefined;
43
+ Count: number | undefined;
44
+ }
38
45
  export interface Job {
39
46
  AccelerationSettings?: AccelerationSettings | undefined;
40
47
  AccelerationStatus?: AccelerationStatus | undefined;
package/package.json CHANGED
@@ -1,7 +1,7 @@
1
1
  {
2
2
  "name": "@aws-sdk/client-mediaconvert",
3
3
  "description": "AWS SDK for JavaScript Mediaconvert Client for Node.js, Browser and React Native",
4
- "version": "3.808.0",
4
+ "version": "3.810.0",
5
5
  "scripts": {
6
6
  "build": "concurrently 'yarn:build:cjs' 'yarn:build:es' 'yarn:build:types'",
7
7
  "build:cjs": "node ../../scripts/compilation/inline client-mediaconvert",
@@ -20,38 +20,38 @@
20
20
  "dependencies": {
21
21
  "@aws-crypto/sha256-browser": "5.2.0",
22
22
  "@aws-crypto/sha256-js": "5.2.0",
23
- "@aws-sdk/core": "3.808.0",
24
- "@aws-sdk/credential-provider-node": "3.808.0",
23
+ "@aws-sdk/core": "3.810.0",
24
+ "@aws-sdk/credential-provider-node": "3.810.0",
25
25
  "@aws-sdk/middleware-host-header": "3.804.0",
26
26
  "@aws-sdk/middleware-logger": "3.804.0",
27
27
  "@aws-sdk/middleware-recursion-detection": "3.804.0",
28
- "@aws-sdk/middleware-user-agent": "3.808.0",
28
+ "@aws-sdk/middleware-user-agent": "3.810.0",
29
29
  "@aws-sdk/region-config-resolver": "3.808.0",
30
30
  "@aws-sdk/types": "3.804.0",
31
31
  "@aws-sdk/util-endpoints": "3.808.0",
32
32
  "@aws-sdk/util-user-agent-browser": "3.804.0",
33
- "@aws-sdk/util-user-agent-node": "3.808.0",
33
+ "@aws-sdk/util-user-agent-node": "3.810.0",
34
34
  "@smithy/config-resolver": "^4.1.2",
35
- "@smithy/core": "^3.3.1",
35
+ "@smithy/core": "^3.3.3",
36
36
  "@smithy/fetch-http-handler": "^5.0.2",
37
37
  "@smithy/hash-node": "^4.0.2",
38
38
  "@smithy/invalid-dependency": "^4.0.2",
39
39
  "@smithy/middleware-content-length": "^4.0.2",
40
- "@smithy/middleware-endpoint": "^4.1.4",
41
- "@smithy/middleware-retry": "^4.1.5",
42
- "@smithy/middleware-serde": "^4.0.3",
40
+ "@smithy/middleware-endpoint": "^4.1.6",
41
+ "@smithy/middleware-retry": "^4.1.7",
42
+ "@smithy/middleware-serde": "^4.0.5",
43
43
  "@smithy/middleware-stack": "^4.0.2",
44
44
  "@smithy/node-config-provider": "^4.1.1",
45
45
  "@smithy/node-http-handler": "^4.0.4",
46
46
  "@smithy/protocol-http": "^5.1.0",
47
- "@smithy/smithy-client": "^4.2.4",
47
+ "@smithy/smithy-client": "^4.2.6",
48
48
  "@smithy/types": "^4.2.0",
49
49
  "@smithy/url-parser": "^4.0.2",
50
50
  "@smithy/util-base64": "^4.0.0",
51
51
  "@smithy/util-body-length-browser": "^4.0.0",
52
52
  "@smithy/util-body-length-node": "^4.0.0",
53
- "@smithy/util-defaults-mode-browser": "^4.0.12",
54
- "@smithy/util-defaults-mode-node": "^4.0.12",
53
+ "@smithy/util-defaults-mode-browser": "^4.0.14",
54
+ "@smithy/util-defaults-mode-node": "^4.0.14",
55
55
  "@smithy/util-endpoints": "^3.0.4",
56
56
  "@smithy/util-middleware": "^4.0.2",
57
57
  "@smithy/util-retry": "^4.0.3",