fal-endpoint-types 1.3.38 → 1.3.39

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "fal-endpoint-types",
3
- "version": "1.3.38",
3
+ "version": "1.3.39",
4
4
  "description": "TypeScript types for Fal AI endpoints generated from the OpenAPI schemas.",
5
5
  "homepage": "https://github.com/rawpixel-vincent/fal-endpoint-types#readme",
6
6
  "bugs": {
@@ -4264,33 +4264,6 @@ export interface ForceParams {
4264
4264
  y: number;
4265
4265
  }
4266
4266
 
4267
- export interface File_2 {
4268
- /**
4269
- * Content Type
4270
- * @description The mime type of the file.
4271
- * @example image/png
4272
- */
4273
- content_type: string;
4274
- /**
4275
- * File Name
4276
- * @description The name of the file. It will be auto-generated if not provided.
4277
- * @example z9RV14K95DvU.png
4278
- */
4279
- file_name: string;
4280
- /**
4281
- * File Size
4282
- * @description The size of the file in bytes.
4283
- * @example 4404019
4284
- */
4285
- file_size: number;
4286
- /**
4287
- * Url
4288
- * @description The URL where the file can be downloaded from.
4289
- * @example https://url.to/generated/file/z9RV14K95DvU.png
4290
- */
4291
- url: string;
4292
- }
4293
-
4294
4267
  export interface File_1 {
4295
4268
  /**
4296
4269
  * Content Type
@@ -6174,11 +6174,6 @@ declare global {
6174
6174
  output: falEndpoints.BeatovenSoundEffectGenerationOutput;
6175
6175
  };
6176
6176
 
6177
- 'beatoven/music-generation': {
6178
- input: falEndpoints.BeatovenMusicGenerationInput;
6179
- output: falEndpoints.BeatovenMusicGenerationOutput;
6180
- };
6181
-
6182
6177
  'argil/avatars/text-to-video': {
6183
6178
  input: falEndpoints.ArgilAvatarsTextToVideoInput;
6184
6179
  output: falEndpoints.ArgilAvatarsTextToVideoOutput;
@@ -94,8 +94,8 @@ export interface XaiGrokImagineVideoTextToVideoOutput {
94
94
  * "fps": 24,
95
95
  * "width": 1280,
96
96
  * "file_name": "RUAbFYlssdqnbjNLmE8qP_IX7BNYGP.mp4",
97
- * "content_type": "video/mp4",
98
- * "num_frames": 145
97
+ * "num_frames": 145,
98
+ * "content_type": "video/mp4"
99
99
  * }
100
100
  */
101
101
  video: Components.VideoFile;
@@ -148,8 +148,8 @@ export interface XaiGrokImagineVideoReferenceToVideoOutput {
148
148
  * "fps": 24,
149
149
  * "width": 1280,
150
150
  * "file_name": "r2v_output.mp4",
151
- * "content_type": "video/mp4",
152
- * "num_frames": 192
151
+ * "num_frames": 192,
152
+ * "content_type": "video/mp4"
153
153
  * }
154
154
  */
155
155
  video: Components.VideoFile;
@@ -199,8 +199,8 @@ export interface XaiGrokImagineVideoImageToVideoOutput {
199
199
  * "fps": 24,
200
200
  * "width": 1280,
201
201
  * "file_name": "0Ci1dviuSnEyUZzBUq-_5_nu7MrAAa.mp4",
202
- * "content_type": "video/mp4",
203
- * "num_frames": 145
202
+ * "num_frames": 145,
203
+ * "content_type": "video/mp4"
204
204
  * }
205
205
  */
206
206
  video: Components.VideoFile;
@@ -237,8 +237,8 @@ export interface XaiGrokImagineVideoExtendVideoOutput {
237
237
  * "fps": 24,
238
238
  * "width": 1280,
239
239
  * "file_name": "extended_video.mp4",
240
- * "content_type": "video/mp4",
241
- * "num_frames": 384
240
+ * "num_frames": 384,
241
+ * "content_type": "video/mp4"
242
242
  * }
243
243
  */
244
244
  video: Components.VideoFile;
@@ -276,8 +276,8 @@ export interface XaiGrokImagineVideoEditVideoOutput {
276
276
  * "fps": 24,
277
277
  * "width": 1280,
278
278
  * "file_name": "EuDrZuQTW9m1phBXOsauz_EpJH3s8X.mp4",
279
- * "content_type": "video/mp4",
280
- * "num_frames": 121
279
+ * "num_frames": 121,
280
+ * "content_type": "video/mp4"
281
281
  * }
282
282
  */
283
283
  video: Components.VideoFile;
@@ -27160,7 +27160,7 @@ export interface SeedvrUpscaleImageSeamlessInput {
27160
27160
  /**
27161
27161
  * Image Url
27162
27162
  * @description The input image to be processed
27163
- * @example https://storage.googleapis.com/falserverless/example_inputs/seedvr2/image_in.png
27163
+ * @example https://v3b.fal.media/files/b/0a93eb33/dQ_ZwLyeu8P2UJYsa8RDW_seamless_image_in.png
27164
27164
  */
27165
27165
  image_url: string;
27166
27166
  /**
@@ -35240,6 +35240,12 @@ export interface PixverseLipsyncOutput {
35240
35240
  }
35241
35241
 
35242
35242
  export interface PixverseExtendFastInput {
35243
+ /**
35244
+ * Generate Audio Switch
35245
+ * @description Enable audio generation (BGM, SFX, dialogue). Supported in v5.6+ models.
35246
+ * @default false
35247
+ */
35248
+ generate_audio_switch?: boolean;
35243
35249
  /**
35244
35250
  * Model
35245
35251
  * @description The model version to use for generation
@@ -35294,6 +35300,12 @@ export interface PixverseExtendInput {
35294
35300
  * @enum {string}
35295
35301
  */
35296
35302
  duration?: '5' | '8';
35303
+ /**
35304
+ * Generate Audio Switch
35305
+ * @description Enable audio generation (BGM, SFX, dialogue). Supported in v5.6+ models.
35306
+ * @default false
35307
+ */
35308
+ generate_audio_switch?: boolean;
35297
35309
  /**
35298
35310
  * Model
35299
35311
  * @description The model version to use for generation
@@ -36365,7 +36377,7 @@ export interface PhotaEditInput {
36365
36377
  profile_ids?: string[];
36366
36378
  /**
36367
36379
  * Prompt
36368
- * @description Text description of the desired image. To refer to specific profiles, use @Profile1, @Profile2, etc. Profiles will not be applied if not referenced in the prompt.
36380
+ * @description Text description of the desired image. To refer to specific profiles, use [[profile_id_1]], [[profile_id_2]], etc.
36369
36381
  * @example Make this scene more realistic but still keep the game vibes
36370
36382
  */
36371
36383
  prompt: string;
@@ -36434,14 +36446,9 @@ export interface PhotaInput {
36434
36446
  * @enum {string}
36435
36447
  */
36436
36448
  output_format?: 'jpeg' | 'png' | 'webp';
36437
- /**
36438
- * Profile Ids
36439
- * @description List of profile IDs to use for the image generation. Profiles may be tagged in the prompt as @Profile1, @Profile2, etc.
36440
- */
36441
- profile_ids?: string[];
36442
36449
  /**
36443
36450
  * Prompt
36444
- * @description Text description of the desired image. In case you wish to use specific profiles, refer to them as @Profile1, @Profile2, etc. Profiles will not be applied if not referenced in the prompt.
36451
+ * @description Text description of the desired image. In case you wish to use specific profiles, refer to them as [[profile_id_1]], [[profile_id_2]], etc.
36445
36452
  * @example Middle Eastern man in traditional clothing sitting in a cool tent in the desert with a laptop
36446
36453
  */
36447
36454
  prompt: string;
@@ -52918,6 +52925,8 @@ export interface KlingImageV3ImageToImageInput {
52918
52925
  /**
52919
52926
  * Image Url
52920
52927
  * @description Reference image for image-to-image generation.
52928
+ *
52929
+ * Max file size: 10.0MB, Min width: 300px, Min height: 300px, Min aspect ratio: 0.40, Max aspect ratio: 2.50, Timeout: 20.0s
52921
52930
  * @example https://v3b.fal.media/files/b/0a8d06b9/3zxm2qoj2xYWSNwEe5Vd9_a74d767fc42e47a0bf657117fbcf8b90.png
52922
52931
  */
52923
52932
  image_url: string;
@@ -65251,15 +65260,15 @@ export interface FluxVisionUpscalerInput {
65251
65260
  export interface FluxVisionUpscalerOutput {
65252
65261
  /**
65253
65262
  * Caption
65254
- * @description The VLM-generated caption describing the upscaled image.
65255
- * @example A highly detailed upscaled photograph featuring sharp edges and enhanced textures. The image shows improved clarity in fine details with natural color preservation and minimal artifacts, demonstrating the AI-enhanced resolution increase.
65263
+ * @description The VLM-generated caption used for upscaling.
65264
+ * @example A highly detailed upscaled photograph featuring sharp edges and enhanced textures. The image shows improved clarity in fine details with natural color preservation and minimal artifacts.
65256
65265
  */
65257
65266
  caption: string;
65258
65267
  /**
65259
65268
  * @description The URL of the generated image.
65260
65269
  * @example {
65261
- * "height": 2048,
65262
65270
  * "file_size": 8842156,
65271
+ * "height": 2048,
65263
65272
  * "file_name": "20TZeUQtQ8oKgsCKXSL81_StableSR_00002_.png",
65264
65273
  * "content_type": "image/png",
65265
65274
  * "url": "https://v3b.fal.media/files/b/panda/20TZeUQtQ8oKgsCKXSL81_StableSR_00002_.png",
@@ -75651,7 +75660,7 @@ export interface FaceToStickerOutput {
75651
75660
  export interface F5TtsInput {
75652
75661
  /**
75653
75662
  * Text to be converted to speech
75654
- * @description The text to be converted to speech.
75663
+ * @description The text to be converted to speech. Maximum 5000 characters.
75655
75664
  * @example I don't really care what you call me. I've been a silent spectator, watching species evolve, empires rise and fall. But always remember, I am mighty and enduring. Respect me and I'll nurture you; ignore me and you shall face the consequences.
75656
75665
  */
75657
75666
  gen_text: string;
@@ -76308,220 +76317,220 @@ export interface ElevenlabsSpeechToTextScribeV2Output {
76308
76317
  * Words
76309
76318
  * @description Word-level transcription details
76310
76319
  * @example {
76311
- * "text": "Hey,",
76320
+ * "end": 0.539,
76312
76321
  * "start": 0.079,
76313
76322
  * "type": "word",
76314
- * "end": 0.539,
76323
+ * "text": "Hey,",
76315
76324
  * "speaker_id": "speaker_0"
76316
76325
  * }
76317
76326
  * @example {
76318
- * "text": " ",
76327
+ * "end": 0.599,
76319
76328
  * "start": 0.539,
76320
76329
  * "type": "spacing",
76321
- * "end": 0.599,
76330
+ * "text": " ",
76322
76331
  * "speaker_id": "speaker_0"
76323
76332
  * }
76324
76333
  * @example {
76325
- * "text": "this",
76334
+ * "end": 0.679,
76326
76335
  * "start": 0.599,
76327
76336
  * "type": "word",
76328
- * "end": 0.679,
76337
+ * "text": "this",
76329
76338
  * "speaker_id": "speaker_0"
76330
76339
  * }
76331
76340
  * @example {
76332
- * "text": " ",
76341
+ * "end": 0.739,
76333
76342
  * "start": 0.679,
76334
76343
  * "type": "spacing",
76335
- * "end": 0.739,
76344
+ * "text": " ",
76336
76345
  * "speaker_id": "speaker_0"
76337
76346
  * }
76338
76347
  * @example {
76339
- * "text": "is",
76348
+ * "end": 0.799,
76340
76349
  * "start": 0.739,
76341
76350
  * "type": "word",
76342
- * "end": 0.799,
76351
+ * "text": "is",
76343
76352
  * "speaker_id": "speaker_0"
76344
76353
  * }
76345
76354
  * @example {
76346
- * "text": " ",
76355
+ * "end": 0.939,
76347
76356
  * "start": 0.799,
76348
76357
  * "type": "spacing",
76349
- * "end": 0.939,
76358
+ * "text": " ",
76350
76359
  * "speaker_id": "speaker_0"
76351
76360
  * }
76352
76361
  * @example {
76353
- * "text": "a",
76362
+ * "end": 0.939,
76354
76363
  * "start": 0.939,
76355
76364
  * "type": "word",
76356
- * "end": 0.939,
76365
+ * "text": "a",
76357
76366
  * "speaker_id": "speaker_0"
76358
76367
  * }
76359
76368
  * @example {
76360
- * "text": " ",
76369
+ * "end": 0.959,
76361
76370
  * "start": 0.939,
76362
76371
  * "type": "spacing",
76363
- * "end": 0.959,
76372
+ * "text": " ",
76364
76373
  * "speaker_id": "speaker_0"
76365
76374
  * }
76366
76375
  * @example {
76367
- * "text": "test",
76376
+ * "end": 1.179,
76368
76377
  * "start": 0.959,
76369
76378
  * "type": "word",
76370
- * "end": 1.179,
76379
+ * "text": "test",
76371
76380
  * "speaker_id": "speaker_0"
76372
76381
  * }
76373
76382
  * @example {
76374
- * "text": " ",
76383
+ * "end": 1.219,
76375
76384
  * "start": 1.179,
76376
76385
  * "type": "spacing",
76377
- * "end": 1.219,
76386
+ * "text": " ",
76378
76387
  * "speaker_id": "speaker_0"
76379
76388
  * }
76380
76389
  * @example {
76381
- * "text": "recording",
76390
+ * "end": 1.719,
76382
76391
  * "start": 1.22,
76383
76392
  * "type": "word",
76384
- * "end": 1.719,
76393
+ * "text": "recording",
76385
76394
  * "speaker_id": "speaker_0"
76386
76395
  * }
76387
76396
  * @example {
76388
- * "text": " ",
76397
+ * "end": 1.719,
76389
76398
  * "start": 1.719,
76390
76399
  * "type": "spacing",
76391
- * "end": 1.719,
76400
+ * "text": " ",
76392
76401
  * "speaker_id": "speaker_0"
76393
76402
  * }
76394
76403
  * @example {
76395
- * "text": "for",
76404
+ * "end": 1.86,
76396
76405
  * "start": 1.719,
76397
76406
  * "type": "word",
76398
- * "end": 1.86,
76407
+ * "text": "for",
76399
76408
  * "speaker_id": "speaker_0"
76400
76409
  * }
76401
76410
  * @example {
76402
- * "text": " ",
76411
+ * "end": 1.879,
76403
76412
  * "start": 1.86,
76404
76413
  * "type": "spacing",
76405
- * "end": 1.879,
76414
+ * "text": " ",
76406
76415
  * "speaker_id": "speaker_0"
76407
76416
  * }
76408
76417
  * @example {
76409
- * "text": "Scribe",
76418
+ * "end": 2.24,
76410
76419
  * "start": 1.879,
76411
76420
  * "type": "word",
76412
- * "end": 2.24,
76421
+ * "text": "Scribe",
76413
76422
  * "speaker_id": "speaker_0"
76414
76423
  * }
76415
76424
  * @example {
76416
- * "text": " ",
76425
+ * "end": 2.319,
76417
76426
  * "start": 2.24,
76418
76427
  * "type": "spacing",
76419
- * "end": 2.319,
76428
+ * "text": " ",
76420
76429
  * "speaker_id": "speaker_0"
76421
76430
  * }
76422
76431
  * @example {
76423
- * "text": "version",
76432
+ * "end": 2.759,
76424
76433
  * "start": 2.319,
76425
76434
  * "type": "word",
76426
- * "end": 2.759,
76435
+ * "text": "version",
76427
76436
  * "speaker_id": "speaker_0"
76428
76437
  * }
76429
76438
  * @example {
76430
- * "text": " ",
76439
+ * "end": 2.779,
76431
76440
  * "start": 2.759,
76432
76441
  * "type": "spacing",
76433
- * "end": 2.779,
76442
+ * "text": " ",
76434
76443
  * "speaker_id": "speaker_0"
76435
76444
  * }
76436
76445
  * @example {
76437
- * "text": "two,",
76446
+ * "end": 3.379,
76438
76447
  * "start": 2.779,
76439
76448
  * "type": "word",
76440
- * "end": 3.379,
76449
+ * "text": "two,",
76441
76450
  * "speaker_id": "speaker_0"
76442
76451
  * }
76443
76452
  * @example {
76444
- * "text": " ",
76453
+ * "end": 3.399,
76445
76454
  * "start": 3.379,
76446
76455
  * "type": "spacing",
76447
- * "end": 3.399,
76456
+ * "text": " ",
76448
76457
  * "speaker_id": "speaker_0"
76449
76458
  * }
76450
76459
  * @example {
76451
- * "text": "which",
76460
+ * "end": 3.519,
76452
76461
  * "start": 3.399,
76453
76462
  * "type": "word",
76454
- * "end": 3.519,
76463
+ * "text": "which",
76455
76464
  * "speaker_id": "speaker_0"
76456
76465
  * }
76457
76466
  * @example {
76458
- * "text": " ",
76467
+ * "end": 3.539,
76459
76468
  * "start": 3.519,
76460
76469
  * "type": "spacing",
76461
- * "end": 3.539,
76470
+ * "text": " ",
76462
76471
  * "speaker_id": "speaker_0"
76463
76472
  * }
76464
76473
  * @example {
76465
- * "text": "is",
76474
+ * "end": 3.659,
76466
76475
  * "start": 3.539,
76467
76476
  * "type": "word",
76468
- * "end": 3.659,
76477
+ * "text": "is",
76469
76478
  * "speaker_id": "speaker_0"
76470
76479
  * }
76471
76480
  * @example {
76472
- * "text": " ",
76481
+ * "end": 3.699,
76473
76482
  * "start": 3.659,
76474
76483
  * "type": "spacing",
76475
- * "end": 3.699,
76484
+ * "text": " ",
76476
76485
  * "speaker_id": "speaker_0"
76477
76486
  * }
76478
76487
  * @example {
76479
- * "text": "now",
76488
+ * "end": 3.839,
76480
76489
  * "start": 3.699,
76481
76490
  * "type": "word",
76482
- * "end": 3.839,
76491
+ * "text": "now",
76483
76492
  * "speaker_id": "speaker_0"
76484
76493
  * }
76485
76494
  * @example {
76486
- * "text": " ",
76495
+ * "end": 3.839,
76487
76496
  * "start": 3.839,
76488
76497
  * "type": "spacing",
76489
- * "end": 3.839,
76498
+ * "text": " ",
76490
76499
  * "speaker_id": "speaker_0"
76491
76500
  * }
76492
76501
  * @example {
76493
- * "text": "available",
76502
+ * "end": 4.319,
76494
76503
  * "start": 3.839,
76495
76504
  * "type": "word",
76496
- * "end": 4.319,
76505
+ * "text": "available",
76497
76506
  * "speaker_id": "speaker_0"
76498
76507
  * }
76499
76508
  * @example {
76500
- * "text": " ",
76509
+ * "end": 4.339,
76501
76510
  * "start": 4.319,
76502
76511
  * "type": "spacing",
76503
- * "end": 4.339,
76512
+ * "text": " ",
76504
76513
  * "speaker_id": "speaker_0"
76505
76514
  * }
76506
76515
  * @example {
76507
- * "text": "on",
76516
+ * "end": 4.579,
76508
76517
  * "start": 4.339,
76509
76518
  * "type": "word",
76510
- * "end": 4.579,
76519
+ * "text": "on",
76511
76520
  * "speaker_id": "speaker_0"
76512
76521
  * }
76513
76522
  * @example {
76514
- * "text": " ",
76523
+ * "end": 4.599,
76515
76524
  * "start": 4.579,
76516
76525
  * "type": "spacing",
76517
- * "end": 4.599,
76526
+ * "text": " ",
76518
76527
  * "speaker_id": "speaker_0"
76519
76528
  * }
76520
76529
  * @example {
76521
- * "text": "fal.ai.",
76530
+ * "end": 5.699,
76522
76531
  * "start": 4.599,
76523
76532
  * "type": "word",
76524
- * "end": 5.699,
76533
+ * "text": "fal.ai.",
76525
76534
  * "speaker_id": "speaker_0"
76526
76535
  * }
76527
76536
  */
@@ -82702,11 +82711,7 @@ export interface AnimatediffSparsectrlLcmOutput {
82702
82711
  * @description The seed used to generate the video.
82703
82712
  */
82704
82713
  seed: number;
82705
- /**
82706
- * Video
82707
- * @description Generated video file.
82708
- */
82709
- video: Components.File_2;
82714
+ video: Components.File;
82710
82715
  }
82711
82716
 
82712
82717
  export interface AmtInterpolationFrameInterpolationInput {
@@ -85241,82 +85246,6 @@ export interface BeatovenSoundEffectGenerationOutput {
85241
85246
  prompt: string;
85242
85247
  }
85243
85248
 
85244
- export interface BeatovenMusicGenerationInput {
85245
- /**
85246
- * Creativity
85247
- * @description Creativity level - higher values allow more creative interpretation of the prompt
85248
- * @default 16
85249
- * @example 16
85250
- * @example 14
85251
- * @example 11
85252
- */
85253
- creativity?: number;
85254
- /**
85255
- * Duration
85256
- * @description Length of the generated music in seconds
85257
- * @default 90
85258
- * @example 90
85259
- * @example 47
85260
- * @example 150
85261
- */
85262
- duration?: number;
85263
- /**
85264
- * Negative Prompt
85265
- * @description Describe what you want to avoid in the music (instruments, styles, moods). Leave blank for none.
85266
- * @default
85267
- * @example noise
85268
- * @example distortion
85269
- * @example heavy drums
85270
- * @example high-hats
85271
- */
85272
- negative_prompt?: string;
85273
- /**
85274
- * Prompt
85275
- * @description Describe the music you want to generate
85276
- * @example Jazz music for a late-night restaurant setting
85277
- * @example A lush, ambient soundscape featuring serene sounds, and a gentle, melancholic piano melody
85278
- * @example Hip-hop music, mellow keys and vinyl crackle
85279
- * @example House music with synthesizers, driving bass and a steady 4/4 beat
85280
- * @example Classical piano melody with emotional depth and gentle strings
85281
- */
85282
- prompt: string;
85283
- /**
85284
- * Refinement
85285
- * @description Refinement level - higher values may improve quality but take longer
85286
- * @default 100
85287
- * @example 100
85288
- * @example 200
85289
- */
85290
- refinement?: number;
85291
- /**
85292
- * Seed
85293
- * @description Random seed for reproducible results - leave empty for random generation
85294
- */
85295
- seed?: number;
85296
- }
85297
-
85298
- export interface BeatovenMusicGenerationOutput {
85299
- /**
85300
- * @description Generated audio file in WAV format
85301
- * @example {
85302
- * "url": "https://v3b.fal.media/files/b/rabbit/DBesSNPP6NwfhwMftene-_music_ZfniDF.wav"
85303
- * }
85304
- */
85305
- audio: Components.File;
85306
- /**
85307
- * Metadata
85308
- * @description Generation metadata including duration, sample rate, and parameters
85309
- */
85310
- metadata: {
85311
- [key: string]: { [x: string]: any } | null;
85312
- };
85313
- /**
85314
- * Prompt
85315
- * @description The processed prompt used for generation
85316
- */
85317
- prompt: string;
85318
- }
85319
-
85320
85249
  export interface ArgilAvatarsTextToVideoInput {
85321
85250
  /**
85322
85251
  * Avatar