fal-endpoint-types 1.3.50 → 1.3.52

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "fal-endpoint-types",
3
- "version": "1.3.50",
3
+ "version": "1.3.52",
4
4
  "description": "TypeScript types for Fal AI endpoints generated from the OpenAPI schemas.",
5
5
  "homepage": "https://github.com/rawpixel-vincent/fal-endpoint-types#readme",
6
6
  "bugs": {
@@ -1758,6 +1758,19 @@ export interface V2VValidation {
1758
1758
  reference_video_url: string;
1759
1759
  }
1760
1760
 
1761
+ export interface UsageInfo_2 {
1762
+ /**
1763
+ * Input Tokens
1764
+ * @description Number of input tokens processed.
1765
+ */
1766
+ input_tokens: number;
1767
+ /**
1768
+ * Output Tokens
1769
+ * @description Number of output tokens generated.
1770
+ */
1771
+ output_tokens: number;
1772
+ }
1773
+
1761
1774
  export interface UsageInfo_1 {
1762
1775
  /**
1763
1776
  * Decode Time Ms
@@ -2453,21 +2466,46 @@ export interface SAM3DBodyPersonMetadata {
2453
2466
  * @description Bounding box [x_min, y_min, x_max, y_max]
2454
2467
  */
2455
2468
  bbox: number[];
2469
+ /**
2470
+ * Body Pose Params
2471
+ * @description Per-joint body pose parameters (axis-angle form) from the MHR model.
2472
+ */
2473
+ body_pose_params?: number[][];
2474
+ /**
2475
+ * Expr Params
2476
+ * @description MHR facial-expression parameters.
2477
+ */
2478
+ expr_params?: number[];
2456
2479
  /**
2457
2480
  * Focal Length
2458
2481
  * @description Estimated focal length
2459
2482
  */
2460
2483
  focal_length: number;
2484
+ /**
2485
+ * Global Rot
2486
+ * @description Global root rotation produced by MHR. Shape matches the upstream tensor (axis-angle [3] or rotation matrix [3, 3]).
2487
+ */
2488
+ global_rot?: { [x: string]: any }[];
2489
+ /**
2490
+ * Hand Pose Params
2491
+ * @description Per-joint hand pose parameters (axis-angle form) from the MHR model.
2492
+ */
2493
+ hand_pose_params?: number[][];
2461
2494
  /**
2462
2495
  * Keypoints 2D
2463
- * @description 2D keypoints [[x, y], ...] - 70 body keypoints
2496
+ * @description 2D keypoints [[x, y], ...] - 70 MHR body keypoints in image coordinates. See `SAM3DBodyMetadata.keypoint_names` for the ordered name of each index.
2464
2497
  */
2465
2498
  keypoints_2d: number[][];
2466
2499
  /**
2467
2500
  * Keypoints 3D
2468
- * @description 3D keypoints [[x, y, z], ...] - 70 body keypoints in camera space
2501
+ * @description 3D keypoints [[x, y, z], ...] - 70 MHR body keypoints in camera space. Ordering matches `keypoints_2d` / `SAM3DBodyMetadata.keypoint_names`.
2469
2502
  */
2470
2503
  keypoints_3d?: number[][];
2504
+ /**
2505
+ * Mhr Model Params
2506
+ * @description Packed MHR parameter vector (concatenated shape/pose/expression/scale). Shape is forwarded as-is from the upstream model.
2507
+ */
2508
+ mhr_model_params?: { [x: string]: any }[];
2471
2509
  /**
2472
2510
  * Person Id
2473
2511
  * @description Index of the person in the scene
@@ -2478,9 +2516,39 @@ export interface SAM3DBodyPersonMetadata {
2478
2516
  * @description Predicted camera translation [tx, ty, tz]
2479
2517
  */
2480
2518
  pred_cam_t: number[];
2519
+ /**
2520
+ * Pred Global Rots
2521
+ * @description Per-joint global rotations (world-space), typically [N_joints, 3, 3] rotation matrices. Needed for inverse linear-blend skinning / un-posing clients.
2522
+ */
2523
+ pred_global_rots?: { [x: string]: any }[];
2524
+ /**
2525
+ * Pred Joint Coords
2526
+ * @description Skeleton joint positions in world space [[x, y, z], ...]. One row per MHR joint.
2527
+ */
2528
+ pred_joint_coords?: number[][];
2529
+ /**
2530
+ * Pred Pose Raw
2531
+ * @description Raw pose transforms produced by the MHR decoder (pre-FK), forwarded as-is from the upstream model.
2532
+ */
2533
+ pred_pose_raw?: { [x: string]: any }[];
2534
+ /**
2535
+ * Scale Params
2536
+ * @description MHR scale parameters (isotropic or per-axis).
2537
+ */
2538
+ scale_params?: number[];
2539
+ /**
2540
+ * Shape Params
2541
+ * @description MHR identity (β) shape parameters. Enables local canonical-pose mesh reconstruction when combined with the MHR model.
2542
+ */
2543
+ shape_params?: number[];
2481
2544
  }
2482
2545
 
2483
2546
  export interface SAM3DBodyMetadata {
2547
+ /**
2548
+ * Keypoint Names
2549
+ * @description Ordered names of the 70 MHR keypoints returned by this endpoint. Index `i` in this list corresponds to index `i` in every person's `keypoints_2d` and `keypoints_3d` arrays. Sourced from facebookresearch/sam-3d-body `mhr70.py`.
2550
+ */
2551
+ keypoint_names?: string[];
2484
2552
  /**
2485
2553
  * Num People
2486
2554
  * @description Number of people detected
@@ -4608,12 +4676,12 @@ export interface DynamicMask {
4608
4676
  * @description List of trajectories
4609
4677
  * @example [
4610
4678
  * {
4611
- * "x": 279,
4612
- * "y": 219
4679
+ * "y": 219,
4680
+ * "x": 279
4613
4681
  * },
4614
4682
  * {
4615
- * "x": 417,
4616
- * "y": 65
4683
+ * "y": 65,
4684
+ * "x": 417
4617
4685
  * }
4618
4686
  * ]
4619
4687
  */
@@ -294,6 +294,26 @@ declare global {
294
294
  output: falEndpoints.OpenaiGptImage2Output;
295
295
  };
296
296
 
297
+ 'nvidia/nemotron-3-nano-omni/vision': {
298
+ input: falEndpoints.NvidiaNemotron3NanoOmniVisionInput;
299
+ output: falEndpoints.NvidiaNemotron3NanoOmniVisionOutput;
300
+ };
301
+
302
+ 'nvidia/nemotron-3-nano-omni/video': {
303
+ input: falEndpoints.NvidiaNemotron3NanoOmniVideoInput;
304
+ output: falEndpoints.NvidiaNemotron3NanoOmniVideoOutput;
305
+ };
306
+
307
+ 'nvidia/nemotron-3-nano-omni/audio': {
308
+ input: falEndpoints.NvidiaNemotron3NanoOmniAudioInput;
309
+ output: falEndpoints.NvidiaNemotron3NanoOmniAudioOutput;
310
+ };
311
+
312
+ 'nvidia/nemotron-3-nano-omni': {
313
+ input: falEndpoints.NvidiaNemotron3NanoOmniInput;
314
+ output: falEndpoints.NvidiaNemotron3NanoOmniOutput;
315
+ };
316
+
297
317
  'moonvalley/marey/t2v': {
298
318
  input: falEndpoints.MoonvalleyMareyT2vInput;
299
319
  output: falEndpoints.MoonvalleyMareyT2vOutput;
@@ -6618,6 +6638,26 @@ declare global {
6618
6638
  input: falEndpoints.ArgilAvatarsAudioToVideoInput;
6619
6639
  output: falEndpoints.ArgilAvatarsAudioToVideoOutput;
6620
6640
  };
6641
+
6642
+ 'alibaba/happy-horse/video-edit': {
6643
+ input: falEndpoints.AlibabaHappyHorseVideoEditInput;
6644
+ output: falEndpoints.AlibabaHappyHorseVideoEditOutput;
6645
+ };
6646
+
6647
+ 'alibaba/happy-horse/text-to-video': {
6648
+ input: falEndpoints.AlibabaHappyHorseTextToVideoInput;
6649
+ output: falEndpoints.AlibabaHappyHorseTextToVideoOutput;
6650
+ };
6651
+
6652
+ 'alibaba/happy-horse/reference-to-video': {
6653
+ input: falEndpoints.AlibabaHappyHorseReferenceToVideoInput;
6654
+ output: falEndpoints.AlibabaHappyHorseReferenceToVideoOutput;
6655
+ };
6656
+
6657
+ 'alibaba/happy-horse/image-to-video': {
6658
+ input: falEndpoints.AlibabaHappyHorseImageToVideoInput;
6659
+ output: falEndpoints.AlibabaHappyHorseImageToVideoOutput;
6660
+ };
6621
6661
  }
6622
6662
  }
6623
6663
  }