@neuracore/types 6.12.1 → 7.0.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -188,6 +188,8 @@ export interface DataFormat {
188
188
  distance_units?: DistanceUnitsConfig;
189
189
  pose_type?: PoseConfig;
190
190
  orientation?: OrientationConfig | null;
191
+ scale_position?: number;
192
+ scale_orientation?: number;
191
193
  ee_pose_input_type?: EndEffectorPoseInputTypeConfig;
192
194
  language_type?: LanguageConfig;
193
195
  normalize?: NormalizeConfig | null;
@@ -209,6 +211,9 @@ export interface OrientationConfig {
209
211
  quaternion_order?: QuaternionOrderConfig;
210
212
  euler_order?: EulerOrderConfig;
211
213
  angle_units?: AngleConfig1;
214
+ align_frame_roll?: number;
215
+ align_frame_pitch?: number;
216
+ align_frame_yaw?: number;
212
217
  [k: string]: unknown;
213
218
  }
214
219
  /**
@@ -350,6 +355,8 @@ export interface DataFormat1 {
350
355
  distance_units?: DistanceUnitsConfig;
351
356
  pose_type?: PoseConfig;
352
357
  orientation?: OrientationConfig | null;
358
+ scale_position?: number;
359
+ scale_orientation?: number;
353
360
  ee_pose_input_type?: EndEffectorPoseInputTypeConfig;
354
361
  language_type?: LanguageConfig;
355
362
  normalize?: NormalizeConfig | null;
@@ -403,6 +410,8 @@ export interface DataFormat2 {
403
410
  distance_units?: DistanceUnitsConfig;
404
411
  pose_type?: PoseConfig;
405
412
  orientation?: OrientationConfig | null;
413
+ scale_position?: number;
414
+ scale_orientation?: number;
406
415
  ee_pose_input_type?: EndEffectorPoseInputTypeConfig;
407
416
  language_type?: LanguageConfig;
408
417
  normalize?: NormalizeConfig | null;
@@ -456,6 +465,8 @@ export interface DataFormat3 {
456
465
  distance_units?: DistanceUnitsConfig;
457
466
  pose_type?: PoseConfig;
458
467
  orientation?: OrientationConfig | null;
468
+ scale_position?: number;
469
+ scale_orientation?: number;
459
470
  ee_pose_input_type?: EndEffectorPoseInputTypeConfig;
460
471
  language_type?: LanguageConfig;
461
472
  normalize?: NormalizeConfig | null;
@@ -492,6 +503,8 @@ export interface DataFormat4 {
492
503
  distance_units?: DistanceUnitsConfig;
493
504
  pose_type?: PoseConfig;
494
505
  orientation?: OrientationConfig | null;
506
+ scale_position?: number;
507
+ scale_orientation?: number;
495
508
  ee_pose_input_type?: EndEffectorPoseInputTypeConfig;
496
509
  language_type?: LanguageConfig;
497
510
  normalize?: NormalizeConfig | null;
@@ -528,6 +541,8 @@ export interface DataFormat5 {
528
541
  distance_units?: DistanceUnitsConfig;
529
542
  pose_type?: PoseConfig;
530
543
  orientation?: OrientationConfig | null;
544
+ scale_position?: number;
545
+ scale_orientation?: number;
531
546
  ee_pose_input_type?: EndEffectorPoseInputTypeConfig;
532
547
  language_type?: LanguageConfig;
533
548
  normalize?: NormalizeConfig | null;
@@ -564,6 +579,8 @@ export interface DataFormat6 {
564
579
  distance_units?: DistanceUnitsConfig;
565
580
  pose_type?: PoseConfig;
566
581
  orientation?: OrientationConfig | null;
582
+ scale_position?: number;
583
+ scale_orientation?: number;
567
584
  ee_pose_input_type?: EndEffectorPoseInputTypeConfig;
568
585
  language_type?: LanguageConfig;
569
586
  normalize?: NormalizeConfig | null;
@@ -600,6 +617,8 @@ export interface DataFormat7 {
600
617
  distance_units?: DistanceUnitsConfig;
601
618
  pose_type?: PoseConfig;
602
619
  orientation?: OrientationConfig | null;
620
+ scale_position?: number;
621
+ scale_orientation?: number;
603
622
  ee_pose_input_type?: EndEffectorPoseInputTypeConfig;
604
623
  language_type?: LanguageConfig;
605
624
  normalize?: NormalizeConfig | null;
@@ -636,6 +655,8 @@ export interface DataFormat8 {
636
655
  distance_units?: DistanceUnitsConfig;
637
656
  pose_type?: PoseConfig;
638
657
  orientation?: OrientationConfig | null;
658
+ scale_position?: number;
659
+ scale_orientation?: number;
639
660
  ee_pose_input_type?: EndEffectorPoseInputTypeConfig;
640
661
  language_type?: LanguageConfig;
641
662
  normalize?: NormalizeConfig | null;
@@ -672,6 +693,8 @@ export interface DataFormat9 {
672
693
  distance_units?: DistanceUnitsConfig;
673
694
  pose_type?: PoseConfig;
674
695
  orientation?: OrientationConfig | null;
696
+ scale_position?: number;
697
+ scale_orientation?: number;
675
698
  ee_pose_input_type?: EndEffectorPoseInputTypeConfig;
676
699
  language_type?: LanguageConfig;
677
700
  normalize?: NormalizeConfig | null;
@@ -708,6 +731,8 @@ export interface DataFormat10 {
708
731
  distance_units?: DistanceUnitsConfig;
709
732
  pose_type?: PoseConfig;
710
733
  orientation?: OrientationConfig | null;
734
+ scale_position?: number;
735
+ scale_orientation?: number;
711
736
  ee_pose_input_type?: EndEffectorPoseInputTypeConfig;
712
737
  language_type?: LanguageConfig;
713
738
  normalize?: NormalizeConfig | null;
@@ -755,19 +780,25 @@ export interface DeploymentConfig {
755
780
  * training_id: Identifier of the trained model to deploy.
756
781
  * name: Optional name for the endpoint.
757
782
  * ttl: Optional time-to-live in seconds for the endpoint.
758
- * model_input_order: Specification of the model input data order.
759
- * model_output_order: Specification of the model output data order.
783
+ * input_embodiment_description: Specification of the model input data
784
+ * order.
785
+ * output_embodiment_description: Specification of the model output data
786
+ * order.
760
787
  * config: Deployment configuration parameters.
761
788
  */
762
789
  export interface DeploymentRequest {
763
790
  training_id: string;
764
791
  name?: string | null;
765
792
  ttl?: number | null;
766
- model_input_order: {
767
- [k: string]: string[];
793
+ input_embodiment_description: {
794
+ [k: string]: {
795
+ [k: string]: string;
796
+ };
768
797
  };
769
- model_output_order: {
770
- [k: string]: string[];
798
+ output_embodiment_description: {
799
+ [k: string]: {
800
+ [k: string]: string;
801
+ };
771
802
  };
772
803
  config?: DeploymentConfig;
773
804
  }
@@ -887,8 +918,10 @@ export interface HandshakeMessage {
887
918
  * algorithm_config: Configuration parameters for the algorithm.
888
919
  * gpu_type: The type of GPU to use for the job.
889
920
  * num_gpus: The number of GPUs to allocate for the job.
890
- * input_robot_data_spec: Mapping of robot to input data types to names
891
- * output_robot_data_spec: Mapping of robot to output data types to names
921
+ * input_cross_embodiment_description: Mapping of robot to input data
922
+ * types to names
923
+ * output_cross_embodiment_description: Mapping of robot to output data
924
+ * types to names
892
925
  */
893
926
  export interface InternalStartTrainingJobRequest {
894
927
  org_id: string;
@@ -903,14 +936,18 @@ export interface InternalStartTrainingJobRequest {
903
936
  gpu_type: GPUType;
904
937
  num_gpus: number;
905
938
  synchronization_details: SynchronizationDetails;
906
- input_robot_data_spec: {
939
+ input_cross_embodiment_description: {
907
940
  [k: string]: {
908
- [k: string]: string[];
941
+ [k: string]: {
942
+ [k: string]: string;
943
+ };
909
944
  };
910
945
  };
911
- output_robot_data_spec: {
946
+ output_cross_embodiment_description: {
912
947
  [k: string]: {
913
- [k: string]: string[];
948
+ [k: string]: {
949
+ [k: string]: string;
950
+ };
914
951
  };
915
952
  };
916
953
  }
@@ -919,7 +956,8 @@ export interface InternalStartTrainingJobRequest {
919
956
  *
920
957
  * Attributes:
921
958
  * frequency: Synchronization frequency in Hz.
922
- * robot_data_spec: Specification of robot data to include in the synchronization.
959
+ * cross_embodiment_description: Specification of robot data to include
960
+ * in the synchronization.
923
961
  * max_delay_s: Maximum allowable delay (in seconds) for synchronization.
924
962
  * allow_duplicates: Whether to allow duplicate data points in the synchronization.
925
963
  * trim_start_end: Whether to trim the start and end of the episode
@@ -927,7 +965,7 @@ export interface InternalStartTrainingJobRequest {
927
965
  */
928
966
  export interface SynchronizationDetails {
929
967
  frequency: number;
930
- robot_data_spec: {
968
+ cross_embodiment_union: {
931
969
  [k: string]: {
932
970
  [k: string]: string[];
933
971
  };
@@ -1007,9 +1045,11 @@ export interface MetricsData {
1007
1045
  * ModelInitDescription(
1008
1046
  * input_data_types=[DataType.RGB_IMAGES, DataType.JOINT_POSITIONS],
1009
1047
  * output_data_types=[DataType.JOINT_TARGET_POSITIONS],
1010
- * dataset_statistics={
1048
+ * input_dataset_statistics={
1011
1049
  * DataType.RGB_IMAGES: DataItemStats(...),
1012
1050
  * DataType.JOINT_POSITIONS: DataItemStats(...),
1051
+ * },
1052
+ * output_dataset_statistics={
1013
1053
  * DataType.JOINT_TARGET_POSITIONS: DataItemStats(...),
1014
1054
  * },
1015
1055
  * )
@@ -1017,7 +1057,10 @@ export interface MetricsData {
1017
1057
  export interface ModelInitDescription {
1018
1058
  input_data_types: DataType[];
1019
1059
  output_data_types: DataType[];
1020
- dataset_statistics: {
1060
+ input_dataset_statistics: {
1061
+ [k: string]: (JointDataStats | CameraDataStats | PoseDataStats | EndEffectorPoseDataStats | ParallelGripperOpenAmountDataStats | PointCloudDataStats | LanguageDataStats | Custom1DDataStats)[];
1062
+ };
1063
+ output_dataset_statistics: {
1021
1064
  [k: string]: (JointDataStats | CameraDataStats | PoseDataStats | EndEffectorPoseDataStats | ParallelGripperOpenAmountDataStats | PointCloudDataStats | LanguageDataStats | Custom1DDataStats)[];
1022
1065
  };
1023
1066
  output_prediction_horizon: number;
@@ -1078,6 +1121,8 @@ export interface DataFormat11 {
1078
1121
  distance_units?: DistanceUnitsConfig;
1079
1122
  pose_type?: PoseConfig;
1080
1123
  orientation?: OrientationConfig | null;
1124
+ scale_position?: number;
1125
+ scale_orientation?: number;
1081
1126
  ee_pose_input_type?: EndEffectorPoseInputTypeConfig;
1082
1127
  language_type?: LanguageConfig;
1083
1128
  normalize?: NormalizeConfig | null;
@@ -1151,7 +1196,6 @@ export interface ParallelGripperOpenAmountData {
1151
1196
  * progress: Upload progress percentage (0-100)
1152
1197
  * expected_trace_count: Number of traces expected (set by register_traces API)
1153
1198
  * total_bytes: Total bytes expected across all traces (for progress bar)
1154
- * upload_method: Method used to upload recording data (streaming or daemon)
1155
1199
  */
1156
1200
  export interface PendingRecording {
1157
1201
  id: string;
@@ -1169,7 +1213,6 @@ export interface PendingRecording {
1169
1213
  status: PendingRecordingStatus;
1170
1214
  progress: number;
1171
1215
  expected_trace_count: number;
1172
- upload_method: RecordingUploadMethod;
1173
1216
  }
1174
1217
  /**
1175
1218
  * Metadata details for a recording.
@@ -1396,18 +1439,29 @@ export interface SynchronizedDataset {
1396
1439
  *
1397
1440
  * Attributes:
1398
1441
  * synchronized_dataset_id: Unique identifier for the synced dataset.
1399
- * robot_data_spec: Mapping of robot IDs to data type names.
1442
+ * cross_embodiment_description: Mapping of robot IDs to data type names.
1400
1443
  * dataset_statistics: Statistics for each robot and data type.
1401
1444
  */
1402
1445
  export interface SynchronizedDatasetStatistics {
1403
1446
  synchronized_dataset_id: string;
1404
- robot_data_spec: {
1447
+ input_cross_embodiment_description: {
1405
1448
  [k: string]: {
1406
- [k: string]: string[];
1449
+ [k: string]: {
1450
+ [k: string]: string;
1451
+ };
1452
+ };
1453
+ };
1454
+ output_cross_embodiment_description: {
1455
+ [k: string]: {
1456
+ [k: string]: {
1457
+ [k: string]: string;
1458
+ };
1407
1459
  };
1408
1460
  };
1409
1461
  dataset_statistics: {
1410
- [k: string]: (JointDataStats | CameraDataStats | PoseDataStats | EndEffectorPoseDataStats | ParallelGripperOpenAmountDataStats | PointCloudDataStats | LanguageDataStats | Custom1DDataStats)[];
1462
+ [k: string]: {
1463
+ [k: string]: (JointDataStats | CameraDataStats | PoseDataStats | EndEffectorPoseDataStats | ParallelGripperOpenAmountDataStats | PointCloudDataStats | LanguageDataStats | Custom1DDataStats)[];
1464
+ };
1411
1465
  };
1412
1466
  }
1413
1467
  /**
@@ -1471,8 +1525,8 @@ export interface TracesMetadataRequest {
1471
1525
  * previous_training_time: The time spent on the previous training, if applicable.
1472
1526
  * error: Any error message associated with the job, if applicable.
1473
1527
  * resume_points: List of timestamps where the job can be resumed.
1474
- * input_robot_data_spec: List of data types for the input data.
1475
- * output_robot_data_spec: List of data types for the output data.
1528
+ * input_cross_embodiment_description: List of data types for the input data.
1529
+ * output_cross_embodiment_description: List of data types for the output data.
1476
1530
  */
1477
1531
  export interface TrainingJob {
1478
1532
  id: string;
@@ -1498,14 +1552,18 @@ export interface TrainingJob {
1498
1552
  previous_training_time?: number | null;
1499
1553
  error?: string | null;
1500
1554
  resume_points: number[];
1501
- input_robot_data_spec: {
1555
+ input_cross_embodiment_description: {
1502
1556
  [k: string]: {
1503
- [k: string]: string[];
1557
+ [k: string]: {
1558
+ [k: string]: string;
1559
+ };
1504
1560
  };
1505
1561
  };
1506
- output_robot_data_spec: {
1562
+ output_cross_embodiment_description: {
1507
1563
  [k: string]: {
1508
- [k: string]: string[];
1564
+ [k: string]: {
1565
+ [k: string]: string;
1566
+ };
1509
1567
  };
1510
1568
  };
1511
1569
  synchronization_details: SynchronizationDetails;
@@ -1521,9 +1579,9 @@ export interface TrainingJob {
1521
1579
  * frequency: Synchronization frequency in Hz.
1522
1580
  * gpu_type: Type of GPU to use for training.
1523
1581
  * num_gpus: Number of GPUs to use for training.
1524
- * input_robot_data_spec: The robust mapping of robot to input
1582
+ * input_cross_embodiment_description: The robust mapping of robot to input
1525
1583
  * data types to names
1526
- * output_robot_data_spec: The robust mapping of robot to output
1584
+ * output_cross_embodiment_description: The robust mapping of robot to output
1527
1585
  * data types to names
1528
1586
  */
1529
1587
  export interface TrainingJobRequest {
@@ -1536,14 +1594,18 @@ export interface TrainingJobRequest {
1536
1594
  gpu_type: GPUType;
1537
1595
  num_gpus: number;
1538
1596
  synchronization_details: SynchronizationDetails;
1539
- input_robot_data_spec: {
1597
+ input_cross_embodiment_description: {
1540
1598
  [k: string]: {
1541
- [k: string]: string[];
1599
+ [k: string]: {
1600
+ [k: string]: string;
1601
+ };
1542
1602
  };
1543
1603
  };
1544
- output_robot_data_spec: {
1604
+ output_cross_embodiment_description: {
1545
1605
  [k: string]: {
1546
- [k: string]: string[];
1606
+ [k: string]: {
1607
+ [k: string]: string;
1608
+ };
1547
1609
  };
1548
1610
  };
1549
1611
  }
@@ -1842,15 +1904,6 @@ export declare enum PendingRecordingStatus {
1842
1904
  UPLOADING = "UPLOADING",
1843
1905
  UPLOADED = "UPLOADED"
1844
1906
  }
1845
- /**
1846
- * Method used to upload recording data.
1847
- *
1848
- * Indicates how the recording data is being uploaded to the backend.
1849
- */
1850
- export declare enum RecordingUploadMethod {
1851
- streaming = "streaming",
1852
- data_daemon = "data_daemon"
1853
- }
1854
1907
  export declare enum Type14 {
1855
1908
  PointCloudData = "PointCloudData"
1856
1909
  }
@@ -6,7 +6,7 @@
6
6
  /* Do not modify it by hand - just update the pydantic models and then re-run the script
7
7
  */
8
8
  Object.defineProperty(exports, "__esModule", { value: true });
9
- exports.GPUType2 = exports.TrainingJobStatus = exports.RecordingNotificationType = exports.RecordingDataTraceStatus = exports.Type16 = exports.Type15 = exports.Type14 = exports.RecordingUploadMethod = exports.PendingRecordingStatus = exports.RecordingStatus = exports.Type13 = exports.VideoFormat = exports.Type12 = exports.Type11 = exports.Type10 = exports.Type9 = exports.Type8 = exports.Type7 = exports.Type6 = exports.MessageType = exports.GPUType1 = exports.EndpointStatus = exports.Type5 = exports.Type4 = exports.Type3 = exports.GPUType = exports.DatasetTypeConfig = exports.Type2 = exports.IntrinsicsConfig = exports.PoseConfig1 = exports.ActionSpaceConfig = exports.ActionTypeConfig = exports.JointPositionInputTypeConfig = exports.VisualJointInputTypeConfig = exports.LanguageConfig = exports.EndEffectorPoseInputTypeConfig = exports.AngleConfig1 = exports.EulerOrderConfig = exports.QuaternionOrderConfig = exports.RotationConfig = exports.PoseConfig = exports.DistanceUnitsConfig = exports.TorqueUnitsConfig = exports.AngleConfig = exports.ImageChannelOrderConfig = exports.ImageConventionConfig = exports.Type1 = exports.Type = exports.DataType = void 0;
9
+ exports.GPUType2 = exports.TrainingJobStatus = exports.RecordingNotificationType = exports.RecordingDataTraceStatus = exports.Type16 = exports.Type15 = exports.Type14 = exports.PendingRecordingStatus = exports.RecordingStatus = exports.Type13 = exports.VideoFormat = exports.Type12 = exports.Type11 = exports.Type10 = exports.Type9 = exports.Type8 = exports.Type7 = exports.Type6 = exports.MessageType = exports.GPUType1 = exports.EndpointStatus = exports.Type5 = exports.Type4 = exports.Type3 = exports.GPUType = exports.DatasetTypeConfig = exports.Type2 = exports.IntrinsicsConfig = exports.PoseConfig1 = exports.ActionSpaceConfig = exports.ActionTypeConfig = exports.JointPositionInputTypeConfig = exports.VisualJointInputTypeConfig = exports.LanguageConfig = exports.EndEffectorPoseInputTypeConfig = exports.AngleConfig1 = exports.EulerOrderConfig = exports.QuaternionOrderConfig = exports.RotationConfig = exports.PoseConfig = exports.DistanceUnitsConfig = exports.TorqueUnitsConfig = exports.AngleConfig = exports.ImageChannelOrderConfig = exports.ImageConventionConfig = exports.Type1 = exports.Type = exports.DataType = void 0;
10
10
  /**
11
11
  * Enumeration of supported data types in the Neuracore system.
12
12
  *
@@ -343,16 +343,6 @@ var PendingRecordingStatus;
343
343
  PendingRecordingStatus["UPLOADING"] = "UPLOADING";
344
344
  PendingRecordingStatus["UPLOADED"] = "UPLOADED";
345
345
  })(PendingRecordingStatus || (exports.PendingRecordingStatus = PendingRecordingStatus = {}));
346
- /**
347
- * Method used to upload recording data.
348
- *
349
- * Indicates how the recording data is being uploaded to the backend.
350
- */
351
- var RecordingUploadMethod;
352
- (function (RecordingUploadMethod) {
353
- RecordingUploadMethod["streaming"] = "streaming";
354
- RecordingUploadMethod["data_daemon"] = "data_daemon";
355
- })(RecordingUploadMethod || (exports.RecordingUploadMethod = RecordingUploadMethod = {}));
356
346
  var Type14;
357
347
  (function (Type14) {
358
348
  Type14["PointCloudData"] = "PointCloudData";
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@neuracore/types",
3
- "version": "6.12.1",
3
+ "version": "7.0.0",
4
4
  "description": "Shared TypeScript type definitions for Neuracore robotics platform",
5
5
  "main": "dist/index.js",
6
6
  "types": "dist/index.d.ts",