@neuracore/types 6.8.0 → 6.10.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -117,6 +117,8 @@ export interface DataItemStats {
117
117
  count: unknown[] | null;
118
118
  min: unknown[] | null;
119
119
  max: unknown[] | null;
120
+ q01: unknown[] | null;
121
+ q99: unknown[] | null;
120
122
  }
121
123
  /**
122
124
  * Custom 1D numerical data for specialized applications.
@@ -694,6 +696,39 @@ export interface EndEffectorPoseDataStats {
694
696
  type: Type5;
695
697
  pose: DataItemStats;
696
698
  }
699
+ /**
700
+ * Response model for endpoint operations.
701
+ *
702
+ * Attributes:
703
+ * id: Unique identifier for the endpoint
704
+ * name: Display name of the endpoint
705
+ * training_job_id: ID of the training job used for this endpoint
706
+ * status: Current status of the endpoint
707
+ * gpu_type: Type of GPU used by the endpoint
708
+ * create_time: Timestamp when the endpoint was created
709
+ * update_time: Timestamp when the endpoint was last updated
710
+ * endpoint_url: URL for accessing the endpoint (if active)
711
+ * error: Error message if the endpoint failed
712
+ * error_code: Error code if the endpoint failed
713
+ * expiration_time: Timestamp when the endpoint expires (if set)
714
+ * cloud_compute_job_id: ID of the cloud compute job running this endpoint
715
+ * zone: The GCP zone where the endpoint is running
716
+ */
717
+ export interface Endpoint {
718
+ id: string;
719
+ name: string;
720
+ training_job_id: string;
721
+ zone?: string | null;
722
+ status: EndpointStatus;
723
+ gpu_type?: GPUType1;
724
+ create_time: number;
725
+ update_time?: number | null;
726
+ endpoint_url?: string | null;
727
+ error?: string | null;
728
+ error_code?: number | null;
729
+ expiration_time?: number | null;
730
+ cloud_compute_job_id?: string | null;
731
+ }
697
732
  /**
698
733
  * Description of a single episode with statistics and counts.
699
734
  *
@@ -1346,7 +1381,7 @@ export interface TrainingJob {
1346
1381
  algorithm_config: {
1347
1382
  [k: string]: unknown;
1348
1383
  };
1349
- gpu_type: GPUType1;
1384
+ gpu_type: GPUType2;
1350
1385
  num_gpus: number;
1351
1386
  resumed_at?: number | null;
1352
1387
  previous_training_time?: number | null;
@@ -1538,7 +1573,7 @@ export declare enum DatasetTypeConfig {
1538
1573
  TFDS = "TFDS"
1539
1574
  }
1540
1575
  /**
1541
- * GPU types available for training.
1576
+ * GPU types available for training and deployment.
1542
1577
  */
1543
1578
  export declare enum GPUType {
1544
1579
  NVIDIA_H100_80GB = "NVIDIA_H100_80GB",
@@ -1559,6 +1594,28 @@ export declare enum Type4 {
1559
1594
  export declare enum Type5 {
1560
1595
  EndEffectorPoseDataStats = "EndEffectorPoseDataStats"
1561
1596
  }
1597
+ /**
1598
+ * Model endpoint status.
1599
+ */
1600
+ export declare enum EndpointStatus {
1601
+ creating = "creating",
1602
+ active = "active",
1603
+ failed = "failed",
1604
+ deactivated = "deactivated"
1605
+ }
1606
+ /**
1607
+ * GPU types available for training and deployment.
1608
+ */
1609
+ export declare enum GPUType1 {
1610
+ NVIDIA_H100_80GB = "NVIDIA_H100_80GB",
1611
+ NVIDIA_A100_80GB = "NVIDIA_A100_80GB",
1612
+ NVIDIA_TESLA_A100 = "NVIDIA_TESLA_A100",
1613
+ NVIDIA_TESLA_V100 = "NVIDIA_TESLA_V100",
1614
+ NVIDIA_TESLA_P100 = "NVIDIA_TESLA_P100",
1615
+ NVIDIA_TESLA_T4 = "NVIDIA_TESLA_T4",
1616
+ NVIDIA_TESLA_P4 = "NVIDIA_TESLA_P4",
1617
+ NVIDIA_L4 = "NVIDIA_L4"
1618
+ }
1562
1619
  /**
1563
1620
  * Enumerates the types of signaling messages for WebRTC handshakes.
1564
1621
  *
@@ -1667,12 +1724,13 @@ export declare enum TrainingJobStatus {
1667
1724
  RUNNING = "RUNNING",
1668
1725
  COMPLETED = "COMPLETED",
1669
1726
  FAILED = "FAILED",
1670
- CANCELLED = "CANCELLED"
1727
+ CANCELLED = "CANCELLED",
1728
+ CANCELLING = "CANCELLING"
1671
1729
  }
1672
1730
  /**
1673
- * GPU types available for training.
1731
+ * GPU types available for training and deployment.
1674
1732
  */
1675
- export declare enum GPUType1 {
1733
+ export declare enum GPUType2 {
1676
1734
  NVIDIA_H100_80GB = "NVIDIA_H100_80GB",
1677
1735
  NVIDIA_A100_80GB = "NVIDIA_A100_80GB",
1678
1736
  NVIDIA_TESLA_A100 = "NVIDIA_TESLA_A100",
@@ -6,7 +6,7 @@
6
6
  /* Do not modify it by hand - just update the pydantic models and then re-run the script
7
7
  */
8
8
  Object.defineProperty(exports, "__esModule", { value: true });
9
- exports.GPUType1 = exports.TrainingJobStatus = exports.RecordingNotificationType = exports.RecordingDataTraceStatus = exports.Type16 = exports.Type15 = exports.Type14 = exports.RecordingUploadMethod = exports.PendingRecordingStatus = exports.RecordingStatus = exports.Type13 = exports.VideoFormat = exports.Type12 = exports.Type11 = exports.Type10 = exports.Type9 = exports.Type8 = exports.Type7 = exports.Type6 = exports.MessageType = exports.Type5 = exports.Type4 = exports.Type3 = exports.GPUType = exports.DatasetTypeConfig = exports.Type2 = exports.JointPositionTypeConfig = exports.VisualJointTypeConfig = exports.LanguageConfig = exports.AngleConfig1 = exports.EulerOrderConfig = exports.QuaternionOrderConfig = exports.RotationConfig = exports.PoseConfig = exports.DistanceUnitsConfig = exports.TorqueUnitsConfig = exports.AngleConfig = exports.ImageChannelOrderConfig = exports.ImageConventionConfig = exports.Type1 = exports.Type = exports.DataType = void 0;
9
+ exports.GPUType2 = exports.TrainingJobStatus = exports.RecordingNotificationType = exports.RecordingDataTraceStatus = exports.Type16 = exports.Type15 = exports.Type14 = exports.RecordingUploadMethod = exports.PendingRecordingStatus = exports.RecordingStatus = exports.Type13 = exports.VideoFormat = exports.Type12 = exports.Type11 = exports.Type10 = exports.Type9 = exports.Type8 = exports.Type7 = exports.Type6 = exports.MessageType = exports.GPUType1 = exports.EndpointStatus = exports.Type5 = exports.Type4 = exports.Type3 = exports.GPUType = exports.DatasetTypeConfig = exports.Type2 = exports.JointPositionTypeConfig = exports.VisualJointTypeConfig = exports.LanguageConfig = exports.AngleConfig1 = exports.EulerOrderConfig = exports.QuaternionOrderConfig = exports.RotationConfig = exports.PoseConfig = exports.DistanceUnitsConfig = exports.TorqueUnitsConfig = exports.AngleConfig = exports.ImageChannelOrderConfig = exports.ImageConventionConfig = exports.Type1 = exports.Type = exports.DataType = void 0;
10
10
  /**
11
11
  * Enumeration of supported data types in the Neuracore system.
12
12
  *
@@ -162,7 +162,7 @@ var DatasetTypeConfig;
162
162
  DatasetTypeConfig["TFDS"] = "TFDS";
163
163
  })(DatasetTypeConfig || (exports.DatasetTypeConfig = DatasetTypeConfig = {}));
164
164
  /**
165
- * GPU types available for training.
165
+ * GPU types available for training and deployment.
166
166
  */
167
167
  var GPUType;
168
168
  (function (GPUType) {
@@ -187,6 +187,30 @@ var Type5;
187
187
  (function (Type5) {
188
188
  Type5["EndEffectorPoseDataStats"] = "EndEffectorPoseDataStats";
189
189
  })(Type5 || (exports.Type5 = Type5 = {}));
190
+ /**
191
+ * Model endpoint status.
192
+ */
193
+ var EndpointStatus;
194
+ (function (EndpointStatus) {
195
+ EndpointStatus["creating"] = "creating";
196
+ EndpointStatus["active"] = "active";
197
+ EndpointStatus["failed"] = "failed";
198
+ EndpointStatus["deactivated"] = "deactivated";
199
+ })(EndpointStatus || (exports.EndpointStatus = EndpointStatus = {}));
200
+ /**
201
+ * GPU types available for training and deployment.
202
+ */
203
+ var GPUType1;
204
+ (function (GPUType1) {
205
+ GPUType1["NVIDIA_H100_80GB"] = "NVIDIA_H100_80GB";
206
+ GPUType1["NVIDIA_A100_80GB"] = "NVIDIA_A100_80GB";
207
+ GPUType1["NVIDIA_TESLA_A100"] = "NVIDIA_TESLA_A100";
208
+ GPUType1["NVIDIA_TESLA_V100"] = "NVIDIA_TESLA_V100";
209
+ GPUType1["NVIDIA_TESLA_P100"] = "NVIDIA_TESLA_P100";
210
+ GPUType1["NVIDIA_TESLA_T4"] = "NVIDIA_TESLA_T4";
211
+ GPUType1["NVIDIA_TESLA_P4"] = "NVIDIA_TESLA_P4";
212
+ GPUType1["NVIDIA_L4"] = "NVIDIA_L4";
213
+ })(GPUType1 || (exports.GPUType1 = GPUType1 = {}));
190
214
  /**
191
215
  * Enumerates the types of signaling messages for WebRTC handshakes.
192
216
  *
@@ -315,18 +339,19 @@ var TrainingJobStatus;
315
339
  TrainingJobStatus["COMPLETED"] = "COMPLETED";
316
340
  TrainingJobStatus["FAILED"] = "FAILED";
317
341
  TrainingJobStatus["CANCELLED"] = "CANCELLED";
342
+ TrainingJobStatus["CANCELLING"] = "CANCELLING";
318
343
  })(TrainingJobStatus || (exports.TrainingJobStatus = TrainingJobStatus = {}));
319
344
  /**
320
- * GPU types available for training.
345
+ * GPU types available for training and deployment.
321
346
  */
322
- var GPUType1;
323
- (function (GPUType1) {
324
- GPUType1["NVIDIA_H100_80GB"] = "NVIDIA_H100_80GB";
325
- GPUType1["NVIDIA_A100_80GB"] = "NVIDIA_A100_80GB";
326
- GPUType1["NVIDIA_TESLA_A100"] = "NVIDIA_TESLA_A100";
327
- GPUType1["NVIDIA_TESLA_V100"] = "NVIDIA_TESLA_V100";
328
- GPUType1["NVIDIA_TESLA_P100"] = "NVIDIA_TESLA_P100";
329
- GPUType1["NVIDIA_TESLA_T4"] = "NVIDIA_TESLA_T4";
330
- GPUType1["NVIDIA_TESLA_P4"] = "NVIDIA_TESLA_P4";
331
- GPUType1["NVIDIA_L4"] = "NVIDIA_L4";
332
- })(GPUType1 || (exports.GPUType1 = GPUType1 = {}));
347
+ var GPUType2;
348
+ (function (GPUType2) {
349
+ GPUType2["NVIDIA_H100_80GB"] = "NVIDIA_H100_80GB";
350
+ GPUType2["NVIDIA_A100_80GB"] = "NVIDIA_A100_80GB";
351
+ GPUType2["NVIDIA_TESLA_A100"] = "NVIDIA_TESLA_A100";
352
+ GPUType2["NVIDIA_TESLA_V100"] = "NVIDIA_TESLA_V100";
353
+ GPUType2["NVIDIA_TESLA_P100"] = "NVIDIA_TESLA_P100";
354
+ GPUType2["NVIDIA_TESLA_T4"] = "NVIDIA_TESLA_T4";
355
+ GPUType2["NVIDIA_TESLA_P4"] = "NVIDIA_TESLA_P4";
356
+ GPUType2["NVIDIA_L4"] = "NVIDIA_L4";
357
+ })(GPUType2 || (exports.GPUType2 = GPUType2 = {}));
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@neuracore/types",
3
- "version": "6.8.0",
3
+ "version": "6.10.0",
4
4
  "description": "Shared TypeScript type definitions for Neuracore robotics platform",
5
5
  "main": "dist/index.js",
6
6
  "types": "dist/index.d.ts",