gcore 0.4.0__py3-none-any.whl → 0.5.0__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of gcore might be problematic. Click here for more details.

Files changed (125) hide show
  1. gcore/_base_client.py +9 -2
  2. gcore/_client.py +9 -0
  3. gcore/_constants.py +2 -2
  4. gcore/_models.py +8 -5
  5. gcore/_version.py +1 -1
  6. gcore/pagination.py +192 -1
  7. gcore/resources/__init__.py +14 -0
  8. gcore/resources/cloud/baremetal/servers.py +2 -2
  9. gcore/resources/cloud/gpu_baremetal_clusters/gpu_baremetal_clusters.py +2 -2
  10. gcore/resources/cloud/inference/deployments/deployments.py +66 -33
  11. gcore/resources/cloud/inference/deployments/logs.py +7 -7
  12. gcore/resources/cloud/inference/inference.py +5 -5
  13. gcore/resources/cloud/inference/models.py +16 -15
  14. gcore/resources/cloud/inference/registry_credentials.py +16 -16
  15. gcore/resources/cloud/inference/secrets.py +4 -5
  16. gcore/resources/cloud/instances/instances.py +2 -2
  17. gcore/resources/cloud/secrets.py +14 -224
  18. gcore/resources/cloud/tasks.py +30 -28
  19. gcore/resources/fastedge/__init__.py +103 -0
  20. gcore/resources/fastedge/apps/__init__.py +33 -0
  21. gcore/resources/fastedge/apps/apps.py +932 -0
  22. gcore/resources/fastedge/apps/logs.py +248 -0
  23. gcore/resources/fastedge/binaries.py +286 -0
  24. gcore/resources/fastedge/fastedge.py +327 -0
  25. gcore/resources/fastedge/kv_stores.py +523 -0
  26. gcore/resources/fastedge/secrets.py +687 -0
  27. gcore/resources/fastedge/statistics.py +347 -0
  28. gcore/resources/fastedge/templates.py +652 -0
  29. gcore/types/cloud/__init__.py +2 -28
  30. gcore/types/cloud/baremetal/server_create_params.py +1 -1
  31. gcore/types/cloud/gpu_baremetal_cluster_create_params.py +1 -1
  32. gcore/types/cloud/inference/__init__.py +12 -8
  33. gcore/types/cloud/inference/deployment_create_params.py +224 -7
  34. gcore/types/cloud/inference/deployment_update_params.py +24 -3
  35. gcore/types/cloud/inference/deployments/__init__.py +1 -0
  36. gcore/types/cloud/inference/{inference_log.py → deployments/inference_deployment_log.py} +3 -3
  37. gcore/types/cloud/inference/inference_deployment.py +266 -0
  38. gcore/types/cloud/inference/{inference_apikey_secret.py → inference_deployment_api_key.py} +2 -2
  39. gcore/types/cloud/inference/{mlcatalog_model_card.py → inference_model.py} +2 -2
  40. gcore/types/cloud/inference/{inference_registry_credential.py → inference_registry_credentials.py} +2 -2
  41. gcore/types/cloud/inference/{inference_registry_credential_full.py → inference_registry_credentials_create.py} +2 -2
  42. gcore/types/cloud/inference/inference_secret.py +10 -3
  43. gcore/types/cloud/inference/model_list_params.py +2 -4
  44. gcore/types/cloud/{container_probe.py → inference/probe.py} +9 -9
  45. gcore/types/cloud/{container_probe_config.py → inference/probe_config.py} +5 -5
  46. gcore/types/cloud/{container_probe_exec.py → inference/probe_exec.py} +3 -3
  47. gcore/types/cloud/{container_probe_http_get.py → inference/probe_http_get.py} +3 -3
  48. gcore/types/cloud/{container_probe_tcp_socket.py → inference/probe_tcp_socket.py} +3 -3
  49. gcore/types/cloud/inference/secret_create_params.py +10 -4
  50. gcore/types/cloud/inference/secret_replace_params.py +10 -4
  51. gcore/types/cloud/{region_capacity.py → inference_region_capacity.py} +10 -3
  52. gcore/types/cloud/{region_capacity_list.py → inference_region_capacity_list.py} +4 -4
  53. gcore/types/cloud/instance_create_params.py +1 -1
  54. gcore/types/cloud/task_list_params.py +15 -14
  55. gcore/types/fastedge/__init__.py +48 -0
  56. gcore/types/fastedge/app.py +81 -0
  57. gcore/types/fastedge/app_create_params.py +56 -0
  58. gcore/types/fastedge/app_list_params.py +50 -0
  59. gcore/types/fastedge/app_param.py +56 -0
  60. gcore/types/fastedge/app_replace_params.py +17 -0
  61. gcore/types/fastedge/app_short.py +60 -0
  62. gcore/types/fastedge/app_update_params.py +56 -0
  63. gcore/types/fastedge/apps/__init__.py +6 -0
  64. gcore/types/fastedge/apps/log.py +28 -0
  65. gcore/types/fastedge/apps/log_list_params.py +37 -0
  66. gcore/types/fastedge/binary.py +40 -0
  67. gcore/types/fastedge/binary_list_response.py +12 -0
  68. gcore/types/fastedge/binary_short.py +32 -0
  69. gcore/types/fastedge/call_status.py +24 -0
  70. gcore/types/fastedge/client.py +57 -0
  71. gcore/types/fastedge/duration_stats.py +30 -0
  72. gcore/types/fastedge/kv_store.py +33 -0
  73. gcore/types/fastedge/kv_store_create_params.py +23 -0
  74. gcore/types/fastedge/kv_store_get_response.py +10 -0
  75. gcore/types/fastedge/kv_store_list_params.py +12 -0
  76. gcore/types/fastedge/kv_store_list_response.py +15 -0
  77. gcore/types/fastedge/kv_store_replace_params.py +23 -0
  78. gcore/types/fastedge/kv_store_short.py +19 -0
  79. gcore/types/fastedge/kv_store_stats.py +26 -0
  80. gcore/types/fastedge/secret.py +29 -0
  81. gcore/types/fastedge/secret_create_params.py +27 -0
  82. gcore/types/fastedge/secret_create_response.py +12 -0
  83. gcore/types/fastedge/secret_delete_params.py +12 -0
  84. gcore/types/fastedge/secret_list_params.py +15 -0
  85. gcore/types/{cloud → fastedge}/secret_list_response.py +2 -6
  86. gcore/types/fastedge/secret_replace_params.py +27 -0
  87. gcore/types/fastedge/secret_short.py +21 -0
  88. gcore/types/fastedge/secret_update_params.py +27 -0
  89. gcore/types/fastedge/statistic_get_call_series_params.py +28 -0
  90. gcore/types/fastedge/statistic_get_call_series_response.py +12 -0
  91. gcore/types/fastedge/statistic_get_duration_series_params.py +28 -0
  92. gcore/types/fastedge/statistic_get_duration_series_response.py +12 -0
  93. gcore/types/fastedge/template.py +31 -0
  94. gcore/types/fastedge/template_create_params.py +30 -0
  95. gcore/types/fastedge/template_delete_params.py +12 -0
  96. gcore/types/fastedge/template_list_params.py +25 -0
  97. gcore/types/fastedge/template_parameter.py +22 -0
  98. gcore/types/fastedge/template_parameter_param.py +21 -0
  99. gcore/types/fastedge/template_replace_params.py +30 -0
  100. gcore/types/fastedge/template_short.py +27 -0
  101. {gcore-0.4.0.dist-info → gcore-0.5.0.dist-info}/METADATA +7 -5
  102. {gcore-0.4.0.dist-info → gcore-0.5.0.dist-info}/RECORD +104 -69
  103. gcore/types/cloud/aws_iam_data.py +0 -13
  104. gcore/types/cloud/aws_iam_data_param.py +0 -15
  105. gcore/types/cloud/capacity.py +0 -13
  106. gcore/types/cloud/container_probe_config_create_param.py +0 -17
  107. gcore/types/cloud/container_probe_create_param.py +0 -38
  108. gcore/types/cloud/container_probe_exec_create_param.py +0 -13
  109. gcore/types/cloud/container_probe_http_get_create_param.py +0 -25
  110. gcore/types/cloud/container_probe_tcp_socket_create_param.py +0 -12
  111. gcore/types/cloud/container_scale.py +0 -25
  112. gcore/types/cloud/container_scale_trigger_rate.py +0 -13
  113. gcore/types/cloud/container_scale_trigger_sqs.py +0 -33
  114. gcore/types/cloud/container_scale_trigger_threshold.py +0 -10
  115. gcore/types/cloud/container_scale_triggers.py +0 -36
  116. gcore/types/cloud/deploy_status.py +0 -13
  117. gcore/types/cloud/inference/container.py +0 -26
  118. gcore/types/cloud/inference/inference.py +0 -95
  119. gcore/types/cloud/inference/mlcatalog_order_by_choices.py +0 -7
  120. gcore/types/cloud/inference_probes.py +0 -19
  121. gcore/types/cloud/ingress_opts_out.py +0 -16
  122. gcore/types/cloud/ingress_opts_param.py +0 -18
  123. gcore/types/cloud/secret_create_params.py +0 -66
  124. {gcore-0.4.0.dist-info → gcore-0.5.0.dist-info}/WHEEL +0 -0
  125. {gcore-0.4.0.dist-info → gcore-0.5.0.dist-info}/licenses/LICENSE +0 -0
@@ -0,0 +1,266 @@
1
+ # File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.
2
+
3
+ from typing import Dict, List, Optional
4
+ from typing_extensions import Literal
5
+
6
+ from ..logging import Logging
7
+ from ...._models import BaseModel
8
+ from .probe_config import ProbeConfig
9
+
10
+ __all__ = [
11
+ "InferenceDeployment",
12
+ "Container",
13
+ "ContainerDeployStatus",
14
+ "ContainerScale",
15
+ "ContainerScaleTriggers",
16
+ "ContainerScaleTriggersCPU",
17
+ "ContainerScaleTriggersGPUMemory",
18
+ "ContainerScaleTriggersGPUUtilization",
19
+ "ContainerScaleTriggersHTTP",
20
+ "ContainerScaleTriggersMemory",
21
+ "ContainerScaleTriggersSqs",
22
+ "IngressOpts",
23
+ "ObjectReference",
24
+ "Probes",
25
+ ]
26
+
27
+
28
+ class ContainerDeployStatus(BaseModel):
29
+ ready: int
30
+ """Number of ready instances"""
31
+
32
+ total: int
33
+ """Total number of instances"""
34
+
35
+
36
+ class ContainerScaleTriggersCPU(BaseModel):
37
+ threshold: int
38
+ """Threshold value for the trigger in percentage"""
39
+
40
+
41
+ class ContainerScaleTriggersGPUMemory(BaseModel):
42
+ threshold: int
43
+ """Threshold value for the trigger in percentage"""
44
+
45
+
46
+ class ContainerScaleTriggersGPUUtilization(BaseModel):
47
+ threshold: int
48
+ """Threshold value for the trigger in percentage"""
49
+
50
+
51
+ class ContainerScaleTriggersHTTP(BaseModel):
52
+ rate: int
53
+ """Request count per 'window' seconds for the http trigger"""
54
+
55
+ window: int
56
+ """Time window for rate calculation in seconds"""
57
+
58
+
59
+ class ContainerScaleTriggersMemory(BaseModel):
60
+ threshold: int
61
+ """Threshold value for the trigger in percentage"""
62
+
63
+
64
+ class ContainerScaleTriggersSqs(BaseModel):
65
+ activation_queue_length: int
66
+ """Number of messages for activation"""
67
+
68
+ aws_endpoint: Optional[str] = None
69
+ """Custom AWS endpoint"""
70
+
71
+ aws_region: str
72
+ """AWS region"""
73
+
74
+ queue_length: int
75
+ """Number of messages for one replica"""
76
+
77
+ queue_url: str
78
+ """SQS queue URL"""
79
+
80
+ scale_on_delayed: bool
81
+ """Scale on delayed messages"""
82
+
83
+ scale_on_flight: bool
84
+ """Scale on in-flight messages"""
85
+
86
+ secret_name: str
87
+ """Auth secret name"""
88
+
89
+
90
+ class ContainerScaleTriggers(BaseModel):
91
+ cpu: Optional[ContainerScaleTriggersCPU] = None
92
+ """CPU trigger configuration"""
93
+
94
+ gpu_memory: Optional[ContainerScaleTriggersGPUMemory] = None
95
+ """GPU memory trigger configuration.
96
+
97
+ Calculated by `DCGM_FI_DEV_MEM_COPY_UTIL` metric
98
+ """
99
+
100
+ gpu_utilization: Optional[ContainerScaleTriggersGPUUtilization] = None
101
+ """GPU utilization trigger configuration.
102
+
103
+ Calculated by `DCGM_FI_DEV_GPU_UTIL` metric
104
+ """
105
+
106
+ http: Optional[ContainerScaleTriggersHTTP] = None
107
+ """HTTP trigger configuration"""
108
+
109
+ memory: Optional[ContainerScaleTriggersMemory] = None
110
+ """Memory trigger configuration"""
111
+
112
+ sqs: Optional[ContainerScaleTriggersSqs] = None
113
+ """SQS trigger configuration"""
114
+
115
+
116
+ class ContainerScale(BaseModel):
117
+ cooldown_period: Optional[int] = None
118
+ """Cooldown period between scaling actions in seconds"""
119
+
120
+ max: int
121
+ """Maximum scale for the container"""
122
+
123
+ min: int
124
+ """Minimum scale for the container"""
125
+
126
+ polling_interval: Optional[int] = None
127
+ """Polling interval for scaling triggers in seconds"""
128
+
129
+ triggers: ContainerScaleTriggers
130
+ """Triggers for scaling actions"""
131
+
132
+
133
+ class Container(BaseModel):
134
+ address: Optional[str] = None
135
+ """Address of the inference instance"""
136
+
137
+ deploy_status: ContainerDeployStatus
138
+ """Status of the containers deployment"""
139
+
140
+ error_message: Optional[str] = None
141
+ """Error message if the container deployment failed"""
142
+
143
+ region_id: int
144
+ """Region name for the container"""
145
+
146
+ scale: ContainerScale
147
+ """Scale for the container"""
148
+
149
+
150
+ class IngressOpts(BaseModel):
151
+ disable_response_buffering: bool
152
+ """Disable response buffering if true.
153
+
154
+ A client usually has a much slower connection and can not consume the response
155
+ data as fast as it is produced by an upstream application. Ingress tries to
156
+ buffer the whole response in order to release the upstream application as soon
157
+ as possible.By default, the response buffering is enabled.
158
+ """
159
+
160
+
161
+ class ObjectReference(BaseModel):
162
+ kind: Literal["AppDeployment"]
163
+ """Kind of the inference object to be referenced"""
164
+
165
+ name: str
166
+ """Name of the inference object to be referenced"""
167
+
168
+
169
+ class Probes(BaseModel):
170
+ liveness_probe: Optional[ProbeConfig] = None
171
+ """Liveness probe configuration"""
172
+
173
+ readiness_probe: Optional[ProbeConfig] = None
174
+ """Readiness probe configuration"""
175
+
176
+ startup_probe: Optional[ProbeConfig] = None
177
+ """Startup probe configuration"""
178
+
179
+
180
+ class InferenceDeployment(BaseModel):
181
+ address: Optional[str] = None
182
+ """Address of the inference instance"""
183
+
184
+ auth_enabled: bool
185
+ """`true` if instance uses API key authentication.
186
+
187
+ `"Authorization": "Bearer ****\\**"` or `"X-Api-Key": "****\\**"` header is required
188
+ for the requests to the instance if enabled.
189
+ """
190
+
191
+ command: Optional[str] = None
192
+ """Command to be executed when running a container from an image."""
193
+
194
+ containers: List[Container]
195
+ """List of containers for the inference instance"""
196
+
197
+ created_at: Optional[str] = None
198
+ """Inference instance creation date in ISO 8601 format."""
199
+
200
+ credentials_name: str
201
+ """Registry credentials name"""
202
+
203
+ description: str
204
+ """Inference instance description."""
205
+
206
+ envs: Optional[Dict[str, str]] = None
207
+ """Environment variables for the inference instance"""
208
+
209
+ flavor_name: str
210
+ """Flavor name for the inference instance"""
211
+
212
+ image: str
213
+ """Docker image for the inference instance.
214
+
215
+ This field should contain the image name and tag in the format 'name:tag', e.g.,
216
+ 'nginx:latest'. It defaults to Docker Hub as the image registry, but any
217
+ accessible Docker image URL can be specified.
218
+ """
219
+
220
+ ingress_opts: Optional[IngressOpts] = None
221
+ """Ingress options for the inference instance"""
222
+
223
+ listening_port: int
224
+ """Listening port for the inference instance."""
225
+
226
+ logging: Optional[Logging] = None
227
+ """Logging configuration for the inference instance"""
228
+
229
+ name: str
230
+ """Inference instance name."""
231
+
232
+ object_references: List[ObjectReference]
233
+ """Indicates to which parent object this inference belongs to."""
234
+
235
+ probes: Optional[Probes] = None
236
+ """Probes configured for all containers of the inference instance."""
237
+
238
+ project_id: int
239
+ """Project ID. If not provided, your default project ID will be used."""
240
+
241
+ status: Literal["ACTIVE", "DELETING", "DEPLOYING", "DISABLED", "PARTIALLYDEPLOYED", "PENDING"]
242
+ """Inference instance status. Value can be one of the following:
243
+
244
+ - `DEPLOYING` - The instance is being deployed. Containers are not yet created.
245
+ - `PARTIALLYDEPLOYED` - All containers have been created, but some may not be
246
+ ready yet. Instances stuck in this state typically indicate either image being
247
+ pulled, or a failure of some kind. In the latter case, the `error_message`
248
+ field of the respective container object in the `containers` collection
249
+ explains the failure reason.
250
+ - `ACTIVE` - The instance is running and ready to accept requests.
251
+ - `DISABLED` - The instance is disabled and not accepting any requests.
252
+ - `PENDING` - The instance is running but scaled to zero. It will be
253
+ automatically scaled up when a request is made.
254
+ - `DELETING` - The instance is being deleted.
255
+ """
256
+
257
+ timeout: Optional[int] = None
258
+ """
259
+ Specifies the duration in seconds without any requests after which the
260
+ containers will be downscaled to their minimum scale value as defined by
261
+ `scale.min`. If set, this helps in optimizing resource usage by reducing the
262
+ number of container instances during periods of inactivity.
263
+ """
264
+
265
+ api_keys: Optional[List[str]] = None
266
+ """List of API keys for the inference instance"""
@@ -4,10 +4,10 @@ from typing_extensions import Literal
4
4
 
5
5
  from ...._models import BaseModel
6
6
 
7
- __all__ = ["InferenceApikeySecret"]
7
+ __all__ = ["InferenceDeploymentAPIKey"]
8
8
 
9
9
 
10
- class InferenceApikeySecret(BaseModel):
10
+ class InferenceDeploymentAPIKey(BaseModel):
11
11
  secret: str
12
12
  """API key secret"""
13
13
 
@@ -6,10 +6,10 @@ from pydantic import Field as FieldInfo
6
6
 
7
7
  from ...._models import BaseModel
8
8
 
9
- __all__ = ["MlcatalogModelCard"]
9
+ __all__ = ["InferenceModel"]
10
10
 
11
11
 
12
- class MlcatalogModelCard(BaseModel):
12
+ class InferenceModel(BaseModel):
13
13
  id: str
14
14
  """Model ID."""
15
15
 
@@ -2,10 +2,10 @@
2
2
 
3
3
  from ...._models import BaseModel
4
4
 
5
- __all__ = ["InferenceRegistryCredential"]
5
+ __all__ = ["InferenceRegistryCredentials"]
6
6
 
7
7
 
8
- class InferenceRegistryCredential(BaseModel):
8
+ class InferenceRegistryCredentials(BaseModel):
9
9
  name: str
10
10
  """Registry credential name."""
11
11
 
@@ -2,10 +2,10 @@
2
2
 
3
3
  from ...._models import BaseModel
4
4
 
5
- __all__ = ["InferenceRegistryCredentialFull"]
5
+ __all__ = ["InferenceRegistryCredentialsCreate"]
6
6
 
7
7
 
8
- class InferenceRegistryCredentialFull(BaseModel):
8
+ class InferenceRegistryCredentialsCreate(BaseModel):
9
9
  name: str
10
10
  """Registry credential name."""
11
11
 
@@ -1,13 +1,20 @@
1
1
  # File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.
2
2
 
3
3
  from ...._models import BaseModel
4
- from ..aws_iam_data import AwsIamData
5
4
 
6
- __all__ = ["InferenceSecret"]
5
+ __all__ = ["InferenceSecret", "Data"]
6
+
7
+
8
+ class Data(BaseModel):
9
+ aws_access_key_id: str
10
+ """AWS IAM key ID."""
11
+
12
+ aws_secret_access_key: str
13
+ """AWS IAM secret key."""
7
14
 
8
15
 
9
16
  class InferenceSecret(BaseModel):
10
- data: AwsIamData
17
+ data: Data
11
18
  """Secret data."""
12
19
 
13
20
  name: str
@@ -2,9 +2,7 @@
2
2
 
3
3
  from __future__ import annotations
4
4
 
5
- from typing_extensions import TypedDict
6
-
7
- from .mlcatalog_order_by_choices import MlcatalogOrderByChoices
5
+ from typing_extensions import Literal, TypedDict
8
6
 
9
7
  __all__ = ["ModelListParams"]
10
8
 
@@ -19,5 +17,5 @@ class ModelListParams(TypedDict, total=False):
19
17
  Offset value is used to exclude the first set of records from the result
20
18
  """
21
19
 
22
- order_by: MlcatalogOrderByChoices
20
+ order_by: Literal["name.asc", "name.desc"]
23
21
  """Order instances by transmitted fields and directions"""
@@ -2,22 +2,22 @@
2
2
 
3
3
  from typing import Optional
4
4
 
5
- from ..._models import BaseModel
6
- from .container_probe_exec import ContainerProbeExec
7
- from .container_probe_http_get import ContainerProbeHTTPGet
8
- from .container_probe_tcp_socket import ContainerProbeTcpSocket
5
+ from ...._models import BaseModel
6
+ from .probe_exec import ProbeExec
7
+ from .probe_http_get import ProbeHTTPGet
8
+ from .probe_tcp_socket import ProbeTcpSocket
9
9
 
10
- __all__ = ["ContainerProbe"]
10
+ __all__ = ["Probe"]
11
11
 
12
12
 
13
- class ContainerProbe(BaseModel):
14
- exec: Optional[ContainerProbeExec] = None
13
+ class Probe(BaseModel):
14
+ exec: Optional[ProbeExec] = None
15
15
  """Exec probe configuration"""
16
16
 
17
17
  failure_threshold: int
18
18
  """The number of consecutive probe failures that mark the container as unhealthy."""
19
19
 
20
- http_get: Optional[ContainerProbeHTTPGet] = None
20
+ http_get: Optional[ProbeHTTPGet] = None
21
21
  """HTTP GET probe configuration"""
22
22
 
23
23
  initial_delay_seconds: int
@@ -29,7 +29,7 @@ class ContainerProbe(BaseModel):
29
29
  success_threshold: int
30
30
  """The number of consecutive successful probes that mark the container as healthy."""
31
31
 
32
- tcp_socket: Optional[ContainerProbeTcpSocket] = None
32
+ tcp_socket: Optional[ProbeTcpSocket] = None
33
33
  """TCP socket probe configuration"""
34
34
 
35
35
  timeout_seconds: int
@@ -2,15 +2,15 @@
2
2
 
3
3
  from typing import Optional
4
4
 
5
- from ..._models import BaseModel
6
- from .container_probe import ContainerProbe
5
+ from .probe import Probe
6
+ from ...._models import BaseModel
7
7
 
8
- __all__ = ["ContainerProbeConfig"]
8
+ __all__ = ["ProbeConfig"]
9
9
 
10
10
 
11
- class ContainerProbeConfig(BaseModel):
11
+ class ProbeConfig(BaseModel):
12
12
  enabled: bool
13
13
  """Whether the probe is enabled or not."""
14
14
 
15
- probe: Optional[ContainerProbe] = None
15
+ probe: Optional[Probe] = None
16
16
  """Probe configuration (exec, `http_get` or `tcp_socket`)"""
@@ -2,11 +2,11 @@
2
2
 
3
3
  from typing import List
4
4
 
5
- from ..._models import BaseModel
5
+ from ...._models import BaseModel
6
6
 
7
- __all__ = ["ContainerProbeExec"]
7
+ __all__ = ["ProbeExec"]
8
8
 
9
9
 
10
- class ContainerProbeExec(BaseModel):
10
+ class ProbeExec(BaseModel):
11
11
  command: List[str]
12
12
  """Command to be executed inside the running container."""
@@ -4,12 +4,12 @@ from typing import Dict, Optional
4
4
 
5
5
  from pydantic import Field as FieldInfo
6
6
 
7
- from ..._models import BaseModel
7
+ from ...._models import BaseModel
8
8
 
9
- __all__ = ["ContainerProbeHTTPGet"]
9
+ __all__ = ["ProbeHTTPGet"]
10
10
 
11
11
 
12
- class ContainerProbeHTTPGet(BaseModel):
12
+ class ProbeHTTPGet(BaseModel):
13
13
  headers: Dict[str, str]
14
14
  """HTTP headers to be sent with the request."""
15
15
 
@@ -1,10 +1,10 @@
1
1
  # File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.
2
2
 
3
- from ..._models import BaseModel
3
+ from ...._models import BaseModel
4
4
 
5
- __all__ = ["ContainerProbeTcpSocket"]
5
+ __all__ = ["ProbeTcpSocket"]
6
6
 
7
7
 
8
- class ContainerProbeTcpSocket(BaseModel):
8
+ class ProbeTcpSocket(BaseModel):
9
9
  port: int
10
10
  """Port number to check if it's open."""
@@ -4,16 +4,14 @@ from __future__ import annotations
4
4
 
5
5
  from typing_extensions import Required, TypedDict
6
6
 
7
- from ..aws_iam_data_param import AwsIamDataParam
8
-
9
- __all__ = ["SecretCreateParams"]
7
+ __all__ = ["SecretCreateParams", "Data"]
10
8
 
11
9
 
12
10
  class SecretCreateParams(TypedDict, total=False):
13
11
  project_id: int
14
12
  """Project ID"""
15
13
 
16
- data: Required[AwsIamDataParam]
14
+ data: Required[Data]
17
15
  """Secret data."""
18
16
 
19
17
  name: Required[str]
@@ -21,3 +19,11 @@ class SecretCreateParams(TypedDict, total=False):
21
19
 
22
20
  type: Required[str]
23
21
  """Secret type. Currently only `aws-iam` is supported."""
22
+
23
+
24
+ class Data(TypedDict, total=False):
25
+ aws_access_key_id: Required[str]
26
+ """AWS IAM key ID."""
27
+
28
+ aws_secret_access_key: Required[str]
29
+ """AWS IAM secret key."""
@@ -4,17 +4,23 @@ from __future__ import annotations
4
4
 
5
5
  from typing_extensions import Required, TypedDict
6
6
 
7
- from ..aws_iam_data_param import AwsIamDataParam
8
-
9
- __all__ = ["SecretReplaceParams"]
7
+ __all__ = ["SecretReplaceParams", "Data"]
10
8
 
11
9
 
12
10
  class SecretReplaceParams(TypedDict, total=False):
13
11
  project_id: int
14
12
  """Project ID"""
15
13
 
16
- data: Required[AwsIamDataParam]
14
+ data: Required[Data]
17
15
  """Secret data."""
18
16
 
19
17
  type: Required[str]
20
18
  """Secret type."""
19
+
20
+
21
+ class Data(TypedDict, total=False):
22
+ aws_access_key_id: Required[str]
23
+ """AWS IAM key ID."""
24
+
25
+ aws_secret_access_key: Required[str]
26
+ """AWS IAM secret key."""
@@ -2,13 +2,20 @@
2
2
 
3
3
  from typing import List
4
4
 
5
- from .capacity import Capacity
6
5
  from ..._models import BaseModel
7
6
 
8
- __all__ = ["RegionCapacity"]
7
+ __all__ = ["InferenceRegionCapacity", "Capacity"]
9
8
 
10
9
 
11
- class RegionCapacity(BaseModel):
10
+ class Capacity(BaseModel):
11
+ capacity: int
12
+ """Available capacity."""
13
+
14
+ flavor_name: str
15
+ """Flavor name."""
16
+
17
+
18
+ class InferenceRegionCapacity(BaseModel):
12
19
  capacity: List[Capacity]
13
20
  """List of capacities by flavor."""
14
21
 
@@ -3,14 +3,14 @@
3
3
  from typing import List
4
4
 
5
5
  from ..._models import BaseModel
6
- from .region_capacity import RegionCapacity
6
+ from .inference_region_capacity import InferenceRegionCapacity
7
7
 
8
- __all__ = ["RegionCapacityList"]
8
+ __all__ = ["InferenceRegionCapacityList"]
9
9
 
10
10
 
11
- class RegionCapacityList(BaseModel):
11
+ class InferenceRegionCapacityList(BaseModel):
12
12
  count: int
13
13
  """Number of objects"""
14
14
 
15
- results: List[RegionCapacity]
15
+ results: List[InferenceRegionCapacity]
16
16
  """Objects"""
@@ -109,7 +109,7 @@ class InstanceCreateParams(TypedDict, total=False):
109
109
  ssh_key_name: Optional[str]
110
110
  """
111
111
  Specifies the name of the SSH keypair, created via the
112
- [/v1/`ssh_keys` endpoint](/docs/api-reference/ssh-keys/add-or-generate-ssh-key).
112
+ [/v1/`ssh_keys` endpoint](/docs/api-reference/cloud/ssh-keys/add-or-generate-ssh-key).
113
113
  """
114
114
 
115
115
  tags: Dict[str, str]
@@ -65,7 +65,8 @@ class TaskListParams(TypedDict, total=False):
65
65
  '`create_ai_cluster_gpu`', '`create_bm`', '`create_caas_container`',
66
66
  '`create_dbaas_postgres_cluster`', '`create_ddos_profile`',
67
67
  '`create_faas_function`', '`create_faas_namespace`', '`create_fip`',
68
- '`create_gpu_virtual_cluster`', '`create_image`', '`create_inference_instance`',
68
+ '`create_gpu_virtual_cluster`', '`create_image`',
69
+ '`create_inference_application`', '`create_inference_instance`',
69
70
  '`create_inference_instance_key`', '`create_k8s_cluster_pool_v2`',
70
71
  '`create_k8s_cluster_v2`', '`create_l7policy`', '`create_l7rule`',
71
72
  '`create_lblistener`', '`create_lbmember`', '`create_lbpool`',
@@ -77,14 +78,14 @@ class TaskListParams(TypedDict, total=False):
77
78
  '`delete_dbaas_postgres_cluster`', '`delete_ddos_profile`',
78
79
  '`delete_faas_function`', '`delete_faas_namespace`', '`delete_fip`',
79
80
  '`delete_gpu_virtual_cluster`', '`delete_gpu_virtual_server`', '`delete_image`',
80
- '`delete_inference_instance`', '`delete_k8s_cluster_pool_v2`',
81
- '`delete_k8s_cluster_v2`', '`delete_l7policy`', '`delete_l7rule`',
82
- '`delete_lblistener`', '`delete_lbmember`', '`delete_lbmetadata`',
83
- '`delete_lbpool`', '`delete_loadbalancer`', '`delete_network`',
84
- '`delete_reserved_fixed_ip`', '`delete_router`', '`delete_secret`',
85
- '`delete_servergroup`', '`delete_sfs`', '`delete_snapshot`', '`delete_subnet`',
86
- '`delete_vm`', '`delete_volume`', '`detach_vm_interface`', '`detach_volume`',
87
- '`download_image`', '`downscale_ai_cluster_gpu`',
81
+ '`delete_inference_application`', '`delete_inference_instance`',
82
+ '`delete_k8s_cluster_pool_v2`', '`delete_k8s_cluster_v2`', '`delete_l7policy`',
83
+ '`delete_l7rule`', '`delete_lblistener`', '`delete_lbmember`',
84
+ '`delete_lbmetadata`', '`delete_lbpool`', '`delete_loadbalancer`',
85
+ '`delete_network`', '`delete_reserved_fixed_ip`', '`delete_router`',
86
+ '`delete_secret`', '`delete_servergroup`', '`delete_sfs`', '`delete_snapshot`',
87
+ '`delete_subnet`', '`delete_vm`', '`delete_volume`', '`detach_vm_interface`',
88
+ '`detach_volume`', '`download_image`', '`downscale_ai_cluster_gpu`',
88
89
  '`downscale_gpu_virtual_cluster`', '`extend_sfs`', '`extend_volume`',
89
90
  '`failover_loadbalancer`', '`hard_reboot_gpu_baremetal_server`',
90
91
  '`hard_reboot_gpu_virtual_cluster`', '`hard_reboot_gpu_virtual_server`',
@@ -100,11 +101,11 @@ class TaskListParams(TypedDict, total=False):
100
101
  '`start_gpu_virtual_server`', '`start_vm`', '`stop_gpu_baremetal_server`',
101
102
  '`stop_gpu_virtual_cluster`', '`stop_gpu_virtual_server`', '`stop_vm`',
102
103
  '`suspend_vm`', '`sync_private_flavors`', '`update_ddos_profile`',
103
- '`update_inference_instance`', '`update_inference_instance_key`',
104
- '`update_k8s_cluster_v2`', '`update_lbmetadata`',
105
- '`update_port_allowed_address_pairs`', '`update_tags_gpu_virtual_cluster`',
106
- '`upgrade_k8s_cluster_v2`', '`upscale_ai_cluster_gpu`',
107
- '`upscale_gpu_virtual_cluster`']
104
+ '`update_inference_application`', '`update_inference_instance`',
105
+ '`update_inference_instance_key`', '`update_k8s_cluster_v2`',
106
+ '`update_lbmetadata`', '`update_port_allowed_address_pairs`',
107
+ '`update_tags_gpu_virtual_cluster`', '`upgrade_k8s_cluster_v2`',
108
+ '`upscale_ai_cluster_gpu`', '`upscale_gpu_virtual_cluster`']
108
109
  """
109
110
 
110
111
  to_timestamp: Annotated[Union[str, datetime, None], PropertyInfo(format="iso8601")]