gcore 0.4.0__py3-none-any.whl → 0.6.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of gcore might be problematic. Click here for more details.
- gcore/_base_client.py +9 -2
- gcore/_client.py +9 -0
- gcore/_constants.py +2 -2
- gcore/_models.py +8 -5
- gcore/_version.py +1 -1
- gcore/pagination.py +192 -1
- gcore/resources/__init__.py +14 -0
- gcore/resources/cloud/__init__.py +14 -0
- gcore/resources/cloud/audit_logs.py +480 -0
- gcore/resources/cloud/baremetal/servers.py +2 -2
- gcore/resources/cloud/cloud.py +32 -0
- gcore/resources/cloud/file_shares/file_shares.py +64 -9
- gcore/resources/cloud/gpu_baremetal_clusters/gpu_baremetal_clusters.py +2 -2
- gcore/resources/cloud/inference/__init__.py +14 -0
- gcore/resources/cloud/inference/api_keys.py +621 -0
- gcore/resources/cloud/inference/deployments/deployments.py +66 -33
- gcore/resources/cloud/inference/deployments/logs.py +7 -7
- gcore/resources/cloud/inference/inference.py +37 -5
- gcore/resources/cloud/inference/models.py +16 -15
- gcore/resources/cloud/inference/registry_credentials.py +16 -16
- gcore/resources/cloud/inference/secrets.py +4 -5
- gcore/resources/cloud/instances/instances.py +2 -2
- gcore/resources/cloud/load_balancers/pools/members.py +22 -6
- gcore/resources/cloud/secrets.py +14 -224
- gcore/resources/cloud/tasks.py +30 -28
- gcore/resources/fastedge/__init__.py +103 -0
- gcore/resources/fastedge/apps/__init__.py +33 -0
- gcore/resources/fastedge/apps/apps.py +932 -0
- gcore/resources/fastedge/apps/logs.py +248 -0
- gcore/resources/fastedge/binaries.py +286 -0
- gcore/resources/fastedge/fastedge.py +327 -0
- gcore/resources/fastedge/kv_stores.py +523 -0
- gcore/resources/fastedge/secrets.py +687 -0
- gcore/resources/fastedge/statistics.py +347 -0
- gcore/resources/fastedge/templates.py +652 -0
- gcore/types/cloud/__init__.py +4 -28
- gcore/types/cloud/audit_log_entry.py +254 -0
- gcore/types/cloud/audit_log_list_params.py +158 -0
- gcore/types/cloud/baremetal/server_create_params.py +1 -1
- gcore/types/cloud/file_share_update_params.py +29 -3
- gcore/types/cloud/gpu_baremetal_cluster_create_params.py +1 -1
- gcore/types/cloud/inference/__init__.py +17 -8
- gcore/types/cloud/inference/api_key_create_params.py +21 -0
- gcore/types/cloud/inference/api_key_list_params.py +21 -0
- gcore/types/cloud/inference/api_key_update_params.py +16 -0
- gcore/types/cloud/inference/deployment_create_params.py +224 -7
- gcore/types/cloud/inference/deployment_update_params.py +24 -3
- gcore/types/cloud/inference/deployments/__init__.py +1 -0
- gcore/types/cloud/inference/{inference_log.py → deployments/inference_deployment_log.py} +3 -3
- gcore/types/cloud/inference/inference_api_key.py +24 -0
- gcore/types/cloud/inference/inference_api_key_create.py +27 -0
- gcore/types/cloud/inference/inference_deployment.py +266 -0
- gcore/types/cloud/inference/{inference_apikey_secret.py → inference_deployment_api_key.py} +2 -2
- gcore/types/cloud/inference/{mlcatalog_model_card.py → inference_model.py} +2 -2
- gcore/types/cloud/inference/{inference_registry_credential.py → inference_registry_credentials.py} +2 -2
- gcore/types/cloud/inference/{inference_registry_credential_full.py → inference_registry_credentials_create.py} +2 -2
- gcore/types/cloud/inference/inference_secret.py +10 -3
- gcore/types/cloud/inference/model_list_params.py +2 -4
- gcore/types/cloud/{container_probe.py → inference/probe.py} +9 -9
- gcore/types/cloud/{container_probe_config.py → inference/probe_config.py} +5 -5
- gcore/types/cloud/{container_probe_exec.py → inference/probe_exec.py} +3 -3
- gcore/types/cloud/{container_probe_http_get.py → inference/probe_http_get.py} +3 -3
- gcore/types/cloud/{container_probe_tcp_socket.py → inference/probe_tcp_socket.py} +3 -3
- gcore/types/cloud/inference/secret_create_params.py +10 -4
- gcore/types/cloud/inference/secret_replace_params.py +10 -4
- gcore/types/cloud/{region_capacity.py → inference_region_capacity.py} +10 -3
- gcore/types/cloud/{region_capacity_list.py → inference_region_capacity_list.py} +4 -4
- gcore/types/cloud/instance_create_params.py +1 -1
- gcore/types/cloud/load_balancer_create_params.py +14 -3
- gcore/types/cloud/load_balancers/pool_create_params.py +14 -3
- gcore/types/cloud/load_balancers/pool_update_params.py +14 -3
- gcore/types/cloud/load_balancers/pools/member_add_params.py +14 -3
- gcore/types/cloud/member.py +12 -4
- gcore/types/cloud/task_list_params.py +15 -14
- gcore/types/fastedge/__init__.py +48 -0
- gcore/types/fastedge/app.py +81 -0
- gcore/types/fastedge/app_create_params.py +56 -0
- gcore/types/fastedge/app_list_params.py +50 -0
- gcore/types/fastedge/app_param.py +56 -0
- gcore/types/fastedge/app_replace_params.py +17 -0
- gcore/types/fastedge/app_short.py +60 -0
- gcore/types/fastedge/app_update_params.py +56 -0
- gcore/types/fastedge/apps/__init__.py +6 -0
- gcore/types/fastedge/apps/log.py +28 -0
- gcore/types/fastedge/apps/log_list_params.py +37 -0
- gcore/types/fastedge/binary.py +40 -0
- gcore/types/fastedge/binary_list_response.py +12 -0
- gcore/types/fastedge/binary_short.py +32 -0
- gcore/types/fastedge/call_status.py +24 -0
- gcore/types/fastedge/client.py +57 -0
- gcore/types/fastedge/duration_stats.py +30 -0
- gcore/types/fastedge/kv_store.py +33 -0
- gcore/types/fastedge/kv_store_create_params.py +23 -0
- gcore/types/fastedge/kv_store_get_response.py +10 -0
- gcore/types/fastedge/kv_store_list_params.py +12 -0
- gcore/types/fastedge/kv_store_list_response.py +15 -0
- gcore/types/fastedge/kv_store_replace_params.py +23 -0
- gcore/types/fastedge/kv_store_short.py +19 -0
- gcore/types/fastedge/kv_store_stats.py +26 -0
- gcore/types/fastedge/secret.py +29 -0
- gcore/types/fastedge/secret_create_params.py +27 -0
- gcore/types/fastedge/secret_create_response.py +12 -0
- gcore/types/fastedge/secret_delete_params.py +12 -0
- gcore/types/fastedge/secret_list_params.py +15 -0
- gcore/types/{cloud → fastedge}/secret_list_response.py +2 -6
- gcore/types/fastedge/secret_replace_params.py +27 -0
- gcore/types/fastedge/secret_short.py +21 -0
- gcore/types/fastedge/secret_update_params.py +27 -0
- gcore/types/fastedge/statistic_get_call_series_params.py +28 -0
- gcore/types/fastedge/statistic_get_call_series_response.py +12 -0
- gcore/types/fastedge/statistic_get_duration_series_params.py +28 -0
- gcore/types/fastedge/statistic_get_duration_series_response.py +12 -0
- gcore/types/fastedge/template.py +31 -0
- gcore/types/fastedge/template_create_params.py +30 -0
- gcore/types/fastedge/template_delete_params.py +12 -0
- gcore/types/fastedge/template_list_params.py +25 -0
- gcore/types/fastedge/template_parameter.py +22 -0
- gcore/types/fastedge/template_parameter_param.py +21 -0
- gcore/types/fastedge/template_replace_params.py +30 -0
- gcore/types/fastedge/template_short.py +27 -0
- {gcore-0.4.0.dist-info → gcore-0.6.0.dist-info}/METADATA +8 -7
- {gcore-0.4.0.dist-info → gcore-0.6.0.dist-info}/RECORD +124 -80
- gcore/types/cloud/aws_iam_data.py +0 -13
- gcore/types/cloud/aws_iam_data_param.py +0 -15
- gcore/types/cloud/capacity.py +0 -13
- gcore/types/cloud/container_probe_config_create_param.py +0 -17
- gcore/types/cloud/container_probe_create_param.py +0 -38
- gcore/types/cloud/container_probe_exec_create_param.py +0 -13
- gcore/types/cloud/container_probe_http_get_create_param.py +0 -25
- gcore/types/cloud/container_probe_tcp_socket_create_param.py +0 -12
- gcore/types/cloud/container_scale.py +0 -25
- gcore/types/cloud/container_scale_trigger_rate.py +0 -13
- gcore/types/cloud/container_scale_trigger_sqs.py +0 -33
- gcore/types/cloud/container_scale_trigger_threshold.py +0 -10
- gcore/types/cloud/container_scale_triggers.py +0 -36
- gcore/types/cloud/deploy_status.py +0 -13
- gcore/types/cloud/inference/container.py +0 -26
- gcore/types/cloud/inference/inference.py +0 -95
- gcore/types/cloud/inference/mlcatalog_order_by_choices.py +0 -7
- gcore/types/cloud/inference_probes.py +0 -19
- gcore/types/cloud/ingress_opts_out.py +0 -16
- gcore/types/cloud/ingress_opts_param.py +0 -18
- gcore/types/cloud/secret_create_params.py +0 -66
- {gcore-0.4.0.dist-info → gcore-0.6.0.dist-info}/WHEEL +0 -0
- {gcore-0.4.0.dist-info → gcore-0.6.0.dist-info}/licenses/LICENSE +0 -0
|
@@ -0,0 +1,266 @@
|
|
|
1
|
+
# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.
|
|
2
|
+
|
|
3
|
+
from typing import Dict, List, Optional
|
|
4
|
+
from typing_extensions import Literal
|
|
5
|
+
|
|
6
|
+
from ..logging import Logging
|
|
7
|
+
from ...._models import BaseModel
|
|
8
|
+
from .probe_config import ProbeConfig
|
|
9
|
+
|
|
10
|
+
__all__ = [
|
|
11
|
+
"InferenceDeployment",
|
|
12
|
+
"Container",
|
|
13
|
+
"ContainerDeployStatus",
|
|
14
|
+
"ContainerScale",
|
|
15
|
+
"ContainerScaleTriggers",
|
|
16
|
+
"ContainerScaleTriggersCPU",
|
|
17
|
+
"ContainerScaleTriggersGPUMemory",
|
|
18
|
+
"ContainerScaleTriggersGPUUtilization",
|
|
19
|
+
"ContainerScaleTriggersHTTP",
|
|
20
|
+
"ContainerScaleTriggersMemory",
|
|
21
|
+
"ContainerScaleTriggersSqs",
|
|
22
|
+
"IngressOpts",
|
|
23
|
+
"ObjectReference",
|
|
24
|
+
"Probes",
|
|
25
|
+
]
|
|
26
|
+
|
|
27
|
+
|
|
28
|
+
class ContainerDeployStatus(BaseModel):
|
|
29
|
+
ready: int
|
|
30
|
+
"""Number of ready instances"""
|
|
31
|
+
|
|
32
|
+
total: int
|
|
33
|
+
"""Total number of instances"""
|
|
34
|
+
|
|
35
|
+
|
|
36
|
+
class ContainerScaleTriggersCPU(BaseModel):
|
|
37
|
+
threshold: int
|
|
38
|
+
"""Threshold value for the trigger in percentage"""
|
|
39
|
+
|
|
40
|
+
|
|
41
|
+
class ContainerScaleTriggersGPUMemory(BaseModel):
|
|
42
|
+
threshold: int
|
|
43
|
+
"""Threshold value for the trigger in percentage"""
|
|
44
|
+
|
|
45
|
+
|
|
46
|
+
class ContainerScaleTriggersGPUUtilization(BaseModel):
|
|
47
|
+
threshold: int
|
|
48
|
+
"""Threshold value for the trigger in percentage"""
|
|
49
|
+
|
|
50
|
+
|
|
51
|
+
class ContainerScaleTriggersHTTP(BaseModel):
|
|
52
|
+
rate: int
|
|
53
|
+
"""Request count per 'window' seconds for the http trigger"""
|
|
54
|
+
|
|
55
|
+
window: int
|
|
56
|
+
"""Time window for rate calculation in seconds"""
|
|
57
|
+
|
|
58
|
+
|
|
59
|
+
class ContainerScaleTriggersMemory(BaseModel):
|
|
60
|
+
threshold: int
|
|
61
|
+
"""Threshold value for the trigger in percentage"""
|
|
62
|
+
|
|
63
|
+
|
|
64
|
+
class ContainerScaleTriggersSqs(BaseModel):
|
|
65
|
+
activation_queue_length: int
|
|
66
|
+
"""Number of messages for activation"""
|
|
67
|
+
|
|
68
|
+
aws_endpoint: Optional[str] = None
|
|
69
|
+
"""Custom AWS endpoint"""
|
|
70
|
+
|
|
71
|
+
aws_region: str
|
|
72
|
+
"""AWS region"""
|
|
73
|
+
|
|
74
|
+
queue_length: int
|
|
75
|
+
"""Number of messages for one replica"""
|
|
76
|
+
|
|
77
|
+
queue_url: str
|
|
78
|
+
"""SQS queue URL"""
|
|
79
|
+
|
|
80
|
+
scale_on_delayed: bool
|
|
81
|
+
"""Scale on delayed messages"""
|
|
82
|
+
|
|
83
|
+
scale_on_flight: bool
|
|
84
|
+
"""Scale on in-flight messages"""
|
|
85
|
+
|
|
86
|
+
secret_name: str
|
|
87
|
+
"""Auth secret name"""
|
|
88
|
+
|
|
89
|
+
|
|
90
|
+
class ContainerScaleTriggers(BaseModel):
|
|
91
|
+
cpu: Optional[ContainerScaleTriggersCPU] = None
|
|
92
|
+
"""CPU trigger configuration"""
|
|
93
|
+
|
|
94
|
+
gpu_memory: Optional[ContainerScaleTriggersGPUMemory] = None
|
|
95
|
+
"""GPU memory trigger configuration.
|
|
96
|
+
|
|
97
|
+
Calculated by `DCGM_FI_DEV_MEM_COPY_UTIL` metric
|
|
98
|
+
"""
|
|
99
|
+
|
|
100
|
+
gpu_utilization: Optional[ContainerScaleTriggersGPUUtilization] = None
|
|
101
|
+
"""GPU utilization trigger configuration.
|
|
102
|
+
|
|
103
|
+
Calculated by `DCGM_FI_DEV_GPU_UTIL` metric
|
|
104
|
+
"""
|
|
105
|
+
|
|
106
|
+
http: Optional[ContainerScaleTriggersHTTP] = None
|
|
107
|
+
"""HTTP trigger configuration"""
|
|
108
|
+
|
|
109
|
+
memory: Optional[ContainerScaleTriggersMemory] = None
|
|
110
|
+
"""Memory trigger configuration"""
|
|
111
|
+
|
|
112
|
+
sqs: Optional[ContainerScaleTriggersSqs] = None
|
|
113
|
+
"""SQS trigger configuration"""
|
|
114
|
+
|
|
115
|
+
|
|
116
|
+
class ContainerScale(BaseModel):
|
|
117
|
+
cooldown_period: Optional[int] = None
|
|
118
|
+
"""Cooldown period between scaling actions in seconds"""
|
|
119
|
+
|
|
120
|
+
max: int
|
|
121
|
+
"""Maximum scale for the container"""
|
|
122
|
+
|
|
123
|
+
min: int
|
|
124
|
+
"""Minimum scale for the container"""
|
|
125
|
+
|
|
126
|
+
polling_interval: Optional[int] = None
|
|
127
|
+
"""Polling interval for scaling triggers in seconds"""
|
|
128
|
+
|
|
129
|
+
triggers: ContainerScaleTriggers
|
|
130
|
+
"""Triggers for scaling actions"""
|
|
131
|
+
|
|
132
|
+
|
|
133
|
+
class Container(BaseModel):
|
|
134
|
+
address: Optional[str] = None
|
|
135
|
+
"""Address of the inference instance"""
|
|
136
|
+
|
|
137
|
+
deploy_status: ContainerDeployStatus
|
|
138
|
+
"""Status of the containers deployment"""
|
|
139
|
+
|
|
140
|
+
error_message: Optional[str] = None
|
|
141
|
+
"""Error message if the container deployment failed"""
|
|
142
|
+
|
|
143
|
+
region_id: int
|
|
144
|
+
"""Region name for the container"""
|
|
145
|
+
|
|
146
|
+
scale: ContainerScale
|
|
147
|
+
"""Scale for the container"""
|
|
148
|
+
|
|
149
|
+
|
|
150
|
+
class IngressOpts(BaseModel):
|
|
151
|
+
disable_response_buffering: bool
|
|
152
|
+
"""Disable response buffering if true.
|
|
153
|
+
|
|
154
|
+
A client usually has a much slower connection and can not consume the response
|
|
155
|
+
data as fast as it is produced by an upstream application. Ingress tries to
|
|
156
|
+
buffer the whole response in order to release the upstream application as soon
|
|
157
|
+
as possible.By default, the response buffering is enabled.
|
|
158
|
+
"""
|
|
159
|
+
|
|
160
|
+
|
|
161
|
+
class ObjectReference(BaseModel):
|
|
162
|
+
kind: Literal["AppDeployment"]
|
|
163
|
+
"""Kind of the inference object to be referenced"""
|
|
164
|
+
|
|
165
|
+
name: str
|
|
166
|
+
"""Name of the inference object to be referenced"""
|
|
167
|
+
|
|
168
|
+
|
|
169
|
+
class Probes(BaseModel):
|
|
170
|
+
liveness_probe: Optional[ProbeConfig] = None
|
|
171
|
+
"""Liveness probe configuration"""
|
|
172
|
+
|
|
173
|
+
readiness_probe: Optional[ProbeConfig] = None
|
|
174
|
+
"""Readiness probe configuration"""
|
|
175
|
+
|
|
176
|
+
startup_probe: Optional[ProbeConfig] = None
|
|
177
|
+
"""Startup probe configuration"""
|
|
178
|
+
|
|
179
|
+
|
|
180
|
+
class InferenceDeployment(BaseModel):
|
|
181
|
+
address: Optional[str] = None
|
|
182
|
+
"""Address of the inference instance"""
|
|
183
|
+
|
|
184
|
+
auth_enabled: bool
|
|
185
|
+
"""`true` if instance uses API key authentication.
|
|
186
|
+
|
|
187
|
+
`"Authorization": "Bearer ****\\**"` or `"X-Api-Key": "****\\**"` header is required
|
|
188
|
+
for the requests to the instance if enabled.
|
|
189
|
+
"""
|
|
190
|
+
|
|
191
|
+
command: Optional[str] = None
|
|
192
|
+
"""Command to be executed when running a container from an image."""
|
|
193
|
+
|
|
194
|
+
containers: List[Container]
|
|
195
|
+
"""List of containers for the inference instance"""
|
|
196
|
+
|
|
197
|
+
created_at: Optional[str] = None
|
|
198
|
+
"""Inference instance creation date in ISO 8601 format."""
|
|
199
|
+
|
|
200
|
+
credentials_name: str
|
|
201
|
+
"""Registry credentials name"""
|
|
202
|
+
|
|
203
|
+
description: str
|
|
204
|
+
"""Inference instance description."""
|
|
205
|
+
|
|
206
|
+
envs: Optional[Dict[str, str]] = None
|
|
207
|
+
"""Environment variables for the inference instance"""
|
|
208
|
+
|
|
209
|
+
flavor_name: str
|
|
210
|
+
"""Flavor name for the inference instance"""
|
|
211
|
+
|
|
212
|
+
image: str
|
|
213
|
+
"""Docker image for the inference instance.
|
|
214
|
+
|
|
215
|
+
This field should contain the image name and tag in the format 'name:tag', e.g.,
|
|
216
|
+
'nginx:latest'. It defaults to Docker Hub as the image registry, but any
|
|
217
|
+
accessible Docker image URL can be specified.
|
|
218
|
+
"""
|
|
219
|
+
|
|
220
|
+
ingress_opts: Optional[IngressOpts] = None
|
|
221
|
+
"""Ingress options for the inference instance"""
|
|
222
|
+
|
|
223
|
+
listening_port: int
|
|
224
|
+
"""Listening port for the inference instance."""
|
|
225
|
+
|
|
226
|
+
logging: Optional[Logging] = None
|
|
227
|
+
"""Logging configuration for the inference instance"""
|
|
228
|
+
|
|
229
|
+
name: str
|
|
230
|
+
"""Inference instance name."""
|
|
231
|
+
|
|
232
|
+
object_references: List[ObjectReference]
|
|
233
|
+
"""Indicates to which parent object this inference belongs to."""
|
|
234
|
+
|
|
235
|
+
probes: Optional[Probes] = None
|
|
236
|
+
"""Probes configured for all containers of the inference instance."""
|
|
237
|
+
|
|
238
|
+
project_id: int
|
|
239
|
+
"""Project ID. If not provided, your default project ID will be used."""
|
|
240
|
+
|
|
241
|
+
status: Literal["ACTIVE", "DELETING", "DEPLOYING", "DISABLED", "PARTIALLYDEPLOYED", "PENDING"]
|
|
242
|
+
"""Inference instance status. Value can be one of the following:
|
|
243
|
+
|
|
244
|
+
- `DEPLOYING` - The instance is being deployed. Containers are not yet created.
|
|
245
|
+
- `PARTIALLYDEPLOYED` - All containers have been created, but some may not be
|
|
246
|
+
ready yet. Instances stuck in this state typically indicate either image being
|
|
247
|
+
pulled, or a failure of some kind. In the latter case, the `error_message`
|
|
248
|
+
field of the respective container object in the `containers` collection
|
|
249
|
+
explains the failure reason.
|
|
250
|
+
- `ACTIVE` - The instance is running and ready to accept requests.
|
|
251
|
+
- `DISABLED` - The instance is disabled and not accepting any requests.
|
|
252
|
+
- `PENDING` - The instance is running but scaled to zero. It will be
|
|
253
|
+
automatically scaled up when a request is made.
|
|
254
|
+
- `DELETING` - The instance is being deleted.
|
|
255
|
+
"""
|
|
256
|
+
|
|
257
|
+
timeout: Optional[int] = None
|
|
258
|
+
"""
|
|
259
|
+
Specifies the duration in seconds without any requests after which the
|
|
260
|
+
containers will be downscaled to their minimum scale value as defined by
|
|
261
|
+
`scale.min`. If set, this helps in optimizing resource usage by reducing the
|
|
262
|
+
number of container instances during periods of inactivity.
|
|
263
|
+
"""
|
|
264
|
+
|
|
265
|
+
api_keys: Optional[List[str]] = None
|
|
266
|
+
"""List of API keys for the inference instance"""
|
|
@@ -4,10 +4,10 @@ from typing_extensions import Literal
|
|
|
4
4
|
|
|
5
5
|
from ...._models import BaseModel
|
|
6
6
|
|
|
7
|
-
__all__ = ["
|
|
7
|
+
__all__ = ["InferenceDeploymentAPIKey"]
|
|
8
8
|
|
|
9
9
|
|
|
10
|
-
class
|
|
10
|
+
class InferenceDeploymentAPIKey(BaseModel):
|
|
11
11
|
secret: str
|
|
12
12
|
"""API key secret"""
|
|
13
13
|
|
|
@@ -6,10 +6,10 @@ from pydantic import Field as FieldInfo
|
|
|
6
6
|
|
|
7
7
|
from ...._models import BaseModel
|
|
8
8
|
|
|
9
|
-
__all__ = ["
|
|
9
|
+
__all__ = ["InferenceModel"]
|
|
10
10
|
|
|
11
11
|
|
|
12
|
-
class
|
|
12
|
+
class InferenceModel(BaseModel):
|
|
13
13
|
id: str
|
|
14
14
|
"""Model ID."""
|
|
15
15
|
|
gcore/types/cloud/inference/{inference_registry_credential.py → inference_registry_credentials.py}
RENAMED
|
@@ -2,10 +2,10 @@
|
|
|
2
2
|
|
|
3
3
|
from ...._models import BaseModel
|
|
4
4
|
|
|
5
|
-
__all__ = ["
|
|
5
|
+
__all__ = ["InferenceRegistryCredentials"]
|
|
6
6
|
|
|
7
7
|
|
|
8
|
-
class
|
|
8
|
+
class InferenceRegistryCredentials(BaseModel):
|
|
9
9
|
name: str
|
|
10
10
|
"""Registry credential name."""
|
|
11
11
|
|
|
@@ -2,10 +2,10 @@
|
|
|
2
2
|
|
|
3
3
|
from ...._models import BaseModel
|
|
4
4
|
|
|
5
|
-
__all__ = ["
|
|
5
|
+
__all__ = ["InferenceRegistryCredentialsCreate"]
|
|
6
6
|
|
|
7
7
|
|
|
8
|
-
class
|
|
8
|
+
class InferenceRegistryCredentialsCreate(BaseModel):
|
|
9
9
|
name: str
|
|
10
10
|
"""Registry credential name."""
|
|
11
11
|
|
|
@@ -1,13 +1,20 @@
|
|
|
1
1
|
# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.
|
|
2
2
|
|
|
3
3
|
from ...._models import BaseModel
|
|
4
|
-
from ..aws_iam_data import AwsIamData
|
|
5
4
|
|
|
6
|
-
__all__ = ["InferenceSecret"]
|
|
5
|
+
__all__ = ["InferenceSecret", "Data"]
|
|
6
|
+
|
|
7
|
+
|
|
8
|
+
class Data(BaseModel):
|
|
9
|
+
aws_access_key_id: str
|
|
10
|
+
"""AWS IAM key ID."""
|
|
11
|
+
|
|
12
|
+
aws_secret_access_key: str
|
|
13
|
+
"""AWS IAM secret key."""
|
|
7
14
|
|
|
8
15
|
|
|
9
16
|
class InferenceSecret(BaseModel):
|
|
10
|
-
data:
|
|
17
|
+
data: Data
|
|
11
18
|
"""Secret data."""
|
|
12
19
|
|
|
13
20
|
name: str
|
|
@@ -2,9 +2,7 @@
|
|
|
2
2
|
|
|
3
3
|
from __future__ import annotations
|
|
4
4
|
|
|
5
|
-
from typing_extensions import TypedDict
|
|
6
|
-
|
|
7
|
-
from .mlcatalog_order_by_choices import MlcatalogOrderByChoices
|
|
5
|
+
from typing_extensions import Literal, TypedDict
|
|
8
6
|
|
|
9
7
|
__all__ = ["ModelListParams"]
|
|
10
8
|
|
|
@@ -19,5 +17,5 @@ class ModelListParams(TypedDict, total=False):
|
|
|
19
17
|
Offset value is used to exclude the first set of records from the result
|
|
20
18
|
"""
|
|
21
19
|
|
|
22
|
-
order_by:
|
|
20
|
+
order_by: Literal["name.asc", "name.desc"]
|
|
23
21
|
"""Order instances by transmitted fields and directions"""
|
|
@@ -2,22 +2,22 @@
|
|
|
2
2
|
|
|
3
3
|
from typing import Optional
|
|
4
4
|
|
|
5
|
-
from
|
|
6
|
-
from .
|
|
7
|
-
from .
|
|
8
|
-
from .
|
|
5
|
+
from ...._models import BaseModel
|
|
6
|
+
from .probe_exec import ProbeExec
|
|
7
|
+
from .probe_http_get import ProbeHTTPGet
|
|
8
|
+
from .probe_tcp_socket import ProbeTcpSocket
|
|
9
9
|
|
|
10
|
-
__all__ = ["
|
|
10
|
+
__all__ = ["Probe"]
|
|
11
11
|
|
|
12
12
|
|
|
13
|
-
class
|
|
14
|
-
exec: Optional[
|
|
13
|
+
class Probe(BaseModel):
|
|
14
|
+
exec: Optional[ProbeExec] = None
|
|
15
15
|
"""Exec probe configuration"""
|
|
16
16
|
|
|
17
17
|
failure_threshold: int
|
|
18
18
|
"""The number of consecutive probe failures that mark the container as unhealthy."""
|
|
19
19
|
|
|
20
|
-
http_get: Optional[
|
|
20
|
+
http_get: Optional[ProbeHTTPGet] = None
|
|
21
21
|
"""HTTP GET probe configuration"""
|
|
22
22
|
|
|
23
23
|
initial_delay_seconds: int
|
|
@@ -29,7 +29,7 @@ class ContainerProbe(BaseModel):
|
|
|
29
29
|
success_threshold: int
|
|
30
30
|
"""The number of consecutive successful probes that mark the container as healthy."""
|
|
31
31
|
|
|
32
|
-
tcp_socket: Optional[
|
|
32
|
+
tcp_socket: Optional[ProbeTcpSocket] = None
|
|
33
33
|
"""TCP socket probe configuration"""
|
|
34
34
|
|
|
35
35
|
timeout_seconds: int
|
|
@@ -2,15 +2,15 @@
|
|
|
2
2
|
|
|
3
3
|
from typing import Optional
|
|
4
4
|
|
|
5
|
-
from
|
|
6
|
-
from
|
|
5
|
+
from .probe import Probe
|
|
6
|
+
from ...._models import BaseModel
|
|
7
7
|
|
|
8
|
-
__all__ = ["
|
|
8
|
+
__all__ = ["ProbeConfig"]
|
|
9
9
|
|
|
10
10
|
|
|
11
|
-
class
|
|
11
|
+
class ProbeConfig(BaseModel):
|
|
12
12
|
enabled: bool
|
|
13
13
|
"""Whether the probe is enabled or not."""
|
|
14
14
|
|
|
15
|
-
probe: Optional[
|
|
15
|
+
probe: Optional[Probe] = None
|
|
16
16
|
"""Probe configuration (exec, `http_get` or `tcp_socket`)"""
|
|
@@ -2,11 +2,11 @@
|
|
|
2
2
|
|
|
3
3
|
from typing import List
|
|
4
4
|
|
|
5
|
-
from
|
|
5
|
+
from ...._models import BaseModel
|
|
6
6
|
|
|
7
|
-
__all__ = ["
|
|
7
|
+
__all__ = ["ProbeExec"]
|
|
8
8
|
|
|
9
9
|
|
|
10
|
-
class
|
|
10
|
+
class ProbeExec(BaseModel):
|
|
11
11
|
command: List[str]
|
|
12
12
|
"""Command to be executed inside the running container."""
|
|
@@ -4,12 +4,12 @@ from typing import Dict, Optional
|
|
|
4
4
|
|
|
5
5
|
from pydantic import Field as FieldInfo
|
|
6
6
|
|
|
7
|
-
from
|
|
7
|
+
from ...._models import BaseModel
|
|
8
8
|
|
|
9
|
-
__all__ = ["
|
|
9
|
+
__all__ = ["ProbeHTTPGet"]
|
|
10
10
|
|
|
11
11
|
|
|
12
|
-
class
|
|
12
|
+
class ProbeHTTPGet(BaseModel):
|
|
13
13
|
headers: Dict[str, str]
|
|
14
14
|
"""HTTP headers to be sent with the request."""
|
|
15
15
|
|
|
@@ -1,10 +1,10 @@
|
|
|
1
1
|
# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.
|
|
2
2
|
|
|
3
|
-
from
|
|
3
|
+
from ...._models import BaseModel
|
|
4
4
|
|
|
5
|
-
__all__ = ["
|
|
5
|
+
__all__ = ["ProbeTcpSocket"]
|
|
6
6
|
|
|
7
7
|
|
|
8
|
-
class
|
|
8
|
+
class ProbeTcpSocket(BaseModel):
|
|
9
9
|
port: int
|
|
10
10
|
"""Port number to check if it's open."""
|
|
@@ -4,16 +4,14 @@ from __future__ import annotations
|
|
|
4
4
|
|
|
5
5
|
from typing_extensions import Required, TypedDict
|
|
6
6
|
|
|
7
|
-
|
|
8
|
-
|
|
9
|
-
__all__ = ["SecretCreateParams"]
|
|
7
|
+
__all__ = ["SecretCreateParams", "Data"]
|
|
10
8
|
|
|
11
9
|
|
|
12
10
|
class SecretCreateParams(TypedDict, total=False):
|
|
13
11
|
project_id: int
|
|
14
12
|
"""Project ID"""
|
|
15
13
|
|
|
16
|
-
data: Required[
|
|
14
|
+
data: Required[Data]
|
|
17
15
|
"""Secret data."""
|
|
18
16
|
|
|
19
17
|
name: Required[str]
|
|
@@ -21,3 +19,11 @@ class SecretCreateParams(TypedDict, total=False):
|
|
|
21
19
|
|
|
22
20
|
type: Required[str]
|
|
23
21
|
"""Secret type. Currently only `aws-iam` is supported."""
|
|
22
|
+
|
|
23
|
+
|
|
24
|
+
class Data(TypedDict, total=False):
|
|
25
|
+
aws_access_key_id: Required[str]
|
|
26
|
+
"""AWS IAM key ID."""
|
|
27
|
+
|
|
28
|
+
aws_secret_access_key: Required[str]
|
|
29
|
+
"""AWS IAM secret key."""
|
|
@@ -4,17 +4,23 @@ from __future__ import annotations
|
|
|
4
4
|
|
|
5
5
|
from typing_extensions import Required, TypedDict
|
|
6
6
|
|
|
7
|
-
|
|
8
|
-
|
|
9
|
-
__all__ = ["SecretReplaceParams"]
|
|
7
|
+
__all__ = ["SecretReplaceParams", "Data"]
|
|
10
8
|
|
|
11
9
|
|
|
12
10
|
class SecretReplaceParams(TypedDict, total=False):
|
|
13
11
|
project_id: int
|
|
14
12
|
"""Project ID"""
|
|
15
13
|
|
|
16
|
-
data: Required[
|
|
14
|
+
data: Required[Data]
|
|
17
15
|
"""Secret data."""
|
|
18
16
|
|
|
19
17
|
type: Required[str]
|
|
20
18
|
"""Secret type."""
|
|
19
|
+
|
|
20
|
+
|
|
21
|
+
class Data(TypedDict, total=False):
|
|
22
|
+
aws_access_key_id: Required[str]
|
|
23
|
+
"""AWS IAM key ID."""
|
|
24
|
+
|
|
25
|
+
aws_secret_access_key: Required[str]
|
|
26
|
+
"""AWS IAM secret key."""
|
|
@@ -2,13 +2,20 @@
|
|
|
2
2
|
|
|
3
3
|
from typing import List
|
|
4
4
|
|
|
5
|
-
from .capacity import Capacity
|
|
6
5
|
from ..._models import BaseModel
|
|
7
6
|
|
|
8
|
-
__all__ = ["
|
|
7
|
+
__all__ = ["InferenceRegionCapacity", "Capacity"]
|
|
9
8
|
|
|
10
9
|
|
|
11
|
-
class
|
|
10
|
+
class Capacity(BaseModel):
|
|
11
|
+
capacity: int
|
|
12
|
+
"""Available capacity."""
|
|
13
|
+
|
|
14
|
+
flavor_name: str
|
|
15
|
+
"""Flavor name."""
|
|
16
|
+
|
|
17
|
+
|
|
18
|
+
class InferenceRegionCapacity(BaseModel):
|
|
12
19
|
capacity: List[Capacity]
|
|
13
20
|
"""List of capacities by flavor."""
|
|
14
21
|
|
|
@@ -3,14 +3,14 @@
|
|
|
3
3
|
from typing import List
|
|
4
4
|
|
|
5
5
|
from ..._models import BaseModel
|
|
6
|
-
from .
|
|
6
|
+
from .inference_region_capacity import InferenceRegionCapacity
|
|
7
7
|
|
|
8
|
-
__all__ = ["
|
|
8
|
+
__all__ = ["InferenceRegionCapacityList"]
|
|
9
9
|
|
|
10
10
|
|
|
11
|
-
class
|
|
11
|
+
class InferenceRegionCapacityList(BaseModel):
|
|
12
12
|
count: int
|
|
13
13
|
"""Number of objects"""
|
|
14
14
|
|
|
15
|
-
results: List[
|
|
15
|
+
results: List[InferenceRegionCapacity]
|
|
16
16
|
"""Objects"""
|
|
@@ -109,7 +109,7 @@ class InstanceCreateParams(TypedDict, total=False):
|
|
|
109
109
|
ssh_key_name: Optional[str]
|
|
110
110
|
"""
|
|
111
111
|
Specifies the name of the SSH keypair, created via the
|
|
112
|
-
[/v1/`ssh_keys` endpoint](/docs/api-reference/ssh-keys/add-or-generate-ssh-key).
|
|
112
|
+
[/v1/`ssh_keys` endpoint](/docs/api-reference/cloud/ssh-keys/add-or-generate-ssh-key).
|
|
113
113
|
"""
|
|
114
114
|
|
|
115
115
|
tags: Dict[str, str]
|
|
@@ -173,7 +173,7 @@ class ListenerPoolMember(TypedDict, total=False):
|
|
|
173
173
|
protocol_port: Required[int]
|
|
174
174
|
"""Member IP port"""
|
|
175
175
|
|
|
176
|
-
admin_state_up:
|
|
176
|
+
admin_state_up: bool
|
|
177
177
|
"""Administrative state of the resource.
|
|
178
178
|
|
|
179
179
|
When set to true, the resource is enabled and operational. When set to false,
|
|
@@ -181,6 +181,14 @@ class ListenerPoolMember(TypedDict, total=False):
|
|
|
181
181
|
value is skipped and defaults to true.
|
|
182
182
|
"""
|
|
183
183
|
|
|
184
|
+
backup: bool
|
|
185
|
+
"""
|
|
186
|
+
Set to true if the member is a backup member, to which traffic will be sent
|
|
187
|
+
exclusively when all non-backup members will be unreachable. It allows to
|
|
188
|
+
realize ACTIVE-BACKUP load balancing without thinking about VRRP and VIP
|
|
189
|
+
configuration. Default is false.
|
|
190
|
+
"""
|
|
191
|
+
|
|
184
192
|
instance_id: Optional[str]
|
|
185
193
|
"""Either `subnet_id` or `instance_id` should be provided"""
|
|
186
194
|
|
|
@@ -197,10 +205,13 @@ class ListenerPoolMember(TypedDict, total=False):
|
|
|
197
205
|
"""
|
|
198
206
|
|
|
199
207
|
subnet_id: Optional[str]
|
|
200
|
-
"""
|
|
208
|
+
"""`subnet_id` in which `address` is present.
|
|
209
|
+
|
|
210
|
+
Either `subnet_id` or `instance_id` should be provided
|
|
211
|
+
"""
|
|
201
212
|
|
|
202
213
|
weight: Optional[int]
|
|
203
|
-
"""Member weight. Valid values
|
|
214
|
+
"""Member weight. Valid values are 0 < `weight` <= 256, defaults to 1."""
|
|
204
215
|
|
|
205
216
|
|
|
206
217
|
class ListenerPoolSessionPersistence(TypedDict, total=False):
|
|
@@ -104,7 +104,7 @@ class Member(TypedDict, total=False):
|
|
|
104
104
|
protocol_port: Required[int]
|
|
105
105
|
"""Member IP port"""
|
|
106
106
|
|
|
107
|
-
admin_state_up:
|
|
107
|
+
admin_state_up: bool
|
|
108
108
|
"""Administrative state of the resource.
|
|
109
109
|
|
|
110
110
|
When set to true, the resource is enabled and operational. When set to false,
|
|
@@ -112,6 +112,14 @@ class Member(TypedDict, total=False):
|
|
|
112
112
|
value is skipped and defaults to true.
|
|
113
113
|
"""
|
|
114
114
|
|
|
115
|
+
backup: bool
|
|
116
|
+
"""
|
|
117
|
+
Set to true if the member is a backup member, to which traffic will be sent
|
|
118
|
+
exclusively when all non-backup members will be unreachable. It allows to
|
|
119
|
+
realize ACTIVE-BACKUP load balancing without thinking about VRRP and VIP
|
|
120
|
+
configuration. Default is false.
|
|
121
|
+
"""
|
|
122
|
+
|
|
115
123
|
instance_id: Optional[str]
|
|
116
124
|
"""Either `subnet_id` or `instance_id` should be provided"""
|
|
117
125
|
|
|
@@ -128,10 +136,13 @@ class Member(TypedDict, total=False):
|
|
|
128
136
|
"""
|
|
129
137
|
|
|
130
138
|
subnet_id: Optional[str]
|
|
131
|
-
"""
|
|
139
|
+
"""`subnet_id` in which `address` is present.
|
|
140
|
+
|
|
141
|
+
Either `subnet_id` or `instance_id` should be provided
|
|
142
|
+
"""
|
|
132
143
|
|
|
133
144
|
weight: Optional[int]
|
|
134
|
-
"""Member weight. Valid values
|
|
145
|
+
"""Member weight. Valid values are 0 < `weight` <= 256, defaults to 1."""
|
|
135
146
|
|
|
136
147
|
|
|
137
148
|
class SessionPersistence(TypedDict, total=False):
|