weave-python 0.28.1__py3-none-any.whl → 0.30.3__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- weave/weaveapi/auth/v1/service_pb2.py +194 -98
- weave/weaveapi/auth/v1/service_pb2.pyi +1288 -1134
- weave/weaveapi/auth/v1/service_pb2_grpc.pyi +248 -335
- weave/weaveapi/auth/v1/session_pb2.py +14 -6
- weave/weaveapi/auth/v1/session_pb2.pyi +109 -89
- weave/weaveapi/auth/v1/session_pb2_grpc.pyi +11 -10
- weave/weaveapi/auth/v1/usage_pb2.py +18 -14
- weave/weaveapi/auth/v1/usage_pb2.pyi +251 -197
- weave/weaveapi/auth/v1/usage_pb2_grpc.pyi +11 -10
- weave/weaveapi/auth/v1/user_pb2.py +60 -18
- weave/weaveapi/auth/v1/user_pb2.pyi +588 -505
- weave/weaveapi/auth/v1/user_pb2_grpc.pyi +11 -10
- weave/weaveapi/generate/v1/configuration_pb2.py +31 -19
- weave/weaveapi/generate/v1/configuration_pb2.pyi +277 -241
- weave/weaveapi/generate/v1/configuration_pb2_grpc.pyi +11 -10
- weave/weaveapi/generate/v1/generate_pb2.py +32 -6
- weave/weaveapi/generate/v1/generate_pb2.pyi +138 -40
- weave/weaveapi/generate/v1/generate_pb2_grpc.pyi +11 -10
- weave/weaveapi/generate/v1/service_pb2.py +52 -22
- weave/weaveapi/generate/v1/service_pb2.pyi +264 -135
- weave/weaveapi/generate/v1/service_pb2_grpc.py +47 -0
- weave/weaveapi/generate/v1/service_pb2_grpc.pyi +89 -78
- weave/weaveapi/llmx/v1/architecture_pb2.py +42 -42
- weave/weaveapi/llmx/v1/architecture_pb2.pyi +664 -555
- weave/weaveapi/llmx/v1/architecture_pb2_grpc.pyi +11 -10
- weave/weaveapi/llmx/v1/capabilities_pb2.py +88 -56
- weave/weaveapi/llmx/v1/capabilities_pb2.pyi +967 -852
- weave/weaveapi/llmx/v1/capabilities_pb2_grpc.pyi +11 -10
- weave/weaveapi/llmx/v1/model_pb2.py +56 -20
- weave/weaveapi/llmx/v1/model_pb2.pyi +533 -455
- weave/weaveapi/llmx/v1/model_pb2_grpc.pyi +11 -10
- weave/weaveapi/llmx/v1/pricing_pb2.py +24 -20
- weave/weaveapi/llmx/v1/pricing_pb2.pyi +194 -172
- weave/weaveapi/llmx/v1/pricing_pb2_grpc.pyi +11 -10
- weave/weaveapi/llmx/v1/provider_pb2.py +6 -6
- weave/weaveapi/llmx/v1/provider_pb2.pyi +84 -59
- weave/weaveapi/llmx/v1/provider_pb2_grpc.pyi +11 -10
- weave/weaveapi/llmx/v1/service_pb2.py +228 -90
- weave/weaveapi/llmx/v1/service_pb2.pyi +1651 -1403
- weave/weaveapi/llmx/v1/service_pb2_grpc.pyi +160 -203
- weave/weaveapi/mcpregistry/v1/server_pb2.py +16 -10
- weave/weaveapi/mcpregistry/v1/server_pb2.pyi +143 -121
- weave/weaveapi/mcpregistry/v1/server_pb2_grpc.pyi +11 -10
- weave/weaveapi/mcpregistry/v1/service_pb2.py +51 -31
- weave/weaveapi/mcpregistry/v1/service_pb2.pyi +132 -122
- weave/weaveapi/mcpregistry/v1/service_pb2_grpc.pyi +87 -112
- weave/weaveapi/payment/v1/invoice_pb2.py +38 -12
- weave/weaveapi/payment/v1/invoice_pb2.pyi +352 -291
- weave/weaveapi/payment/v1/invoice_pb2_grpc.pyi +11 -10
- weave/weaveapi/payment/v1/service_pb2.py +264 -98
- weave/weaveapi/payment/v1/service_pb2.pyi +1381 -1242
- weave/weaveapi/payment/v1/service_pb2_grpc.pyi +229 -319
- weave/weaveapi/payment/v1/subscription_pb2.py +99 -23
- weave/weaveapi/payment/v1/subscription_pb2.pyi +727 -611
- weave/weaveapi/payment/v1/subscription_pb2_grpc.pyi +11 -10
- weave/weaveapi/storage/v1/auth_pb2.py +6 -6
- weave/weaveapi/storage/v1/auth_pb2.pyi +42 -29
- weave/weaveapi/storage/v1/auth_pb2_grpc.pyi +11 -10
- weave/weaveapi/storage/v1/nosql_database_pb2.py +47 -23
- weave/weaveapi/storage/v1/nosql_database_pb2.pyi +438 -372
- weave/weaveapi/storage/v1/nosql_database_pb2_grpc.pyi +11 -10
- weave/weaveapi/storage/v1/object_store_pb2.py +27 -13
- weave/weaveapi/storage/v1/object_store_pb2.pyi +203 -187
- weave/weaveapi/storage/v1/object_store_pb2_grpc.pyi +11 -10
- weave/weaveapi/storage/v1/service_pb2.py +96 -36
- weave/weaveapi/storage/v1/service_pb2.pyi +414 -357
- weave/weaveapi/storage/v1/service_pb2_grpc.pyi +88 -107
- weave/weaveapi/storage/v1/sql_database_pb2.py +39 -23
- weave/weaveapi/storage/v1/sql_database_pb2.pyi +481 -400
- weave/weaveapi/storage/v1/sql_database_pb2_grpc.pyi +11 -10
- weave/weaveapi/storage/v1/storage_pb2.py +20 -6
- weave/weaveapi/storage/v1/storage_pb2.pyi +79 -70
- weave/weaveapi/storage/v1/storage_pb2_grpc.pyi +11 -10
- weave/weaveapi/synthesize/v1/dataset_pb2.py +14 -12
- weave/weaveapi/synthesize/v1/dataset_pb2.pyi +158 -128
- weave/weaveapi/synthesize/v1/dataset_pb2_grpc.pyi +11 -10
- weave/weaveapi/synthesize/v1/inline_data_pb2.py +9 -9
- weave/weaveapi/synthesize/v1/inline_data_pb2.pyi +31 -27
- weave/weaveapi/synthesize/v1/inline_data_pb2_grpc.pyi +11 -10
- weave/weaveapi/synthesize/v1/relationship_pb2.py +19 -11
- weave/weaveapi/synthesize/v1/relationship_pb2.pyi +67 -64
- weave/weaveapi/synthesize/v1/relationship_pb2_grpc.pyi +11 -10
- weave/weaveapi/synthesize/v1/service_pb2.py +42 -24
- weave/weaveapi/synthesize/v1/service_pb2.pyi +202 -168
- weave/weaveapi/synthesize/v1/service_pb2_grpc.pyi +67 -79
- weave/weaveapi/synthesize/v1/training_pb2.py +19 -13
- weave/weaveapi/synthesize/v1/training_pb2.pyi +119 -106
- weave/weaveapi/synthesize/v1/training_pb2_grpc.pyi +11 -10
- weave/weavesql/llmxdb/capabilities.py +487 -0
- weave/weavesql/llmxdb/changes.py +297 -0
- weave/weavesql/llmxdb/models.py +594 -0
- weave/weavesql/llmxdb/providers.py +348 -0
- weave/weavesql/llmxdb/scraper_runs.py +287 -0
- weave/weavesql/llmxdb/search.py +721 -0
- weave/weavesql/weavedb/dataset.py +75 -0
- weave/weavesql/weavedb/models.py +135 -0
- weave/weavesql/weavedb/relationships.py +72 -0
- weave/weavesql/weavedb/storage.py +113 -0
- weave/weavesql/weavedb/synthesizer.py +107 -0
- weave_python-0.30.3.dist-info/METADATA +52 -0
- weave_python-0.30.3.dist-info/RECORD +131 -0
- {weave_python-0.28.1.dist-info → weave_python-0.30.3.dist-info}/WHEEL +1 -1
- weave_python-0.30.3.dist-info/licenses/LICENSE +120 -0
- weave_python-0.28.1.dist-info/METADATA +0 -6
- weave_python-0.28.1.dist-info/RECORD +0 -119
|
@@ -3,33 +3,30 @@
|
|
|
3
3
|
isort:skip_file
|
|
4
4
|
"""
|
|
5
5
|
|
|
6
|
-
import
|
|
7
|
-
import
|
|
8
|
-
|
|
9
|
-
|
|
10
|
-
|
|
11
|
-
import
|
|
6
|
+
from collections import abc as _abc
|
|
7
|
+
from google.protobuf import descriptor as _descriptor
|
|
8
|
+
from google.protobuf import message as _message
|
|
9
|
+
from google.protobuf.internal import containers as _containers
|
|
10
|
+
from google.protobuf.internal import enum_type_wrapper as _enum_type_wrapper
|
|
11
|
+
import builtins as _builtins
|
|
12
12
|
import sys
|
|
13
|
-
import typing
|
|
13
|
+
import typing as _typing
|
|
14
14
|
|
|
15
15
|
if sys.version_info >= (3, 10):
|
|
16
|
-
|
|
16
|
+
from typing import TypeAlias as _TypeAlias
|
|
17
17
|
else:
|
|
18
|
-
import
|
|
18
|
+
from typing_extensions import TypeAlias as _TypeAlias
|
|
19
19
|
|
|
20
|
-
DESCRIPTOR:
|
|
20
|
+
DESCRIPTOR: _descriptor.FileDescriptor
|
|
21
21
|
|
|
22
22
|
class _CapabilityType:
|
|
23
|
-
ValueType =
|
|
24
|
-
V:
|
|
23
|
+
ValueType = _typing.NewType("ValueType", _builtins.int)
|
|
24
|
+
V: _TypeAlias = ValueType # noqa: Y015
|
|
25
25
|
|
|
26
26
|
class _CapabilityTypeEnumTypeWrapper(
|
|
27
|
-
|
|
28
|
-
_CapabilityType.ValueType
|
|
29
|
-
],
|
|
30
|
-
builtins.type,
|
|
27
|
+
_enum_type_wrapper._EnumTypeWrapper[_CapabilityType.ValueType], _builtins.type
|
|
31
28
|
):
|
|
32
|
-
DESCRIPTOR:
|
|
29
|
+
DESCRIPTOR: _descriptor.EnumDescriptor
|
|
33
30
|
CAPABILITY_TYPE_UNSPECIFIED: _CapabilityType.ValueType # 0
|
|
34
31
|
CAPABILITY_TYPE_TEXT: _CapabilityType.ValueType # 1
|
|
35
32
|
"""Basic text input/output capability"""
|
|
@@ -95,17 +92,16 @@ CAPABILITY_TYPE_EMBEDDINGS: CapabilityType.ValueType # 12
|
|
|
95
92
|
"""Vector embedding generation"""
|
|
96
93
|
CAPABILITY_TYPE_FINE_TUNING: CapabilityType.ValueType # 13
|
|
97
94
|
"""Custom model training/fine-tuning"""
|
|
98
|
-
|
|
95
|
+
Global___CapabilityType: _TypeAlias = CapabilityType # noqa: Y015
|
|
99
96
|
|
|
100
97
|
class _DataFormat:
|
|
101
|
-
ValueType =
|
|
102
|
-
V:
|
|
98
|
+
ValueType = _typing.NewType("ValueType", _builtins.int)
|
|
99
|
+
V: _TypeAlias = ValueType # noqa: Y015
|
|
103
100
|
|
|
104
101
|
class _DataFormatEnumTypeWrapper(
|
|
105
|
-
|
|
106
|
-
builtins.type,
|
|
102
|
+
_enum_type_wrapper._EnumTypeWrapper[_DataFormat.ValueType], _builtins.type
|
|
107
103
|
):
|
|
108
|
-
DESCRIPTOR:
|
|
104
|
+
DESCRIPTOR: _descriptor.EnumDescriptor
|
|
109
105
|
DATA_FORMAT_UNSPECIFIED: _DataFormat.ValueType # 0
|
|
110
106
|
DATA_FORMAT_JSON: _DataFormat.ValueType # 1
|
|
111
107
|
DATA_FORMAT_YAML: _DataFormat.ValueType # 2
|
|
@@ -130,19 +126,16 @@ DATA_FORMAT_PARQUET: DataFormat.ValueType # 6
|
|
|
130
126
|
DATA_FORMAT_PLAIN: DataFormat.ValueType # 7
|
|
131
127
|
DATA_FORMAT_MARKDOWN: DataFormat.ValueType # 8
|
|
132
128
|
DATA_FORMAT_STRUCTURED: DataFormat.ValueType # 9
|
|
133
|
-
|
|
129
|
+
Global___DataFormat: _TypeAlias = DataFormat # noqa: Y015
|
|
134
130
|
|
|
135
131
|
class _JsonSchemaType:
|
|
136
|
-
ValueType =
|
|
137
|
-
V:
|
|
132
|
+
ValueType = _typing.NewType("ValueType", _builtins.int)
|
|
133
|
+
V: _TypeAlias = ValueType # noqa: Y015
|
|
138
134
|
|
|
139
135
|
class _JsonSchemaTypeEnumTypeWrapper(
|
|
140
|
-
|
|
141
|
-
_JsonSchemaType.ValueType
|
|
142
|
-
],
|
|
143
|
-
builtins.type,
|
|
136
|
+
_enum_type_wrapper._EnumTypeWrapper[_JsonSchemaType.ValueType], _builtins.type
|
|
144
137
|
):
|
|
145
|
-
DESCRIPTOR:
|
|
138
|
+
DESCRIPTOR: _descriptor.EnumDescriptor
|
|
146
139
|
JSON_SCHEMA_TYPE_UNSPECIFIED: _JsonSchemaType.ValueType # 0
|
|
147
140
|
JSON_SCHEMA_TYPE_OBJECT: _JsonSchemaType.ValueType # 1
|
|
148
141
|
JSON_SCHEMA_TYPE_ARRAY: _JsonSchemaType.ValueType # 2
|
|
@@ -163,17 +156,16 @@ JSON_SCHEMA_TYPE_NUMBER: JsonSchemaType.ValueType # 4
|
|
|
163
156
|
JSON_SCHEMA_TYPE_BOOLEAN: JsonSchemaType.ValueType # 5
|
|
164
157
|
JSON_SCHEMA_TYPE_NULL: JsonSchemaType.ValueType # 6
|
|
165
158
|
JSON_SCHEMA_TYPE_INTEGER: JsonSchemaType.ValueType # 7
|
|
166
|
-
|
|
159
|
+
Global___JsonSchemaType: _TypeAlias = JsonSchemaType # noqa: Y015
|
|
167
160
|
|
|
168
161
|
class _ImageFormat:
|
|
169
|
-
ValueType =
|
|
170
|
-
V:
|
|
162
|
+
ValueType = _typing.NewType("ValueType", _builtins.int)
|
|
163
|
+
V: _TypeAlias = ValueType # noqa: Y015
|
|
171
164
|
|
|
172
165
|
class _ImageFormatEnumTypeWrapper(
|
|
173
|
-
|
|
174
|
-
builtins.type,
|
|
166
|
+
_enum_type_wrapper._EnumTypeWrapper[_ImageFormat.ValueType], _builtins.type
|
|
175
167
|
):
|
|
176
|
-
DESCRIPTOR:
|
|
168
|
+
DESCRIPTOR: _descriptor.EnumDescriptor
|
|
177
169
|
IMAGE_FORMAT_UNSPECIFIED: _ImageFormat.ValueType # 0
|
|
178
170
|
IMAGE_FORMAT_JPEG: _ImageFormat.ValueType # 1
|
|
179
171
|
IMAGE_FORMAT_PNG: _ImageFormat.ValueType # 2
|
|
@@ -194,17 +186,16 @@ IMAGE_FORMAT_WEBP: ImageFormat.ValueType # 4
|
|
|
194
186
|
IMAGE_FORMAT_BMP: ImageFormat.ValueType # 5
|
|
195
187
|
IMAGE_FORMAT_TIFF: ImageFormat.ValueType # 6
|
|
196
188
|
IMAGE_FORMAT_SVG: ImageFormat.ValueType # 7
|
|
197
|
-
|
|
189
|
+
Global___ImageFormat: _TypeAlias = ImageFormat # noqa: Y015
|
|
198
190
|
|
|
199
191
|
class _AudioFormat:
|
|
200
|
-
ValueType =
|
|
201
|
-
V:
|
|
192
|
+
ValueType = _typing.NewType("ValueType", _builtins.int)
|
|
193
|
+
V: _TypeAlias = ValueType # noqa: Y015
|
|
202
194
|
|
|
203
195
|
class _AudioFormatEnumTypeWrapper(
|
|
204
|
-
|
|
205
|
-
builtins.type,
|
|
196
|
+
_enum_type_wrapper._EnumTypeWrapper[_AudioFormat.ValueType], _builtins.type
|
|
206
197
|
):
|
|
207
|
-
DESCRIPTOR:
|
|
198
|
+
DESCRIPTOR: _descriptor.EnumDescriptor
|
|
208
199
|
AUDIO_FORMAT_UNSPECIFIED: _AudioFormat.ValueType # 0
|
|
209
200
|
AUDIO_FORMAT_MP3: _AudioFormat.ValueType # 1
|
|
210
201
|
AUDIO_FORMAT_WAV: _AudioFormat.ValueType # 2
|
|
@@ -227,17 +218,16 @@ AUDIO_FORMAT_FLAC: AudioFormat.ValueType # 5
|
|
|
227
218
|
AUDIO_FORMAT_AAC: AudioFormat.ValueType # 6
|
|
228
219
|
AUDIO_FORMAT_WMA: AudioFormat.ValueType # 7
|
|
229
220
|
AUDIO_FORMAT_OPUS: AudioFormat.ValueType # 8
|
|
230
|
-
|
|
221
|
+
Global___AudioFormat: _TypeAlias = AudioFormat # noqa: Y015
|
|
231
222
|
|
|
232
223
|
class _VideoFormat:
|
|
233
|
-
ValueType =
|
|
234
|
-
V:
|
|
224
|
+
ValueType = _typing.NewType("ValueType", _builtins.int)
|
|
225
|
+
V: _TypeAlias = ValueType # noqa: Y015
|
|
235
226
|
|
|
236
227
|
class _VideoFormatEnumTypeWrapper(
|
|
237
|
-
|
|
238
|
-
builtins.type,
|
|
228
|
+
_enum_type_wrapper._EnumTypeWrapper[_VideoFormat.ValueType], _builtins.type
|
|
239
229
|
):
|
|
240
|
-
DESCRIPTOR:
|
|
230
|
+
DESCRIPTOR: _descriptor.EnumDescriptor
|
|
241
231
|
VIDEO_FORMAT_UNSPECIFIED: _VideoFormat.ValueType # 0
|
|
242
232
|
VIDEO_FORMAT_MP4: _VideoFormat.ValueType # 1
|
|
243
233
|
VIDEO_FORMAT_AVI: _VideoFormat.ValueType # 2
|
|
@@ -258,17 +248,16 @@ VIDEO_FORMAT_MKV: VideoFormat.ValueType # 4
|
|
|
258
248
|
VIDEO_FORMAT_WEBM: VideoFormat.ValueType # 5
|
|
259
249
|
VIDEO_FORMAT_FLV: VideoFormat.ValueType # 6
|
|
260
250
|
VIDEO_FORMAT_WMV: VideoFormat.ValueType # 7
|
|
261
|
-
|
|
251
|
+
Global___VideoFormat: _TypeAlias = VideoFormat # noqa: Y015
|
|
262
252
|
|
|
263
253
|
class _ToolType:
|
|
264
|
-
ValueType =
|
|
265
|
-
V:
|
|
254
|
+
ValueType = _typing.NewType("ValueType", _builtins.int)
|
|
255
|
+
V: _TypeAlias = ValueType # noqa: Y015
|
|
266
256
|
|
|
267
257
|
class _ToolTypeEnumTypeWrapper(
|
|
268
|
-
|
|
269
|
-
builtins.type,
|
|
258
|
+
_enum_type_wrapper._EnumTypeWrapper[_ToolType.ValueType], _builtins.type
|
|
270
259
|
):
|
|
271
|
-
DESCRIPTOR:
|
|
260
|
+
DESCRIPTOR: _descriptor.EnumDescriptor
|
|
272
261
|
TOOL_TYPE_UNSPECIFIED: _ToolType.ValueType # 0
|
|
273
262
|
TOOL_TYPE_FUNCTION: _ToolType.ValueType # 1
|
|
274
263
|
TOOL_TYPE_RETRIEVAL: _ToolType.ValueType # 2
|
|
@@ -289,19 +278,16 @@ TOOL_TYPE_WEB_BROWSER: ToolType.ValueType # 4
|
|
|
289
278
|
TOOL_TYPE_DATABASE: ToolType.ValueType # 5
|
|
290
279
|
TOOL_TYPE_API: ToolType.ValueType # 6
|
|
291
280
|
TOOL_TYPE_CUSTOM: ToolType.ValueType # 7
|
|
292
|
-
|
|
281
|
+
Global___ToolType: _TypeAlias = ToolType # noqa: Y015
|
|
293
282
|
|
|
294
283
|
class _CacheStrategy:
|
|
295
|
-
ValueType =
|
|
296
|
-
V:
|
|
284
|
+
ValueType = _typing.NewType("ValueType", _builtins.int)
|
|
285
|
+
V: _TypeAlias = ValueType # noqa: Y015
|
|
297
286
|
|
|
298
287
|
class _CacheStrategyEnumTypeWrapper(
|
|
299
|
-
|
|
300
|
-
_CacheStrategy.ValueType
|
|
301
|
-
],
|
|
302
|
-
builtins.type,
|
|
288
|
+
_enum_type_wrapper._EnumTypeWrapper[_CacheStrategy.ValueType], _builtins.type
|
|
303
289
|
):
|
|
304
|
-
DESCRIPTOR:
|
|
290
|
+
DESCRIPTOR: _descriptor.EnumDescriptor
|
|
305
291
|
CACHE_STRATEGY_UNSPECIFIED: _CacheStrategy.ValueType # 0
|
|
306
292
|
CACHE_STRATEGY_HASH: _CacheStrategy.ValueType # 1
|
|
307
293
|
CACHE_STRATEGY_SEMANTIC: _CacheStrategy.ValueType # 2
|
|
@@ -318,19 +304,16 @@ CACHE_STRATEGY_SEMANTIC: CacheStrategy.ValueType # 2
|
|
|
318
304
|
CACHE_STRATEGY_CUSTOM: CacheStrategy.ValueType # 3
|
|
319
305
|
CACHE_STRATEGY_PREFIX: CacheStrategy.ValueType # 4
|
|
320
306
|
CACHE_STRATEGY_SUFFIX: CacheStrategy.ValueType # 5
|
|
321
|
-
|
|
307
|
+
Global___CacheStrategy: _TypeAlias = CacheStrategy # noqa: Y015
|
|
322
308
|
|
|
323
309
|
class _ReasoningStrategy:
|
|
324
|
-
ValueType =
|
|
325
|
-
V:
|
|
310
|
+
ValueType = _typing.NewType("ValueType", _builtins.int)
|
|
311
|
+
V: _TypeAlias = ValueType # noqa: Y015
|
|
326
312
|
|
|
327
313
|
class _ReasoningStrategyEnumTypeWrapper(
|
|
328
|
-
|
|
329
|
-
_ReasoningStrategy.ValueType
|
|
330
|
-
],
|
|
331
|
-
builtins.type,
|
|
314
|
+
_enum_type_wrapper._EnumTypeWrapper[_ReasoningStrategy.ValueType], _builtins.type
|
|
332
315
|
):
|
|
333
|
-
DESCRIPTOR:
|
|
316
|
+
DESCRIPTOR: _descriptor.EnumDescriptor
|
|
334
317
|
REASONING_STRATEGY_UNSPECIFIED: _ReasoningStrategy.ValueType # 0
|
|
335
318
|
REASONING_STRATEGY_CHAIN_OF_THOUGHT: _ReasoningStrategy.ValueType # 1
|
|
336
319
|
REASONING_STRATEGY_TREE_OF_THOUGHTS: _ReasoningStrategy.ValueType # 2
|
|
@@ -351,19 +334,16 @@ REASONING_STRATEGY_GRAPH_OF_THOUGHTS: ReasoningStrategy.ValueType # 3
|
|
|
351
334
|
REASONING_STRATEGY_STEP_BY_STEP: ReasoningStrategy.ValueType # 4
|
|
352
335
|
REASONING_STRATEGY_SELF_CONSISTENCY: ReasoningStrategy.ValueType # 5
|
|
353
336
|
REASONING_STRATEGY_LEAST_TO_MOST: ReasoningStrategy.ValueType # 6
|
|
354
|
-
|
|
337
|
+
Global___ReasoningStrategy: _TypeAlias = ReasoningStrategy # noqa: Y015
|
|
355
338
|
|
|
356
339
|
class _DistanceMetric:
|
|
357
|
-
ValueType =
|
|
358
|
-
V:
|
|
340
|
+
ValueType = _typing.NewType("ValueType", _builtins.int)
|
|
341
|
+
V: _TypeAlias = ValueType # noqa: Y015
|
|
359
342
|
|
|
360
343
|
class _DistanceMetricEnumTypeWrapper(
|
|
361
|
-
|
|
362
|
-
_DistanceMetric.ValueType
|
|
363
|
-
],
|
|
364
|
-
builtins.type,
|
|
344
|
+
_enum_type_wrapper._EnumTypeWrapper[_DistanceMetric.ValueType], _builtins.type
|
|
365
345
|
):
|
|
366
|
-
DESCRIPTOR:
|
|
346
|
+
DESCRIPTOR: _descriptor.EnumDescriptor
|
|
367
347
|
DISTANCE_METRIC_UNSPECIFIED: _DistanceMetric.ValueType # 0
|
|
368
348
|
DISTANCE_METRIC_COSINE: _DistanceMetric.ValueType # 1
|
|
369
349
|
DISTANCE_METRIC_EUCLIDEAN: _DistanceMetric.ValueType # 2
|
|
@@ -380,19 +360,16 @@ DISTANCE_METRIC_EUCLIDEAN: DistanceMetric.ValueType # 2
|
|
|
380
360
|
DISTANCE_METRIC_DOT_PRODUCT: DistanceMetric.ValueType # 3
|
|
381
361
|
DISTANCE_METRIC_MANHATTAN: DistanceMetric.ValueType # 4
|
|
382
362
|
DISTANCE_METRIC_HAMMING: DistanceMetric.ValueType # 5
|
|
383
|
-
|
|
363
|
+
Global___DistanceMetric: _TypeAlias = DistanceMetric # noqa: Y015
|
|
384
364
|
|
|
385
365
|
class _Hyperparameter:
|
|
386
|
-
ValueType =
|
|
387
|
-
V:
|
|
366
|
+
ValueType = _typing.NewType("ValueType", _builtins.int)
|
|
367
|
+
V: _TypeAlias = ValueType # noqa: Y015
|
|
388
368
|
|
|
389
369
|
class _HyperparameterEnumTypeWrapper(
|
|
390
|
-
|
|
391
|
-
_Hyperparameter.ValueType
|
|
392
|
-
],
|
|
393
|
-
builtins.type,
|
|
370
|
+
_enum_type_wrapper._EnumTypeWrapper[_Hyperparameter.ValueType], _builtins.type
|
|
394
371
|
):
|
|
395
|
-
DESCRIPTOR:
|
|
372
|
+
DESCRIPTOR: _descriptor.EnumDescriptor
|
|
396
373
|
HYPERPARAMETER_UNSPECIFIED: _Hyperparameter.ValueType # 0
|
|
397
374
|
HYPERPARAMETER_LEARNING_RATE: _Hyperparameter.ValueType # 1
|
|
398
375
|
HYPERPARAMETER_BATCH_SIZE: _Hyperparameter.ValueType # 2
|
|
@@ -419,19 +396,16 @@ HYPERPARAMETER_LR_SCHEDULER: Hyperparameter.ValueType # 7
|
|
|
419
396
|
HYPERPARAMETER_OPTIMIZER: Hyperparameter.ValueType # 8
|
|
420
397
|
HYPERPARAMETER_DROPOUT: Hyperparameter.ValueType # 9
|
|
421
398
|
HYPERPARAMETER_MAX_SEQUENCE_LENGTH: Hyperparameter.ValueType # 10
|
|
422
|
-
|
|
399
|
+
Global___Hyperparameter: _TypeAlias = Hyperparameter # noqa: Y015
|
|
423
400
|
|
|
424
401
|
class _ModalityDirection:
|
|
425
|
-
ValueType =
|
|
426
|
-
V:
|
|
402
|
+
ValueType = _typing.NewType("ValueType", _builtins.int)
|
|
403
|
+
V: _TypeAlias = ValueType # noqa: Y015
|
|
427
404
|
|
|
428
405
|
class _ModalityDirectionEnumTypeWrapper(
|
|
429
|
-
|
|
430
|
-
_ModalityDirection.ValueType
|
|
431
|
-
],
|
|
432
|
-
builtins.type,
|
|
406
|
+
_enum_type_wrapper._EnumTypeWrapper[_ModalityDirection.ValueType], _builtins.type
|
|
433
407
|
):
|
|
434
|
-
DESCRIPTOR:
|
|
408
|
+
DESCRIPTOR: _descriptor.EnumDescriptor
|
|
435
409
|
MODALITY_DIRECTION_UNSPECIFIED: _ModalityDirection.ValueType # 0
|
|
436
410
|
MODALITY_DIRECTION_INPUT_ONLY: _ModalityDirection.ValueType # 1
|
|
437
411
|
"""Can only process/analyze (e.g., image analysis)"""
|
|
@@ -452,224 +426,225 @@ MODALITY_DIRECTION_OUTPUT_ONLY: ModalityDirection.ValueType # 2
|
|
|
452
426
|
"""Can only generate (e.g., TTS without STT)"""
|
|
453
427
|
MODALITY_DIRECTION_INPUT_OUTPUT: ModalityDirection.ValueType # 3
|
|
454
428
|
"""Can both process and generate"""
|
|
455
|
-
|
|
429
|
+
Global___ModalityDirection: _TypeAlias = ModalityDirection # noqa: Y015
|
|
456
430
|
|
|
457
|
-
@
|
|
458
|
-
class Capability(
|
|
431
|
+
@_typing.final
|
|
432
|
+
class Capability(_message.Message):
|
|
459
433
|
"""Capability represents a specific feature/ability of a model with its configuration.
|
|
460
434
|
Each capability has a type and optional detailed configuration.
|
|
461
435
|
"""
|
|
462
436
|
|
|
463
|
-
DESCRIPTOR:
|
|
464
|
-
|
|
465
|
-
TYPE_FIELD_NUMBER:
|
|
466
|
-
ENABLED_FIELD_NUMBER:
|
|
467
|
-
TEXT_FIELD_NUMBER:
|
|
468
|
-
STRUCTURED_RESPONSE_FIELD_NUMBER:
|
|
469
|
-
STREAMING_FIELD_NUMBER:
|
|
470
|
-
FUNCTION_CALLING_FIELD_NUMBER:
|
|
471
|
-
VISION_FIELD_NUMBER:
|
|
472
|
-
TOOL_USE_FIELD_NUMBER:
|
|
473
|
-
SYSTEM_PROMPT_FIELD_NUMBER:
|
|
474
|
-
CACHING_FIELD_NUMBER:
|
|
475
|
-
REASONING_FIELD_NUMBER:
|
|
476
|
-
AUDIO_FIELD_NUMBER:
|
|
477
|
-
VIDEO_FIELD_NUMBER:
|
|
478
|
-
EMBEDDINGS_FIELD_NUMBER:
|
|
479
|
-
FINE_TUNING_FIELD_NUMBER:
|
|
480
|
-
ADDITIONAL_INFO_FIELD_NUMBER:
|
|
481
|
-
type:
|
|
437
|
+
DESCRIPTOR: _descriptor.Descriptor
|
|
438
|
+
|
|
439
|
+
TYPE_FIELD_NUMBER: _builtins.int
|
|
440
|
+
ENABLED_FIELD_NUMBER: _builtins.int
|
|
441
|
+
TEXT_FIELD_NUMBER: _builtins.int
|
|
442
|
+
STRUCTURED_RESPONSE_FIELD_NUMBER: _builtins.int
|
|
443
|
+
STREAMING_FIELD_NUMBER: _builtins.int
|
|
444
|
+
FUNCTION_CALLING_FIELD_NUMBER: _builtins.int
|
|
445
|
+
VISION_FIELD_NUMBER: _builtins.int
|
|
446
|
+
TOOL_USE_FIELD_NUMBER: _builtins.int
|
|
447
|
+
SYSTEM_PROMPT_FIELD_NUMBER: _builtins.int
|
|
448
|
+
CACHING_FIELD_NUMBER: _builtins.int
|
|
449
|
+
REASONING_FIELD_NUMBER: _builtins.int
|
|
450
|
+
AUDIO_FIELD_NUMBER: _builtins.int
|
|
451
|
+
VIDEO_FIELD_NUMBER: _builtins.int
|
|
452
|
+
EMBEDDINGS_FIELD_NUMBER: _builtins.int
|
|
453
|
+
FINE_TUNING_FIELD_NUMBER: _builtins.int
|
|
454
|
+
ADDITIONAL_INFO_FIELD_NUMBER: _builtins.int
|
|
455
|
+
type: Global___CapabilityType.ValueType
|
|
482
456
|
"""The type of capability this represents.
|
|
483
457
|
Example: CAPABILITY_TYPE_FUNCTION_CALLING for function/tool calling
|
|
484
458
|
"""
|
|
485
|
-
enabled:
|
|
459
|
+
enabled: _builtins.bool
|
|
486
460
|
"""Whether this capability is currently enabled/available.
|
|
487
461
|
Example: true if the model supports and has this feature active
|
|
488
462
|
"""
|
|
489
|
-
additional_info:
|
|
463
|
+
additional_info: _builtins.str
|
|
490
464
|
"""Unstructured additional information about this capability.
|
|
491
465
|
Used for provider-specific details that don't fit the structured fields.
|
|
492
466
|
Example: "Supports up to 10 parallel function calls with automatic retry"
|
|
493
467
|
Example: "Beta feature - may have unexpected behavior"
|
|
494
468
|
Example: "Optimized for conversational use cases"
|
|
495
469
|
"""
|
|
496
|
-
@property
|
|
497
|
-
def text(self) ->
|
|
498
|
-
@property
|
|
499
|
-
def structured_response(self) ->
|
|
500
|
-
@property
|
|
501
|
-
def streaming(self) ->
|
|
502
|
-
@property
|
|
503
|
-
def function_calling(self) ->
|
|
504
|
-
@property
|
|
505
|
-
def vision(self) ->
|
|
506
|
-
@property
|
|
507
|
-
def tool_use(self) ->
|
|
508
|
-
@property
|
|
509
|
-
def system_prompt(self) ->
|
|
510
|
-
@property
|
|
511
|
-
def caching(self) ->
|
|
512
|
-
@property
|
|
513
|
-
def reasoning(self) ->
|
|
514
|
-
@property
|
|
515
|
-
def audio(self) ->
|
|
516
|
-
@property
|
|
517
|
-
def video(self) ->
|
|
518
|
-
@property
|
|
519
|
-
def embeddings(self) ->
|
|
520
|
-
@property
|
|
521
|
-
def fine_tuning(self) ->
|
|
470
|
+
@_builtins.property
|
|
471
|
+
def text(self) -> Global___Text: ...
|
|
472
|
+
@_builtins.property
|
|
473
|
+
def structured_response(self) -> Global___StructuredResponse: ...
|
|
474
|
+
@_builtins.property
|
|
475
|
+
def streaming(self) -> Global___Streaming: ...
|
|
476
|
+
@_builtins.property
|
|
477
|
+
def function_calling(self) -> Global___FunctionCalling: ...
|
|
478
|
+
@_builtins.property
|
|
479
|
+
def vision(self) -> Global___Vision: ...
|
|
480
|
+
@_builtins.property
|
|
481
|
+
def tool_use(self) -> Global___ToolUse: ...
|
|
482
|
+
@_builtins.property
|
|
483
|
+
def system_prompt(self) -> Global___SystemPrompt: ...
|
|
484
|
+
@_builtins.property
|
|
485
|
+
def caching(self) -> Global___Caching: ...
|
|
486
|
+
@_builtins.property
|
|
487
|
+
def reasoning(self) -> Global___Reasoning: ...
|
|
488
|
+
@_builtins.property
|
|
489
|
+
def audio(self) -> Global___Audio: ...
|
|
490
|
+
@_builtins.property
|
|
491
|
+
def video(self) -> Global___Video: ...
|
|
492
|
+
@_builtins.property
|
|
493
|
+
def embeddings(self) -> Global___Embeddings: ...
|
|
494
|
+
@_builtins.property
|
|
495
|
+
def fine_tuning(self) -> Global___FineTuning: ...
|
|
522
496
|
def __init__(
|
|
523
497
|
self,
|
|
524
498
|
*,
|
|
525
|
-
type:
|
|
526
|
-
enabled:
|
|
527
|
-
text:
|
|
528
|
-
structured_response:
|
|
529
|
-
streaming:
|
|
530
|
-
function_calling:
|
|
531
|
-
vision:
|
|
532
|
-
tool_use:
|
|
533
|
-
system_prompt:
|
|
534
|
-
caching:
|
|
535
|
-
reasoning:
|
|
536
|
-
audio:
|
|
537
|
-
video:
|
|
538
|
-
embeddings:
|
|
539
|
-
fine_tuning:
|
|
540
|
-
additional_info:
|
|
541
|
-
) -> None: ...
|
|
542
|
-
def HasField(
|
|
543
|
-
self,
|
|
544
|
-
field_name: typing.Literal[
|
|
545
|
-
"audio",
|
|
546
|
-
b"audio",
|
|
547
|
-
"caching",
|
|
548
|
-
b"caching",
|
|
549
|
-
"config",
|
|
550
|
-
b"config",
|
|
551
|
-
"embeddings",
|
|
552
|
-
b"embeddings",
|
|
553
|
-
"fine_tuning",
|
|
554
|
-
b"fine_tuning",
|
|
555
|
-
"function_calling",
|
|
556
|
-
b"function_calling",
|
|
557
|
-
"reasoning",
|
|
558
|
-
b"reasoning",
|
|
559
|
-
"streaming",
|
|
560
|
-
b"streaming",
|
|
561
|
-
"structured_response",
|
|
562
|
-
b"structured_response",
|
|
563
|
-
"system_prompt",
|
|
564
|
-
b"system_prompt",
|
|
565
|
-
"text",
|
|
566
|
-
b"text",
|
|
567
|
-
"tool_use",
|
|
568
|
-
b"tool_use",
|
|
569
|
-
"video",
|
|
570
|
-
b"video",
|
|
571
|
-
"vision",
|
|
572
|
-
b"vision",
|
|
573
|
-
],
|
|
574
|
-
) -> builtins.bool: ...
|
|
575
|
-
def ClearField(
|
|
576
|
-
self,
|
|
577
|
-
field_name: typing.Literal[
|
|
578
|
-
"additional_info",
|
|
579
|
-
b"additional_info",
|
|
580
|
-
"audio",
|
|
581
|
-
b"audio",
|
|
582
|
-
"caching",
|
|
583
|
-
b"caching",
|
|
584
|
-
"config",
|
|
585
|
-
b"config",
|
|
586
|
-
"embeddings",
|
|
587
|
-
b"embeddings",
|
|
588
|
-
"enabled",
|
|
589
|
-
b"enabled",
|
|
590
|
-
"fine_tuning",
|
|
591
|
-
b"fine_tuning",
|
|
592
|
-
"function_calling",
|
|
593
|
-
b"function_calling",
|
|
594
|
-
"reasoning",
|
|
595
|
-
b"reasoning",
|
|
596
|
-
"streaming",
|
|
597
|
-
b"streaming",
|
|
598
|
-
"structured_response",
|
|
599
|
-
b"structured_response",
|
|
600
|
-
"system_prompt",
|
|
601
|
-
b"system_prompt",
|
|
602
|
-
"text",
|
|
603
|
-
b"text",
|
|
604
|
-
"tool_use",
|
|
605
|
-
b"tool_use",
|
|
606
|
-
"type",
|
|
607
|
-
b"type",
|
|
608
|
-
"video",
|
|
609
|
-
b"video",
|
|
610
|
-
"vision",
|
|
611
|
-
b"vision",
|
|
612
|
-
],
|
|
499
|
+
type: Global___CapabilityType.ValueType | None = ...,
|
|
500
|
+
enabled: _builtins.bool | None = ...,
|
|
501
|
+
text: Global___Text | None = ...,
|
|
502
|
+
structured_response: Global___StructuredResponse | None = ...,
|
|
503
|
+
streaming: Global___Streaming | None = ...,
|
|
504
|
+
function_calling: Global___FunctionCalling | None = ...,
|
|
505
|
+
vision: Global___Vision | None = ...,
|
|
506
|
+
tool_use: Global___ToolUse | None = ...,
|
|
507
|
+
system_prompt: Global___SystemPrompt | None = ...,
|
|
508
|
+
caching: Global___Caching | None = ...,
|
|
509
|
+
reasoning: Global___Reasoning | None = ...,
|
|
510
|
+
audio: Global___Audio | None = ...,
|
|
511
|
+
video: Global___Video | None = ...,
|
|
512
|
+
embeddings: Global___Embeddings | None = ...,
|
|
513
|
+
fine_tuning: Global___FineTuning | None = ...,
|
|
514
|
+
additional_info: _builtins.str | None = ...,
|
|
613
515
|
) -> None: ...
|
|
516
|
+
_HasFieldArgType: _TypeAlias = _typing.Literal[
|
|
517
|
+
"additional_info",
|
|
518
|
+
b"additional_info",
|
|
519
|
+
"audio",
|
|
520
|
+
b"audio",
|
|
521
|
+
"caching",
|
|
522
|
+
b"caching",
|
|
523
|
+
"config",
|
|
524
|
+
b"config",
|
|
525
|
+
"embeddings",
|
|
526
|
+
b"embeddings",
|
|
527
|
+
"enabled",
|
|
528
|
+
b"enabled",
|
|
529
|
+
"fine_tuning",
|
|
530
|
+
b"fine_tuning",
|
|
531
|
+
"function_calling",
|
|
532
|
+
b"function_calling",
|
|
533
|
+
"reasoning",
|
|
534
|
+
b"reasoning",
|
|
535
|
+
"streaming",
|
|
536
|
+
b"streaming",
|
|
537
|
+
"structured_response",
|
|
538
|
+
b"structured_response",
|
|
539
|
+
"system_prompt",
|
|
540
|
+
b"system_prompt",
|
|
541
|
+
"text",
|
|
542
|
+
b"text",
|
|
543
|
+
"tool_use",
|
|
544
|
+
b"tool_use",
|
|
545
|
+
"type",
|
|
546
|
+
b"type",
|
|
547
|
+
"video",
|
|
548
|
+
b"video",
|
|
549
|
+
"vision",
|
|
550
|
+
b"vision",
|
|
551
|
+
] # noqa: Y015
|
|
552
|
+
def HasField(self, field_name: _HasFieldArgType) -> _builtins.bool: ...
|
|
553
|
+
_ClearFieldArgType: _TypeAlias = _typing.Literal[
|
|
554
|
+
"additional_info",
|
|
555
|
+
b"additional_info",
|
|
556
|
+
"audio",
|
|
557
|
+
b"audio",
|
|
558
|
+
"caching",
|
|
559
|
+
b"caching",
|
|
560
|
+
"config",
|
|
561
|
+
b"config",
|
|
562
|
+
"embeddings",
|
|
563
|
+
b"embeddings",
|
|
564
|
+
"enabled",
|
|
565
|
+
b"enabled",
|
|
566
|
+
"fine_tuning",
|
|
567
|
+
b"fine_tuning",
|
|
568
|
+
"function_calling",
|
|
569
|
+
b"function_calling",
|
|
570
|
+
"reasoning",
|
|
571
|
+
b"reasoning",
|
|
572
|
+
"streaming",
|
|
573
|
+
b"streaming",
|
|
574
|
+
"structured_response",
|
|
575
|
+
b"structured_response",
|
|
576
|
+
"system_prompt",
|
|
577
|
+
b"system_prompt",
|
|
578
|
+
"text",
|
|
579
|
+
b"text",
|
|
580
|
+
"tool_use",
|
|
581
|
+
b"tool_use",
|
|
582
|
+
"type",
|
|
583
|
+
b"type",
|
|
584
|
+
"video",
|
|
585
|
+
b"video",
|
|
586
|
+
"vision",
|
|
587
|
+
b"vision",
|
|
588
|
+
] # noqa: Y015
|
|
589
|
+
def ClearField(self, field_name: _ClearFieldArgType) -> None: ...
|
|
590
|
+
_WhichOneofReturnType_config: _TypeAlias = _typing.Literal[
|
|
591
|
+
"text",
|
|
592
|
+
"structured_response",
|
|
593
|
+
"streaming",
|
|
594
|
+
"function_calling",
|
|
595
|
+
"vision",
|
|
596
|
+
"tool_use",
|
|
597
|
+
"system_prompt",
|
|
598
|
+
"caching",
|
|
599
|
+
"reasoning",
|
|
600
|
+
"audio",
|
|
601
|
+
"video",
|
|
602
|
+
"embeddings",
|
|
603
|
+
"fine_tuning",
|
|
604
|
+
] # noqa: Y015
|
|
605
|
+
_WhichOneofArgType_config: _TypeAlias = _typing.Literal["config", b"config"] # noqa: Y015
|
|
614
606
|
def WhichOneof(
|
|
615
|
-
self, oneof_group:
|
|
616
|
-
) ->
|
|
617
|
-
|
|
618
|
-
|
|
619
|
-
|
|
620
|
-
|
|
621
|
-
|
|
622
|
-
"vision",
|
|
623
|
-
"tool_use",
|
|
624
|
-
"system_prompt",
|
|
625
|
-
"caching",
|
|
626
|
-
"reasoning",
|
|
627
|
-
"audio",
|
|
628
|
-
"video",
|
|
629
|
-
"embeddings",
|
|
630
|
-
"fine_tuning",
|
|
631
|
-
]
|
|
632
|
-
| None
|
|
633
|
-
): ...
|
|
634
|
-
|
|
635
|
-
global___Capability = Capability
|
|
636
|
-
|
|
637
|
-
@typing.final
|
|
638
|
-
class Text(google.protobuf.message.Message):
|
|
607
|
+
self, oneof_group: _WhichOneofArgType_config
|
|
608
|
+
) -> _WhichOneofReturnType_config | None: ...
|
|
609
|
+
|
|
610
|
+
Global___Capability: _TypeAlias = Capability # noqa: Y015
|
|
611
|
+
|
|
612
|
+
@_typing.final
|
|
613
|
+
class Text(_message.Message):
|
|
639
614
|
"""Text capability configuration for basic text input/output"""
|
|
640
615
|
|
|
641
|
-
DESCRIPTOR:
|
|
616
|
+
DESCRIPTOR: _descriptor.Descriptor
|
|
642
617
|
|
|
643
|
-
DIRECTION_FIELD_NUMBER:
|
|
644
|
-
MAX_INPUT_LENGTH_FIELD_NUMBER:
|
|
645
|
-
MAX_OUTPUT_LENGTH_FIELD_NUMBER:
|
|
646
|
-
SUPPORTED_LANGUAGES_FIELD_NUMBER:
|
|
647
|
-
SUPPORTS_CONVERSATION_FIELD_NUMBER:
|
|
648
|
-
SUPPORTS_CONTEXT_FIELD_NUMBER:
|
|
649
|
-
direction:
|
|
618
|
+
DIRECTION_FIELD_NUMBER: _builtins.int
|
|
619
|
+
MAX_INPUT_LENGTH_FIELD_NUMBER: _builtins.int
|
|
620
|
+
MAX_OUTPUT_LENGTH_FIELD_NUMBER: _builtins.int
|
|
621
|
+
SUPPORTED_LANGUAGES_FIELD_NUMBER: _builtins.int
|
|
622
|
+
SUPPORTS_CONVERSATION_FIELD_NUMBER: _builtins.int
|
|
623
|
+
SUPPORTS_CONTEXT_FIELD_NUMBER: _builtins.int
|
|
624
|
+
direction: Global___ModalityDirection.ValueType
|
|
650
625
|
"""Direction of text support (input, output, or both)
|
|
651
626
|
Example: MODALITY_DIRECTION_INPUT_OUTPUT for chat models
|
|
652
627
|
"""
|
|
653
|
-
max_input_length:
|
|
628
|
+
max_input_length: _builtins.int
|
|
654
629
|
"""Maximum input text length in characters (if limited)
|
|
655
630
|
Example: 32000 for models with character limits
|
|
656
631
|
"""
|
|
657
|
-
max_output_length:
|
|
632
|
+
max_output_length: _builtins.int
|
|
658
633
|
"""Maximum output text length in characters (if limited)
|
|
659
634
|
Example: 4096 for models with output limits
|
|
660
635
|
"""
|
|
661
|
-
supports_conversation:
|
|
636
|
+
supports_conversation: _builtins.bool
|
|
662
637
|
"""Whether the model supports multi-turn conversations
|
|
663
638
|
Example: true for chat models, false for completion-only models
|
|
664
639
|
"""
|
|
665
|
-
supports_context:
|
|
640
|
+
supports_context: _builtins.bool
|
|
666
641
|
"""Whether the model can maintain context across messages
|
|
667
642
|
Example: true for stateful models
|
|
668
643
|
"""
|
|
669
|
-
@property
|
|
644
|
+
@_builtins.property
|
|
670
645
|
def supported_languages(
|
|
671
646
|
self,
|
|
672
|
-
) ->
|
|
647
|
+
) -> _containers.RepeatedScalarFieldContainer[_builtins.str]:
|
|
673
648
|
"""Supported languages for text processing
|
|
674
649
|
Examples: ["en", "es", "fr", "de", "zh", "ja"]
|
|
675
650
|
"""
|
|
@@ -677,88 +652,97 @@ class Text(google.protobuf.message.Message):
|
|
|
677
652
|
def __init__(
|
|
678
653
|
self,
|
|
679
654
|
*,
|
|
680
|
-
direction:
|
|
681
|
-
max_input_length:
|
|
682
|
-
max_output_length:
|
|
683
|
-
supported_languages:
|
|
684
|
-
supports_conversation:
|
|
685
|
-
supports_context:
|
|
655
|
+
direction: Global___ModalityDirection.ValueType | None = ...,
|
|
656
|
+
max_input_length: _builtins.int | None = ...,
|
|
657
|
+
max_output_length: _builtins.int | None = ...,
|
|
658
|
+
supported_languages: _abc.Iterable[_builtins.str] | None = ...,
|
|
659
|
+
supports_conversation: _builtins.bool | None = ...,
|
|
660
|
+
supports_context: _builtins.bool | None = ...,
|
|
686
661
|
) -> None: ...
|
|
687
|
-
|
|
688
|
-
|
|
689
|
-
|
|
690
|
-
|
|
691
|
-
|
|
692
|
-
|
|
693
|
-
|
|
694
|
-
|
|
695
|
-
|
|
696
|
-
|
|
697
|
-
|
|
698
|
-
|
|
699
|
-
|
|
700
|
-
|
|
701
|
-
|
|
702
|
-
|
|
703
|
-
|
|
704
|
-
|
|
705
|
-
|
|
706
|
-
|
|
707
|
-
|
|
708
|
-
|
|
662
|
+
_HasFieldArgType: _TypeAlias = _typing.Literal[
|
|
663
|
+
"direction",
|
|
664
|
+
b"direction",
|
|
665
|
+
"max_input_length",
|
|
666
|
+
b"max_input_length",
|
|
667
|
+
"max_output_length",
|
|
668
|
+
b"max_output_length",
|
|
669
|
+
"supported_languages",
|
|
670
|
+
b"supported_languages",
|
|
671
|
+
"supports_context",
|
|
672
|
+
b"supports_context",
|
|
673
|
+
"supports_conversation",
|
|
674
|
+
b"supports_conversation",
|
|
675
|
+
] # noqa: Y015
|
|
676
|
+
def HasField(self, field_name: _HasFieldArgType) -> _builtins.bool: ...
|
|
677
|
+
_ClearFieldArgType: _TypeAlias = _typing.Literal[
|
|
678
|
+
"direction",
|
|
679
|
+
b"direction",
|
|
680
|
+
"max_input_length",
|
|
681
|
+
b"max_input_length",
|
|
682
|
+
"max_output_length",
|
|
683
|
+
b"max_output_length",
|
|
684
|
+
"supported_languages",
|
|
685
|
+
b"supported_languages",
|
|
686
|
+
"supports_context",
|
|
687
|
+
b"supports_context",
|
|
688
|
+
"supports_conversation",
|
|
689
|
+
b"supports_conversation",
|
|
690
|
+
] # noqa: Y015
|
|
691
|
+
def ClearField(self, field_name: _ClearFieldArgType) -> None: ...
|
|
692
|
+
|
|
693
|
+
Global___Text: _TypeAlias = Text # noqa: Y015
|
|
694
|
+
|
|
695
|
+
@_typing.final
|
|
696
|
+
class StructuredResponse(_message.Message):
|
|
709
697
|
"""Structured response capability configuration"""
|
|
710
698
|
|
|
711
|
-
DESCRIPTOR:
|
|
712
|
-
|
|
713
|
-
SYSTEM_PROMPT_HINT_FIELD_NUMBER:
|
|
714
|
-
SUPPORTED_FORMATS_FIELD_NUMBER:
|
|
715
|
-
MAX_SCHEMA_DEPTH_FIELD_NUMBER:
|
|
716
|
-
REQUIRES_TOOL_USE_FIELD_NUMBER:
|
|
717
|
-
REQUIRES_JSON_MODE_FIELD_NUMBER:
|
|
718
|
-
MAX_PROPERTIES_FIELD_NUMBER:
|
|
719
|
-
SUPPORTED_TYPES_FIELD_NUMBER:
|
|
720
|
-
SUPPORTS_STREAMING_FIELD_NUMBER:
|
|
721
|
-
system_prompt_hint:
|
|
699
|
+
DESCRIPTOR: _descriptor.Descriptor
|
|
700
|
+
|
|
701
|
+
SYSTEM_PROMPT_HINT_FIELD_NUMBER: _builtins.int
|
|
702
|
+
SUPPORTED_FORMATS_FIELD_NUMBER: _builtins.int
|
|
703
|
+
MAX_SCHEMA_DEPTH_FIELD_NUMBER: _builtins.int
|
|
704
|
+
REQUIRES_TOOL_USE_FIELD_NUMBER: _builtins.int
|
|
705
|
+
REQUIRES_JSON_MODE_FIELD_NUMBER: _builtins.int
|
|
706
|
+
MAX_PROPERTIES_FIELD_NUMBER: _builtins.int
|
|
707
|
+
SUPPORTED_TYPES_FIELD_NUMBER: _builtins.int
|
|
708
|
+
SUPPORTS_STREAMING_FIELD_NUMBER: _builtins.int
|
|
709
|
+
system_prompt_hint: _builtins.str
|
|
722
710
|
"""Optional guidance to include in the system prompt to elicit structured output
|
|
723
711
|
Example: "Always return valid JSON matching the provided schema"
|
|
724
712
|
"""
|
|
725
|
-
max_schema_depth:
|
|
713
|
+
max_schema_depth: _builtins.int
|
|
726
714
|
"""Maximum allowed nesting depth for JSON schema objects/arrays
|
|
727
715
|
Example: 5 for moderately complex schemas
|
|
728
716
|
"""
|
|
729
|
-
requires_tool_use:
|
|
717
|
+
requires_tool_use: _builtins.bool
|
|
730
718
|
"""Whether structured output requires tool/function use (vs. direct generation)
|
|
731
719
|
Example: true for providers that only return JSON via tool calls
|
|
732
720
|
"""
|
|
733
|
-
requires_json_mode:
|
|
721
|
+
requires_json_mode: _builtins.bool
|
|
734
722
|
"""Whether the model must be in a special "JSON mode" to honor schemas
|
|
735
723
|
Example: true for models that enforce strict JSON output when enabled
|
|
736
724
|
"""
|
|
737
|
-
max_properties:
|
|
725
|
+
max_properties: _builtins.int
|
|
738
726
|
"""Maximum number of top-level properties in object schemas
|
|
739
727
|
Example: 50
|
|
740
728
|
"""
|
|
741
|
-
supports_streaming:
|
|
729
|
+
supports_streaming: _builtins.bool
|
|
742
730
|
"""Can stream structured responses (not just generate them)
|
|
743
731
|
Example: true if partial JSON chunks are streamed
|
|
744
732
|
Can stream structured responses (not just generate them)
|
|
745
733
|
"""
|
|
746
|
-
@property
|
|
734
|
+
@_builtins.property
|
|
747
735
|
def supported_formats(
|
|
748
736
|
self,
|
|
749
|
-
) ->
|
|
750
|
-
global___DataFormat.ValueType
|
|
751
|
-
]:
|
|
737
|
+
) -> _containers.RepeatedScalarFieldContainer[Global___DataFormat.ValueType]:
|
|
752
738
|
"""Supported output formats for structured responses
|
|
753
739
|
Examples: [DATA_FORMAT_JSON, DATA_FORMAT_YAML]
|
|
754
740
|
"""
|
|
755
741
|
|
|
756
|
-
@property
|
|
742
|
+
@_builtins.property
|
|
757
743
|
def supported_types(
|
|
758
744
|
self,
|
|
759
|
-
) ->
|
|
760
|
-
global___JsonSchemaType.ValueType
|
|
761
|
-
]:
|
|
745
|
+
) -> _containers.RepeatedScalarFieldContainer[Global___JsonSchemaType.ValueType]:
|
|
762
746
|
"""JSON Schema primitive/types supported by the model
|
|
763
747
|
Examples: [JSON_SCHEMA_TYPE_OBJECT, JSON_SCHEMA_TYPE_ARRAY, JSON_SCHEMA_TYPE_STRING]
|
|
764
748
|
"""
|
|
@@ -766,150 +750,176 @@ class StructuredResponse(google.protobuf.message.Message):
|
|
|
766
750
|
def __init__(
|
|
767
751
|
self,
|
|
768
752
|
*,
|
|
769
|
-
system_prompt_hint:
|
|
770
|
-
supported_formats:
|
|
771
|
-
| None = ...,
|
|
772
|
-
|
|
773
|
-
|
|
774
|
-
|
|
775
|
-
|
|
776
|
-
|
|
777
|
-
| None = ...,
|
|
778
|
-
supports_streaming: builtins.bool = ...,
|
|
779
|
-
) -> None: ...
|
|
780
|
-
def ClearField(
|
|
781
|
-
self,
|
|
782
|
-
field_name: typing.Literal[
|
|
783
|
-
"max_properties",
|
|
784
|
-
b"max_properties",
|
|
785
|
-
"max_schema_depth",
|
|
786
|
-
b"max_schema_depth",
|
|
787
|
-
"requires_json_mode",
|
|
788
|
-
b"requires_json_mode",
|
|
789
|
-
"requires_tool_use",
|
|
790
|
-
b"requires_tool_use",
|
|
791
|
-
"supported_formats",
|
|
792
|
-
b"supported_formats",
|
|
793
|
-
"supported_types",
|
|
794
|
-
b"supported_types",
|
|
795
|
-
"supports_streaming",
|
|
796
|
-
b"supports_streaming",
|
|
797
|
-
"system_prompt_hint",
|
|
798
|
-
b"system_prompt_hint",
|
|
799
|
-
],
|
|
753
|
+
system_prompt_hint: _builtins.str | None = ...,
|
|
754
|
+
supported_formats: _abc.Iterable[Global___DataFormat.ValueType] | None = ...,
|
|
755
|
+
max_schema_depth: _builtins.int | None = ...,
|
|
756
|
+
requires_tool_use: _builtins.bool | None = ...,
|
|
757
|
+
requires_json_mode: _builtins.bool | None = ...,
|
|
758
|
+
max_properties: _builtins.int | None = ...,
|
|
759
|
+
supported_types: _abc.Iterable[Global___JsonSchemaType.ValueType] | None = ...,
|
|
760
|
+
supports_streaming: _builtins.bool | None = ...,
|
|
800
761
|
) -> None: ...
|
|
801
|
-
|
|
802
|
-
|
|
803
|
-
|
|
804
|
-
|
|
805
|
-
|
|
762
|
+
_HasFieldArgType: _TypeAlias = _typing.Literal[
|
|
763
|
+
"max_properties",
|
|
764
|
+
b"max_properties",
|
|
765
|
+
"max_schema_depth",
|
|
766
|
+
b"max_schema_depth",
|
|
767
|
+
"requires_json_mode",
|
|
768
|
+
b"requires_json_mode",
|
|
769
|
+
"requires_tool_use",
|
|
770
|
+
b"requires_tool_use",
|
|
771
|
+
"supported_formats",
|
|
772
|
+
b"supported_formats",
|
|
773
|
+
"supported_types",
|
|
774
|
+
b"supported_types",
|
|
775
|
+
"supports_streaming",
|
|
776
|
+
b"supports_streaming",
|
|
777
|
+
"system_prompt_hint",
|
|
778
|
+
b"system_prompt_hint",
|
|
779
|
+
] # noqa: Y015
|
|
780
|
+
def HasField(self, field_name: _HasFieldArgType) -> _builtins.bool: ...
|
|
781
|
+
_ClearFieldArgType: _TypeAlias = _typing.Literal[
|
|
782
|
+
"max_properties",
|
|
783
|
+
b"max_properties",
|
|
784
|
+
"max_schema_depth",
|
|
785
|
+
b"max_schema_depth",
|
|
786
|
+
"requires_json_mode",
|
|
787
|
+
b"requires_json_mode",
|
|
788
|
+
"requires_tool_use",
|
|
789
|
+
b"requires_tool_use",
|
|
790
|
+
"supported_formats",
|
|
791
|
+
b"supported_formats",
|
|
792
|
+
"supported_types",
|
|
793
|
+
b"supported_types",
|
|
794
|
+
"supports_streaming",
|
|
795
|
+
b"supports_streaming",
|
|
796
|
+
"system_prompt_hint",
|
|
797
|
+
b"system_prompt_hint",
|
|
798
|
+
] # noqa: Y015
|
|
799
|
+
def ClearField(self, field_name: _ClearFieldArgType) -> None: ...
|
|
800
|
+
|
|
801
|
+
Global___StructuredResponse: _TypeAlias = StructuredResponse # noqa: Y015
|
|
802
|
+
|
|
803
|
+
@_typing.final
|
|
804
|
+
class Streaming(_message.Message):
|
|
806
805
|
"""Streaming capability configuration"""
|
|
807
806
|
|
|
808
|
-
DESCRIPTOR:
|
|
807
|
+
DESCRIPTOR: _descriptor.Descriptor
|
|
809
808
|
|
|
810
|
-
CHUNK_DELIMITER_FIELD_NUMBER:
|
|
811
|
-
BUFFER_SIZE_FIELD_NUMBER:
|
|
812
|
-
SUPPORTS_SSE_FIELD_NUMBER:
|
|
813
|
-
SUPPORTS_USAGE_FIELD_NUMBER:
|
|
814
|
-
AVG_CHUNK_SIZE_BYTES_FIELD_NUMBER:
|
|
815
|
-
MAX_CHUNK_DELAY_MS_FIELD_NUMBER:
|
|
816
|
-
chunk_delimiter:
|
|
809
|
+
CHUNK_DELIMITER_FIELD_NUMBER: _builtins.int
|
|
810
|
+
BUFFER_SIZE_FIELD_NUMBER: _builtins.int
|
|
811
|
+
SUPPORTS_SSE_FIELD_NUMBER: _builtins.int
|
|
812
|
+
SUPPORTS_USAGE_FIELD_NUMBER: _builtins.int
|
|
813
|
+
AVG_CHUNK_SIZE_BYTES_FIELD_NUMBER: _builtins.int
|
|
814
|
+
MAX_CHUNK_DELAY_MS_FIELD_NUMBER: _builtins.int
|
|
815
|
+
chunk_delimiter: _builtins.str
|
|
817
816
|
"""Delimiter used between streamed chunks, if any
|
|
818
817
|
Example: "\\n\\n" for event-stream boundaries
|
|
819
818
|
"""
|
|
820
|
-
buffer_size:
|
|
819
|
+
buffer_size: _builtins.int
|
|
821
820
|
"""Preferred buffer size used by the provider when batching output
|
|
822
821
|
Example: 1024 (bytes)
|
|
823
822
|
"""
|
|
824
|
-
supports_sse:
|
|
823
|
+
supports_sse: _builtins.bool
|
|
825
824
|
"""Whether Server-Sent Events (SSE) is supported
|
|
826
825
|
Example: true if provider uses text/event-stream
|
|
827
826
|
"""
|
|
828
|
-
supports_usage:
|
|
827
|
+
supports_usage: _builtins.bool
|
|
829
828
|
"""Whether token usage info can be emitted incrementally during streaming
|
|
830
829
|
Example: true if usage deltas are included in stream
|
|
831
830
|
"""
|
|
832
|
-
avg_chunk_size_bytes:
|
|
831
|
+
avg_chunk_size_bytes: _builtins.int
|
|
833
832
|
"""Typical size of individual streamed chunks in bytes
|
|
834
833
|
Example: 256
|
|
835
834
|
"""
|
|
836
|
-
max_chunk_delay_ms:
|
|
835
|
+
max_chunk_delay_ms: _builtins.int
|
|
837
836
|
"""Maximum delay between streamed chunks in milliseconds
|
|
838
837
|
Example: 200
|
|
839
838
|
"""
|
|
840
839
|
def __init__(
|
|
841
840
|
self,
|
|
842
841
|
*,
|
|
843
|
-
chunk_delimiter:
|
|
844
|
-
buffer_size:
|
|
845
|
-
supports_sse:
|
|
846
|
-
supports_usage:
|
|
847
|
-
avg_chunk_size_bytes:
|
|
848
|
-
max_chunk_delay_ms:
|
|
849
|
-
) -> None: ...
|
|
850
|
-
def ClearField(
|
|
851
|
-
self,
|
|
852
|
-
field_name: typing.Literal[
|
|
853
|
-
"avg_chunk_size_bytes",
|
|
854
|
-
b"avg_chunk_size_bytes",
|
|
855
|
-
"buffer_size",
|
|
856
|
-
b"buffer_size",
|
|
857
|
-
"chunk_delimiter",
|
|
858
|
-
b"chunk_delimiter",
|
|
859
|
-
"max_chunk_delay_ms",
|
|
860
|
-
b"max_chunk_delay_ms",
|
|
861
|
-
"supports_sse",
|
|
862
|
-
b"supports_sse",
|
|
863
|
-
"supports_usage",
|
|
864
|
-
b"supports_usage",
|
|
865
|
-
],
|
|
842
|
+
chunk_delimiter: _builtins.str | None = ...,
|
|
843
|
+
buffer_size: _builtins.int | None = ...,
|
|
844
|
+
supports_sse: _builtins.bool | None = ...,
|
|
845
|
+
supports_usage: _builtins.bool | None = ...,
|
|
846
|
+
avg_chunk_size_bytes: _builtins.int | None = ...,
|
|
847
|
+
max_chunk_delay_ms: _builtins.int | None = ...,
|
|
866
848
|
) -> None: ...
|
|
867
|
-
|
|
868
|
-
|
|
869
|
-
|
|
870
|
-
|
|
871
|
-
|
|
849
|
+
_HasFieldArgType: _TypeAlias = _typing.Literal[
|
|
850
|
+
"avg_chunk_size_bytes",
|
|
851
|
+
b"avg_chunk_size_bytes",
|
|
852
|
+
"buffer_size",
|
|
853
|
+
b"buffer_size",
|
|
854
|
+
"chunk_delimiter",
|
|
855
|
+
b"chunk_delimiter",
|
|
856
|
+
"max_chunk_delay_ms",
|
|
857
|
+
b"max_chunk_delay_ms",
|
|
858
|
+
"supports_sse",
|
|
859
|
+
b"supports_sse",
|
|
860
|
+
"supports_usage",
|
|
861
|
+
b"supports_usage",
|
|
862
|
+
] # noqa: Y015
|
|
863
|
+
def HasField(self, field_name: _HasFieldArgType) -> _builtins.bool: ...
|
|
864
|
+
_ClearFieldArgType: _TypeAlias = _typing.Literal[
|
|
865
|
+
"avg_chunk_size_bytes",
|
|
866
|
+
b"avg_chunk_size_bytes",
|
|
867
|
+
"buffer_size",
|
|
868
|
+
b"buffer_size",
|
|
869
|
+
"chunk_delimiter",
|
|
870
|
+
b"chunk_delimiter",
|
|
871
|
+
"max_chunk_delay_ms",
|
|
872
|
+
b"max_chunk_delay_ms",
|
|
873
|
+
"supports_sse",
|
|
874
|
+
b"supports_sse",
|
|
875
|
+
"supports_usage",
|
|
876
|
+
b"supports_usage",
|
|
877
|
+
] # noqa: Y015
|
|
878
|
+
def ClearField(self, field_name: _ClearFieldArgType) -> None: ...
|
|
879
|
+
|
|
880
|
+
Global___Streaming: _TypeAlias = Streaming # noqa: Y015
|
|
881
|
+
|
|
882
|
+
@_typing.final
|
|
883
|
+
class FunctionCalling(_message.Message):
|
|
872
884
|
"""Function calling capability configuration"""
|
|
873
885
|
|
|
874
|
-
DESCRIPTOR:
|
|
886
|
+
DESCRIPTOR: _descriptor.Descriptor
|
|
875
887
|
|
|
876
|
-
MAX_FUNCTIONS_FIELD_NUMBER:
|
|
877
|
-
MAX_PARALLEL_CALLS_FIELD_NUMBER:
|
|
878
|
-
SUPPORTS_PARALLEL_FIELD_NUMBER:
|
|
879
|
-
REQUIRES_TOOL_ROLE_FIELD_NUMBER:
|
|
880
|
-
SUPPORTS_STREAMING_FIELD_NUMBER:
|
|
881
|
-
SUPPORTED_PARAMETER_TYPES_FIELD_NUMBER:
|
|
882
|
-
MAX_NESTING_DEPTH_FIELD_NUMBER:
|
|
883
|
-
max_functions:
|
|
888
|
+
MAX_FUNCTIONS_FIELD_NUMBER: _builtins.int
|
|
889
|
+
MAX_PARALLEL_CALLS_FIELD_NUMBER: _builtins.int
|
|
890
|
+
SUPPORTS_PARALLEL_FIELD_NUMBER: _builtins.int
|
|
891
|
+
REQUIRES_TOOL_ROLE_FIELD_NUMBER: _builtins.int
|
|
892
|
+
SUPPORTS_STREAMING_FIELD_NUMBER: _builtins.int
|
|
893
|
+
SUPPORTED_PARAMETER_TYPES_FIELD_NUMBER: _builtins.int
|
|
894
|
+
MAX_NESTING_DEPTH_FIELD_NUMBER: _builtins.int
|
|
895
|
+
max_functions: _builtins.int
|
|
884
896
|
"""Maximum number of functions that can be registered per request
|
|
885
897
|
Example: 128
|
|
886
898
|
"""
|
|
887
|
-
max_parallel_calls:
|
|
899
|
+
max_parallel_calls: _builtins.int
|
|
888
900
|
"""Maximum number of tool/function calls that may be executed in parallel
|
|
889
901
|
Example: 5
|
|
890
902
|
"""
|
|
891
|
-
supports_parallel:
|
|
903
|
+
supports_parallel: _builtins.bool
|
|
892
904
|
"""Whether the model can plan and invoke multiple tools concurrently
|
|
893
905
|
Example: true for models with parallel tool-use orchestration
|
|
894
906
|
"""
|
|
895
|
-
requires_tool_role:
|
|
907
|
+
requires_tool_role: _builtins.bool
|
|
896
908
|
"""Whether requests must use a special "tool" role or channel for calls
|
|
897
909
|
Example: true if provider requires tool role messages
|
|
898
910
|
"""
|
|
899
|
-
supports_streaming:
|
|
911
|
+
supports_streaming: _builtins.bool
|
|
900
912
|
"""Whether intermediate tool call tokens/results are streamable
|
|
901
913
|
Example: true if tool call arguments/results are streamed as they are produced
|
|
902
914
|
"""
|
|
903
|
-
max_nesting_depth:
|
|
915
|
+
max_nesting_depth: _builtins.int
|
|
904
916
|
"""Maximum nesting depth allowed for function/tool call plans
|
|
905
917
|
Example: 3 for limited recursion
|
|
906
918
|
"""
|
|
907
|
-
@property
|
|
919
|
+
@_builtins.property
|
|
908
920
|
def supported_parameter_types(
|
|
909
921
|
self,
|
|
910
|
-
) ->
|
|
911
|
-
global___JsonSchemaType.ValueType
|
|
912
|
-
]:
|
|
922
|
+
) -> _containers.RepeatedScalarFieldContainer[Global___JsonSchemaType.ValueType]:
|
|
913
923
|
"""JSON schema types supported for function parameters
|
|
914
924
|
Examples: [JSON_SCHEMA_TYPE_OBJECT, JSON_SCHEMA_TYPE_ARRAY, JSON_SCHEMA_TYPE_STRING]
|
|
915
925
|
"""
|
|
@@ -917,94 +927,105 @@ class FunctionCalling(google.protobuf.message.Message):
|
|
|
917
927
|
def __init__(
|
|
918
928
|
self,
|
|
919
929
|
*,
|
|
920
|
-
max_functions:
|
|
921
|
-
max_parallel_calls:
|
|
922
|
-
supports_parallel:
|
|
923
|
-
requires_tool_role:
|
|
924
|
-
supports_streaming:
|
|
925
|
-
supported_parameter_types:
|
|
926
|
-
global___JsonSchemaType.ValueType
|
|
927
|
-
]
|
|
930
|
+
max_functions: _builtins.int | None = ...,
|
|
931
|
+
max_parallel_calls: _builtins.int | None = ...,
|
|
932
|
+
supports_parallel: _builtins.bool | None = ...,
|
|
933
|
+
requires_tool_role: _builtins.bool | None = ...,
|
|
934
|
+
supports_streaming: _builtins.bool | None = ...,
|
|
935
|
+
supported_parameter_types: _abc.Iterable[Global___JsonSchemaType.ValueType]
|
|
928
936
|
| None = ...,
|
|
929
|
-
max_nesting_depth:
|
|
930
|
-
) -> None: ...
|
|
931
|
-
def ClearField(
|
|
932
|
-
self,
|
|
933
|
-
field_name: typing.Literal[
|
|
934
|
-
"max_functions",
|
|
935
|
-
b"max_functions",
|
|
936
|
-
"max_nesting_depth",
|
|
937
|
-
b"max_nesting_depth",
|
|
938
|
-
"max_parallel_calls",
|
|
939
|
-
b"max_parallel_calls",
|
|
940
|
-
"requires_tool_role",
|
|
941
|
-
b"requires_tool_role",
|
|
942
|
-
"supported_parameter_types",
|
|
943
|
-
b"supported_parameter_types",
|
|
944
|
-
"supports_parallel",
|
|
945
|
-
b"supports_parallel",
|
|
946
|
-
"supports_streaming",
|
|
947
|
-
b"supports_streaming",
|
|
948
|
-
],
|
|
937
|
+
max_nesting_depth: _builtins.int | None = ...,
|
|
949
938
|
) -> None: ...
|
|
950
|
-
|
|
951
|
-
|
|
952
|
-
|
|
953
|
-
|
|
954
|
-
|
|
939
|
+
_HasFieldArgType: _TypeAlias = _typing.Literal[
|
|
940
|
+
"max_functions",
|
|
941
|
+
b"max_functions",
|
|
942
|
+
"max_nesting_depth",
|
|
943
|
+
b"max_nesting_depth",
|
|
944
|
+
"max_parallel_calls",
|
|
945
|
+
b"max_parallel_calls",
|
|
946
|
+
"requires_tool_role",
|
|
947
|
+
b"requires_tool_role",
|
|
948
|
+
"supported_parameter_types",
|
|
949
|
+
b"supported_parameter_types",
|
|
950
|
+
"supports_parallel",
|
|
951
|
+
b"supports_parallel",
|
|
952
|
+
"supports_streaming",
|
|
953
|
+
b"supports_streaming",
|
|
954
|
+
] # noqa: Y015
|
|
955
|
+
def HasField(self, field_name: _HasFieldArgType) -> _builtins.bool: ...
|
|
956
|
+
_ClearFieldArgType: _TypeAlias = _typing.Literal[
|
|
957
|
+
"max_functions",
|
|
958
|
+
b"max_functions",
|
|
959
|
+
"max_nesting_depth",
|
|
960
|
+
b"max_nesting_depth",
|
|
961
|
+
"max_parallel_calls",
|
|
962
|
+
b"max_parallel_calls",
|
|
963
|
+
"requires_tool_role",
|
|
964
|
+
b"requires_tool_role",
|
|
965
|
+
"supported_parameter_types",
|
|
966
|
+
b"supported_parameter_types",
|
|
967
|
+
"supports_parallel",
|
|
968
|
+
b"supports_parallel",
|
|
969
|
+
"supports_streaming",
|
|
970
|
+
b"supports_streaming",
|
|
971
|
+
] # noqa: Y015
|
|
972
|
+
def ClearField(self, field_name: _ClearFieldArgType) -> None: ...
|
|
973
|
+
|
|
974
|
+
Global___FunctionCalling: _TypeAlias = FunctionCalling # noqa: Y015
|
|
975
|
+
|
|
976
|
+
@_typing.final
|
|
977
|
+
class Vision(_message.Message):
|
|
955
978
|
"""Vision capability configuration"""
|
|
956
979
|
|
|
957
|
-
DESCRIPTOR:
|
|
958
|
-
|
|
959
|
-
DIRECTION_FIELD_NUMBER:
|
|
960
|
-
SUPPORTED_FORMATS_FIELD_NUMBER:
|
|
961
|
-
MAX_IMAGE_SIZE_BYTES_FIELD_NUMBER:
|
|
962
|
-
MAX_IMAGES_PER_REQUEST_FIELD_NUMBER:
|
|
963
|
-
MAX_RESOLUTION_WIDTH_FIELD_NUMBER:
|
|
964
|
-
MAX_RESOLUTION_HEIGHT_FIELD_NUMBER:
|
|
965
|
-
SUPPORTS_OCR_FIELD_NUMBER:
|
|
966
|
-
SUPPORTS_OBJECT_DETECTION_FIELD_NUMBER:
|
|
967
|
-
SUPPORTS_VIDEO_FRAMES_FIELD_NUMBER:
|
|
968
|
-
direction:
|
|
980
|
+
DESCRIPTOR: _descriptor.Descriptor
|
|
981
|
+
|
|
982
|
+
DIRECTION_FIELD_NUMBER: _builtins.int
|
|
983
|
+
SUPPORTED_FORMATS_FIELD_NUMBER: _builtins.int
|
|
984
|
+
MAX_IMAGE_SIZE_BYTES_FIELD_NUMBER: _builtins.int
|
|
985
|
+
MAX_IMAGES_PER_REQUEST_FIELD_NUMBER: _builtins.int
|
|
986
|
+
MAX_RESOLUTION_WIDTH_FIELD_NUMBER: _builtins.int
|
|
987
|
+
MAX_RESOLUTION_HEIGHT_FIELD_NUMBER: _builtins.int
|
|
988
|
+
SUPPORTS_OCR_FIELD_NUMBER: _builtins.int
|
|
989
|
+
SUPPORTS_OBJECT_DETECTION_FIELD_NUMBER: _builtins.int
|
|
990
|
+
SUPPORTS_VIDEO_FRAMES_FIELD_NUMBER: _builtins.int
|
|
991
|
+
direction: Global___ModalityDirection.ValueType
|
|
969
992
|
"""Direction of vision support
|
|
970
993
|
Example: MODALITY_DIRECTION_INPUT_ONLY for analysis-only models
|
|
971
994
|
Example: MODALITY_DIRECTION_OUTPUT_ONLY for image generation models (DALL-E)
|
|
972
995
|
Example: MODALITY_DIRECTION_INPUT_OUTPUT for models that can both analyze and generate
|
|
973
996
|
"""
|
|
974
|
-
max_image_size_bytes:
|
|
997
|
+
max_image_size_bytes: _builtins.int
|
|
975
998
|
"""Maximum size per image in bytes
|
|
976
999
|
Example: 20971520 (20MB) for GPT-4-vision
|
|
977
1000
|
"""
|
|
978
|
-
max_images_per_request:
|
|
1001
|
+
max_images_per_request: _builtins.int
|
|
979
1002
|
"""Maximum images per API request (for input)
|
|
980
1003
|
Example: 10 for GPT-4-vision, 1 for some models
|
|
981
1004
|
"""
|
|
982
|
-
max_resolution_width:
|
|
1005
|
+
max_resolution_width: _builtins.int
|
|
983
1006
|
"""Maximum image width in pixels
|
|
984
1007
|
Example: 4096 for high-resolution support
|
|
985
1008
|
"""
|
|
986
|
-
max_resolution_height:
|
|
1009
|
+
max_resolution_height: _builtins.int
|
|
987
1010
|
"""Maximum image height in pixels
|
|
988
1011
|
Example: 4096 for high-resolution support
|
|
989
1012
|
"""
|
|
990
|
-
supports_ocr:
|
|
1013
|
+
supports_ocr: _builtins.bool
|
|
991
1014
|
"""Supports optical character recognition
|
|
992
1015
|
Example: true if model can extract text from images
|
|
993
1016
|
"""
|
|
994
|
-
supports_object_detection:
|
|
1017
|
+
supports_object_detection: _builtins.bool
|
|
995
1018
|
"""Supports object detection/localization
|
|
996
1019
|
Example: true if model can identify and locate objects
|
|
997
1020
|
"""
|
|
998
|
-
supports_video_frames:
|
|
1021
|
+
supports_video_frames: _builtins.bool
|
|
999
1022
|
"""Can process video frames as images
|
|
1000
1023
|
Example: true if model accepts video frame extraction
|
|
1001
1024
|
"""
|
|
1002
|
-
@property
|
|
1025
|
+
@_builtins.property
|
|
1003
1026
|
def supported_formats(
|
|
1004
1027
|
self,
|
|
1005
|
-
) ->
|
|
1006
|
-
global___ImageFormat.ValueType
|
|
1007
|
-
]:
|
|
1028
|
+
) -> _containers.RepeatedScalarFieldContainer[Global___ImageFormat.ValueType]:
|
|
1008
1029
|
"""Supported image file formats
|
|
1009
1030
|
Examples: [IMAGE_FORMAT_JPEG, IMAGE_FORMAT_PNG, IMAGE_FORMAT_WEBP]
|
|
1010
1031
|
"""
|
|
@@ -1012,76 +1033,92 @@ class Vision(google.protobuf.message.Message):
|
|
|
1012
1033
|
def __init__(
|
|
1013
1034
|
self,
|
|
1014
1035
|
*,
|
|
1015
|
-
direction:
|
|
1016
|
-
supported_formats:
|
|
1017
|
-
| None = ...,
|
|
1018
|
-
|
|
1019
|
-
|
|
1020
|
-
|
|
1021
|
-
|
|
1022
|
-
|
|
1023
|
-
|
|
1024
|
-
supports_video_frames: builtins.bool = ...,
|
|
1025
|
-
) -> None: ...
|
|
1026
|
-
def ClearField(
|
|
1027
|
-
self,
|
|
1028
|
-
field_name: typing.Literal[
|
|
1029
|
-
"direction",
|
|
1030
|
-
b"direction",
|
|
1031
|
-
"max_image_size_bytes",
|
|
1032
|
-
b"max_image_size_bytes",
|
|
1033
|
-
"max_images_per_request",
|
|
1034
|
-
b"max_images_per_request",
|
|
1035
|
-
"max_resolution_height",
|
|
1036
|
-
b"max_resolution_height",
|
|
1037
|
-
"max_resolution_width",
|
|
1038
|
-
b"max_resolution_width",
|
|
1039
|
-
"supported_formats",
|
|
1040
|
-
b"supported_formats",
|
|
1041
|
-
"supports_object_detection",
|
|
1042
|
-
b"supports_object_detection",
|
|
1043
|
-
"supports_ocr",
|
|
1044
|
-
b"supports_ocr",
|
|
1045
|
-
"supports_video_frames",
|
|
1046
|
-
b"supports_video_frames",
|
|
1047
|
-
],
|
|
1036
|
+
direction: Global___ModalityDirection.ValueType | None = ...,
|
|
1037
|
+
supported_formats: _abc.Iterable[Global___ImageFormat.ValueType] | None = ...,
|
|
1038
|
+
max_image_size_bytes: _builtins.int | None = ...,
|
|
1039
|
+
max_images_per_request: _builtins.int | None = ...,
|
|
1040
|
+
max_resolution_width: _builtins.int | None = ...,
|
|
1041
|
+
max_resolution_height: _builtins.int | None = ...,
|
|
1042
|
+
supports_ocr: _builtins.bool | None = ...,
|
|
1043
|
+
supports_object_detection: _builtins.bool | None = ...,
|
|
1044
|
+
supports_video_frames: _builtins.bool | None = ...,
|
|
1048
1045
|
) -> None: ...
|
|
1049
|
-
|
|
1050
|
-
|
|
1051
|
-
|
|
1052
|
-
|
|
1053
|
-
|
|
1046
|
+
_HasFieldArgType: _TypeAlias = _typing.Literal[
|
|
1047
|
+
"direction",
|
|
1048
|
+
b"direction",
|
|
1049
|
+
"max_image_size_bytes",
|
|
1050
|
+
b"max_image_size_bytes",
|
|
1051
|
+
"max_images_per_request",
|
|
1052
|
+
b"max_images_per_request",
|
|
1053
|
+
"max_resolution_height",
|
|
1054
|
+
b"max_resolution_height",
|
|
1055
|
+
"max_resolution_width",
|
|
1056
|
+
b"max_resolution_width",
|
|
1057
|
+
"supported_formats",
|
|
1058
|
+
b"supported_formats",
|
|
1059
|
+
"supports_object_detection",
|
|
1060
|
+
b"supports_object_detection",
|
|
1061
|
+
"supports_ocr",
|
|
1062
|
+
b"supports_ocr",
|
|
1063
|
+
"supports_video_frames",
|
|
1064
|
+
b"supports_video_frames",
|
|
1065
|
+
] # noqa: Y015
|
|
1066
|
+
def HasField(self, field_name: _HasFieldArgType) -> _builtins.bool: ...
|
|
1067
|
+
_ClearFieldArgType: _TypeAlias = _typing.Literal[
|
|
1068
|
+
"direction",
|
|
1069
|
+
b"direction",
|
|
1070
|
+
"max_image_size_bytes",
|
|
1071
|
+
b"max_image_size_bytes",
|
|
1072
|
+
"max_images_per_request",
|
|
1073
|
+
b"max_images_per_request",
|
|
1074
|
+
"max_resolution_height",
|
|
1075
|
+
b"max_resolution_height",
|
|
1076
|
+
"max_resolution_width",
|
|
1077
|
+
b"max_resolution_width",
|
|
1078
|
+
"supported_formats",
|
|
1079
|
+
b"supported_formats",
|
|
1080
|
+
"supports_object_detection",
|
|
1081
|
+
b"supports_object_detection",
|
|
1082
|
+
"supports_ocr",
|
|
1083
|
+
b"supports_ocr",
|
|
1084
|
+
"supports_video_frames",
|
|
1085
|
+
b"supports_video_frames",
|
|
1086
|
+
] # noqa: Y015
|
|
1087
|
+
def ClearField(self, field_name: _ClearFieldArgType) -> None: ...
|
|
1088
|
+
|
|
1089
|
+
Global___Vision: _TypeAlias = Vision # noqa: Y015
|
|
1090
|
+
|
|
1091
|
+
@_typing.final
|
|
1092
|
+
class ToolUse(_message.Message):
|
|
1054
1093
|
"""Tool use capability configuration"""
|
|
1055
1094
|
|
|
1056
|
-
DESCRIPTOR:
|
|
1095
|
+
DESCRIPTOR: _descriptor.Descriptor
|
|
1057
1096
|
|
|
1058
|
-
MAX_TOOLS_FIELD_NUMBER:
|
|
1059
|
-
SUPPORTS_SEQUENTIAL_FIELD_NUMBER:
|
|
1060
|
-
SUPPORTS_PARALLEL_FIELD_NUMBER:
|
|
1061
|
-
MAX_TOOL_ROUNDS_FIELD_NUMBER:
|
|
1062
|
-
SUPPORTED_TOOL_TYPES_FIELD_NUMBER:
|
|
1063
|
-
max_tools:
|
|
1097
|
+
MAX_TOOLS_FIELD_NUMBER: _builtins.int
|
|
1098
|
+
SUPPORTS_SEQUENTIAL_FIELD_NUMBER: _builtins.int
|
|
1099
|
+
SUPPORTS_PARALLEL_FIELD_NUMBER: _builtins.int
|
|
1100
|
+
MAX_TOOL_ROUNDS_FIELD_NUMBER: _builtins.int
|
|
1101
|
+
SUPPORTED_TOOL_TYPES_FIELD_NUMBER: _builtins.int
|
|
1102
|
+
max_tools: _builtins.int
|
|
1064
1103
|
"""Maximum number of tools that can be registered or considered
|
|
1065
1104
|
Example: 64
|
|
1066
1105
|
"""
|
|
1067
|
-
supports_sequential:
|
|
1106
|
+
supports_sequential: _builtins.bool
|
|
1068
1107
|
"""Whether the model can chain tool invocations one after another
|
|
1069
1108
|
Example: true for stepwise tool planning
|
|
1070
1109
|
"""
|
|
1071
|
-
supports_parallel:
|
|
1110
|
+
supports_parallel: _builtins.bool
|
|
1072
1111
|
"""Whether the model can invoke multiple tools concurrently
|
|
1073
1112
|
Example: true for parallelized tool execution
|
|
1074
1113
|
"""
|
|
1075
|
-
max_tool_rounds:
|
|
1114
|
+
max_tool_rounds: _builtins.int
|
|
1076
1115
|
"""Maximum number of tool-use rounds allowed in a single request
|
|
1077
1116
|
Example: 10
|
|
1078
1117
|
"""
|
|
1079
|
-
@property
|
|
1118
|
+
@_builtins.property
|
|
1080
1119
|
def supported_tool_types(
|
|
1081
1120
|
self,
|
|
1082
|
-
) ->
|
|
1083
|
-
global___ToolType.ValueType
|
|
1084
|
-
]:
|
|
1121
|
+
) -> _containers.RepeatedScalarFieldContainer[Global___ToolType.ValueType]:
|
|
1085
1122
|
"""Types of tools supported by the provider/model
|
|
1086
1123
|
Examples: [TOOL_TYPE_FUNCTION, TOOL_TYPE_RETRIEVAL]
|
|
1087
1124
|
"""
|
|
@@ -1089,180 +1126,210 @@ class ToolUse(google.protobuf.message.Message):
|
|
|
1089
1126
|
def __init__(
|
|
1090
1127
|
self,
|
|
1091
1128
|
*,
|
|
1092
|
-
max_tools:
|
|
1093
|
-
supports_sequential:
|
|
1094
|
-
supports_parallel:
|
|
1095
|
-
max_tool_rounds:
|
|
1096
|
-
supported_tool_types:
|
|
1097
|
-
| None = ...,
|
|
1129
|
+
max_tools: _builtins.int | None = ...,
|
|
1130
|
+
supports_sequential: _builtins.bool | None = ...,
|
|
1131
|
+
supports_parallel: _builtins.bool | None = ...,
|
|
1132
|
+
max_tool_rounds: _builtins.int | None = ...,
|
|
1133
|
+
supported_tool_types: _abc.Iterable[Global___ToolType.ValueType] | None = ...,
|
|
1098
1134
|
) -> None: ...
|
|
1099
|
-
|
|
1100
|
-
|
|
1101
|
-
|
|
1102
|
-
|
|
1103
|
-
|
|
1104
|
-
|
|
1105
|
-
|
|
1106
|
-
|
|
1107
|
-
|
|
1108
|
-
|
|
1109
|
-
|
|
1110
|
-
|
|
1111
|
-
|
|
1112
|
-
|
|
1113
|
-
|
|
1114
|
-
|
|
1115
|
-
|
|
1116
|
-
|
|
1117
|
-
|
|
1118
|
-
|
|
1135
|
+
_HasFieldArgType: _TypeAlias = _typing.Literal[
|
|
1136
|
+
"max_tool_rounds",
|
|
1137
|
+
b"max_tool_rounds",
|
|
1138
|
+
"max_tools",
|
|
1139
|
+
b"max_tools",
|
|
1140
|
+
"supported_tool_types",
|
|
1141
|
+
b"supported_tool_types",
|
|
1142
|
+
"supports_parallel",
|
|
1143
|
+
b"supports_parallel",
|
|
1144
|
+
"supports_sequential",
|
|
1145
|
+
b"supports_sequential",
|
|
1146
|
+
] # noqa: Y015
|
|
1147
|
+
def HasField(self, field_name: _HasFieldArgType) -> _builtins.bool: ...
|
|
1148
|
+
_ClearFieldArgType: _TypeAlias = _typing.Literal[
|
|
1149
|
+
"max_tool_rounds",
|
|
1150
|
+
b"max_tool_rounds",
|
|
1151
|
+
"max_tools",
|
|
1152
|
+
b"max_tools",
|
|
1153
|
+
"supported_tool_types",
|
|
1154
|
+
b"supported_tool_types",
|
|
1155
|
+
"supports_parallel",
|
|
1156
|
+
b"supports_parallel",
|
|
1157
|
+
"supports_sequential",
|
|
1158
|
+
b"supports_sequential",
|
|
1159
|
+
] # noqa: Y015
|
|
1160
|
+
def ClearField(self, field_name: _ClearFieldArgType) -> None: ...
|
|
1161
|
+
|
|
1162
|
+
Global___ToolUse: _TypeAlias = ToolUse # noqa: Y015
|
|
1163
|
+
|
|
1164
|
+
@_typing.final
|
|
1165
|
+
class SystemPrompt(_message.Message):
|
|
1119
1166
|
"""System prompt capability configuration"""
|
|
1120
1167
|
|
|
1121
|
-
DESCRIPTOR:
|
|
1168
|
+
DESCRIPTOR: _descriptor.Descriptor
|
|
1122
1169
|
|
|
1123
|
-
MAX_LENGTH_FIELD_NUMBER:
|
|
1124
|
-
SUPPORTS_MULTIPLE_FIELD_NUMBER:
|
|
1125
|
-
SUPPORTS_CACHING_FIELD_NUMBER:
|
|
1126
|
-
FORMAT_FIELD_NUMBER:
|
|
1127
|
-
max_length:
|
|
1170
|
+
MAX_LENGTH_FIELD_NUMBER: _builtins.int
|
|
1171
|
+
SUPPORTS_MULTIPLE_FIELD_NUMBER: _builtins.int
|
|
1172
|
+
SUPPORTS_CACHING_FIELD_NUMBER: _builtins.int
|
|
1173
|
+
FORMAT_FIELD_NUMBER: _builtins.int
|
|
1174
|
+
max_length: _builtins.int
|
|
1128
1175
|
"""Maximum allowed length of system prompt content (characters or tokens)
|
|
1129
1176
|
Example: 8000 (characters)
|
|
1130
1177
|
"""
|
|
1131
|
-
supports_multiple:
|
|
1178
|
+
supports_multiple: _builtins.bool
|
|
1132
1179
|
"""Whether multiple system prompts (or segments) can be supplied
|
|
1133
1180
|
Example: true if the provider supports multiple system messages
|
|
1134
1181
|
"""
|
|
1135
|
-
supports_caching:
|
|
1182
|
+
supports_caching: _builtins.bool
|
|
1136
1183
|
"""Whether system prompts can be cached for re-use
|
|
1137
1184
|
Example: true if provider supports prompt caching hints
|
|
1138
1185
|
"""
|
|
1139
|
-
format:
|
|
1186
|
+
format: Global___DataFormat.ValueType
|
|
1140
1187
|
"""Preferred/required format for system prompts (plain, markdown, etc.)
|
|
1141
1188
|
Example: DATA_FORMAT_MARKDOWN
|
|
1142
1189
|
"""
|
|
1143
1190
|
def __init__(
|
|
1144
1191
|
self,
|
|
1145
1192
|
*,
|
|
1146
|
-
max_length:
|
|
1147
|
-
supports_multiple:
|
|
1148
|
-
supports_caching:
|
|
1149
|
-
format:
|
|
1193
|
+
max_length: _builtins.int | None = ...,
|
|
1194
|
+
supports_multiple: _builtins.bool | None = ...,
|
|
1195
|
+
supports_caching: _builtins.bool | None = ...,
|
|
1196
|
+
format: Global___DataFormat.ValueType | None = ...,
|
|
1150
1197
|
) -> None: ...
|
|
1151
|
-
|
|
1152
|
-
|
|
1153
|
-
|
|
1154
|
-
|
|
1155
|
-
|
|
1156
|
-
|
|
1157
|
-
|
|
1158
|
-
|
|
1159
|
-
|
|
1160
|
-
|
|
1161
|
-
|
|
1162
|
-
|
|
1163
|
-
|
|
1164
|
-
|
|
1165
|
-
|
|
1166
|
-
|
|
1167
|
-
|
|
1168
|
-
|
|
1198
|
+
_HasFieldArgType: _TypeAlias = _typing.Literal[
|
|
1199
|
+
"format",
|
|
1200
|
+
b"format",
|
|
1201
|
+
"max_length",
|
|
1202
|
+
b"max_length",
|
|
1203
|
+
"supports_caching",
|
|
1204
|
+
b"supports_caching",
|
|
1205
|
+
"supports_multiple",
|
|
1206
|
+
b"supports_multiple",
|
|
1207
|
+
] # noqa: Y015
|
|
1208
|
+
def HasField(self, field_name: _HasFieldArgType) -> _builtins.bool: ...
|
|
1209
|
+
_ClearFieldArgType: _TypeAlias = _typing.Literal[
|
|
1210
|
+
"format",
|
|
1211
|
+
b"format",
|
|
1212
|
+
"max_length",
|
|
1213
|
+
b"max_length",
|
|
1214
|
+
"supports_caching",
|
|
1215
|
+
b"supports_caching",
|
|
1216
|
+
"supports_multiple",
|
|
1217
|
+
b"supports_multiple",
|
|
1218
|
+
] # noqa: Y015
|
|
1219
|
+
def ClearField(self, field_name: _ClearFieldArgType) -> None: ...
|
|
1220
|
+
|
|
1221
|
+
Global___SystemPrompt: _TypeAlias = SystemPrompt # noqa: Y015
|
|
1222
|
+
|
|
1223
|
+
@_typing.final
|
|
1224
|
+
class Caching(_message.Message):
|
|
1169
1225
|
"""Caching capability configuration"""
|
|
1170
1226
|
|
|
1171
|
-
DESCRIPTOR:
|
|
1227
|
+
DESCRIPTOR: _descriptor.Descriptor
|
|
1172
1228
|
|
|
1173
|
-
CACHE_KEY_STRATEGY_FIELD_NUMBER:
|
|
1174
|
-
MAX_CACHE_SIZE_BYTES_FIELD_NUMBER:
|
|
1175
|
-
CACHE_TTL_SECONDS_FIELD_NUMBER:
|
|
1176
|
-
SUPPORTS_CONTEXT_CACHING_FIELD_NUMBER:
|
|
1177
|
-
SUPPORTS_PROMPT_CACHING_FIELD_NUMBER:
|
|
1178
|
-
MIN_CACHEABLE_TOKENS_FIELD_NUMBER:
|
|
1179
|
-
cache_key_strategy:
|
|
1229
|
+
CACHE_KEY_STRATEGY_FIELD_NUMBER: _builtins.int
|
|
1230
|
+
MAX_CACHE_SIZE_BYTES_FIELD_NUMBER: _builtins.int
|
|
1231
|
+
CACHE_TTL_SECONDS_FIELD_NUMBER: _builtins.int
|
|
1232
|
+
SUPPORTS_CONTEXT_CACHING_FIELD_NUMBER: _builtins.int
|
|
1233
|
+
SUPPORTS_PROMPT_CACHING_FIELD_NUMBER: _builtins.int
|
|
1234
|
+
MIN_CACHEABLE_TOKENS_FIELD_NUMBER: _builtins.int
|
|
1235
|
+
cache_key_strategy: Global___CacheStrategy.ValueType
|
|
1180
1236
|
"""Strategy used to compute cache keys for requests/responses
|
|
1181
1237
|
Examples: CACHE_STRATEGY_HASH, CACHE_STRATEGY_SEMANTIC
|
|
1182
1238
|
"""
|
|
1183
|
-
max_cache_size_bytes:
|
|
1239
|
+
max_cache_size_bytes: _builtins.int
|
|
1184
1240
|
"""Maximum total cache capacity (bytes)
|
|
1185
1241
|
Example: 1073741824 for 1GB
|
|
1186
1242
|
"""
|
|
1187
|
-
cache_ttl_seconds:
|
|
1243
|
+
cache_ttl_seconds: _builtins.int
|
|
1188
1244
|
"""Default time-to-live for cache entries (seconds)
|
|
1189
1245
|
Example: 3600 for 1 hour
|
|
1190
1246
|
"""
|
|
1191
|
-
supports_context_caching:
|
|
1247
|
+
supports_context_caching: _builtins.bool
|
|
1192
1248
|
"""Whether the model/provider supports caching of conversation context
|
|
1193
1249
|
Example: true for context window reuse
|
|
1194
1250
|
"""
|
|
1195
|
-
supports_prompt_caching:
|
|
1251
|
+
supports_prompt_caching: _builtins.bool
|
|
1196
1252
|
"""Whether the model/provider supports caching of prompts/prefixes
|
|
1197
1253
|
Example: true for prefix caching/token reuse
|
|
1198
1254
|
"""
|
|
1199
|
-
min_cacheable_tokens:
|
|
1255
|
+
min_cacheable_tokens: _builtins.int
|
|
1200
1256
|
"""Minimum number of tokens required for an entry to be cacheable
|
|
1201
1257
|
Example: 128
|
|
1202
1258
|
"""
|
|
1203
1259
|
def __init__(
|
|
1204
1260
|
self,
|
|
1205
1261
|
*,
|
|
1206
|
-
cache_key_strategy:
|
|
1207
|
-
max_cache_size_bytes:
|
|
1208
|
-
cache_ttl_seconds:
|
|
1209
|
-
supports_context_caching:
|
|
1210
|
-
supports_prompt_caching:
|
|
1211
|
-
min_cacheable_tokens:
|
|
1262
|
+
cache_key_strategy: Global___CacheStrategy.ValueType | None = ...,
|
|
1263
|
+
max_cache_size_bytes: _builtins.int | None = ...,
|
|
1264
|
+
cache_ttl_seconds: _builtins.int | None = ...,
|
|
1265
|
+
supports_context_caching: _builtins.bool | None = ...,
|
|
1266
|
+
supports_prompt_caching: _builtins.bool | None = ...,
|
|
1267
|
+
min_cacheable_tokens: _builtins.int | None = ...,
|
|
1212
1268
|
) -> None: ...
|
|
1213
|
-
|
|
1214
|
-
|
|
1215
|
-
|
|
1216
|
-
|
|
1217
|
-
|
|
1218
|
-
|
|
1219
|
-
|
|
1220
|
-
|
|
1221
|
-
|
|
1222
|
-
|
|
1223
|
-
|
|
1224
|
-
|
|
1225
|
-
|
|
1226
|
-
|
|
1227
|
-
|
|
1228
|
-
|
|
1229
|
-
|
|
1230
|
-
|
|
1231
|
-
|
|
1232
|
-
|
|
1233
|
-
|
|
1234
|
-
|
|
1269
|
+
_HasFieldArgType: _TypeAlias = _typing.Literal[
|
|
1270
|
+
"cache_key_strategy",
|
|
1271
|
+
b"cache_key_strategy",
|
|
1272
|
+
"cache_ttl_seconds",
|
|
1273
|
+
b"cache_ttl_seconds",
|
|
1274
|
+
"max_cache_size_bytes",
|
|
1275
|
+
b"max_cache_size_bytes",
|
|
1276
|
+
"min_cacheable_tokens",
|
|
1277
|
+
b"min_cacheable_tokens",
|
|
1278
|
+
"supports_context_caching",
|
|
1279
|
+
b"supports_context_caching",
|
|
1280
|
+
"supports_prompt_caching",
|
|
1281
|
+
b"supports_prompt_caching",
|
|
1282
|
+
] # noqa: Y015
|
|
1283
|
+
def HasField(self, field_name: _HasFieldArgType) -> _builtins.bool: ...
|
|
1284
|
+
_ClearFieldArgType: _TypeAlias = _typing.Literal[
|
|
1285
|
+
"cache_key_strategy",
|
|
1286
|
+
b"cache_key_strategy",
|
|
1287
|
+
"cache_ttl_seconds",
|
|
1288
|
+
b"cache_ttl_seconds",
|
|
1289
|
+
"max_cache_size_bytes",
|
|
1290
|
+
b"max_cache_size_bytes",
|
|
1291
|
+
"min_cacheable_tokens",
|
|
1292
|
+
b"min_cacheable_tokens",
|
|
1293
|
+
"supports_context_caching",
|
|
1294
|
+
b"supports_context_caching",
|
|
1295
|
+
"supports_prompt_caching",
|
|
1296
|
+
b"supports_prompt_caching",
|
|
1297
|
+
] # noqa: Y015
|
|
1298
|
+
def ClearField(self, field_name: _ClearFieldArgType) -> None: ...
|
|
1299
|
+
|
|
1300
|
+
Global___Caching: _TypeAlias = Caching # noqa: Y015
|
|
1301
|
+
|
|
1302
|
+
@_typing.final
|
|
1303
|
+
class Reasoning(_message.Message):
|
|
1235
1304
|
"""Reasoning capability configuration"""
|
|
1236
1305
|
|
|
1237
|
-
DESCRIPTOR:
|
|
1306
|
+
DESCRIPTOR: _descriptor.Descriptor
|
|
1238
1307
|
|
|
1239
|
-
SUPPORTS_CHAIN_OF_THOUGHT_FIELD_NUMBER:
|
|
1240
|
-
SUPPORTS_STEP_TRACKING_FIELD_NUMBER:
|
|
1241
|
-
MAX_REASONING_STEPS_FIELD_NUMBER:
|
|
1242
|
-
SUPPORTS_SELF_CORRECTION_FIELD_NUMBER:
|
|
1243
|
-
REASONING_STRATEGIES_FIELD_NUMBER:
|
|
1244
|
-
supports_chain_of_thought:
|
|
1308
|
+
SUPPORTS_CHAIN_OF_THOUGHT_FIELD_NUMBER: _builtins.int
|
|
1309
|
+
SUPPORTS_STEP_TRACKING_FIELD_NUMBER: _builtins.int
|
|
1310
|
+
MAX_REASONING_STEPS_FIELD_NUMBER: _builtins.int
|
|
1311
|
+
SUPPORTS_SELF_CORRECTION_FIELD_NUMBER: _builtins.int
|
|
1312
|
+
REASONING_STRATEGIES_FIELD_NUMBER: _builtins.int
|
|
1313
|
+
supports_chain_of_thought: _builtins.bool
|
|
1245
1314
|
"""Whether the model can produce hidden chain-of-thought/assistant reasoning
|
|
1246
1315
|
Example: true for models supporting private CoT traces
|
|
1247
1316
|
"""
|
|
1248
|
-
supports_step_tracking:
|
|
1317
|
+
supports_step_tracking: _builtins.bool
|
|
1249
1318
|
"""Whether the model can expose discrete step indices or markers
|
|
1250
1319
|
Example: true for models that number reasoning steps
|
|
1251
1320
|
"""
|
|
1252
|
-
max_reasoning_steps:
|
|
1321
|
+
max_reasoning_steps: _builtins.int
|
|
1253
1322
|
"""Maximum number of internal reasoning steps supported
|
|
1254
1323
|
Example: 64
|
|
1255
1324
|
"""
|
|
1256
|
-
supports_self_correction:
|
|
1325
|
+
supports_self_correction: _builtins.bool
|
|
1257
1326
|
"""Whether the model can revise earlier steps based on later insights
|
|
1258
1327
|
Example: true for self-correction/self-reflection
|
|
1259
1328
|
"""
|
|
1260
|
-
@property
|
|
1329
|
+
@_builtins.property
|
|
1261
1330
|
def reasoning_strategies(
|
|
1262
1331
|
self,
|
|
1263
|
-
) ->
|
|
1264
|
-
global___ReasoningStrategy.ValueType
|
|
1265
|
-
]:
|
|
1332
|
+
) -> _containers.RepeatedScalarFieldContainer[Global___ReasoningStrategy.ValueType]:
|
|
1266
1333
|
"""Supported high-level reasoning strategies
|
|
1267
1334
|
Examples: [REASONING_STRATEGY_STEP_BY_STEP, REASONING_STRATEGY_SELF_CONSISTENCY]
|
|
1268
1335
|
"""
|
|
@@ -1270,82 +1337,89 @@ class Reasoning(google.protobuf.message.Message):
|
|
|
1270
1337
|
def __init__(
|
|
1271
1338
|
self,
|
|
1272
1339
|
*,
|
|
1273
|
-
supports_chain_of_thought:
|
|
1274
|
-
supports_step_tracking:
|
|
1275
|
-
max_reasoning_steps:
|
|
1276
|
-
supports_self_correction:
|
|
1277
|
-
reasoning_strategies:
|
|
1278
|
-
global___ReasoningStrategy.ValueType
|
|
1279
|
-
]
|
|
1340
|
+
supports_chain_of_thought: _builtins.bool | None = ...,
|
|
1341
|
+
supports_step_tracking: _builtins.bool | None = ...,
|
|
1342
|
+
max_reasoning_steps: _builtins.int | None = ...,
|
|
1343
|
+
supports_self_correction: _builtins.bool | None = ...,
|
|
1344
|
+
reasoning_strategies: _abc.Iterable[Global___ReasoningStrategy.ValueType]
|
|
1280
1345
|
| None = ...,
|
|
1281
1346
|
) -> None: ...
|
|
1282
|
-
|
|
1283
|
-
|
|
1284
|
-
|
|
1285
|
-
|
|
1286
|
-
|
|
1287
|
-
|
|
1288
|
-
|
|
1289
|
-
|
|
1290
|
-
|
|
1291
|
-
|
|
1292
|
-
|
|
1293
|
-
|
|
1294
|
-
|
|
1295
|
-
|
|
1296
|
-
|
|
1297
|
-
|
|
1298
|
-
|
|
1299
|
-
|
|
1300
|
-
|
|
1301
|
-
|
|
1347
|
+
_HasFieldArgType: _TypeAlias = _typing.Literal[
|
|
1348
|
+
"max_reasoning_steps",
|
|
1349
|
+
b"max_reasoning_steps",
|
|
1350
|
+
"reasoning_strategies",
|
|
1351
|
+
b"reasoning_strategies",
|
|
1352
|
+
"supports_chain_of_thought",
|
|
1353
|
+
b"supports_chain_of_thought",
|
|
1354
|
+
"supports_self_correction",
|
|
1355
|
+
b"supports_self_correction",
|
|
1356
|
+
"supports_step_tracking",
|
|
1357
|
+
b"supports_step_tracking",
|
|
1358
|
+
] # noqa: Y015
|
|
1359
|
+
def HasField(self, field_name: _HasFieldArgType) -> _builtins.bool: ...
|
|
1360
|
+
_ClearFieldArgType: _TypeAlias = _typing.Literal[
|
|
1361
|
+
"max_reasoning_steps",
|
|
1362
|
+
b"max_reasoning_steps",
|
|
1363
|
+
"reasoning_strategies",
|
|
1364
|
+
b"reasoning_strategies",
|
|
1365
|
+
"supports_chain_of_thought",
|
|
1366
|
+
b"supports_chain_of_thought",
|
|
1367
|
+
"supports_self_correction",
|
|
1368
|
+
b"supports_self_correction",
|
|
1369
|
+
"supports_step_tracking",
|
|
1370
|
+
b"supports_step_tracking",
|
|
1371
|
+
] # noqa: Y015
|
|
1372
|
+
def ClearField(self, field_name: _ClearFieldArgType) -> None: ...
|
|
1373
|
+
|
|
1374
|
+
Global___Reasoning: _TypeAlias = Reasoning # noqa: Y015
|
|
1375
|
+
|
|
1376
|
+
@_typing.final
|
|
1377
|
+
class Audio(_message.Message):
|
|
1302
1378
|
"""Audio capability configuration"""
|
|
1303
1379
|
|
|
1304
|
-
DESCRIPTOR:
|
|
1380
|
+
DESCRIPTOR: _descriptor.Descriptor
|
|
1305
1381
|
|
|
1306
|
-
DIRECTION_FIELD_NUMBER:
|
|
1307
|
-
SUPPORTED_FORMATS_FIELD_NUMBER:
|
|
1308
|
-
MAX_DURATION_SECONDS_FIELD_NUMBER:
|
|
1309
|
-
MAX_FILE_SIZE_BYTES_FIELD_NUMBER:
|
|
1310
|
-
SUPPORTED_LANGUAGES_FIELD_NUMBER:
|
|
1311
|
-
SUPPORTS_STREAMING_FIELD_NUMBER:
|
|
1312
|
-
SUPPORTS_VOICE_SELECTION_FIELD_NUMBER:
|
|
1313
|
-
direction:
|
|
1382
|
+
DIRECTION_FIELD_NUMBER: _builtins.int
|
|
1383
|
+
SUPPORTED_FORMATS_FIELD_NUMBER: _builtins.int
|
|
1384
|
+
MAX_DURATION_SECONDS_FIELD_NUMBER: _builtins.int
|
|
1385
|
+
MAX_FILE_SIZE_BYTES_FIELD_NUMBER: _builtins.int
|
|
1386
|
+
SUPPORTED_LANGUAGES_FIELD_NUMBER: _builtins.int
|
|
1387
|
+
SUPPORTS_STREAMING_FIELD_NUMBER: _builtins.int
|
|
1388
|
+
SUPPORTS_VOICE_SELECTION_FIELD_NUMBER: _builtins.int
|
|
1389
|
+
direction: Global___ModalityDirection.ValueType
|
|
1314
1390
|
"""Direction of audio support
|
|
1315
1391
|
Example: MODALITY_DIRECTION_INPUT_ONLY for speech-to-text only
|
|
1316
1392
|
Example: MODALITY_DIRECTION_OUTPUT_ONLY for text-to-speech only
|
|
1317
1393
|
Example: MODALITY_DIRECTION_INPUT_OUTPUT for models supporting both STT and TTS
|
|
1318
1394
|
"""
|
|
1319
|
-
max_duration_seconds:
|
|
1395
|
+
max_duration_seconds: _builtins.int
|
|
1320
1396
|
"""Maximum audio duration in seconds
|
|
1321
1397
|
Example: 600 for 10-minute limit
|
|
1322
1398
|
"""
|
|
1323
|
-
max_file_size_bytes:
|
|
1399
|
+
max_file_size_bytes: _builtins.int
|
|
1324
1400
|
"""Maximum audio file size in bytes
|
|
1325
1401
|
Example: 26214400 (25MB) limit
|
|
1326
1402
|
"""
|
|
1327
|
-
supports_streaming:
|
|
1403
|
+
supports_streaming: _builtins.bool
|
|
1328
1404
|
"""Supports real-time streaming (for live audio)
|
|
1329
1405
|
Example: true for real-time voice models
|
|
1330
1406
|
"""
|
|
1331
|
-
supports_voice_selection:
|
|
1407
|
+
supports_voice_selection: _builtins.bool
|
|
1332
1408
|
"""Supports voice cloning or voice selection
|
|
1333
1409
|
Example: true if TTS can use different voices
|
|
1334
1410
|
"""
|
|
1335
|
-
@property
|
|
1411
|
+
@_builtins.property
|
|
1336
1412
|
def supported_formats(
|
|
1337
1413
|
self,
|
|
1338
|
-
) ->
|
|
1339
|
-
global___AudioFormat.ValueType
|
|
1340
|
-
]:
|
|
1414
|
+
) -> _containers.RepeatedScalarFieldContainer[Global___AudioFormat.ValueType]:
|
|
1341
1415
|
"""Supported audio file formats
|
|
1342
1416
|
Examples: [AUDIO_FORMAT_MP3, AUDIO_FORMAT_WAV, AUDIO_FORMAT_M4A]
|
|
1343
1417
|
"""
|
|
1344
1418
|
|
|
1345
|
-
@property
|
|
1419
|
+
@_builtins.property
|
|
1346
1420
|
def supported_languages(
|
|
1347
1421
|
self,
|
|
1348
|
-
) ->
|
|
1422
|
+
) -> _containers.RepeatedScalarFieldContainer[_builtins.str]:
|
|
1349
1423
|
"""Supported languages for audio processing
|
|
1350
1424
|
Examples: ["en", "es", "fr", "de", "zh", "ja"]
|
|
1351
1425
|
"""
|
|
@@ -1353,82 +1427,94 @@ class Audio(google.protobuf.message.Message):
|
|
|
1353
1427
|
def __init__(
|
|
1354
1428
|
self,
|
|
1355
1429
|
*,
|
|
1356
|
-
direction:
|
|
1357
|
-
supported_formats:
|
|
1358
|
-
| None = ...,
|
|
1359
|
-
|
|
1360
|
-
|
|
1361
|
-
|
|
1362
|
-
|
|
1363
|
-
supports_voice_selection: builtins.bool = ...,
|
|
1364
|
-
) -> None: ...
|
|
1365
|
-
def ClearField(
|
|
1366
|
-
self,
|
|
1367
|
-
field_name: typing.Literal[
|
|
1368
|
-
"direction",
|
|
1369
|
-
b"direction",
|
|
1370
|
-
"max_duration_seconds",
|
|
1371
|
-
b"max_duration_seconds",
|
|
1372
|
-
"max_file_size_bytes",
|
|
1373
|
-
b"max_file_size_bytes",
|
|
1374
|
-
"supported_formats",
|
|
1375
|
-
b"supported_formats",
|
|
1376
|
-
"supported_languages",
|
|
1377
|
-
b"supported_languages",
|
|
1378
|
-
"supports_streaming",
|
|
1379
|
-
b"supports_streaming",
|
|
1380
|
-
"supports_voice_selection",
|
|
1381
|
-
b"supports_voice_selection",
|
|
1382
|
-
],
|
|
1430
|
+
direction: Global___ModalityDirection.ValueType | None = ...,
|
|
1431
|
+
supported_formats: _abc.Iterable[Global___AudioFormat.ValueType] | None = ...,
|
|
1432
|
+
max_duration_seconds: _builtins.int | None = ...,
|
|
1433
|
+
max_file_size_bytes: _builtins.int | None = ...,
|
|
1434
|
+
supported_languages: _abc.Iterable[_builtins.str] | None = ...,
|
|
1435
|
+
supports_streaming: _builtins.bool | None = ...,
|
|
1436
|
+
supports_voice_selection: _builtins.bool | None = ...,
|
|
1383
1437
|
) -> None: ...
|
|
1384
|
-
|
|
1385
|
-
|
|
1386
|
-
|
|
1387
|
-
|
|
1388
|
-
|
|
1438
|
+
_HasFieldArgType: _TypeAlias = _typing.Literal[
|
|
1439
|
+
"direction",
|
|
1440
|
+
b"direction",
|
|
1441
|
+
"max_duration_seconds",
|
|
1442
|
+
b"max_duration_seconds",
|
|
1443
|
+
"max_file_size_bytes",
|
|
1444
|
+
b"max_file_size_bytes",
|
|
1445
|
+
"supported_formats",
|
|
1446
|
+
b"supported_formats",
|
|
1447
|
+
"supported_languages",
|
|
1448
|
+
b"supported_languages",
|
|
1449
|
+
"supports_streaming",
|
|
1450
|
+
b"supports_streaming",
|
|
1451
|
+
"supports_voice_selection",
|
|
1452
|
+
b"supports_voice_selection",
|
|
1453
|
+
] # noqa: Y015
|
|
1454
|
+
def HasField(self, field_name: _HasFieldArgType) -> _builtins.bool: ...
|
|
1455
|
+
_ClearFieldArgType: _TypeAlias = _typing.Literal[
|
|
1456
|
+
"direction",
|
|
1457
|
+
b"direction",
|
|
1458
|
+
"max_duration_seconds",
|
|
1459
|
+
b"max_duration_seconds",
|
|
1460
|
+
"max_file_size_bytes",
|
|
1461
|
+
b"max_file_size_bytes",
|
|
1462
|
+
"supported_formats",
|
|
1463
|
+
b"supported_formats",
|
|
1464
|
+
"supported_languages",
|
|
1465
|
+
b"supported_languages",
|
|
1466
|
+
"supports_streaming",
|
|
1467
|
+
b"supports_streaming",
|
|
1468
|
+
"supports_voice_selection",
|
|
1469
|
+
b"supports_voice_selection",
|
|
1470
|
+
] # noqa: Y015
|
|
1471
|
+
def ClearField(self, field_name: _ClearFieldArgType) -> None: ...
|
|
1472
|
+
|
|
1473
|
+
Global___Audio: _TypeAlias = Audio # noqa: Y015
|
|
1474
|
+
|
|
1475
|
+
@_typing.final
|
|
1476
|
+
class Video(_message.Message):
|
|
1389
1477
|
"""Video capability configuration"""
|
|
1390
1478
|
|
|
1391
|
-
DESCRIPTOR:
|
|
1479
|
+
DESCRIPTOR: _descriptor.Descriptor
|
|
1392
1480
|
|
|
1393
|
-
DIRECTION_FIELD_NUMBER:
|
|
1394
|
-
SUPPORTED_FORMATS_FIELD_NUMBER:
|
|
1395
|
-
MAX_DURATION_SECONDS_FIELD_NUMBER:
|
|
1396
|
-
MAX_FILE_SIZE_BYTES_FIELD_NUMBER:
|
|
1397
|
-
MAX_FPS_FIELD_NUMBER:
|
|
1398
|
-
SUPPORTS_FRAME_EXTRACTION_FIELD_NUMBER:
|
|
1399
|
-
MAX_FRAMES_FIELD_NUMBER:
|
|
1400
|
-
direction:
|
|
1481
|
+
DIRECTION_FIELD_NUMBER: _builtins.int
|
|
1482
|
+
SUPPORTED_FORMATS_FIELD_NUMBER: _builtins.int
|
|
1483
|
+
MAX_DURATION_SECONDS_FIELD_NUMBER: _builtins.int
|
|
1484
|
+
MAX_FILE_SIZE_BYTES_FIELD_NUMBER: _builtins.int
|
|
1485
|
+
MAX_FPS_FIELD_NUMBER: _builtins.int
|
|
1486
|
+
SUPPORTS_FRAME_EXTRACTION_FIELD_NUMBER: _builtins.int
|
|
1487
|
+
MAX_FRAMES_FIELD_NUMBER: _builtins.int
|
|
1488
|
+
direction: Global___ModalityDirection.ValueType
|
|
1401
1489
|
"""Direction of video support
|
|
1402
1490
|
Example: MODALITY_DIRECTION_INPUT_ONLY for video analysis/understanding
|
|
1403
1491
|
Example: MODALITY_DIRECTION_OUTPUT_ONLY for video generation
|
|
1404
1492
|
Example: MODALITY_DIRECTION_INPUT_OUTPUT for models that can both analyze and generate
|
|
1405
1493
|
"""
|
|
1406
|
-
max_duration_seconds:
|
|
1494
|
+
max_duration_seconds: _builtins.int
|
|
1407
1495
|
"""Maximum video duration in seconds
|
|
1408
1496
|
Example: 120 for 2-minute limit
|
|
1409
1497
|
"""
|
|
1410
|
-
max_file_size_bytes:
|
|
1498
|
+
max_file_size_bytes: _builtins.int
|
|
1411
1499
|
"""Maximum video file size in bytes
|
|
1412
1500
|
Example: 1073741824 (1GB) limit
|
|
1413
1501
|
"""
|
|
1414
|
-
max_fps:
|
|
1502
|
+
max_fps: _builtins.int
|
|
1415
1503
|
"""Maximum frames per second supported
|
|
1416
1504
|
Example: 30 for standard frame rate
|
|
1417
1505
|
"""
|
|
1418
|
-
supports_frame_extraction:
|
|
1506
|
+
supports_frame_extraction: _builtins.bool
|
|
1419
1507
|
"""Supports extracting and analyzing individual frames
|
|
1420
1508
|
Example: true if model can process video as a sequence of images
|
|
1421
1509
|
"""
|
|
1422
|
-
max_frames:
|
|
1510
|
+
max_frames: _builtins.int
|
|
1423
1511
|
"""Maximum number of frames that can be analyzed
|
|
1424
1512
|
Example: 100 for frame-by-frame analysis limit
|
|
1425
1513
|
"""
|
|
1426
|
-
@property
|
|
1514
|
+
@_builtins.property
|
|
1427
1515
|
def supported_formats(
|
|
1428
1516
|
self,
|
|
1429
|
-
) ->
|
|
1430
|
-
global___VideoFormat.ValueType
|
|
1431
|
-
]:
|
|
1517
|
+
) -> _containers.RepeatedScalarFieldContainer[Global___VideoFormat.ValueType]:
|
|
1432
1518
|
"""Supported video file formats
|
|
1433
1519
|
Examples: [VIDEO_FORMAT_MP4, VIDEO_FORMAT_MOV, VIDEO_FORMAT_AVI]
|
|
1434
1520
|
"""
|
|
@@ -1436,70 +1522,82 @@ class Video(google.protobuf.message.Message):
|
|
|
1436
1522
|
def __init__(
|
|
1437
1523
|
self,
|
|
1438
1524
|
*,
|
|
1439
|
-
direction:
|
|
1440
|
-
supported_formats:
|
|
1441
|
-
| None = ...,
|
|
1442
|
-
|
|
1443
|
-
|
|
1444
|
-
|
|
1445
|
-
|
|
1446
|
-
max_frames: builtins.int = ...,
|
|
1447
|
-
) -> None: ...
|
|
1448
|
-
def ClearField(
|
|
1449
|
-
self,
|
|
1450
|
-
field_name: typing.Literal[
|
|
1451
|
-
"direction",
|
|
1452
|
-
b"direction",
|
|
1453
|
-
"max_duration_seconds",
|
|
1454
|
-
b"max_duration_seconds",
|
|
1455
|
-
"max_file_size_bytes",
|
|
1456
|
-
b"max_file_size_bytes",
|
|
1457
|
-
"max_fps",
|
|
1458
|
-
b"max_fps",
|
|
1459
|
-
"max_frames",
|
|
1460
|
-
b"max_frames",
|
|
1461
|
-
"supported_formats",
|
|
1462
|
-
b"supported_formats",
|
|
1463
|
-
"supports_frame_extraction",
|
|
1464
|
-
b"supports_frame_extraction",
|
|
1465
|
-
],
|
|
1525
|
+
direction: Global___ModalityDirection.ValueType | None = ...,
|
|
1526
|
+
supported_formats: _abc.Iterable[Global___VideoFormat.ValueType] | None = ...,
|
|
1527
|
+
max_duration_seconds: _builtins.int | None = ...,
|
|
1528
|
+
max_file_size_bytes: _builtins.int | None = ...,
|
|
1529
|
+
max_fps: _builtins.int | None = ...,
|
|
1530
|
+
supports_frame_extraction: _builtins.bool | None = ...,
|
|
1531
|
+
max_frames: _builtins.int | None = ...,
|
|
1466
1532
|
) -> None: ...
|
|
1467
|
-
|
|
1468
|
-
|
|
1469
|
-
|
|
1470
|
-
|
|
1471
|
-
|
|
1533
|
+
_HasFieldArgType: _TypeAlias = _typing.Literal[
|
|
1534
|
+
"direction",
|
|
1535
|
+
b"direction",
|
|
1536
|
+
"max_duration_seconds",
|
|
1537
|
+
b"max_duration_seconds",
|
|
1538
|
+
"max_file_size_bytes",
|
|
1539
|
+
b"max_file_size_bytes",
|
|
1540
|
+
"max_fps",
|
|
1541
|
+
b"max_fps",
|
|
1542
|
+
"max_frames",
|
|
1543
|
+
b"max_frames",
|
|
1544
|
+
"supported_formats",
|
|
1545
|
+
b"supported_formats",
|
|
1546
|
+
"supports_frame_extraction",
|
|
1547
|
+
b"supports_frame_extraction",
|
|
1548
|
+
] # noqa: Y015
|
|
1549
|
+
def HasField(self, field_name: _HasFieldArgType) -> _builtins.bool: ...
|
|
1550
|
+
_ClearFieldArgType: _TypeAlias = _typing.Literal[
|
|
1551
|
+
"direction",
|
|
1552
|
+
b"direction",
|
|
1553
|
+
"max_duration_seconds",
|
|
1554
|
+
b"max_duration_seconds",
|
|
1555
|
+
"max_file_size_bytes",
|
|
1556
|
+
b"max_file_size_bytes",
|
|
1557
|
+
"max_fps",
|
|
1558
|
+
b"max_fps",
|
|
1559
|
+
"max_frames",
|
|
1560
|
+
b"max_frames",
|
|
1561
|
+
"supported_formats",
|
|
1562
|
+
b"supported_formats",
|
|
1563
|
+
"supports_frame_extraction",
|
|
1564
|
+
b"supports_frame_extraction",
|
|
1565
|
+
] # noqa: Y015
|
|
1566
|
+
def ClearField(self, field_name: _ClearFieldArgType) -> None: ...
|
|
1567
|
+
|
|
1568
|
+
Global___Video: _TypeAlias = Video # noqa: Y015
|
|
1569
|
+
|
|
1570
|
+
@_typing.final
|
|
1571
|
+
class Embeddings(_message.Message):
|
|
1472
1572
|
"""Embeddings capability configuration"""
|
|
1473
1573
|
|
|
1474
|
-
DESCRIPTOR:
|
|
1574
|
+
DESCRIPTOR: _descriptor.Descriptor
|
|
1475
1575
|
|
|
1476
|
-
EMBEDDING_DIMENSIONS_FIELD_NUMBER:
|
|
1477
|
-
MAX_INPUT_TOKENS_FIELD_NUMBER:
|
|
1478
|
-
SUPPORTS_BATCH_FIELD_NUMBER:
|
|
1479
|
-
MAX_BATCH_SIZE_FIELD_NUMBER:
|
|
1480
|
-
DISTANCE_METRICS_FIELD_NUMBER:
|
|
1481
|
-
embedding_dimensions:
|
|
1576
|
+
EMBEDDING_DIMENSIONS_FIELD_NUMBER: _builtins.int
|
|
1577
|
+
MAX_INPUT_TOKENS_FIELD_NUMBER: _builtins.int
|
|
1578
|
+
SUPPORTS_BATCH_FIELD_NUMBER: _builtins.int
|
|
1579
|
+
MAX_BATCH_SIZE_FIELD_NUMBER: _builtins.int
|
|
1580
|
+
DISTANCE_METRICS_FIELD_NUMBER: _builtins.int
|
|
1581
|
+
embedding_dimensions: _builtins.int
|
|
1482
1582
|
"""Dimensionality of the vector space produced by the model
|
|
1483
1583
|
Example: 1536
|
|
1484
1584
|
"""
|
|
1485
|
-
max_input_tokens:
|
|
1585
|
+
max_input_tokens: _builtins.int
|
|
1486
1586
|
"""Maximum number of tokens accepted per input item
|
|
1487
1587
|
Example: 8192
|
|
1488
1588
|
"""
|
|
1489
|
-
supports_batch:
|
|
1589
|
+
supports_batch: _builtins.bool
|
|
1490
1590
|
"""Whether the API supports batching multiple inputs in one call
|
|
1491
1591
|
Example: true
|
|
1492
1592
|
"""
|
|
1493
|
-
max_batch_size:
|
|
1593
|
+
max_batch_size: _builtins.int
|
|
1494
1594
|
"""Maximum number of items allowed per batch
|
|
1495
1595
|
Example: 128
|
|
1496
1596
|
"""
|
|
1497
|
-
@property
|
|
1597
|
+
@_builtins.property
|
|
1498
1598
|
def distance_metrics(
|
|
1499
1599
|
self,
|
|
1500
|
-
) ->
|
|
1501
|
-
global___DistanceMetric.ValueType
|
|
1502
|
-
]:
|
|
1600
|
+
) -> _containers.RepeatedScalarFieldContainer[Global___DistanceMetric.ValueType]:
|
|
1503
1601
|
"""Supported similarity/distance metrics for index/search
|
|
1504
1602
|
Examples: [DISTANCE_METRIC_COSINE, DISTANCE_METRIC_DOT_PRODUCT]
|
|
1505
1603
|
"""
|
|
@@ -1507,75 +1605,81 @@ class Embeddings(google.protobuf.message.Message):
|
|
|
1507
1605
|
def __init__(
|
|
1508
1606
|
self,
|
|
1509
1607
|
*,
|
|
1510
|
-
embedding_dimensions:
|
|
1511
|
-
max_input_tokens:
|
|
1512
|
-
supports_batch:
|
|
1513
|
-
max_batch_size:
|
|
1514
|
-
distance_metrics:
|
|
1515
|
-
| None = ...,
|
|
1516
|
-
) -> None: ...
|
|
1517
|
-
def ClearField(
|
|
1518
|
-
self,
|
|
1519
|
-
field_name: typing.Literal[
|
|
1520
|
-
"distance_metrics",
|
|
1521
|
-
b"distance_metrics",
|
|
1522
|
-
"embedding_dimensions",
|
|
1523
|
-
b"embedding_dimensions",
|
|
1524
|
-
"max_batch_size",
|
|
1525
|
-
b"max_batch_size",
|
|
1526
|
-
"max_input_tokens",
|
|
1527
|
-
b"max_input_tokens",
|
|
1528
|
-
"supports_batch",
|
|
1529
|
-
b"supports_batch",
|
|
1530
|
-
],
|
|
1608
|
+
embedding_dimensions: _builtins.int | None = ...,
|
|
1609
|
+
max_input_tokens: _builtins.int | None = ...,
|
|
1610
|
+
supports_batch: _builtins.bool | None = ...,
|
|
1611
|
+
max_batch_size: _builtins.int | None = ...,
|
|
1612
|
+
distance_metrics: _abc.Iterable[Global___DistanceMetric.ValueType] | None = ...,
|
|
1531
1613
|
) -> None: ...
|
|
1532
|
-
|
|
1533
|
-
|
|
1534
|
-
|
|
1535
|
-
|
|
1536
|
-
|
|
1614
|
+
_HasFieldArgType: _TypeAlias = _typing.Literal[
|
|
1615
|
+
"distance_metrics",
|
|
1616
|
+
b"distance_metrics",
|
|
1617
|
+
"embedding_dimensions",
|
|
1618
|
+
b"embedding_dimensions",
|
|
1619
|
+
"max_batch_size",
|
|
1620
|
+
b"max_batch_size",
|
|
1621
|
+
"max_input_tokens",
|
|
1622
|
+
b"max_input_tokens",
|
|
1623
|
+
"supports_batch",
|
|
1624
|
+
b"supports_batch",
|
|
1625
|
+
] # noqa: Y015
|
|
1626
|
+
def HasField(self, field_name: _HasFieldArgType) -> _builtins.bool: ...
|
|
1627
|
+
_ClearFieldArgType: _TypeAlias = _typing.Literal[
|
|
1628
|
+
"distance_metrics",
|
|
1629
|
+
b"distance_metrics",
|
|
1630
|
+
"embedding_dimensions",
|
|
1631
|
+
b"embedding_dimensions",
|
|
1632
|
+
"max_batch_size",
|
|
1633
|
+
b"max_batch_size",
|
|
1634
|
+
"max_input_tokens",
|
|
1635
|
+
b"max_input_tokens",
|
|
1636
|
+
"supports_batch",
|
|
1637
|
+
b"supports_batch",
|
|
1638
|
+
] # noqa: Y015
|
|
1639
|
+
def ClearField(self, field_name: _ClearFieldArgType) -> None: ...
|
|
1640
|
+
|
|
1641
|
+
Global___Embeddings: _TypeAlias = Embeddings # noqa: Y015
|
|
1642
|
+
|
|
1643
|
+
@_typing.final
|
|
1644
|
+
class FineTuning(_message.Message):
|
|
1537
1645
|
"""Fine-tuning capability configuration"""
|
|
1538
1646
|
|
|
1539
|
-
DESCRIPTOR:
|
|
1647
|
+
DESCRIPTOR: _descriptor.Descriptor
|
|
1540
1648
|
|
|
1541
|
-
MIN_EXAMPLES_FIELD_NUMBER:
|
|
1542
|
-
MAX_EXAMPLES_FIELD_NUMBER:
|
|
1543
|
-
SUPPORTED_FORMATS_FIELD_NUMBER:
|
|
1544
|
-
MAX_FILE_SIZE_MB_FIELD_NUMBER:
|
|
1545
|
-
SUPPORTS_VALIDATION_SET_FIELD_NUMBER:
|
|
1546
|
-
HYPERPARAMETERS_FIELD_NUMBER:
|
|
1547
|
-
min_examples:
|
|
1649
|
+
MIN_EXAMPLES_FIELD_NUMBER: _builtins.int
|
|
1650
|
+
MAX_EXAMPLES_FIELD_NUMBER: _builtins.int
|
|
1651
|
+
SUPPORTED_FORMATS_FIELD_NUMBER: _builtins.int
|
|
1652
|
+
MAX_FILE_SIZE_MB_FIELD_NUMBER: _builtins.int
|
|
1653
|
+
SUPPORTS_VALIDATION_SET_FIELD_NUMBER: _builtins.int
|
|
1654
|
+
HYPERPARAMETERS_FIELD_NUMBER: _builtins.int
|
|
1655
|
+
min_examples: _builtins.int
|
|
1548
1656
|
"""Minimum number of training examples required to start fine-tuning
|
|
1549
1657
|
Example: 50
|
|
1550
1658
|
"""
|
|
1551
|
-
max_examples:
|
|
1659
|
+
max_examples: _builtins.int
|
|
1552
1660
|
"""Maximum number of training examples supported in a single job
|
|
1553
1661
|
Example: 500000
|
|
1554
1662
|
"""
|
|
1555
|
-
max_file_size_mb:
|
|
1663
|
+
max_file_size_mb: _builtins.int
|
|
1556
1664
|
"""Maximum size of individual training files (MB)
|
|
1557
1665
|
Example: 512
|
|
1558
1666
|
"""
|
|
1559
|
-
supports_validation_set:
|
|
1667
|
+
supports_validation_set: _builtins.bool
|
|
1560
1668
|
"""Whether a separate validation set can be provided
|
|
1561
1669
|
Example: true
|
|
1562
1670
|
"""
|
|
1563
|
-
@property
|
|
1671
|
+
@_builtins.property
|
|
1564
1672
|
def supported_formats(
|
|
1565
1673
|
self,
|
|
1566
|
-
) ->
|
|
1567
|
-
global___DataFormat.ValueType
|
|
1568
|
-
]:
|
|
1674
|
+
) -> _containers.RepeatedScalarFieldContainer[Global___DataFormat.ValueType]:
|
|
1569
1675
|
"""Supported dataset/file formats for fine-tuning
|
|
1570
1676
|
Examples: [DATA_FORMAT_JSONL, DATA_FORMAT_CSV]
|
|
1571
1677
|
"""
|
|
1572
1678
|
|
|
1573
|
-
@property
|
|
1679
|
+
@_builtins.property
|
|
1574
1680
|
def hyperparameters(
|
|
1575
1681
|
self,
|
|
1576
|
-
) ->
|
|
1577
|
-
global___Hyperparameter.ValueType
|
|
1578
|
-
]:
|
|
1682
|
+
) -> _containers.RepeatedScalarFieldContainer[Global___Hyperparameter.ValueType]:
|
|
1579
1683
|
"""Hyperparameters that can be configured for training
|
|
1580
1684
|
Examples: [HYPERPARAMETER_LEARNING_RATE, HYPERPARAMETER_EPOCHS]
|
|
1581
1685
|
"""
|
|
@@ -1583,31 +1687,42 @@ class FineTuning(google.protobuf.message.Message):
|
|
|
1583
1687
|
def __init__(
|
|
1584
1688
|
self,
|
|
1585
1689
|
*,
|
|
1586
|
-
min_examples:
|
|
1587
|
-
max_examples:
|
|
1588
|
-
supported_formats:
|
|
1589
|
-
| None = ...,
|
|
1590
|
-
|
|
1591
|
-
|
|
1592
|
-
hyperparameters: collections.abc.Iterable[global___Hyperparameter.ValueType]
|
|
1593
|
-
| None = ...,
|
|
1594
|
-
) -> None: ...
|
|
1595
|
-
def ClearField(
|
|
1596
|
-
self,
|
|
1597
|
-
field_name: typing.Literal[
|
|
1598
|
-
"hyperparameters",
|
|
1599
|
-
b"hyperparameters",
|
|
1600
|
-
"max_examples",
|
|
1601
|
-
b"max_examples",
|
|
1602
|
-
"max_file_size_mb",
|
|
1603
|
-
b"max_file_size_mb",
|
|
1604
|
-
"min_examples",
|
|
1605
|
-
b"min_examples",
|
|
1606
|
-
"supported_formats",
|
|
1607
|
-
b"supported_formats",
|
|
1608
|
-
"supports_validation_set",
|
|
1609
|
-
b"supports_validation_set",
|
|
1610
|
-
],
|
|
1690
|
+
min_examples: _builtins.int | None = ...,
|
|
1691
|
+
max_examples: _builtins.int | None = ...,
|
|
1692
|
+
supported_formats: _abc.Iterable[Global___DataFormat.ValueType] | None = ...,
|
|
1693
|
+
max_file_size_mb: _builtins.int | None = ...,
|
|
1694
|
+
supports_validation_set: _builtins.bool | None = ...,
|
|
1695
|
+
hyperparameters: _abc.Iterable[Global___Hyperparameter.ValueType] | None = ...,
|
|
1611
1696
|
) -> None: ...
|
|
1612
|
-
|
|
1613
|
-
|
|
1697
|
+
_HasFieldArgType: _TypeAlias = _typing.Literal[
|
|
1698
|
+
"hyperparameters",
|
|
1699
|
+
b"hyperparameters",
|
|
1700
|
+
"max_examples",
|
|
1701
|
+
b"max_examples",
|
|
1702
|
+
"max_file_size_mb",
|
|
1703
|
+
b"max_file_size_mb",
|
|
1704
|
+
"min_examples",
|
|
1705
|
+
b"min_examples",
|
|
1706
|
+
"supported_formats",
|
|
1707
|
+
b"supported_formats",
|
|
1708
|
+
"supports_validation_set",
|
|
1709
|
+
b"supports_validation_set",
|
|
1710
|
+
] # noqa: Y015
|
|
1711
|
+
def HasField(self, field_name: _HasFieldArgType) -> _builtins.bool: ...
|
|
1712
|
+
_ClearFieldArgType: _TypeAlias = _typing.Literal[
|
|
1713
|
+
"hyperparameters",
|
|
1714
|
+
b"hyperparameters",
|
|
1715
|
+
"max_examples",
|
|
1716
|
+
b"max_examples",
|
|
1717
|
+
"max_file_size_mb",
|
|
1718
|
+
b"max_file_size_mb",
|
|
1719
|
+
"min_examples",
|
|
1720
|
+
b"min_examples",
|
|
1721
|
+
"supported_formats",
|
|
1722
|
+
b"supported_formats",
|
|
1723
|
+
"supports_validation_set",
|
|
1724
|
+
b"supports_validation_set",
|
|
1725
|
+
] # noqa: Y015
|
|
1726
|
+
def ClearField(self, field_name: _ClearFieldArgType) -> None: ...
|
|
1727
|
+
|
|
1728
|
+
Global___FineTuning: _TypeAlias = FineTuning # noqa: Y015
|