vellum-ai 0.9.6__py3-none-any.whl → 0.9.7__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- vellum/__init__.py +14 -46
- vellum/core/client_wrapper.py +1 -1
- vellum/resources/ad_hoc/client.py +75 -108
- vellum/types/__init__.py +14 -46
- vellum/types/{ad_hoc_expand_meta_request.py → ad_hoc_expand_meta.py} +1 -1
- vellum/types/{prompt_parameters_request.py → prompt_parameters.py} +1 -1
- vellum/types/{prompt_request_chat_history_input_request.py → prompt_request_chat_history_input.py} +3 -3
- vellum/types/prompt_request_input.py +8 -0
- vellum/types/{prompt_request_json_input_request.py → prompt_request_json_input.py} +1 -1
- vellum/types/{prompt_request_string_input_request.py → prompt_request_string_input.py} +1 -1
- vellum/types/{prompt_settings_request.py → prompt_settings.py} +1 -1
- {vellum_ai-0.9.6.dist-info → vellum_ai-0.9.7.dist-info}/METADATA +1 -1
- {vellum_ai-0.9.6.dist-info → vellum_ai-0.9.7.dist-info}/RECORD +15 -31
- vellum/types/chat_message_prompt_block_properties_request.py +0 -38
- vellum/types/chat_message_prompt_block_request.py +0 -38
- vellum/types/ephemeral_prompt_cache_config_request.py +0 -20
- vellum/types/ephemeral_prompt_cache_config_type_enum.py +0 -5
- vellum/types/function_definition_prompt_block_properties_request.py +0 -42
- vellum/types/function_definition_prompt_block_request.py +0 -29
- vellum/types/jinja_prompt_block_properties_request.py +0 -21
- vellum/types/jinja_prompt_block_request.py +0 -29
- vellum/types/plain_text_prompt_block_request.py +0 -28
- vellum/types/prompt_block_request.py +0 -19
- vellum/types/prompt_block_state.py +0 -5
- vellum/types/prompt_request_input_request.py +0 -10
- vellum/types/rich_text_child_block_request.py +0 -7
- vellum/types/rich_text_prompt_block_request.py +0 -30
- vellum/types/variable_prompt_block_request.py +0 -28
- vellum/types/vellum_variable_extensions_request.py +0 -23
- vellum/types/vellum_variable_request.py +0 -33
- {vellum_ai-0.9.6.dist-info → vellum_ai-0.9.7.dist-info}/LICENSE +0 -0
- {vellum_ai-0.9.6.dist-info → vellum_ai-0.9.7.dist-info}/WHEEL +0 -0
vellum/__init__.py
CHANGED
@@ -2,7 +2,7 @@
|
|
2
2
|
|
3
3
|
from .types import (
|
4
4
|
AdHocExecutePromptEvent,
|
5
|
-
|
5
|
+
AdHocExpandMeta,
|
6
6
|
AdHocFulfilledPromptExecutionMeta,
|
7
7
|
AdHocInitiatedPromptExecutionMeta,
|
8
8
|
AdHocRejectedPromptExecutionMeta,
|
@@ -37,8 +37,6 @@ from .types import (
|
|
37
37
|
ChatMessage,
|
38
38
|
ChatMessageContent,
|
39
39
|
ChatMessageContentRequest,
|
40
|
-
ChatMessagePromptBlockPropertiesRequest,
|
41
|
-
ChatMessagePromptBlockRequest,
|
42
40
|
ChatMessageRequest,
|
43
41
|
ChatMessageRole,
|
44
42
|
CodeExecutionNodeArrayResult,
|
@@ -84,8 +82,6 @@ from .types import (
|
|
84
82
|
EntityStatus,
|
85
83
|
EntityVisibility,
|
86
84
|
EnvironmentEnum,
|
87
|
-
EphemeralPromptCacheConfigRequest,
|
88
|
-
EphemeralPromptCacheConfigTypeEnum,
|
89
85
|
ErrorInputRequest,
|
90
86
|
ErrorVariableValue,
|
91
87
|
ErrorVellumValue,
|
@@ -134,8 +130,6 @@ from .types import (
|
|
134
130
|
FunctionCallVariableValue,
|
135
131
|
FunctionCallVellumValue,
|
136
132
|
FunctionCallVellumValueRequest,
|
137
|
-
FunctionDefinitionPromptBlockPropertiesRequest,
|
138
|
-
FunctionDefinitionPromptBlockRequest,
|
139
133
|
GenerateOptionsRequest,
|
140
134
|
GenerateRequest,
|
141
135
|
GenerateResponse,
|
@@ -168,8 +162,6 @@ from .types import (
|
|
168
162
|
InstructorVectorizerConfig,
|
169
163
|
InstructorVectorizerConfigRequest,
|
170
164
|
IterationStateEnum,
|
171
|
-
JinjaPromptBlockPropertiesRequest,
|
172
|
-
JinjaPromptBlockRequest,
|
173
165
|
JsonInputRequest,
|
174
166
|
JsonVariableValue,
|
175
167
|
JsonVellumValue,
|
@@ -254,12 +246,9 @@ from .types import (
|
|
254
246
|
PaginatedWorkflowReleaseTagReadList,
|
255
247
|
PdfSearchResultMetaSource,
|
256
248
|
PdfSearchResultMetaSourceRequest,
|
257
|
-
PlainTextPromptBlockRequest,
|
258
249
|
Price,
|
259
250
|
ProcessingFailureReasonEnum,
|
260
251
|
ProcessingStateEnum,
|
261
|
-
PromptBlockRequest,
|
262
|
-
PromptBlockState,
|
263
252
|
PromptDeploymentExpandMetaRequest,
|
264
253
|
PromptDeploymentInputRequest,
|
265
254
|
PromptExecutionMeta,
|
@@ -267,12 +256,12 @@ from .types import (
|
|
267
256
|
PromptNodeResult,
|
268
257
|
PromptNodeResultData,
|
269
258
|
PromptOutput,
|
270
|
-
|
271
|
-
|
272
|
-
|
273
|
-
|
274
|
-
|
275
|
-
|
259
|
+
PromptParameters,
|
260
|
+
PromptRequestChatHistoryInput,
|
261
|
+
PromptRequestInput,
|
262
|
+
PromptRequestJsonInput,
|
263
|
+
PromptRequestStringInput,
|
264
|
+
PromptSettings,
|
276
265
|
RawPromptExecutionOverridesRequest,
|
277
266
|
ReductoChunkerConfig,
|
278
267
|
ReductoChunkerConfigRequest,
|
@@ -286,8 +275,6 @@ from .types import (
|
|
286
275
|
RejectedWorkflowNodeResultEvent,
|
287
276
|
ReleaseTagSource,
|
288
277
|
ReplaceTestSuiteTestCaseRequest,
|
289
|
-
RichTextChildBlockRequest,
|
290
|
-
RichTextPromptBlockRequest,
|
291
278
|
SandboxScenario,
|
292
279
|
ScenarioInput,
|
293
280
|
ScenarioInputChatHistoryVariableValue,
|
@@ -419,7 +406,6 @@ from .types import (
|
|
419
406
|
UnitEnum,
|
420
407
|
UploadDocumentResponse,
|
421
408
|
UpsertTestSuiteTestCaseRequest,
|
422
|
-
VariablePromptBlockRequest,
|
423
409
|
VellumAudio,
|
424
410
|
VellumAudioRequest,
|
425
411
|
VellumError,
|
@@ -434,8 +420,6 @@ from .types import (
|
|
434
420
|
VellumValueRequest,
|
435
421
|
VellumVariable,
|
436
422
|
VellumVariableExtensions,
|
437
|
-
VellumVariableExtensionsRequest,
|
438
|
-
VellumVariableRequest,
|
439
423
|
VellumVariableType,
|
440
424
|
WorkflowDeploymentRead,
|
441
425
|
WorkflowEventError,
|
@@ -513,7 +497,7 @@ from .version import __version__
|
|
513
497
|
|
514
498
|
__all__ = [
|
515
499
|
"AdHocExecutePromptEvent",
|
516
|
-
"
|
500
|
+
"AdHocExpandMeta",
|
517
501
|
"AdHocFulfilledPromptExecutionMeta",
|
518
502
|
"AdHocInitiatedPromptExecutionMeta",
|
519
503
|
"AdHocRejectedPromptExecutionMeta",
|
@@ -550,8 +534,6 @@ __all__ = [
|
|
550
534
|
"ChatMessage",
|
551
535
|
"ChatMessageContent",
|
552
536
|
"ChatMessageContentRequest",
|
553
|
-
"ChatMessagePromptBlockPropertiesRequest",
|
554
|
-
"ChatMessagePromptBlockRequest",
|
555
537
|
"ChatMessageRequest",
|
556
538
|
"ChatMessageRole",
|
557
539
|
"CodeExecutionNodeArrayResult",
|
@@ -599,8 +581,6 @@ __all__ = [
|
|
599
581
|
"EntityStatus",
|
600
582
|
"EntityVisibility",
|
601
583
|
"EnvironmentEnum",
|
602
|
-
"EphemeralPromptCacheConfigRequest",
|
603
|
-
"EphemeralPromptCacheConfigTypeEnum",
|
604
584
|
"ErrorInputRequest",
|
605
585
|
"ErrorVariableValue",
|
606
586
|
"ErrorVellumValue",
|
@@ -651,8 +631,6 @@ __all__ = [
|
|
651
631
|
"FunctionCallVariableValue",
|
652
632
|
"FunctionCallVellumValue",
|
653
633
|
"FunctionCallVellumValueRequest",
|
654
|
-
"FunctionDefinitionPromptBlockPropertiesRequest",
|
655
|
-
"FunctionDefinitionPromptBlockRequest",
|
656
634
|
"GenerateOptionsRequest",
|
657
635
|
"GenerateRequest",
|
658
636
|
"GenerateResponse",
|
@@ -686,8 +664,6 @@ __all__ = [
|
|
686
664
|
"InstructorVectorizerConfigRequest",
|
687
665
|
"InternalServerError",
|
688
666
|
"IterationStateEnum",
|
689
|
-
"JinjaPromptBlockPropertiesRequest",
|
690
|
-
"JinjaPromptBlockRequest",
|
691
667
|
"JsonInputRequest",
|
692
668
|
"JsonVariableValue",
|
693
669
|
"JsonVellumValue",
|
@@ -775,12 +751,9 @@ __all__ = [
|
|
775
751
|
"PaginatedWorkflowReleaseTagReadList",
|
776
752
|
"PdfSearchResultMetaSource",
|
777
753
|
"PdfSearchResultMetaSourceRequest",
|
778
|
-
"PlainTextPromptBlockRequest",
|
779
754
|
"Price",
|
780
755
|
"ProcessingFailureReasonEnum",
|
781
756
|
"ProcessingStateEnum",
|
782
|
-
"PromptBlockRequest",
|
783
|
-
"PromptBlockState",
|
784
757
|
"PromptDeploymentExpandMetaRequest",
|
785
758
|
"PromptDeploymentInputRequest",
|
786
759
|
"PromptExecutionMeta",
|
@@ -788,12 +761,12 @@ __all__ = [
|
|
788
761
|
"PromptNodeResult",
|
789
762
|
"PromptNodeResultData",
|
790
763
|
"PromptOutput",
|
791
|
-
"
|
792
|
-
"
|
793
|
-
"
|
794
|
-
"
|
795
|
-
"
|
796
|
-
"
|
764
|
+
"PromptParameters",
|
765
|
+
"PromptRequestChatHistoryInput",
|
766
|
+
"PromptRequestInput",
|
767
|
+
"PromptRequestJsonInput",
|
768
|
+
"PromptRequestStringInput",
|
769
|
+
"PromptSettings",
|
797
770
|
"RawPromptExecutionOverridesRequest",
|
798
771
|
"ReductoChunkerConfig",
|
799
772
|
"ReductoChunkerConfigRequest",
|
@@ -807,8 +780,6 @@ __all__ = [
|
|
807
780
|
"RejectedWorkflowNodeResultEvent",
|
808
781
|
"ReleaseTagSource",
|
809
782
|
"ReplaceTestSuiteTestCaseRequest",
|
810
|
-
"RichTextChildBlockRequest",
|
811
|
-
"RichTextPromptBlockRequest",
|
812
783
|
"SandboxScenario",
|
813
784
|
"ScenarioInput",
|
814
785
|
"ScenarioInputChatHistoryVariableValue",
|
@@ -940,7 +911,6 @@ __all__ = [
|
|
940
911
|
"UnitEnum",
|
941
912
|
"UploadDocumentResponse",
|
942
913
|
"UpsertTestSuiteTestCaseRequest",
|
943
|
-
"VariablePromptBlockRequest",
|
944
914
|
"Vellum",
|
945
915
|
"VellumAudio",
|
946
916
|
"VellumAudioRequest",
|
@@ -957,8 +927,6 @@ __all__ = [
|
|
957
927
|
"VellumValueRequest",
|
958
928
|
"VellumVariable",
|
959
929
|
"VellumVariableExtensions",
|
960
|
-
"VellumVariableExtensionsRequest",
|
961
|
-
"VellumVariableRequest",
|
962
930
|
"VellumVariableType",
|
963
931
|
"WorkflowDeploymentRead",
|
964
932
|
"WorkflowDeploymentsListRequestStatus",
|
vellum/core/client_wrapper.py
CHANGED
@@ -2,12 +2,11 @@
|
|
2
2
|
|
3
3
|
import typing
|
4
4
|
from ...core.client_wrapper import SyncClientWrapper
|
5
|
-
from ...types.
|
6
|
-
from ...types.
|
7
|
-
from ...types.
|
8
|
-
from ...types.
|
9
|
-
from ...types.
|
10
|
-
from ...types.ad_hoc_expand_meta_request import AdHocExpandMetaRequest
|
5
|
+
from ...types.prompt_request_input import PromptRequestInput
|
6
|
+
from ...types.vellum_variable import VellumVariable
|
7
|
+
from ...types.prompt_parameters import PromptParameters
|
8
|
+
from ...types.prompt_settings import PromptSettings
|
9
|
+
from ...types.ad_hoc_expand_meta import AdHocExpandMeta
|
11
10
|
from ...core.request_options import RequestOptions
|
12
11
|
from ...types.ad_hoc_execute_prompt_event import AdHocExecutePromptEvent
|
13
12
|
from ...core.serialization import convert_and_respect_annotation_metadata
|
@@ -32,12 +31,12 @@ class AdHocClient:
|
|
32
31
|
self,
|
33
32
|
*,
|
34
33
|
ml_model: str,
|
35
|
-
input_values: typing.Sequence[
|
36
|
-
input_variables: typing.Sequence[
|
37
|
-
parameters:
|
38
|
-
blocks: typing.Sequence[
|
39
|
-
settings: typing.Optional[
|
40
|
-
expand_meta: typing.Optional[
|
34
|
+
input_values: typing.Sequence[PromptRequestInput],
|
35
|
+
input_variables: typing.Sequence[VellumVariable],
|
36
|
+
parameters: PromptParameters,
|
37
|
+
blocks: typing.Sequence[typing.Optional[typing.Any]],
|
38
|
+
settings: typing.Optional[PromptSettings] = OMIT,
|
39
|
+
expand_meta: typing.Optional[AdHocExpandMeta] = OMIT,
|
41
40
|
request_options: typing.Optional[RequestOptions] = None,
|
42
41
|
) -> typing.Iterator[AdHocExecutePromptEvent]:
|
43
42
|
"""
|
@@ -47,17 +46,17 @@ class AdHocClient:
|
|
47
46
|
----------
|
48
47
|
ml_model : str
|
49
48
|
|
50
|
-
input_values : typing.Sequence[
|
49
|
+
input_values : typing.Sequence[PromptRequestInput]
|
51
50
|
|
52
|
-
input_variables : typing.Sequence[
|
51
|
+
input_variables : typing.Sequence[VellumVariable]
|
53
52
|
|
54
|
-
parameters :
|
53
|
+
parameters : PromptParameters
|
55
54
|
|
56
|
-
blocks : typing.Sequence[
|
55
|
+
blocks : typing.Sequence[typing.Optional[typing.Any]]
|
57
56
|
|
58
|
-
settings : typing.Optional[
|
57
|
+
settings : typing.Optional[PromptSettings]
|
59
58
|
|
60
|
-
expand_meta : typing.Optional[
|
59
|
+
expand_meta : typing.Optional[AdHocExpandMeta]
|
61
60
|
|
62
61
|
request_options : typing.Optional[RequestOptions]
|
63
62
|
Request-specific configuration.
|
@@ -70,17 +69,14 @@ class AdHocClient:
|
|
70
69
|
Examples
|
71
70
|
--------
|
72
71
|
from vellum import (
|
73
|
-
|
74
|
-
|
75
|
-
|
76
|
-
|
77
|
-
|
78
|
-
PromptRequestStringInputRequest,
|
79
|
-
PromptSettingsRequest,
|
80
|
-
StringVellumValueRequest,
|
72
|
+
AdHocExpandMeta,
|
73
|
+
PromptParameters,
|
74
|
+
PromptRequestStringInput,
|
75
|
+
PromptSettings,
|
76
|
+
StringVellumValue,
|
81
77
|
Vellum,
|
82
|
-
|
83
|
-
|
78
|
+
VellumVariable,
|
79
|
+
VellumVariableExtensions,
|
84
80
|
)
|
85
81
|
|
86
82
|
client = Vellum(
|
@@ -89,26 +85,26 @@ class AdHocClient:
|
|
89
85
|
response = client.ad_hoc.adhoc_execute_prompt_stream(
|
90
86
|
ml_model="string",
|
91
87
|
input_values=[
|
92
|
-
|
88
|
+
PromptRequestStringInput(
|
93
89
|
key="string",
|
94
90
|
value="string",
|
95
91
|
)
|
96
92
|
],
|
97
93
|
input_variables=[
|
98
|
-
|
94
|
+
VellumVariable(
|
99
95
|
id="string",
|
100
96
|
key="string",
|
101
97
|
type="STRING",
|
102
98
|
required=True,
|
103
|
-
default=
|
104
|
-
value="
|
99
|
+
default=StringVellumValue(
|
100
|
+
value={"key": "value"},
|
105
101
|
),
|
106
|
-
extensions=
|
107
|
-
color="
|
102
|
+
extensions=VellumVariableExtensions(
|
103
|
+
color={"key": "value"},
|
108
104
|
),
|
109
105
|
)
|
110
106
|
],
|
111
|
-
parameters=
|
107
|
+
parameters=PromptParameters(
|
112
108
|
stop=["string"],
|
113
109
|
temperature=1.1,
|
114
110
|
max_tokens=1,
|
@@ -119,22 +115,11 @@ class AdHocClient:
|
|
119
115
|
logit_bias={"string": {"key": "value"}},
|
120
116
|
custom_parameters={"string": {"key": "value"}},
|
121
117
|
),
|
122
|
-
settings=
|
118
|
+
settings=PromptSettings(
|
123
119
|
timeout=1.1,
|
124
120
|
),
|
125
|
-
blocks=[
|
126
|
-
|
127
|
-
state="ENABLED",
|
128
|
-
cache_config=EphemeralPromptCacheConfigRequest(
|
129
|
-
type={"key": "value"},
|
130
|
-
),
|
131
|
-
properties=JinjaPromptBlockPropertiesRequest(
|
132
|
-
template="string",
|
133
|
-
template_type="STRING",
|
134
|
-
),
|
135
|
-
)
|
136
|
-
],
|
137
|
-
expand_meta=AdHocExpandMetaRequest(
|
121
|
+
blocks=[{"key": "value"}],
|
122
|
+
expand_meta=AdHocExpandMeta(
|
138
123
|
cost=True,
|
139
124
|
model_name=True,
|
140
125
|
usage=True,
|
@@ -151,22 +136,20 @@ class AdHocClient:
|
|
151
136
|
json={
|
152
137
|
"ml_model": ml_model,
|
153
138
|
"input_values": convert_and_respect_annotation_metadata(
|
154
|
-
object_=input_values, annotation=typing.Sequence[
|
139
|
+
object_=input_values, annotation=typing.Sequence[PromptRequestInput], direction="write"
|
155
140
|
),
|
156
141
|
"input_variables": convert_and_respect_annotation_metadata(
|
157
|
-
object_=input_variables, annotation=typing.Sequence[
|
142
|
+
object_=input_variables, annotation=typing.Sequence[VellumVariable], direction="write"
|
158
143
|
),
|
159
144
|
"parameters": convert_and_respect_annotation_metadata(
|
160
|
-
object_=parameters, annotation=
|
145
|
+
object_=parameters, annotation=PromptParameters, direction="write"
|
161
146
|
),
|
162
147
|
"settings": convert_and_respect_annotation_metadata(
|
163
|
-
object_=settings, annotation=
|
164
|
-
),
|
165
|
-
"blocks": convert_and_respect_annotation_metadata(
|
166
|
-
object_=blocks, annotation=typing.Sequence[PromptBlockRequest], direction="write"
|
148
|
+
object_=settings, annotation=PromptSettings, direction="write"
|
167
149
|
),
|
150
|
+
"blocks": blocks,
|
168
151
|
"expand_meta": convert_and_respect_annotation_metadata(
|
169
|
-
object_=expand_meta, annotation=
|
152
|
+
object_=expand_meta, annotation=AdHocExpandMeta, direction="write"
|
170
153
|
),
|
171
154
|
},
|
172
155
|
request_options=request_options,
|
@@ -233,12 +216,12 @@ class AsyncAdHocClient:
|
|
233
216
|
self,
|
234
217
|
*,
|
235
218
|
ml_model: str,
|
236
|
-
input_values: typing.Sequence[
|
237
|
-
input_variables: typing.Sequence[
|
238
|
-
parameters:
|
239
|
-
blocks: typing.Sequence[
|
240
|
-
settings: typing.Optional[
|
241
|
-
expand_meta: typing.Optional[
|
219
|
+
input_values: typing.Sequence[PromptRequestInput],
|
220
|
+
input_variables: typing.Sequence[VellumVariable],
|
221
|
+
parameters: PromptParameters,
|
222
|
+
blocks: typing.Sequence[typing.Optional[typing.Any]],
|
223
|
+
settings: typing.Optional[PromptSettings] = OMIT,
|
224
|
+
expand_meta: typing.Optional[AdHocExpandMeta] = OMIT,
|
242
225
|
request_options: typing.Optional[RequestOptions] = None,
|
243
226
|
) -> typing.AsyncIterator[AdHocExecutePromptEvent]:
|
244
227
|
"""
|
@@ -248,17 +231,17 @@ class AsyncAdHocClient:
|
|
248
231
|
----------
|
249
232
|
ml_model : str
|
250
233
|
|
251
|
-
input_values : typing.Sequence[
|
234
|
+
input_values : typing.Sequence[PromptRequestInput]
|
252
235
|
|
253
|
-
input_variables : typing.Sequence[
|
236
|
+
input_variables : typing.Sequence[VellumVariable]
|
254
237
|
|
255
|
-
parameters :
|
238
|
+
parameters : PromptParameters
|
256
239
|
|
257
|
-
blocks : typing.Sequence[
|
240
|
+
blocks : typing.Sequence[typing.Optional[typing.Any]]
|
258
241
|
|
259
|
-
settings : typing.Optional[
|
242
|
+
settings : typing.Optional[PromptSettings]
|
260
243
|
|
261
|
-
expand_meta : typing.Optional[
|
244
|
+
expand_meta : typing.Optional[AdHocExpandMeta]
|
262
245
|
|
263
246
|
request_options : typing.Optional[RequestOptions]
|
264
247
|
Request-specific configuration.
|
@@ -273,17 +256,14 @@ class AsyncAdHocClient:
|
|
273
256
|
import asyncio
|
274
257
|
|
275
258
|
from vellum import (
|
276
|
-
|
259
|
+
AdHocExpandMeta,
|
277
260
|
AsyncVellum,
|
278
|
-
|
279
|
-
|
280
|
-
|
281
|
-
|
282
|
-
|
283
|
-
|
284
|
-
StringVellumValueRequest,
|
285
|
-
VellumVariableExtensionsRequest,
|
286
|
-
VellumVariableRequest,
|
261
|
+
PromptParameters,
|
262
|
+
PromptRequestStringInput,
|
263
|
+
PromptSettings,
|
264
|
+
StringVellumValue,
|
265
|
+
VellumVariable,
|
266
|
+
VellumVariableExtensions,
|
287
267
|
)
|
288
268
|
|
289
269
|
client = AsyncVellum(
|
@@ -295,26 +275,26 @@ class AsyncAdHocClient:
|
|
295
275
|
response = await client.ad_hoc.adhoc_execute_prompt_stream(
|
296
276
|
ml_model="string",
|
297
277
|
input_values=[
|
298
|
-
|
278
|
+
PromptRequestStringInput(
|
299
279
|
key="string",
|
300
280
|
value="string",
|
301
281
|
)
|
302
282
|
],
|
303
283
|
input_variables=[
|
304
|
-
|
284
|
+
VellumVariable(
|
305
285
|
id="string",
|
306
286
|
key="string",
|
307
287
|
type="STRING",
|
308
288
|
required=True,
|
309
|
-
default=
|
310
|
-
value="
|
289
|
+
default=StringVellumValue(
|
290
|
+
value={"key": "value"},
|
311
291
|
),
|
312
|
-
extensions=
|
313
|
-
color="
|
292
|
+
extensions=VellumVariableExtensions(
|
293
|
+
color={"key": "value"},
|
314
294
|
),
|
315
295
|
)
|
316
296
|
],
|
317
|
-
parameters=
|
297
|
+
parameters=PromptParameters(
|
318
298
|
stop=["string"],
|
319
299
|
temperature=1.1,
|
320
300
|
max_tokens=1,
|
@@ -325,22 +305,11 @@ class AsyncAdHocClient:
|
|
325
305
|
logit_bias={"string": {"key": "value"}},
|
326
306
|
custom_parameters={"string": {"key": "value"}},
|
327
307
|
),
|
328
|
-
settings=
|
308
|
+
settings=PromptSettings(
|
329
309
|
timeout=1.1,
|
330
310
|
),
|
331
|
-
blocks=[
|
332
|
-
|
333
|
-
state="ENABLED",
|
334
|
-
cache_config=EphemeralPromptCacheConfigRequest(
|
335
|
-
type={"key": "value"},
|
336
|
-
),
|
337
|
-
properties=JinjaPromptBlockPropertiesRequest(
|
338
|
-
template="string",
|
339
|
-
template_type="STRING",
|
340
|
-
),
|
341
|
-
)
|
342
|
-
],
|
343
|
-
expand_meta=AdHocExpandMetaRequest(
|
311
|
+
blocks=[{"key": "value"}],
|
312
|
+
expand_meta=AdHocExpandMeta(
|
344
313
|
cost=True,
|
345
314
|
model_name=True,
|
346
315
|
usage=True,
|
@@ -360,22 +329,20 @@ class AsyncAdHocClient:
|
|
360
329
|
json={
|
361
330
|
"ml_model": ml_model,
|
362
331
|
"input_values": convert_and_respect_annotation_metadata(
|
363
|
-
object_=input_values, annotation=typing.Sequence[
|
332
|
+
object_=input_values, annotation=typing.Sequence[PromptRequestInput], direction="write"
|
364
333
|
),
|
365
334
|
"input_variables": convert_and_respect_annotation_metadata(
|
366
|
-
object_=input_variables, annotation=typing.Sequence[
|
335
|
+
object_=input_variables, annotation=typing.Sequence[VellumVariable], direction="write"
|
367
336
|
),
|
368
337
|
"parameters": convert_and_respect_annotation_metadata(
|
369
|
-
object_=parameters, annotation=
|
338
|
+
object_=parameters, annotation=PromptParameters, direction="write"
|
370
339
|
),
|
371
340
|
"settings": convert_and_respect_annotation_metadata(
|
372
|
-
object_=settings, annotation=
|
373
|
-
),
|
374
|
-
"blocks": convert_and_respect_annotation_metadata(
|
375
|
-
object_=blocks, annotation=typing.Sequence[PromptBlockRequest], direction="write"
|
341
|
+
object_=settings, annotation=PromptSettings, direction="write"
|
376
342
|
),
|
343
|
+
"blocks": blocks,
|
377
344
|
"expand_meta": convert_and_respect_annotation_metadata(
|
378
|
-
object_=expand_meta, annotation=
|
345
|
+
object_=expand_meta, annotation=AdHocExpandMeta, direction="write"
|
379
346
|
),
|
380
347
|
},
|
381
348
|
request_options=request_options,
|