vellum-ai 0.9.5__py3-none-any.whl → 0.9.9__py3-none-any.whl
Sign up to get free protection for your applications and to get access to all the features.
- vellum/__init__.py +32 -42
- vellum/core/client_wrapper.py +1 -1
- vellum/resources/ad_hoc/client.py +86 -98
- vellum/types/__init__.py +32 -42
- vellum/types/{ad_hoc_expand_meta_request.py → ad_hoc_expand_meta.py} +1 -1
- vellum/types/{chat_message_prompt_block_request.py → chat_message_prompt_block.py} +11 -11
- vellum/types/{ephemeral_prompt_cache_config_request.py → ephemeral_prompt_cache_config.py} +1 -1
- vellum/types/{function_definition_prompt_block_request.py → function_definition_prompt_block.py} +7 -7
- vellum/types/{jinja_prompt_block_request.py → jinja_prompt_block.py} +7 -7
- vellum/types/{plain_text_prompt_block_request.py → plain_text_prompt_block.py} +4 -5
- vellum/types/prompt_block.py +15 -0
- vellum/types/{prompt_parameters_request.py → prompt_parameters.py} +1 -1
- vellum/types/{prompt_request_chat_history_input_request.py → prompt_request_chat_history_input.py} +3 -3
- vellum/types/prompt_request_input.py +8 -0
- vellum/types/{prompt_request_json_input_request.py → prompt_request_json_input.py} +1 -1
- vellum/types/{prompt_request_string_input_request.py → prompt_request_string_input.py} +1 -1
- vellum/types/{prompt_settings_request.py → prompt_settings.py} +1 -1
- vellum/types/rich_text_child_block.py +7 -0
- vellum/types/{rich_text_prompt_block_request.py → rich_text_prompt_block.py} +6 -7
- vellum/types/{variable_prompt_block_request.py → variable_prompt_block.py} +4 -5
- {vellum_ai-0.9.5.dist-info → vellum_ai-0.9.9.dist-info}/METADATA +1 -1
- {vellum_ai-0.9.5.dist-info → vellum_ai-0.9.9.dist-info}/RECORD +24 -29
- vellum/types/chat_message_prompt_block_properties_request.py +0 -38
- vellum/types/function_definition_prompt_block_properties_request.py +0 -42
- vellum/types/jinja_prompt_block_properties_request.py +0 -21
- vellum/types/prompt_block_request.py +0 -19
- vellum/types/prompt_request_input_request.py +0 -10
- vellum/types/rich_text_child_block_request.py +0 -7
- vellum/types/vellum_variable_extensions_request.py +0 -23
- vellum/types/vellum_variable_request.py +0 -33
- {vellum_ai-0.9.5.dist-info → vellum_ai-0.9.9.dist-info}/LICENSE +0 -0
- {vellum_ai-0.9.5.dist-info → vellum_ai-0.9.9.dist-info}/WHEEL +0 -0
vellum/__init__.py
CHANGED
@@ -2,7 +2,7 @@
|
|
2
2
|
|
3
3
|
from .types import (
|
4
4
|
AdHocExecutePromptEvent,
|
5
|
-
|
5
|
+
AdHocExpandMeta,
|
6
6
|
AdHocFulfilledPromptExecutionMeta,
|
7
7
|
AdHocInitiatedPromptExecutionMeta,
|
8
8
|
AdHocRejectedPromptExecutionMeta,
|
@@ -37,8 +37,7 @@ from .types import (
|
|
37
37
|
ChatMessage,
|
38
38
|
ChatMessageContent,
|
39
39
|
ChatMessageContentRequest,
|
40
|
-
|
41
|
-
ChatMessagePromptBlockRequest,
|
40
|
+
ChatMessagePromptBlock,
|
42
41
|
ChatMessageRequest,
|
43
42
|
ChatMessageRole,
|
44
43
|
CodeExecutionNodeArrayResult,
|
@@ -84,7 +83,7 @@ from .types import (
|
|
84
83
|
EntityStatus,
|
85
84
|
EntityVisibility,
|
86
85
|
EnvironmentEnum,
|
87
|
-
|
86
|
+
EphemeralPromptCacheConfig,
|
88
87
|
EphemeralPromptCacheConfigTypeEnum,
|
89
88
|
ErrorInputRequest,
|
90
89
|
ErrorVariableValue,
|
@@ -134,8 +133,7 @@ from .types import (
|
|
134
133
|
FunctionCallVariableValue,
|
135
134
|
FunctionCallVellumValue,
|
136
135
|
FunctionCallVellumValueRequest,
|
137
|
-
|
138
|
-
FunctionDefinitionPromptBlockRequest,
|
136
|
+
FunctionDefinitionPromptBlock,
|
139
137
|
GenerateOptionsRequest,
|
140
138
|
GenerateRequest,
|
141
139
|
GenerateResponse,
|
@@ -168,8 +166,7 @@ from .types import (
|
|
168
166
|
InstructorVectorizerConfig,
|
169
167
|
InstructorVectorizerConfigRequest,
|
170
168
|
IterationStateEnum,
|
171
|
-
|
172
|
-
JinjaPromptBlockRequest,
|
169
|
+
JinjaPromptBlock,
|
173
170
|
JsonInputRequest,
|
174
171
|
JsonVariableValue,
|
175
172
|
JsonVellumValue,
|
@@ -254,11 +251,11 @@ from .types import (
|
|
254
251
|
PaginatedWorkflowReleaseTagReadList,
|
255
252
|
PdfSearchResultMetaSource,
|
256
253
|
PdfSearchResultMetaSourceRequest,
|
257
|
-
|
254
|
+
PlainTextPromptBlock,
|
258
255
|
Price,
|
259
256
|
ProcessingFailureReasonEnum,
|
260
257
|
ProcessingStateEnum,
|
261
|
-
|
258
|
+
PromptBlock,
|
262
259
|
PromptBlockState,
|
263
260
|
PromptDeploymentExpandMetaRequest,
|
264
261
|
PromptDeploymentInputRequest,
|
@@ -267,12 +264,12 @@ from .types import (
|
|
267
264
|
PromptNodeResult,
|
268
265
|
PromptNodeResultData,
|
269
266
|
PromptOutput,
|
270
|
-
|
271
|
-
|
272
|
-
|
273
|
-
|
274
|
-
|
275
|
-
|
267
|
+
PromptParameters,
|
268
|
+
PromptRequestChatHistoryInput,
|
269
|
+
PromptRequestInput,
|
270
|
+
PromptRequestJsonInput,
|
271
|
+
PromptRequestStringInput,
|
272
|
+
PromptSettings,
|
276
273
|
RawPromptExecutionOverridesRequest,
|
277
274
|
ReductoChunkerConfig,
|
278
275
|
ReductoChunkerConfigRequest,
|
@@ -286,8 +283,8 @@ from .types import (
|
|
286
283
|
RejectedWorkflowNodeResultEvent,
|
287
284
|
ReleaseTagSource,
|
288
285
|
ReplaceTestSuiteTestCaseRequest,
|
289
|
-
|
290
|
-
|
286
|
+
RichTextChildBlock,
|
287
|
+
RichTextPromptBlock,
|
291
288
|
SandboxScenario,
|
292
289
|
ScenarioInput,
|
293
290
|
ScenarioInputChatHistoryVariableValue,
|
@@ -419,7 +416,7 @@ from .types import (
|
|
419
416
|
UnitEnum,
|
420
417
|
UploadDocumentResponse,
|
421
418
|
UpsertTestSuiteTestCaseRequest,
|
422
|
-
|
419
|
+
VariablePromptBlock,
|
423
420
|
VellumAudio,
|
424
421
|
VellumAudioRequest,
|
425
422
|
VellumError,
|
@@ -434,8 +431,6 @@ from .types import (
|
|
434
431
|
VellumValueRequest,
|
435
432
|
VellumVariable,
|
436
433
|
VellumVariableExtensions,
|
437
|
-
VellumVariableExtensionsRequest,
|
438
|
-
VellumVariableRequest,
|
439
434
|
VellumVariableType,
|
440
435
|
WorkflowDeploymentRead,
|
441
436
|
WorkflowEventError,
|
@@ -513,7 +508,7 @@ from .version import __version__
|
|
513
508
|
|
514
509
|
__all__ = [
|
515
510
|
"AdHocExecutePromptEvent",
|
516
|
-
"
|
511
|
+
"AdHocExpandMeta",
|
517
512
|
"AdHocFulfilledPromptExecutionMeta",
|
518
513
|
"AdHocInitiatedPromptExecutionMeta",
|
519
514
|
"AdHocRejectedPromptExecutionMeta",
|
@@ -550,8 +545,7 @@ __all__ = [
|
|
550
545
|
"ChatMessage",
|
551
546
|
"ChatMessageContent",
|
552
547
|
"ChatMessageContentRequest",
|
553
|
-
"
|
554
|
-
"ChatMessagePromptBlockRequest",
|
548
|
+
"ChatMessagePromptBlock",
|
555
549
|
"ChatMessageRequest",
|
556
550
|
"ChatMessageRole",
|
557
551
|
"CodeExecutionNodeArrayResult",
|
@@ -599,7 +593,7 @@ __all__ = [
|
|
599
593
|
"EntityStatus",
|
600
594
|
"EntityVisibility",
|
601
595
|
"EnvironmentEnum",
|
602
|
-
"
|
596
|
+
"EphemeralPromptCacheConfig",
|
603
597
|
"EphemeralPromptCacheConfigTypeEnum",
|
604
598
|
"ErrorInputRequest",
|
605
599
|
"ErrorVariableValue",
|
@@ -651,8 +645,7 @@ __all__ = [
|
|
651
645
|
"FunctionCallVariableValue",
|
652
646
|
"FunctionCallVellumValue",
|
653
647
|
"FunctionCallVellumValueRequest",
|
654
|
-
"
|
655
|
-
"FunctionDefinitionPromptBlockRequest",
|
648
|
+
"FunctionDefinitionPromptBlock",
|
656
649
|
"GenerateOptionsRequest",
|
657
650
|
"GenerateRequest",
|
658
651
|
"GenerateResponse",
|
@@ -686,8 +679,7 @@ __all__ = [
|
|
686
679
|
"InstructorVectorizerConfigRequest",
|
687
680
|
"InternalServerError",
|
688
681
|
"IterationStateEnum",
|
689
|
-
"
|
690
|
-
"JinjaPromptBlockRequest",
|
682
|
+
"JinjaPromptBlock",
|
691
683
|
"JsonInputRequest",
|
692
684
|
"JsonVariableValue",
|
693
685
|
"JsonVellumValue",
|
@@ -775,11 +767,11 @@ __all__ = [
|
|
775
767
|
"PaginatedWorkflowReleaseTagReadList",
|
776
768
|
"PdfSearchResultMetaSource",
|
777
769
|
"PdfSearchResultMetaSourceRequest",
|
778
|
-
"
|
770
|
+
"PlainTextPromptBlock",
|
779
771
|
"Price",
|
780
772
|
"ProcessingFailureReasonEnum",
|
781
773
|
"ProcessingStateEnum",
|
782
|
-
"
|
774
|
+
"PromptBlock",
|
783
775
|
"PromptBlockState",
|
784
776
|
"PromptDeploymentExpandMetaRequest",
|
785
777
|
"PromptDeploymentInputRequest",
|
@@ -788,12 +780,12 @@ __all__ = [
|
|
788
780
|
"PromptNodeResult",
|
789
781
|
"PromptNodeResultData",
|
790
782
|
"PromptOutput",
|
791
|
-
"
|
792
|
-
"
|
793
|
-
"
|
794
|
-
"
|
795
|
-
"
|
796
|
-
"
|
783
|
+
"PromptParameters",
|
784
|
+
"PromptRequestChatHistoryInput",
|
785
|
+
"PromptRequestInput",
|
786
|
+
"PromptRequestJsonInput",
|
787
|
+
"PromptRequestStringInput",
|
788
|
+
"PromptSettings",
|
797
789
|
"RawPromptExecutionOverridesRequest",
|
798
790
|
"ReductoChunkerConfig",
|
799
791
|
"ReductoChunkerConfigRequest",
|
@@ -807,8 +799,8 @@ __all__ = [
|
|
807
799
|
"RejectedWorkflowNodeResultEvent",
|
808
800
|
"ReleaseTagSource",
|
809
801
|
"ReplaceTestSuiteTestCaseRequest",
|
810
|
-
"
|
811
|
-
"
|
802
|
+
"RichTextChildBlock",
|
803
|
+
"RichTextPromptBlock",
|
812
804
|
"SandboxScenario",
|
813
805
|
"ScenarioInput",
|
814
806
|
"ScenarioInputChatHistoryVariableValue",
|
@@ -940,7 +932,7 @@ __all__ = [
|
|
940
932
|
"UnitEnum",
|
941
933
|
"UploadDocumentResponse",
|
942
934
|
"UpsertTestSuiteTestCaseRequest",
|
943
|
-
"
|
935
|
+
"VariablePromptBlock",
|
944
936
|
"Vellum",
|
945
937
|
"VellumAudio",
|
946
938
|
"VellumAudioRequest",
|
@@ -957,8 +949,6 @@ __all__ = [
|
|
957
949
|
"VellumValueRequest",
|
958
950
|
"VellumVariable",
|
959
951
|
"VellumVariableExtensions",
|
960
|
-
"VellumVariableExtensionsRequest",
|
961
|
-
"VellumVariableRequest",
|
962
952
|
"VellumVariableType",
|
963
953
|
"WorkflowDeploymentRead",
|
964
954
|
"WorkflowDeploymentsListRequestStatus",
|
vellum/core/client_wrapper.py
CHANGED
@@ -2,12 +2,12 @@
|
|
2
2
|
|
3
3
|
import typing
|
4
4
|
from ...core.client_wrapper import SyncClientWrapper
|
5
|
-
from ...types.
|
6
|
-
from ...types.
|
7
|
-
from ...types.
|
8
|
-
from ...types.
|
9
|
-
from ...types.
|
10
|
-
from ...types.
|
5
|
+
from ...types.prompt_request_input import PromptRequestInput
|
6
|
+
from ...types.vellum_variable import VellumVariable
|
7
|
+
from ...types.prompt_parameters import PromptParameters
|
8
|
+
from ...types.prompt_block import PromptBlock
|
9
|
+
from ...types.prompt_settings import PromptSettings
|
10
|
+
from ...types.ad_hoc_expand_meta import AdHocExpandMeta
|
11
11
|
from ...core.request_options import RequestOptions
|
12
12
|
from ...types.ad_hoc_execute_prompt_event import AdHocExecutePromptEvent
|
13
13
|
from ...core.serialization import convert_and_respect_annotation_metadata
|
@@ -32,12 +32,12 @@ class AdHocClient:
|
|
32
32
|
self,
|
33
33
|
*,
|
34
34
|
ml_model: str,
|
35
|
-
input_values: typing.Sequence[
|
36
|
-
input_variables: typing.Sequence[
|
37
|
-
parameters:
|
38
|
-
blocks: typing.Sequence[
|
39
|
-
settings: typing.Optional[
|
40
|
-
expand_meta: typing.Optional[
|
35
|
+
input_values: typing.Sequence[PromptRequestInput],
|
36
|
+
input_variables: typing.Sequence[VellumVariable],
|
37
|
+
parameters: PromptParameters,
|
38
|
+
blocks: typing.Sequence[PromptBlock],
|
39
|
+
settings: typing.Optional[PromptSettings] = OMIT,
|
40
|
+
expand_meta: typing.Optional[AdHocExpandMeta] = OMIT,
|
41
41
|
request_options: typing.Optional[RequestOptions] = None,
|
42
42
|
) -> typing.Iterator[AdHocExecutePromptEvent]:
|
43
43
|
"""
|
@@ -47,17 +47,17 @@ class AdHocClient:
|
|
47
47
|
----------
|
48
48
|
ml_model : str
|
49
49
|
|
50
|
-
input_values : typing.Sequence[
|
50
|
+
input_values : typing.Sequence[PromptRequestInput]
|
51
51
|
|
52
|
-
input_variables : typing.Sequence[
|
52
|
+
input_variables : typing.Sequence[VellumVariable]
|
53
53
|
|
54
|
-
parameters :
|
54
|
+
parameters : PromptParameters
|
55
55
|
|
56
|
-
blocks : typing.Sequence[
|
56
|
+
blocks : typing.Sequence[PromptBlock]
|
57
57
|
|
58
|
-
settings : typing.Optional[
|
58
|
+
settings : typing.Optional[PromptSettings]
|
59
59
|
|
60
|
-
expand_meta : typing.Optional[
|
60
|
+
expand_meta : typing.Optional[AdHocExpandMeta]
|
61
61
|
|
62
62
|
request_options : typing.Optional[RequestOptions]
|
63
63
|
Request-specific configuration.
|
@@ -70,17 +70,16 @@ class AdHocClient:
|
|
70
70
|
Examples
|
71
71
|
--------
|
72
72
|
from vellum import (
|
73
|
-
|
74
|
-
|
75
|
-
|
76
|
-
|
77
|
-
|
78
|
-
|
79
|
-
|
80
|
-
StringVellumValueRequest,
|
73
|
+
AdHocExpandMeta,
|
74
|
+
EphemeralPromptCacheConfig,
|
75
|
+
JinjaPromptBlock,
|
76
|
+
PromptParameters,
|
77
|
+
PromptRequestStringInput,
|
78
|
+
PromptSettings,
|
79
|
+
StringVellumValue,
|
81
80
|
Vellum,
|
82
|
-
|
83
|
-
|
81
|
+
VellumVariable,
|
82
|
+
VellumVariableExtensions,
|
84
83
|
)
|
85
84
|
|
86
85
|
client = Vellum(
|
@@ -89,26 +88,26 @@ class AdHocClient:
|
|
89
88
|
response = client.ad_hoc.adhoc_execute_prompt_stream(
|
90
89
|
ml_model="string",
|
91
90
|
input_values=[
|
92
|
-
|
91
|
+
PromptRequestStringInput(
|
93
92
|
key="string",
|
94
93
|
value="string",
|
95
94
|
)
|
96
95
|
],
|
97
96
|
input_variables=[
|
98
|
-
|
97
|
+
VellumVariable(
|
99
98
|
id="string",
|
100
99
|
key="string",
|
101
100
|
type="STRING",
|
102
101
|
required=True,
|
103
|
-
default=
|
104
|
-
value="
|
102
|
+
default=StringVellumValue(
|
103
|
+
value={"key": "value"},
|
105
104
|
),
|
106
|
-
extensions=
|
107
|
-
color="
|
105
|
+
extensions=VellumVariableExtensions(
|
106
|
+
color={"key": "value"},
|
108
107
|
),
|
109
108
|
)
|
110
109
|
],
|
111
|
-
parameters=
|
110
|
+
parameters=PromptParameters(
|
112
111
|
stop=["string"],
|
113
112
|
temperature=1.1,
|
114
113
|
max_tokens=1,
|
@@ -119,23 +118,18 @@ class AdHocClient:
|
|
119
118
|
logit_bias={"string": {"key": "value"}},
|
120
119
|
custom_parameters={"string": {"key": "value"}},
|
121
120
|
),
|
122
|
-
settings=
|
121
|
+
settings=PromptSettings(
|
123
122
|
timeout=1.1,
|
124
123
|
),
|
125
124
|
blocks=[
|
126
|
-
|
127
|
-
properties=JinjaPromptBlockPropertiesRequest(
|
128
|
-
template="string",
|
129
|
-
template_type="STRING",
|
130
|
-
),
|
131
|
-
id="string",
|
125
|
+
JinjaPromptBlock(
|
132
126
|
state="ENABLED",
|
133
|
-
cache_config=
|
134
|
-
|
135
|
-
|
127
|
+
cache_config=EphemeralPromptCacheConfig(),
|
128
|
+
template="string",
|
129
|
+
template_type="STRING",
|
136
130
|
)
|
137
131
|
],
|
138
|
-
expand_meta=
|
132
|
+
expand_meta=AdHocExpandMeta(
|
139
133
|
cost=True,
|
140
134
|
model_name=True,
|
141
135
|
usage=True,
|
@@ -152,22 +146,22 @@ class AdHocClient:
|
|
152
146
|
json={
|
153
147
|
"ml_model": ml_model,
|
154
148
|
"input_values": convert_and_respect_annotation_metadata(
|
155
|
-
object_=input_values, annotation=typing.Sequence[
|
149
|
+
object_=input_values, annotation=typing.Sequence[PromptRequestInput], direction="write"
|
156
150
|
),
|
157
151
|
"input_variables": convert_and_respect_annotation_metadata(
|
158
|
-
object_=input_variables, annotation=typing.Sequence[
|
152
|
+
object_=input_variables, annotation=typing.Sequence[VellumVariable], direction="write"
|
159
153
|
),
|
160
154
|
"parameters": convert_and_respect_annotation_metadata(
|
161
|
-
object_=parameters, annotation=
|
155
|
+
object_=parameters, annotation=PromptParameters, direction="write"
|
162
156
|
),
|
163
157
|
"settings": convert_and_respect_annotation_metadata(
|
164
|
-
object_=settings, annotation=
|
158
|
+
object_=settings, annotation=PromptSettings, direction="write"
|
165
159
|
),
|
166
160
|
"blocks": convert_and_respect_annotation_metadata(
|
167
|
-
object_=blocks, annotation=typing.Sequence[
|
161
|
+
object_=blocks, annotation=typing.Sequence[PromptBlock], direction="write"
|
168
162
|
),
|
169
163
|
"expand_meta": convert_and_respect_annotation_metadata(
|
170
|
-
object_=expand_meta, annotation=
|
164
|
+
object_=expand_meta, annotation=AdHocExpandMeta, direction="write"
|
171
165
|
),
|
172
166
|
},
|
173
167
|
request_options=request_options,
|
@@ -234,12 +228,12 @@ class AsyncAdHocClient:
|
|
234
228
|
self,
|
235
229
|
*,
|
236
230
|
ml_model: str,
|
237
|
-
input_values: typing.Sequence[
|
238
|
-
input_variables: typing.Sequence[
|
239
|
-
parameters:
|
240
|
-
blocks: typing.Sequence[
|
241
|
-
settings: typing.Optional[
|
242
|
-
expand_meta: typing.Optional[
|
231
|
+
input_values: typing.Sequence[PromptRequestInput],
|
232
|
+
input_variables: typing.Sequence[VellumVariable],
|
233
|
+
parameters: PromptParameters,
|
234
|
+
blocks: typing.Sequence[PromptBlock],
|
235
|
+
settings: typing.Optional[PromptSettings] = OMIT,
|
236
|
+
expand_meta: typing.Optional[AdHocExpandMeta] = OMIT,
|
243
237
|
request_options: typing.Optional[RequestOptions] = None,
|
244
238
|
) -> typing.AsyncIterator[AdHocExecutePromptEvent]:
|
245
239
|
"""
|
@@ -249,17 +243,17 @@ class AsyncAdHocClient:
|
|
249
243
|
----------
|
250
244
|
ml_model : str
|
251
245
|
|
252
|
-
input_values : typing.Sequence[
|
246
|
+
input_values : typing.Sequence[PromptRequestInput]
|
253
247
|
|
254
|
-
input_variables : typing.Sequence[
|
248
|
+
input_variables : typing.Sequence[VellumVariable]
|
255
249
|
|
256
|
-
parameters :
|
250
|
+
parameters : PromptParameters
|
257
251
|
|
258
|
-
blocks : typing.Sequence[
|
252
|
+
blocks : typing.Sequence[PromptBlock]
|
259
253
|
|
260
|
-
settings : typing.Optional[
|
254
|
+
settings : typing.Optional[PromptSettings]
|
261
255
|
|
262
|
-
expand_meta : typing.Optional[
|
256
|
+
expand_meta : typing.Optional[AdHocExpandMeta]
|
263
257
|
|
264
258
|
request_options : typing.Optional[RequestOptions]
|
265
259
|
Request-specific configuration.
|
@@ -274,17 +268,16 @@ class AsyncAdHocClient:
|
|
274
268
|
import asyncio
|
275
269
|
|
276
270
|
from vellum import (
|
277
|
-
|
271
|
+
AdHocExpandMeta,
|
278
272
|
AsyncVellum,
|
279
|
-
|
280
|
-
|
281
|
-
|
282
|
-
|
283
|
-
|
284
|
-
|
285
|
-
|
286
|
-
|
287
|
-
VellumVariableRequest,
|
273
|
+
EphemeralPromptCacheConfig,
|
274
|
+
JinjaPromptBlock,
|
275
|
+
PromptParameters,
|
276
|
+
PromptRequestStringInput,
|
277
|
+
PromptSettings,
|
278
|
+
StringVellumValue,
|
279
|
+
VellumVariable,
|
280
|
+
VellumVariableExtensions,
|
288
281
|
)
|
289
282
|
|
290
283
|
client = AsyncVellum(
|
@@ -296,26 +289,26 @@ class AsyncAdHocClient:
|
|
296
289
|
response = await client.ad_hoc.adhoc_execute_prompt_stream(
|
297
290
|
ml_model="string",
|
298
291
|
input_values=[
|
299
|
-
|
292
|
+
PromptRequestStringInput(
|
300
293
|
key="string",
|
301
294
|
value="string",
|
302
295
|
)
|
303
296
|
],
|
304
297
|
input_variables=[
|
305
|
-
|
298
|
+
VellumVariable(
|
306
299
|
id="string",
|
307
300
|
key="string",
|
308
301
|
type="STRING",
|
309
302
|
required=True,
|
310
|
-
default=
|
311
|
-
value="
|
303
|
+
default=StringVellumValue(
|
304
|
+
value={"key": "value"},
|
312
305
|
),
|
313
|
-
extensions=
|
314
|
-
color="
|
306
|
+
extensions=VellumVariableExtensions(
|
307
|
+
color={"key": "value"},
|
315
308
|
),
|
316
309
|
)
|
317
310
|
],
|
318
|
-
parameters=
|
311
|
+
parameters=PromptParameters(
|
319
312
|
stop=["string"],
|
320
313
|
temperature=1.1,
|
321
314
|
max_tokens=1,
|
@@ -326,23 +319,18 @@ class AsyncAdHocClient:
|
|
326
319
|
logit_bias={"string": {"key": "value"}},
|
327
320
|
custom_parameters={"string": {"key": "value"}},
|
328
321
|
),
|
329
|
-
settings=
|
322
|
+
settings=PromptSettings(
|
330
323
|
timeout=1.1,
|
331
324
|
),
|
332
325
|
blocks=[
|
333
|
-
|
334
|
-
properties=JinjaPromptBlockPropertiesRequest(
|
335
|
-
template="string",
|
336
|
-
template_type="STRING",
|
337
|
-
),
|
338
|
-
id="string",
|
326
|
+
JinjaPromptBlock(
|
339
327
|
state="ENABLED",
|
340
|
-
cache_config=
|
341
|
-
|
342
|
-
|
328
|
+
cache_config=EphemeralPromptCacheConfig(),
|
329
|
+
template="string",
|
330
|
+
template_type="STRING",
|
343
331
|
)
|
344
332
|
],
|
345
|
-
expand_meta=
|
333
|
+
expand_meta=AdHocExpandMeta(
|
346
334
|
cost=True,
|
347
335
|
model_name=True,
|
348
336
|
usage=True,
|
@@ -362,22 +350,22 @@ class AsyncAdHocClient:
|
|
362
350
|
json={
|
363
351
|
"ml_model": ml_model,
|
364
352
|
"input_values": convert_and_respect_annotation_metadata(
|
365
|
-
object_=input_values, annotation=typing.Sequence[
|
353
|
+
object_=input_values, annotation=typing.Sequence[PromptRequestInput], direction="write"
|
366
354
|
),
|
367
355
|
"input_variables": convert_and_respect_annotation_metadata(
|
368
|
-
object_=input_variables, annotation=typing.Sequence[
|
356
|
+
object_=input_variables, annotation=typing.Sequence[VellumVariable], direction="write"
|
369
357
|
),
|
370
358
|
"parameters": convert_and_respect_annotation_metadata(
|
371
|
-
object_=parameters, annotation=
|
359
|
+
object_=parameters, annotation=PromptParameters, direction="write"
|
372
360
|
),
|
373
361
|
"settings": convert_and_respect_annotation_metadata(
|
374
|
-
object_=settings, annotation=
|
362
|
+
object_=settings, annotation=PromptSettings, direction="write"
|
375
363
|
),
|
376
364
|
"blocks": convert_and_respect_annotation_metadata(
|
377
|
-
object_=blocks, annotation=typing.Sequence[
|
365
|
+
object_=blocks, annotation=typing.Sequence[PromptBlock], direction="write"
|
378
366
|
),
|
379
367
|
"expand_meta": convert_and_respect_annotation_metadata(
|
380
|
-
object_=expand_meta, annotation=
|
368
|
+
object_=expand_meta, annotation=AdHocExpandMeta, direction="write"
|
381
369
|
),
|
382
370
|
},
|
383
371
|
request_options=request_options,
|