vellum-ai 0.9.5__py3-none-any.whl → 0.9.9__py3-none-any.whl

Sign up to get free protection for your applications and to get access to all the features.
Files changed (32) hide show
  1. vellum/__init__.py +32 -42
  2. vellum/core/client_wrapper.py +1 -1
  3. vellum/resources/ad_hoc/client.py +86 -98
  4. vellum/types/__init__.py +32 -42
  5. vellum/types/{ad_hoc_expand_meta_request.py → ad_hoc_expand_meta.py} +1 -1
  6. vellum/types/{chat_message_prompt_block_request.py → chat_message_prompt_block.py} +11 -11
  7. vellum/types/{ephemeral_prompt_cache_config_request.py → ephemeral_prompt_cache_config.py} +1 -1
  8. vellum/types/{function_definition_prompt_block_request.py → function_definition_prompt_block.py} +7 -7
  9. vellum/types/{jinja_prompt_block_request.py → jinja_prompt_block.py} +7 -7
  10. vellum/types/{plain_text_prompt_block_request.py → plain_text_prompt_block.py} +4 -5
  11. vellum/types/prompt_block.py +15 -0
  12. vellum/types/{prompt_parameters_request.py → prompt_parameters.py} +1 -1
  13. vellum/types/{prompt_request_chat_history_input_request.py → prompt_request_chat_history_input.py} +3 -3
  14. vellum/types/prompt_request_input.py +8 -0
  15. vellum/types/{prompt_request_json_input_request.py → prompt_request_json_input.py} +1 -1
  16. vellum/types/{prompt_request_string_input_request.py → prompt_request_string_input.py} +1 -1
  17. vellum/types/{prompt_settings_request.py → prompt_settings.py} +1 -1
  18. vellum/types/rich_text_child_block.py +7 -0
  19. vellum/types/{rich_text_prompt_block_request.py → rich_text_prompt_block.py} +6 -7
  20. vellum/types/{variable_prompt_block_request.py → variable_prompt_block.py} +4 -5
  21. {vellum_ai-0.9.5.dist-info → vellum_ai-0.9.9.dist-info}/METADATA +1 -1
  22. {vellum_ai-0.9.5.dist-info → vellum_ai-0.9.9.dist-info}/RECORD +24 -29
  23. vellum/types/chat_message_prompt_block_properties_request.py +0 -38
  24. vellum/types/function_definition_prompt_block_properties_request.py +0 -42
  25. vellum/types/jinja_prompt_block_properties_request.py +0 -21
  26. vellum/types/prompt_block_request.py +0 -19
  27. vellum/types/prompt_request_input_request.py +0 -10
  28. vellum/types/rich_text_child_block_request.py +0 -7
  29. vellum/types/vellum_variable_extensions_request.py +0 -23
  30. vellum/types/vellum_variable_request.py +0 -33
  31. {vellum_ai-0.9.5.dist-info → vellum_ai-0.9.9.dist-info}/LICENSE +0 -0
  32. {vellum_ai-0.9.5.dist-info → vellum_ai-0.9.9.dist-info}/WHEEL +0 -0
vellum/__init__.py CHANGED
@@ -2,7 +2,7 @@
2
2
 
3
3
  from .types import (
4
4
  AdHocExecutePromptEvent,
5
- AdHocExpandMetaRequest,
5
+ AdHocExpandMeta,
6
6
  AdHocFulfilledPromptExecutionMeta,
7
7
  AdHocInitiatedPromptExecutionMeta,
8
8
  AdHocRejectedPromptExecutionMeta,
@@ -37,8 +37,7 @@ from .types import (
37
37
  ChatMessage,
38
38
  ChatMessageContent,
39
39
  ChatMessageContentRequest,
40
- ChatMessagePromptBlockPropertiesRequest,
41
- ChatMessagePromptBlockRequest,
40
+ ChatMessagePromptBlock,
42
41
  ChatMessageRequest,
43
42
  ChatMessageRole,
44
43
  CodeExecutionNodeArrayResult,
@@ -84,7 +83,7 @@ from .types import (
84
83
  EntityStatus,
85
84
  EntityVisibility,
86
85
  EnvironmentEnum,
87
- EphemeralPromptCacheConfigRequest,
86
+ EphemeralPromptCacheConfig,
88
87
  EphemeralPromptCacheConfigTypeEnum,
89
88
  ErrorInputRequest,
90
89
  ErrorVariableValue,
@@ -134,8 +133,7 @@ from .types import (
134
133
  FunctionCallVariableValue,
135
134
  FunctionCallVellumValue,
136
135
  FunctionCallVellumValueRequest,
137
- FunctionDefinitionPromptBlockPropertiesRequest,
138
- FunctionDefinitionPromptBlockRequest,
136
+ FunctionDefinitionPromptBlock,
139
137
  GenerateOptionsRequest,
140
138
  GenerateRequest,
141
139
  GenerateResponse,
@@ -168,8 +166,7 @@ from .types import (
168
166
  InstructorVectorizerConfig,
169
167
  InstructorVectorizerConfigRequest,
170
168
  IterationStateEnum,
171
- JinjaPromptBlockPropertiesRequest,
172
- JinjaPromptBlockRequest,
169
+ JinjaPromptBlock,
173
170
  JsonInputRequest,
174
171
  JsonVariableValue,
175
172
  JsonVellumValue,
@@ -254,11 +251,11 @@ from .types import (
254
251
  PaginatedWorkflowReleaseTagReadList,
255
252
  PdfSearchResultMetaSource,
256
253
  PdfSearchResultMetaSourceRequest,
257
- PlainTextPromptBlockRequest,
254
+ PlainTextPromptBlock,
258
255
  Price,
259
256
  ProcessingFailureReasonEnum,
260
257
  ProcessingStateEnum,
261
- PromptBlockRequest,
258
+ PromptBlock,
262
259
  PromptBlockState,
263
260
  PromptDeploymentExpandMetaRequest,
264
261
  PromptDeploymentInputRequest,
@@ -267,12 +264,12 @@ from .types import (
267
264
  PromptNodeResult,
268
265
  PromptNodeResultData,
269
266
  PromptOutput,
270
- PromptParametersRequest,
271
- PromptRequestChatHistoryInputRequest,
272
- PromptRequestInputRequest,
273
- PromptRequestJsonInputRequest,
274
- PromptRequestStringInputRequest,
275
- PromptSettingsRequest,
267
+ PromptParameters,
268
+ PromptRequestChatHistoryInput,
269
+ PromptRequestInput,
270
+ PromptRequestJsonInput,
271
+ PromptRequestStringInput,
272
+ PromptSettings,
276
273
  RawPromptExecutionOverridesRequest,
277
274
  ReductoChunkerConfig,
278
275
  ReductoChunkerConfigRequest,
@@ -286,8 +283,8 @@ from .types import (
286
283
  RejectedWorkflowNodeResultEvent,
287
284
  ReleaseTagSource,
288
285
  ReplaceTestSuiteTestCaseRequest,
289
- RichTextChildBlockRequest,
290
- RichTextPromptBlockRequest,
286
+ RichTextChildBlock,
287
+ RichTextPromptBlock,
291
288
  SandboxScenario,
292
289
  ScenarioInput,
293
290
  ScenarioInputChatHistoryVariableValue,
@@ -419,7 +416,7 @@ from .types import (
419
416
  UnitEnum,
420
417
  UploadDocumentResponse,
421
418
  UpsertTestSuiteTestCaseRequest,
422
- VariablePromptBlockRequest,
419
+ VariablePromptBlock,
423
420
  VellumAudio,
424
421
  VellumAudioRequest,
425
422
  VellumError,
@@ -434,8 +431,6 @@ from .types import (
434
431
  VellumValueRequest,
435
432
  VellumVariable,
436
433
  VellumVariableExtensions,
437
- VellumVariableExtensionsRequest,
438
- VellumVariableRequest,
439
434
  VellumVariableType,
440
435
  WorkflowDeploymentRead,
441
436
  WorkflowEventError,
@@ -513,7 +508,7 @@ from .version import __version__
513
508
 
514
509
  __all__ = [
515
510
  "AdHocExecutePromptEvent",
516
- "AdHocExpandMetaRequest",
511
+ "AdHocExpandMeta",
517
512
  "AdHocFulfilledPromptExecutionMeta",
518
513
  "AdHocInitiatedPromptExecutionMeta",
519
514
  "AdHocRejectedPromptExecutionMeta",
@@ -550,8 +545,7 @@ __all__ = [
550
545
  "ChatMessage",
551
546
  "ChatMessageContent",
552
547
  "ChatMessageContentRequest",
553
- "ChatMessagePromptBlockPropertiesRequest",
554
- "ChatMessagePromptBlockRequest",
548
+ "ChatMessagePromptBlock",
555
549
  "ChatMessageRequest",
556
550
  "ChatMessageRole",
557
551
  "CodeExecutionNodeArrayResult",
@@ -599,7 +593,7 @@ __all__ = [
599
593
  "EntityStatus",
600
594
  "EntityVisibility",
601
595
  "EnvironmentEnum",
602
- "EphemeralPromptCacheConfigRequest",
596
+ "EphemeralPromptCacheConfig",
603
597
  "EphemeralPromptCacheConfigTypeEnum",
604
598
  "ErrorInputRequest",
605
599
  "ErrorVariableValue",
@@ -651,8 +645,7 @@ __all__ = [
651
645
  "FunctionCallVariableValue",
652
646
  "FunctionCallVellumValue",
653
647
  "FunctionCallVellumValueRequest",
654
- "FunctionDefinitionPromptBlockPropertiesRequest",
655
- "FunctionDefinitionPromptBlockRequest",
648
+ "FunctionDefinitionPromptBlock",
656
649
  "GenerateOptionsRequest",
657
650
  "GenerateRequest",
658
651
  "GenerateResponse",
@@ -686,8 +679,7 @@ __all__ = [
686
679
  "InstructorVectorizerConfigRequest",
687
680
  "InternalServerError",
688
681
  "IterationStateEnum",
689
- "JinjaPromptBlockPropertiesRequest",
690
- "JinjaPromptBlockRequest",
682
+ "JinjaPromptBlock",
691
683
  "JsonInputRequest",
692
684
  "JsonVariableValue",
693
685
  "JsonVellumValue",
@@ -775,11 +767,11 @@ __all__ = [
775
767
  "PaginatedWorkflowReleaseTagReadList",
776
768
  "PdfSearchResultMetaSource",
777
769
  "PdfSearchResultMetaSourceRequest",
778
- "PlainTextPromptBlockRequest",
770
+ "PlainTextPromptBlock",
779
771
  "Price",
780
772
  "ProcessingFailureReasonEnum",
781
773
  "ProcessingStateEnum",
782
- "PromptBlockRequest",
774
+ "PromptBlock",
783
775
  "PromptBlockState",
784
776
  "PromptDeploymentExpandMetaRequest",
785
777
  "PromptDeploymentInputRequest",
@@ -788,12 +780,12 @@ __all__ = [
788
780
  "PromptNodeResult",
789
781
  "PromptNodeResultData",
790
782
  "PromptOutput",
791
- "PromptParametersRequest",
792
- "PromptRequestChatHistoryInputRequest",
793
- "PromptRequestInputRequest",
794
- "PromptRequestJsonInputRequest",
795
- "PromptRequestStringInputRequest",
796
- "PromptSettingsRequest",
783
+ "PromptParameters",
784
+ "PromptRequestChatHistoryInput",
785
+ "PromptRequestInput",
786
+ "PromptRequestJsonInput",
787
+ "PromptRequestStringInput",
788
+ "PromptSettings",
797
789
  "RawPromptExecutionOverridesRequest",
798
790
  "ReductoChunkerConfig",
799
791
  "ReductoChunkerConfigRequest",
@@ -807,8 +799,8 @@ __all__ = [
807
799
  "RejectedWorkflowNodeResultEvent",
808
800
  "ReleaseTagSource",
809
801
  "ReplaceTestSuiteTestCaseRequest",
810
- "RichTextChildBlockRequest",
811
- "RichTextPromptBlockRequest",
802
+ "RichTextChildBlock",
803
+ "RichTextPromptBlock",
812
804
  "SandboxScenario",
813
805
  "ScenarioInput",
814
806
  "ScenarioInputChatHistoryVariableValue",
@@ -940,7 +932,7 @@ __all__ = [
940
932
  "UnitEnum",
941
933
  "UploadDocumentResponse",
942
934
  "UpsertTestSuiteTestCaseRequest",
943
- "VariablePromptBlockRequest",
935
+ "VariablePromptBlock",
944
936
  "Vellum",
945
937
  "VellumAudio",
946
938
  "VellumAudioRequest",
@@ -957,8 +949,6 @@ __all__ = [
957
949
  "VellumValueRequest",
958
950
  "VellumVariable",
959
951
  "VellumVariableExtensions",
960
- "VellumVariableExtensionsRequest",
961
- "VellumVariableRequest",
962
952
  "VellumVariableType",
963
953
  "WorkflowDeploymentRead",
964
954
  "WorkflowDeploymentsListRequestStatus",
@@ -17,7 +17,7 @@ class BaseClientWrapper:
17
17
  headers: typing.Dict[str, str] = {
18
18
  "X-Fern-Language": "Python",
19
19
  "X-Fern-SDK-Name": "vellum-ai",
20
- "X-Fern-SDK-Version": "0.9.5",
20
+ "X-Fern-SDK-Version": "0.9.9",
21
21
  }
22
22
  headers["X_API_KEY"] = self.api_key
23
23
  return headers
@@ -2,12 +2,12 @@
2
2
 
3
3
  import typing
4
4
  from ...core.client_wrapper import SyncClientWrapper
5
- from ...types.prompt_request_input_request import PromptRequestInputRequest
6
- from ...types.vellum_variable_request import VellumVariableRequest
7
- from ...types.prompt_parameters_request import PromptParametersRequest
8
- from ...types.prompt_block_request import PromptBlockRequest
9
- from ...types.prompt_settings_request import PromptSettingsRequest
10
- from ...types.ad_hoc_expand_meta_request import AdHocExpandMetaRequest
5
+ from ...types.prompt_request_input import PromptRequestInput
6
+ from ...types.vellum_variable import VellumVariable
7
+ from ...types.prompt_parameters import PromptParameters
8
+ from ...types.prompt_block import PromptBlock
9
+ from ...types.prompt_settings import PromptSettings
10
+ from ...types.ad_hoc_expand_meta import AdHocExpandMeta
11
11
  from ...core.request_options import RequestOptions
12
12
  from ...types.ad_hoc_execute_prompt_event import AdHocExecutePromptEvent
13
13
  from ...core.serialization import convert_and_respect_annotation_metadata
@@ -32,12 +32,12 @@ class AdHocClient:
32
32
  self,
33
33
  *,
34
34
  ml_model: str,
35
- input_values: typing.Sequence[PromptRequestInputRequest],
36
- input_variables: typing.Sequence[VellumVariableRequest],
37
- parameters: PromptParametersRequest,
38
- blocks: typing.Sequence[PromptBlockRequest],
39
- settings: typing.Optional[PromptSettingsRequest] = OMIT,
40
- expand_meta: typing.Optional[AdHocExpandMetaRequest] = OMIT,
35
+ input_values: typing.Sequence[PromptRequestInput],
36
+ input_variables: typing.Sequence[VellumVariable],
37
+ parameters: PromptParameters,
38
+ blocks: typing.Sequence[PromptBlock],
39
+ settings: typing.Optional[PromptSettings] = OMIT,
40
+ expand_meta: typing.Optional[AdHocExpandMeta] = OMIT,
41
41
  request_options: typing.Optional[RequestOptions] = None,
42
42
  ) -> typing.Iterator[AdHocExecutePromptEvent]:
43
43
  """
@@ -47,17 +47,17 @@ class AdHocClient:
47
47
  ----------
48
48
  ml_model : str
49
49
 
50
- input_values : typing.Sequence[PromptRequestInputRequest]
50
+ input_values : typing.Sequence[PromptRequestInput]
51
51
 
52
- input_variables : typing.Sequence[VellumVariableRequest]
52
+ input_variables : typing.Sequence[VellumVariable]
53
53
 
54
- parameters : PromptParametersRequest
54
+ parameters : PromptParameters
55
55
 
56
- blocks : typing.Sequence[PromptBlockRequest]
56
+ blocks : typing.Sequence[PromptBlock]
57
57
 
58
- settings : typing.Optional[PromptSettingsRequest]
58
+ settings : typing.Optional[PromptSettings]
59
59
 
60
- expand_meta : typing.Optional[AdHocExpandMetaRequest]
60
+ expand_meta : typing.Optional[AdHocExpandMeta]
61
61
 
62
62
  request_options : typing.Optional[RequestOptions]
63
63
  Request-specific configuration.
@@ -70,17 +70,16 @@ class AdHocClient:
70
70
  Examples
71
71
  --------
72
72
  from vellum import (
73
- AdHocExpandMetaRequest,
74
- EphemeralPromptCacheConfigRequest,
75
- JinjaPromptBlockPropertiesRequest,
76
- JinjaPromptBlockRequest,
77
- PromptParametersRequest,
78
- PromptRequestStringInputRequest,
79
- PromptSettingsRequest,
80
- StringVellumValueRequest,
73
+ AdHocExpandMeta,
74
+ EphemeralPromptCacheConfig,
75
+ JinjaPromptBlock,
76
+ PromptParameters,
77
+ PromptRequestStringInput,
78
+ PromptSettings,
79
+ StringVellumValue,
81
80
  Vellum,
82
- VellumVariableExtensionsRequest,
83
- VellumVariableRequest,
81
+ VellumVariable,
82
+ VellumVariableExtensions,
84
83
  )
85
84
 
86
85
  client = Vellum(
@@ -89,26 +88,26 @@ class AdHocClient:
89
88
  response = client.ad_hoc.adhoc_execute_prompt_stream(
90
89
  ml_model="string",
91
90
  input_values=[
92
- PromptRequestStringInputRequest(
91
+ PromptRequestStringInput(
93
92
  key="string",
94
93
  value="string",
95
94
  )
96
95
  ],
97
96
  input_variables=[
98
- VellumVariableRequest(
97
+ VellumVariable(
99
98
  id="string",
100
99
  key="string",
101
100
  type="STRING",
102
101
  required=True,
103
- default=StringVellumValueRequest(
104
- value="string",
102
+ default=StringVellumValue(
103
+ value={"key": "value"},
105
104
  ),
106
- extensions=VellumVariableExtensionsRequest(
107
- color="string",
105
+ extensions=VellumVariableExtensions(
106
+ color={"key": "value"},
108
107
  ),
109
108
  )
110
109
  ],
111
- parameters=PromptParametersRequest(
110
+ parameters=PromptParameters(
112
111
  stop=["string"],
113
112
  temperature=1.1,
114
113
  max_tokens=1,
@@ -119,23 +118,18 @@ class AdHocClient:
119
118
  logit_bias={"string": {"key": "value"}},
120
119
  custom_parameters={"string": {"key": "value"}},
121
120
  ),
122
- settings=PromptSettingsRequest(
121
+ settings=PromptSettings(
123
122
  timeout=1.1,
124
123
  ),
125
124
  blocks=[
126
- JinjaPromptBlockRequest(
127
- properties=JinjaPromptBlockPropertiesRequest(
128
- template="string",
129
- template_type="STRING",
130
- ),
131
- id="string",
125
+ JinjaPromptBlock(
132
126
  state="ENABLED",
133
- cache_config=EphemeralPromptCacheConfigRequest(
134
- type={"key": "value"},
135
- ),
127
+ cache_config=EphemeralPromptCacheConfig(),
128
+ template="string",
129
+ template_type="STRING",
136
130
  )
137
131
  ],
138
- expand_meta=AdHocExpandMetaRequest(
132
+ expand_meta=AdHocExpandMeta(
139
133
  cost=True,
140
134
  model_name=True,
141
135
  usage=True,
@@ -152,22 +146,22 @@ class AdHocClient:
152
146
  json={
153
147
  "ml_model": ml_model,
154
148
  "input_values": convert_and_respect_annotation_metadata(
155
- object_=input_values, annotation=typing.Sequence[PromptRequestInputRequest], direction="write"
149
+ object_=input_values, annotation=typing.Sequence[PromptRequestInput], direction="write"
156
150
  ),
157
151
  "input_variables": convert_and_respect_annotation_metadata(
158
- object_=input_variables, annotation=typing.Sequence[VellumVariableRequest], direction="write"
152
+ object_=input_variables, annotation=typing.Sequence[VellumVariable], direction="write"
159
153
  ),
160
154
  "parameters": convert_and_respect_annotation_metadata(
161
- object_=parameters, annotation=PromptParametersRequest, direction="write"
155
+ object_=parameters, annotation=PromptParameters, direction="write"
162
156
  ),
163
157
  "settings": convert_and_respect_annotation_metadata(
164
- object_=settings, annotation=PromptSettingsRequest, direction="write"
158
+ object_=settings, annotation=PromptSettings, direction="write"
165
159
  ),
166
160
  "blocks": convert_and_respect_annotation_metadata(
167
- object_=blocks, annotation=typing.Sequence[PromptBlockRequest], direction="write"
161
+ object_=blocks, annotation=typing.Sequence[PromptBlock], direction="write"
168
162
  ),
169
163
  "expand_meta": convert_and_respect_annotation_metadata(
170
- object_=expand_meta, annotation=AdHocExpandMetaRequest, direction="write"
164
+ object_=expand_meta, annotation=AdHocExpandMeta, direction="write"
171
165
  ),
172
166
  },
173
167
  request_options=request_options,
@@ -234,12 +228,12 @@ class AsyncAdHocClient:
234
228
  self,
235
229
  *,
236
230
  ml_model: str,
237
- input_values: typing.Sequence[PromptRequestInputRequest],
238
- input_variables: typing.Sequence[VellumVariableRequest],
239
- parameters: PromptParametersRequest,
240
- blocks: typing.Sequence[PromptBlockRequest],
241
- settings: typing.Optional[PromptSettingsRequest] = OMIT,
242
- expand_meta: typing.Optional[AdHocExpandMetaRequest] = OMIT,
231
+ input_values: typing.Sequence[PromptRequestInput],
232
+ input_variables: typing.Sequence[VellumVariable],
233
+ parameters: PromptParameters,
234
+ blocks: typing.Sequence[PromptBlock],
235
+ settings: typing.Optional[PromptSettings] = OMIT,
236
+ expand_meta: typing.Optional[AdHocExpandMeta] = OMIT,
243
237
  request_options: typing.Optional[RequestOptions] = None,
244
238
  ) -> typing.AsyncIterator[AdHocExecutePromptEvent]:
245
239
  """
@@ -249,17 +243,17 @@ class AsyncAdHocClient:
249
243
  ----------
250
244
  ml_model : str
251
245
 
252
- input_values : typing.Sequence[PromptRequestInputRequest]
246
+ input_values : typing.Sequence[PromptRequestInput]
253
247
 
254
- input_variables : typing.Sequence[VellumVariableRequest]
248
+ input_variables : typing.Sequence[VellumVariable]
255
249
 
256
- parameters : PromptParametersRequest
250
+ parameters : PromptParameters
257
251
 
258
- blocks : typing.Sequence[PromptBlockRequest]
252
+ blocks : typing.Sequence[PromptBlock]
259
253
 
260
- settings : typing.Optional[PromptSettingsRequest]
254
+ settings : typing.Optional[PromptSettings]
261
255
 
262
- expand_meta : typing.Optional[AdHocExpandMetaRequest]
256
+ expand_meta : typing.Optional[AdHocExpandMeta]
263
257
 
264
258
  request_options : typing.Optional[RequestOptions]
265
259
  Request-specific configuration.
@@ -274,17 +268,16 @@ class AsyncAdHocClient:
274
268
  import asyncio
275
269
 
276
270
  from vellum import (
277
- AdHocExpandMetaRequest,
271
+ AdHocExpandMeta,
278
272
  AsyncVellum,
279
- EphemeralPromptCacheConfigRequest,
280
- JinjaPromptBlockPropertiesRequest,
281
- JinjaPromptBlockRequest,
282
- PromptParametersRequest,
283
- PromptRequestStringInputRequest,
284
- PromptSettingsRequest,
285
- StringVellumValueRequest,
286
- VellumVariableExtensionsRequest,
287
- VellumVariableRequest,
273
+ EphemeralPromptCacheConfig,
274
+ JinjaPromptBlock,
275
+ PromptParameters,
276
+ PromptRequestStringInput,
277
+ PromptSettings,
278
+ StringVellumValue,
279
+ VellumVariable,
280
+ VellumVariableExtensions,
288
281
  )
289
282
 
290
283
  client = AsyncVellum(
@@ -296,26 +289,26 @@ class AsyncAdHocClient:
296
289
  response = await client.ad_hoc.adhoc_execute_prompt_stream(
297
290
  ml_model="string",
298
291
  input_values=[
299
- PromptRequestStringInputRequest(
292
+ PromptRequestStringInput(
300
293
  key="string",
301
294
  value="string",
302
295
  )
303
296
  ],
304
297
  input_variables=[
305
- VellumVariableRequest(
298
+ VellumVariable(
306
299
  id="string",
307
300
  key="string",
308
301
  type="STRING",
309
302
  required=True,
310
- default=StringVellumValueRequest(
311
- value="string",
303
+ default=StringVellumValue(
304
+ value={"key": "value"},
312
305
  ),
313
- extensions=VellumVariableExtensionsRequest(
314
- color="string",
306
+ extensions=VellumVariableExtensions(
307
+ color={"key": "value"},
315
308
  ),
316
309
  )
317
310
  ],
318
- parameters=PromptParametersRequest(
311
+ parameters=PromptParameters(
319
312
  stop=["string"],
320
313
  temperature=1.1,
321
314
  max_tokens=1,
@@ -326,23 +319,18 @@ class AsyncAdHocClient:
326
319
  logit_bias={"string": {"key": "value"}},
327
320
  custom_parameters={"string": {"key": "value"}},
328
321
  ),
329
- settings=PromptSettingsRequest(
322
+ settings=PromptSettings(
330
323
  timeout=1.1,
331
324
  ),
332
325
  blocks=[
333
- JinjaPromptBlockRequest(
334
- properties=JinjaPromptBlockPropertiesRequest(
335
- template="string",
336
- template_type="STRING",
337
- ),
338
- id="string",
326
+ JinjaPromptBlock(
339
327
  state="ENABLED",
340
- cache_config=EphemeralPromptCacheConfigRequest(
341
- type={"key": "value"},
342
- ),
328
+ cache_config=EphemeralPromptCacheConfig(),
329
+ template="string",
330
+ template_type="STRING",
343
331
  )
344
332
  ],
345
- expand_meta=AdHocExpandMetaRequest(
333
+ expand_meta=AdHocExpandMeta(
346
334
  cost=True,
347
335
  model_name=True,
348
336
  usage=True,
@@ -362,22 +350,22 @@ class AsyncAdHocClient:
362
350
  json={
363
351
  "ml_model": ml_model,
364
352
  "input_values": convert_and_respect_annotation_metadata(
365
- object_=input_values, annotation=typing.Sequence[PromptRequestInputRequest], direction="write"
353
+ object_=input_values, annotation=typing.Sequence[PromptRequestInput], direction="write"
366
354
  ),
367
355
  "input_variables": convert_and_respect_annotation_metadata(
368
- object_=input_variables, annotation=typing.Sequence[VellumVariableRequest], direction="write"
356
+ object_=input_variables, annotation=typing.Sequence[VellumVariable], direction="write"
369
357
  ),
370
358
  "parameters": convert_and_respect_annotation_metadata(
371
- object_=parameters, annotation=PromptParametersRequest, direction="write"
359
+ object_=parameters, annotation=PromptParameters, direction="write"
372
360
  ),
373
361
  "settings": convert_and_respect_annotation_metadata(
374
- object_=settings, annotation=PromptSettingsRequest, direction="write"
362
+ object_=settings, annotation=PromptSettings, direction="write"
375
363
  ),
376
364
  "blocks": convert_and_respect_annotation_metadata(
377
- object_=blocks, annotation=typing.Sequence[PromptBlockRequest], direction="write"
365
+ object_=blocks, annotation=typing.Sequence[PromptBlock], direction="write"
378
366
  ),
379
367
  "expand_meta": convert_and_respect_annotation_metadata(
380
- object_=expand_meta, annotation=AdHocExpandMetaRequest, direction="write"
368
+ object_=expand_meta, annotation=AdHocExpandMeta, direction="write"
381
369
  ),
382
370
  },
383
371
  request_options=request_options,