vellum-ai 0.7.1__py3-none-any.whl → 0.7.3__py3-none-any.whl

Sign up to get free protection for your applications and to get access to all the features.
Files changed (30) hide show
  1. vellum/__init__.py +46 -0
  2. vellum/client.py +45 -12
  3. vellum/core/client_wrapper.py +1 -1
  4. vellum/resources/deployments/client.py +11 -0
  5. vellum/types/__init__.py +48 -0
  6. vellum/types/array_vellum_value_item_request.py +82 -0
  7. vellum/types/compile_prompt_deployment_expand_meta_request.py +38 -0
  8. vellum/types/compile_prompt_meta.py +31 -0
  9. vellum/types/deployment_provider_payload_response.py +2 -0
  10. vellum/types/error_vellum_value_request.py +30 -0
  11. vellum/types/function_call_vellum_value_request.py +30 -0
  12. vellum/types/image_vellum_value_request.py +30 -0
  13. vellum/types/json_vellum_value_request.py +29 -0
  14. vellum/types/named_test_case_array_variable_value.py +31 -0
  15. vellum/types/named_test_case_array_variable_value_request.py +31 -0
  16. vellum/types/named_test_case_variable_value.py +12 -0
  17. vellum/types/named_test_case_variable_value_request.py +12 -0
  18. vellum/types/number_vellum_value_request.py +29 -0
  19. vellum/types/prompt_deployment_expand_meta_request_request.py +11 -11
  20. vellum/types/prompt_execution_meta.py +1 -1
  21. vellum/types/prompt_node_execution_meta.py +30 -0
  22. vellum/types/prompt_node_result_data.py +2 -0
  23. vellum/types/string_vellum_value_request.py +29 -0
  24. vellum/types/test_case_array_variable_value.py +32 -0
  25. vellum/types/test_case_variable_value.py +12 -0
  26. vellum/types/workflow_expand_meta_request.py +28 -0
  27. {vellum_ai-0.7.1.dist-info → vellum_ai-0.7.3.dist-info}/METADATA +1 -1
  28. {vellum_ai-0.7.1.dist-info → vellum_ai-0.7.3.dist-info}/RECORD +30 -16
  29. {vellum_ai-0.7.1.dist-info → vellum_ai-0.7.3.dist-info}/LICENSE +0 -0
  30. {vellum_ai-0.7.1.dist-info → vellum_ai-0.7.3.dist-info}/WHEEL +0 -0
vellum/__init__.py CHANGED
@@ -23,6 +23,13 @@ from .types import (
23
23
  ArrayVariableValueItem_Number,
24
24
  ArrayVariableValueItem_String,
25
25
  ArrayVellumValueItem,
26
+ ArrayVellumValueItemRequest,
27
+ ArrayVellumValueItemRequest_Error,
28
+ ArrayVellumValueItemRequest_FunctionCall,
29
+ ArrayVellumValueItemRequest_Image,
30
+ ArrayVellumValueItemRequest_Json,
31
+ ArrayVellumValueItemRequest_Number,
32
+ ArrayVellumValueItemRequest_String,
26
33
  ArrayVellumValueItem_Error,
27
34
  ArrayVellumValueItem_FunctionCall,
28
35
  ArrayVellumValueItem_Image,
@@ -69,6 +76,8 @@ from .types import (
69
76
  CodeExecutionNodeResultOutput_String,
70
77
  CodeExecutionNodeSearchResultsResult,
71
78
  CodeExecutionNodeStringResult,
79
+ CompilePromptDeploymentExpandMetaRequest,
80
+ CompilePromptMeta,
72
81
  ConditionalNodeResult,
73
82
  ConditionalNodeResultData,
74
83
  CreateEnum,
@@ -101,6 +110,7 @@ from .types import (
101
110
  ErrorEnum,
102
111
  ErrorVariableValue,
103
112
  ErrorVellumValue,
113
+ ErrorVellumValueRequest,
104
114
  ExecutePromptApiErrorResponse,
105
115
  ExecutePromptEvent,
106
116
  ExecutePromptEvent_Fulfilled,
@@ -151,6 +161,7 @@ from .types import (
151
161
  FunctionCallRequest,
152
162
  FunctionCallVariableValue,
153
163
  FunctionCallVellumValue,
164
+ FunctionCallVellumValueRequest,
154
165
  GenerateErrorResponse,
155
166
  GenerateOptionsRequest,
156
167
  GenerateRequest,
@@ -169,6 +180,7 @@ from .types import (
169
180
  ImageEnum,
170
181
  ImageVariableValue,
171
182
  ImageVellumValue,
183
+ ImageVellumValueRequest,
172
184
  IndexingConfigVectorizer,
173
185
  IndexingConfigVectorizerRequest,
174
186
  IndexingConfigVectorizerRequest_HkunlpInstructorXl,
@@ -198,6 +210,7 @@ from .types import (
198
210
  JsonInputRequest,
199
211
  JsonVariableValue,
200
212
  JsonVellumValue,
213
+ JsonVellumValueRequest,
201
214
  LogicalOperator,
202
215
  LogprobsEnum,
203
216
  MapEnum,
@@ -217,6 +230,8 @@ from .types import (
217
230
  NamedScenarioInputRequest_ChatHistory,
218
231
  NamedScenarioInputRequest_String,
219
232
  NamedScenarioInputStringVariableValueRequest,
233
+ NamedTestCaseArrayVariableValue,
234
+ NamedTestCaseArrayVariableValueRequest,
220
235
  NamedTestCaseChatHistoryVariableValue,
221
236
  NamedTestCaseChatHistoryVariableValueRequest,
222
237
  NamedTestCaseErrorVariableValue,
@@ -233,6 +248,7 @@ from .types import (
233
248
  NamedTestCaseStringVariableValueRequest,
234
249
  NamedTestCaseVariableValue,
235
250
  NamedTestCaseVariableValueRequest,
251
+ NamedTestCaseVariableValueRequest_Array,
236
252
  NamedTestCaseVariableValueRequest_ChatHistory,
237
253
  NamedTestCaseVariableValueRequest_Error,
238
254
  NamedTestCaseVariableValueRequest_FunctionCall,
@@ -240,6 +256,7 @@ from .types import (
240
256
  NamedTestCaseVariableValueRequest_Number,
241
257
  NamedTestCaseVariableValueRequest_SearchResults,
242
258
  NamedTestCaseVariableValueRequest_String,
259
+ NamedTestCaseVariableValue_Array,
243
260
  NamedTestCaseVariableValue_ChatHistory,
244
261
  NamedTestCaseVariableValue_Error,
245
262
  NamedTestCaseVariableValue_FunctionCall,
@@ -286,6 +303,7 @@ from .types import (
286
303
  NumberEnum,
287
304
  NumberVariableValue,
288
305
  NumberVellumValue,
306
+ NumberVellumValueRequest,
289
307
  OpenAiVectorizerConfig,
290
308
  OpenAiVectorizerConfigRequest,
291
309
  OpenAiVectorizerTextEmbedding3Large,
@@ -311,6 +329,7 @@ from .types import (
311
329
  PromptDeploymentInputRequest_Json,
312
330
  PromptDeploymentInputRequest_String,
313
331
  PromptExecutionMeta,
332
+ PromptNodeExecutionMeta,
314
333
  PromptNodeResult,
315
334
  PromptNodeResultData,
316
335
  PromptOutput,
@@ -379,6 +398,7 @@ from .types import (
379
398
  StringInputRequest,
380
399
  StringVariableValue,
381
400
  StringVellumValue,
401
+ StringVellumValueRequest,
382
402
  SubmitCompletionActualRequest,
383
403
  SubmitCompletionActualsErrorResponse,
384
404
  SubmitWorkflowExecutionActualRequest,
@@ -426,6 +446,7 @@ from .types import (
426
446
  TerminalNodeResultOutput_String,
427
447
  TerminalNodeSearchResultsResult,
428
448
  TerminalNodeStringResult,
449
+ TestCaseArrayVariableValue,
429
450
  TestCaseChatHistoryVariableValue,
430
451
  TestCaseErrorVariableValue,
431
452
  TestCaseFunctionCallVariableValue,
@@ -434,6 +455,7 @@ from .types import (
434
455
  TestCaseSearchResultsVariableValue,
435
456
  TestCaseStringVariableValue,
436
457
  TestCaseVariableValue,
458
+ TestCaseVariableValue_Array,
437
459
  TestCaseVariableValue_ChatHistory,
438
460
  TestCaseVariableValue_Error,
439
461
  TestCaseVariableValue_FunctionCall,
@@ -548,6 +570,7 @@ from .types import (
548
570
  WorkflowExecutionEventType,
549
571
  WorkflowExecutionNodeResultEvent,
550
572
  WorkflowExecutionWorkflowResultEvent,
573
+ WorkflowExpandMetaRequest,
551
574
  WorkflowNodeResultData,
552
575
  WorkflowNodeResultData_Api,
553
576
  WorkflowNodeResultData_CodeExecution,
@@ -659,6 +682,13 @@ __all__ = [
659
682
  "ArrayVariableValueItem_Number",
660
683
  "ArrayVariableValueItem_String",
661
684
  "ArrayVellumValueItem",
685
+ "ArrayVellumValueItemRequest",
686
+ "ArrayVellumValueItemRequest_Error",
687
+ "ArrayVellumValueItemRequest_FunctionCall",
688
+ "ArrayVellumValueItemRequest_Image",
689
+ "ArrayVellumValueItemRequest_Json",
690
+ "ArrayVellumValueItemRequest_Number",
691
+ "ArrayVellumValueItemRequest_String",
662
692
  "ArrayVellumValueItem_Error",
663
693
  "ArrayVellumValueItem_FunctionCall",
664
694
  "ArrayVellumValueItem_Image",
@@ -706,6 +736,8 @@ __all__ = [
706
736
  "CodeExecutionNodeResultOutput_String",
707
737
  "CodeExecutionNodeSearchResultsResult",
708
738
  "CodeExecutionNodeStringResult",
739
+ "CompilePromptDeploymentExpandMetaRequest",
740
+ "CompilePromptMeta",
709
741
  "ConditionalNodeResult",
710
742
  "ConditionalNodeResultData",
711
743
  "CreateEnum",
@@ -740,6 +772,7 @@ __all__ = [
740
772
  "ErrorEnum",
741
773
  "ErrorVariableValue",
742
774
  "ErrorVellumValue",
775
+ "ErrorVellumValueRequest",
743
776
  "ExecutePromptApiErrorResponse",
744
777
  "ExecutePromptEvent",
745
778
  "ExecutePromptEvent_Fulfilled",
@@ -791,6 +824,7 @@ __all__ = [
791
824
  "FunctionCallRequest",
792
825
  "FunctionCallVariableValue",
793
826
  "FunctionCallVellumValue",
827
+ "FunctionCallVellumValueRequest",
794
828
  "GenerateErrorResponse",
795
829
  "GenerateOptionsRequest",
796
830
  "GenerateRequest",
@@ -809,6 +843,7 @@ __all__ = [
809
843
  "ImageEnum",
810
844
  "ImageVariableValue",
811
845
  "ImageVellumValue",
846
+ "ImageVellumValueRequest",
812
847
  "IndexingConfigVectorizer",
813
848
  "IndexingConfigVectorizerRequest",
814
849
  "IndexingConfigVectorizerRequest_HkunlpInstructorXl",
@@ -839,6 +874,7 @@ __all__ = [
839
874
  "JsonInputRequest",
840
875
  "JsonVariableValue",
841
876
  "JsonVellumValue",
877
+ "JsonVellumValueRequest",
842
878
  "LogicalOperator",
843
879
  "LogprobsEnum",
844
880
  "MapEnum",
@@ -858,6 +894,8 @@ __all__ = [
858
894
  "NamedScenarioInputRequest_ChatHistory",
859
895
  "NamedScenarioInputRequest_String",
860
896
  "NamedScenarioInputStringVariableValueRequest",
897
+ "NamedTestCaseArrayVariableValue",
898
+ "NamedTestCaseArrayVariableValueRequest",
861
899
  "NamedTestCaseChatHistoryVariableValue",
862
900
  "NamedTestCaseChatHistoryVariableValueRequest",
863
901
  "NamedTestCaseErrorVariableValue",
@@ -874,6 +912,7 @@ __all__ = [
874
912
  "NamedTestCaseStringVariableValueRequest",
875
913
  "NamedTestCaseVariableValue",
876
914
  "NamedTestCaseVariableValueRequest",
915
+ "NamedTestCaseVariableValueRequest_Array",
877
916
  "NamedTestCaseVariableValueRequest_ChatHistory",
878
917
  "NamedTestCaseVariableValueRequest_Error",
879
918
  "NamedTestCaseVariableValueRequest_FunctionCall",
@@ -881,6 +920,7 @@ __all__ = [
881
920
  "NamedTestCaseVariableValueRequest_Number",
882
921
  "NamedTestCaseVariableValueRequest_SearchResults",
883
922
  "NamedTestCaseVariableValueRequest_String",
923
+ "NamedTestCaseVariableValue_Array",
884
924
  "NamedTestCaseVariableValue_ChatHistory",
885
925
  "NamedTestCaseVariableValue_Error",
886
926
  "NamedTestCaseVariableValue_FunctionCall",
@@ -928,6 +968,7 @@ __all__ = [
928
968
  "NumberEnum",
929
969
  "NumberVariableValue",
930
970
  "NumberVellumValue",
971
+ "NumberVellumValueRequest",
931
972
  "OpenAiVectorizerConfig",
932
973
  "OpenAiVectorizerConfigRequest",
933
974
  "OpenAiVectorizerTextEmbedding3Large",
@@ -953,6 +994,7 @@ __all__ = [
953
994
  "PromptDeploymentInputRequest_Json",
954
995
  "PromptDeploymentInputRequest_String",
955
996
  "PromptExecutionMeta",
997
+ "PromptNodeExecutionMeta",
956
998
  "PromptNodeResult",
957
999
  "PromptNodeResultData",
958
1000
  "PromptOutput",
@@ -1021,6 +1063,7 @@ __all__ = [
1021
1063
  "StringInputRequest",
1022
1064
  "StringVariableValue",
1023
1065
  "StringVellumValue",
1066
+ "StringVellumValueRequest",
1024
1067
  "SubmitCompletionActualRequest",
1025
1068
  "SubmitCompletionActualsErrorResponse",
1026
1069
  "SubmitWorkflowExecutionActualRequest",
@@ -1068,6 +1111,7 @@ __all__ = [
1068
1111
  "TerminalNodeResultOutput_String",
1069
1112
  "TerminalNodeSearchResultsResult",
1070
1113
  "TerminalNodeStringResult",
1114
+ "TestCaseArrayVariableValue",
1071
1115
  "TestCaseChatHistoryVariableValue",
1072
1116
  "TestCaseErrorVariableValue",
1073
1117
  "TestCaseFunctionCallVariableValue",
@@ -1076,6 +1120,7 @@ __all__ = [
1076
1120
  "TestCaseSearchResultsVariableValue",
1077
1121
  "TestCaseStringVariableValue",
1078
1122
  "TestCaseVariableValue",
1123
+ "TestCaseVariableValue_Array",
1079
1124
  "TestCaseVariableValue_ChatHistory",
1080
1125
  "TestCaseVariableValue_Error",
1081
1126
  "TestCaseVariableValue_FunctionCall",
@@ -1192,6 +1237,7 @@ __all__ = [
1192
1237
  "WorkflowExecutionEventType",
1193
1238
  "WorkflowExecutionNodeResultEvent",
1194
1239
  "WorkflowExecutionWorkflowResultEvent",
1240
+ "WorkflowExpandMetaRequest",
1195
1241
  "WorkflowNodeResultData",
1196
1242
  "WorkflowNodeResultData_Api",
1197
1243
  "WorkflowNodeResultData_CodeExecution",
vellum/client.py CHANGED
@@ -42,6 +42,7 @@ from .types.search_response import SearchResponse
42
42
  from .types.submit_completion_actual_request import SubmitCompletionActualRequest
43
43
  from .types.submit_workflow_execution_actual_request import SubmitWorkflowExecutionActualRequest
44
44
  from .types.workflow_execution_event_type import WorkflowExecutionEventType
45
+ from .types.workflow_expand_meta_request import WorkflowExpandMetaRequest
45
46
  from .types.workflow_request_input_request import WorkflowRequestInputRequest
46
47
  from .types.workflow_stream_event import WorkflowStreamEvent
47
48
 
@@ -164,11 +165,11 @@ class Vellum:
164
165
  external_id="string",
165
166
  expand_meta=PromptDeploymentExpandMetaRequestRequest(
166
167
  model_name=True,
168
+ usage=True,
169
+ finish_reason=True,
167
170
  latency=True,
168
171
  deployment_release_tag=True,
169
172
  prompt_version_id=True,
170
- finish_reason=True,
171
- usage=True,
172
173
  ),
173
174
  raw_overrides=RawPromptExecutionOverridesRequest(
174
175
  body={"string": {"key": "value"}},
@@ -299,11 +300,11 @@ class Vellum:
299
300
  external_id="string",
300
301
  expand_meta=PromptDeploymentExpandMetaRequestRequest(
301
302
  model_name=True,
303
+ usage=True,
304
+ finish_reason=True,
302
305
  latency=True,
303
306
  deployment_release_tag=True,
304
307
  prompt_version_id=True,
305
- finish_reason=True,
306
- usage=True,
307
308
  ),
308
309
  raw_overrides=RawPromptExecutionOverridesRequest(
309
310
  body={"string": {"key": "value"}},
@@ -382,6 +383,7 @@ class Vellum:
382
383
  self,
383
384
  *,
384
385
  inputs: typing.Sequence[WorkflowRequestInputRequest],
386
+ expand_meta: typing.Optional[WorkflowExpandMetaRequest] = OMIT,
385
387
  workflow_deployment_id: typing.Optional[str] = OMIT,
386
388
  workflow_deployment_name: typing.Optional[str] = OMIT,
387
389
  release_tag: typing.Optional[str] = OMIT,
@@ -394,6 +396,8 @@ class Vellum:
394
396
  Parameters:
395
397
  - inputs: typing.Sequence[WorkflowRequestInputRequest]. The list of inputs defined in the Workflow's Deployment with their corresponding values.
396
398
 
399
+ - expand_meta: typing.Optional[WorkflowExpandMetaRequest]. An optionally specified configuration used to opt in to including additional metadata about this workflow execution in the API response. Corresponding values will be returned under the `execution_meta` key within NODE events in the response stream.
400
+
397
401
  - workflow_deployment_id: typing.Optional[str]. The ID of the Workflow Deployment. Must provide either this or workflow_deployment_name.
398
402
 
399
403
  - workflow_deployment_name: typing.Optional[str]. The name of the Workflow Deployment. Must provide either this or workflow_deployment_id.
@@ -404,7 +408,7 @@ class Vellum:
404
408
 
405
409
  - request_options: typing.Optional[RequestOptions]. Request-specific configuration.
406
410
  ---
407
- from vellum import WorkflowRequestInputRequest_String
411
+ from vellum import WorkflowExpandMetaRequest, WorkflowRequestInputRequest_String
408
412
  from vellum.client import Vellum
409
413
 
410
414
  client = Vellum(
@@ -417,6 +421,9 @@ class Vellum:
417
421
  value="string",
418
422
  )
419
423
  ],
424
+ expand_meta=WorkflowExpandMetaRequest(
425
+ usage=True,
426
+ ),
420
427
  workflow_deployment_id="string",
421
428
  workflow_deployment_name="string",
422
429
  release_tag="string",
@@ -424,6 +431,8 @@ class Vellum:
424
431
  )
425
432
  """
426
433
  _request: typing.Dict[str, typing.Any] = {"inputs": inputs}
434
+ if expand_meta is not OMIT:
435
+ _request["expand_meta"] = expand_meta
427
436
  if workflow_deployment_id is not OMIT:
428
437
  _request["workflow_deployment_id"] = workflow_deployment_id
429
438
  if workflow_deployment_name is not OMIT:
@@ -476,6 +485,7 @@ class Vellum:
476
485
  self,
477
486
  *,
478
487
  inputs: typing.Sequence[WorkflowRequestInputRequest],
488
+ expand_meta: typing.Optional[WorkflowExpandMetaRequest] = OMIT,
479
489
  workflow_deployment_id: typing.Optional[str] = OMIT,
480
490
  workflow_deployment_name: typing.Optional[str] = OMIT,
481
491
  release_tag: typing.Optional[str] = OMIT,
@@ -489,6 +499,8 @@ class Vellum:
489
499
  Parameters:
490
500
  - inputs: typing.Sequence[WorkflowRequestInputRequest]. The list of inputs defined in the Workflow's Deployment with their corresponding values.
491
501
 
502
+ - expand_meta: typing.Optional[WorkflowExpandMetaRequest]. An optionally specified configuration used to opt in to including additional metadata about this workflow execution in the API response. Corresponding values will be returned under the `execution_meta` key within NODE events in the response stream.
503
+
492
504
  - workflow_deployment_id: typing.Optional[str]. The ID of the Workflow Deployment. Must provide either this or workflow_deployment_name.
493
505
 
494
506
  - workflow_deployment_name: typing.Optional[str]. The name of the Workflow Deployment. Must provide either this or workflow_deployment_id.
@@ -501,7 +513,7 @@ class Vellum:
501
513
 
502
514
  - request_options: typing.Optional[RequestOptions]. Request-specific configuration.
503
515
  ---
504
- from vellum import WorkflowRequestInputRequest_String
516
+ from vellum import WorkflowExpandMetaRequest, WorkflowRequestInputRequest_String
505
517
  from vellum.client import Vellum
506
518
 
507
519
  client = Vellum(
@@ -514,6 +526,9 @@ class Vellum:
514
526
  value="string",
515
527
  )
516
528
  ],
529
+ expand_meta=WorkflowExpandMetaRequest(
530
+ usage=True,
531
+ ),
517
532
  workflow_deployment_id="string",
518
533
  workflow_deployment_name="string",
519
534
  release_tag="string",
@@ -522,6 +537,8 @@ class Vellum:
522
537
  )
523
538
  """
524
539
  _request: typing.Dict[str, typing.Any] = {"inputs": inputs}
540
+ if expand_meta is not OMIT:
541
+ _request["expand_meta"] = expand_meta
525
542
  if workflow_deployment_id is not OMIT:
526
543
  _request["workflow_deployment_id"] = workflow_deployment_id
527
544
  if workflow_deployment_name is not OMIT:
@@ -1125,11 +1142,11 @@ class AsyncVellum:
1125
1142
  external_id="string",
1126
1143
  expand_meta=PromptDeploymentExpandMetaRequestRequest(
1127
1144
  model_name=True,
1145
+ usage=True,
1146
+ finish_reason=True,
1128
1147
  latency=True,
1129
1148
  deployment_release_tag=True,
1130
1149
  prompt_version_id=True,
1131
- finish_reason=True,
1132
- usage=True,
1133
1150
  ),
1134
1151
  raw_overrides=RawPromptExecutionOverridesRequest(
1135
1152
  body={"string": {"key": "value"}},
@@ -1260,11 +1277,11 @@ class AsyncVellum:
1260
1277
  external_id="string",
1261
1278
  expand_meta=PromptDeploymentExpandMetaRequestRequest(
1262
1279
  model_name=True,
1280
+ usage=True,
1281
+ finish_reason=True,
1263
1282
  latency=True,
1264
1283
  deployment_release_tag=True,
1265
1284
  prompt_version_id=True,
1266
- finish_reason=True,
1267
- usage=True,
1268
1285
  ),
1269
1286
  raw_overrides=RawPromptExecutionOverridesRequest(
1270
1287
  body={"string": {"key": "value"}},
@@ -1343,6 +1360,7 @@ class AsyncVellum:
1343
1360
  self,
1344
1361
  *,
1345
1362
  inputs: typing.Sequence[WorkflowRequestInputRequest],
1363
+ expand_meta: typing.Optional[WorkflowExpandMetaRequest] = OMIT,
1346
1364
  workflow_deployment_id: typing.Optional[str] = OMIT,
1347
1365
  workflow_deployment_name: typing.Optional[str] = OMIT,
1348
1366
  release_tag: typing.Optional[str] = OMIT,
@@ -1355,6 +1373,8 @@ class AsyncVellum:
1355
1373
  Parameters:
1356
1374
  - inputs: typing.Sequence[WorkflowRequestInputRequest]. The list of inputs defined in the Workflow's Deployment with their corresponding values.
1357
1375
 
1376
+ - expand_meta: typing.Optional[WorkflowExpandMetaRequest]. An optionally specified configuration used to opt in to including additional metadata about this workflow execution in the API response. Corresponding values will be returned under the `execution_meta` key within NODE events in the response stream.
1377
+
1358
1378
  - workflow_deployment_id: typing.Optional[str]. The ID of the Workflow Deployment. Must provide either this or workflow_deployment_name.
1359
1379
 
1360
1380
  - workflow_deployment_name: typing.Optional[str]. The name of the Workflow Deployment. Must provide either this or workflow_deployment_id.
@@ -1365,7 +1385,7 @@ class AsyncVellum:
1365
1385
 
1366
1386
  - request_options: typing.Optional[RequestOptions]. Request-specific configuration.
1367
1387
  ---
1368
- from vellum import WorkflowRequestInputRequest_String
1388
+ from vellum import WorkflowExpandMetaRequest, WorkflowRequestInputRequest_String
1369
1389
  from vellum.client import AsyncVellum
1370
1390
 
1371
1391
  client = AsyncVellum(
@@ -1378,6 +1398,9 @@ class AsyncVellum:
1378
1398
  value="string",
1379
1399
  )
1380
1400
  ],
1401
+ expand_meta=WorkflowExpandMetaRequest(
1402
+ usage=True,
1403
+ ),
1381
1404
  workflow_deployment_id="string",
1382
1405
  workflow_deployment_name="string",
1383
1406
  release_tag="string",
@@ -1385,6 +1408,8 @@ class AsyncVellum:
1385
1408
  )
1386
1409
  """
1387
1410
  _request: typing.Dict[str, typing.Any] = {"inputs": inputs}
1411
+ if expand_meta is not OMIT:
1412
+ _request["expand_meta"] = expand_meta
1388
1413
  if workflow_deployment_id is not OMIT:
1389
1414
  _request["workflow_deployment_id"] = workflow_deployment_id
1390
1415
  if workflow_deployment_name is not OMIT:
@@ -1437,6 +1462,7 @@ class AsyncVellum:
1437
1462
  self,
1438
1463
  *,
1439
1464
  inputs: typing.Sequence[WorkflowRequestInputRequest],
1465
+ expand_meta: typing.Optional[WorkflowExpandMetaRequest] = OMIT,
1440
1466
  workflow_deployment_id: typing.Optional[str] = OMIT,
1441
1467
  workflow_deployment_name: typing.Optional[str] = OMIT,
1442
1468
  release_tag: typing.Optional[str] = OMIT,
@@ -1450,6 +1476,8 @@ class AsyncVellum:
1450
1476
  Parameters:
1451
1477
  - inputs: typing.Sequence[WorkflowRequestInputRequest]. The list of inputs defined in the Workflow's Deployment with their corresponding values.
1452
1478
 
1479
+ - expand_meta: typing.Optional[WorkflowExpandMetaRequest]. An optionally specified configuration used to opt in to including additional metadata about this workflow execution in the API response. Corresponding values will be returned under the `execution_meta` key within NODE events in the response stream.
1480
+
1453
1481
  - workflow_deployment_id: typing.Optional[str]. The ID of the Workflow Deployment. Must provide either this or workflow_deployment_name.
1454
1482
 
1455
1483
  - workflow_deployment_name: typing.Optional[str]. The name of the Workflow Deployment. Must provide either this or workflow_deployment_id.
@@ -1462,7 +1490,7 @@ class AsyncVellum:
1462
1490
 
1463
1491
  - request_options: typing.Optional[RequestOptions]. Request-specific configuration.
1464
1492
  ---
1465
- from vellum import WorkflowRequestInputRequest_String
1493
+ from vellum import WorkflowExpandMetaRequest, WorkflowRequestInputRequest_String
1466
1494
  from vellum.client import AsyncVellum
1467
1495
 
1468
1496
  client = AsyncVellum(
@@ -1475,6 +1503,9 @@ class AsyncVellum:
1475
1503
  value="string",
1476
1504
  )
1477
1505
  ],
1506
+ expand_meta=WorkflowExpandMetaRequest(
1507
+ usage=True,
1508
+ ),
1478
1509
  workflow_deployment_id="string",
1479
1510
  workflow_deployment_name="string",
1480
1511
  release_tag="string",
@@ -1483,6 +1514,8 @@ class AsyncVellum:
1483
1514
  )
1484
1515
  """
1485
1516
  _request: typing.Dict[str, typing.Any] = {"inputs": inputs}
1517
+ if expand_meta is not OMIT:
1518
+ _request["expand_meta"] = expand_meta
1486
1519
  if workflow_deployment_id is not OMIT:
1487
1520
  _request["workflow_deployment_id"] = workflow_deployment_id
1488
1521
  if workflow_deployment_name is not OMIT:
@@ -18,7 +18,7 @@ class BaseClientWrapper:
18
18
  headers: typing.Dict[str, str] = {
19
19
  "X-Fern-Language": "Python",
20
20
  "X-Fern-SDK-Name": "vellum-ai",
21
- "X-Fern-SDK-Version": "0.7.1",
21
+ "X-Fern-SDK-Version": "0.7.3",
22
22
  }
23
23
  headers["X_API_KEY"] = self.api_key
24
24
  return headers
@@ -14,6 +14,7 @@ from ...errors.bad_request_error import BadRequestError
14
14
  from ...errors.forbidden_error import ForbiddenError
15
15
  from ...errors.internal_server_error import InternalServerError
16
16
  from ...errors.not_found_error import NotFoundError
17
+ from ...types.compile_prompt_deployment_expand_meta_request import CompilePromptDeploymentExpandMetaRequest
17
18
  from ...types.deployment_provider_payload_response import DeploymentProviderPayloadResponse
18
19
  from ...types.deployment_read import DeploymentRead
19
20
  from ...types.deployment_release_tag_read import DeploymentReleaseTagRead
@@ -276,6 +277,7 @@ class DeploymentsClient:
276
277
  deployment_name: typing.Optional[str] = OMIT,
277
278
  inputs: typing.Sequence[PromptDeploymentInputRequest],
278
279
  release_tag: typing.Optional[str] = OMIT,
280
+ expand_meta: typing.Optional[CompilePromptDeploymentExpandMetaRequest] = OMIT,
279
281
  request_options: typing.Optional[RequestOptions] = None,
280
282
  ) -> DeploymentProviderPayloadResponse:
281
283
  """
@@ -288,6 +290,8 @@ class DeploymentsClient:
288
290
 
289
291
  - release_tag: typing.Optional[str]. Optionally specify a release tag if you want to pin to a specific release of the Workflow Deployment
290
292
 
293
+ - expand_meta: typing.Optional[CompilePromptDeploymentExpandMetaRequest].
294
+
291
295
  - request_options: typing.Optional[RequestOptions]. Request-specific configuration.
292
296
  ---
293
297
  from vellum.client import Vellum
@@ -306,6 +310,8 @@ class DeploymentsClient:
306
310
  _request["deployment_name"] = deployment_name
307
311
  if release_tag is not OMIT:
308
312
  _request["release_tag"] = release_tag
313
+ if expand_meta is not OMIT:
314
+ _request["expand_meta"] = expand_meta
309
315
  _response = self._client_wrapper.httpx_client.request(
310
316
  method="POST",
311
317
  url=urllib.parse.urljoin(
@@ -602,6 +608,7 @@ class AsyncDeploymentsClient:
602
608
  deployment_name: typing.Optional[str] = OMIT,
603
609
  inputs: typing.Sequence[PromptDeploymentInputRequest],
604
610
  release_tag: typing.Optional[str] = OMIT,
611
+ expand_meta: typing.Optional[CompilePromptDeploymentExpandMetaRequest] = OMIT,
605
612
  request_options: typing.Optional[RequestOptions] = None,
606
613
  ) -> DeploymentProviderPayloadResponse:
607
614
  """
@@ -614,6 +621,8 @@ class AsyncDeploymentsClient:
614
621
 
615
622
  - release_tag: typing.Optional[str]. Optionally specify a release tag if you want to pin to a specific release of the Workflow Deployment
616
623
 
624
+ - expand_meta: typing.Optional[CompilePromptDeploymentExpandMetaRequest].
625
+
617
626
  - request_options: typing.Optional[RequestOptions]. Request-specific configuration.
618
627
  ---
619
628
  from vellum.client import AsyncVellum
@@ -632,6 +641,8 @@ class AsyncDeploymentsClient:
632
641
  _request["deployment_name"] = deployment_name
633
642
  if release_tag is not OMIT:
634
643
  _request["release_tag"] = release_tag
644
+ if expand_meta is not OMIT:
645
+ _request["expand_meta"] = expand_meta
635
646
  _response = await self._client_wrapper.httpx_client.request(
636
647
  method="POST",
637
648
  url=urllib.parse.urljoin(