vellum-ai 0.7.6__py3-none-any.whl → 0.7.8__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
vellum/__init__.py CHANGED
@@ -236,8 +236,10 @@ from .types import (
236
236
  MlModelTokenizerConfig_Tiktoken,
237
237
  MlModelUsage,
238
238
  NamedScenarioInputChatHistoryVariableValueRequest,
239
+ NamedScenarioInputJsonVariableValueRequest,
239
240
  NamedScenarioInputRequest,
240
241
  NamedScenarioInputRequest_ChatHistory,
242
+ NamedScenarioInputRequest_Json,
241
243
  NamedScenarioInputRequest_String,
242
244
  NamedScenarioInputStringVariableValueRequest,
243
245
  NamedTestCaseArrayVariableValue,
@@ -395,8 +397,10 @@ from .types import (
395
397
  SandboxScenario,
396
398
  ScenarioInput,
397
399
  ScenarioInputChatHistoryVariableValue,
400
+ ScenarioInputJsonVariableValue,
398
401
  ScenarioInputStringVariableValue,
399
402
  ScenarioInput_ChatHistory,
403
+ ScenarioInput_Json,
400
404
  ScenarioInput_String,
401
405
  SearchFiltersRequest,
402
406
  SearchNodeResult,
@@ -922,8 +926,10 @@ __all__ = [
922
926
  "MlModelTokenizerConfig_Tiktoken",
923
927
  "MlModelUsage",
924
928
  "NamedScenarioInputChatHistoryVariableValueRequest",
929
+ "NamedScenarioInputJsonVariableValueRequest",
925
930
  "NamedScenarioInputRequest",
926
931
  "NamedScenarioInputRequest_ChatHistory",
932
+ "NamedScenarioInputRequest_Json",
927
933
  "NamedScenarioInputRequest_String",
928
934
  "NamedScenarioInputStringVariableValueRequest",
929
935
  "NamedTestCaseArrayVariableValue",
@@ -1082,8 +1088,10 @@ __all__ = [
1082
1088
  "SandboxScenario",
1083
1089
  "ScenarioInput",
1084
1090
  "ScenarioInputChatHistoryVariableValue",
1091
+ "ScenarioInputJsonVariableValue",
1085
1092
  "ScenarioInputStringVariableValue",
1086
1093
  "ScenarioInput_ChatHistory",
1094
+ "ScenarioInput_Json",
1087
1095
  "ScenarioInput_String",
1088
1096
  "SearchFiltersRequest",
1089
1097
  "SearchNodeResult",
@@ -18,7 +18,7 @@ class BaseClientWrapper:
18
18
  headers: typing.Dict[str, str] = {
19
19
  "X-Fern-Language": "Python",
20
20
  "X-Fern-SDK-Name": "vellum-ai",
21
- "X-Fern-SDK-Version": "0.7.6",
21
+ "X-Fern-SDK-Version": "0.7.8",
22
22
  }
23
23
  headers["X_API_KEY"] = self.api_key
24
24
  return headers
@@ -85,9 +85,9 @@ class MlModelsClient:
85
85
  *,
86
86
  name: str,
87
87
  family: MlModelFamily,
88
+ hosted_by: HostedByEnum,
89
+ developed_by: MlModelDeveloper,
88
90
  exec_config: MlModelExecConfigRequest,
89
- hosted_by: typing.Optional[HostedByEnum] = OMIT,
90
- developed_by: typing.Optional[MlModelDeveloper] = OMIT,
91
91
  parameter_config: typing.Optional[MlModelParameterConfigRequest] = OMIT,
92
92
  display_config: typing.Optional[MlModelDisplayConfigRequest] = OMIT,
93
93
  visibility: typing.Optional[VisibilityEnum] = OMIT,
@@ -125,10 +125,7 @@ class MlModelsClient:
125
125
  * `YI` - Yi
126
126
  * `ZEPHYR` - Zephyr
127
127
 
128
- exec_config : MlModelExecConfigRequest
129
- Configuration for how to execute the ML Model.
130
-
131
- hosted_by : typing.Optional[HostedByEnum]
128
+ hosted_by : HostedByEnum
132
129
  The organization hosting the ML Model.
133
130
 
134
131
  * `ANTHROPIC` - ANTHROPIC
@@ -149,7 +146,7 @@ class MlModelsClient:
149
146
  * `PYQ` - PYQ
150
147
  * `REPLICATE` - REPLICATE
151
148
 
152
- developed_by : typing.Optional[MlModelDeveloper]
149
+ developed_by : MlModelDeveloper
153
150
  The organization that developed the ML Model.
154
151
 
155
152
  * `01_AI` - 01_AI
@@ -171,6 +168,9 @@ class MlModelsClient:
171
168
  * `TII` - TII
172
169
  * `WIZARDLM` - WIZARDLM
173
170
 
171
+ exec_config : MlModelExecConfigRequest
172
+ Configuration for how to execute the ML Model.
173
+
174
174
  parameter_config : typing.Optional[MlModelParameterConfigRequest]
175
175
  Configuration for the ML Model's parameters.
176
176
 
@@ -204,6 +204,8 @@ class MlModelsClient:
204
204
  client.ml_models.create(
205
205
  name="name",
206
206
  family="CAPYBARA",
207
+ hosted_by="ANTHROPIC",
208
+ developed_by="01_AI",
207
209
  exec_config=MlModelExecConfigRequest(
208
210
  model_identifier="model_identifier",
209
211
  base_url="base_url",
@@ -244,7 +246,7 @@ class MlModelsClient:
244
246
  Parameters
245
247
  ----------
246
248
  id : str
247
- A UUID string identifying this ml model.
249
+ Either the ML Model's ID or its unique name
248
250
 
249
251
  request_options : typing.Optional[RequestOptions]
250
252
  Request-specific configuration.
@@ -293,7 +295,7 @@ class MlModelsClient:
293
295
  Parameters
294
296
  ----------
295
297
  id : str
296
- A UUID string identifying this ml model.
298
+ Either the ML Model's ID or its unique name
297
299
 
298
300
  display_config : typing.Optional[MlModelDisplayConfigRequest]
299
301
  Configuration for how to display the ML Model.
@@ -355,7 +357,7 @@ class MlModelsClient:
355
357
  Parameters
356
358
  ----------
357
359
  id : str
358
- A UUID string identifying this ml model.
360
+ Either the ML Model's ID or its unique name
359
361
 
360
362
  display_config : typing.Optional[MlModelDisplayConfigRequest]
361
363
  Configuration for how to display the ML Model.
@@ -475,9 +477,9 @@ class AsyncMlModelsClient:
475
477
  *,
476
478
  name: str,
477
479
  family: MlModelFamily,
480
+ hosted_by: HostedByEnum,
481
+ developed_by: MlModelDeveloper,
478
482
  exec_config: MlModelExecConfigRequest,
479
- hosted_by: typing.Optional[HostedByEnum] = OMIT,
480
- developed_by: typing.Optional[MlModelDeveloper] = OMIT,
481
483
  parameter_config: typing.Optional[MlModelParameterConfigRequest] = OMIT,
482
484
  display_config: typing.Optional[MlModelDisplayConfigRequest] = OMIT,
483
485
  visibility: typing.Optional[VisibilityEnum] = OMIT,
@@ -515,10 +517,7 @@ class AsyncMlModelsClient:
515
517
  * `YI` - Yi
516
518
  * `ZEPHYR` - Zephyr
517
519
 
518
- exec_config : MlModelExecConfigRequest
519
- Configuration for how to execute the ML Model.
520
-
521
- hosted_by : typing.Optional[HostedByEnum]
520
+ hosted_by : HostedByEnum
522
521
  The organization hosting the ML Model.
523
522
 
524
523
  * `ANTHROPIC` - ANTHROPIC
@@ -539,7 +538,7 @@ class AsyncMlModelsClient:
539
538
  * `PYQ` - PYQ
540
539
  * `REPLICATE` - REPLICATE
541
540
 
542
- developed_by : typing.Optional[MlModelDeveloper]
541
+ developed_by : MlModelDeveloper
543
542
  The organization that developed the ML Model.
544
543
 
545
544
  * `01_AI` - 01_AI
@@ -561,6 +560,9 @@ class AsyncMlModelsClient:
561
560
  * `TII` - TII
562
561
  * `WIZARDLM` - WIZARDLM
563
562
 
563
+ exec_config : MlModelExecConfigRequest
564
+ Configuration for how to execute the ML Model.
565
+
564
566
  parameter_config : typing.Optional[MlModelParameterConfigRequest]
565
567
  Configuration for the ML Model's parameters.
566
568
 
@@ -599,6 +601,8 @@ class AsyncMlModelsClient:
599
601
  await client.ml_models.create(
600
602
  name="name",
601
603
  family="CAPYBARA",
604
+ hosted_by="ANTHROPIC",
605
+ developed_by="01_AI",
602
606
  exec_config=MlModelExecConfigRequest(
603
607
  model_identifier="model_identifier",
604
608
  base_url="base_url",
@@ -642,7 +646,7 @@ class AsyncMlModelsClient:
642
646
  Parameters
643
647
  ----------
644
648
  id : str
645
- A UUID string identifying this ml model.
649
+ Either the ML Model's ID or its unique name
646
650
 
647
651
  request_options : typing.Optional[RequestOptions]
648
652
  Request-specific configuration.
@@ -699,7 +703,7 @@ class AsyncMlModelsClient:
699
703
  Parameters
700
704
  ----------
701
705
  id : str
702
- A UUID string identifying this ml model.
706
+ Either the ML Model's ID or its unique name
703
707
 
704
708
  display_config : typing.Optional[MlModelDisplayConfigRequest]
705
709
  Configuration for how to display the ML Model.
@@ -769,7 +773,7 @@ class AsyncMlModelsClient:
769
773
  Parameters
770
774
  ----------
771
775
  id : str
772
- A UUID string identifying this ml model.
776
+ Either the ML Model's ID or its unique name
773
777
 
774
778
  display_config : typing.Optional[MlModelDisplayConfigRequest]
775
779
  Configuration for how to display the ML Model.
vellum/types/__init__.py CHANGED
@@ -279,9 +279,11 @@ from .ml_model_tokenizer_config_request import (
279
279
  )
280
280
  from .ml_model_usage import MlModelUsage
281
281
  from .named_scenario_input_chat_history_variable_value_request import NamedScenarioInputChatHistoryVariableValueRequest
282
+ from .named_scenario_input_json_variable_value_request import NamedScenarioInputJsonVariableValueRequest
282
283
  from .named_scenario_input_request import (
283
284
  NamedScenarioInputRequest,
284
285
  NamedScenarioInputRequest_ChatHistory,
286
+ NamedScenarioInputRequest_Json,
285
287
  NamedScenarioInputRequest_String,
286
288
  )
287
289
  from .named_scenario_input_string_variable_value_request import NamedScenarioInputStringVariableValueRequest
@@ -454,8 +456,9 @@ from .rejected_workflow_node_result_event import RejectedWorkflowNodeResultEvent
454
456
  from .release_tag_source import ReleaseTagSource
455
457
  from .replace_test_suite_test_case_request import ReplaceTestSuiteTestCaseRequest
456
458
  from .sandbox_scenario import SandboxScenario
457
- from .scenario_input import ScenarioInput, ScenarioInput_ChatHistory, ScenarioInput_String
459
+ from .scenario_input import ScenarioInput, ScenarioInput_ChatHistory, ScenarioInput_Json, ScenarioInput_String
458
460
  from .scenario_input_chat_history_variable_value import ScenarioInputChatHistoryVariableValue
461
+ from .scenario_input_json_variable_value import ScenarioInputJsonVariableValue
459
462
  from .scenario_input_string_variable_value import ScenarioInputStringVariableValue
460
463
  from .search_filters_request import SearchFiltersRequest
461
464
  from .search_node_result import SearchNodeResult
@@ -987,8 +990,10 @@ __all__ = [
987
990
  "MlModelTokenizerConfig_Tiktoken",
988
991
  "MlModelUsage",
989
992
  "NamedScenarioInputChatHistoryVariableValueRequest",
993
+ "NamedScenarioInputJsonVariableValueRequest",
990
994
  "NamedScenarioInputRequest",
991
995
  "NamedScenarioInputRequest_ChatHistory",
996
+ "NamedScenarioInputRequest_Json",
992
997
  "NamedScenarioInputRequest_String",
993
998
  "NamedScenarioInputStringVariableValueRequest",
994
999
  "NamedTestCaseArrayVariableValue",
@@ -1146,8 +1151,10 @@ __all__ = [
1146
1151
  "SandboxScenario",
1147
1152
  "ScenarioInput",
1148
1153
  "ScenarioInputChatHistoryVariableValue",
1154
+ "ScenarioInputJsonVariableValue",
1149
1155
  "ScenarioInputStringVariableValue",
1150
1156
  "ScenarioInput_ChatHistory",
1157
+ "ScenarioInput_Json",
1151
1158
  "ScenarioInput_String",
1152
1159
  "SearchFiltersRequest",
1153
1160
  "SearchNodeResult",
@@ -16,7 +16,6 @@ class MlModelExecConfig(pydantic_v1.BaseModel):
16
16
  base_url: str
17
17
  metadata: typing.Dict[str, typing.Any]
18
18
  features: typing.List[MlModelFeature]
19
- force_system_credentials: typing.Optional[bool] = None
20
19
  tokenizer_config: typing.Optional[MlModelTokenizerConfig] = None
21
20
  request_config: typing.Optional[MlModelRequestConfig] = None
22
21
  response_config: typing.Optional[MlModelResponseConfig] = None
@@ -16,7 +16,6 @@ class MlModelExecConfigRequest(pydantic_v1.BaseModel):
16
16
  base_url: str
17
17
  metadata: typing.Dict[str, typing.Any]
18
18
  features: typing.List[MlModelFeature]
19
- force_system_credentials: typing.Optional[bool] = None
20
19
  tokenizer_config: typing.Optional[MlModelTokenizerConfigRequest] = None
21
20
  request_config: typing.Optional[MlModelRequestConfigRequest] = None
22
21
  response_config: typing.Optional[MlModelResponseConfigRequest] = None
@@ -0,0 +1,34 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import datetime as dt
4
+ import typing
5
+
6
+ from ..core.datetime_utils import serialize_datetime
7
+ from ..core.pydantic_utilities import deep_union_pydantic_dicts, pydantic_v1
8
+
9
+
10
+ class NamedScenarioInputJsonVariableValueRequest(pydantic_v1.BaseModel):
11
+ """
12
+ Named Prompt Sandbox Scenario input value that is of type JSON
13
+ """
14
+
15
+ value: typing.Any
16
+ name: str
17
+
18
+ def json(self, **kwargs: typing.Any) -> str:
19
+ kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
20
+ return super().json(**kwargs_with_defaults)
21
+
22
+ def dict(self, **kwargs: typing.Any) -> typing.Dict[str, typing.Any]:
23
+ kwargs_with_defaults_exclude_unset: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
24
+ kwargs_with_defaults_exclude_none: typing.Any = {"by_alias": True, "exclude_none": True, **kwargs}
25
+
26
+ return deep_union_pydantic_dicts(
27
+ super().dict(**kwargs_with_defaults_exclude_unset), super().dict(**kwargs_with_defaults_exclude_none)
28
+ )
29
+
30
+ class Config:
31
+ frozen = True
32
+ smart_union = True
33
+ extra = pydantic_v1.Extra.allow
34
+ json_encoders = {dt.datetime: serialize_datetime}
@@ -34,6 +34,30 @@ class NamedScenarioInputRequest_String(pydantic_v1.BaseModel):
34
34
  json_encoders = {dt.datetime: serialize_datetime}
35
35
 
36
36
 
37
+ class NamedScenarioInputRequest_Json(pydantic_v1.BaseModel):
38
+ value: typing.Any
39
+ name: str
40
+ type: typing.Literal["JSON"] = "JSON"
41
+
42
+ def json(self, **kwargs: typing.Any) -> str:
43
+ kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
44
+ return super().json(**kwargs_with_defaults)
45
+
46
+ def dict(self, **kwargs: typing.Any) -> typing.Dict[str, typing.Any]:
47
+ kwargs_with_defaults_exclude_unset: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
48
+ kwargs_with_defaults_exclude_none: typing.Any = {"by_alias": True, "exclude_none": True, **kwargs}
49
+
50
+ return deep_union_pydantic_dicts(
51
+ super().dict(**kwargs_with_defaults_exclude_unset), super().dict(**kwargs_with_defaults_exclude_none)
52
+ )
53
+
54
+ class Config:
55
+ frozen = True
56
+ smart_union = True
57
+ extra = pydantic_v1.Extra.allow
58
+ json_encoders = {dt.datetime: serialize_datetime}
59
+
60
+
37
61
  class NamedScenarioInputRequest_ChatHistory(pydantic_v1.BaseModel):
38
62
  value: typing.Optional[typing.List[ChatMessageRequest]] = None
39
63
  name: str
@@ -58,4 +82,6 @@ class NamedScenarioInputRequest_ChatHistory(pydantic_v1.BaseModel):
58
82
  json_encoders = {dt.datetime: serialize_datetime}
59
83
 
60
84
 
61
- NamedScenarioInputRequest = typing.Union[NamedScenarioInputRequest_String, NamedScenarioInputRequest_ChatHistory]
85
+ NamedScenarioInputRequest = typing.Union[
86
+ NamedScenarioInputRequest_String, NamedScenarioInputRequest_Json, NamedScenarioInputRequest_ChatHistory
87
+ ]
@@ -34,6 +34,30 @@ class ScenarioInput_String(pydantic_v1.BaseModel):
34
34
  json_encoders = {dt.datetime: serialize_datetime}
35
35
 
36
36
 
37
+ class ScenarioInput_Json(pydantic_v1.BaseModel):
38
+ value: typing.Any
39
+ input_variable_id: str
40
+ type: typing.Literal["JSON"] = "JSON"
41
+
42
+ def json(self, **kwargs: typing.Any) -> str:
43
+ kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
44
+ return super().json(**kwargs_with_defaults)
45
+
46
+ def dict(self, **kwargs: typing.Any) -> typing.Dict[str, typing.Any]:
47
+ kwargs_with_defaults_exclude_unset: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
48
+ kwargs_with_defaults_exclude_none: typing.Any = {"by_alias": True, "exclude_none": True, **kwargs}
49
+
50
+ return deep_union_pydantic_dicts(
51
+ super().dict(**kwargs_with_defaults_exclude_unset), super().dict(**kwargs_with_defaults_exclude_none)
52
+ )
53
+
54
+ class Config:
55
+ frozen = True
56
+ smart_union = True
57
+ extra = pydantic_v1.Extra.allow
58
+ json_encoders = {dt.datetime: serialize_datetime}
59
+
60
+
37
61
  class ScenarioInput_ChatHistory(pydantic_v1.BaseModel):
38
62
  value: typing.Optional[typing.List[ChatMessage]] = None
39
63
  input_variable_id: str
@@ -58,4 +82,4 @@ class ScenarioInput_ChatHistory(pydantic_v1.BaseModel):
58
82
  json_encoders = {dt.datetime: serialize_datetime}
59
83
 
60
84
 
61
- ScenarioInput = typing.Union[ScenarioInput_String, ScenarioInput_ChatHistory]
85
+ ScenarioInput = typing.Union[ScenarioInput_String, ScenarioInput_Json, ScenarioInput_ChatHistory]
@@ -0,0 +1,34 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import datetime as dt
4
+ import typing
5
+
6
+ from ..core.datetime_utils import serialize_datetime
7
+ from ..core.pydantic_utilities import deep_union_pydantic_dicts, pydantic_v1
8
+
9
+
10
+ class ScenarioInputJsonVariableValue(pydantic_v1.BaseModel):
11
+ """
12
+ Prompt Sandbox Scenario input value that is of type JSON
13
+ """
14
+
15
+ value: typing.Any
16
+ input_variable_id: str
17
+
18
+ def json(self, **kwargs: typing.Any) -> str:
19
+ kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
20
+ return super().json(**kwargs_with_defaults)
21
+
22
+ def dict(self, **kwargs: typing.Any) -> typing.Dict[str, typing.Any]:
23
+ kwargs_with_defaults_exclude_unset: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
24
+ kwargs_with_defaults_exclude_none: typing.Any = {"by_alias": True, "exclude_none": True, **kwargs}
25
+
26
+ return deep_union_pydantic_dicts(
27
+ super().dict(**kwargs_with_defaults_exclude_unset), super().dict(**kwargs_with_defaults_exclude_none)
28
+ )
29
+
30
+ class Config:
31
+ frozen = True
32
+ smart_union = True
33
+ extra = pydantic_v1.Extra.allow
34
+ json_encoders = {dt.datetime: serialize_datetime}
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: vellum-ai
3
- Version: 0.7.6
3
+ Version: 0.7.8
4
4
  Summary:
5
5
  License: MIT
6
6
  Requires-Python: >=3.8,<4.0
@@ -1,8 +1,8 @@
1
- vellum/__init__.py,sha256=eqkGPne9ahxtbVtLlx7i1BxCA-RBaWg3qGRTCLw6lTA,49657
1
+ vellum/__init__.py,sha256=Y08IiuMhi07kjouldi6dHJvZRt8cwdEVSKrG0N9BzGM,49953
2
2
  vellum/client.py,sha256=FEelOptuh8ylBnqSznSXvIUj2LWGTEPDTPrK5sgQkSE,83651
3
3
  vellum/core/__init__.py,sha256=UFXpYzcGxWQUucU1TkjOQ9mGWN3A5JohluOIWVYKU4I,973
4
4
  vellum/core/api_error.py,sha256=RE8LELok2QCjABadECTvtDp7qejA1VmINCh6TbqPwSE,426
5
- vellum/core/client_wrapper.py,sha256=Z9KoCufJbg73Hf7nVHFk9DnB5JNBULg1oQsf7_0BQbE,1873
5
+ vellum/core/client_wrapper.py,sha256=zAmOvT3hIgcE7oJ1_UMRHNE5WUsLBY-bplg8t5xrwlo,1873
6
6
  vellum/core/datetime_utils.py,sha256=nBys2IsYrhPdszxGKCNRPSOCwa-5DWOHG95FB8G9PKo,1047
7
7
  vellum/core/file.py,sha256=sy1RUGZ3aJYuw998bZytxxo6QdgKmlnlgBaMvwEKCGg,1480
8
8
  vellum/core/http_client.py,sha256=46CyqS5Y8MwWTclAXnb1z5-ODJfwfHYbyhvjhb7RY1c,18753
@@ -41,7 +41,7 @@ vellum/resources/documents/client.py,sha256=lrRR9wp5nnMnENycYm-FrWwKIy7tKrfpHQ5L
41
41
  vellum/resources/folder_entities/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
42
42
  vellum/resources/folder_entities/client.py,sha256=EZ_RjrB87rPLoaqNC44Dkrhp7aWEqEqI2pm5bekMqLw,4359
43
43
  vellum/resources/ml_models/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
44
- vellum/resources/ml_models/client.py,sha256=1pXtLSTr3J7Jdu5INAGqyRFN8FEEZZ2DZ54k02HXR48,26644
44
+ vellum/resources/ml_models/client.py,sha256=ygrjdLSoUBJIHjT4AtofxcWezx-E1pO5-tgpt4XvwvI,26644
45
45
  vellum/resources/sandboxes/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
46
46
  vellum/resources/sandboxes/client.py,sha256=Vn80xkXWKZ8llBQSSoSqs9NU62mP1BBpNxgRBpDdLy8,15204
47
47
  vellum/resources/test_suite_runs/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
@@ -63,7 +63,7 @@ vellum/terraform/document_index/__init__.py,sha256=GY4Sn8X8-TgNiW_2Rph2uvY6tmJ6q
63
63
  vellum/terraform/provider/__init__.py,sha256=YYQLWWJDslcjc1eN0N719A3wqMdbNR2c3WuqGtX1U_I,12684
64
64
  vellum/terraform/py.typed,sha256=AbpHGcgLb-kRsJGnwFEktk7uzpZOCcBY74-YBdrKVGs,1
65
65
  vellum/terraform/versions.json,sha256=L-eKdEx1tp0hyZY8TGQ0Gy1a5R7q9hHAMLrCXICurZo,56
66
- vellum/types/__init__.py,sha256=OymATkVwsBVnU8eA1JZ9okU8N07VwDMUsWWl9rdGY5s,66253
66
+ vellum/types/__init__.py,sha256=74SSEdsGGOEXMsJZJwqRExTDFUxadMjmPdS_YpK5710,66645
67
67
  vellum/types/add_openai_api_key_enum.py,sha256=GB7sLK_Ou7-Xn73sKJHUo6Gx3TjyhU7uJvWZAg4UeaI,92
68
68
  vellum/types/api_node_result.py,sha256=H25_pCtZ9xT6GTbQG1Gae89yjK-45yKipSLNqfc2PTk,1246
69
69
  vellum/types/api_node_result_data.py,sha256=1MIuC_rN_4ul7PQVWPbQySrLKzwwVb5sjXO6G6v13z4,1424
@@ -200,8 +200,8 @@ vellum/types/ml_model_display_config_labelled.py,sha256=SleUaRCUTc3nFAqU6R8UOOzR
200
200
  vellum/types/ml_model_display_config_request.py,sha256=WfPdaLzX0X0w4V-3hQ1A8JOuej1mqsevEzkj8pW-ry4,1242
201
201
  vellum/types/ml_model_display_tag.py,sha256=mvxWDmFB7nxMiO9f9XEueWoXL3YmkPfqvroDVzVXNyE,229
202
202
  vellum/types/ml_model_display_tag_enum_value_label.py,sha256=r8_Nxz3TOBcEhWmVVgCbrI0h2x85Q8RsGA-TLZOBBe0,1213
203
- vellum/types/ml_model_exec_config.py,sha256=dveJCP_BOy4VcNP9Uv15ISUTyRA2Z5DRauDWT2weSos,1717
204
- vellum/types/ml_model_exec_config_request.py,sha256=ul1FDz4JS-G94E9EP4LCFGi3c-Us9kEReKlXYOfKu9M,1790
203
+ vellum/types/ml_model_exec_config.py,sha256=bZiGQ2XQ2M67eJHzOM9wZIE3oq2Mee9B-9InNj5-Ew4,1658
204
+ vellum/types/ml_model_exec_config_request.py,sha256=Fy5nGCj7zXCnB6N1fs0cmdRCRV_hvYYNzS2NMXH47jE,1731
205
205
  vellum/types/ml_model_family.py,sha256=96a6cigZeoGoM4VLvwvSmIKnEVBAI8RsW7xaS3EZZDI,513
206
206
  vellum/types/ml_model_family_enum_value_label.py,sha256=ldXOZN0ZV277Q9qSjli3kJJ47ozzDe50d1DanB_l9v4,1196
207
207
  vellum/types/ml_model_feature.py,sha256=s3DTCKk5m8MBdVV4-gHGe3Oj4B38fS6I1LLj8pps4Lo,441
@@ -219,7 +219,8 @@ vellum/types/ml_model_tokenizer_config.py,sha256=4E01CgGUU5FlHja7wnMio8pSpfnHStH
219
219
  vellum/types/ml_model_tokenizer_config_request.py,sha256=mOLzJIidqLnkjg3Sj1JvHhX-e7ATK_tGuZt7oZ5W-Hc,2365
220
220
  vellum/types/ml_model_usage.py,sha256=IFbXxMXf-4bkGuOzPIgfm0acZ4lVRI9C6uQpeO4_0o8,1349
221
221
  vellum/types/named_scenario_input_chat_history_variable_value_request.py,sha256=CfGCHcFOEWL3UV-VjA6pYyExnpAabrlQ44c6LxdhjyA,1361
222
- vellum/types/named_scenario_input_request.py,sha256=-GRkDZGwkhlaIwDcTenP_-u4fH79bRE0rRG5Ddi9bSs,2452
222
+ vellum/types/named_scenario_input_json_variable_value_request.py,sha256=r6jsE_wRPQHudeTaKDrSlKYVH2Nu4F6PlSaKx21o4GM,1248
223
+ vellum/types/named_scenario_input_request.py,sha256=KGuT0bFtt3rmBJHmWkTlD8zKqRtmZKfP865aHifi6XM,3452
223
224
  vellum/types/named_scenario_input_string_variable_value_request.py,sha256=tlJ8G8ru1BQHxdjfBpMiwKhQOHOw3mROuyJzG3KcGow,1269
224
225
  vellum/types/named_test_case_array_variable_value.py,sha256=reOwe29S23QepGANR5yGWpKpUUQ_SUDPg5TCBBEnEDA,1323
225
226
  vellum/types/named_test_case_array_variable_value_request.py,sha256=-_F7ymtx32IfSCSgoHjHXDE1kcloeQvZdy8lhRY1aCI,1352
@@ -319,8 +320,9 @@ vellum/types/rejected_workflow_node_result_event.py,sha256=n0yp5qdEyTEsnSVRAWPFU
319
320
  vellum/types/release_tag_source.py,sha256=YavosOXZ976yfXTNWRTZwh2HhRiYmSDk0bQCkl-jCoQ,158
320
321
  vellum/types/replace_test_suite_test_case_request.py,sha256=70JYolE2hwB52LU5gpkEgF3bKC-epaeDeuJrgPzJDmE,2209
321
322
  vellum/types/sandbox_scenario.py,sha256=1vp9eQhMMPpiHfSWoAtTUpDmx0jy4VMWW9ZzHlaj7Yk,1407
322
- vellum/types/scenario_input.py,sha256=o98RTH7IDbXBCBKF2JKiCPvUkkloYDhWJjvGxMgQVKA,2396
323
+ vellum/types/scenario_input.py,sha256=fMFI-FPKq6WBA1I_h_XrheeVN27JoqU_YB5vCGlD-ho,3379
323
324
  vellum/types/scenario_input_chat_history_variable_value.py,sha256=l63tBZzYwgtcBVB_-Iph-ylRRQMV7FkD7fZVbadLk2c,1334
325
+ vellum/types/scenario_input_json_variable_value.py,sha256=SXZ7stmJ2DmXcYdmiwTkvKahP87EiCcau3QoRTn-q9I,1243
324
326
  vellum/types/scenario_input_string_variable_value.py,sha256=QEeWZyJWvEeEParxhVs4n5sJ6fztkz4uA1CVqnIrVEw,1264
325
327
  vellum/types/search_filters_request.py,sha256=3k90Kq2KWu8p9a-CYtkcybS7ODNosNLgBIk-ZL0uQkM,1481
326
328
  vellum/types/search_node_result.py,sha256=Yz9S17nCtOTQDkhWM_hPUOHG8ZttG1TQukzSVjnWGH4,1260
@@ -494,7 +496,7 @@ vellum/types/workflow_result_event_output_data_search_results.py,sha256=_C4ueKK8
494
496
  vellum/types/workflow_result_event_output_data_string.py,sha256=AAWHZT3X9HOIRA3UuIqw0VpfSGwGemsJM71WDNbWYTc,1745
495
497
  vellum/types/workflow_stream_event.py,sha256=5K-Mtn9fvJDq8m5nhURDbChL01PXIiuIZDkfAC1d6fU,2610
496
498
  vellum/version.py,sha256=neLt8HBHHUtDF9M5fsyUzHT-pKooEPvceaLDqqIGb0s,77
497
- vellum_ai-0.7.6.dist-info/LICENSE,sha256=CcaljEIoOBaU-wItPH4PmM_mDCGpyuUY0Er1BGu5Ti8,1073
498
- vellum_ai-0.7.6.dist-info/METADATA,sha256=Px_DEw78UFucYoZd9Z3V9a8pjtuPXhl8E2AcMkjRNvc,4398
499
- vellum_ai-0.7.6.dist-info/WHEEL,sha256=Zb28QaM1gQi8f4VCBhsUklF61CTlNYfs9YAZn-TOGFk,88
500
- vellum_ai-0.7.6.dist-info/RECORD,,
499
+ vellum_ai-0.7.8.dist-info/LICENSE,sha256=CcaljEIoOBaU-wItPH4PmM_mDCGpyuUY0Er1BGu5Ti8,1073
500
+ vellum_ai-0.7.8.dist-info/METADATA,sha256=R5nnOWBhtOMhaXi9kk1yndyxQnvBrb-nkVdWAnHmrEs,4398
501
+ vellum_ai-0.7.8.dist-info/WHEEL,sha256=Zb28QaM1gQi8f4VCBhsUklF61CTlNYfs9YAZn-TOGFk,88
502
+ vellum_ai-0.7.8.dist-info/RECORD,,