vellum-ai 0.3.20__py3-none-any.whl → 0.3.21__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- vellum/__init__.py +2 -0
 - vellum/client.py +4 -0
 - vellum/core/client_wrapper.py +1 -1
 - vellum/types/__init__.py +2 -0
 - vellum/types/fulfilled_prompt_execution_meta.py +2 -0
 - vellum/types/ml_model_usage.py +33 -0
 - vellum/types/prompt_deployment_expand_meta_request_request.py +5 -0
 - vellum/types/prompt_execution_meta.py +2 -0
 - {vellum_ai-0.3.20.dist-info → vellum_ai-0.3.21.dist-info}/METADATA +1 -1
 - {vellum_ai-0.3.20.dist-info → vellum_ai-0.3.21.dist-info}/RECORD +12 -11
 - {vellum_ai-0.3.20.dist-info → vellum_ai-0.3.21.dist-info}/LICENSE +0 -0
 - {vellum_ai-0.3.20.dist-info → vellum_ai-0.3.21.dist-info}/WHEEL +0 -0
 
    
        vellum/__init__.py
    CHANGED
    
    | 
         @@ -148,6 +148,7 @@ from .types import ( 
     | 
|
| 
       148 
148 
     | 
    
         
             
                MetadataFilterConfigRequest,
         
     | 
| 
       149 
149 
     | 
    
         
             
                MetadataFilterRuleCombinator,
         
     | 
| 
       150 
150 
     | 
    
         
             
                MetadataFilterRuleRequest,
         
     | 
| 
      
 151 
     | 
    
         
            +
                MlModelUsage,
         
     | 
| 
       151 
152 
     | 
    
         
             
                ModelVersionBuildConfig,
         
     | 
| 
       152 
153 
     | 
    
         
             
                ModelVersionExecConfig,
         
     | 
| 
       153 
154 
     | 
    
         
             
                ModelVersionExecConfigParameters,
         
     | 
| 
         @@ -641,6 +642,7 @@ __all__ = [ 
     | 
|
| 
       641 
642 
     | 
    
         
             
                "MetadataFilterConfigRequest",
         
     | 
| 
       642 
643 
     | 
    
         
             
                "MetadataFilterRuleCombinator",
         
     | 
| 
       643 
644 
     | 
    
         
             
                "MetadataFilterRuleRequest",
         
     | 
| 
      
 645 
     | 
    
         
            +
                "MlModelUsage",
         
     | 
| 
       644 
646 
     | 
    
         
             
                "ModelVersionBuildConfig",
         
     | 
| 
       645 
647 
     | 
    
         
             
                "ModelVersionExecConfig",
         
     | 
| 
       646 
648 
     | 
    
         
             
                "ModelVersionExecConfigParameters",
         
     | 
    
        vellum/client.py
    CHANGED
    
    | 
         @@ -167,6 +167,7 @@ class Vellum: 
     | 
|
| 
       167 
167 
     | 
    
         
             
                            deployment_release_tag=True,
         
     | 
| 
       168 
168 
     | 
    
         
             
                            prompt_version_id=True,
         
     | 
| 
       169 
169 
     | 
    
         
             
                            finish_reason=True,
         
     | 
| 
      
 170 
     | 
    
         
            +
                            usage=True,
         
     | 
| 
       170 
171 
     | 
    
         
             
                        ),
         
     | 
| 
       171 
172 
     | 
    
         
             
                        raw_overrides=RawPromptExecutionOverridesRequest(
         
     | 
| 
       172 
173 
     | 
    
         
             
                            body={"string": {"key": "value"}},
         
     | 
| 
         @@ -301,6 +302,7 @@ class Vellum: 
     | 
|
| 
       301 
302 
     | 
    
         
             
                            deployment_release_tag=True,
         
     | 
| 
       302 
303 
     | 
    
         
             
                            prompt_version_id=True,
         
     | 
| 
       303 
304 
     | 
    
         
             
                            finish_reason=True,
         
     | 
| 
      
 305 
     | 
    
         
            +
                            usage=True,
         
     | 
| 
       304 
306 
     | 
    
         
             
                        ),
         
     | 
| 
       305 
307 
     | 
    
         
             
                        raw_overrides=RawPromptExecutionOverridesRequest(
         
     | 
| 
       306 
308 
     | 
    
         
             
                            body={"string": {"key": "value"}},
         
     | 
| 
         @@ -1123,6 +1125,7 @@ class AsyncVellum: 
     | 
|
| 
       1123 
1125 
     | 
    
         
             
                            deployment_release_tag=True,
         
     | 
| 
       1124 
1126 
     | 
    
         
             
                            prompt_version_id=True,
         
     | 
| 
       1125 
1127 
     | 
    
         
             
                            finish_reason=True,
         
     | 
| 
      
 1128 
     | 
    
         
            +
                            usage=True,
         
     | 
| 
       1126 
1129 
     | 
    
         
             
                        ),
         
     | 
| 
       1127 
1130 
     | 
    
         
             
                        raw_overrides=RawPromptExecutionOverridesRequest(
         
     | 
| 
       1128 
1131 
     | 
    
         
             
                            body={"string": {"key": "value"}},
         
     | 
| 
         @@ -1257,6 +1260,7 @@ class AsyncVellum: 
     | 
|
| 
       1257 
1260 
     | 
    
         
             
                            deployment_release_tag=True,
         
     | 
| 
       1258 
1261 
     | 
    
         
             
                            prompt_version_id=True,
         
     | 
| 
       1259 
1262 
     | 
    
         
             
                            finish_reason=True,
         
     | 
| 
      
 1263 
     | 
    
         
            +
                            usage=True,
         
     | 
| 
       1260 
1264 
     | 
    
         
             
                        ),
         
     | 
| 
       1261 
1265 
     | 
    
         
             
                        raw_overrides=RawPromptExecutionOverridesRequest(
         
     | 
| 
       1262 
1266 
     | 
    
         
             
                            body={"string": {"key": "value"}},
         
     | 
    
        vellum/core/client_wrapper.py
    CHANGED
    
    | 
         @@ -18,7 +18,7 @@ class BaseClientWrapper: 
     | 
|
| 
       18 
18 
     | 
    
         
             
                    headers: typing.Dict[str, str] = {
         
     | 
| 
       19 
19 
     | 
    
         
             
                        "X-Fern-Language": "Python",
         
     | 
| 
       20 
20 
     | 
    
         
             
                        "X-Fern-SDK-Name": "vellum-ai",
         
     | 
| 
       21 
     | 
    
         
            -
                        "X-Fern-SDK-Version": "0.3. 
     | 
| 
      
 21 
     | 
    
         
            +
                        "X-Fern-SDK-Version": "0.3.21",
         
     | 
| 
       22 
22 
     | 
    
         
             
                    }
         
     | 
| 
       23 
23 
     | 
    
         
             
                    headers["X_API_KEY"] = self.api_key
         
     | 
| 
       24 
24 
     | 
    
         
             
                    return headers
         
     | 
    
        vellum/types/__init__.py
    CHANGED
    
    | 
         @@ -165,6 +165,7 @@ from .logprobs_enum import LogprobsEnum 
     | 
|
| 
       165 
165 
     | 
    
         
             
            from .metadata_filter_config_request import MetadataFilterConfigRequest
         
     | 
| 
       166 
166 
     | 
    
         
             
            from .metadata_filter_rule_combinator import MetadataFilterRuleCombinator
         
     | 
| 
       167 
167 
     | 
    
         
             
            from .metadata_filter_rule_request import MetadataFilterRuleRequest
         
     | 
| 
      
 168 
     | 
    
         
            +
            from .ml_model_usage import MlModelUsage
         
     | 
| 
       168 
169 
     | 
    
         
             
            from .model_version_build_config import ModelVersionBuildConfig
         
     | 
| 
       169 
170 
     | 
    
         
             
            from .model_version_exec_config import ModelVersionExecConfig
         
     | 
| 
       170 
171 
     | 
    
         
             
            from .model_version_exec_config_parameters import ModelVersionExecConfigParameters
         
     | 
| 
         @@ -673,6 +674,7 @@ __all__ = [ 
     | 
|
| 
       673 
674 
     | 
    
         
             
                "MetadataFilterConfigRequest",
         
     | 
| 
       674 
675 
     | 
    
         
             
                "MetadataFilterRuleCombinator",
         
     | 
| 
       675 
676 
     | 
    
         
             
                "MetadataFilterRuleRequest",
         
     | 
| 
      
 677 
     | 
    
         
            +
                "MlModelUsage",
         
     | 
| 
       676 
678 
     | 
    
         
             
                "ModelVersionBuildConfig",
         
     | 
| 
       677 
679 
     | 
    
         
             
                "ModelVersionExecConfig",
         
     | 
| 
       678 
680 
     | 
    
         
             
                "ModelVersionExecConfigParameters",
         
     | 
| 
         @@ -5,6 +5,7 @@ import typing 
     | 
|
| 
       5 
5 
     | 
    
         | 
| 
       6 
6 
     | 
    
         
             
            from ..core.datetime_utils import serialize_datetime
         
     | 
| 
       7 
7 
     | 
    
         
             
            from .finish_reason_enum import FinishReasonEnum
         
     | 
| 
      
 8 
     | 
    
         
            +
            from .ml_model_usage import MlModelUsage
         
     | 
| 
       8 
9 
     | 
    
         | 
| 
       9 
10 
     | 
    
         
             
            try:
         
     | 
| 
       10 
11 
     | 
    
         
             
                import pydantic.v1 as pydantic  # type: ignore
         
     | 
| 
         @@ -19,6 +20,7 @@ class FulfilledPromptExecutionMeta(pydantic.BaseModel): 
     | 
|
| 
       19 
20 
     | 
    
         | 
| 
       20 
21 
     | 
    
         
             
                latency: typing.Optional[int] = None
         
     | 
| 
       21 
22 
     | 
    
         
             
                finish_reason: typing.Optional[FinishReasonEnum] = None
         
     | 
| 
      
 23 
     | 
    
         
            +
                usage: typing.Optional[MlModelUsage] = None
         
     | 
| 
       22 
24 
     | 
    
         | 
| 
       23 
25 
     | 
    
         
             
                def json(self, **kwargs: typing.Any) -> str:
         
     | 
| 
       24 
26 
     | 
    
         
             
                    kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
         
     | 
| 
         @@ -0,0 +1,33 @@ 
     | 
|
| 
      
 1 
     | 
    
         
            +
            # This file was auto-generated by Fern from our API Definition.
         
     | 
| 
      
 2 
     | 
    
         
            +
             
     | 
| 
      
 3 
     | 
    
         
            +
            import datetime as dt
         
     | 
| 
      
 4 
     | 
    
         
            +
            import typing
         
     | 
| 
      
 5 
     | 
    
         
            +
             
     | 
| 
      
 6 
     | 
    
         
            +
            from ..core.datetime_utils import serialize_datetime
         
     | 
| 
      
 7 
     | 
    
         
            +
             
     | 
| 
      
 8 
     | 
    
         
            +
            try:
         
     | 
| 
      
 9 
     | 
    
         
            +
                import pydantic.v1 as pydantic  # type: ignore
         
     | 
| 
      
 10 
     | 
    
         
            +
            except ImportError:
         
     | 
| 
      
 11 
     | 
    
         
            +
                import pydantic  # type: ignore
         
     | 
| 
      
 12 
     | 
    
         
            +
             
     | 
| 
      
 13 
     | 
    
         
            +
             
     | 
| 
      
 14 
     | 
    
         
            +
            class MlModelUsage(pydantic.BaseModel):
         
     | 
| 
      
 15 
     | 
    
         
            +
                output_token_count: typing.Optional[int] = None
         
     | 
| 
      
 16 
     | 
    
         
            +
                input_token_count: typing.Optional[int] = None
         
     | 
| 
      
 17 
     | 
    
         
            +
                input_char_count: typing.Optional[int] = None
         
     | 
| 
      
 18 
     | 
    
         
            +
                output_char_count: typing.Optional[int] = None
         
     | 
| 
      
 19 
     | 
    
         
            +
                compute_nanos: typing.Optional[int] = None
         
     | 
| 
      
 20 
     | 
    
         
            +
             
     | 
| 
      
 21 
     | 
    
         
            +
                def json(self, **kwargs: typing.Any) -> str:
         
     | 
| 
      
 22 
     | 
    
         
            +
                    kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
         
     | 
| 
      
 23 
     | 
    
         
            +
                    return super().json(**kwargs_with_defaults)
         
     | 
| 
      
 24 
     | 
    
         
            +
             
     | 
| 
      
 25 
     | 
    
         
            +
                def dict(self, **kwargs: typing.Any) -> typing.Dict[str, typing.Any]:
         
     | 
| 
      
 26 
     | 
    
         
            +
                    kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
         
     | 
| 
      
 27 
     | 
    
         
            +
                    return super().dict(**kwargs_with_defaults)
         
     | 
| 
      
 28 
     | 
    
         
            +
             
     | 
| 
      
 29 
     | 
    
         
            +
                class Config:
         
     | 
| 
      
 30 
     | 
    
         
            +
                    frozen = True
         
     | 
| 
      
 31 
     | 
    
         
            +
                    smart_union = True
         
     | 
| 
      
 32 
     | 
    
         
            +
                    extra = pydantic.Extra.allow
         
     | 
| 
      
 33 
     | 
    
         
            +
                    json_encoders = {dt.datetime: serialize_datetime}
         
     | 
| 
         @@ -37,6 +37,11 @@ class PromptDeploymentExpandMetaRequestRequest(pydantic.BaseModel): 
     | 
|
| 
       37 
37 
     | 
    
         
             
                If enabled, the response will include the reason provided by the model for why the execution finished.
         
     | 
| 
       38 
38 
     | 
    
         
             
                """
         
     | 
| 
       39 
39 
     | 
    
         | 
| 
      
 40 
     | 
    
         
            +
                usage: typing.Optional[bool] = pydantic.Field(default=None)
         
     | 
| 
      
 41 
     | 
    
         
            +
                """
         
     | 
| 
      
 42 
     | 
    
         
            +
                If enabled, the response will include model host usage tracking. This may increase latency for some model hosts.
         
     | 
| 
      
 43 
     | 
    
         
            +
                """
         
     | 
| 
      
 44 
     | 
    
         
            +
             
     | 
| 
       40 
45 
     | 
    
         
             
                def json(self, **kwargs: typing.Any) -> str:
         
     | 
| 
       41 
46 
     | 
    
         
             
                    kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
         
     | 
| 
       42 
47 
     | 
    
         
             
                    return super().json(**kwargs_with_defaults)
         
     | 
| 
         @@ -5,6 +5,7 @@ import typing 
     | 
|
| 
       5 
5 
     | 
    
         | 
| 
       6 
6 
     | 
    
         
             
            from ..core.datetime_utils import serialize_datetime
         
     | 
| 
       7 
7 
     | 
    
         
             
            from .finish_reason_enum import FinishReasonEnum
         
     | 
| 
      
 8 
     | 
    
         
            +
            from .ml_model_usage import MlModelUsage
         
     | 
| 
       8 
9 
     | 
    
         | 
| 
       9 
10 
     | 
    
         
             
            try:
         
     | 
| 
       10 
11 
     | 
    
         
             
                import pydantic.v1 as pydantic  # type: ignore
         
     | 
| 
         @@ -17,6 +18,7 @@ class PromptExecutionMeta(pydantic.BaseModel): 
     | 
|
| 
       17 
18 
     | 
    
         
             
                The subset of the metadata tracked by Vellum during prompt execution that the request opted into with `expand_meta`.
         
     | 
| 
       18 
19 
     | 
    
         
             
                """
         
     | 
| 
       19 
20 
     | 
    
         | 
| 
      
 21 
     | 
    
         
            +
                usage: typing.Optional[MlModelUsage] = None
         
     | 
| 
       20 
22 
     | 
    
         
             
                model_name: typing.Optional[str] = None
         
     | 
| 
       21 
23 
     | 
    
         
             
                latency: typing.Optional[int] = None
         
     | 
| 
       22 
24 
     | 
    
         
             
                deployment_release_tag: typing.Optional[str] = None
         
     | 
| 
         @@ -1,8 +1,8 @@ 
     | 
|
| 
       1 
     | 
    
         
            -
            vellum/__init__.py,sha256= 
     | 
| 
       2 
     | 
    
         
            -
            vellum/client.py,sha256= 
     | 
| 
      
 1 
     | 
    
         
            +
            vellum/__init__.py,sha256=h2QBLHkAs9OTsNI_W0fYfaUrQUmLd37wNbkK63TQcGY,33979
         
     | 
| 
      
 2 
     | 
    
         
            +
            vellum/client.py,sha256=ViOMnbiP_kKm3ayXW6DY1wC7Ge8tnI389qC8XVffyB8,96748
         
     | 
| 
       3 
3 
     | 
    
         
             
            vellum/core/__init__.py,sha256=RWfyDqkzWsf8e3VGc3NV60MovfJbg5XWzNFGB2DZ0hA,790
         
     | 
| 
       4 
4 
     | 
    
         
             
            vellum/core/api_error.py,sha256=RE8LELok2QCjABadECTvtDp7qejA1VmINCh6TbqPwSE,426
         
     | 
| 
       5 
     | 
    
         
            -
            vellum/core/client_wrapper.py,sha256= 
     | 
| 
      
 5 
     | 
    
         
            +
            vellum/core/client_wrapper.py,sha256=zOXh9ZnRKMmvfCIpadlblmkfFSVgnNdLnEaRVillKps,1698
         
     | 
| 
       6 
6 
     | 
    
         
             
            vellum/core/datetime_utils.py,sha256=nBys2IsYrhPdszxGKCNRPSOCwa-5DWOHG95FB8G9PKo,1047
         
     | 
| 
       7 
7 
     | 
    
         
             
            vellum/core/file.py,sha256=sy1RUGZ3aJYuw998bZytxxo6QdgKmlnlgBaMvwEKCGg,1480
         
     | 
| 
       8 
8 
     | 
    
         
             
            vellum/core/http_client.py,sha256=5ok6hqgZDJhg57EHvMnr0BBaHdG50QxFPKaCZ9aVWTc,5059
         
     | 
| 
         @@ -44,7 +44,7 @@ vellum/resources/workflow_deployments/__init__.py,sha256=-5BCA0kSmW6WUh4gqLuQtHv 
     | 
|
| 
       44 
44 
     | 
    
         
             
            vellum/resources/workflow_deployments/client.py,sha256=DUY1h8mvXqwP-zilmwoek_XbLNTFeeVbxgd7M-k2IzA,11164
         
     | 
| 
       45 
45 
     | 
    
         
             
            vellum/resources/workflow_deployments/types/__init__.py,sha256=rmS_4dtbgLHGNQJ_pOloygrjl4sNbKZjTEKBxbMyz6E,208
         
     | 
| 
       46 
46 
     | 
    
         
             
            vellum/resources/workflow_deployments/types/workflow_deployments_list_request_status.py,sha256=8-G1SalBR6-AfRnb6POOR9M3tvZa3CGwFIs1ArZb6uw,542
         
     | 
| 
       47 
     | 
    
         
            -
            vellum/types/__init__.py,sha256= 
     | 
| 
      
 47 
     | 
    
         
            +
            vellum/types/__init__.py,sha256=aXYiDbcBGEypul3Ae4DcgwNT55UtGftQsVxTPRpVAZU,45312
         
     | 
| 
       48 
48 
     | 
    
         
             
            vellum/types/api_node_result.py,sha256=aLOooryBCfglTeZwOuyyqRoHz5oPTja2mYYY7Gj8XxA,1040
         
     | 
| 
       49 
49 
     | 
    
         
             
            vellum/types/api_node_result_data.py,sha256=cQcZiwI1hjRvQFlkVEGxyaD360vEklIfjBmhBAeI75U,1215
         
     | 
| 
       50 
50 
     | 
    
         
             
            vellum/types/array_chat_message_content.py,sha256=Z7nKH8RNgot0i4-159w95OHywLjc-376gBPOovbw2hk,1087
         
     | 
| 
         @@ -108,7 +108,7 @@ vellum/types/fulfilled_execute_prompt_event.py,sha256=2_jNUOaSg12WKggFiczulVoHFM 
     | 
|
| 
       108 
108 
     | 
    
         
             
            vellum/types/fulfilled_execute_prompt_response.py,sha256=LyVpUlphfBrqSWzo1QyIzdTjAJ8cgW60U98VnMd83ws,1508
         
     | 
| 
       109 
109 
     | 
    
         
             
            vellum/types/fulfilled_execute_workflow_workflow_result_event.py,sha256=ROLCpeU2FoqyisPDvgumUlYFDUbwOr93pEfII-ST3NI,1146
         
     | 
| 
       110 
110 
     | 
    
         
             
            vellum/types/fulfilled_function_call.py,sha256=wIk8JVy9HyjSbY7rPOhrXPh3OuKy3mtpUvPSOx2FcUc,1056
         
     | 
| 
       111 
     | 
    
         
            -
            vellum/types/fulfilled_prompt_execution_meta.py,sha256= 
     | 
| 
      
 111 
     | 
    
         
            +
            vellum/types/fulfilled_prompt_execution_meta.py,sha256=FHivWXQbtMMoud9vMC4rc_ik1JWPZG6HfJOkuevCnCs,1285
         
     | 
| 
       112 
112 
     | 
    
         
             
            vellum/types/fulfilled_workflow_node_result_event.py,sha256=MGxts5L-VBaYP7xbL5ojjtKWEXUv45XSPqT7Y2RtvjI,1454
         
     | 
| 
       113 
113 
     | 
    
         
             
            vellum/types/function_call.py,sha256=U1_fLhZ_VGnHYPX8rb7iqyyUhBzzbRAGR-ai4wLDrs4,813
         
     | 
| 
       114 
114 
     | 
    
         
             
            vellum/types/function_call_chat_message_content.py,sha256=z9bTybrPXuhztOUw65PXGws-obJkCJWhc0Zfv1ZwA-k,1122
         
     | 
| 
         @@ -144,6 +144,7 @@ vellum/types/logprobs_enum.py,sha256=cw9yroKgrZzpPHniPV4hg7_21PM9_FSX1GArHHC_-_o 
     | 
|
| 
       144 
144 
     | 
    
         
             
            vellum/types/metadata_filter_config_request.py,sha256=JH5EBNES30DuRIhgad6IYwXZcuAW6bRajSnWUqs4hG4,1412
         
     | 
| 
       145 
145 
     | 
    
         
             
            vellum/types/metadata_filter_rule_combinator.py,sha256=hmfQ3drGlGYqRA78iZr7ek8qZTFN-WENrwL3_OlexMM,528
         
     | 
| 
       146 
146 
     | 
    
         
             
            vellum/types/metadata_filter_rule_request.py,sha256=JGPkjXzMS9I3dlEQUAJ0HllJ3F-Znp-0x-4hp06miPs,1428
         
     | 
| 
      
 147 
     | 
    
         
            +
            vellum/types/ml_model_usage.py,sha256=2o6Yy9lMDTqRHQNpUzQSoQMfA5_my79LA1pBZDcGLV0,1143
         
     | 
| 
       147 
148 
     | 
    
         
             
            vellum/types/model_version_build_config.py,sha256=0dOGfizN30eYJtjs-fJc32TUBGdpYuMFo57a_tEutSg,1400
         
     | 
| 
       148 
149 
     | 
    
         
             
            vellum/types/model_version_exec_config.py,sha256=tJJ1Kb8rBdbkD4z9fVwP94bjR9gee4I9N-i9o9w8vVs,1677
         
     | 
| 
       149 
150 
     | 
    
         
             
            vellum/types/model_version_exec_config_parameters.py,sha256=jF4wm_GIeITFavMNGSvBd-fa5v9fEjWxoHaZhVFx0CI,1326
         
     | 
| 
         @@ -187,9 +188,9 @@ vellum/types/paginated_test_suite_run_execution_list.py,sha256=pwT-FeIxOBnP_S3kp 
     | 
|
| 
       187 
188 
     | 
    
         
             
            vellum/types/paginated_test_suite_test_case_list.py,sha256=0XlghGkiSubI41oR1UGSdYOKDUM4XLCgmcElKaYFdkA,1101
         
     | 
| 
       188 
189 
     | 
    
         
             
            vellum/types/processing_failure_reason_enum.py,sha256=MDj2vNyO1Y-2WHuolkrGez8F1cZqS6ultfsqvGI4Fg8,752
         
     | 
| 
       189 
190 
     | 
    
         
             
            vellum/types/processing_state_enum.py,sha256=rMhw5oLZNfhR4QkIgFfLMWRSLXgHt9qhiguEqWQNz5k,962
         
     | 
| 
       190 
     | 
    
         
            -
            vellum/types/prompt_deployment_expand_meta_request_request.py,sha256= 
     | 
| 
      
 191 
     | 
    
         
            +
            vellum/types/prompt_deployment_expand_meta_request_request.py,sha256=bpA4eO3wIwg5fZK89eoC8B8QsRrKFddIN69o04YEBPk,2083
         
     | 
| 
       191 
192 
     | 
    
         
             
            vellum/types/prompt_deployment_input_request.py,sha256=o8PCi7TsitddNxiAfDblxaXkJ1-WjuDTAW0LnKbMDY4,1233
         
     | 
| 
       192 
     | 
    
         
            -
            vellum/types/prompt_execution_meta.py,sha256= 
     | 
| 
      
 193 
     | 
    
         
            +
            vellum/types/prompt_execution_meta.py,sha256=4RNkaZ2QjUtZwGUc9574Y4KuoZnzW2UCWj2RNiNiGKQ,1427
         
     | 
| 
       193 
194 
     | 
    
         
             
            vellum/types/prompt_node_result.py,sha256=s6jusFXMD1calmDVdL0lWO0NIqILWpdLP8xFRqf-5-k,1054
         
     | 
| 
       194 
195 
     | 
    
         
             
            vellum/types/prompt_node_result_data.py,sha256=q6y32EyaaAoIxM1od6-5RQAmPTitBQ-Hlkw5QNdeo6g,1045
         
     | 
| 
       195 
196 
     | 
    
         
             
            vellum/types/prompt_output.py,sha256=Df5UnLyBRf79_oUnYiMl9vi_h41yKEDLw1KgkgiX1Uk,1442
         
     | 
| 
         @@ -362,7 +363,7 @@ vellum/types/workflow_result_event_output_data_search_results.py,sha256=wYc76-DW 
     | 
|
| 
       362 
363 
     | 
    
         
             
            vellum/types/workflow_result_event_output_data_string.py,sha256=7SR46tKt-I_cGd_eVeKU8ymQ7R3vKkA_dFHCgAU1GcQ,1533
         
     | 
| 
       363 
364 
     | 
    
         
             
            vellum/types/workflow_stream_event.py,sha256=KA6Bkk_XA6AIPWR-1vKnwF1A8l_Bm5y0arQCWWWRpsk,911
         
     | 
| 
       364 
365 
     | 
    
         
             
            vellum/version.py,sha256=neLt8HBHHUtDF9M5fsyUzHT-pKooEPvceaLDqqIGb0s,77
         
     | 
| 
       365 
     | 
    
         
            -
            vellum_ai-0.3. 
     | 
| 
       366 
     | 
    
         
            -
            vellum_ai-0.3. 
     | 
| 
       367 
     | 
    
         
            -
            vellum_ai-0.3. 
     | 
| 
       368 
     | 
    
         
            -
            vellum_ai-0.3. 
     | 
| 
      
 366 
     | 
    
         
            +
            vellum_ai-0.3.21.dist-info/LICENSE,sha256=CcaljEIoOBaU-wItPH4PmM_mDCGpyuUY0Er1BGu5Ti8,1073
         
     | 
| 
      
 367 
     | 
    
         
            +
            vellum_ai-0.3.21.dist-info/METADATA,sha256=BPMxESrgGDvdc1CRhBOdYCRugknRd_NwOR5D5NzmRI4,3550
         
     | 
| 
      
 368 
     | 
    
         
            +
            vellum_ai-0.3.21.dist-info/WHEEL,sha256=Zb28QaM1gQi8f4VCBhsUklF61CTlNYfs9YAZn-TOGFk,88
         
     | 
| 
      
 369 
     | 
    
         
            +
            vellum_ai-0.3.21.dist-info/RECORD,,
         
     | 
| 
         
            File without changes
         
     | 
| 
         
            File without changes
         
     |