vellum-ai 0.1.4__py3-none-any.whl → 0.1.5__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- vellum/__init__.py +30 -0
- vellum/client.py +116 -5
- vellum/core/client_wrapper.py +1 -1
- vellum/errors/forbidden_error.py +3 -2
- vellum/types/__init__.py +34 -0
- vellum/types/chat_history_input_request.py +30 -0
- vellum/types/error_execute_prompt_response.py +30 -0
- vellum/types/execute_prompt_api_error_response.py +28 -0
- vellum/types/execute_prompt_response.py +43 -0
- vellum/types/json_execute_prompt_response.py +29 -0
- vellum/types/json_input_request.py +29 -0
- vellum/types/prompt_deployment_input_request.py +43 -0
- vellum/types/string_execute_prompt_response.py +29 -0
- vellum/types/string_input_request.py +29 -0
- {vellum_ai-0.1.4.dist-info → vellum_ai-0.1.5.dist-info}/METADATA +1 -1
- {vellum_ai-0.1.4.dist-info → vellum_ai-0.1.5.dist-info}/RECORD +17 -8
- {vellum_ai-0.1.4.dist-info → vellum_ai-0.1.5.dist-info}/WHEEL +0 -0
vellum/__init__.py
CHANGED
@@ -4,6 +4,7 @@ from .types import (
|
|
4
4
|
ApiNodeResult,
|
5
5
|
ApiNodeResultData,
|
6
6
|
BlockTypeEnum,
|
7
|
+
ChatHistoryInputRequest,
|
7
8
|
ChatMessage,
|
8
9
|
ChatMessageRequest,
|
9
10
|
ChatMessageRole,
|
@@ -18,8 +19,14 @@ from .types import (
|
|
18
19
|
DocumentStatus,
|
19
20
|
EnrichedNormalizedCompletion,
|
20
21
|
EnvironmentEnum,
|
22
|
+
ErrorExecutePromptResponse,
|
21
23
|
EvaluationParams,
|
22
24
|
EvaluationParamsRequest,
|
25
|
+
ExecutePromptApiErrorResponse,
|
26
|
+
ExecutePromptResponse,
|
27
|
+
ExecutePromptResponse_Error,
|
28
|
+
ExecutePromptResponse_Json,
|
29
|
+
ExecutePromptResponse_String,
|
23
30
|
ExecuteWorkflowStreamErrorResponse,
|
24
31
|
FinishReasonEnum,
|
25
32
|
GenerateErrorResponse,
|
@@ -33,6 +40,8 @@ from .types import (
|
|
33
40
|
GenerateStreamResult,
|
34
41
|
GenerateStreamResultData,
|
35
42
|
IndexingStateEnum,
|
43
|
+
JsonExecutePromptResponse,
|
44
|
+
JsonInputRequest,
|
36
45
|
LogicalOperator,
|
37
46
|
LogprobsEnum,
|
38
47
|
MetadataFilterConfigRequest,
|
@@ -62,6 +71,10 @@ from .types import (
|
|
62
71
|
PaginatedSlimDocumentList,
|
63
72
|
ProcessingFailureReasonEnum,
|
64
73
|
ProcessingStateEnum,
|
74
|
+
PromptDeploymentInputRequest,
|
75
|
+
PromptDeploymentInputRequest_ChatHistory,
|
76
|
+
PromptDeploymentInputRequest_Json,
|
77
|
+
PromptDeploymentInputRequest_String,
|
65
78
|
PromptNodeResult,
|
66
79
|
PromptNodeResultData,
|
67
80
|
PromptTemplateBlock,
|
@@ -100,6 +113,8 @@ from .types import (
|
|
100
113
|
SearchResultRequest,
|
101
114
|
SearchWeightsRequest,
|
102
115
|
SlimDocument,
|
116
|
+
StringExecutePromptResponse,
|
117
|
+
StringInputRequest,
|
103
118
|
SubmitCompletionActualRequest,
|
104
119
|
SubmitCompletionActualsErrorResponse,
|
105
120
|
SubmitWorkflowExecutionActualRequest,
|
@@ -229,6 +244,7 @@ __all__ = [
|
|
229
244
|
"ApiNodeResultData",
|
230
245
|
"BadRequestError",
|
231
246
|
"BlockTypeEnum",
|
247
|
+
"ChatHistoryInputRequest",
|
232
248
|
"ChatMessage",
|
233
249
|
"ChatMessageRequest",
|
234
250
|
"ChatMessageRole",
|
@@ -244,8 +260,14 @@ __all__ = [
|
|
244
260
|
"DocumentStatus",
|
245
261
|
"EnrichedNormalizedCompletion",
|
246
262
|
"EnvironmentEnum",
|
263
|
+
"ErrorExecutePromptResponse",
|
247
264
|
"EvaluationParams",
|
248
265
|
"EvaluationParamsRequest",
|
266
|
+
"ExecutePromptApiErrorResponse",
|
267
|
+
"ExecutePromptResponse",
|
268
|
+
"ExecutePromptResponse_Error",
|
269
|
+
"ExecutePromptResponse_Json",
|
270
|
+
"ExecutePromptResponse_String",
|
249
271
|
"ExecuteWorkflowStreamErrorResponse",
|
250
272
|
"FinishReasonEnum",
|
251
273
|
"ForbiddenError",
|
@@ -261,6 +283,8 @@ __all__ = [
|
|
261
283
|
"GenerateStreamResultData",
|
262
284
|
"IndexingStateEnum",
|
263
285
|
"InternalServerError",
|
286
|
+
"JsonExecutePromptResponse",
|
287
|
+
"JsonInputRequest",
|
264
288
|
"LogicalOperator",
|
265
289
|
"LogprobsEnum",
|
266
290
|
"MetadataFilterConfigRequest",
|
@@ -291,6 +315,10 @@ __all__ = [
|
|
291
315
|
"PaginatedSlimDocumentList",
|
292
316
|
"ProcessingFailureReasonEnum",
|
293
317
|
"ProcessingStateEnum",
|
318
|
+
"PromptDeploymentInputRequest",
|
319
|
+
"PromptDeploymentInputRequest_ChatHistory",
|
320
|
+
"PromptDeploymentInputRequest_Json",
|
321
|
+
"PromptDeploymentInputRequest_String",
|
294
322
|
"PromptNodeResult",
|
295
323
|
"PromptNodeResultData",
|
296
324
|
"PromptTemplateBlock",
|
@@ -329,6 +357,8 @@ __all__ = [
|
|
329
357
|
"SearchResultRequest",
|
330
358
|
"SearchWeightsRequest",
|
331
359
|
"SlimDocument",
|
360
|
+
"StringExecutePromptResponse",
|
361
|
+
"StringInputRequest",
|
332
362
|
"SubmitCompletionActualRequest",
|
333
363
|
"SubmitCompletionActualsErrorResponse",
|
334
364
|
"SubmitWorkflowExecutionActualRequest",
|
vellum/client.py
CHANGED
@@ -22,11 +22,12 @@ from .resources.model_versions.client import AsyncModelVersionsClient, ModelVers
|
|
22
22
|
from .resources.registered_prompts.client import AsyncRegisteredPromptsClient, RegisteredPromptsClient
|
23
23
|
from .resources.sandboxes.client import AsyncSandboxesClient, SandboxesClient
|
24
24
|
from .resources.test_suites.client import AsyncTestSuitesClient, TestSuitesClient
|
25
|
-
from .types.
|
25
|
+
from .types.execute_prompt_response import ExecutePromptResponse
|
26
26
|
from .types.generate_options_request import GenerateOptionsRequest
|
27
27
|
from .types.generate_request import GenerateRequest
|
28
28
|
from .types.generate_response import GenerateResponse
|
29
29
|
from .types.generate_stream_response import GenerateStreamResponse
|
30
|
+
from .types.prompt_deployment_input_request import PromptDeploymentInputRequest
|
30
31
|
from .types.search_request_options_request import SearchRequestOptionsRequest
|
31
32
|
from .types.search_response import SearchResponse
|
32
33
|
from .types.submit_completion_actual_request import SubmitCompletionActualRequest
|
@@ -66,6 +67,61 @@ class Vellum:
|
|
66
67
|
self.sandboxes = SandboxesClient(client_wrapper=self._client_wrapper)
|
67
68
|
self.test_suites = TestSuitesClient(client_wrapper=self._client_wrapper)
|
68
69
|
|
70
|
+
def execute_prompt(
|
71
|
+
self,
|
72
|
+
*,
|
73
|
+
inputs: typing.List[PromptDeploymentInputRequest],
|
74
|
+
prompt_deployment_id: typing.Optional[str] = OMIT,
|
75
|
+
prompt_deployment_name: typing.Optional[str] = OMIT,
|
76
|
+
release_tag: typing.Optional[str] = OMIT,
|
77
|
+
external_id: typing.Optional[str] = OMIT,
|
78
|
+
) -> ExecutePromptResponse:
|
79
|
+
"""
|
80
|
+
Executes a deployed Prompt and returns the result.
|
81
|
+
|
82
|
+
Parameters:
|
83
|
+
- inputs: typing.List[PromptDeploymentInputRequest].
|
84
|
+
|
85
|
+
- prompt_deployment_id: typing.Optional[str]. The ID of the Prompt Deployment. Must provide either this or prompt_deployment_name.
|
86
|
+
|
87
|
+
- prompt_deployment_name: typing.Optional[str]. The name of the Prompt Deployment. Must provide either this or prompt_deployment_id.
|
88
|
+
|
89
|
+
- release_tag: typing.Optional[str]. Optionally specify a release tag if you want to pin to a specific release of the Prompt Deployment
|
90
|
+
|
91
|
+
- external_id: typing.Optional[str].
|
92
|
+
"""
|
93
|
+
_request: typing.Dict[str, typing.Any] = {"inputs": inputs}
|
94
|
+
if prompt_deployment_id is not OMIT:
|
95
|
+
_request["prompt_deployment_id"] = prompt_deployment_id
|
96
|
+
if prompt_deployment_name is not OMIT:
|
97
|
+
_request["prompt_deployment_name"] = prompt_deployment_name
|
98
|
+
if release_tag is not OMIT:
|
99
|
+
_request["release_tag"] = release_tag
|
100
|
+
if external_id is not OMIT:
|
101
|
+
_request["external_id"] = external_id
|
102
|
+
_response = self._client_wrapper.httpx_client.request(
|
103
|
+
"POST",
|
104
|
+
urllib.parse.urljoin(f"{self._client_wrapper.get_environment().default}/", "v1/execute-prompt"),
|
105
|
+
json=jsonable_encoder(_request),
|
106
|
+
headers=self._client_wrapper.get_headers(),
|
107
|
+
timeout=None,
|
108
|
+
)
|
109
|
+
if 200 <= _response.status_code < 300:
|
110
|
+
return pydantic.parse_obj_as(ExecutePromptResponse, _response.json()) # type: ignore
|
111
|
+
if _response.status_code == 400:
|
112
|
+
raise BadRequestError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
|
113
|
+
if _response.status_code == 403:
|
114
|
+
raise ForbiddenError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
|
115
|
+
if _response.status_code == 404:
|
116
|
+
raise NotFoundError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
|
117
|
+
if _response.status_code == 500:
|
118
|
+
raise InternalServerError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
|
119
|
+
try:
|
120
|
+
_response_json = _response.json()
|
121
|
+
except JSONDecodeError:
|
122
|
+
raise ApiError(status_code=_response.status_code, body=_response.text)
|
123
|
+
raise ApiError(status_code=_response.status_code, body=_response_json)
|
124
|
+
|
69
125
|
def execute_workflow_stream(
|
70
126
|
self,
|
71
127
|
*,
|
@@ -187,7 +243,7 @@ class Vellum:
|
|
187
243
|
if _response.status_code == 400:
|
188
244
|
raise BadRequestError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
|
189
245
|
if _response.status_code == 403:
|
190
|
-
raise ForbiddenError(pydantic.parse_obj_as(
|
246
|
+
raise ForbiddenError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
|
191
247
|
if _response.status_code == 404:
|
192
248
|
raise NotFoundError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
|
193
249
|
if _response.status_code == 500:
|
@@ -244,7 +300,7 @@ class Vellum:
|
|
244
300
|
if _response.status_code == 400:
|
245
301
|
raise BadRequestError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
|
246
302
|
if _response.status_code == 403:
|
247
|
-
raise ForbiddenError(pydantic.parse_obj_as(
|
303
|
+
raise ForbiddenError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
|
248
304
|
if _response.status_code == 404:
|
249
305
|
raise NotFoundError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
|
250
306
|
if _response.status_code == 500:
|
@@ -432,6 +488,61 @@ class AsyncVellum:
|
|
432
488
|
self.sandboxes = AsyncSandboxesClient(client_wrapper=self._client_wrapper)
|
433
489
|
self.test_suites = AsyncTestSuitesClient(client_wrapper=self._client_wrapper)
|
434
490
|
|
491
|
+
async def execute_prompt(
|
492
|
+
self,
|
493
|
+
*,
|
494
|
+
inputs: typing.List[PromptDeploymentInputRequest],
|
495
|
+
prompt_deployment_id: typing.Optional[str] = OMIT,
|
496
|
+
prompt_deployment_name: typing.Optional[str] = OMIT,
|
497
|
+
release_tag: typing.Optional[str] = OMIT,
|
498
|
+
external_id: typing.Optional[str] = OMIT,
|
499
|
+
) -> ExecutePromptResponse:
|
500
|
+
"""
|
501
|
+
Executes a deployed Prompt and returns the result.
|
502
|
+
|
503
|
+
Parameters:
|
504
|
+
- inputs: typing.List[PromptDeploymentInputRequest].
|
505
|
+
|
506
|
+
- prompt_deployment_id: typing.Optional[str]. The ID of the Prompt Deployment. Must provide either this or prompt_deployment_name.
|
507
|
+
|
508
|
+
- prompt_deployment_name: typing.Optional[str]. The name of the Prompt Deployment. Must provide either this or prompt_deployment_id.
|
509
|
+
|
510
|
+
- release_tag: typing.Optional[str]. Optionally specify a release tag if you want to pin to a specific release of the Prompt Deployment
|
511
|
+
|
512
|
+
- external_id: typing.Optional[str].
|
513
|
+
"""
|
514
|
+
_request: typing.Dict[str, typing.Any] = {"inputs": inputs}
|
515
|
+
if prompt_deployment_id is not OMIT:
|
516
|
+
_request["prompt_deployment_id"] = prompt_deployment_id
|
517
|
+
if prompt_deployment_name is not OMIT:
|
518
|
+
_request["prompt_deployment_name"] = prompt_deployment_name
|
519
|
+
if release_tag is not OMIT:
|
520
|
+
_request["release_tag"] = release_tag
|
521
|
+
if external_id is not OMIT:
|
522
|
+
_request["external_id"] = external_id
|
523
|
+
_response = await self._client_wrapper.httpx_client.request(
|
524
|
+
"POST",
|
525
|
+
urllib.parse.urljoin(f"{self._client_wrapper.get_environment().default}/", "v1/execute-prompt"),
|
526
|
+
json=jsonable_encoder(_request),
|
527
|
+
headers=self._client_wrapper.get_headers(),
|
528
|
+
timeout=None,
|
529
|
+
)
|
530
|
+
if 200 <= _response.status_code < 300:
|
531
|
+
return pydantic.parse_obj_as(ExecutePromptResponse, _response.json()) # type: ignore
|
532
|
+
if _response.status_code == 400:
|
533
|
+
raise BadRequestError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
|
534
|
+
if _response.status_code == 403:
|
535
|
+
raise ForbiddenError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
|
536
|
+
if _response.status_code == 404:
|
537
|
+
raise NotFoundError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
|
538
|
+
if _response.status_code == 500:
|
539
|
+
raise InternalServerError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
|
540
|
+
try:
|
541
|
+
_response_json = _response.json()
|
542
|
+
except JSONDecodeError:
|
543
|
+
raise ApiError(status_code=_response.status_code, body=_response.text)
|
544
|
+
raise ApiError(status_code=_response.status_code, body=_response_json)
|
545
|
+
|
435
546
|
async def execute_workflow_stream(
|
436
547
|
self,
|
437
548
|
*,
|
@@ -553,7 +664,7 @@ class AsyncVellum:
|
|
553
664
|
if _response.status_code == 400:
|
554
665
|
raise BadRequestError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
|
555
666
|
if _response.status_code == 403:
|
556
|
-
raise ForbiddenError(pydantic.parse_obj_as(
|
667
|
+
raise ForbiddenError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
|
557
668
|
if _response.status_code == 404:
|
558
669
|
raise NotFoundError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
|
559
670
|
if _response.status_code == 500:
|
@@ -610,7 +721,7 @@ class AsyncVellum:
|
|
610
721
|
if _response.status_code == 400:
|
611
722
|
raise BadRequestError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
|
612
723
|
if _response.status_code == 403:
|
613
|
-
raise ForbiddenError(pydantic.parse_obj_as(
|
724
|
+
raise ForbiddenError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
|
614
725
|
if _response.status_code == 404:
|
615
726
|
raise NotFoundError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
|
616
727
|
if _response.status_code == 500:
|
vellum/core/client_wrapper.py
CHANGED
@@ -16,7 +16,7 @@ class BaseClientWrapper:
|
|
16
16
|
headers: typing.Dict[str, str] = {
|
17
17
|
"X-Fern-Language": "Python",
|
18
18
|
"X-Fern-SDK-Name": "vellum-ai",
|
19
|
-
"X-Fern-SDK-Version": "v0.1.
|
19
|
+
"X-Fern-SDK-Version": "v0.1.5",
|
20
20
|
}
|
21
21
|
headers["X_API_KEY"] = self.api_key
|
22
22
|
return headers
|
vellum/errors/forbidden_error.py
CHANGED
@@ -1,9 +1,10 @@
|
|
1
1
|
# This file was auto-generated by Fern from our API Definition.
|
2
2
|
|
3
|
+
import typing
|
4
|
+
|
3
5
|
from ..core.api_error import ApiError
|
4
|
-
from ..types.generate_error_response import GenerateErrorResponse
|
5
6
|
|
6
7
|
|
7
8
|
class ForbiddenError(ApiError):
|
8
|
-
def __init__(self, body:
|
9
|
+
def __init__(self, body: typing.Any):
|
9
10
|
super().__init__(status_code=403, body=body)
|
vellum/types/__init__.py
CHANGED
@@ -3,6 +3,7 @@
|
|
3
3
|
from .api_node_result import ApiNodeResult
|
4
4
|
from .api_node_result_data import ApiNodeResultData
|
5
5
|
from .block_type_enum import BlockTypeEnum
|
6
|
+
from .chat_history_input_request import ChatHistoryInputRequest
|
6
7
|
from .chat_message import ChatMessage
|
7
8
|
from .chat_message_request import ChatMessageRequest
|
8
9
|
from .chat_message_role import ChatMessageRole
|
@@ -17,8 +18,16 @@ from .document_read import DocumentRead
|
|
17
18
|
from .document_status import DocumentStatus
|
18
19
|
from .enriched_normalized_completion import EnrichedNormalizedCompletion
|
19
20
|
from .environment_enum import EnvironmentEnum
|
21
|
+
from .error_execute_prompt_response import ErrorExecutePromptResponse
|
20
22
|
from .evaluation_params import EvaluationParams
|
21
23
|
from .evaluation_params_request import EvaluationParamsRequest
|
24
|
+
from .execute_prompt_api_error_response import ExecutePromptApiErrorResponse
|
25
|
+
from .execute_prompt_response import (
|
26
|
+
ExecutePromptResponse,
|
27
|
+
ExecutePromptResponse_Error,
|
28
|
+
ExecutePromptResponse_Json,
|
29
|
+
ExecutePromptResponse_String,
|
30
|
+
)
|
22
31
|
from .execute_workflow_stream_error_response import ExecuteWorkflowStreamErrorResponse
|
23
32
|
from .finish_reason_enum import FinishReasonEnum
|
24
33
|
from .generate_error_response import GenerateErrorResponse
|
@@ -32,6 +41,8 @@ from .generate_stream_response import GenerateStreamResponse
|
|
32
41
|
from .generate_stream_result import GenerateStreamResult
|
33
42
|
from .generate_stream_result_data import GenerateStreamResultData
|
34
43
|
from .indexing_state_enum import IndexingStateEnum
|
44
|
+
from .json_execute_prompt_response import JsonExecutePromptResponse
|
45
|
+
from .json_input_request import JsonInputRequest
|
35
46
|
from .logical_operator import LogicalOperator
|
36
47
|
from .logprobs_enum import LogprobsEnum
|
37
48
|
from .metadata_filter_config_request import MetadataFilterConfigRequest
|
@@ -63,6 +74,12 @@ from .normalized_token_log_probs import NormalizedTokenLogProbs
|
|
63
74
|
from .paginated_slim_document_list import PaginatedSlimDocumentList
|
64
75
|
from .processing_failure_reason_enum import ProcessingFailureReasonEnum
|
65
76
|
from .processing_state_enum import ProcessingStateEnum
|
77
|
+
from .prompt_deployment_input_request import (
|
78
|
+
PromptDeploymentInputRequest,
|
79
|
+
PromptDeploymentInputRequest_ChatHistory,
|
80
|
+
PromptDeploymentInputRequest_Json,
|
81
|
+
PromptDeploymentInputRequest_String,
|
82
|
+
)
|
66
83
|
from .prompt_node_result import PromptNodeResult
|
67
84
|
from .prompt_node_result_data import PromptNodeResultData
|
68
85
|
from .prompt_template_block import PromptTemplateBlock
|
@@ -101,6 +118,8 @@ from .search_result_merging_request import SearchResultMergingRequest
|
|
101
118
|
from .search_result_request import SearchResultRequest
|
102
119
|
from .search_weights_request import SearchWeightsRequest
|
103
120
|
from .slim_document import SlimDocument
|
121
|
+
from .string_execute_prompt_response import StringExecutePromptResponse
|
122
|
+
from .string_input_request import StringInputRequest
|
104
123
|
from .submit_completion_actual_request import SubmitCompletionActualRequest
|
105
124
|
from .submit_completion_actuals_error_response import SubmitCompletionActualsErrorResponse
|
106
125
|
from .submit_workflow_execution_actual_request import (
|
@@ -231,6 +250,7 @@ __all__ = [
|
|
231
250
|
"ApiNodeResult",
|
232
251
|
"ApiNodeResultData",
|
233
252
|
"BlockTypeEnum",
|
253
|
+
"ChatHistoryInputRequest",
|
234
254
|
"ChatMessage",
|
235
255
|
"ChatMessageRequest",
|
236
256
|
"ChatMessageRole",
|
@@ -245,8 +265,14 @@ __all__ = [
|
|
245
265
|
"DocumentStatus",
|
246
266
|
"EnrichedNormalizedCompletion",
|
247
267
|
"EnvironmentEnum",
|
268
|
+
"ErrorExecutePromptResponse",
|
248
269
|
"EvaluationParams",
|
249
270
|
"EvaluationParamsRequest",
|
271
|
+
"ExecutePromptApiErrorResponse",
|
272
|
+
"ExecutePromptResponse",
|
273
|
+
"ExecutePromptResponse_Error",
|
274
|
+
"ExecutePromptResponse_Json",
|
275
|
+
"ExecutePromptResponse_String",
|
250
276
|
"ExecuteWorkflowStreamErrorResponse",
|
251
277
|
"FinishReasonEnum",
|
252
278
|
"GenerateErrorResponse",
|
@@ -260,6 +286,8 @@ __all__ = [
|
|
260
286
|
"GenerateStreamResult",
|
261
287
|
"GenerateStreamResultData",
|
262
288
|
"IndexingStateEnum",
|
289
|
+
"JsonExecutePromptResponse",
|
290
|
+
"JsonInputRequest",
|
263
291
|
"LogicalOperator",
|
264
292
|
"LogprobsEnum",
|
265
293
|
"MetadataFilterConfigRequest",
|
@@ -289,6 +317,10 @@ __all__ = [
|
|
289
317
|
"PaginatedSlimDocumentList",
|
290
318
|
"ProcessingFailureReasonEnum",
|
291
319
|
"ProcessingStateEnum",
|
320
|
+
"PromptDeploymentInputRequest",
|
321
|
+
"PromptDeploymentInputRequest_ChatHistory",
|
322
|
+
"PromptDeploymentInputRequest_Json",
|
323
|
+
"PromptDeploymentInputRequest_String",
|
292
324
|
"PromptNodeResult",
|
293
325
|
"PromptNodeResultData",
|
294
326
|
"PromptTemplateBlock",
|
@@ -327,6 +359,8 @@ __all__ = [
|
|
327
359
|
"SearchResultRequest",
|
328
360
|
"SearchWeightsRequest",
|
329
361
|
"SlimDocument",
|
362
|
+
"StringExecutePromptResponse",
|
363
|
+
"StringInputRequest",
|
330
364
|
"SubmitCompletionActualRequest",
|
331
365
|
"SubmitCompletionActualsErrorResponse",
|
332
366
|
"SubmitWorkflowExecutionActualRequest",
|
@@ -0,0 +1,30 @@
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
2
|
+
|
3
|
+
import datetime as dt
|
4
|
+
import typing
|
5
|
+
|
6
|
+
from ..core.datetime_utils import serialize_datetime
|
7
|
+
from .chat_message_request import ChatMessageRequest
|
8
|
+
|
9
|
+
try:
|
10
|
+
import pydantic.v1 as pydantic # type: ignore
|
11
|
+
except ImportError:
|
12
|
+
import pydantic # type: ignore
|
13
|
+
|
14
|
+
|
15
|
+
class ChatHistoryInputRequest(pydantic.BaseModel):
|
16
|
+
name: str = pydantic.Field(description="The variable's name, as defined in the deployment.")
|
17
|
+
value: typing.List[ChatMessageRequest]
|
18
|
+
|
19
|
+
def json(self, **kwargs: typing.Any) -> str:
|
20
|
+
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
21
|
+
return super().json(**kwargs_with_defaults)
|
22
|
+
|
23
|
+
def dict(self, **kwargs: typing.Any) -> typing.Dict[str, typing.Any]:
|
24
|
+
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
25
|
+
return super().dict(**kwargs_with_defaults)
|
26
|
+
|
27
|
+
class Config:
|
28
|
+
frozen = True
|
29
|
+
smart_union = True
|
30
|
+
json_encoders = {dt.datetime: serialize_datetime}
|
@@ -0,0 +1,30 @@
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
2
|
+
|
3
|
+
import datetime as dt
|
4
|
+
import typing
|
5
|
+
|
6
|
+
from ..core.datetime_utils import serialize_datetime
|
7
|
+
from .vellum_error import VellumError
|
8
|
+
|
9
|
+
try:
|
10
|
+
import pydantic.v1 as pydantic # type: ignore
|
11
|
+
except ImportError:
|
12
|
+
import pydantic # type: ignore
|
13
|
+
|
14
|
+
|
15
|
+
class ErrorExecutePromptResponse(pydantic.BaseModel):
|
16
|
+
value: VellumError
|
17
|
+
execution_id: str
|
18
|
+
|
19
|
+
def json(self, **kwargs: typing.Any) -> str:
|
20
|
+
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
21
|
+
return super().json(**kwargs_with_defaults)
|
22
|
+
|
23
|
+
def dict(self, **kwargs: typing.Any) -> typing.Dict[str, typing.Any]:
|
24
|
+
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
25
|
+
return super().dict(**kwargs_with_defaults)
|
26
|
+
|
27
|
+
class Config:
|
28
|
+
frozen = True
|
29
|
+
smart_union = True
|
30
|
+
json_encoders = {dt.datetime: serialize_datetime}
|
@@ -0,0 +1,28 @@
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
2
|
+
|
3
|
+
import datetime as dt
|
4
|
+
import typing
|
5
|
+
|
6
|
+
from ..core.datetime_utils import serialize_datetime
|
7
|
+
|
8
|
+
try:
|
9
|
+
import pydantic.v1 as pydantic # type: ignore
|
10
|
+
except ImportError:
|
11
|
+
import pydantic # type: ignore
|
12
|
+
|
13
|
+
|
14
|
+
class ExecutePromptApiErrorResponse(pydantic.BaseModel):
|
15
|
+
detail: str = pydantic.Field(description="Details about why the request failed.")
|
16
|
+
|
17
|
+
def json(self, **kwargs: typing.Any) -> str:
|
18
|
+
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
19
|
+
return super().json(**kwargs_with_defaults)
|
20
|
+
|
21
|
+
def dict(self, **kwargs: typing.Any) -> typing.Dict[str, typing.Any]:
|
22
|
+
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
23
|
+
return super().dict(**kwargs_with_defaults)
|
24
|
+
|
25
|
+
class Config:
|
26
|
+
frozen = True
|
27
|
+
smart_union = True
|
28
|
+
json_encoders = {dt.datetime: serialize_datetime}
|
@@ -0,0 +1,43 @@
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
2
|
+
|
3
|
+
from __future__ import annotations
|
4
|
+
|
5
|
+
import typing
|
6
|
+
|
7
|
+
import typing_extensions
|
8
|
+
|
9
|
+
from .error_execute_prompt_response import ErrorExecutePromptResponse
|
10
|
+
from .json_execute_prompt_response import JsonExecutePromptResponse
|
11
|
+
from .string_execute_prompt_response import StringExecutePromptResponse
|
12
|
+
|
13
|
+
|
14
|
+
class ExecutePromptResponse_Error(ErrorExecutePromptResponse):
|
15
|
+
type: typing_extensions.Literal["ERROR"]
|
16
|
+
|
17
|
+
class Config:
|
18
|
+
frozen = True
|
19
|
+
smart_union = True
|
20
|
+
allow_population_by_field_name = True
|
21
|
+
|
22
|
+
|
23
|
+
class ExecutePromptResponse_Json(JsonExecutePromptResponse):
|
24
|
+
type: typing_extensions.Literal["JSON"]
|
25
|
+
|
26
|
+
class Config:
|
27
|
+
frozen = True
|
28
|
+
smart_union = True
|
29
|
+
allow_population_by_field_name = True
|
30
|
+
|
31
|
+
|
32
|
+
class ExecutePromptResponse_String(StringExecutePromptResponse):
|
33
|
+
type: typing_extensions.Literal["STRING"]
|
34
|
+
|
35
|
+
class Config:
|
36
|
+
frozen = True
|
37
|
+
smart_union = True
|
38
|
+
allow_population_by_field_name = True
|
39
|
+
|
40
|
+
|
41
|
+
ExecutePromptResponse = typing.Union[
|
42
|
+
ExecutePromptResponse_Error, ExecutePromptResponse_Json, ExecutePromptResponse_String
|
43
|
+
]
|
@@ -0,0 +1,29 @@
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
2
|
+
|
3
|
+
import datetime as dt
|
4
|
+
import typing
|
5
|
+
|
6
|
+
from ..core.datetime_utils import serialize_datetime
|
7
|
+
|
8
|
+
try:
|
9
|
+
import pydantic.v1 as pydantic # type: ignore
|
10
|
+
except ImportError:
|
11
|
+
import pydantic # type: ignore
|
12
|
+
|
13
|
+
|
14
|
+
class JsonExecutePromptResponse(pydantic.BaseModel):
|
15
|
+
value: typing.Dict[str, typing.Any]
|
16
|
+
execution_id: str
|
17
|
+
|
18
|
+
def json(self, **kwargs: typing.Any) -> str:
|
19
|
+
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
20
|
+
return super().json(**kwargs_with_defaults)
|
21
|
+
|
22
|
+
def dict(self, **kwargs: typing.Any) -> typing.Dict[str, typing.Any]:
|
23
|
+
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
24
|
+
return super().dict(**kwargs_with_defaults)
|
25
|
+
|
26
|
+
class Config:
|
27
|
+
frozen = True
|
28
|
+
smart_union = True
|
29
|
+
json_encoders = {dt.datetime: serialize_datetime}
|
@@ -0,0 +1,29 @@
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
2
|
+
|
3
|
+
import datetime as dt
|
4
|
+
import typing
|
5
|
+
|
6
|
+
from ..core.datetime_utils import serialize_datetime
|
7
|
+
|
8
|
+
try:
|
9
|
+
import pydantic.v1 as pydantic # type: ignore
|
10
|
+
except ImportError:
|
11
|
+
import pydantic # type: ignore
|
12
|
+
|
13
|
+
|
14
|
+
class JsonInputRequest(pydantic.BaseModel):
|
15
|
+
name: str = pydantic.Field(description="The variable's name, as defined in the deployment.")
|
16
|
+
value: typing.Dict[str, typing.Any]
|
17
|
+
|
18
|
+
def json(self, **kwargs: typing.Any) -> str:
|
19
|
+
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
20
|
+
return super().json(**kwargs_with_defaults)
|
21
|
+
|
22
|
+
def dict(self, **kwargs: typing.Any) -> typing.Dict[str, typing.Any]:
|
23
|
+
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
24
|
+
return super().dict(**kwargs_with_defaults)
|
25
|
+
|
26
|
+
class Config:
|
27
|
+
frozen = True
|
28
|
+
smart_union = True
|
29
|
+
json_encoders = {dt.datetime: serialize_datetime}
|
@@ -0,0 +1,43 @@
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
2
|
+
|
3
|
+
from __future__ import annotations
|
4
|
+
|
5
|
+
import typing
|
6
|
+
|
7
|
+
import typing_extensions
|
8
|
+
|
9
|
+
from .chat_history_input_request import ChatHistoryInputRequest
|
10
|
+
from .json_input_request import JsonInputRequest
|
11
|
+
from .string_input_request import StringInputRequest
|
12
|
+
|
13
|
+
|
14
|
+
class PromptDeploymentInputRequest_String(StringInputRequest):
|
15
|
+
type: typing_extensions.Literal["STRING"]
|
16
|
+
|
17
|
+
class Config:
|
18
|
+
frozen = True
|
19
|
+
smart_union = True
|
20
|
+
allow_population_by_field_name = True
|
21
|
+
|
22
|
+
|
23
|
+
class PromptDeploymentInputRequest_Json(JsonInputRequest):
|
24
|
+
type: typing_extensions.Literal["JSON"]
|
25
|
+
|
26
|
+
class Config:
|
27
|
+
frozen = True
|
28
|
+
smart_union = True
|
29
|
+
allow_population_by_field_name = True
|
30
|
+
|
31
|
+
|
32
|
+
class PromptDeploymentInputRequest_ChatHistory(ChatHistoryInputRequest):
|
33
|
+
type: typing_extensions.Literal["CHAT_HISTORY"]
|
34
|
+
|
35
|
+
class Config:
|
36
|
+
frozen = True
|
37
|
+
smart_union = True
|
38
|
+
allow_population_by_field_name = True
|
39
|
+
|
40
|
+
|
41
|
+
PromptDeploymentInputRequest = typing.Union[
|
42
|
+
PromptDeploymentInputRequest_String, PromptDeploymentInputRequest_Json, PromptDeploymentInputRequest_ChatHistory
|
43
|
+
]
|
@@ -0,0 +1,29 @@
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
2
|
+
|
3
|
+
import datetime as dt
|
4
|
+
import typing
|
5
|
+
|
6
|
+
from ..core.datetime_utils import serialize_datetime
|
7
|
+
|
8
|
+
try:
|
9
|
+
import pydantic.v1 as pydantic # type: ignore
|
10
|
+
except ImportError:
|
11
|
+
import pydantic # type: ignore
|
12
|
+
|
13
|
+
|
14
|
+
class StringExecutePromptResponse(pydantic.BaseModel):
|
15
|
+
value: str
|
16
|
+
execution_id: str
|
17
|
+
|
18
|
+
def json(self, **kwargs: typing.Any) -> str:
|
19
|
+
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
20
|
+
return super().json(**kwargs_with_defaults)
|
21
|
+
|
22
|
+
def dict(self, **kwargs: typing.Any) -> typing.Dict[str, typing.Any]:
|
23
|
+
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
24
|
+
return super().dict(**kwargs_with_defaults)
|
25
|
+
|
26
|
+
class Config:
|
27
|
+
frozen = True
|
28
|
+
smart_union = True
|
29
|
+
json_encoders = {dt.datetime: serialize_datetime}
|
@@ -0,0 +1,29 @@
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
2
|
+
|
3
|
+
import datetime as dt
|
4
|
+
import typing
|
5
|
+
|
6
|
+
from ..core.datetime_utils import serialize_datetime
|
7
|
+
|
8
|
+
try:
|
9
|
+
import pydantic.v1 as pydantic # type: ignore
|
10
|
+
except ImportError:
|
11
|
+
import pydantic # type: ignore
|
12
|
+
|
13
|
+
|
14
|
+
class StringInputRequest(pydantic.BaseModel):
|
15
|
+
name: str = pydantic.Field(description="The variable's name, as defined in the deployment.")
|
16
|
+
value: str
|
17
|
+
|
18
|
+
def json(self, **kwargs: typing.Any) -> str:
|
19
|
+
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
20
|
+
return super().json(**kwargs_with_defaults)
|
21
|
+
|
22
|
+
def dict(self, **kwargs: typing.Any) -> typing.Dict[str, typing.Any]:
|
23
|
+
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
24
|
+
return super().dict(**kwargs_with_defaults)
|
25
|
+
|
26
|
+
class Config:
|
27
|
+
frozen = True
|
28
|
+
smart_union = True
|
29
|
+
json_encoders = {dt.datetime: serialize_datetime}
|
@@ -1,8 +1,8 @@
|
|
1
|
-
vellum/__init__.py,sha256=
|
2
|
-
vellum/client.py,sha256=
|
1
|
+
vellum/__init__.py,sha256=G549OSXyIsjzL4hNZ1xFsFSA19Qznr5CzviTS1TRT_o,15911
|
2
|
+
vellum/client.py,sha256=VyCnmohtSxlZBzzR-Tpg9ApsFLh6xlcVCnKyLaQT_q4,43212
|
3
3
|
vellum/core/__init__.py,sha256=QJS3CJ2TYP2E1Tge0CS6Z7r8LTNzJHQVX1hD3558eP0,519
|
4
4
|
vellum/core/api_error.py,sha256=RE8LELok2QCjABadECTvtDp7qejA1VmINCh6TbqPwSE,426
|
5
|
-
vellum/core/client_wrapper.py,sha256=
|
5
|
+
vellum/core/client_wrapper.py,sha256=_edGK6g1DZbPsbUeNrzqmKJdj6IgcwSqadtDD5JVKWI,1213
|
6
6
|
vellum/core/datetime_utils.py,sha256=nBys2IsYrhPdszxGKCNRPSOCwa-5DWOHG95FB8G9PKo,1047
|
7
7
|
vellum/core/jsonable_encoder.py,sha256=MTYkDov2EryHgee4QM46uZiBOuOXK9KTHlBdBwU-CpU,3799
|
8
8
|
vellum/core/remove_none_from_dict.py,sha256=8m91FC3YuVem0Gm9_sXhJ2tGvP33owJJdrqCLEdowGw,330
|
@@ -10,7 +10,7 @@ vellum/environment.py,sha256=bcAFjoE9XXd7tiysYS90Os669IJmUMZS2JZ_ZQn0Dpg,498
|
|
10
10
|
vellum/errors/__init__.py,sha256=57E7ykyjmPTSDm-CJKS6exWPndqZBUVslrmzLbtQUYg,402
|
11
11
|
vellum/errors/bad_request_error.py,sha256=jvBFMSUKdTc6ya16S5O7zwaFnPhOm8N4DjIom_pw5pc,248
|
12
12
|
vellum/errors/conflict_error.py,sha256=DFxSC0nkPmb6_3lR018RksrzHH5-LOStlLnTetamGfs,327
|
13
|
-
vellum/errors/forbidden_error.py,sha256=
|
13
|
+
vellum/errors/forbidden_error.py,sha256=dgnatOGair3CvxljCE45_qwN_yefzcw2G0vw88wrrOA,247
|
14
14
|
vellum/errors/internal_server_error.py,sha256=E0rgqJC0-LcetLi1HmSi92KpvNkGSRCIdBeEqT_ln1s,252
|
15
15
|
vellum/errors/not_found_error.py,sha256=P65k-Lm2RuefAVSNLER5hH-4P99SGohKy2cOPSrIxNk,246
|
16
16
|
vellum/py.typed,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
@@ -29,10 +29,11 @@ vellum/resources/sandboxes/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roe
|
|
29
29
|
vellum/resources/sandboxes/client.py,sha256=RkMHXuNGblYEMDdbUeFHzJDGEIxJd9HGiThXUbR1MNg,8201
|
30
30
|
vellum/resources/test_suites/__init__.py,sha256=FTtvy8EDg9nNNg9WCatVgKTRYV8-_v1roeGPAKoa_pw,65
|
31
31
|
vellum/resources/test_suites/client.py,sha256=TW8piGrmeEqYmMeWtuOyj2lH5_erQ1WVMmDcCOuGdG8,8586
|
32
|
-
vellum/types/__init__.py,sha256=
|
32
|
+
vellum/types/__init__.py,sha256=BdaLv4JlnA9fVetO8ZgcwFqC1GZW2jfbdhjNxno-NRY,21535
|
33
33
|
vellum/types/api_node_result.py,sha256=1zAcNznjyVkTQM2V--UE1lYFzRWbYEogoChZBGcsliE,936
|
34
34
|
vellum/types/api_node_result_data.py,sha256=HvpZaAKYXsoBOnobACIYCmIdxbRc7Zp-ibIohiz_Nzc,1125
|
35
35
|
vellum/types/block_type_enum.py,sha256=BQ8ZN1OKhOdmHeQ3-bhMfkLLohirsNhHPljBJX8Bbmo,1051
|
36
|
+
vellum/types/chat_history_input_request.py,sha256=zZ3_qj0vPCBcAc5fMaxJjGibGCqYXv6yeSQy6VCHIc0,1059
|
36
37
|
vellum/types/chat_message.py,sha256=dvII2EkChDp2zuPnSU5ycmqm2SMGoqttGha00HC9ETs,941
|
37
38
|
vellum/types/chat_message_request.py,sha256=9wgnG0Q36jUNTr7vS0FASRFrsXzgIccrjNiC7M_PwkM,948
|
38
39
|
vellum/types/chat_message_role.py,sha256=VscYXloHNYjiThf7g6OsMxO0S8p0ZUa0G6sakaWhc1g,914
|
@@ -47,8 +48,11 @@ vellum/types/document_read.py,sha256=aeAwY8-Si0nT2Ifh1J9ChLaNFJaker1jlfbJkuoA-xc
|
|
47
48
|
vellum/types/document_status.py,sha256=dkotle2rykOxF1QxqcM8af2ORzmFjZR5a7pHXP6SjDQ,144
|
48
49
|
vellum/types/enriched_normalized_completion.py,sha256=z3MzEF7T-JcyTQRNlZlnB9GpFcLFo0gy-hF3LYW7Pfg,2135
|
49
50
|
vellum/types/environment_enum.py,sha256=16gr0xkzNIO5dhrOnJNjcQ34z9Xk1oE-egW17P0RWDs,799
|
51
|
+
vellum/types/error_execute_prompt_response.py,sha256=EJpJ5llClytDgVx5BaAKuoMySHwQ_YNQrSN3jR-A3xs,952
|
50
52
|
vellum/types/evaluation_params.py,sha256=Ey-RgV4uKVv8G4Dj6lmX_qiBsLcPo1NmoY2xcFW_qDQ,1048
|
51
53
|
vellum/types/evaluation_params_request.py,sha256=izDbrQHT7QFYjxXX9bkeZnoCle3teShaodbnV_11sJo,1055
|
54
|
+
vellum/types/execute_prompt_api_error_response.py,sha256=f0peA0yIyF1dqh2PclsVEaiCCoLK73KEwYq32cj82Nw,958
|
55
|
+
vellum/types/execute_prompt_response.py,sha256=XjY80wf3-_zZQp4BeEP4OzedkEPzhLCr3pSJZ3vv-qo,1156
|
52
56
|
vellum/types/execute_workflow_stream_error_response.py,sha256=WHSu333maVV2Nbxs43QCxlW6wjbkPdm-97DP1tzglAA,963
|
53
57
|
vellum/types/finish_reason_enum.py,sha256=JawC1tsGIUCwfGoIjrnD2c4wmbsO5Dx39WPAbwYIWm0,726
|
54
58
|
vellum/types/generate_error_response.py,sha256=auz4NPbyceabCe0eN89lavEwotWbqsTSbkJtI_lgXJs,950
|
@@ -62,6 +66,8 @@ vellum/types/generate_stream_response.py,sha256=mbaw11ScmPmSgih02qE2pLDN0fjILc3L
|
|
62
66
|
vellum/types/generate_stream_result.py,sha256=NevSf70_mbODDRPnauYobl41PNbPVcNKyWTe2gg-lxI,1107
|
63
67
|
vellum/types/generate_stream_result_data.py,sha256=lyAvxY714Zo5W1i4WkuIbKjyP3rLGJAQ_pL2jkDL7Pg,1011
|
64
68
|
vellum/types/indexing_state_enum.py,sha256=7eEvjuZi3dy6XJIYhHgca4XfRDqntaoYTFOOl9AXe78,1181
|
69
|
+
vellum/types/json_execute_prompt_response.py,sha256=ESvK2oJfNop3VWotbW7LHR2djcJjoIan70X-11Ivnoo,930
|
70
|
+
vellum/types/json_input_request.py,sha256=x3aGIGpjqG8aPxRWfneeWalR2mOdQJKlviF72URrnMg,996
|
65
71
|
vellum/types/logical_operator.py,sha256=4AVyWKNKyL8NxaqUKdS8hV7nfRpUlItKDNdC_hkJU9E,4282
|
66
72
|
vellum/types/logprobs_enum.py,sha256=v3VTlzw1bpuU0hBhcWJTN7JMBS2nWqwertJlaCgxmuY,490
|
67
73
|
vellum/types/metadata_filter_config_request.py,sha256=tUkflYOkwBYmouX_9OfO2mJixeylSns4gHoDuKFgpmo,1333
|
@@ -85,6 +91,7 @@ vellum/types/normalized_token_log_probs.py,sha256=ZIEr6evXULAv5GFhkekr1Qc6QsJjr2
|
|
85
91
|
vellum/types/paginated_slim_document_list.py,sha256=rfm_k539tWn6jVBjtgUG1M0AqtPvxtwPXwBJLPBiE6Q,1062
|
86
92
|
vellum/types/processing_failure_reason_enum.py,sha256=YxYs-fZd957LUlO0tNRY-jSNkki8quIJX8ElzNhEnlw,752
|
87
93
|
vellum/types/processing_state_enum.py,sha256=c3eTs4vo-7053YYXL6-n74qPCYOsxgunY15qgFMl3X0,962
|
94
|
+
vellum/types/prompt_deployment_input_request.py,sha256=pwI4mGEz88s8gJr1j_5-4t501tAVwsFNZrEVfPYWLmk,1159
|
88
95
|
vellum/types/prompt_node_result.py,sha256=TxFRE6-2LFrG1Mo9uA7dOfNs8_Jk6u7bOWhmOzYQiJY,948
|
89
96
|
vellum/types/prompt_node_result_data.py,sha256=LyNkZz-9lYejVtIaej1m7u5CiWmXVFG8IAr6hVQBgZg,945
|
90
97
|
vellum/types/prompt_template_block.py,sha256=2UfBd8ktRrvVp_gZFLWevrV5D1UjXCji4YbeHPAVl0c,1164
|
@@ -123,6 +130,8 @@ vellum/types/search_result_merging_request.py,sha256=lLgp8q1WwvdUJcR1YWpLRjmZ3Sg
|
|
123
130
|
vellum/types/search_result_request.py,sha256=SMjICHnxr3l8-KCe7ftWMZiKPdlvdMAPZd4zRXNxpQ8,1335
|
124
131
|
vellum/types/search_weights_request.py,sha256=guPgN3N0b5Mg0sIpw8ts80iBOT_nUsRNZqjDMKJIphI,1117
|
125
132
|
vellum/types/slim_document.py,sha256=-wzURGcLcnoh5vdTlbtkPLKfP3yAShdkkK6AIIi6W48,3074
|
133
|
+
vellum/types/string_execute_prompt_response.py,sha256=qsOhsJYzbZ4IjcsrNW8qSF6B0HAkWRJsRKi7xc3k_sk,907
|
134
|
+
vellum/types/string_input_request.py,sha256=aUzb_zuLK9G7tHBRHVWCJD4o-UbcVcMVtIrgAoES8dw,973
|
126
135
|
vellum/types/submit_completion_actual_request.py,sha256=krKZefK_-0LAEJYIKilUemEwc6v0RELpusQvyj1XUJQ,1787
|
127
136
|
vellum/types/submit_completion_actuals_error_response.py,sha256=AJTkBM60F8rxGKXfz0TcjdP5v_OgRnF3zUrDG1CtNDo,895
|
128
137
|
vellum/types/submit_workflow_execution_actual_request.py,sha256=hazWw_rBxap8g0QAqlrO-qYFu3gJV_XXVgquWkDVsi0,1413
|
@@ -190,6 +199,6 @@ vellum/types/workflow_result_event_output_data_number.py,sha256=zWtQor27iaDDHmY5
|
|
190
199
|
vellum/types/workflow_result_event_output_data_search_results.py,sha256=frCaJ5kWrIqCeV-waBNfd7rO4fqWe5aYpSI8PM4-oRw,1323
|
191
200
|
vellum/types/workflow_result_event_output_data_string.py,sha256=TByZxyQh9ci4UIdEmoEi_JK1U_JwYCnVZeB_4kGuXKM,1405
|
192
201
|
vellum/types/workflow_stream_event.py,sha256=OQUSzwoM-OCfWxNzeOVVLsjCue_WWqin3tGMtwvp_rc,873
|
193
|
-
vellum_ai-0.1.
|
194
|
-
vellum_ai-0.1.
|
195
|
-
vellum_ai-0.1.
|
202
|
+
vellum_ai-0.1.5.dist-info/METADATA,sha256=vd81SnK9c0nlCZ6NM7hpZM_lBSXJ8aLXsalahlewW5w,3486
|
203
|
+
vellum_ai-0.1.5.dist-info/WHEEL,sha256=Zb28QaM1gQi8f4VCBhsUklF61CTlNYfs9YAZn-TOGFk,88
|
204
|
+
vellum_ai-0.1.5.dist-info/RECORD,,
|
File without changes
|