vellum-ai 0.8.5__py3-none-any.whl → 0.8.6__py3-none-any.whl
Sign up to get free protection for your applications and to get access to all the features.
- vellum/__init__.py +64 -104
- vellum/client.py +17 -17
- vellum/core/client_wrapper.py +1 -1
- vellum/resources/__init__.py +2 -2
- vellum/resources/ad_hoc/client.py +372 -0
- vellum/types/__init__.py +62 -102
- vellum/types/ad_hoc_execute_prompt_event.py +14 -0
- vellum/types/ad_hoc_expand_meta_request.py +33 -0
- vellum/types/ad_hoc_fulfilled_prompt_execution_meta.py +27 -0
- vellum/types/{open_api_const_property_request.py → ad_hoc_initiated_prompt_execution_meta.py} +4 -6
- vellum/types/{open_api_boolean_property_request.py → ad_hoc_rejected_prompt_execution_meta.py} +5 -6
- vellum/types/{tik_token_tokenizer_config_request.py → ad_hoc_streaming_prompt_execution_meta.py} +3 -4
- vellum/types/{open_api_one_of_property.py → chat_message_prompt_block_properties_request.py} +10 -9
- vellum/types/chat_message_prompt_block_request.py +36 -0
- vellum/types/{ml_model_developer_enum_value_label.py → ephemeral_prompt_cache_config_request.py} +4 -5
- vellum/types/ephemeral_prompt_cache_config_type_enum.py +5 -0
- vellum/types/fulfilled_ad_hoc_execute_prompt_event.py +28 -0
- vellum/types/function_definition_prompt_block_properties_request.py +42 -0
- vellum/types/function_definition_prompt_block_request.py +30 -0
- vellum/types/{hugging_face_tokenizer_config.py → initiated_ad_hoc_execute_prompt_event.py} +6 -5
- vellum/types/{ml_model_display_tag_enum_value_label.py → jinja_prompt_block_properties_request.py} +5 -5
- vellum/types/jinja_prompt_block_request.py +30 -0
- vellum/types/ml_model_usage.py +2 -0
- vellum/types/plain_text_prompt_block_request.py +29 -0
- vellum/types/prompt_block_request.py +19 -0
- vellum/types/prompt_block_state.py +5 -0
- vellum/types/{prompt_deployment_expand_meta_request_request.py → prompt_deployment_expand_meta_request.py} +1 -1
- vellum/types/prompt_parameters_request.py +27 -0
- vellum/types/{open_api_boolean_property.py → prompt_request_chat_history_input_request.py} +7 -7
- vellum/types/prompt_request_input_request.py +10 -0
- vellum/types/{hugging_face_tokenizer_config_request.py → prompt_request_json_input_request.py} +6 -6
- vellum/types/{tik_token_tokenizer_config.py → prompt_request_string_input_request.py} +6 -5
- vellum/types/rejected_ad_hoc_execute_prompt_event.py +28 -0
- vellum/types/rich_text_child_block_request.py +7 -0
- vellum/types/rich_text_prompt_block_request.py +30 -0
- vellum/types/streaming_ad_hoc_execute_prompt_event.py +33 -0
- vellum/types/variable_prompt_block_request.py +29 -0
- vellum/types/{ml_model_family_enum_value_label.py → vellum_variable_request.py} +5 -4
- {vellum_ai-0.8.5.dist-info → vellum_ai-0.8.6.dist-info}/METADATA +1 -1
- {vellum_ai-0.8.5.dist-info → vellum_ai-0.8.6.dist-info}/RECORD +43 -63
- vellum/resources/ml_models/client.py +0 -948
- vellum/types/hosted_by_enum.py +0 -26
- vellum/types/ml_model_developer.py +0 -27
- vellum/types/ml_model_display_config_labelled.py +0 -26
- vellum/types/ml_model_display_config_request.py +0 -26
- vellum/types/ml_model_display_tag.py +0 -7
- vellum/types/ml_model_exec_config.py +0 -34
- vellum/types/ml_model_exec_config_request.py +0 -34
- vellum/types/ml_model_family.py +0 -29
- vellum/types/ml_model_feature.py +0 -18
- vellum/types/ml_model_parameter_config.py +0 -32
- vellum/types/ml_model_parameter_config_request.py +0 -32
- vellum/types/ml_model_read.py +0 -92
- vellum/types/ml_model_request_authorization_config.py +0 -20
- vellum/types/ml_model_request_authorization_config_request.py +0 -20
- vellum/types/ml_model_request_authorization_config_type_enum.py +0 -5
- vellum/types/ml_model_request_config.py +0 -22
- vellum/types/ml_model_request_config_request.py +0 -22
- vellum/types/ml_model_response_config.py +0 -21
- vellum/types/ml_model_response_config_request.py +0 -21
- vellum/types/ml_model_tokenizer_config.py +0 -7
- vellum/types/ml_model_tokenizer_config_request.py +0 -7
- vellum/types/open_api_array_property.py +0 -41
- vellum/types/open_api_array_property_request.py +0 -41
- vellum/types/open_api_const_property.py +0 -26
- vellum/types/open_api_integer_property.py +0 -30
- vellum/types/open_api_integer_property_request.py +0 -30
- vellum/types/open_api_number_property.py +0 -31
- vellum/types/open_api_number_property_request.py +0 -31
- vellum/types/open_api_object_property.py +0 -40
- vellum/types/open_api_object_property_request.py +0 -40
- vellum/types/open_api_one_of_property_request.py +0 -33
- vellum/types/open_api_property.py +0 -27
- vellum/types/open_api_property_request.py +0 -27
- vellum/types/open_api_ref_property.py +0 -26
- vellum/types/open_api_ref_property_request.py +0 -26
- vellum/types/open_api_string_property.py +0 -30
- vellum/types/open_api_string_property_request.py +0 -30
- vellum/types/paginated_ml_model_read_list.py +0 -23
- vellum/types/visibility_enum.py +0 -5
- /vellum/resources/{ml_models → ad_hoc}/__init__.py +0 -0
- {vellum_ai-0.8.5.dist-info → vellum_ai-0.8.6.dist-info}/LICENSE +0 -0
- {vellum_ai-0.8.5.dist-info → vellum_ai-0.8.6.dist-info}/WHEEL +0 -0
vellum/types/__init__.py
CHANGED
@@ -1,5 +1,11 @@
|
|
1
1
|
# This file was auto-generated by Fern from our API Definition.
|
2
2
|
|
3
|
+
from .ad_hoc_execute_prompt_event import AdHocExecutePromptEvent
|
4
|
+
from .ad_hoc_expand_meta_request import AdHocExpandMetaRequest
|
5
|
+
from .ad_hoc_fulfilled_prompt_execution_meta import AdHocFulfilledPromptExecutionMeta
|
6
|
+
from .ad_hoc_initiated_prompt_execution_meta import AdHocInitiatedPromptExecutionMeta
|
7
|
+
from .ad_hoc_rejected_prompt_execution_meta import AdHocRejectedPromptExecutionMeta
|
8
|
+
from .ad_hoc_streaming_prompt_execution_meta import AdHocStreamingPromptExecutionMeta
|
3
9
|
from .add_openai_api_key_enum import AddOpenaiApiKeyEnum
|
4
10
|
from .api_node_result import ApiNodeResult
|
5
11
|
from .api_node_result_data import ApiNodeResultData
|
@@ -28,6 +34,8 @@ from .chat_history_input_request import ChatHistoryInputRequest
|
|
28
34
|
from .chat_message import ChatMessage
|
29
35
|
from .chat_message_content import ChatMessageContent
|
30
36
|
from .chat_message_content_request import ChatMessageContentRequest
|
37
|
+
from .chat_message_prompt_block_properties_request import ChatMessagePromptBlockPropertiesRequest
|
38
|
+
from .chat_message_prompt_block_request import ChatMessagePromptBlockRequest
|
31
39
|
from .chat_message_request import ChatMessageRequest
|
32
40
|
from .chat_message_role import ChatMessageRole
|
33
41
|
from .code_execution_node_array_result import CodeExecutionNodeArrayResult
|
@@ -64,6 +72,8 @@ from .document_status import DocumentStatus
|
|
64
72
|
from .enriched_normalized_completion import EnrichedNormalizedCompletion
|
65
73
|
from .entity_status import EntityStatus
|
66
74
|
from .environment_enum import EnvironmentEnum
|
75
|
+
from .ephemeral_prompt_cache_config_request import EphemeralPromptCacheConfigRequest
|
76
|
+
from .ephemeral_prompt_cache_config_type_enum import EphemeralPromptCacheConfigTypeEnum
|
67
77
|
from .error_variable_value import ErrorVariableValue
|
68
78
|
from .error_vellum_value import ErrorVellumValue
|
69
79
|
from .error_vellum_value_request import ErrorVellumValueRequest
|
@@ -83,6 +93,7 @@ from .execution_vellum_value import ExecutionVellumValue
|
|
83
93
|
from .external_test_case_execution import ExternalTestCaseExecution
|
84
94
|
from .external_test_case_execution_request import ExternalTestCaseExecutionRequest
|
85
95
|
from .finish_reason_enum import FinishReasonEnum
|
96
|
+
from .fulfilled_ad_hoc_execute_prompt_event import FulfilledAdHocExecutePromptEvent
|
86
97
|
from .fulfilled_enum import FulfilledEnum
|
87
98
|
from .fulfilled_execute_prompt_event import FulfilledExecutePromptEvent
|
88
99
|
from .fulfilled_execute_prompt_response import FulfilledExecutePromptResponse
|
@@ -98,6 +109,8 @@ from .function_call_request import FunctionCallRequest
|
|
98
109
|
from .function_call_variable_value import FunctionCallVariableValue
|
99
110
|
from .function_call_vellum_value import FunctionCallVellumValue
|
100
111
|
from .function_call_vellum_value_request import FunctionCallVellumValueRequest
|
112
|
+
from .function_definition_prompt_block_properties_request import FunctionDefinitionPromptBlockPropertiesRequest
|
113
|
+
from .function_definition_prompt_block_request import FunctionDefinitionPromptBlockRequest
|
101
114
|
from .generate_options_request import GenerateOptionsRequest
|
102
115
|
from .generate_request import GenerateRequest
|
103
116
|
from .generate_response import GenerateResponse
|
@@ -109,9 +122,6 @@ from .generate_stream_result import GenerateStreamResult
|
|
109
122
|
from .generate_stream_result_data import GenerateStreamResultData
|
110
123
|
from .hkunlp_instructor_xl_vectorizer import HkunlpInstructorXlVectorizer
|
111
124
|
from .hkunlp_instructor_xl_vectorizer_request import HkunlpInstructorXlVectorizerRequest
|
112
|
-
from .hosted_by_enum import HostedByEnum
|
113
|
-
from .hugging_face_tokenizer_config import HuggingFaceTokenizerConfig
|
114
|
-
from .hugging_face_tokenizer_config_request import HuggingFaceTokenizerConfigRequest
|
115
125
|
from .image_chat_message_content import ImageChatMessageContent
|
116
126
|
from .image_chat_message_content_request import ImageChatMessageContentRequest
|
117
127
|
from .image_variable_value import ImageVariableValue
|
@@ -120,12 +130,15 @@ from .image_vellum_value_request import ImageVellumValueRequest
|
|
120
130
|
from .indexing_config_vectorizer import IndexingConfigVectorizer
|
121
131
|
from .indexing_config_vectorizer_request import IndexingConfigVectorizerRequest
|
122
132
|
from .indexing_state_enum import IndexingStateEnum
|
133
|
+
from .initiated_ad_hoc_execute_prompt_event import InitiatedAdHocExecutePromptEvent
|
123
134
|
from .initiated_execute_prompt_event import InitiatedExecutePromptEvent
|
124
135
|
from .initiated_prompt_execution_meta import InitiatedPromptExecutionMeta
|
125
136
|
from .initiated_workflow_node_result_event import InitiatedWorkflowNodeResultEvent
|
126
137
|
from .instructor_vectorizer_config import InstructorVectorizerConfig
|
127
138
|
from .instructor_vectorizer_config_request import InstructorVectorizerConfigRequest
|
128
139
|
from .iteration_state_enum import IterationStateEnum
|
140
|
+
from .jinja_prompt_block_properties_request import JinjaPromptBlockPropertiesRequest
|
141
|
+
from .jinja_prompt_block_request import JinjaPromptBlockRequest
|
129
142
|
from .json_input_request import JsonInputRequest
|
130
143
|
from .json_variable_value import JsonVariableValue
|
131
144
|
from .json_vellum_value import JsonVellumValue
|
@@ -140,29 +153,6 @@ from .metadata_filter_config_request import MetadataFilterConfigRequest
|
|
140
153
|
from .metadata_filter_rule_combinator import MetadataFilterRuleCombinator
|
141
154
|
from .metadata_filter_rule_request import MetadataFilterRuleRequest
|
142
155
|
from .metric_node_result import MetricNodeResult
|
143
|
-
from .ml_model_developer import MlModelDeveloper
|
144
|
-
from .ml_model_developer_enum_value_label import MlModelDeveloperEnumValueLabel
|
145
|
-
from .ml_model_display_config_labelled import MlModelDisplayConfigLabelled
|
146
|
-
from .ml_model_display_config_request import MlModelDisplayConfigRequest
|
147
|
-
from .ml_model_display_tag import MlModelDisplayTag
|
148
|
-
from .ml_model_display_tag_enum_value_label import MlModelDisplayTagEnumValueLabel
|
149
|
-
from .ml_model_exec_config import MlModelExecConfig
|
150
|
-
from .ml_model_exec_config_request import MlModelExecConfigRequest
|
151
|
-
from .ml_model_family import MlModelFamily
|
152
|
-
from .ml_model_family_enum_value_label import MlModelFamilyEnumValueLabel
|
153
|
-
from .ml_model_feature import MlModelFeature
|
154
|
-
from .ml_model_parameter_config import MlModelParameterConfig
|
155
|
-
from .ml_model_parameter_config_request import MlModelParameterConfigRequest
|
156
|
-
from .ml_model_read import MlModelRead
|
157
|
-
from .ml_model_request_authorization_config import MlModelRequestAuthorizationConfig
|
158
|
-
from .ml_model_request_authorization_config_request import MlModelRequestAuthorizationConfigRequest
|
159
|
-
from .ml_model_request_authorization_config_type_enum import MlModelRequestAuthorizationConfigTypeEnum
|
160
|
-
from .ml_model_request_config import MlModelRequestConfig
|
161
|
-
from .ml_model_request_config_request import MlModelRequestConfigRequest
|
162
|
-
from .ml_model_response_config import MlModelResponseConfig
|
163
|
-
from .ml_model_response_config_request import MlModelResponseConfigRequest
|
164
|
-
from .ml_model_tokenizer_config import MlModelTokenizerConfig
|
165
|
-
from .ml_model_tokenizer_config_request import MlModelTokenizerConfigRequest
|
166
156
|
from .ml_model_usage import MlModelUsage
|
167
157
|
from .named_scenario_input_chat_history_variable_value_request import NamedScenarioInputChatHistoryVariableValueRequest
|
168
158
|
from .named_scenario_input_json_variable_value_request import NamedScenarioInputJsonVariableValueRequest
|
@@ -217,28 +207,7 @@ from .open_ai_vectorizer_text_embedding_3_small import OpenAiVectorizerTextEmbed
|
|
217
207
|
from .open_ai_vectorizer_text_embedding_3_small_request import OpenAiVectorizerTextEmbedding3SmallRequest
|
218
208
|
from .open_ai_vectorizer_text_embedding_ada_002 import OpenAiVectorizerTextEmbeddingAda002
|
219
209
|
from .open_ai_vectorizer_text_embedding_ada_002_request import OpenAiVectorizerTextEmbeddingAda002Request
|
220
|
-
from .open_api_array_property import OpenApiArrayProperty
|
221
|
-
from .open_api_array_property_request import OpenApiArrayPropertyRequest
|
222
|
-
from .open_api_boolean_property import OpenApiBooleanProperty
|
223
|
-
from .open_api_boolean_property_request import OpenApiBooleanPropertyRequest
|
224
|
-
from .open_api_const_property import OpenApiConstProperty
|
225
|
-
from .open_api_const_property_request import OpenApiConstPropertyRequest
|
226
|
-
from .open_api_integer_property import OpenApiIntegerProperty
|
227
|
-
from .open_api_integer_property_request import OpenApiIntegerPropertyRequest
|
228
|
-
from .open_api_number_property import OpenApiNumberProperty
|
229
|
-
from .open_api_number_property_request import OpenApiNumberPropertyRequest
|
230
|
-
from .open_api_object_property import OpenApiObjectProperty
|
231
|
-
from .open_api_object_property_request import OpenApiObjectPropertyRequest
|
232
|
-
from .open_api_one_of_property import OpenApiOneOfProperty
|
233
|
-
from .open_api_one_of_property_request import OpenApiOneOfPropertyRequest
|
234
|
-
from .open_api_property import OpenApiProperty
|
235
|
-
from .open_api_property_request import OpenApiPropertyRequest
|
236
|
-
from .open_api_ref_property import OpenApiRefProperty
|
237
|
-
from .open_api_ref_property_request import OpenApiRefPropertyRequest
|
238
|
-
from .open_api_string_property import OpenApiStringProperty
|
239
|
-
from .open_api_string_property_request import OpenApiStringPropertyRequest
|
240
210
|
from .paginated_document_index_read_list import PaginatedDocumentIndexReadList
|
241
|
-
from .paginated_ml_model_read_list import PaginatedMlModelReadList
|
242
211
|
from .paginated_slim_deployment_read_list import PaginatedSlimDeploymentReadList
|
243
212
|
from .paginated_slim_document_list import PaginatedSlimDocumentList
|
244
213
|
from .paginated_slim_workflow_deployment_list import PaginatedSlimWorkflowDeploymentList
|
@@ -246,20 +215,29 @@ from .paginated_test_suite_run_execution_list import PaginatedTestSuiteRunExecut
|
|
246
215
|
from .paginated_test_suite_test_case_list import PaginatedTestSuiteTestCaseList
|
247
216
|
from .pdf_search_result_meta_source import PdfSearchResultMetaSource
|
248
217
|
from .pdf_search_result_meta_source_request import PdfSearchResultMetaSourceRequest
|
218
|
+
from .plain_text_prompt_block_request import PlainTextPromptBlockRequest
|
249
219
|
from .processing_failure_reason_enum import ProcessingFailureReasonEnum
|
250
220
|
from .processing_state_enum import ProcessingStateEnum
|
251
|
-
from .
|
221
|
+
from .prompt_block_request import PromptBlockRequest
|
222
|
+
from .prompt_block_state import PromptBlockState
|
223
|
+
from .prompt_deployment_expand_meta_request import PromptDeploymentExpandMetaRequest
|
252
224
|
from .prompt_deployment_input_request import PromptDeploymentInputRequest
|
253
225
|
from .prompt_execution_meta import PromptExecutionMeta
|
254
226
|
from .prompt_node_execution_meta import PromptNodeExecutionMeta
|
255
227
|
from .prompt_node_result import PromptNodeResult
|
256
228
|
from .prompt_node_result_data import PromptNodeResultData
|
257
229
|
from .prompt_output import PromptOutput
|
230
|
+
from .prompt_parameters_request import PromptParametersRequest
|
231
|
+
from .prompt_request_chat_history_input_request import PromptRequestChatHistoryInputRequest
|
232
|
+
from .prompt_request_input_request import PromptRequestInputRequest
|
233
|
+
from .prompt_request_json_input_request import PromptRequestJsonInputRequest
|
234
|
+
from .prompt_request_string_input_request import PromptRequestStringInputRequest
|
258
235
|
from .raw_prompt_execution_overrides_request import RawPromptExecutionOverridesRequest
|
259
236
|
from .reducto_chunker_config import ReductoChunkerConfig
|
260
237
|
from .reducto_chunker_config_request import ReductoChunkerConfigRequest
|
261
238
|
from .reducto_chunking import ReductoChunking
|
262
239
|
from .reducto_chunking_request import ReductoChunkingRequest
|
240
|
+
from .rejected_ad_hoc_execute_prompt_event import RejectedAdHocExecutePromptEvent
|
263
241
|
from .rejected_execute_prompt_event import RejectedExecutePromptEvent
|
264
242
|
from .rejected_execute_prompt_response import RejectedExecutePromptResponse
|
265
243
|
from .rejected_execute_workflow_workflow_result_event import RejectedExecuteWorkflowWorkflowResultEvent
|
@@ -267,6 +245,8 @@ from .rejected_prompt_execution_meta import RejectedPromptExecutionMeta
|
|
267
245
|
from .rejected_workflow_node_result_event import RejectedWorkflowNodeResultEvent
|
268
246
|
from .release_tag_source import ReleaseTagSource
|
269
247
|
from .replace_test_suite_test_case_request import ReplaceTestSuiteTestCaseRequest
|
248
|
+
from .rich_text_child_block_request import RichTextChildBlockRequest
|
249
|
+
from .rich_text_prompt_block_request import RichTextPromptBlockRequest
|
270
250
|
from .sandbox_scenario import SandboxScenario
|
271
251
|
from .scenario_input import ScenarioInput
|
272
252
|
from .scenario_input_chat_history_variable_value import ScenarioInputChatHistoryVariableValue
|
@@ -292,6 +272,7 @@ from .sentence_chunking_request import SentenceChunkingRequest
|
|
292
272
|
from .slim_deployment_read import SlimDeploymentRead
|
293
273
|
from .slim_document import SlimDocument
|
294
274
|
from .slim_workflow_deployment import SlimWorkflowDeployment
|
275
|
+
from .streaming_ad_hoc_execute_prompt_event import StreamingAdHocExecutePromptEvent
|
295
276
|
from .streaming_execute_prompt_event import StreamingExecutePromptEvent
|
296
277
|
from .streaming_prompt_execution_meta import StreamingPromptExecutionMeta
|
297
278
|
from .streaming_workflow_node_result_event import StreamingWorkflowNodeResultEvent
|
@@ -388,22 +369,21 @@ from .test_suite_test_case_replace_bulk_operation_request import TestSuiteTestCa
|
|
388
369
|
from .test_suite_test_case_replaced_bulk_result import TestSuiteTestCaseReplacedBulkResult
|
389
370
|
from .test_suite_test_case_replaced_bulk_result_data import TestSuiteTestCaseReplacedBulkResultData
|
390
371
|
from .test_suite_test_case_upsert_bulk_operation_request import TestSuiteTestCaseUpsertBulkOperationRequest
|
391
|
-
from .tik_token_tokenizer_config import TikTokenTokenizerConfig
|
392
|
-
from .tik_token_tokenizer_config_request import TikTokenTokenizerConfigRequest
|
393
372
|
from .token_overlapping_window_chunker_config import TokenOverlappingWindowChunkerConfig
|
394
373
|
from .token_overlapping_window_chunker_config_request import TokenOverlappingWindowChunkerConfigRequest
|
395
374
|
from .token_overlapping_window_chunking import TokenOverlappingWindowChunking
|
396
375
|
from .token_overlapping_window_chunking_request import TokenOverlappingWindowChunkingRequest
|
397
376
|
from .upload_document_response import UploadDocumentResponse
|
398
377
|
from .upsert_test_suite_test_case_request import UpsertTestSuiteTestCaseRequest
|
378
|
+
from .variable_prompt_block_request import VariablePromptBlockRequest
|
399
379
|
from .vellum_error import VellumError
|
400
380
|
from .vellum_error_code_enum import VellumErrorCodeEnum
|
401
381
|
from .vellum_error_request import VellumErrorRequest
|
402
382
|
from .vellum_image import VellumImage
|
403
383
|
from .vellum_image_request import VellumImageRequest
|
404
384
|
from .vellum_variable import VellumVariable
|
385
|
+
from .vellum_variable_request import VellumVariableRequest
|
405
386
|
from .vellum_variable_type import VellumVariableType
|
406
|
-
from .visibility_enum import VisibilityEnum
|
407
387
|
from .workflow_deployment_read import WorkflowDeploymentRead
|
408
388
|
from .workflow_event_error import WorkflowEventError
|
409
389
|
from .workflow_execution_actual_chat_history_request import WorkflowExecutionActualChatHistoryRequest
|
@@ -447,6 +427,12 @@ from .workflow_result_event_output_data_string import WorkflowResultEventOutputD
|
|
447
427
|
from .workflow_stream_event import WorkflowStreamEvent
|
448
428
|
|
449
429
|
__all__ = [
|
430
|
+
"AdHocExecutePromptEvent",
|
431
|
+
"AdHocExpandMetaRequest",
|
432
|
+
"AdHocFulfilledPromptExecutionMeta",
|
433
|
+
"AdHocInitiatedPromptExecutionMeta",
|
434
|
+
"AdHocRejectedPromptExecutionMeta",
|
435
|
+
"AdHocStreamingPromptExecutionMeta",
|
450
436
|
"AddOpenaiApiKeyEnum",
|
451
437
|
"ApiNodeResult",
|
452
438
|
"ApiNodeResultData",
|
@@ -467,6 +453,8 @@ __all__ = [
|
|
467
453
|
"ChatMessage",
|
468
454
|
"ChatMessageContent",
|
469
455
|
"ChatMessageContentRequest",
|
456
|
+
"ChatMessagePromptBlockPropertiesRequest",
|
457
|
+
"ChatMessagePromptBlockRequest",
|
470
458
|
"ChatMessageRequest",
|
471
459
|
"ChatMessageRole",
|
472
460
|
"CodeExecutionNodeArrayResult",
|
@@ -503,6 +491,8 @@ __all__ = [
|
|
503
491
|
"EnrichedNormalizedCompletion",
|
504
492
|
"EntityStatus",
|
505
493
|
"EnvironmentEnum",
|
494
|
+
"EphemeralPromptCacheConfigRequest",
|
495
|
+
"EphemeralPromptCacheConfigTypeEnum",
|
506
496
|
"ErrorVariableValue",
|
507
497
|
"ErrorVellumValue",
|
508
498
|
"ErrorVellumValueRequest",
|
@@ -522,6 +512,7 @@ __all__ = [
|
|
522
512
|
"ExternalTestCaseExecution",
|
523
513
|
"ExternalTestCaseExecutionRequest",
|
524
514
|
"FinishReasonEnum",
|
515
|
+
"FulfilledAdHocExecutePromptEvent",
|
525
516
|
"FulfilledEnum",
|
526
517
|
"FulfilledExecutePromptEvent",
|
527
518
|
"FulfilledExecutePromptResponse",
|
@@ -537,6 +528,8 @@ __all__ = [
|
|
537
528
|
"FunctionCallVariableValue",
|
538
529
|
"FunctionCallVellumValue",
|
539
530
|
"FunctionCallVellumValueRequest",
|
531
|
+
"FunctionDefinitionPromptBlockPropertiesRequest",
|
532
|
+
"FunctionDefinitionPromptBlockRequest",
|
540
533
|
"GenerateOptionsRequest",
|
541
534
|
"GenerateRequest",
|
542
535
|
"GenerateResponse",
|
@@ -548,9 +541,6 @@ __all__ = [
|
|
548
541
|
"GenerateStreamResultData",
|
549
542
|
"HkunlpInstructorXlVectorizer",
|
550
543
|
"HkunlpInstructorXlVectorizerRequest",
|
551
|
-
"HostedByEnum",
|
552
|
-
"HuggingFaceTokenizerConfig",
|
553
|
-
"HuggingFaceTokenizerConfigRequest",
|
554
544
|
"ImageChatMessageContent",
|
555
545
|
"ImageChatMessageContentRequest",
|
556
546
|
"ImageVariableValue",
|
@@ -559,12 +549,15 @@ __all__ = [
|
|
559
549
|
"IndexingConfigVectorizer",
|
560
550
|
"IndexingConfigVectorizerRequest",
|
561
551
|
"IndexingStateEnum",
|
552
|
+
"InitiatedAdHocExecutePromptEvent",
|
562
553
|
"InitiatedExecutePromptEvent",
|
563
554
|
"InitiatedPromptExecutionMeta",
|
564
555
|
"InitiatedWorkflowNodeResultEvent",
|
565
556
|
"InstructorVectorizerConfig",
|
566
557
|
"InstructorVectorizerConfigRequest",
|
567
558
|
"IterationStateEnum",
|
559
|
+
"JinjaPromptBlockPropertiesRequest",
|
560
|
+
"JinjaPromptBlockRequest",
|
568
561
|
"JsonInputRequest",
|
569
562
|
"JsonVariableValue",
|
570
563
|
"JsonVellumValue",
|
@@ -579,29 +572,6 @@ __all__ = [
|
|
579
572
|
"MetadataFilterRuleCombinator",
|
580
573
|
"MetadataFilterRuleRequest",
|
581
574
|
"MetricNodeResult",
|
582
|
-
"MlModelDeveloper",
|
583
|
-
"MlModelDeveloperEnumValueLabel",
|
584
|
-
"MlModelDisplayConfigLabelled",
|
585
|
-
"MlModelDisplayConfigRequest",
|
586
|
-
"MlModelDisplayTag",
|
587
|
-
"MlModelDisplayTagEnumValueLabel",
|
588
|
-
"MlModelExecConfig",
|
589
|
-
"MlModelExecConfigRequest",
|
590
|
-
"MlModelFamily",
|
591
|
-
"MlModelFamilyEnumValueLabel",
|
592
|
-
"MlModelFeature",
|
593
|
-
"MlModelParameterConfig",
|
594
|
-
"MlModelParameterConfigRequest",
|
595
|
-
"MlModelRead",
|
596
|
-
"MlModelRequestAuthorizationConfig",
|
597
|
-
"MlModelRequestAuthorizationConfigRequest",
|
598
|
-
"MlModelRequestAuthorizationConfigTypeEnum",
|
599
|
-
"MlModelRequestConfig",
|
600
|
-
"MlModelRequestConfigRequest",
|
601
|
-
"MlModelResponseConfig",
|
602
|
-
"MlModelResponseConfigRequest",
|
603
|
-
"MlModelTokenizerConfig",
|
604
|
-
"MlModelTokenizerConfigRequest",
|
605
575
|
"MlModelUsage",
|
606
576
|
"NamedScenarioInputChatHistoryVariableValueRequest",
|
607
577
|
"NamedScenarioInputJsonVariableValueRequest",
|
@@ -656,28 +626,7 @@ __all__ = [
|
|
656
626
|
"OpenAiVectorizerTextEmbedding3SmallRequest",
|
657
627
|
"OpenAiVectorizerTextEmbeddingAda002",
|
658
628
|
"OpenAiVectorizerTextEmbeddingAda002Request",
|
659
|
-
"OpenApiArrayProperty",
|
660
|
-
"OpenApiArrayPropertyRequest",
|
661
|
-
"OpenApiBooleanProperty",
|
662
|
-
"OpenApiBooleanPropertyRequest",
|
663
|
-
"OpenApiConstProperty",
|
664
|
-
"OpenApiConstPropertyRequest",
|
665
|
-
"OpenApiIntegerProperty",
|
666
|
-
"OpenApiIntegerPropertyRequest",
|
667
|
-
"OpenApiNumberProperty",
|
668
|
-
"OpenApiNumberPropertyRequest",
|
669
|
-
"OpenApiObjectProperty",
|
670
|
-
"OpenApiObjectPropertyRequest",
|
671
|
-
"OpenApiOneOfProperty",
|
672
|
-
"OpenApiOneOfPropertyRequest",
|
673
|
-
"OpenApiProperty",
|
674
|
-
"OpenApiPropertyRequest",
|
675
|
-
"OpenApiRefProperty",
|
676
|
-
"OpenApiRefPropertyRequest",
|
677
|
-
"OpenApiStringProperty",
|
678
|
-
"OpenApiStringPropertyRequest",
|
679
629
|
"PaginatedDocumentIndexReadList",
|
680
|
-
"PaginatedMlModelReadList",
|
681
630
|
"PaginatedSlimDeploymentReadList",
|
682
631
|
"PaginatedSlimDocumentList",
|
683
632
|
"PaginatedSlimWorkflowDeploymentList",
|
@@ -685,20 +634,29 @@ __all__ = [
|
|
685
634
|
"PaginatedTestSuiteTestCaseList",
|
686
635
|
"PdfSearchResultMetaSource",
|
687
636
|
"PdfSearchResultMetaSourceRequest",
|
637
|
+
"PlainTextPromptBlockRequest",
|
688
638
|
"ProcessingFailureReasonEnum",
|
689
639
|
"ProcessingStateEnum",
|
690
|
-
"
|
640
|
+
"PromptBlockRequest",
|
641
|
+
"PromptBlockState",
|
642
|
+
"PromptDeploymentExpandMetaRequest",
|
691
643
|
"PromptDeploymentInputRequest",
|
692
644
|
"PromptExecutionMeta",
|
693
645
|
"PromptNodeExecutionMeta",
|
694
646
|
"PromptNodeResult",
|
695
647
|
"PromptNodeResultData",
|
696
648
|
"PromptOutput",
|
649
|
+
"PromptParametersRequest",
|
650
|
+
"PromptRequestChatHistoryInputRequest",
|
651
|
+
"PromptRequestInputRequest",
|
652
|
+
"PromptRequestJsonInputRequest",
|
653
|
+
"PromptRequestStringInputRequest",
|
697
654
|
"RawPromptExecutionOverridesRequest",
|
698
655
|
"ReductoChunkerConfig",
|
699
656
|
"ReductoChunkerConfigRequest",
|
700
657
|
"ReductoChunking",
|
701
658
|
"ReductoChunkingRequest",
|
659
|
+
"RejectedAdHocExecutePromptEvent",
|
702
660
|
"RejectedExecutePromptEvent",
|
703
661
|
"RejectedExecutePromptResponse",
|
704
662
|
"RejectedExecuteWorkflowWorkflowResultEvent",
|
@@ -706,6 +664,8 @@ __all__ = [
|
|
706
664
|
"RejectedWorkflowNodeResultEvent",
|
707
665
|
"ReleaseTagSource",
|
708
666
|
"ReplaceTestSuiteTestCaseRequest",
|
667
|
+
"RichTextChildBlockRequest",
|
668
|
+
"RichTextPromptBlockRequest",
|
709
669
|
"SandboxScenario",
|
710
670
|
"ScenarioInput",
|
711
671
|
"ScenarioInputChatHistoryVariableValue",
|
@@ -731,6 +691,7 @@ __all__ = [
|
|
731
691
|
"SlimDeploymentRead",
|
732
692
|
"SlimDocument",
|
733
693
|
"SlimWorkflowDeployment",
|
694
|
+
"StreamingAdHocExecutePromptEvent",
|
734
695
|
"StreamingExecutePromptEvent",
|
735
696
|
"StreamingPromptExecutionMeta",
|
736
697
|
"StreamingWorkflowNodeResultEvent",
|
@@ -823,22 +784,21 @@ __all__ = [
|
|
823
784
|
"TestSuiteTestCaseReplacedBulkResult",
|
824
785
|
"TestSuiteTestCaseReplacedBulkResultData",
|
825
786
|
"TestSuiteTestCaseUpsertBulkOperationRequest",
|
826
|
-
"TikTokenTokenizerConfig",
|
827
|
-
"TikTokenTokenizerConfigRequest",
|
828
787
|
"TokenOverlappingWindowChunkerConfig",
|
829
788
|
"TokenOverlappingWindowChunkerConfigRequest",
|
830
789
|
"TokenOverlappingWindowChunking",
|
831
790
|
"TokenOverlappingWindowChunkingRequest",
|
832
791
|
"UploadDocumentResponse",
|
833
792
|
"UpsertTestSuiteTestCaseRequest",
|
793
|
+
"VariablePromptBlockRequest",
|
834
794
|
"VellumError",
|
835
795
|
"VellumErrorCodeEnum",
|
836
796
|
"VellumErrorRequest",
|
837
797
|
"VellumImage",
|
838
798
|
"VellumImageRequest",
|
839
799
|
"VellumVariable",
|
800
|
+
"VellumVariableRequest",
|
840
801
|
"VellumVariableType",
|
841
|
-
"VisibilityEnum",
|
842
802
|
"WorkflowDeploymentRead",
|
843
803
|
"WorkflowEventError",
|
844
804
|
"WorkflowExecutionActualChatHistoryRequest",
|
@@ -0,0 +1,14 @@
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
2
|
+
|
3
|
+
import typing
|
4
|
+
from .initiated_ad_hoc_execute_prompt_event import InitiatedAdHocExecutePromptEvent
|
5
|
+
from .streaming_ad_hoc_execute_prompt_event import StreamingAdHocExecutePromptEvent
|
6
|
+
from .fulfilled_ad_hoc_execute_prompt_event import FulfilledAdHocExecutePromptEvent
|
7
|
+
from .rejected_ad_hoc_execute_prompt_event import RejectedAdHocExecutePromptEvent
|
8
|
+
|
9
|
+
AdHocExecutePromptEvent = typing.Union[
|
10
|
+
InitiatedAdHocExecutePromptEvent,
|
11
|
+
StreamingAdHocExecutePromptEvent,
|
12
|
+
FulfilledAdHocExecutePromptEvent,
|
13
|
+
RejectedAdHocExecutePromptEvent,
|
14
|
+
]
|
@@ -0,0 +1,33 @@
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
2
|
+
|
3
|
+
from ..core.pydantic_utilities import UniversalBaseModel
|
4
|
+
import typing
|
5
|
+
import pydantic
|
6
|
+
from ..core.pydantic_utilities import IS_PYDANTIC_V2
|
7
|
+
|
8
|
+
|
9
|
+
class AdHocExpandMetaRequest(UniversalBaseModel):
|
10
|
+
cost: typing.Optional[bool] = None
|
11
|
+
model_name: typing.Optional[bool] = pydantic.Field(default=None)
|
12
|
+
"""
|
13
|
+
If enabled, the response will include the model identifier representing the ML Model invoked by the Prompt.
|
14
|
+
"""
|
15
|
+
|
16
|
+
usage: typing.Optional[bool] = pydantic.Field(default=None)
|
17
|
+
"""
|
18
|
+
If enabled, the response will include model host usage tracking. This may increase latency for some model hosts.
|
19
|
+
"""
|
20
|
+
|
21
|
+
finish_reason: typing.Optional[bool] = pydantic.Field(default=None)
|
22
|
+
"""
|
23
|
+
If enabled, the response will include the reason provided by the model for why the execution finished.
|
24
|
+
"""
|
25
|
+
|
26
|
+
if IS_PYDANTIC_V2:
|
27
|
+
model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2
|
28
|
+
else:
|
29
|
+
|
30
|
+
class Config:
|
31
|
+
frozen = True
|
32
|
+
smart_union = True
|
33
|
+
extra = pydantic.Extra.allow
|
@@ -0,0 +1,27 @@
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
2
|
+
|
3
|
+
from ..core.pydantic_utilities import UniversalBaseModel
|
4
|
+
import typing
|
5
|
+
from .finish_reason_enum import FinishReasonEnum
|
6
|
+
from .ml_model_usage import MlModelUsage
|
7
|
+
from ..core.pydantic_utilities import IS_PYDANTIC_V2
|
8
|
+
import pydantic
|
9
|
+
|
10
|
+
|
11
|
+
class AdHocFulfilledPromptExecutionMeta(UniversalBaseModel):
|
12
|
+
"""
|
13
|
+
The subset of the metadata tracked by Vellum during prompt execution that the request opted into with `expand_meta`.
|
14
|
+
"""
|
15
|
+
|
16
|
+
latency: typing.Optional[int] = None
|
17
|
+
finish_reason: typing.Optional[FinishReasonEnum] = None
|
18
|
+
usage: typing.Optional[MlModelUsage] = None
|
19
|
+
|
20
|
+
if IS_PYDANTIC_V2:
|
21
|
+
model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2
|
22
|
+
else:
|
23
|
+
|
24
|
+
class Config:
|
25
|
+
frozen = True
|
26
|
+
smart_union = True
|
27
|
+
extra = pydantic.Extra.allow
|
vellum/types/{open_api_const_property_request.py → ad_hoc_initiated_prompt_execution_meta.py}
RENAMED
@@ -6,15 +6,13 @@ from ..core.pydantic_utilities import IS_PYDANTIC_V2
|
|
6
6
|
import pydantic
|
7
7
|
|
8
8
|
|
9
|
-
class
|
9
|
+
class AdHocInitiatedPromptExecutionMeta(UniversalBaseModel):
|
10
10
|
"""
|
11
|
-
|
11
|
+
The subset of the metadata tracked by Vellum during prompt execution that the request opted into with `expand_meta`.
|
12
12
|
"""
|
13
13
|
|
14
|
-
|
15
|
-
|
16
|
-
description: typing.Optional[str] = None
|
17
|
-
const: str
|
14
|
+
model_name: typing.Optional[str] = None
|
15
|
+
latency: typing.Optional[int] = None
|
18
16
|
|
19
17
|
if IS_PYDANTIC_V2:
|
20
18
|
model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2
|
vellum/types/{open_api_boolean_property_request.py → ad_hoc_rejected_prompt_execution_meta.py}
RENAMED
@@ -2,19 +2,18 @@
|
|
2
2
|
|
3
3
|
from ..core.pydantic_utilities import UniversalBaseModel
|
4
4
|
import typing
|
5
|
+
from .finish_reason_enum import FinishReasonEnum
|
5
6
|
from ..core.pydantic_utilities import IS_PYDANTIC_V2
|
6
7
|
import pydantic
|
7
8
|
|
8
9
|
|
9
|
-
class
|
10
|
+
class AdHocRejectedPromptExecutionMeta(UniversalBaseModel):
|
10
11
|
"""
|
11
|
-
|
12
|
+
The subset of the metadata tracked by Vellum during prompt execution that the request opted into with `expand_meta`.
|
12
13
|
"""
|
13
14
|
|
14
|
-
|
15
|
-
|
16
|
-
title: typing.Optional[str] = None
|
17
|
-
description: typing.Optional[str] = None
|
15
|
+
latency: typing.Optional[int] = None
|
16
|
+
finish_reason: typing.Optional[FinishReasonEnum] = None
|
18
17
|
|
19
18
|
if IS_PYDANTIC_V2:
|
20
19
|
model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2
|
vellum/types/{tik_token_tokenizer_config_request.py → ad_hoc_streaming_prompt_execution_meta.py}
RENAMED
@@ -6,13 +6,12 @@ from ..core.pydantic_utilities import IS_PYDANTIC_V2
|
|
6
6
|
import pydantic
|
7
7
|
|
8
8
|
|
9
|
-
class
|
9
|
+
class AdHocStreamingPromptExecutionMeta(UniversalBaseModel):
|
10
10
|
"""
|
11
|
-
|
11
|
+
The subset of the metadata tracked by Vellum during prompt execution that the request opted into with `expand_meta`.
|
12
12
|
"""
|
13
13
|
|
14
|
-
|
15
|
-
name: str
|
14
|
+
latency: typing.Optional[int] = None
|
16
15
|
|
17
16
|
if IS_PYDANTIC_V2:
|
18
17
|
model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2
|
vellum/types/{open_api_one_of_property.py → chat_message_prompt_block_properties_request.py}
RENAMED
@@ -3,20 +3,21 @@
|
|
3
3
|
from __future__ import annotations
|
4
4
|
from ..core.pydantic_utilities import UniversalBaseModel
|
5
5
|
import typing
|
6
|
-
import
|
6
|
+
from .chat_message_role import ChatMessageRole
|
7
7
|
from ..core.pydantic_utilities import IS_PYDANTIC_V2
|
8
|
+
import pydantic
|
8
9
|
from ..core.pydantic_utilities import update_forward_refs
|
9
10
|
|
10
11
|
|
11
|
-
class
|
12
|
+
class ChatMessagePromptBlockPropertiesRequest(UniversalBaseModel):
|
12
13
|
"""
|
13
|
-
|
14
|
+
The properties of a ChatMessagePromptTemplateBlock
|
14
15
|
"""
|
15
16
|
|
16
|
-
|
17
|
-
|
18
|
-
|
19
|
-
|
17
|
+
blocks: typing.List["PromptBlockRequest"]
|
18
|
+
chat_role: typing.Optional[ChatMessageRole] = None
|
19
|
+
chat_source: typing.Optional[str] = None
|
20
|
+
chat_message_unterminated: typing.Optional[bool] = None
|
20
21
|
|
21
22
|
if IS_PYDANTIC_V2:
|
22
23
|
model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2
|
@@ -28,6 +29,6 @@ class OpenApiOneOfProperty(UniversalBaseModel):
|
|
28
29
|
extra = pydantic.Extra.allow
|
29
30
|
|
30
31
|
|
31
|
-
from .
|
32
|
+
from .prompt_block_request import PromptBlockRequest # noqa: E402
|
32
33
|
|
33
|
-
update_forward_refs(
|
34
|
+
update_forward_refs(ChatMessagePromptBlockPropertiesRequest)
|
@@ -0,0 +1,36 @@
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
2
|
+
|
3
|
+
from __future__ import annotations
|
4
|
+
from ..core.pydantic_utilities import UniversalBaseModel
|
5
|
+
import typing
|
6
|
+
from .prompt_block_state import PromptBlockState
|
7
|
+
from .ephemeral_prompt_cache_config_request import EphemeralPromptCacheConfigRequest
|
8
|
+
from ..core.pydantic_utilities import IS_PYDANTIC_V2
|
9
|
+
import pydantic
|
10
|
+
from ..core.pydantic_utilities import update_forward_refs
|
11
|
+
|
12
|
+
|
13
|
+
class ChatMessagePromptBlockRequest(UniversalBaseModel):
|
14
|
+
"""
|
15
|
+
A block that represents a chat message in a prompt template.
|
16
|
+
"""
|
17
|
+
|
18
|
+
block_type: typing.Literal["CHAT_MESSAGE"] = "CHAT_MESSAGE"
|
19
|
+
properties: "ChatMessagePromptBlockPropertiesRequest"
|
20
|
+
id: str
|
21
|
+
state: typing.Optional[PromptBlockState] = None
|
22
|
+
cache_config: typing.Optional[EphemeralPromptCacheConfigRequest] = None
|
23
|
+
|
24
|
+
if IS_PYDANTIC_V2:
|
25
|
+
model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2
|
26
|
+
else:
|
27
|
+
|
28
|
+
class Config:
|
29
|
+
frozen = True
|
30
|
+
smart_union = True
|
31
|
+
extra = pydantic.Extra.allow
|
32
|
+
|
33
|
+
|
34
|
+
from .chat_message_prompt_block_properties_request import ChatMessagePromptBlockPropertiesRequest # noqa: E402
|
35
|
+
|
36
|
+
update_forward_refs(ChatMessagePromptBlockRequest)
|
vellum/types/{ml_model_developer_enum_value_label.py → ephemeral_prompt_cache_config_request.py}
RENAMED
@@ -1,15 +1,14 @@
|
|
1
1
|
# This file was auto-generated by Fern from our API Definition.
|
2
2
|
|
3
3
|
from ..core.pydantic_utilities import UniversalBaseModel
|
4
|
-
from .ml_model_developer import MlModelDeveloper
|
5
|
-
from ..core.pydantic_utilities import IS_PYDANTIC_V2
|
6
4
|
import typing
|
5
|
+
from .ephemeral_prompt_cache_config_type_enum import EphemeralPromptCacheConfigTypeEnum
|
6
|
+
from ..core.pydantic_utilities import IS_PYDANTIC_V2
|
7
7
|
import pydantic
|
8
8
|
|
9
9
|
|
10
|
-
class
|
11
|
-
|
12
|
-
value: MlModelDeveloper
|
10
|
+
class EphemeralPromptCacheConfigRequest(UniversalBaseModel):
|
11
|
+
type: typing.Optional[EphemeralPromptCacheConfigTypeEnum] = None
|
13
12
|
|
14
13
|
if IS_PYDANTIC_V2:
|
15
14
|
model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2
|