vellum-ai 0.1.5__tar.gz → 0.1.7__tar.gz
Sign up to get free protection for your applications and to get access to all the features.
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/PKG-INFO +1 -1
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/pyproject.toml +1 -1
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/__init__.py +26 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/client.py +68 -57
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/core/client_wrapper.py +1 -1
- vellum_ai-0.1.7/src/vellum/resources/deployments/client.py +164 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/resources/model_versions/client.py +2 -2
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/__init__.py +28 -0
- vellum_ai-0.1.7/src/vellum/types/deployment_provider_payload_response.py +28 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/deployment_read.py +3 -1
- vellum_ai-0.1.7/src/vellum/types/execute_prompt_request.py +39 -0
- vellum_ai-0.1.7/src/vellum/types/execute_prompt_streaming_response.py +56 -0
- vellum_ai-0.1.7/src/vellum/types/fulfilled_execute_prompt_streaming_response.py +29 -0
- vellum_ai-0.1.7/src/vellum/types/initiated_execute_prompt_response.py +30 -0
- vellum_ai-0.1.7/src/vellum/types/initiated_execute_prompt_streaming_response.py +29 -0
- vellum_ai-0.1.7/src/vellum/types/rejected_execute_prompt_response.py +31 -0
- vellum_ai-0.1.7/src/vellum/types/rejected_execute_prompt_streaming_response.py +29 -0
- vellum_ai-0.1.7/src/vellum/types/streaming_execute_prompt_streaming_response.py +29 -0
- vellum_ai-0.1.5/src/vellum/resources/deployments/client.py +0 -65
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/README.md +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/core/__init__.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/core/api_error.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/core/datetime_utils.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/core/jsonable_encoder.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/core/remove_none_from_dict.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/environment.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/errors/__init__.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/errors/bad_request_error.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/errors/conflict_error.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/errors/forbidden_error.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/errors/internal_server_error.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/errors/not_found_error.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/py.typed +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/resources/__init__.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/resources/deployments/__init__.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/resources/document_indexes/__init__.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/resources/document_indexes/client.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/resources/documents/__init__.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/resources/documents/client.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/resources/model_versions/__init__.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/resources/registered_prompts/__init__.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/resources/registered_prompts/client.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/resources/sandboxes/__init__.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/resources/sandboxes/client.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/resources/test_suites/__init__.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/resources/test_suites/client.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/api_node_result.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/api_node_result_data.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/block_type_enum.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/chat_history_input_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/chat_message.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/chat_message_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/chat_message_role.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/conditional_node_result.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/conditional_node_result_data.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/deployment_status.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/document_document_to_document_index.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/document_index_read.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/document_index_status.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/document_read.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/document_status.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/enriched_normalized_completion.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/environment_enum.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/error_execute_prompt_response.py +1 -1
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/evaluation_params.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/evaluation_params_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/execute_prompt_api_error_response.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/execute_prompt_response.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/execute_workflow_stream_error_response.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/finish_reason_enum.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/generate_error_response.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/generate_options_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/generate_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/generate_response.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/generate_result.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/generate_result_data.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/generate_result_error.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/generate_stream_response.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/generate_stream_result.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/generate_stream_result_data.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/indexing_state_enum.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/json_execute_prompt_response.py +1 -1
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/json_input_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/logical_operator.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/logprobs_enum.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/metadata_filter_config_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/metadata_filter_rule_combinator.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/metadata_filter_rule_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/model_version_build_config.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/model_version_exec_config.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/model_version_exec_config_parameters.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/model_version_read.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/model_version_read_status_enum.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/model_version_sandbox_snapshot.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/node_input_compiled_chat_history_value.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/node_input_compiled_error_value.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/node_input_compiled_json_value.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/node_input_compiled_number_value.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/node_input_compiled_search_results_value.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/node_input_compiled_string_value.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/node_input_variable_compiled_value.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/normalized_log_probs.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/normalized_token_log_probs.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/paginated_slim_document_list.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/processing_failure_reason_enum.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/processing_state_enum.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/prompt_deployment_input_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/prompt_node_result.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/prompt_node_result_data.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/prompt_template_block.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/prompt_template_block_data.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/prompt_template_block_data_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/prompt_template_block_properties.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/prompt_template_block_properties_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/prompt_template_block_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/provider_enum.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/register_prompt_error_response.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/register_prompt_model_parameters_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/register_prompt_prompt.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/register_prompt_prompt_info_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/register_prompt_response.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/registered_prompt_deployment.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/registered_prompt_input_variable_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/registered_prompt_model_version.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/registered_prompt_sandbox.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/registered_prompt_sandbox_snapshot.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/sandbox_metric_input_params.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/sandbox_metric_input_params_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/sandbox_scenario.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/scenario_input.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/scenario_input_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/scenario_input_type_enum.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/search_error_response.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/search_filters_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/search_node_result.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/search_node_result_data.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/search_request_options_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/search_response.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/search_result.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/search_result_document.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/search_result_document_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/search_result_merging_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/search_result_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/search_weights_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/slim_document.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/string_execute_prompt_response.py +1 -1
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/string_input_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/submit_completion_actual_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/submit_completion_actuals_error_response.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/submit_workflow_execution_actual_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/templating_node_chat_history_result.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/templating_node_error_result.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/templating_node_json_result.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/templating_node_number_result.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/templating_node_result.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/templating_node_result_data.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/templating_node_result_output.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/templating_node_search_results_result.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/templating_node_string_result.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/terminal_node_chat_history_result.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/terminal_node_error_result.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/terminal_node_json_result.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/terminal_node_number_result.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/terminal_node_result.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/terminal_node_result_data.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/terminal_node_result_output.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/terminal_node_search_results_result.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/terminal_node_string_result.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/test_case_chat_history_variable_value.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/test_case_chat_history_variable_value_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/test_case_error_variable_value.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/test_case_error_variable_value_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/test_case_json_variable_value.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/test_case_json_variable_value_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/test_case_number_variable_value.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/test_case_number_variable_value_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/test_case_search_results_variable_value.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/test_case_search_results_variable_value_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/test_case_string_variable_value.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/test_case_string_variable_value_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/test_case_variable_value.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/test_case_variable_value_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/test_suite_test_case.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/upload_document_error_response.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/upload_document_response.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/vellum_error.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/vellum_error_code_enum.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/vellum_error_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/vellum_variable.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/vellum_variable_type.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_event_error.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_execution_actual_chat_history_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_execution_actual_json_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_execution_actual_string_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_execution_event_error_code.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_execution_event_type.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_execution_node_result_event.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_execution_workflow_result_event.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_node_result_data.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_node_result_event.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_node_result_event_state.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_request_chat_history_input_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_request_input_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_request_json_input_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_request_string_input_request.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_result_event.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_result_event_output_data.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_result_event_output_data_chat_history.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_result_event_output_data_error.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_result_event_output_data_json.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_result_event_output_data_number.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_result_event_output_data_search_results.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_result_event_output_data_string.py +0 -0
- {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_stream_event.py +0 -0
@@ -10,6 +10,7 @@ from .types import (
|
|
10
10
|
ChatMessageRole,
|
11
11
|
ConditionalNodeResult,
|
12
12
|
ConditionalNodeResultData,
|
13
|
+
DeploymentProviderPayloadResponse,
|
13
14
|
DeploymentRead,
|
14
15
|
DeploymentStatus,
|
15
16
|
DocumentDocumentToDocumentIndex,
|
@@ -23,12 +24,19 @@ from .types import (
|
|
23
24
|
EvaluationParams,
|
24
25
|
EvaluationParamsRequest,
|
25
26
|
ExecutePromptApiErrorResponse,
|
27
|
+
ExecutePromptRequest,
|
26
28
|
ExecutePromptResponse,
|
27
29
|
ExecutePromptResponse_Error,
|
28
30
|
ExecutePromptResponse_Json,
|
29
31
|
ExecutePromptResponse_String,
|
32
|
+
ExecutePromptStreamingResponse,
|
33
|
+
ExecutePromptStreamingResponse_Fulfilled,
|
34
|
+
ExecutePromptStreamingResponse_Initiated,
|
35
|
+
ExecutePromptStreamingResponse_Rejected,
|
36
|
+
ExecutePromptStreamingResponse_Streaming,
|
30
37
|
ExecuteWorkflowStreamErrorResponse,
|
31
38
|
FinishReasonEnum,
|
39
|
+
FulfilledExecutePromptStreamingResponse,
|
32
40
|
GenerateErrorResponse,
|
33
41
|
GenerateOptionsRequest,
|
34
42
|
GenerateRequest,
|
@@ -40,6 +48,8 @@ from .types import (
|
|
40
48
|
GenerateStreamResult,
|
41
49
|
GenerateStreamResultData,
|
42
50
|
IndexingStateEnum,
|
51
|
+
InitiatedExecutePromptResponse,
|
52
|
+
InitiatedExecutePromptStreamingResponse,
|
43
53
|
JsonExecutePromptResponse,
|
44
54
|
JsonInputRequest,
|
45
55
|
LogicalOperator,
|
@@ -94,6 +104,8 @@ from .types import (
|
|
94
104
|
RegisteredPromptModelVersion,
|
95
105
|
RegisteredPromptSandbox,
|
96
106
|
RegisteredPromptSandboxSnapshot,
|
107
|
+
RejectedExecutePromptResponse,
|
108
|
+
RejectedExecutePromptStreamingResponse,
|
97
109
|
SandboxMetricInputParams,
|
98
110
|
SandboxMetricInputParamsRequest,
|
99
111
|
SandboxScenario,
|
@@ -113,6 +125,7 @@ from .types import (
|
|
113
125
|
SearchResultRequest,
|
114
126
|
SearchWeightsRequest,
|
115
127
|
SlimDocument,
|
128
|
+
StreamingExecutePromptStreamingResponse,
|
116
129
|
StringExecutePromptResponse,
|
117
130
|
StringInputRequest,
|
118
131
|
SubmitCompletionActualRequest,
|
@@ -251,6 +264,7 @@ __all__ = [
|
|
251
264
|
"ConditionalNodeResult",
|
252
265
|
"ConditionalNodeResultData",
|
253
266
|
"ConflictError",
|
267
|
+
"DeploymentProviderPayloadResponse",
|
254
268
|
"DeploymentRead",
|
255
269
|
"DeploymentStatus",
|
256
270
|
"DocumentDocumentToDocumentIndex",
|
@@ -264,13 +278,20 @@ __all__ = [
|
|
264
278
|
"EvaluationParams",
|
265
279
|
"EvaluationParamsRequest",
|
266
280
|
"ExecutePromptApiErrorResponse",
|
281
|
+
"ExecutePromptRequest",
|
267
282
|
"ExecutePromptResponse",
|
268
283
|
"ExecutePromptResponse_Error",
|
269
284
|
"ExecutePromptResponse_Json",
|
270
285
|
"ExecutePromptResponse_String",
|
286
|
+
"ExecutePromptStreamingResponse",
|
287
|
+
"ExecutePromptStreamingResponse_Fulfilled",
|
288
|
+
"ExecutePromptStreamingResponse_Initiated",
|
289
|
+
"ExecutePromptStreamingResponse_Rejected",
|
290
|
+
"ExecutePromptStreamingResponse_Streaming",
|
271
291
|
"ExecuteWorkflowStreamErrorResponse",
|
272
292
|
"FinishReasonEnum",
|
273
293
|
"ForbiddenError",
|
294
|
+
"FulfilledExecutePromptStreamingResponse",
|
274
295
|
"GenerateErrorResponse",
|
275
296
|
"GenerateOptionsRequest",
|
276
297
|
"GenerateRequest",
|
@@ -282,6 +303,8 @@ __all__ = [
|
|
282
303
|
"GenerateStreamResult",
|
283
304
|
"GenerateStreamResultData",
|
284
305
|
"IndexingStateEnum",
|
306
|
+
"InitiatedExecutePromptResponse",
|
307
|
+
"InitiatedExecutePromptStreamingResponse",
|
285
308
|
"InternalServerError",
|
286
309
|
"JsonExecutePromptResponse",
|
287
310
|
"JsonInputRequest",
|
@@ -338,6 +361,8 @@ __all__ = [
|
|
338
361
|
"RegisteredPromptModelVersion",
|
339
362
|
"RegisteredPromptSandbox",
|
340
363
|
"RegisteredPromptSandboxSnapshot",
|
364
|
+
"RejectedExecutePromptResponse",
|
365
|
+
"RejectedExecutePromptStreamingResponse",
|
341
366
|
"SandboxMetricInputParams",
|
342
367
|
"SandboxMetricInputParamsRequest",
|
343
368
|
"SandboxScenario",
|
@@ -357,6 +382,7 @@ __all__ = [
|
|
357
382
|
"SearchResultRequest",
|
358
383
|
"SearchWeightsRequest",
|
359
384
|
"SlimDocument",
|
385
|
+
"StreamingExecutePromptStreamingResponse",
|
360
386
|
"StringExecutePromptResponse",
|
361
387
|
"StringInputRequest",
|
362
388
|
"SubmitCompletionActualRequest",
|
@@ -22,12 +22,13 @@ from .resources.model_versions.client import AsyncModelVersionsClient, ModelVers
|
|
22
22
|
from .resources.registered_prompts.client import AsyncRegisteredPromptsClient, RegisteredPromptsClient
|
23
23
|
from .resources.sandboxes.client import AsyncSandboxesClient, SandboxesClient
|
24
24
|
from .resources.test_suites.client import AsyncTestSuitesClient, TestSuitesClient
|
25
|
+
from .types.execute_prompt_request import ExecutePromptRequest
|
25
26
|
from .types.execute_prompt_response import ExecutePromptResponse
|
27
|
+
from .types.execute_prompt_streaming_response import ExecutePromptStreamingResponse
|
26
28
|
from .types.generate_options_request import GenerateOptionsRequest
|
27
29
|
from .types.generate_request import GenerateRequest
|
28
30
|
from .types.generate_response import GenerateResponse
|
29
31
|
from .types.generate_stream_response import GenerateStreamResponse
|
30
|
-
from .types.prompt_deployment_input_request import PromptDeploymentInputRequest
|
31
32
|
from .types.search_request_options_request import SearchRequestOptionsRequest
|
32
33
|
from .types.search_response import SearchResponse
|
33
34
|
from .types.submit_completion_actual_request import SubmitCompletionActualRequest
|
@@ -67,42 +68,17 @@ class Vellum:
|
|
67
68
|
self.sandboxes = SandboxesClient(client_wrapper=self._client_wrapper)
|
68
69
|
self.test_suites = TestSuitesClient(client_wrapper=self._client_wrapper)
|
69
70
|
|
70
|
-
def execute_prompt(
|
71
|
-
self,
|
72
|
-
*,
|
73
|
-
inputs: typing.List[PromptDeploymentInputRequest],
|
74
|
-
prompt_deployment_id: typing.Optional[str] = OMIT,
|
75
|
-
prompt_deployment_name: typing.Optional[str] = OMIT,
|
76
|
-
release_tag: typing.Optional[str] = OMIT,
|
77
|
-
external_id: typing.Optional[str] = OMIT,
|
78
|
-
) -> ExecutePromptResponse:
|
71
|
+
def execute_prompt(self, *, request: ExecutePromptRequest) -> ExecutePromptResponse:
|
79
72
|
"""
|
80
73
|
Executes a deployed Prompt and returns the result.
|
81
74
|
|
82
75
|
Parameters:
|
83
|
-
-
|
84
|
-
|
85
|
-
- prompt_deployment_id: typing.Optional[str]. The ID of the Prompt Deployment. Must provide either this or prompt_deployment_name.
|
86
|
-
|
87
|
-
- prompt_deployment_name: typing.Optional[str]. The name of the Prompt Deployment. Must provide either this or prompt_deployment_id.
|
88
|
-
|
89
|
-
- release_tag: typing.Optional[str]. Optionally specify a release tag if you want to pin to a specific release of the Prompt Deployment
|
90
|
-
|
91
|
-
- external_id: typing.Optional[str].
|
76
|
+
- request: ExecutePromptRequest.
|
92
77
|
"""
|
93
|
-
_request: typing.Dict[str, typing.Any] = {"inputs": inputs}
|
94
|
-
if prompt_deployment_id is not OMIT:
|
95
|
-
_request["prompt_deployment_id"] = prompt_deployment_id
|
96
|
-
if prompt_deployment_name is not OMIT:
|
97
|
-
_request["prompt_deployment_name"] = prompt_deployment_name
|
98
|
-
if release_tag is not OMIT:
|
99
|
-
_request["release_tag"] = release_tag
|
100
|
-
if external_id is not OMIT:
|
101
|
-
_request["external_id"] = external_id
|
102
78
|
_response = self._client_wrapper.httpx_client.request(
|
103
79
|
"POST",
|
104
80
|
urllib.parse.urljoin(f"{self._client_wrapper.get_environment().default}/", "v1/execute-prompt"),
|
105
|
-
json=jsonable_encoder(
|
81
|
+
json=jsonable_encoder(request),
|
106
82
|
headers=self._client_wrapper.get_headers(),
|
107
83
|
timeout=None,
|
108
84
|
)
|
@@ -122,6 +98,36 @@ class Vellum:
|
|
122
98
|
raise ApiError(status_code=_response.status_code, body=_response.text)
|
123
99
|
raise ApiError(status_code=_response.status_code, body=_response_json)
|
124
100
|
|
101
|
+
def execute_prompt_stream(self, *, request: ExecutePromptRequest) -> ExecutePromptStreamingResponse:
|
102
|
+
"""
|
103
|
+
Executes a deployed Prompt and streams back the results.
|
104
|
+
|
105
|
+
Parameters:
|
106
|
+
- request: ExecutePromptRequest.
|
107
|
+
"""
|
108
|
+
_response = self._client_wrapper.httpx_client.request(
|
109
|
+
"POST",
|
110
|
+
urllib.parse.urljoin(f"{self._client_wrapper.get_environment().default}/", "v1/execute-prompt-stream"),
|
111
|
+
json=jsonable_encoder(request),
|
112
|
+
headers=self._client_wrapper.get_headers(),
|
113
|
+
timeout=None,
|
114
|
+
)
|
115
|
+
if 200 <= _response.status_code < 300:
|
116
|
+
return pydantic.parse_obj_as(ExecutePromptStreamingResponse, _response.json()) # type: ignore
|
117
|
+
if _response.status_code == 400:
|
118
|
+
raise BadRequestError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
|
119
|
+
if _response.status_code == 403:
|
120
|
+
raise ForbiddenError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
|
121
|
+
if _response.status_code == 404:
|
122
|
+
raise NotFoundError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
|
123
|
+
if _response.status_code == 500:
|
124
|
+
raise InternalServerError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
|
125
|
+
try:
|
126
|
+
_response_json = _response.json()
|
127
|
+
except JSONDecodeError:
|
128
|
+
raise ApiError(status_code=_response.status_code, body=_response.text)
|
129
|
+
raise ApiError(status_code=_response.status_code, body=_response_json)
|
130
|
+
|
125
131
|
def execute_workflow_stream(
|
126
132
|
self,
|
127
133
|
*,
|
@@ -488,42 +494,17 @@ class AsyncVellum:
|
|
488
494
|
self.sandboxes = AsyncSandboxesClient(client_wrapper=self._client_wrapper)
|
489
495
|
self.test_suites = AsyncTestSuitesClient(client_wrapper=self._client_wrapper)
|
490
496
|
|
491
|
-
async def execute_prompt(
|
492
|
-
self,
|
493
|
-
*,
|
494
|
-
inputs: typing.List[PromptDeploymentInputRequest],
|
495
|
-
prompt_deployment_id: typing.Optional[str] = OMIT,
|
496
|
-
prompt_deployment_name: typing.Optional[str] = OMIT,
|
497
|
-
release_tag: typing.Optional[str] = OMIT,
|
498
|
-
external_id: typing.Optional[str] = OMIT,
|
499
|
-
) -> ExecutePromptResponse:
|
497
|
+
async def execute_prompt(self, *, request: ExecutePromptRequest) -> ExecutePromptResponse:
|
500
498
|
"""
|
501
499
|
Executes a deployed Prompt and returns the result.
|
502
500
|
|
503
501
|
Parameters:
|
504
|
-
-
|
505
|
-
|
506
|
-
- prompt_deployment_id: typing.Optional[str]. The ID of the Prompt Deployment. Must provide either this or prompt_deployment_name.
|
507
|
-
|
508
|
-
- prompt_deployment_name: typing.Optional[str]. The name of the Prompt Deployment. Must provide either this or prompt_deployment_id.
|
509
|
-
|
510
|
-
- release_tag: typing.Optional[str]. Optionally specify a release tag if you want to pin to a specific release of the Prompt Deployment
|
511
|
-
|
512
|
-
- external_id: typing.Optional[str].
|
502
|
+
- request: ExecutePromptRequest.
|
513
503
|
"""
|
514
|
-
_request: typing.Dict[str, typing.Any] = {"inputs": inputs}
|
515
|
-
if prompt_deployment_id is not OMIT:
|
516
|
-
_request["prompt_deployment_id"] = prompt_deployment_id
|
517
|
-
if prompt_deployment_name is not OMIT:
|
518
|
-
_request["prompt_deployment_name"] = prompt_deployment_name
|
519
|
-
if release_tag is not OMIT:
|
520
|
-
_request["release_tag"] = release_tag
|
521
|
-
if external_id is not OMIT:
|
522
|
-
_request["external_id"] = external_id
|
523
504
|
_response = await self._client_wrapper.httpx_client.request(
|
524
505
|
"POST",
|
525
506
|
urllib.parse.urljoin(f"{self._client_wrapper.get_environment().default}/", "v1/execute-prompt"),
|
526
|
-
json=jsonable_encoder(
|
507
|
+
json=jsonable_encoder(request),
|
527
508
|
headers=self._client_wrapper.get_headers(),
|
528
509
|
timeout=None,
|
529
510
|
)
|
@@ -543,6 +524,36 @@ class AsyncVellum:
|
|
543
524
|
raise ApiError(status_code=_response.status_code, body=_response.text)
|
544
525
|
raise ApiError(status_code=_response.status_code, body=_response_json)
|
545
526
|
|
527
|
+
async def execute_prompt_stream(self, *, request: ExecutePromptRequest) -> ExecutePromptStreamingResponse:
|
528
|
+
"""
|
529
|
+
Executes a deployed Prompt and streams back the results.
|
530
|
+
|
531
|
+
Parameters:
|
532
|
+
- request: ExecutePromptRequest.
|
533
|
+
"""
|
534
|
+
_response = await self._client_wrapper.httpx_client.request(
|
535
|
+
"POST",
|
536
|
+
urllib.parse.urljoin(f"{self._client_wrapper.get_environment().default}/", "v1/execute-prompt-stream"),
|
537
|
+
json=jsonable_encoder(request),
|
538
|
+
headers=self._client_wrapper.get_headers(),
|
539
|
+
timeout=None,
|
540
|
+
)
|
541
|
+
if 200 <= _response.status_code < 300:
|
542
|
+
return pydantic.parse_obj_as(ExecutePromptStreamingResponse, _response.json()) # type: ignore
|
543
|
+
if _response.status_code == 400:
|
544
|
+
raise BadRequestError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
|
545
|
+
if _response.status_code == 403:
|
546
|
+
raise ForbiddenError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
|
547
|
+
if _response.status_code == 404:
|
548
|
+
raise NotFoundError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
|
549
|
+
if _response.status_code == 500:
|
550
|
+
raise InternalServerError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
|
551
|
+
try:
|
552
|
+
_response_json = _response.json()
|
553
|
+
except JSONDecodeError:
|
554
|
+
raise ApiError(status_code=_response.status_code, body=_response.text)
|
555
|
+
raise ApiError(status_code=_response.status_code, body=_response_json)
|
556
|
+
|
546
557
|
async def execute_workflow_stream(
|
547
558
|
self,
|
548
559
|
*,
|
@@ -16,7 +16,7 @@ class BaseClientWrapper:
|
|
16
16
|
headers: typing.Dict[str, str] = {
|
17
17
|
"X-Fern-Language": "Python",
|
18
18
|
"X-Fern-SDK-Name": "vellum-ai",
|
19
|
-
"X-Fern-SDK-Version": "
|
19
|
+
"X-Fern-SDK-Version": "0.1.7",
|
20
20
|
}
|
21
21
|
headers["X_API_KEY"] = self.api_key
|
22
22
|
return headers
|
@@ -0,0 +1,164 @@
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
2
|
+
|
3
|
+
import typing
|
4
|
+
import urllib.parse
|
5
|
+
from json.decoder import JSONDecodeError
|
6
|
+
|
7
|
+
from ...core.api_error import ApiError
|
8
|
+
from ...core.client_wrapper import AsyncClientWrapper, SyncClientWrapper
|
9
|
+
from ...core.jsonable_encoder import jsonable_encoder
|
10
|
+
from ...types.deployment_provider_payload_response import DeploymentProviderPayloadResponse
|
11
|
+
from ...types.deployment_read import DeploymentRead
|
12
|
+
from ...types.prompt_deployment_input_request import PromptDeploymentInputRequest
|
13
|
+
|
14
|
+
try:
|
15
|
+
import pydantic.v1 as pydantic # type: ignore
|
16
|
+
except ImportError:
|
17
|
+
import pydantic # type: ignore
|
18
|
+
|
19
|
+
# this is used as the default value for optional parameters
|
20
|
+
OMIT = typing.cast(typing.Any, ...)
|
21
|
+
|
22
|
+
|
23
|
+
class DeploymentsClient:
|
24
|
+
def __init__(self, *, client_wrapper: SyncClientWrapper):
|
25
|
+
self._client_wrapper = client_wrapper
|
26
|
+
|
27
|
+
def retrieve(self, id: str) -> DeploymentRead:
|
28
|
+
"""
|
29
|
+
Used to retrieve a deployment given its ID or name.
|
30
|
+
|
31
|
+
Parameters:
|
32
|
+
- id: str. Either the Deployment's ID or its unique name
|
33
|
+
"""
|
34
|
+
_response = self._client_wrapper.httpx_client.request(
|
35
|
+
"GET",
|
36
|
+
urllib.parse.urljoin(f"{self._client_wrapper.get_environment().default}/", f"v1/deployments/{id}"),
|
37
|
+
headers=self._client_wrapper.get_headers(),
|
38
|
+
timeout=None,
|
39
|
+
)
|
40
|
+
if 200 <= _response.status_code < 300:
|
41
|
+
return pydantic.parse_obj_as(DeploymentRead, _response.json()) # type: ignore
|
42
|
+
try:
|
43
|
+
_response_json = _response.json()
|
44
|
+
except JSONDecodeError:
|
45
|
+
raise ApiError(status_code=_response.status_code, body=_response.text)
|
46
|
+
raise ApiError(status_code=_response.status_code, body=_response_json)
|
47
|
+
|
48
|
+
def retrieve_provider_payload(
|
49
|
+
self,
|
50
|
+
*,
|
51
|
+
deployment_id: typing.Optional[str] = OMIT,
|
52
|
+
deployment_name: typing.Optional[str] = OMIT,
|
53
|
+
inputs: typing.List[PromptDeploymentInputRequest],
|
54
|
+
) -> DeploymentProviderPayloadResponse:
|
55
|
+
"""
|
56
|
+
Parameters:
|
57
|
+
- deployment_id: typing.Optional[str]. The ID of the deployment. Must provide either this or deployment_name.
|
58
|
+
|
59
|
+
- deployment_name: typing.Optional[str]. The name of the deployment. Must provide either this or deployment_id.
|
60
|
+
|
61
|
+
- inputs: typing.List[PromptDeploymentInputRequest].
|
62
|
+
---
|
63
|
+
from vellum.client import Vellum
|
64
|
+
|
65
|
+
client = Vellum(
|
66
|
+
api_key="YOUR_API_KEY",
|
67
|
+
)
|
68
|
+
client.deployments.retrieve_provider_payload(
|
69
|
+
inputs=[],
|
70
|
+
)
|
71
|
+
"""
|
72
|
+
_request: typing.Dict[str, typing.Any] = {"inputs": inputs}
|
73
|
+
if deployment_id is not OMIT:
|
74
|
+
_request["deployment_id"] = deployment_id
|
75
|
+
if deployment_name is not OMIT:
|
76
|
+
_request["deployment_name"] = deployment_name
|
77
|
+
_response = self._client_wrapper.httpx_client.request(
|
78
|
+
"POST",
|
79
|
+
urllib.parse.urljoin(
|
80
|
+
f"{self._client_wrapper.get_environment().default}/", "v1/deployments/provider-payload"
|
81
|
+
),
|
82
|
+
json=jsonable_encoder(_request),
|
83
|
+
headers=self._client_wrapper.get_headers(),
|
84
|
+
timeout=None,
|
85
|
+
)
|
86
|
+
if 200 <= _response.status_code < 300:
|
87
|
+
return pydantic.parse_obj_as(DeploymentProviderPayloadResponse, _response.json()) # type: ignore
|
88
|
+
try:
|
89
|
+
_response_json = _response.json()
|
90
|
+
except JSONDecodeError:
|
91
|
+
raise ApiError(status_code=_response.status_code, body=_response.text)
|
92
|
+
raise ApiError(status_code=_response.status_code, body=_response_json)
|
93
|
+
|
94
|
+
|
95
|
+
class AsyncDeploymentsClient:
|
96
|
+
def __init__(self, *, client_wrapper: AsyncClientWrapper):
|
97
|
+
self._client_wrapper = client_wrapper
|
98
|
+
|
99
|
+
async def retrieve(self, id: str) -> DeploymentRead:
|
100
|
+
"""
|
101
|
+
Used to retrieve a deployment given its ID or name.
|
102
|
+
|
103
|
+
Parameters:
|
104
|
+
- id: str. Either the Deployment's ID or its unique name
|
105
|
+
"""
|
106
|
+
_response = await self._client_wrapper.httpx_client.request(
|
107
|
+
"GET",
|
108
|
+
urllib.parse.urljoin(f"{self._client_wrapper.get_environment().default}/", f"v1/deployments/{id}"),
|
109
|
+
headers=self._client_wrapper.get_headers(),
|
110
|
+
timeout=None,
|
111
|
+
)
|
112
|
+
if 200 <= _response.status_code < 300:
|
113
|
+
return pydantic.parse_obj_as(DeploymentRead, _response.json()) # type: ignore
|
114
|
+
try:
|
115
|
+
_response_json = _response.json()
|
116
|
+
except JSONDecodeError:
|
117
|
+
raise ApiError(status_code=_response.status_code, body=_response.text)
|
118
|
+
raise ApiError(status_code=_response.status_code, body=_response_json)
|
119
|
+
|
120
|
+
async def retrieve_provider_payload(
|
121
|
+
self,
|
122
|
+
*,
|
123
|
+
deployment_id: typing.Optional[str] = OMIT,
|
124
|
+
deployment_name: typing.Optional[str] = OMIT,
|
125
|
+
inputs: typing.List[PromptDeploymentInputRequest],
|
126
|
+
) -> DeploymentProviderPayloadResponse:
|
127
|
+
"""
|
128
|
+
Parameters:
|
129
|
+
- deployment_id: typing.Optional[str]. The ID of the deployment. Must provide either this or deployment_name.
|
130
|
+
|
131
|
+
- deployment_name: typing.Optional[str]. The name of the deployment. Must provide either this or deployment_id.
|
132
|
+
|
133
|
+
- inputs: typing.List[PromptDeploymentInputRequest].
|
134
|
+
---
|
135
|
+
from vellum.client import AsyncVellum
|
136
|
+
|
137
|
+
client = AsyncVellum(
|
138
|
+
api_key="YOUR_API_KEY",
|
139
|
+
)
|
140
|
+
await client.deployments.retrieve_provider_payload(
|
141
|
+
inputs=[],
|
142
|
+
)
|
143
|
+
"""
|
144
|
+
_request: typing.Dict[str, typing.Any] = {"inputs": inputs}
|
145
|
+
if deployment_id is not OMIT:
|
146
|
+
_request["deployment_id"] = deployment_id
|
147
|
+
if deployment_name is not OMIT:
|
148
|
+
_request["deployment_name"] = deployment_name
|
149
|
+
_response = await self._client_wrapper.httpx_client.request(
|
150
|
+
"POST",
|
151
|
+
urllib.parse.urljoin(
|
152
|
+
f"{self._client_wrapper.get_environment().default}/", "v1/deployments/provider-payload"
|
153
|
+
),
|
154
|
+
json=jsonable_encoder(_request),
|
155
|
+
headers=self._client_wrapper.get_headers(),
|
156
|
+
timeout=None,
|
157
|
+
)
|
158
|
+
if 200 <= _response.status_code < 300:
|
159
|
+
return pydantic.parse_obj_as(DeploymentProviderPayloadResponse, _response.json()) # type: ignore
|
160
|
+
try:
|
161
|
+
_response_json = _response.json()
|
162
|
+
except JSONDecodeError:
|
163
|
+
raise ApiError(status_code=_response.status_code, body=_response.text)
|
164
|
+
raise ApiError(status_code=_response.status_code, body=_response_json)
|
@@ -19,7 +19,7 @@ class ModelVersionsClient:
|
|
19
19
|
|
20
20
|
def retrieve(self, id: str) -> ModelVersionRead:
|
21
21
|
"""
|
22
|
-
|
22
|
+
Deprecated. Use the `deployments/provider-payload` endpoint to fetch information that we send to Model providers.
|
23
23
|
|
24
24
|
Parameters:
|
25
25
|
- id: str. A UUID string identifying this model version.
|
@@ -45,7 +45,7 @@ class AsyncModelVersionsClient:
|
|
45
45
|
|
46
46
|
async def retrieve(self, id: str) -> ModelVersionRead:
|
47
47
|
"""
|
48
|
-
|
48
|
+
Deprecated. Use the `deployments/provider-payload` endpoint to fetch information that we send to Model providers.
|
49
49
|
|
50
50
|
Parameters:
|
51
51
|
- id: str. A UUID string identifying this model version.
|
@@ -9,6 +9,7 @@ from .chat_message_request import ChatMessageRequest
|
|
9
9
|
from .chat_message_role import ChatMessageRole
|
10
10
|
from .conditional_node_result import ConditionalNodeResult
|
11
11
|
from .conditional_node_result_data import ConditionalNodeResultData
|
12
|
+
from .deployment_provider_payload_response import DeploymentProviderPayloadResponse
|
12
13
|
from .deployment_read import DeploymentRead
|
13
14
|
from .deployment_status import DeploymentStatus
|
14
15
|
from .document_document_to_document_index import DocumentDocumentToDocumentIndex
|
@@ -22,14 +23,23 @@ from .error_execute_prompt_response import ErrorExecutePromptResponse
|
|
22
23
|
from .evaluation_params import EvaluationParams
|
23
24
|
from .evaluation_params_request import EvaluationParamsRequest
|
24
25
|
from .execute_prompt_api_error_response import ExecutePromptApiErrorResponse
|
26
|
+
from .execute_prompt_request import ExecutePromptRequest
|
25
27
|
from .execute_prompt_response import (
|
26
28
|
ExecutePromptResponse,
|
27
29
|
ExecutePromptResponse_Error,
|
28
30
|
ExecutePromptResponse_Json,
|
29
31
|
ExecutePromptResponse_String,
|
30
32
|
)
|
33
|
+
from .execute_prompt_streaming_response import (
|
34
|
+
ExecutePromptStreamingResponse,
|
35
|
+
ExecutePromptStreamingResponse_Fulfilled,
|
36
|
+
ExecutePromptStreamingResponse_Initiated,
|
37
|
+
ExecutePromptStreamingResponse_Rejected,
|
38
|
+
ExecutePromptStreamingResponse_Streaming,
|
39
|
+
)
|
31
40
|
from .execute_workflow_stream_error_response import ExecuteWorkflowStreamErrorResponse
|
32
41
|
from .finish_reason_enum import FinishReasonEnum
|
42
|
+
from .fulfilled_execute_prompt_streaming_response import FulfilledExecutePromptStreamingResponse
|
33
43
|
from .generate_error_response import GenerateErrorResponse
|
34
44
|
from .generate_options_request import GenerateOptionsRequest
|
35
45
|
from .generate_request import GenerateRequest
|
@@ -41,6 +51,8 @@ from .generate_stream_response import GenerateStreamResponse
|
|
41
51
|
from .generate_stream_result import GenerateStreamResult
|
42
52
|
from .generate_stream_result_data import GenerateStreamResultData
|
43
53
|
from .indexing_state_enum import IndexingStateEnum
|
54
|
+
from .initiated_execute_prompt_response import InitiatedExecutePromptResponse
|
55
|
+
from .initiated_execute_prompt_streaming_response import InitiatedExecutePromptStreamingResponse
|
44
56
|
from .json_execute_prompt_response import JsonExecutePromptResponse
|
45
57
|
from .json_input_request import JsonInputRequest
|
46
58
|
from .logical_operator import LogicalOperator
|
@@ -99,6 +111,8 @@ from .registered_prompt_input_variable_request import RegisteredPromptInputVaria
|
|
99
111
|
from .registered_prompt_model_version import RegisteredPromptModelVersion
|
100
112
|
from .registered_prompt_sandbox import RegisteredPromptSandbox
|
101
113
|
from .registered_prompt_sandbox_snapshot import RegisteredPromptSandboxSnapshot
|
114
|
+
from .rejected_execute_prompt_response import RejectedExecutePromptResponse
|
115
|
+
from .rejected_execute_prompt_streaming_response import RejectedExecutePromptStreamingResponse
|
102
116
|
from .sandbox_metric_input_params import SandboxMetricInputParams
|
103
117
|
from .sandbox_metric_input_params_request import SandboxMetricInputParamsRequest
|
104
118
|
from .sandbox_scenario import SandboxScenario
|
@@ -118,6 +132,7 @@ from .search_result_merging_request import SearchResultMergingRequest
|
|
118
132
|
from .search_result_request import SearchResultRequest
|
119
133
|
from .search_weights_request import SearchWeightsRequest
|
120
134
|
from .slim_document import SlimDocument
|
135
|
+
from .streaming_execute_prompt_streaming_response import StreamingExecutePromptStreamingResponse
|
121
136
|
from .string_execute_prompt_response import StringExecutePromptResponse
|
122
137
|
from .string_input_request import StringInputRequest
|
123
138
|
from .submit_completion_actual_request import SubmitCompletionActualRequest
|
@@ -256,6 +271,7 @@ __all__ = [
|
|
256
271
|
"ChatMessageRole",
|
257
272
|
"ConditionalNodeResult",
|
258
273
|
"ConditionalNodeResultData",
|
274
|
+
"DeploymentProviderPayloadResponse",
|
259
275
|
"DeploymentRead",
|
260
276
|
"DeploymentStatus",
|
261
277
|
"DocumentDocumentToDocumentIndex",
|
@@ -269,12 +285,19 @@ __all__ = [
|
|
269
285
|
"EvaluationParams",
|
270
286
|
"EvaluationParamsRequest",
|
271
287
|
"ExecutePromptApiErrorResponse",
|
288
|
+
"ExecutePromptRequest",
|
272
289
|
"ExecutePromptResponse",
|
273
290
|
"ExecutePromptResponse_Error",
|
274
291
|
"ExecutePromptResponse_Json",
|
275
292
|
"ExecutePromptResponse_String",
|
293
|
+
"ExecutePromptStreamingResponse",
|
294
|
+
"ExecutePromptStreamingResponse_Fulfilled",
|
295
|
+
"ExecutePromptStreamingResponse_Initiated",
|
296
|
+
"ExecutePromptStreamingResponse_Rejected",
|
297
|
+
"ExecutePromptStreamingResponse_Streaming",
|
276
298
|
"ExecuteWorkflowStreamErrorResponse",
|
277
299
|
"FinishReasonEnum",
|
300
|
+
"FulfilledExecutePromptStreamingResponse",
|
278
301
|
"GenerateErrorResponse",
|
279
302
|
"GenerateOptionsRequest",
|
280
303
|
"GenerateRequest",
|
@@ -286,6 +309,8 @@ __all__ = [
|
|
286
309
|
"GenerateStreamResult",
|
287
310
|
"GenerateStreamResultData",
|
288
311
|
"IndexingStateEnum",
|
312
|
+
"InitiatedExecutePromptResponse",
|
313
|
+
"InitiatedExecutePromptStreamingResponse",
|
289
314
|
"JsonExecutePromptResponse",
|
290
315
|
"JsonInputRequest",
|
291
316
|
"LogicalOperator",
|
@@ -340,6 +365,8 @@ __all__ = [
|
|
340
365
|
"RegisteredPromptModelVersion",
|
341
366
|
"RegisteredPromptSandbox",
|
342
367
|
"RegisteredPromptSandboxSnapshot",
|
368
|
+
"RejectedExecutePromptResponse",
|
369
|
+
"RejectedExecutePromptStreamingResponse",
|
343
370
|
"SandboxMetricInputParams",
|
344
371
|
"SandboxMetricInputParamsRequest",
|
345
372
|
"SandboxScenario",
|
@@ -359,6 +386,7 @@ __all__ = [
|
|
359
386
|
"SearchResultRequest",
|
360
387
|
"SearchWeightsRequest",
|
361
388
|
"SlimDocument",
|
389
|
+
"StreamingExecutePromptStreamingResponse",
|
362
390
|
"StringExecutePromptResponse",
|
363
391
|
"StringInputRequest",
|
364
392
|
"SubmitCompletionActualRequest",
|
@@ -0,0 +1,28 @@
|
|
1
|
+
# This file was auto-generated by Fern from our API Definition.
|
2
|
+
|
3
|
+
import datetime as dt
|
4
|
+
import typing
|
5
|
+
|
6
|
+
from ..core.datetime_utils import serialize_datetime
|
7
|
+
|
8
|
+
try:
|
9
|
+
import pydantic.v1 as pydantic # type: ignore
|
10
|
+
except ImportError:
|
11
|
+
import pydantic # type: ignore
|
12
|
+
|
13
|
+
|
14
|
+
class DeploymentProviderPayloadResponse(pydantic.BaseModel):
|
15
|
+
payload: typing.Dict[str, typing.Any]
|
16
|
+
|
17
|
+
def json(self, **kwargs: typing.Any) -> str:
|
18
|
+
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
19
|
+
return super().json(**kwargs_with_defaults)
|
20
|
+
|
21
|
+
def dict(self, **kwargs: typing.Any) -> typing.Dict[str, typing.Any]:
|
22
|
+
kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
|
23
|
+
return super().dict(**kwargs_with_defaults)
|
24
|
+
|
25
|
+
class Config:
|
26
|
+
frozen = True
|
27
|
+
smart_union = True
|
28
|
+
json_encoders = {dt.datetime: serialize_datetime}
|
@@ -37,7 +37,9 @@ class DeploymentRead(pydantic.BaseModel):
|
|
37
37
|
"* `PRODUCTION` - Production\n"
|
38
38
|
)
|
39
39
|
)
|
40
|
-
active_model_version_ids: typing.List[str]
|
40
|
+
active_model_version_ids: typing.List[str] = pydantic.Field(
|
41
|
+
description="Deprecated. The Prompt execution endpoints return a `prompt_version_id` that could be used instead."
|
42
|
+
)
|
41
43
|
last_deployed_on: dt.datetime
|
42
44
|
input_variables: typing.List[VellumVariable]
|
43
45
|
|