vellum-ai 0.1.5__tar.gz → 0.1.7__tar.gz

Sign up to get free protection for your applications and to get access to all the features.
Files changed (214) hide show
  1. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/PKG-INFO +1 -1
  2. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/pyproject.toml +1 -1
  3. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/__init__.py +26 -0
  4. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/client.py +68 -57
  5. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/core/client_wrapper.py +1 -1
  6. vellum_ai-0.1.7/src/vellum/resources/deployments/client.py +164 -0
  7. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/resources/model_versions/client.py +2 -2
  8. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/__init__.py +28 -0
  9. vellum_ai-0.1.7/src/vellum/types/deployment_provider_payload_response.py +28 -0
  10. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/deployment_read.py +3 -1
  11. vellum_ai-0.1.7/src/vellum/types/execute_prompt_request.py +39 -0
  12. vellum_ai-0.1.7/src/vellum/types/execute_prompt_streaming_response.py +56 -0
  13. vellum_ai-0.1.7/src/vellum/types/fulfilled_execute_prompt_streaming_response.py +29 -0
  14. vellum_ai-0.1.7/src/vellum/types/initiated_execute_prompt_response.py +30 -0
  15. vellum_ai-0.1.7/src/vellum/types/initiated_execute_prompt_streaming_response.py +29 -0
  16. vellum_ai-0.1.7/src/vellum/types/rejected_execute_prompt_response.py +31 -0
  17. vellum_ai-0.1.7/src/vellum/types/rejected_execute_prompt_streaming_response.py +29 -0
  18. vellum_ai-0.1.7/src/vellum/types/streaming_execute_prompt_streaming_response.py +29 -0
  19. vellum_ai-0.1.5/src/vellum/resources/deployments/client.py +0 -65
  20. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/README.md +0 -0
  21. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/core/__init__.py +0 -0
  22. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/core/api_error.py +0 -0
  23. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/core/datetime_utils.py +0 -0
  24. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/core/jsonable_encoder.py +0 -0
  25. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/core/remove_none_from_dict.py +0 -0
  26. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/environment.py +0 -0
  27. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/errors/__init__.py +0 -0
  28. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/errors/bad_request_error.py +0 -0
  29. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/errors/conflict_error.py +0 -0
  30. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/errors/forbidden_error.py +0 -0
  31. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/errors/internal_server_error.py +0 -0
  32. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/errors/not_found_error.py +0 -0
  33. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/py.typed +0 -0
  34. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/resources/__init__.py +0 -0
  35. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/resources/deployments/__init__.py +0 -0
  36. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/resources/document_indexes/__init__.py +0 -0
  37. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/resources/document_indexes/client.py +0 -0
  38. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/resources/documents/__init__.py +0 -0
  39. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/resources/documents/client.py +0 -0
  40. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/resources/model_versions/__init__.py +0 -0
  41. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/resources/registered_prompts/__init__.py +0 -0
  42. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/resources/registered_prompts/client.py +0 -0
  43. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/resources/sandboxes/__init__.py +0 -0
  44. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/resources/sandboxes/client.py +0 -0
  45. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/resources/test_suites/__init__.py +0 -0
  46. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/resources/test_suites/client.py +0 -0
  47. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/api_node_result.py +0 -0
  48. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/api_node_result_data.py +0 -0
  49. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/block_type_enum.py +0 -0
  50. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/chat_history_input_request.py +0 -0
  51. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/chat_message.py +0 -0
  52. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/chat_message_request.py +0 -0
  53. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/chat_message_role.py +0 -0
  54. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/conditional_node_result.py +0 -0
  55. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/conditional_node_result_data.py +0 -0
  56. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/deployment_status.py +0 -0
  57. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/document_document_to_document_index.py +0 -0
  58. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/document_index_read.py +0 -0
  59. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/document_index_status.py +0 -0
  60. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/document_read.py +0 -0
  61. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/document_status.py +0 -0
  62. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/enriched_normalized_completion.py +0 -0
  63. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/environment_enum.py +0 -0
  64. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/error_execute_prompt_response.py +1 -1
  65. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/evaluation_params.py +0 -0
  66. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/evaluation_params_request.py +0 -0
  67. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/execute_prompt_api_error_response.py +0 -0
  68. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/execute_prompt_response.py +0 -0
  69. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/execute_workflow_stream_error_response.py +0 -0
  70. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/finish_reason_enum.py +0 -0
  71. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/generate_error_response.py +0 -0
  72. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/generate_options_request.py +0 -0
  73. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/generate_request.py +0 -0
  74. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/generate_response.py +0 -0
  75. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/generate_result.py +0 -0
  76. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/generate_result_data.py +0 -0
  77. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/generate_result_error.py +0 -0
  78. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/generate_stream_response.py +0 -0
  79. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/generate_stream_result.py +0 -0
  80. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/generate_stream_result_data.py +0 -0
  81. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/indexing_state_enum.py +0 -0
  82. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/json_execute_prompt_response.py +1 -1
  83. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/json_input_request.py +0 -0
  84. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/logical_operator.py +0 -0
  85. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/logprobs_enum.py +0 -0
  86. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/metadata_filter_config_request.py +0 -0
  87. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/metadata_filter_rule_combinator.py +0 -0
  88. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/metadata_filter_rule_request.py +0 -0
  89. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/model_version_build_config.py +0 -0
  90. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/model_version_exec_config.py +0 -0
  91. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/model_version_exec_config_parameters.py +0 -0
  92. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/model_version_read.py +0 -0
  93. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/model_version_read_status_enum.py +0 -0
  94. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/model_version_sandbox_snapshot.py +0 -0
  95. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/node_input_compiled_chat_history_value.py +0 -0
  96. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/node_input_compiled_error_value.py +0 -0
  97. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/node_input_compiled_json_value.py +0 -0
  98. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/node_input_compiled_number_value.py +0 -0
  99. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/node_input_compiled_search_results_value.py +0 -0
  100. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/node_input_compiled_string_value.py +0 -0
  101. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/node_input_variable_compiled_value.py +0 -0
  102. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/normalized_log_probs.py +0 -0
  103. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/normalized_token_log_probs.py +0 -0
  104. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/paginated_slim_document_list.py +0 -0
  105. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/processing_failure_reason_enum.py +0 -0
  106. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/processing_state_enum.py +0 -0
  107. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/prompt_deployment_input_request.py +0 -0
  108. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/prompt_node_result.py +0 -0
  109. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/prompt_node_result_data.py +0 -0
  110. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/prompt_template_block.py +0 -0
  111. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/prompt_template_block_data.py +0 -0
  112. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/prompt_template_block_data_request.py +0 -0
  113. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/prompt_template_block_properties.py +0 -0
  114. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/prompt_template_block_properties_request.py +0 -0
  115. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/prompt_template_block_request.py +0 -0
  116. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/provider_enum.py +0 -0
  117. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/register_prompt_error_response.py +0 -0
  118. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/register_prompt_model_parameters_request.py +0 -0
  119. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/register_prompt_prompt.py +0 -0
  120. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/register_prompt_prompt_info_request.py +0 -0
  121. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/register_prompt_response.py +0 -0
  122. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/registered_prompt_deployment.py +0 -0
  123. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/registered_prompt_input_variable_request.py +0 -0
  124. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/registered_prompt_model_version.py +0 -0
  125. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/registered_prompt_sandbox.py +0 -0
  126. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/registered_prompt_sandbox_snapshot.py +0 -0
  127. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/sandbox_metric_input_params.py +0 -0
  128. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/sandbox_metric_input_params_request.py +0 -0
  129. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/sandbox_scenario.py +0 -0
  130. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/scenario_input.py +0 -0
  131. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/scenario_input_request.py +0 -0
  132. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/scenario_input_type_enum.py +0 -0
  133. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/search_error_response.py +0 -0
  134. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/search_filters_request.py +0 -0
  135. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/search_node_result.py +0 -0
  136. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/search_node_result_data.py +0 -0
  137. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/search_request_options_request.py +0 -0
  138. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/search_response.py +0 -0
  139. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/search_result.py +0 -0
  140. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/search_result_document.py +0 -0
  141. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/search_result_document_request.py +0 -0
  142. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/search_result_merging_request.py +0 -0
  143. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/search_result_request.py +0 -0
  144. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/search_weights_request.py +0 -0
  145. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/slim_document.py +0 -0
  146. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/string_execute_prompt_response.py +1 -1
  147. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/string_input_request.py +0 -0
  148. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/submit_completion_actual_request.py +0 -0
  149. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/submit_completion_actuals_error_response.py +0 -0
  150. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/submit_workflow_execution_actual_request.py +0 -0
  151. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/templating_node_chat_history_result.py +0 -0
  152. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/templating_node_error_result.py +0 -0
  153. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/templating_node_json_result.py +0 -0
  154. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/templating_node_number_result.py +0 -0
  155. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/templating_node_result.py +0 -0
  156. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/templating_node_result_data.py +0 -0
  157. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/templating_node_result_output.py +0 -0
  158. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/templating_node_search_results_result.py +0 -0
  159. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/templating_node_string_result.py +0 -0
  160. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/terminal_node_chat_history_result.py +0 -0
  161. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/terminal_node_error_result.py +0 -0
  162. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/terminal_node_json_result.py +0 -0
  163. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/terminal_node_number_result.py +0 -0
  164. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/terminal_node_result.py +0 -0
  165. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/terminal_node_result_data.py +0 -0
  166. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/terminal_node_result_output.py +0 -0
  167. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/terminal_node_search_results_result.py +0 -0
  168. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/terminal_node_string_result.py +0 -0
  169. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/test_case_chat_history_variable_value.py +0 -0
  170. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/test_case_chat_history_variable_value_request.py +0 -0
  171. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/test_case_error_variable_value.py +0 -0
  172. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/test_case_error_variable_value_request.py +0 -0
  173. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/test_case_json_variable_value.py +0 -0
  174. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/test_case_json_variable_value_request.py +0 -0
  175. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/test_case_number_variable_value.py +0 -0
  176. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/test_case_number_variable_value_request.py +0 -0
  177. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/test_case_search_results_variable_value.py +0 -0
  178. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/test_case_search_results_variable_value_request.py +0 -0
  179. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/test_case_string_variable_value.py +0 -0
  180. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/test_case_string_variable_value_request.py +0 -0
  181. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/test_case_variable_value.py +0 -0
  182. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/test_case_variable_value_request.py +0 -0
  183. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/test_suite_test_case.py +0 -0
  184. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/upload_document_error_response.py +0 -0
  185. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/upload_document_response.py +0 -0
  186. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/vellum_error.py +0 -0
  187. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/vellum_error_code_enum.py +0 -0
  188. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/vellum_error_request.py +0 -0
  189. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/vellum_variable.py +0 -0
  190. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/vellum_variable_type.py +0 -0
  191. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_event_error.py +0 -0
  192. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_execution_actual_chat_history_request.py +0 -0
  193. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_execution_actual_json_request.py +0 -0
  194. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_execution_actual_string_request.py +0 -0
  195. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_execution_event_error_code.py +0 -0
  196. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_execution_event_type.py +0 -0
  197. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_execution_node_result_event.py +0 -0
  198. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_execution_workflow_result_event.py +0 -0
  199. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_node_result_data.py +0 -0
  200. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_node_result_event.py +0 -0
  201. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_node_result_event_state.py +0 -0
  202. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_request_chat_history_input_request.py +0 -0
  203. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_request_input_request.py +0 -0
  204. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_request_json_input_request.py +0 -0
  205. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_request_string_input_request.py +0 -0
  206. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_result_event.py +0 -0
  207. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_result_event_output_data.py +0 -0
  208. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_result_event_output_data_chat_history.py +0 -0
  209. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_result_event_output_data_error.py +0 -0
  210. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_result_event_output_data_json.py +0 -0
  211. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_result_event_output_data_number.py +0 -0
  212. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_result_event_output_data_search_results.py +0 -0
  213. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_result_event_output_data_string.py +0 -0
  214. {vellum_ai-0.1.5 → vellum_ai-0.1.7}/src/vellum/types/workflow_stream_event.py +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: vellum-ai
3
- Version: 0.1.5
3
+ Version: 0.1.7
4
4
  Summary:
5
5
  Requires-Python: >=3.7,<4.0
6
6
  Classifier: Programming Language :: Python :: 3
@@ -1,6 +1,6 @@
1
1
  [tool.poetry]
2
2
  name = "vellum-ai"
3
- version = "v0.1.5"
3
+ version = "0.1.7"
4
4
  description = ""
5
5
  readme = "README.md"
6
6
  authors = []
@@ -10,6 +10,7 @@ from .types import (
10
10
  ChatMessageRole,
11
11
  ConditionalNodeResult,
12
12
  ConditionalNodeResultData,
13
+ DeploymentProviderPayloadResponse,
13
14
  DeploymentRead,
14
15
  DeploymentStatus,
15
16
  DocumentDocumentToDocumentIndex,
@@ -23,12 +24,19 @@ from .types import (
23
24
  EvaluationParams,
24
25
  EvaluationParamsRequest,
25
26
  ExecutePromptApiErrorResponse,
27
+ ExecutePromptRequest,
26
28
  ExecutePromptResponse,
27
29
  ExecutePromptResponse_Error,
28
30
  ExecutePromptResponse_Json,
29
31
  ExecutePromptResponse_String,
32
+ ExecutePromptStreamingResponse,
33
+ ExecutePromptStreamingResponse_Fulfilled,
34
+ ExecutePromptStreamingResponse_Initiated,
35
+ ExecutePromptStreamingResponse_Rejected,
36
+ ExecutePromptStreamingResponse_Streaming,
30
37
  ExecuteWorkflowStreamErrorResponse,
31
38
  FinishReasonEnum,
39
+ FulfilledExecutePromptStreamingResponse,
32
40
  GenerateErrorResponse,
33
41
  GenerateOptionsRequest,
34
42
  GenerateRequest,
@@ -40,6 +48,8 @@ from .types import (
40
48
  GenerateStreamResult,
41
49
  GenerateStreamResultData,
42
50
  IndexingStateEnum,
51
+ InitiatedExecutePromptResponse,
52
+ InitiatedExecutePromptStreamingResponse,
43
53
  JsonExecutePromptResponse,
44
54
  JsonInputRequest,
45
55
  LogicalOperator,
@@ -94,6 +104,8 @@ from .types import (
94
104
  RegisteredPromptModelVersion,
95
105
  RegisteredPromptSandbox,
96
106
  RegisteredPromptSandboxSnapshot,
107
+ RejectedExecutePromptResponse,
108
+ RejectedExecutePromptStreamingResponse,
97
109
  SandboxMetricInputParams,
98
110
  SandboxMetricInputParamsRequest,
99
111
  SandboxScenario,
@@ -113,6 +125,7 @@ from .types import (
113
125
  SearchResultRequest,
114
126
  SearchWeightsRequest,
115
127
  SlimDocument,
128
+ StreamingExecutePromptStreamingResponse,
116
129
  StringExecutePromptResponse,
117
130
  StringInputRequest,
118
131
  SubmitCompletionActualRequest,
@@ -251,6 +264,7 @@ __all__ = [
251
264
  "ConditionalNodeResult",
252
265
  "ConditionalNodeResultData",
253
266
  "ConflictError",
267
+ "DeploymentProviderPayloadResponse",
254
268
  "DeploymentRead",
255
269
  "DeploymentStatus",
256
270
  "DocumentDocumentToDocumentIndex",
@@ -264,13 +278,20 @@ __all__ = [
264
278
  "EvaluationParams",
265
279
  "EvaluationParamsRequest",
266
280
  "ExecutePromptApiErrorResponse",
281
+ "ExecutePromptRequest",
267
282
  "ExecutePromptResponse",
268
283
  "ExecutePromptResponse_Error",
269
284
  "ExecutePromptResponse_Json",
270
285
  "ExecutePromptResponse_String",
286
+ "ExecutePromptStreamingResponse",
287
+ "ExecutePromptStreamingResponse_Fulfilled",
288
+ "ExecutePromptStreamingResponse_Initiated",
289
+ "ExecutePromptStreamingResponse_Rejected",
290
+ "ExecutePromptStreamingResponse_Streaming",
271
291
  "ExecuteWorkflowStreamErrorResponse",
272
292
  "FinishReasonEnum",
273
293
  "ForbiddenError",
294
+ "FulfilledExecutePromptStreamingResponse",
274
295
  "GenerateErrorResponse",
275
296
  "GenerateOptionsRequest",
276
297
  "GenerateRequest",
@@ -282,6 +303,8 @@ __all__ = [
282
303
  "GenerateStreamResult",
283
304
  "GenerateStreamResultData",
284
305
  "IndexingStateEnum",
306
+ "InitiatedExecutePromptResponse",
307
+ "InitiatedExecutePromptStreamingResponse",
285
308
  "InternalServerError",
286
309
  "JsonExecutePromptResponse",
287
310
  "JsonInputRequest",
@@ -338,6 +361,8 @@ __all__ = [
338
361
  "RegisteredPromptModelVersion",
339
362
  "RegisteredPromptSandbox",
340
363
  "RegisteredPromptSandboxSnapshot",
364
+ "RejectedExecutePromptResponse",
365
+ "RejectedExecutePromptStreamingResponse",
341
366
  "SandboxMetricInputParams",
342
367
  "SandboxMetricInputParamsRequest",
343
368
  "SandboxScenario",
@@ -357,6 +382,7 @@ __all__ = [
357
382
  "SearchResultRequest",
358
383
  "SearchWeightsRequest",
359
384
  "SlimDocument",
385
+ "StreamingExecutePromptStreamingResponse",
360
386
  "StringExecutePromptResponse",
361
387
  "StringInputRequest",
362
388
  "SubmitCompletionActualRequest",
@@ -22,12 +22,13 @@ from .resources.model_versions.client import AsyncModelVersionsClient, ModelVers
22
22
  from .resources.registered_prompts.client import AsyncRegisteredPromptsClient, RegisteredPromptsClient
23
23
  from .resources.sandboxes.client import AsyncSandboxesClient, SandboxesClient
24
24
  from .resources.test_suites.client import AsyncTestSuitesClient, TestSuitesClient
25
+ from .types.execute_prompt_request import ExecutePromptRequest
25
26
  from .types.execute_prompt_response import ExecutePromptResponse
27
+ from .types.execute_prompt_streaming_response import ExecutePromptStreamingResponse
26
28
  from .types.generate_options_request import GenerateOptionsRequest
27
29
  from .types.generate_request import GenerateRequest
28
30
  from .types.generate_response import GenerateResponse
29
31
  from .types.generate_stream_response import GenerateStreamResponse
30
- from .types.prompt_deployment_input_request import PromptDeploymentInputRequest
31
32
  from .types.search_request_options_request import SearchRequestOptionsRequest
32
33
  from .types.search_response import SearchResponse
33
34
  from .types.submit_completion_actual_request import SubmitCompletionActualRequest
@@ -67,42 +68,17 @@ class Vellum:
67
68
  self.sandboxes = SandboxesClient(client_wrapper=self._client_wrapper)
68
69
  self.test_suites = TestSuitesClient(client_wrapper=self._client_wrapper)
69
70
 
70
- def execute_prompt(
71
- self,
72
- *,
73
- inputs: typing.List[PromptDeploymentInputRequest],
74
- prompt_deployment_id: typing.Optional[str] = OMIT,
75
- prompt_deployment_name: typing.Optional[str] = OMIT,
76
- release_tag: typing.Optional[str] = OMIT,
77
- external_id: typing.Optional[str] = OMIT,
78
- ) -> ExecutePromptResponse:
71
+ def execute_prompt(self, *, request: ExecutePromptRequest) -> ExecutePromptResponse:
79
72
  """
80
73
  Executes a deployed Prompt and returns the result.
81
74
 
82
75
  Parameters:
83
- - inputs: typing.List[PromptDeploymentInputRequest].
84
-
85
- - prompt_deployment_id: typing.Optional[str]. The ID of the Prompt Deployment. Must provide either this or prompt_deployment_name.
86
-
87
- - prompt_deployment_name: typing.Optional[str]. The name of the Prompt Deployment. Must provide either this or prompt_deployment_id.
88
-
89
- - release_tag: typing.Optional[str]. Optionally specify a release tag if you want to pin to a specific release of the Prompt Deployment
90
-
91
- - external_id: typing.Optional[str].
76
+ - request: ExecutePromptRequest.
92
77
  """
93
- _request: typing.Dict[str, typing.Any] = {"inputs": inputs}
94
- if prompt_deployment_id is not OMIT:
95
- _request["prompt_deployment_id"] = prompt_deployment_id
96
- if prompt_deployment_name is not OMIT:
97
- _request["prompt_deployment_name"] = prompt_deployment_name
98
- if release_tag is not OMIT:
99
- _request["release_tag"] = release_tag
100
- if external_id is not OMIT:
101
- _request["external_id"] = external_id
102
78
  _response = self._client_wrapper.httpx_client.request(
103
79
  "POST",
104
80
  urllib.parse.urljoin(f"{self._client_wrapper.get_environment().default}/", "v1/execute-prompt"),
105
- json=jsonable_encoder(_request),
81
+ json=jsonable_encoder(request),
106
82
  headers=self._client_wrapper.get_headers(),
107
83
  timeout=None,
108
84
  )
@@ -122,6 +98,36 @@ class Vellum:
122
98
  raise ApiError(status_code=_response.status_code, body=_response.text)
123
99
  raise ApiError(status_code=_response.status_code, body=_response_json)
124
100
 
101
+ def execute_prompt_stream(self, *, request: ExecutePromptRequest) -> ExecutePromptStreamingResponse:
102
+ """
103
+ Executes a deployed Prompt and streams back the results.
104
+
105
+ Parameters:
106
+ - request: ExecutePromptRequest.
107
+ """
108
+ _response = self._client_wrapper.httpx_client.request(
109
+ "POST",
110
+ urllib.parse.urljoin(f"{self._client_wrapper.get_environment().default}/", "v1/execute-prompt-stream"),
111
+ json=jsonable_encoder(request),
112
+ headers=self._client_wrapper.get_headers(),
113
+ timeout=None,
114
+ )
115
+ if 200 <= _response.status_code < 300:
116
+ return pydantic.parse_obj_as(ExecutePromptStreamingResponse, _response.json()) # type: ignore
117
+ if _response.status_code == 400:
118
+ raise BadRequestError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
119
+ if _response.status_code == 403:
120
+ raise ForbiddenError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
121
+ if _response.status_code == 404:
122
+ raise NotFoundError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
123
+ if _response.status_code == 500:
124
+ raise InternalServerError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
125
+ try:
126
+ _response_json = _response.json()
127
+ except JSONDecodeError:
128
+ raise ApiError(status_code=_response.status_code, body=_response.text)
129
+ raise ApiError(status_code=_response.status_code, body=_response_json)
130
+
125
131
  def execute_workflow_stream(
126
132
  self,
127
133
  *,
@@ -488,42 +494,17 @@ class AsyncVellum:
488
494
  self.sandboxes = AsyncSandboxesClient(client_wrapper=self._client_wrapper)
489
495
  self.test_suites = AsyncTestSuitesClient(client_wrapper=self._client_wrapper)
490
496
 
491
- async def execute_prompt(
492
- self,
493
- *,
494
- inputs: typing.List[PromptDeploymentInputRequest],
495
- prompt_deployment_id: typing.Optional[str] = OMIT,
496
- prompt_deployment_name: typing.Optional[str] = OMIT,
497
- release_tag: typing.Optional[str] = OMIT,
498
- external_id: typing.Optional[str] = OMIT,
499
- ) -> ExecutePromptResponse:
497
+ async def execute_prompt(self, *, request: ExecutePromptRequest) -> ExecutePromptResponse:
500
498
  """
501
499
  Executes a deployed Prompt and returns the result.
502
500
 
503
501
  Parameters:
504
- - inputs: typing.List[PromptDeploymentInputRequest].
505
-
506
- - prompt_deployment_id: typing.Optional[str]. The ID of the Prompt Deployment. Must provide either this or prompt_deployment_name.
507
-
508
- - prompt_deployment_name: typing.Optional[str]. The name of the Prompt Deployment. Must provide either this or prompt_deployment_id.
509
-
510
- - release_tag: typing.Optional[str]. Optionally specify a release tag if you want to pin to a specific release of the Prompt Deployment
511
-
512
- - external_id: typing.Optional[str].
502
+ - request: ExecutePromptRequest.
513
503
  """
514
- _request: typing.Dict[str, typing.Any] = {"inputs": inputs}
515
- if prompt_deployment_id is not OMIT:
516
- _request["prompt_deployment_id"] = prompt_deployment_id
517
- if prompt_deployment_name is not OMIT:
518
- _request["prompt_deployment_name"] = prompt_deployment_name
519
- if release_tag is not OMIT:
520
- _request["release_tag"] = release_tag
521
- if external_id is not OMIT:
522
- _request["external_id"] = external_id
523
504
  _response = await self._client_wrapper.httpx_client.request(
524
505
  "POST",
525
506
  urllib.parse.urljoin(f"{self._client_wrapper.get_environment().default}/", "v1/execute-prompt"),
526
- json=jsonable_encoder(_request),
507
+ json=jsonable_encoder(request),
527
508
  headers=self._client_wrapper.get_headers(),
528
509
  timeout=None,
529
510
  )
@@ -543,6 +524,36 @@ class AsyncVellum:
543
524
  raise ApiError(status_code=_response.status_code, body=_response.text)
544
525
  raise ApiError(status_code=_response.status_code, body=_response_json)
545
526
 
527
+ async def execute_prompt_stream(self, *, request: ExecutePromptRequest) -> ExecutePromptStreamingResponse:
528
+ """
529
+ Executes a deployed Prompt and streams back the results.
530
+
531
+ Parameters:
532
+ - request: ExecutePromptRequest.
533
+ """
534
+ _response = await self._client_wrapper.httpx_client.request(
535
+ "POST",
536
+ urllib.parse.urljoin(f"{self._client_wrapper.get_environment().default}/", "v1/execute-prompt-stream"),
537
+ json=jsonable_encoder(request),
538
+ headers=self._client_wrapper.get_headers(),
539
+ timeout=None,
540
+ )
541
+ if 200 <= _response.status_code < 300:
542
+ return pydantic.parse_obj_as(ExecutePromptStreamingResponse, _response.json()) # type: ignore
543
+ if _response.status_code == 400:
544
+ raise BadRequestError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
545
+ if _response.status_code == 403:
546
+ raise ForbiddenError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
547
+ if _response.status_code == 404:
548
+ raise NotFoundError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
549
+ if _response.status_code == 500:
550
+ raise InternalServerError(pydantic.parse_obj_as(typing.Any, _response.json())) # type: ignore
551
+ try:
552
+ _response_json = _response.json()
553
+ except JSONDecodeError:
554
+ raise ApiError(status_code=_response.status_code, body=_response.text)
555
+ raise ApiError(status_code=_response.status_code, body=_response_json)
556
+
546
557
  async def execute_workflow_stream(
547
558
  self,
548
559
  *,
@@ -16,7 +16,7 @@ class BaseClientWrapper:
16
16
  headers: typing.Dict[str, str] = {
17
17
  "X-Fern-Language": "Python",
18
18
  "X-Fern-SDK-Name": "vellum-ai",
19
- "X-Fern-SDK-Version": "v0.1.5",
19
+ "X-Fern-SDK-Version": "0.1.7",
20
20
  }
21
21
  headers["X_API_KEY"] = self.api_key
22
22
  return headers
@@ -0,0 +1,164 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import typing
4
+ import urllib.parse
5
+ from json.decoder import JSONDecodeError
6
+
7
+ from ...core.api_error import ApiError
8
+ from ...core.client_wrapper import AsyncClientWrapper, SyncClientWrapper
9
+ from ...core.jsonable_encoder import jsonable_encoder
10
+ from ...types.deployment_provider_payload_response import DeploymentProviderPayloadResponse
11
+ from ...types.deployment_read import DeploymentRead
12
+ from ...types.prompt_deployment_input_request import PromptDeploymentInputRequest
13
+
14
+ try:
15
+ import pydantic.v1 as pydantic # type: ignore
16
+ except ImportError:
17
+ import pydantic # type: ignore
18
+
19
+ # this is used as the default value for optional parameters
20
+ OMIT = typing.cast(typing.Any, ...)
21
+
22
+
23
+ class DeploymentsClient:
24
+ def __init__(self, *, client_wrapper: SyncClientWrapper):
25
+ self._client_wrapper = client_wrapper
26
+
27
+ def retrieve(self, id: str) -> DeploymentRead:
28
+ """
29
+ Used to retrieve a deployment given its ID or name.
30
+
31
+ Parameters:
32
+ - id: str. Either the Deployment's ID or its unique name
33
+ """
34
+ _response = self._client_wrapper.httpx_client.request(
35
+ "GET",
36
+ urllib.parse.urljoin(f"{self._client_wrapper.get_environment().default}/", f"v1/deployments/{id}"),
37
+ headers=self._client_wrapper.get_headers(),
38
+ timeout=None,
39
+ )
40
+ if 200 <= _response.status_code < 300:
41
+ return pydantic.parse_obj_as(DeploymentRead, _response.json()) # type: ignore
42
+ try:
43
+ _response_json = _response.json()
44
+ except JSONDecodeError:
45
+ raise ApiError(status_code=_response.status_code, body=_response.text)
46
+ raise ApiError(status_code=_response.status_code, body=_response_json)
47
+
48
+ def retrieve_provider_payload(
49
+ self,
50
+ *,
51
+ deployment_id: typing.Optional[str] = OMIT,
52
+ deployment_name: typing.Optional[str] = OMIT,
53
+ inputs: typing.List[PromptDeploymentInputRequest],
54
+ ) -> DeploymentProviderPayloadResponse:
55
+ """
56
+ Parameters:
57
+ - deployment_id: typing.Optional[str]. The ID of the deployment. Must provide either this or deployment_name.
58
+
59
+ - deployment_name: typing.Optional[str]. The name of the deployment. Must provide either this or deployment_id.
60
+
61
+ - inputs: typing.List[PromptDeploymentInputRequest].
62
+ ---
63
+ from vellum.client import Vellum
64
+
65
+ client = Vellum(
66
+ api_key="YOUR_API_KEY",
67
+ )
68
+ client.deployments.retrieve_provider_payload(
69
+ inputs=[],
70
+ )
71
+ """
72
+ _request: typing.Dict[str, typing.Any] = {"inputs": inputs}
73
+ if deployment_id is not OMIT:
74
+ _request["deployment_id"] = deployment_id
75
+ if deployment_name is not OMIT:
76
+ _request["deployment_name"] = deployment_name
77
+ _response = self._client_wrapper.httpx_client.request(
78
+ "POST",
79
+ urllib.parse.urljoin(
80
+ f"{self._client_wrapper.get_environment().default}/", "v1/deployments/provider-payload"
81
+ ),
82
+ json=jsonable_encoder(_request),
83
+ headers=self._client_wrapper.get_headers(),
84
+ timeout=None,
85
+ )
86
+ if 200 <= _response.status_code < 300:
87
+ return pydantic.parse_obj_as(DeploymentProviderPayloadResponse, _response.json()) # type: ignore
88
+ try:
89
+ _response_json = _response.json()
90
+ except JSONDecodeError:
91
+ raise ApiError(status_code=_response.status_code, body=_response.text)
92
+ raise ApiError(status_code=_response.status_code, body=_response_json)
93
+
94
+
95
+ class AsyncDeploymentsClient:
96
+ def __init__(self, *, client_wrapper: AsyncClientWrapper):
97
+ self._client_wrapper = client_wrapper
98
+
99
+ async def retrieve(self, id: str) -> DeploymentRead:
100
+ """
101
+ Used to retrieve a deployment given its ID or name.
102
+
103
+ Parameters:
104
+ - id: str. Either the Deployment's ID or its unique name
105
+ """
106
+ _response = await self._client_wrapper.httpx_client.request(
107
+ "GET",
108
+ urllib.parse.urljoin(f"{self._client_wrapper.get_environment().default}/", f"v1/deployments/{id}"),
109
+ headers=self._client_wrapper.get_headers(),
110
+ timeout=None,
111
+ )
112
+ if 200 <= _response.status_code < 300:
113
+ return pydantic.parse_obj_as(DeploymentRead, _response.json()) # type: ignore
114
+ try:
115
+ _response_json = _response.json()
116
+ except JSONDecodeError:
117
+ raise ApiError(status_code=_response.status_code, body=_response.text)
118
+ raise ApiError(status_code=_response.status_code, body=_response_json)
119
+
120
+ async def retrieve_provider_payload(
121
+ self,
122
+ *,
123
+ deployment_id: typing.Optional[str] = OMIT,
124
+ deployment_name: typing.Optional[str] = OMIT,
125
+ inputs: typing.List[PromptDeploymentInputRequest],
126
+ ) -> DeploymentProviderPayloadResponse:
127
+ """
128
+ Parameters:
129
+ - deployment_id: typing.Optional[str]. The ID of the deployment. Must provide either this or deployment_name.
130
+
131
+ - deployment_name: typing.Optional[str]. The name of the deployment. Must provide either this or deployment_id.
132
+
133
+ - inputs: typing.List[PromptDeploymentInputRequest].
134
+ ---
135
+ from vellum.client import AsyncVellum
136
+
137
+ client = AsyncVellum(
138
+ api_key="YOUR_API_KEY",
139
+ )
140
+ await client.deployments.retrieve_provider_payload(
141
+ inputs=[],
142
+ )
143
+ """
144
+ _request: typing.Dict[str, typing.Any] = {"inputs": inputs}
145
+ if deployment_id is not OMIT:
146
+ _request["deployment_id"] = deployment_id
147
+ if deployment_name is not OMIT:
148
+ _request["deployment_name"] = deployment_name
149
+ _response = await self._client_wrapper.httpx_client.request(
150
+ "POST",
151
+ urllib.parse.urljoin(
152
+ f"{self._client_wrapper.get_environment().default}/", "v1/deployments/provider-payload"
153
+ ),
154
+ json=jsonable_encoder(_request),
155
+ headers=self._client_wrapper.get_headers(),
156
+ timeout=None,
157
+ )
158
+ if 200 <= _response.status_code < 300:
159
+ return pydantic.parse_obj_as(DeploymentProviderPayloadResponse, _response.json()) # type: ignore
160
+ try:
161
+ _response_json = _response.json()
162
+ except JSONDecodeError:
163
+ raise ApiError(status_code=_response.status_code, body=_response.text)
164
+ raise ApiError(status_code=_response.status_code, body=_response_json)
@@ -19,7 +19,7 @@ class ModelVersionsClient:
19
19
 
20
20
  def retrieve(self, id: str) -> ModelVersionRead:
21
21
  """
22
- Used to retrieve a model version given its ID.
22
+ Deprecated. Use the `deployments/provider-payload` endpoint to fetch information that we send to Model providers.
23
23
 
24
24
  Parameters:
25
25
  - id: str. A UUID string identifying this model version.
@@ -45,7 +45,7 @@ class AsyncModelVersionsClient:
45
45
 
46
46
  async def retrieve(self, id: str) -> ModelVersionRead:
47
47
  """
48
- Used to retrieve a model version given its ID.
48
+ Deprecated. Use the `deployments/provider-payload` endpoint to fetch information that we send to Model providers.
49
49
 
50
50
  Parameters:
51
51
  - id: str. A UUID string identifying this model version.
@@ -9,6 +9,7 @@ from .chat_message_request import ChatMessageRequest
9
9
  from .chat_message_role import ChatMessageRole
10
10
  from .conditional_node_result import ConditionalNodeResult
11
11
  from .conditional_node_result_data import ConditionalNodeResultData
12
+ from .deployment_provider_payload_response import DeploymentProviderPayloadResponse
12
13
  from .deployment_read import DeploymentRead
13
14
  from .deployment_status import DeploymentStatus
14
15
  from .document_document_to_document_index import DocumentDocumentToDocumentIndex
@@ -22,14 +23,23 @@ from .error_execute_prompt_response import ErrorExecutePromptResponse
22
23
  from .evaluation_params import EvaluationParams
23
24
  from .evaluation_params_request import EvaluationParamsRequest
24
25
  from .execute_prompt_api_error_response import ExecutePromptApiErrorResponse
26
+ from .execute_prompt_request import ExecutePromptRequest
25
27
  from .execute_prompt_response import (
26
28
  ExecutePromptResponse,
27
29
  ExecutePromptResponse_Error,
28
30
  ExecutePromptResponse_Json,
29
31
  ExecutePromptResponse_String,
30
32
  )
33
+ from .execute_prompt_streaming_response import (
34
+ ExecutePromptStreamingResponse,
35
+ ExecutePromptStreamingResponse_Fulfilled,
36
+ ExecutePromptStreamingResponse_Initiated,
37
+ ExecutePromptStreamingResponse_Rejected,
38
+ ExecutePromptStreamingResponse_Streaming,
39
+ )
31
40
  from .execute_workflow_stream_error_response import ExecuteWorkflowStreamErrorResponse
32
41
  from .finish_reason_enum import FinishReasonEnum
42
+ from .fulfilled_execute_prompt_streaming_response import FulfilledExecutePromptStreamingResponse
33
43
  from .generate_error_response import GenerateErrorResponse
34
44
  from .generate_options_request import GenerateOptionsRequest
35
45
  from .generate_request import GenerateRequest
@@ -41,6 +51,8 @@ from .generate_stream_response import GenerateStreamResponse
41
51
  from .generate_stream_result import GenerateStreamResult
42
52
  from .generate_stream_result_data import GenerateStreamResultData
43
53
  from .indexing_state_enum import IndexingStateEnum
54
+ from .initiated_execute_prompt_response import InitiatedExecutePromptResponse
55
+ from .initiated_execute_prompt_streaming_response import InitiatedExecutePromptStreamingResponse
44
56
  from .json_execute_prompt_response import JsonExecutePromptResponse
45
57
  from .json_input_request import JsonInputRequest
46
58
  from .logical_operator import LogicalOperator
@@ -99,6 +111,8 @@ from .registered_prompt_input_variable_request import RegisteredPromptInputVaria
99
111
  from .registered_prompt_model_version import RegisteredPromptModelVersion
100
112
  from .registered_prompt_sandbox import RegisteredPromptSandbox
101
113
  from .registered_prompt_sandbox_snapshot import RegisteredPromptSandboxSnapshot
114
+ from .rejected_execute_prompt_response import RejectedExecutePromptResponse
115
+ from .rejected_execute_prompt_streaming_response import RejectedExecutePromptStreamingResponse
102
116
  from .sandbox_metric_input_params import SandboxMetricInputParams
103
117
  from .sandbox_metric_input_params_request import SandboxMetricInputParamsRequest
104
118
  from .sandbox_scenario import SandboxScenario
@@ -118,6 +132,7 @@ from .search_result_merging_request import SearchResultMergingRequest
118
132
  from .search_result_request import SearchResultRequest
119
133
  from .search_weights_request import SearchWeightsRequest
120
134
  from .slim_document import SlimDocument
135
+ from .streaming_execute_prompt_streaming_response import StreamingExecutePromptStreamingResponse
121
136
  from .string_execute_prompt_response import StringExecutePromptResponse
122
137
  from .string_input_request import StringInputRequest
123
138
  from .submit_completion_actual_request import SubmitCompletionActualRequest
@@ -256,6 +271,7 @@ __all__ = [
256
271
  "ChatMessageRole",
257
272
  "ConditionalNodeResult",
258
273
  "ConditionalNodeResultData",
274
+ "DeploymentProviderPayloadResponse",
259
275
  "DeploymentRead",
260
276
  "DeploymentStatus",
261
277
  "DocumentDocumentToDocumentIndex",
@@ -269,12 +285,19 @@ __all__ = [
269
285
  "EvaluationParams",
270
286
  "EvaluationParamsRequest",
271
287
  "ExecutePromptApiErrorResponse",
288
+ "ExecutePromptRequest",
272
289
  "ExecutePromptResponse",
273
290
  "ExecutePromptResponse_Error",
274
291
  "ExecutePromptResponse_Json",
275
292
  "ExecutePromptResponse_String",
293
+ "ExecutePromptStreamingResponse",
294
+ "ExecutePromptStreamingResponse_Fulfilled",
295
+ "ExecutePromptStreamingResponse_Initiated",
296
+ "ExecutePromptStreamingResponse_Rejected",
297
+ "ExecutePromptStreamingResponse_Streaming",
276
298
  "ExecuteWorkflowStreamErrorResponse",
277
299
  "FinishReasonEnum",
300
+ "FulfilledExecutePromptStreamingResponse",
278
301
  "GenerateErrorResponse",
279
302
  "GenerateOptionsRequest",
280
303
  "GenerateRequest",
@@ -286,6 +309,8 @@ __all__ = [
286
309
  "GenerateStreamResult",
287
310
  "GenerateStreamResultData",
288
311
  "IndexingStateEnum",
312
+ "InitiatedExecutePromptResponse",
313
+ "InitiatedExecutePromptStreamingResponse",
289
314
  "JsonExecutePromptResponse",
290
315
  "JsonInputRequest",
291
316
  "LogicalOperator",
@@ -340,6 +365,8 @@ __all__ = [
340
365
  "RegisteredPromptModelVersion",
341
366
  "RegisteredPromptSandbox",
342
367
  "RegisteredPromptSandboxSnapshot",
368
+ "RejectedExecutePromptResponse",
369
+ "RejectedExecutePromptStreamingResponse",
343
370
  "SandboxMetricInputParams",
344
371
  "SandboxMetricInputParamsRequest",
345
372
  "SandboxScenario",
@@ -359,6 +386,7 @@ __all__ = [
359
386
  "SearchResultRequest",
360
387
  "SearchWeightsRequest",
361
388
  "SlimDocument",
389
+ "StreamingExecutePromptStreamingResponse",
362
390
  "StringExecutePromptResponse",
363
391
  "StringInputRequest",
364
392
  "SubmitCompletionActualRequest",
@@ -0,0 +1,28 @@
1
+ # This file was auto-generated by Fern from our API Definition.
2
+
3
+ import datetime as dt
4
+ import typing
5
+
6
+ from ..core.datetime_utils import serialize_datetime
7
+
8
+ try:
9
+ import pydantic.v1 as pydantic # type: ignore
10
+ except ImportError:
11
+ import pydantic # type: ignore
12
+
13
+
14
+ class DeploymentProviderPayloadResponse(pydantic.BaseModel):
15
+ payload: typing.Dict[str, typing.Any]
16
+
17
+ def json(self, **kwargs: typing.Any) -> str:
18
+ kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
19
+ return super().json(**kwargs_with_defaults)
20
+
21
+ def dict(self, **kwargs: typing.Any) -> typing.Dict[str, typing.Any]:
22
+ kwargs_with_defaults: typing.Any = {"by_alias": True, "exclude_unset": True, **kwargs}
23
+ return super().dict(**kwargs_with_defaults)
24
+
25
+ class Config:
26
+ frozen = True
27
+ smart_union = True
28
+ json_encoders = {dt.datetime: serialize_datetime}
@@ -37,7 +37,9 @@ class DeploymentRead(pydantic.BaseModel):
37
37
  "* `PRODUCTION` - Production\n"
38
38
  )
39
39
  )
40
- active_model_version_ids: typing.List[str]
40
+ active_model_version_ids: typing.List[str] = pydantic.Field(
41
+ description="Deprecated. The Prompt execution endpoints return a `prompt_version_id` that could be used instead."
42
+ )
41
43
  last_deployed_on: dt.datetime
42
44
  input_variables: typing.List[VellumVariable]
43
45