eval-studio-client 1.2.5__py3-none-any.whl → 1.3.0__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (306) hide show
  1. eval_studio_client/api/__init__.py +65 -0
  2. eval_studio_client/api/api/__init__.py +3 -0
  3. eval_studio_client/api/api/dashboard_report_service_api.py +292 -0
  4. eval_studio_client/api/api/dashboard_service_api.py +16 -16
  5. eval_studio_client/api/api/dashboard_test_case_annotation_service_api.py +611 -0
  6. eval_studio_client/api/api/document_service_api.py +16 -16
  7. eval_studio_client/api/api/evaluation_service_api.py +12 -12
  8. eval_studio_client/api/api/evaluator_service_api.py +16 -16
  9. eval_studio_client/api/api/leaderboard_report_service_api.py +304 -17
  10. eval_studio_client/api/api/leaderboard_service_api.py +554 -16
  11. eval_studio_client/api/api/leaderboard_test_case_annotation_service_api.py +611 -0
  12. eval_studio_client/api/api/model_service_api.py +16 -16
  13. eval_studio_client/api/api/operation_service_api.py +821 -17
  14. eval_studio_client/api/api/perturbator_service_api.py +22 -22
  15. eval_studio_client/api/api/test_case_service_api.py +300 -16
  16. eval_studio_client/api/api/test_class_service_api.py +16 -16
  17. eval_studio_client/api/api/test_service_api.py +285 -16
  18. eval_studio_client/api/api/workflow_node_service_api.py +16 -16
  19. eval_studio_client/api/api/workflow_service_api.py +16 -16
  20. eval_studio_client/api/docs/AdversarialInputsServiceTestAdversarialInputsRobustnessRequest.md +2 -1
  21. eval_studio_client/api/docs/DashboardReportServiceApi.md +75 -0
  22. eval_studio_client/api/docs/DashboardServiceApi.md +5 -5
  23. eval_studio_client/api/docs/DashboardTestCaseAnnotationServiceApi.md +149 -0
  24. eval_studio_client/api/docs/DocumentServiceApi.md +5 -5
  25. eval_studio_client/api/docs/EvaluationServiceApi.md +4 -4
  26. eval_studio_client/api/docs/EvaluatorServiceApi.md +5 -5
  27. eval_studio_client/api/docs/LeaderboardReportServiceApi.md +75 -5
  28. eval_studio_client/api/docs/LeaderboardServiceApi.md +141 -5
  29. eval_studio_client/api/docs/LeaderboardTestCaseAnnotationServiceApi.md +149 -0
  30. eval_studio_client/api/docs/ModelServiceApi.md +5 -5
  31. eval_studio_client/api/docs/OperationServiceApi.md +215 -8
  32. eval_studio_client/api/docs/PerturbatorServiceApi.md +7 -7
  33. eval_studio_client/api/docs/RequiredTheDashboardTestCaseAnnotationToUpdate.md +35 -0
  34. eval_studio_client/api/docs/RequiredTheLeaderboardTestCaseAnnotationToUpdate.md +35 -0
  35. eval_studio_client/api/docs/RequiredTheLeaderboardToUpdate.md +1 -0
  36. eval_studio_client/api/docs/RequiredTheOperationToFinalize.md +1 -0
  37. eval_studio_client/api/docs/RequiredTheOperationToUpdate.md +1 -0
  38. eval_studio_client/api/docs/TestCaseServiceApi.md +75 -5
  39. eval_studio_client/api/docs/TestCaseServiceAppendTestCasesRequest.md +30 -0
  40. eval_studio_client/api/docs/TestClassServiceApi.md +5 -5
  41. eval_studio_client/api/docs/TestServiceApi.md +73 -5
  42. eval_studio_client/api/docs/V1ActualOutputMeta.md +30 -0
  43. eval_studio_client/api/docs/V1ActualOutputMetaDiff.md +36 -0
  44. eval_studio_client/api/docs/V1AgentChatActivityDiagram.md +31 -0
  45. eval_studio_client/api/docs/V1AgentChatActivityDiagramEdge.md +32 -0
  46. eval_studio_client/api/docs/V1AgentChatActivityDiagramNode.md +32 -0
  47. eval_studio_client/api/docs/V1AgentChatActivityDiagramRow.md +30 -0
  48. eval_studio_client/api/docs/V1AgentChatScriptUsage.md +33 -0
  49. eval_studio_client/api/docs/V1AgentChatScriptsBarChart.md +30 -0
  50. eval_studio_client/api/docs/V1AgentChatToolUsage.md +33 -0
  51. eval_studio_client/api/docs/V1AgentChatToolsBarChart.md +30 -0
  52. eval_studio_client/api/docs/V1AllMetricScores.md +29 -0
  53. eval_studio_client/api/docs/V1AppendTestCasesResponse.md +29 -0
  54. eval_studio_client/api/docs/V1BatchCreateLeaderboardsWithoutCacheRequest.md +31 -0
  55. eval_studio_client/api/docs/V1BatchCreateLeaderboardsWithoutCacheResponse.md +29 -0
  56. eval_studio_client/api/docs/V1BatchMarkOperationSeenByCreatorResponse.md +29 -0
  57. eval_studio_client/api/docs/V1CmpLeaderboardReportsRequest.md +33 -0
  58. eval_studio_client/api/docs/V1CmpLeaderboardReportsResponse.md +29 -0
  59. eval_studio_client/api/docs/V1ComparisonItem.md +36 -0
  60. eval_studio_client/api/docs/V1ComparisonMetricScore.md +30 -0
  61. eval_studio_client/api/docs/V1ComparisonResult.md +31 -0
  62. eval_studio_client/api/docs/V1ComparisonSummary.md +31 -0
  63. eval_studio_client/api/docs/V1CreateEvaluationRequest.md +1 -0
  64. eval_studio_client/api/docs/V1CreateTestFromTestCasesRequest.md +32 -0
  65. eval_studio_client/api/docs/V1CreateTestFromTestCasesResponse.md +29 -0
  66. eval_studio_client/api/docs/V1DashboardReport.md +31 -0
  67. eval_studio_client/api/docs/V1DashboardReportResult.md +39 -0
  68. eval_studio_client/api/docs/V1DashboardTestCaseAnnotation.md +36 -0
  69. eval_studio_client/api/docs/V1DataFragment.md +31 -0
  70. eval_studio_client/api/docs/V1DeepCompareLeaderboardsRequest.md +33 -0
  71. eval_studio_client/api/docs/V1DeepCompareLeaderboardsResponse.md +29 -0
  72. eval_studio_client/api/docs/V1DiffItem.md +36 -0
  73. eval_studio_client/api/docs/V1EvaluationType.md +12 -0
  74. eval_studio_client/api/docs/V1FlippedMetric.md +31 -0
  75. eval_studio_client/api/docs/V1GetDashboardReportResponse.md +29 -0
  76. eval_studio_client/api/docs/V1HumanDecision.md +12 -0
  77. eval_studio_client/api/docs/V1Info.md +1 -0
  78. eval_studio_client/api/docs/V1Leaderboard.md +1 -0
  79. eval_studio_client/api/docs/V1LeaderboardCmpReport.md +30 -0
  80. eval_studio_client/api/docs/V1LeaderboardComparisonItem.md +31 -0
  81. eval_studio_client/api/docs/V1LeaderboardInfo.md +30 -0
  82. eval_studio_client/api/docs/V1LeaderboardReportActualOutputMeta.md +6 -3
  83. eval_studio_client/api/docs/V1LeaderboardReportResult.md +11 -8
  84. eval_studio_client/api/docs/V1LeaderboardReportResultView.md +12 -0
  85. eval_studio_client/api/docs/V1LeaderboardTestCaseAnnotation.md +36 -0
  86. eval_studio_client/api/docs/V1ListDashboardTestCaseAnnotationsResponse.md +29 -0
  87. eval_studio_client/api/docs/V1ListLeaderboardTestCaseAnnotationsResponse.md +29 -0
  88. eval_studio_client/api/docs/V1ListOperationsResponse.md +1 -0
  89. eval_studio_client/api/docs/V1ListUnseenOperationsResponse.md +30 -0
  90. eval_studio_client/api/docs/V1MarkOperationSeenByCreatorResponse.md +29 -0
  91. eval_studio_client/api/docs/V1Metric.md +30 -0
  92. eval_studio_client/api/docs/V1MetricAverage.md +36 -0
  93. eval_studio_client/api/docs/V1MetricMeta.md +40 -0
  94. eval_studio_client/api/docs/V1MetricScore.md +1 -1
  95. eval_studio_client/api/docs/V1MetricScores.md +1 -1
  96. eval_studio_client/api/docs/V1ModelType.md +1 -1
  97. eval_studio_client/api/docs/V1ModelsComparisons.md +32 -0
  98. eval_studio_client/api/docs/V1ModelsComparisonsMetrics.md +33 -0
  99. eval_studio_client/api/docs/V1ModelsOverview.md +34 -0
  100. eval_studio_client/api/docs/V1Operation.md +1 -0
  101. eval_studio_client/api/docs/V1OperationView.md +12 -0
  102. eval_studio_client/api/docs/V1RetrievedContextDiff.md +36 -0
  103. eval_studio_client/api/docs/V1Stats.md +2 -0
  104. eval_studio_client/api/docs/V1TechnicalMetrics.md +30 -0
  105. eval_studio_client/api/docs/V1TechnicalMetricsDetail.md +33 -0
  106. eval_studio_client/api/docs/V1TestCaseLeaderboardItem.md +31 -0
  107. eval_studio_client/api/docs/V1TestCaseRelationshipInfo.md +31 -0
  108. eval_studio_client/api/docs/V1TestCaseResult.md +48 -0
  109. eval_studio_client/api/docs/V1TextSimilarityMetric.md +12 -0
  110. eval_studio_client/api/docs/V1UpdateDashboardTestCaseAnnotationResponse.md +29 -0
  111. eval_studio_client/api/docs/V1UpdateLeaderboardTestCaseAnnotationResponse.md +29 -0
  112. eval_studio_client/api/docs/WorkflowNodeServiceApi.md +5 -5
  113. eval_studio_client/api/docs/WorkflowServiceApi.md +5 -5
  114. eval_studio_client/api/models/__init__.py +62 -0
  115. eval_studio_client/api/models/adversarial_inputs_service_test_adversarial_inputs_robustness_request.py +17 -2
  116. eval_studio_client/api/models/required_the_dashboard_test_case_annotation_to_update.py +108 -0
  117. eval_studio_client/api/models/required_the_leaderboard_test_case_annotation_to_update.py +108 -0
  118. eval_studio_client/api/models/required_the_leaderboard_to_update.py +5 -2
  119. eval_studio_client/api/models/required_the_operation_to_finalize.py +6 -2
  120. eval_studio_client/api/models/required_the_operation_to_update.py +6 -2
  121. eval_studio_client/api/models/test_case_service_append_test_cases_request.py +89 -0
  122. eval_studio_client/api/models/v1_actual_output_meta.py +97 -0
  123. eval_studio_client/api/models/v1_actual_output_meta_diff.py +101 -0
  124. eval_studio_client/api/models/v1_agent_chat_activity_diagram.py +109 -0
  125. eval_studio_client/api/models/v1_agent_chat_activity_diagram_edge.py +97 -0
  126. eval_studio_client/api/models/v1_agent_chat_activity_diagram_node.py +97 -0
  127. eval_studio_client/api/models/v1_agent_chat_activity_diagram_row.py +97 -0
  128. eval_studio_client/api/models/v1_agent_chat_script_usage.py +101 -0
  129. eval_studio_client/api/models/v1_agent_chat_scripts_bar_chart.py +102 -0
  130. eval_studio_client/api/models/v1_agent_chat_tool_usage.py +101 -0
  131. eval_studio_client/api/models/v1_agent_chat_tools_bar_chart.py +102 -0
  132. eval_studio_client/api/models/v1_all_metric_scores.py +87 -0
  133. eval_studio_client/api/models/v1_append_test_cases_response.py +95 -0
  134. eval_studio_client/api/models/v1_batch_create_leaderboards_without_cache_request.py +99 -0
  135. eval_studio_client/api/models/v1_batch_create_leaderboards_without_cache_response.py +91 -0
  136. eval_studio_client/api/models/v1_batch_mark_operation_seen_by_creator_response.py +95 -0
  137. eval_studio_client/api/models/v1_cmp_leaderboard_reports_request.py +96 -0
  138. eval_studio_client/api/models/v1_cmp_leaderboard_reports_response.py +91 -0
  139. eval_studio_client/api/models/v1_comparison_item.py +130 -0
  140. eval_studio_client/api/models/v1_comparison_metric_score.py +89 -0
  141. eval_studio_client/api/models/v1_comparison_result.py +120 -0
  142. eval_studio_client/api/models/v1_comparison_summary.py +91 -0
  143. eval_studio_client/api/models/v1_create_evaluation_request.py +5 -2
  144. eval_studio_client/api/models/v1_create_test_from_test_cases_request.py +93 -0
  145. eval_studio_client/api/models/v1_create_test_from_test_cases_response.py +91 -0
  146. eval_studio_client/api/models/v1_dashboard_report.py +109 -0
  147. eval_studio_client/api/models/v1_dashboard_report_result.py +139 -0
  148. eval_studio_client/api/models/v1_dashboard_test_case_annotation.py +112 -0
  149. eval_studio_client/api/models/v1_data_fragment.py +91 -0
  150. eval_studio_client/api/models/v1_deep_compare_leaderboards_request.py +96 -0
  151. eval_studio_client/api/models/v1_deep_compare_leaderboards_response.py +91 -0
  152. eval_studio_client/api/models/v1_diff_item.py +137 -0
  153. eval_studio_client/api/models/v1_evaluation_type.py +39 -0
  154. eval_studio_client/api/models/v1_flipped_metric.py +91 -0
  155. eval_studio_client/api/models/v1_get_dashboard_report_response.py +91 -0
  156. eval_studio_client/api/models/v1_human_decision.py +38 -0
  157. eval_studio_client/api/models/v1_info.py +4 -2
  158. eval_studio_client/api/models/v1_leaderboard.py +5 -2
  159. eval_studio_client/api/models/v1_leaderboard_cmp_report.py +93 -0
  160. eval_studio_client/api/models/v1_leaderboard_comparison_item.py +91 -0
  161. eval_studio_client/api/models/v1_leaderboard_info.py +97 -0
  162. eval_studio_client/api/models/v1_leaderboard_report_actual_output_meta.py +23 -9
  163. eval_studio_client/api/models/v1_leaderboard_report_result.py +21 -10
  164. eval_studio_client/api/models/v1_leaderboard_report_result_view.py +38 -0
  165. eval_studio_client/api/models/v1_leaderboard_test_case_annotation.py +112 -0
  166. eval_studio_client/api/models/v1_list_dashboard_test_case_annotations_response.py +95 -0
  167. eval_studio_client/api/models/v1_list_leaderboard_test_case_annotations_response.py +95 -0
  168. eval_studio_client/api/models/v1_list_operations_response.py +5 -3
  169. eval_studio_client/api/models/v1_list_unseen_operations_response.py +97 -0
  170. eval_studio_client/api/models/v1_mark_operation_seen_by_creator_response.py +91 -0
  171. eval_studio_client/api/models/v1_metric.py +89 -0
  172. eval_studio_client/api/models/v1_metric_average.py +101 -0
  173. eval_studio_client/api/models/v1_metric_meta.py +109 -0
  174. eval_studio_client/api/models/v1_metric_score.py +6 -1
  175. eval_studio_client/api/models/v1_metric_scores.py +1 -1
  176. eval_studio_client/api/models/v1_model_type.py +2 -1
  177. eval_studio_client/api/models/v1_models_comparisons.py +93 -0
  178. eval_studio_client/api/models/v1_models_comparisons_metrics.py +103 -0
  179. eval_studio_client/api/models/v1_models_overview.py +97 -0
  180. eval_studio_client/api/models/v1_operation.py +6 -2
  181. eval_studio_client/api/models/v1_operation_view.py +38 -0
  182. eval_studio_client/api/models/v1_retrieved_context_diff.py +101 -0
  183. eval_studio_client/api/models/v1_stats.py +16 -2
  184. eval_studio_client/api/models/v1_technical_metrics.py +96 -0
  185. eval_studio_client/api/models/v1_technical_metrics_detail.py +95 -0
  186. eval_studio_client/api/models/v1_test_case_leaderboard_item.py +91 -0
  187. eval_studio_client/api/models/v1_test_case_relationship_info.py +91 -0
  188. eval_studio_client/api/models/v1_test_case_result.py +157 -0
  189. eval_studio_client/api/models/v1_text_similarity_metric.py +39 -0
  190. eval_studio_client/api/models/v1_update_dashboard_test_case_annotation_response.py +91 -0
  191. eval_studio_client/api/models/v1_update_leaderboard_test_case_annotation_response.py +91 -0
  192. eval_studio_client/api/models/v1_workflow_node_type.py +1 -0
  193. eval_studio_client/api/models/v1_workflow_type.py +1 -0
  194. eval_studio_client/api/test/test_adversarial_inputs_service_test_adversarial_inputs_robustness_request.py +6 -0
  195. eval_studio_client/api/test/test_dashboard_report_service_api.py +37 -0
  196. eval_studio_client/api/test/test_dashboard_test_case_annotation_service_api.py +43 -0
  197. eval_studio_client/api/test/test_leaderboard_report_service_api.py +6 -0
  198. eval_studio_client/api/test/test_leaderboard_service_api.py +12 -0
  199. eval_studio_client/api/test/test_leaderboard_test_case_annotation_service_api.py +43 -0
  200. eval_studio_client/api/test/test_operation_service_api.py +18 -0
  201. eval_studio_client/api/test/test_required_the_dashboard_test_case_annotation_to_update.py +57 -0
  202. eval_studio_client/api/test/test_required_the_leaderboard_test_case_annotation_to_update.py +57 -0
  203. eval_studio_client/api/test/test_required_the_leaderboard_to_update.py +2 -1
  204. eval_studio_client/api/test/test_required_the_operation_to_finalize.py +2 -1
  205. eval_studio_client/api/test/test_required_the_operation_to_update.py +2 -1
  206. eval_studio_client/api/test/test_test_case_service_api.py +6 -0
  207. eval_studio_client/api/test/test_test_case_service_append_test_cases_request.py +52 -0
  208. eval_studio_client/api/test/test_test_service_api.py +6 -0
  209. eval_studio_client/api/test/test_v1_abort_operation_response.py +2 -1
  210. eval_studio_client/api/test/test_v1_actual_output_meta.py +61 -0
  211. eval_studio_client/api/test/test_v1_actual_output_meta_diff.py +66 -0
  212. eval_studio_client/api/test/test_v1_agent_chat_activity_diagram.py +65 -0
  213. eval_studio_client/api/test/test_v1_agent_chat_activity_diagram_edge.py +53 -0
  214. eval_studio_client/api/test/test_v1_agent_chat_activity_diagram_node.py +53 -0
  215. eval_studio_client/api/test/test_v1_agent_chat_activity_diagram_row.py +56 -0
  216. eval_studio_client/api/test/test_v1_agent_chat_script_usage.py +54 -0
  217. eval_studio_client/api/test/test_v1_agent_chat_scripts_bar_chart.py +57 -0
  218. eval_studio_client/api/test/test_v1_agent_chat_tool_usage.py +54 -0
  219. eval_studio_client/api/test/test_v1_agent_chat_tools_bar_chart.py +57 -0
  220. eval_studio_client/api/test/test_v1_all_metric_scores.py +53 -0
  221. eval_studio_client/api/test/test_v1_append_test_cases_response.py +74 -0
  222. eval_studio_client/api/test/test_v1_batch_create_leaderboards_request.py +2 -1
  223. eval_studio_client/api/test/test_v1_batch_create_leaderboards_response.py +2 -1
  224. eval_studio_client/api/test/test_v1_batch_create_leaderboards_without_cache_request.py +120 -0
  225. eval_studio_client/api/test/test_v1_batch_create_leaderboards_without_cache_response.py +72 -0
  226. eval_studio_client/api/test/test_v1_batch_delete_leaderboards_response.py +2 -1
  227. eval_studio_client/api/test/test_v1_batch_get_leaderboards_response.py +2 -1
  228. eval_studio_client/api/test/test_v1_batch_get_operations_response.py +2 -1
  229. eval_studio_client/api/test/test_v1_batch_import_leaderboard_response.py +2 -1
  230. eval_studio_client/api/test/test_v1_batch_mark_operation_seen_by_creator_response.py +74 -0
  231. eval_studio_client/api/test/test_v1_cmp_leaderboard_reports_request.py +55 -0
  232. eval_studio_client/api/test/test_v1_cmp_leaderboard_reports_response.py +255 -0
  233. eval_studio_client/api/test/test_v1_comparison_item.py +233 -0
  234. eval_studio_client/api/test/test_v1_comparison_metric_score.py +52 -0
  235. eval_studio_client/api/test/test_v1_comparison_result.py +258 -0
  236. eval_studio_client/api/test/test_v1_comparison_summary.py +53 -0
  237. eval_studio_client/api/test/test_v1_create_evaluation_request.py +2 -1
  238. eval_studio_client/api/test/test_v1_create_leaderboard_request.py +2 -1
  239. eval_studio_client/api/test/test_v1_create_leaderboard_response.py +2 -1
  240. eval_studio_client/api/test/test_v1_create_leaderboard_without_cache_response.py +2 -1
  241. eval_studio_client/api/test/test_v1_create_test_from_test_cases_request.py +54 -0
  242. eval_studio_client/api/test/test_v1_create_test_from_test_cases_response.py +68 -0
  243. eval_studio_client/api/test/test_v1_dashboard_report.py +142 -0
  244. eval_studio_client/api/test/test_v1_dashboard_report_result.py +72 -0
  245. eval_studio_client/api/test/test_v1_dashboard_test_case_annotation.py +58 -0
  246. eval_studio_client/api/test/test_v1_data_fragment.py +57 -0
  247. eval_studio_client/api/test/test_v1_deep_compare_leaderboards_request.py +55 -0
  248. eval_studio_client/api/test/test_v1_deep_compare_leaderboards_response.py +255 -0
  249. eval_studio_client/api/test/test_v1_delete_leaderboard_response.py +2 -1
  250. eval_studio_client/api/test/test_v1_diff_item.py +226 -0
  251. eval_studio_client/api/test/test_v1_evaluation_type.py +33 -0
  252. eval_studio_client/api/test/test_v1_finalize_operation_response.py +2 -1
  253. eval_studio_client/api/test/test_v1_flipped_metric.py +53 -0
  254. eval_studio_client/api/test/test_v1_generate_test_cases_response.py +2 -1
  255. eval_studio_client/api/test/test_v1_get_dashboard_report_response.py +143 -0
  256. eval_studio_client/api/test/test_v1_get_info_response.py +4 -1
  257. eval_studio_client/api/test/test_v1_get_leaderboard_report_response.py +39 -2
  258. eval_studio_client/api/test/test_v1_get_leaderboard_response.py +2 -1
  259. eval_studio_client/api/test/test_v1_get_operation_response.py +2 -1
  260. eval_studio_client/api/test/test_v1_get_stats_response.py +3 -1
  261. eval_studio_client/api/test/test_v1_human_decision.py +33 -0
  262. eval_studio_client/api/test/test_v1_import_leaderboard_response.py +2 -1
  263. eval_studio_client/api/test/test_v1_import_test_cases_from_library_response.py +2 -1
  264. eval_studio_client/api/test/test_v1_info.py +4 -1
  265. eval_studio_client/api/test/test_v1_leaderboard.py +2 -1
  266. eval_studio_client/api/test/test_v1_leaderboard_cmp_report.py +254 -0
  267. eval_studio_client/api/test/test_v1_leaderboard_comparison_item.py +53 -0
  268. eval_studio_client/api/test/test_v1_leaderboard_info.py +57 -0
  269. eval_studio_client/api/test/test_v1_leaderboard_report.py +39 -2
  270. eval_studio_client/api/test/test_v1_leaderboard_report_actual_output_meta.py +33 -1
  271. eval_studio_client/api/test/test_v1_leaderboard_report_result.py +39 -2
  272. eval_studio_client/api/test/test_v1_leaderboard_report_result_view.py +33 -0
  273. eval_studio_client/api/test/test_v1_leaderboard_test_case_annotation.py +58 -0
  274. eval_studio_client/api/test/test_v1_list_dashboard_test_case_annotations_response.py +61 -0
  275. eval_studio_client/api/test/test_v1_list_leaderboard_test_case_annotations_response.py +61 -0
  276. eval_studio_client/api/test/test_v1_list_leaderboards_response.py +2 -1
  277. eval_studio_client/api/test/test_v1_list_most_recent_leaderboards_response.py +2 -1
  278. eval_studio_client/api/test/test_v1_list_operations_response.py +4 -2
  279. eval_studio_client/api/test/test_v1_list_unseen_operations_response.py +75 -0
  280. eval_studio_client/api/test/test_v1_mark_operation_seen_by_creator_response.py +72 -0
  281. eval_studio_client/api/test/test_v1_metric.py +52 -0
  282. eval_studio_client/api/test/test_v1_metric_average.py +58 -0
  283. eval_studio_client/api/test/test_v1_metric_meta.py +66 -0
  284. eval_studio_client/api/test/test_v1_models_comparisons.py +54 -0
  285. eval_studio_client/api/test/test_v1_models_comparisons_metrics.py +65 -0
  286. eval_studio_client/api/test/test_v1_models_overview.py +60 -0
  287. eval_studio_client/api/test/test_v1_operation.py +2 -1
  288. eval_studio_client/api/test/test_v1_operation_view.py +33 -0
  289. eval_studio_client/api/test/test_v1_process_workflow_node_response.py +2 -1
  290. eval_studio_client/api/test/test_v1_retrieved_context_diff.py +66 -0
  291. eval_studio_client/api/test/test_v1_stats.py +3 -1
  292. eval_studio_client/api/test/test_v1_technical_metrics.py +62 -0
  293. eval_studio_client/api/test/test_v1_technical_metrics_detail.py +55 -0
  294. eval_studio_client/api/test/test_v1_test_case_leaderboard_item.py +53 -0
  295. eval_studio_client/api/test/test_v1_test_case_relationship_info.py +53 -0
  296. eval_studio_client/api/test/test_v1_test_case_result.py +106 -0
  297. eval_studio_client/api/test/test_v1_text_similarity_metric.py +33 -0
  298. eval_studio_client/api/test/test_v1_update_dashboard_test_case_annotation_response.py +59 -0
  299. eval_studio_client/api/test/test_v1_update_leaderboard_response.py +2 -1
  300. eval_studio_client/api/test/test_v1_update_leaderboard_test_case_annotation_response.py +59 -0
  301. eval_studio_client/api/test/test_v1_update_operation_response.py +2 -1
  302. eval_studio_client/gen/openapiv2/eval_studio.swagger.json +2340 -210
  303. eval_studio_client/models.py +18 -6
  304. {eval_studio_client-1.2.5.dist-info → eval_studio_client-1.3.0.dist-info}/METADATA +2 -2
  305. {eval_studio_client-1.2.5.dist-info → eval_studio_client-1.3.0.dist-info}/RECORD +306 -111
  306. {eval_studio_client-1.2.5.dist-info → eval_studio_client-1.3.0.dist-info}/WHEEL +0 -0
@@ -19,8 +19,10 @@ from eval_studio_client.api.models.perturbation_service_create_perturbation_requ
19
19
  from eval_studio_client.api.models.prompt_generation_service_auto_generate_prompts_request import PromptGenerationServiceAutoGeneratePromptsRequest
20
20
  from eval_studio_client.api.models.protobuf_any import ProtobufAny
21
21
  from eval_studio_client.api.models.protobuf_null_value import ProtobufNullValue
22
+ from eval_studio_client.api.models.required_the_dashboard_test_case_annotation_to_update import RequiredTheDashboardTestCaseAnnotationToUpdate
22
23
  from eval_studio_client.api.models.required_the_dashboard_to_update import RequiredTheDashboardToUpdate
23
24
  from eval_studio_client.api.models.required_the_document_to_update import RequiredTheDocumentToUpdate
25
+ from eval_studio_client.api.models.required_the_leaderboard_test_case_annotation_to_update import RequiredTheLeaderboardTestCaseAnnotationToUpdate
24
26
  from eval_studio_client.api.models.required_the_leaderboard_to_update import RequiredTheLeaderboardToUpdate
25
27
  from eval_studio_client.api.models.required_the_model_to_update import RequiredTheModelToUpdate
26
28
  from eval_studio_client.api.models.required_the_operation_to_finalize import RequiredTheOperationToFinalize
@@ -30,6 +32,7 @@ from eval_studio_client.api.models.required_the_test_to_update import RequiredTh
30
32
  from eval_studio_client.api.models.required_the_updated_workflow import RequiredTheUpdatedWorkflow
31
33
  from eval_studio_client.api.models.required_the_updated_workflow_node import RequiredTheUpdatedWorkflowNode
32
34
  from eval_studio_client.api.models.rpc_status import RpcStatus
35
+ from eval_studio_client.api.models.test_case_service_append_test_cases_request import TestCaseServiceAppendTestCasesRequest
33
36
  from eval_studio_client.api.models.test_case_service_batch_delete_test_cases_request import TestCaseServiceBatchDeleteTestCasesRequest
34
37
  from eval_studio_client.api.models.test_service_clone_test_request import TestServiceCloneTestRequest
35
38
  from eval_studio_client.api.models.test_service_generate_test_cases_request import TestServiceGenerateTestCasesRequest
@@ -40,8 +43,22 @@ from eval_studio_client.api.models.test_service_perturb_test_in_place_request im
40
43
  from eval_studio_client.api.models.test_service_perturb_test_request import TestServicePerturbTestRequest
41
44
  from eval_studio_client.api.models.test_service_revoke_test_access_request import TestServiceRevokeTestAccessRequest
42
45
  from eval_studio_client.api.models.v1_abort_operation_response import V1AbortOperationResponse
46
+ from eval_studio_client.api.models.v1_actual_output_meta import V1ActualOutputMeta
47
+ from eval_studio_client.api.models.v1_actual_output_meta_diff import V1ActualOutputMetaDiff
48
+ from eval_studio_client.api.models.v1_agent_chat_activity_diagram import V1AgentChatActivityDiagram
49
+ from eval_studio_client.api.models.v1_agent_chat_activity_diagram_edge import V1AgentChatActivityDiagramEdge
50
+ from eval_studio_client.api.models.v1_agent_chat_activity_diagram_node import V1AgentChatActivityDiagramNode
51
+ from eval_studio_client.api.models.v1_agent_chat_activity_diagram_row import V1AgentChatActivityDiagramRow
52
+ from eval_studio_client.api.models.v1_agent_chat_script_usage import V1AgentChatScriptUsage
53
+ from eval_studio_client.api.models.v1_agent_chat_scripts_bar_chart import V1AgentChatScriptsBarChart
54
+ from eval_studio_client.api.models.v1_agent_chat_tool_usage import V1AgentChatToolUsage
55
+ from eval_studio_client.api.models.v1_agent_chat_tools_bar_chart import V1AgentChatToolsBarChart
56
+ from eval_studio_client.api.models.v1_all_metric_scores import V1AllMetricScores
57
+ from eval_studio_client.api.models.v1_append_test_cases_response import V1AppendTestCasesResponse
43
58
  from eval_studio_client.api.models.v1_batch_create_leaderboards_request import V1BatchCreateLeaderboardsRequest
44
59
  from eval_studio_client.api.models.v1_batch_create_leaderboards_response import V1BatchCreateLeaderboardsResponse
60
+ from eval_studio_client.api.models.v1_batch_create_leaderboards_without_cache_request import V1BatchCreateLeaderboardsWithoutCacheRequest
61
+ from eval_studio_client.api.models.v1_batch_create_leaderboards_without_cache_response import V1BatchCreateLeaderboardsWithoutCacheResponse
45
62
  from eval_studio_client.api.models.v1_batch_delete_dashboards_request import V1BatchDeleteDashboardsRequest
46
63
  from eval_studio_client.api.models.v1_batch_delete_dashboards_response import V1BatchDeleteDashboardsResponse
47
64
  from eval_studio_client.api.models.v1_batch_delete_documents_request import V1BatchDeleteDocumentsRequest
@@ -69,10 +86,17 @@ from eval_studio_client.api.models.v1_batch_import_leaderboard_request import V1
69
86
  from eval_studio_client.api.models.v1_batch_import_leaderboard_response import V1BatchImportLeaderboardResponse
70
87
  from eval_studio_client.api.models.v1_batch_import_tests_request import V1BatchImportTestsRequest
71
88
  from eval_studio_client.api.models.v1_batch_import_tests_response import V1BatchImportTestsResponse
89
+ from eval_studio_client.api.models.v1_batch_mark_operation_seen_by_creator_response import V1BatchMarkOperationSeenByCreatorResponse
72
90
  from eval_studio_client.api.models.v1_check_base_models_response import V1CheckBaseModelsResponse
73
91
  from eval_studio_client.api.models.v1_clone_test_response import V1CloneTestResponse
74
92
  from eval_studio_client.api.models.v1_clone_workflow_response import V1CloneWorkflowResponse
93
+ from eval_studio_client.api.models.v1_cmp_leaderboard_reports_request import V1CmpLeaderboardReportsRequest
94
+ from eval_studio_client.api.models.v1_cmp_leaderboard_reports_response import V1CmpLeaderboardReportsResponse
75
95
  from eval_studio_client.api.models.v1_collection_info import V1CollectionInfo
96
+ from eval_studio_client.api.models.v1_comparison_item import V1ComparisonItem
97
+ from eval_studio_client.api.models.v1_comparison_metric_score import V1ComparisonMetricScore
98
+ from eval_studio_client.api.models.v1_comparison_result import V1ComparisonResult
99
+ from eval_studio_client.api.models.v1_comparison_summary import V1ComparisonSummary
76
100
  from eval_studio_client.api.models.v1_context import V1Context
77
101
  from eval_studio_client.api.models.v1_create_dashboard_response import V1CreateDashboardResponse
78
102
  from eval_studio_client.api.models.v1_create_document_response import V1CreateDocumentResponse
@@ -84,14 +108,22 @@ from eval_studio_client.api.models.v1_create_leaderboard_without_cache_response
84
108
  from eval_studio_client.api.models.v1_create_model_response import V1CreateModelResponse
85
109
  from eval_studio_client.api.models.v1_create_perturbation_response import V1CreatePerturbationResponse
86
110
  from eval_studio_client.api.models.v1_create_test_case_response import V1CreateTestCaseResponse
111
+ from eval_studio_client.api.models.v1_create_test_from_test_cases_request import V1CreateTestFromTestCasesRequest
112
+ from eval_studio_client.api.models.v1_create_test_from_test_cases_response import V1CreateTestFromTestCasesResponse
87
113
  from eval_studio_client.api.models.v1_create_test_lab_response import V1CreateTestLabResponse
88
114
  from eval_studio_client.api.models.v1_create_test_response import V1CreateTestResponse
89
115
  from eval_studio_client.api.models.v1_create_workflow_edge_response import V1CreateWorkflowEdgeResponse
90
116
  from eval_studio_client.api.models.v1_create_workflow_node_response import V1CreateWorkflowNodeResponse
91
117
  from eval_studio_client.api.models.v1_create_workflow_response import V1CreateWorkflowResponse
92
118
  from eval_studio_client.api.models.v1_dashboard import V1Dashboard
119
+ from eval_studio_client.api.models.v1_dashboard_report import V1DashboardReport
120
+ from eval_studio_client.api.models.v1_dashboard_report_result import V1DashboardReportResult
93
121
  from eval_studio_client.api.models.v1_dashboard_status import V1DashboardStatus
122
+ from eval_studio_client.api.models.v1_dashboard_test_case_annotation import V1DashboardTestCaseAnnotation
94
123
  from eval_studio_client.api.models.v1_dashboard_type import V1DashboardType
124
+ from eval_studio_client.api.models.v1_data_fragment import V1DataFragment
125
+ from eval_studio_client.api.models.v1_deep_compare_leaderboards_request import V1DeepCompareLeaderboardsRequest
126
+ from eval_studio_client.api.models.v1_deep_compare_leaderboards_response import V1DeepCompareLeaderboardsResponse
95
127
  from eval_studio_client.api.models.v1_delete_dashboard_response import V1DeleteDashboardResponse
96
128
  from eval_studio_client.api.models.v1_delete_document_response import V1DeleteDocumentResponse
97
129
  from eval_studio_client.api.models.v1_delete_evaluator_response import V1DeleteEvaluatorResponse
@@ -103,9 +135,11 @@ from eval_studio_client.api.models.v1_delete_workflow_edge_response import V1Del
103
135
  from eval_studio_client.api.models.v1_delete_workflow_node_response import V1DeleteWorkflowNodeResponse
104
136
  from eval_studio_client.api.models.v1_delete_workflow_response import V1DeleteWorkflowResponse
105
137
  from eval_studio_client.api.models.v1_dependency_list import V1DependencyList
138
+ from eval_studio_client.api.models.v1_diff_item import V1DiffItem
106
139
  from eval_studio_client.api.models.v1_document import V1Document
107
140
  from eval_studio_client.api.models.v1_estimate_threshold_request import V1EstimateThresholdRequest
108
141
  from eval_studio_client.api.models.v1_evaluation_test import V1EvaluationTest
142
+ from eval_studio_client.api.models.v1_evaluation_type import V1EvaluationType
109
143
  from eval_studio_client.api.models.v1_evaluator import V1Evaluator
110
144
  from eval_studio_client.api.models.v1_evaluator_param_type import V1EvaluatorParamType
111
145
  from eval_studio_client.api.models.v1_evaluator_parameter import V1EvaluatorParameter
@@ -114,7 +148,9 @@ from eval_studio_client.api.models.v1_finalize_operation_response import V1Final
114
148
  from eval_studio_client.api.models.v1_find_all_test_cases_by_id_response import V1FindAllTestCasesByIDResponse
115
149
  from eval_studio_client.api.models.v1_find_test_lab_response import V1FindTestLabResponse
116
150
  from eval_studio_client.api.models.v1_find_workflows_by_collection_id_response import V1FindWorkflowsByCollectionIDResponse
151
+ from eval_studio_client.api.models.v1_flipped_metric import V1FlippedMetric
117
152
  from eval_studio_client.api.models.v1_generate_test_cases_response import V1GenerateTestCasesResponse
153
+ from eval_studio_client.api.models.v1_get_dashboard_report_response import V1GetDashboardReportResponse
118
154
  from eval_studio_client.api.models.v1_get_dashboard_response import V1GetDashboardResponse
119
155
  from eval_studio_client.api.models.v1_get_document_response import V1GetDocumentResponse
120
156
  from eval_studio_client.api.models.v1_get_evaluator_response import V1GetEvaluatorResponse
@@ -137,6 +173,7 @@ from eval_studio_client.api.models.v1_get_workflow_result_corpus_patch_response
137
173
  from eval_studio_client.api.models.v1_get_workflow_result_report_response import V1GetWorkflowResultReportResponse
138
174
  from eval_studio_client.api.models.v1_get_workflow_result_summary_response import V1GetWorkflowResultSummaryResponse
139
175
  from eval_studio_client.api.models.v1_get_workflow_result_system_prompt_patch_response import V1GetWorkflowResultSystemPromptPatchResponse
176
+ from eval_studio_client.api.models.v1_human_decision import V1HumanDecision
140
177
  from eval_studio_client.api.models.v1_import_evaluation_request import V1ImportEvaluationRequest
141
178
  from eval_studio_client.api.models.v1_import_leaderboard_request import V1ImportLeaderboardRequest
142
179
  from eval_studio_client.api.models.v1_import_leaderboard_response import V1ImportLeaderboardResponse
@@ -147,6 +184,9 @@ from eval_studio_client.api.models.v1_init_workflow_node_response import V1InitW
147
184
  from eval_studio_client.api.models.v1_insight import V1Insight
148
185
  from eval_studio_client.api.models.v1_labeled_test_case import V1LabeledTestCase
149
186
  from eval_studio_client.api.models.v1_leaderboard import V1Leaderboard
187
+ from eval_studio_client.api.models.v1_leaderboard_cmp_report import V1LeaderboardCmpReport
188
+ from eval_studio_client.api.models.v1_leaderboard_comparison_item import V1LeaderboardComparisonItem
189
+ from eval_studio_client.api.models.v1_leaderboard_info import V1LeaderboardInfo
150
190
  from eval_studio_client.api.models.v1_leaderboard_report import V1LeaderboardReport
151
191
  from eval_studio_client.api.models.v1_leaderboard_report_actual_output_data import V1LeaderboardReportActualOutputData
152
192
  from eval_studio_client.api.models.v1_leaderboard_report_actual_output_meta import V1LeaderboardReportActualOutputMeta
@@ -157,16 +197,20 @@ from eval_studio_client.api.models.v1_leaderboard_report_metrics_meta_entry impo
157
197
  from eval_studio_client.api.models.v1_leaderboard_report_model import V1LeaderboardReportModel
158
198
  from eval_studio_client.api.models.v1_leaderboard_report_result import V1LeaderboardReportResult
159
199
  from eval_studio_client.api.models.v1_leaderboard_report_result_relationship import V1LeaderboardReportResultRelationship
200
+ from eval_studio_client.api.models.v1_leaderboard_report_result_view import V1LeaderboardReportResultView
160
201
  from eval_studio_client.api.models.v1_leaderboard_status import V1LeaderboardStatus
202
+ from eval_studio_client.api.models.v1_leaderboard_test_case_annotation import V1LeaderboardTestCaseAnnotation
161
203
  from eval_studio_client.api.models.v1_leaderboard_type import V1LeaderboardType
162
204
  from eval_studio_client.api.models.v1_leaderboard_view import V1LeaderboardView
163
205
  from eval_studio_client.api.models.v1_list_base_models_response import V1ListBaseModelsResponse
164
206
  from eval_studio_client.api.models.v1_list_dashboard_access_response import V1ListDashboardAccessResponse
207
+ from eval_studio_client.api.models.v1_list_dashboard_test_case_annotations_response import V1ListDashboardTestCaseAnnotationsResponse
165
208
  from eval_studio_client.api.models.v1_list_dashboards_response import V1ListDashboardsResponse
166
209
  from eval_studio_client.api.models.v1_list_dashboards_shared_with_me_response import V1ListDashboardsSharedWithMeResponse
167
210
  from eval_studio_client.api.models.v1_list_documents_response import V1ListDocumentsResponse
168
211
  from eval_studio_client.api.models.v1_list_evaluators_response import V1ListEvaluatorsResponse
169
212
  from eval_studio_client.api.models.v1_list_llm_models_response import V1ListLLMModelsResponse
213
+ from eval_studio_client.api.models.v1_list_leaderboard_test_case_annotations_response import V1ListLeaderboardTestCaseAnnotationsResponse
170
214
  from eval_studio_client.api.models.v1_list_leaderboards_response import V1ListLeaderboardsResponse
171
215
  from eval_studio_client.api.models.v1_list_model_collections_response import V1ListModelCollectionsResponse
172
216
  from eval_studio_client.api.models.v1_list_models_response import V1ListModelsResponse
@@ -185,16 +229,25 @@ from eval_studio_client.api.models.v1_list_test_cases_response import V1ListTest
185
229
  from eval_studio_client.api.models.v1_list_test_classes_response import V1ListTestClassesResponse
186
230
  from eval_studio_client.api.models.v1_list_tests_response import V1ListTestsResponse
187
231
  from eval_studio_client.api.models.v1_list_tests_shared_with_me_response import V1ListTestsSharedWithMeResponse
232
+ from eval_studio_client.api.models.v1_list_unseen_operations_response import V1ListUnseenOperationsResponse
188
233
  from eval_studio_client.api.models.v1_list_workflow_access_response import V1ListWorkflowAccessResponse
189
234
  from eval_studio_client.api.models.v1_list_workflow_dependencies_response import V1ListWorkflowDependenciesResponse
190
235
  from eval_studio_client.api.models.v1_list_workflows_response import V1ListWorkflowsResponse
191
236
  from eval_studio_client.api.models.v1_list_workflows_shared_with_me_response import V1ListWorkflowsSharedWithMeResponse
237
+ from eval_studio_client.api.models.v1_mark_operation_seen_by_creator_response import V1MarkOperationSeenByCreatorResponse
238
+ from eval_studio_client.api.models.v1_metric import V1Metric
239
+ from eval_studio_client.api.models.v1_metric_average import V1MetricAverage
240
+ from eval_studio_client.api.models.v1_metric_meta import V1MetricMeta
192
241
  from eval_studio_client.api.models.v1_metric_score import V1MetricScore
193
242
  from eval_studio_client.api.models.v1_metric_scores import V1MetricScores
194
243
  from eval_studio_client.api.models.v1_model import V1Model
195
244
  from eval_studio_client.api.models.v1_model_type import V1ModelType
245
+ from eval_studio_client.api.models.v1_models_comparisons import V1ModelsComparisons
246
+ from eval_studio_client.api.models.v1_models_comparisons_metrics import V1ModelsComparisonsMetrics
247
+ from eval_studio_client.api.models.v1_models_overview import V1ModelsOverview
196
248
  from eval_studio_client.api.models.v1_operation import V1Operation
197
249
  from eval_studio_client.api.models.v1_operation_progress import V1OperationProgress
250
+ from eval_studio_client.api.models.v1_operation_view import V1OperationView
198
251
  from eval_studio_client.api.models.v1_perturb_test_in_place_response import V1PerturbTestInPlaceResponse
199
252
  from eval_studio_client.api.models.v1_perturb_test_response import V1PerturbTestResponse
200
253
  from eval_studio_client.api.models.v1_perturbator import V1Perturbator
@@ -206,21 +259,30 @@ from eval_studio_client.api.models.v1_prompt_library_item import V1PromptLibrary
206
259
  from eval_studio_client.api.models.v1_repeated_context import V1RepeatedContext
207
260
  from eval_studio_client.api.models.v1_repeated_string import V1RepeatedString
208
261
  from eval_studio_client.api.models.v1_reset_workflow_node_response import V1ResetWorkflowNodeResponse
262
+ from eval_studio_client.api.models.v1_retrieved_context_diff import V1RetrievedContextDiff
209
263
  from eval_studio_client.api.models.v1_role import V1Role
210
264
  from eval_studio_client.api.models.v1_role_binding import V1RoleBinding
211
265
  from eval_studio_client.api.models.v1_stats import V1Stats
266
+ from eval_studio_client.api.models.v1_technical_metrics import V1TechnicalMetrics
267
+ from eval_studio_client.api.models.v1_technical_metrics_detail import V1TechnicalMetricsDetail
212
268
  from eval_studio_client.api.models.v1_test import V1Test
213
269
  from eval_studio_client.api.models.v1_test_case import V1TestCase
270
+ from eval_studio_client.api.models.v1_test_case_leaderboard_item import V1TestCaseLeaderboardItem
214
271
  from eval_studio_client.api.models.v1_test_case_relationship import V1TestCaseRelationship
272
+ from eval_studio_client.api.models.v1_test_case_relationship_info import V1TestCaseRelationshipInfo
273
+ from eval_studio_client.api.models.v1_test_case_result import V1TestCaseResult
215
274
  from eval_studio_client.api.models.v1_test_cases_generator import V1TestCasesGenerator
216
275
  from eval_studio_client.api.models.v1_test_class import V1TestClass
217
276
  from eval_studio_client.api.models.v1_test_class_type import V1TestClassType
218
277
  from eval_studio_client.api.models.v1_test_lab import V1TestLab
219
278
  from eval_studio_client.api.models.v1_test_suite_evaluates import V1TestSuiteEvaluates
220
279
  from eval_studio_client.api.models.v1_test_type import V1TestType
280
+ from eval_studio_client.api.models.v1_text_similarity_metric import V1TextSimilarityMetric
221
281
  from eval_studio_client.api.models.v1_update_dashboard_response import V1UpdateDashboardResponse
282
+ from eval_studio_client.api.models.v1_update_dashboard_test_case_annotation_response import V1UpdateDashboardTestCaseAnnotationResponse
222
283
  from eval_studio_client.api.models.v1_update_document_response import V1UpdateDocumentResponse
223
284
  from eval_studio_client.api.models.v1_update_leaderboard_response import V1UpdateLeaderboardResponse
285
+ from eval_studio_client.api.models.v1_update_leaderboard_test_case_annotation_response import V1UpdateLeaderboardTestCaseAnnotationResponse
224
286
  from eval_studio_client.api.models.v1_update_model_response import V1UpdateModelResponse
225
287
  from eval_studio_client.api.models.v1_update_operation_response import V1UpdateOperationResponse
226
288
  from eval_studio_client.api.models.v1_update_test_case_response import V1UpdateTestCaseResponse
@@ -19,6 +19,7 @@ import json
19
19
 
20
20
  from pydantic import BaseModel, ConfigDict, Field, StrictInt, StrictStr
21
21
  from typing import Any, ClassVar, Dict, List, Optional
22
+ from eval_studio_client.api.models.v1_all_metric_scores import V1AllMetricScores
22
23
  from eval_studio_client.api.models.v1_metric_scores import V1MetricScores
23
24
  from eval_studio_client.api.models.v1_model import V1Model
24
25
  from eval_studio_client.api.models.v1_test_cases_generator import V1TestCasesGenerator
@@ -45,8 +46,9 @@ class AdversarialInputsServiceTestAdversarialInputsRobustnessRequest(BaseModel):
45
46
  model_parameters: Optional[StrictStr] = Field(default=None, description="Optional. Parameters overrides for the Model host in JSON format.", alias="modelParameters")
46
47
  default_h2ogpte_model: Optional[V1Model] = Field(default=None, alias="defaultH2ogpteModel")
47
48
  baseline_eval: Optional[StrictStr] = Field(default=None, description="Required. Baseline evaluation name.", alias="baselineEval")
48
- baseline_metrics: Optional[Dict[str, V1MetricScores]] = Field(default=None, description="Required. Map of baseline metrics from the evaluator to the metric scores for the evaluator.", alias="baselineMetrics")
49
- __properties: ClassVar[List[str]] = ["operation", "generatorInputTypes", "generatorDocumentUrls", "generatorModel", "generatorBaseLlmModel", "generatorCount", "generatorTopics", "generatorChunks", "generatorH2ogpteCollectionId", "evaluatorIdentifiers", "evaluatorsParameters", "model", "baseLlmModel", "modelParameters", "defaultH2ogpteModel", "baselineEval", "baselineMetrics"]
49
+ baseline_metrics: Optional[Dict[str, V1MetricScores]] = Field(default=None, description="Required. Map of baseline metrics from the evaluator to the average metric scores for the evaluator.", alias="baselineMetrics")
50
+ all_baseline_metrics_scores: Optional[Dict[str, V1AllMetricScores]] = Field(default=None, description="Required. Map of baseline metric to all and every test case metric score.", alias="allBaselineMetricsScores")
51
+ __properties: ClassVar[List[str]] = ["operation", "generatorInputTypes", "generatorDocumentUrls", "generatorModel", "generatorBaseLlmModel", "generatorCount", "generatorTopics", "generatorChunks", "generatorH2ogpteCollectionId", "evaluatorIdentifiers", "evaluatorsParameters", "model", "baseLlmModel", "modelParameters", "defaultH2ogpteModel", "baselineEval", "baselineMetrics", "allBaselineMetricsScores"]
50
52
 
51
53
  model_config = ConfigDict(
52
54
  populate_by_name=True,
@@ -103,6 +105,13 @@ class AdversarialInputsServiceTestAdversarialInputsRobustnessRequest(BaseModel):
103
105
  if self.baseline_metrics[_key_baseline_metrics]:
104
106
  _field_dict[_key_baseline_metrics] = self.baseline_metrics[_key_baseline_metrics].to_dict()
105
107
  _dict['baselineMetrics'] = _field_dict
108
+ # override the default output from pydantic by calling `to_dict()` of each value in all_baseline_metrics_scores (dict)
109
+ _field_dict = {}
110
+ if self.all_baseline_metrics_scores:
111
+ for _key_all_baseline_metrics_scores in self.all_baseline_metrics_scores:
112
+ if self.all_baseline_metrics_scores[_key_all_baseline_metrics_scores]:
113
+ _field_dict[_key_all_baseline_metrics_scores] = self.all_baseline_metrics_scores[_key_all_baseline_metrics_scores].to_dict()
114
+ _dict['allBaselineMetricsScores'] = _field_dict
106
115
  return _dict
107
116
 
108
117
  @classmethod
@@ -136,6 +145,12 @@ class AdversarialInputsServiceTestAdversarialInputsRobustnessRequest(BaseModel):
136
145
  for _k, _v in obj["baselineMetrics"].items()
137
146
  )
138
147
  if obj.get("baselineMetrics") is not None
148
+ else None,
149
+ "allBaselineMetricsScores": dict(
150
+ (_k, V1AllMetricScores.from_dict(_v))
151
+ for _k, _v in obj["allBaselineMetricsScores"].items()
152
+ )
153
+ if obj.get("allBaselineMetricsScores") is not None
139
154
  else None
140
155
  }, strict=False)
141
156
  return _obj
@@ -0,0 +1,108 @@
1
+ # coding: utf-8
2
+
3
+ """
4
+ ai/h2o/eval_studio/v1/insight.proto
5
+
6
+ No description provided (generated by Openapi Generator https://github.com/openapitools/openapi-generator)
7
+
8
+ The version of the OpenAPI document: version not set
9
+ Generated by OpenAPI Generator (https://openapi-generator.tech)
10
+
11
+ Do not edit the class manually.
12
+ """ # noqa: E501
13
+
14
+
15
+ from __future__ import annotations
16
+ import pprint
17
+ import re # noqa: F401
18
+ import json
19
+
20
+ from datetime import datetime
21
+ from pydantic import BaseModel, ConfigDict, Field, StrictStr
22
+ from typing import Any, ClassVar, Dict, List, Optional
23
+ from typing import Optional, Set
24
+ from typing_extensions import Self
25
+
26
+ class RequiredTheDashboardTestCaseAnnotationToUpdate(BaseModel):
27
+ """
28
+ RequiredTheDashboardTestCaseAnnotationToUpdate
29
+ """ # noqa: E501
30
+ create_time: Optional[datetime] = Field(default=None, description="Output only. Timestamp when the DashboardTestCaseAnnotation was created.", alias="createTime")
31
+ creator: Optional[StrictStr] = Field(default=None, description="Output only. Name of the user or service that requested creation of the DashboardTestCaseAnnotation.")
32
+ update_time: Optional[datetime] = Field(default=None, description="Output only. Optional. Timestamp when the DashboardTestCaseAnnotation was last updated.", alias="updateTime")
33
+ updater: Optional[StrictStr] = Field(default=None, description="Output only. Optional. Name of the user or service that requested update of the DashboardTestCaseAnnotation.")
34
+ parent: Optional[StrictStr] = Field(default=None, description="Parent Dashboard Test Case resource name. e.g.: \"dashboards/<UUID>/testCases/<UUID>\".")
35
+ key: Optional[StrictStr] = Field(default=None, description="Immutable. Annotation key.")
36
+ value: Optional[Dict[str, Any]] = Field(default=None, description="Annotation value.")
37
+ __properties: ClassVar[List[str]] = ["createTime", "creator", "updateTime", "updater", "parent", "key", "value"]
38
+
39
+ model_config = ConfigDict(
40
+ populate_by_name=True,
41
+ validate_assignment=True,
42
+ protected_namespaces=(),
43
+ )
44
+
45
+
46
+ def to_str(self) -> str:
47
+ """Returns the string representation of the model using alias"""
48
+ return pprint.pformat(self.model_dump(by_alias=True))
49
+
50
+ def to_json(self) -> str:
51
+ """Returns the JSON representation of the model using alias"""
52
+ # TODO: pydantic v2: use .model_dump_json(by_alias=True, exclude_unset=True) instead
53
+ return json.dumps(self.to_dict())
54
+
55
+ @classmethod
56
+ def from_json(cls, json_str: str) -> Optional[Self]:
57
+ """Create an instance of RequiredTheDashboardTestCaseAnnotationToUpdate from a JSON string"""
58
+ return cls.from_dict(json.loads(json_str))
59
+
60
+ def to_dict(self) -> Dict[str, Any]:
61
+ """Return the dictionary representation of the model using alias.
62
+
63
+ This has the following differences from calling pydantic's
64
+ `self.model_dump(by_alias=True)`:
65
+
66
+ * `None` is only added to the output dict for nullable fields that
67
+ were set at model initialization. Other fields with value `None`
68
+ are ignored.
69
+ * OpenAPI `readOnly` fields are excluded.
70
+ * OpenAPI `readOnly` fields are excluded.
71
+ * OpenAPI `readOnly` fields are excluded.
72
+ * OpenAPI `readOnly` fields are excluded.
73
+ """
74
+ excluded_fields: Set[str] = set([
75
+ "create_time",
76
+ "creator",
77
+ "update_time",
78
+ "updater",
79
+ ])
80
+
81
+ _dict = self.model_dump(
82
+ by_alias=True,
83
+ exclude=excluded_fields,
84
+ exclude_none=True,
85
+ )
86
+ return _dict
87
+
88
+ @classmethod
89
+ def from_dict(cls, obj: Optional[Dict[str, Any]]) -> Optional[Self]:
90
+ """Create an instance of RequiredTheDashboardTestCaseAnnotationToUpdate from a dict"""
91
+ if obj is None:
92
+ return None
93
+
94
+ if not isinstance(obj, dict):
95
+ return cls.model_validate(obj, strict=False)
96
+
97
+ _obj = cls.model_validate({
98
+ "createTime": obj.get("createTime"),
99
+ "creator": obj.get("creator"),
100
+ "updateTime": obj.get("updateTime"),
101
+ "updater": obj.get("updater"),
102
+ "parent": obj.get("parent"),
103
+ "key": obj.get("key"),
104
+ "value": obj.get("value")
105
+ }, strict=False)
106
+ return _obj
107
+
108
+
@@ -0,0 +1,108 @@
1
+ # coding: utf-8
2
+
3
+ """
4
+ ai/h2o/eval_studio/v1/insight.proto
5
+
6
+ No description provided (generated by Openapi Generator https://github.com/openapitools/openapi-generator)
7
+
8
+ The version of the OpenAPI document: version not set
9
+ Generated by OpenAPI Generator (https://openapi-generator.tech)
10
+
11
+ Do not edit the class manually.
12
+ """ # noqa: E501
13
+
14
+
15
+ from __future__ import annotations
16
+ import pprint
17
+ import re # noqa: F401
18
+ import json
19
+
20
+ from datetime import datetime
21
+ from pydantic import BaseModel, ConfigDict, Field, StrictStr
22
+ from typing import Any, ClassVar, Dict, List, Optional
23
+ from typing import Optional, Set
24
+ from typing_extensions import Self
25
+
26
+ class RequiredTheLeaderboardTestCaseAnnotationToUpdate(BaseModel):
27
+ """
28
+ RequiredTheLeaderboardTestCaseAnnotationToUpdate
29
+ """ # noqa: E501
30
+ create_time: Optional[datetime] = Field(default=None, description="Output only. Timestamp when the LeaderboardTestCaseAnnotation was created.", alias="createTime")
31
+ creator: Optional[StrictStr] = Field(default=None, description="Output only. Name of the user or service that requested creation of the LeaderboardTestCaseAnnotation.")
32
+ update_time: Optional[datetime] = Field(default=None, description="Output only. Optional. Timestamp when the LeaderboardTestCaseAnnotation was last updated.", alias="updateTime")
33
+ updater: Optional[StrictStr] = Field(default=None, description="Output only. Optional. Name of the user or service that requested update of the LeaderboardTestCaseAnnotation.")
34
+ parent: Optional[StrictStr] = Field(default=None, description="Parent Leaderboard Test Case resource name. e.g.: \"leaderboards/<UUID>/testCases/<UUID>\".")
35
+ key: Optional[StrictStr] = Field(default=None, description="Immutable. Annotation key.")
36
+ value: Optional[Dict[str, Any]] = Field(default=None, description="Annotation value.")
37
+ __properties: ClassVar[List[str]] = ["createTime", "creator", "updateTime", "updater", "parent", "key", "value"]
38
+
39
+ model_config = ConfigDict(
40
+ populate_by_name=True,
41
+ validate_assignment=True,
42
+ protected_namespaces=(),
43
+ )
44
+
45
+
46
+ def to_str(self) -> str:
47
+ """Returns the string representation of the model using alias"""
48
+ return pprint.pformat(self.model_dump(by_alias=True))
49
+
50
+ def to_json(self) -> str:
51
+ """Returns the JSON representation of the model using alias"""
52
+ # TODO: pydantic v2: use .model_dump_json(by_alias=True, exclude_unset=True) instead
53
+ return json.dumps(self.to_dict())
54
+
55
+ @classmethod
56
+ def from_json(cls, json_str: str) -> Optional[Self]:
57
+ """Create an instance of RequiredTheLeaderboardTestCaseAnnotationToUpdate from a JSON string"""
58
+ return cls.from_dict(json.loads(json_str))
59
+
60
+ def to_dict(self) -> Dict[str, Any]:
61
+ """Return the dictionary representation of the model using alias.
62
+
63
+ This has the following differences from calling pydantic's
64
+ `self.model_dump(by_alias=True)`:
65
+
66
+ * `None` is only added to the output dict for nullable fields that
67
+ were set at model initialization. Other fields with value `None`
68
+ are ignored.
69
+ * OpenAPI `readOnly` fields are excluded.
70
+ * OpenAPI `readOnly` fields are excluded.
71
+ * OpenAPI `readOnly` fields are excluded.
72
+ * OpenAPI `readOnly` fields are excluded.
73
+ """
74
+ excluded_fields: Set[str] = set([
75
+ "create_time",
76
+ "creator",
77
+ "update_time",
78
+ "updater",
79
+ ])
80
+
81
+ _dict = self.model_dump(
82
+ by_alias=True,
83
+ exclude=excluded_fields,
84
+ exclude_none=True,
85
+ )
86
+ return _dict
87
+
88
+ @classmethod
89
+ def from_dict(cls, obj: Optional[Dict[str, Any]]) -> Optional[Self]:
90
+ """Create an instance of RequiredTheLeaderboardTestCaseAnnotationToUpdate from a dict"""
91
+ if obj is None:
92
+ return None
93
+
94
+ if not isinstance(obj, dict):
95
+ return cls.model_validate(obj, strict=False)
96
+
97
+ _obj = cls.model_validate({
98
+ "createTime": obj.get("createTime"),
99
+ "creator": obj.get("creator"),
100
+ "updateTime": obj.get("updateTime"),
101
+ "updater": obj.get("updater"),
102
+ "parent": obj.get("parent"),
103
+ "key": obj.get("key"),
104
+ "value": obj.get("value")
105
+ }, strict=False)
106
+ return _obj
107
+
108
+
@@ -20,6 +20,7 @@ import json
20
20
  from datetime import datetime
21
21
  from pydantic import BaseModel, ConfigDict, Field, StrictBool, StrictStr
22
22
  from typing import Any, ClassVar, Dict, List, Optional
23
+ from eval_studio_client.api.models.v1_evaluation_type import V1EvaluationType
23
24
  from eval_studio_client.api.models.v1_insight import V1Insight
24
25
  from eval_studio_client.api.models.v1_leaderboard_status import V1LeaderboardStatus
25
26
  from eval_studio_client.api.models.v1_leaderboard_type import V1LeaderboardType
@@ -56,7 +57,8 @@ class RequiredTheLeaderboardToUpdate(BaseModel):
56
57
  type: Optional[V1LeaderboardType] = None
57
58
  demo: Optional[StrictBool] = Field(default=None, description="Output only. Whether the Leaderboard is a demo resource or not. Demo resources are read only.")
58
59
  test_lab: Optional[StrictStr] = Field(default=None, description="Optional. Resource name of the TestLab if Leaderboard was created from a imported TestLab.", alias="testLab")
59
- __properties: ClassVar[List[str]] = ["createTime", "creator", "updateTime", "updater", "deleteTime", "deleter", "displayName", "description", "status", "evaluator", "tests", "model", "createOperation", "leaderboardReport", "leaderboardTable", "leaderboardSummary", "llmModels", "leaderboardProblems", "evaluatorParameters", "insights", "modelParameters", "h2ogpteCollection", "type", "demo", "testLab"]
60
+ evaluation_type: Optional[V1EvaluationType] = Field(default=None, alias="evaluationType")
61
+ __properties: ClassVar[List[str]] = ["createTime", "creator", "updateTime", "updater", "deleteTime", "deleter", "displayName", "description", "status", "evaluator", "tests", "model", "createOperation", "leaderboardReport", "leaderboardTable", "leaderboardSummary", "llmModels", "leaderboardProblems", "evaluatorParameters", "insights", "modelParameters", "h2ogpteCollection", "type", "demo", "testLab", "evaluationType"]
60
62
 
61
63
  model_config = ConfigDict(
62
64
  populate_by_name=True,
@@ -171,7 +173,8 @@ class RequiredTheLeaderboardToUpdate(BaseModel):
171
173
  "h2ogpteCollection": obj.get("h2ogpteCollection"),
172
174
  "type": obj.get("type"),
173
175
  "demo": obj.get("demo"),
174
- "testLab": obj.get("testLab")
176
+ "testLab": obj.get("testLab"),
177
+ "evaluationType": obj.get("evaluationType")
175
178
  }, strict=False)
176
179
  return _obj
177
180
 
@@ -39,7 +39,8 @@ class RequiredTheOperationToFinalize(BaseModel):
39
39
  done: Optional[StrictBool] = Field(default=None, description="If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.")
40
40
  error: Optional[RpcStatus] = None
41
41
  response: Optional[ProtobufAny] = None
42
- __properties: ClassVar[List[str]] = ["createTime", "creator", "updateTime", "updater", "deleteTime", "deleter", "metadata", "done", "error", "response"]
42
+ seen_by_creator_time: Optional[datetime] = Field(default=None, description="Output only. Optional. Timestamp when the creator marked the Operation as seen. Once set, this field cannot be changed. Set via MarkOperationSeenByCreator method.", alias="seenByCreatorTime")
43
+ __properties: ClassVar[List[str]] = ["createTime", "creator", "updateTime", "updater", "deleteTime", "deleter", "metadata", "done", "error", "response", "seenByCreatorTime"]
43
44
 
44
45
  model_config = ConfigDict(
45
46
  populate_by_name=True,
@@ -77,6 +78,7 @@ class RequiredTheOperationToFinalize(BaseModel):
77
78
  * OpenAPI `readOnly` fields are excluded.
78
79
  * OpenAPI `readOnly` fields are excluded.
79
80
  * OpenAPI `readOnly` fields are excluded.
81
+ * OpenAPI `readOnly` fields are excluded.
80
82
  """
81
83
  excluded_fields: Set[str] = set([
82
84
  "create_time",
@@ -85,6 +87,7 @@ class RequiredTheOperationToFinalize(BaseModel):
85
87
  "updater",
86
88
  "delete_time",
87
89
  "deleter",
90
+ "seen_by_creator_time",
88
91
  ])
89
92
 
90
93
  _dict = self.model_dump(
@@ -122,7 +125,8 @@ class RequiredTheOperationToFinalize(BaseModel):
122
125
  "metadata": ProtobufAny.from_dict(obj["metadata"]) if obj.get("metadata") is not None else None,
123
126
  "done": obj.get("done"),
124
127
  "error": RpcStatus.from_dict(obj["error"]) if obj.get("error") is not None else None,
125
- "response": ProtobufAny.from_dict(obj["response"]) if obj.get("response") is not None else None
128
+ "response": ProtobufAny.from_dict(obj["response"]) if obj.get("response") is not None else None,
129
+ "seenByCreatorTime": obj.get("seenByCreatorTime")
126
130
  }, strict=False)
127
131
  return _obj
128
132
 
@@ -39,7 +39,8 @@ class RequiredTheOperationToUpdate(BaseModel):
39
39
  done: Optional[StrictBool] = Field(default=None, description="If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.")
40
40
  error: Optional[RpcStatus] = None
41
41
  response: Optional[ProtobufAny] = None
42
- __properties: ClassVar[List[str]] = ["createTime", "creator", "updateTime", "updater", "deleteTime", "deleter", "metadata", "done", "error", "response"]
42
+ seen_by_creator_time: Optional[datetime] = Field(default=None, description="Output only. Optional. Timestamp when the creator marked the Operation as seen. Once set, this field cannot be changed. Set via MarkOperationSeenByCreator method.", alias="seenByCreatorTime")
43
+ __properties: ClassVar[List[str]] = ["createTime", "creator", "updateTime", "updater", "deleteTime", "deleter", "metadata", "done", "error", "response", "seenByCreatorTime"]
43
44
 
44
45
  model_config = ConfigDict(
45
46
  populate_by_name=True,
@@ -77,6 +78,7 @@ class RequiredTheOperationToUpdate(BaseModel):
77
78
  * OpenAPI `readOnly` fields are excluded.
78
79
  * OpenAPI `readOnly` fields are excluded.
79
80
  * OpenAPI `readOnly` fields are excluded.
81
+ * OpenAPI `readOnly` fields are excluded.
80
82
  """
81
83
  excluded_fields: Set[str] = set([
82
84
  "create_time",
@@ -85,6 +87,7 @@ class RequiredTheOperationToUpdate(BaseModel):
85
87
  "updater",
86
88
  "delete_time",
87
89
  "deleter",
90
+ "seen_by_creator_time",
88
91
  ])
89
92
 
90
93
  _dict = self.model_dump(
@@ -122,7 +125,8 @@ class RequiredTheOperationToUpdate(BaseModel):
122
125
  "metadata": ProtobufAny.from_dict(obj["metadata"]) if obj.get("metadata") is not None else None,
123
126
  "done": obj.get("done"),
124
127
  "error": RpcStatus.from_dict(obj["error"]) if obj.get("error") is not None else None,
125
- "response": ProtobufAny.from_dict(obj["response"]) if obj.get("response") is not None else None
128
+ "response": ProtobufAny.from_dict(obj["response"]) if obj.get("response") is not None else None,
129
+ "seenByCreatorTime": obj.get("seenByCreatorTime")
126
130
  }, strict=False)
127
131
  return _obj
128
132