eval-studio-client 1.2.5__py3-none-any.whl → 1.3.0a1__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (306) hide show
  1. eval_studio_client/api/__init__.py +65 -0
  2. eval_studio_client/api/api/__init__.py +3 -0
  3. eval_studio_client/api/api/dashboard_report_service_api.py +292 -0
  4. eval_studio_client/api/api/dashboard_service_api.py +16 -16
  5. eval_studio_client/api/api/dashboard_test_case_annotation_service_api.py +611 -0
  6. eval_studio_client/api/api/document_service_api.py +16 -16
  7. eval_studio_client/api/api/evaluation_service_api.py +12 -12
  8. eval_studio_client/api/api/evaluator_service_api.py +16 -16
  9. eval_studio_client/api/api/leaderboard_report_service_api.py +304 -17
  10. eval_studio_client/api/api/leaderboard_service_api.py +554 -16
  11. eval_studio_client/api/api/leaderboard_test_case_annotation_service_api.py +611 -0
  12. eval_studio_client/api/api/model_service_api.py +16 -16
  13. eval_studio_client/api/api/operation_service_api.py +821 -17
  14. eval_studio_client/api/api/perturbator_service_api.py +22 -22
  15. eval_studio_client/api/api/test_case_service_api.py +300 -16
  16. eval_studio_client/api/api/test_class_service_api.py +16 -16
  17. eval_studio_client/api/api/test_service_api.py +285 -16
  18. eval_studio_client/api/api/workflow_node_service_api.py +16 -16
  19. eval_studio_client/api/api/workflow_service_api.py +16 -16
  20. eval_studio_client/api/docs/AdversarialInputsServiceTestAdversarialInputsRobustnessRequest.md +2 -1
  21. eval_studio_client/api/docs/DashboardReportServiceApi.md +75 -0
  22. eval_studio_client/api/docs/DashboardServiceApi.md +5 -5
  23. eval_studio_client/api/docs/DashboardTestCaseAnnotationServiceApi.md +149 -0
  24. eval_studio_client/api/docs/DocumentServiceApi.md +5 -5
  25. eval_studio_client/api/docs/EvaluationServiceApi.md +4 -4
  26. eval_studio_client/api/docs/EvaluatorServiceApi.md +5 -5
  27. eval_studio_client/api/docs/LeaderboardReportServiceApi.md +75 -5
  28. eval_studio_client/api/docs/LeaderboardServiceApi.md +141 -5
  29. eval_studio_client/api/docs/LeaderboardTestCaseAnnotationServiceApi.md +149 -0
  30. eval_studio_client/api/docs/ModelServiceApi.md +5 -5
  31. eval_studio_client/api/docs/OperationServiceApi.md +215 -8
  32. eval_studio_client/api/docs/PerturbatorServiceApi.md +7 -7
  33. eval_studio_client/api/docs/RequiredTheDashboardTestCaseAnnotationToUpdate.md +35 -0
  34. eval_studio_client/api/docs/RequiredTheLeaderboardTestCaseAnnotationToUpdate.md +35 -0
  35. eval_studio_client/api/docs/RequiredTheLeaderboardToUpdate.md +1 -0
  36. eval_studio_client/api/docs/RequiredTheOperationToFinalize.md +1 -0
  37. eval_studio_client/api/docs/RequiredTheOperationToUpdate.md +1 -0
  38. eval_studio_client/api/docs/TestCaseServiceApi.md +75 -5
  39. eval_studio_client/api/docs/TestCaseServiceAppendTestCasesRequest.md +30 -0
  40. eval_studio_client/api/docs/TestClassServiceApi.md +5 -5
  41. eval_studio_client/api/docs/TestServiceApi.md +73 -5
  42. eval_studio_client/api/docs/V1ActualOutputMeta.md +30 -0
  43. eval_studio_client/api/docs/V1ActualOutputMetaDiff.md +36 -0
  44. eval_studio_client/api/docs/V1AgentChatActivityDiagram.md +31 -0
  45. eval_studio_client/api/docs/V1AgentChatActivityDiagramEdge.md +32 -0
  46. eval_studio_client/api/docs/V1AgentChatActivityDiagramNode.md +32 -0
  47. eval_studio_client/api/docs/V1AgentChatActivityDiagramRow.md +30 -0
  48. eval_studio_client/api/docs/V1AgentChatScriptUsage.md +33 -0
  49. eval_studio_client/api/docs/V1AgentChatScriptsBarChart.md +30 -0
  50. eval_studio_client/api/docs/V1AgentChatToolUsage.md +33 -0
  51. eval_studio_client/api/docs/V1AgentChatToolsBarChart.md +30 -0
  52. eval_studio_client/api/docs/V1AllMetricScores.md +29 -0
  53. eval_studio_client/api/docs/V1AppendTestCasesResponse.md +29 -0
  54. eval_studio_client/api/docs/V1BatchCreateLeaderboardsWithoutCacheRequest.md +31 -0
  55. eval_studio_client/api/docs/V1BatchCreateLeaderboardsWithoutCacheResponse.md +29 -0
  56. eval_studio_client/api/docs/V1BatchMarkOperationSeenByCreatorResponse.md +29 -0
  57. eval_studio_client/api/docs/V1CmpLeaderboardReportsRequest.md +33 -0
  58. eval_studio_client/api/docs/V1CmpLeaderboardReportsResponse.md +29 -0
  59. eval_studio_client/api/docs/V1ComparisonItem.md +36 -0
  60. eval_studio_client/api/docs/V1ComparisonMetricScore.md +30 -0
  61. eval_studio_client/api/docs/V1ComparisonResult.md +31 -0
  62. eval_studio_client/api/docs/V1ComparisonSummary.md +31 -0
  63. eval_studio_client/api/docs/V1CreateEvaluationRequest.md +1 -0
  64. eval_studio_client/api/docs/V1CreateTestFromTestCasesRequest.md +32 -0
  65. eval_studio_client/api/docs/V1CreateTestFromTestCasesResponse.md +29 -0
  66. eval_studio_client/api/docs/V1DashboardReport.md +31 -0
  67. eval_studio_client/api/docs/V1DashboardReportResult.md +39 -0
  68. eval_studio_client/api/docs/V1DashboardTestCaseAnnotation.md +36 -0
  69. eval_studio_client/api/docs/V1DataFragment.md +31 -0
  70. eval_studio_client/api/docs/V1DeepCompareLeaderboardsRequest.md +33 -0
  71. eval_studio_client/api/docs/V1DeepCompareLeaderboardsResponse.md +29 -0
  72. eval_studio_client/api/docs/V1DiffItem.md +36 -0
  73. eval_studio_client/api/docs/V1EvaluationType.md +12 -0
  74. eval_studio_client/api/docs/V1FlippedMetric.md +31 -0
  75. eval_studio_client/api/docs/V1GetDashboardReportResponse.md +29 -0
  76. eval_studio_client/api/docs/V1HumanDecision.md +12 -0
  77. eval_studio_client/api/docs/V1Info.md +1 -0
  78. eval_studio_client/api/docs/V1Leaderboard.md +1 -0
  79. eval_studio_client/api/docs/V1LeaderboardCmpReport.md +30 -0
  80. eval_studio_client/api/docs/V1LeaderboardComparisonItem.md +31 -0
  81. eval_studio_client/api/docs/V1LeaderboardInfo.md +30 -0
  82. eval_studio_client/api/docs/V1LeaderboardReportActualOutputMeta.md +6 -3
  83. eval_studio_client/api/docs/V1LeaderboardReportResult.md +11 -8
  84. eval_studio_client/api/docs/V1LeaderboardReportResultView.md +12 -0
  85. eval_studio_client/api/docs/V1LeaderboardTestCaseAnnotation.md +36 -0
  86. eval_studio_client/api/docs/V1ListDashboardTestCaseAnnotationsResponse.md +29 -0
  87. eval_studio_client/api/docs/V1ListLeaderboardTestCaseAnnotationsResponse.md +29 -0
  88. eval_studio_client/api/docs/V1ListOperationsResponse.md +1 -0
  89. eval_studio_client/api/docs/V1ListUnseenOperationsResponse.md +30 -0
  90. eval_studio_client/api/docs/V1MarkOperationSeenByCreatorResponse.md +29 -0
  91. eval_studio_client/api/docs/V1Metric.md +30 -0
  92. eval_studio_client/api/docs/V1MetricAverage.md +36 -0
  93. eval_studio_client/api/docs/V1MetricMeta.md +40 -0
  94. eval_studio_client/api/docs/V1MetricScore.md +1 -1
  95. eval_studio_client/api/docs/V1MetricScores.md +1 -1
  96. eval_studio_client/api/docs/V1ModelType.md +1 -1
  97. eval_studio_client/api/docs/V1ModelsComparisons.md +32 -0
  98. eval_studio_client/api/docs/V1ModelsComparisonsMetrics.md +33 -0
  99. eval_studio_client/api/docs/V1ModelsOverview.md +34 -0
  100. eval_studio_client/api/docs/V1Operation.md +1 -0
  101. eval_studio_client/api/docs/V1OperationView.md +12 -0
  102. eval_studio_client/api/docs/V1RetrievedContextDiff.md +36 -0
  103. eval_studio_client/api/docs/V1Stats.md +2 -0
  104. eval_studio_client/api/docs/V1TechnicalMetrics.md +30 -0
  105. eval_studio_client/api/docs/V1TechnicalMetricsDetail.md +33 -0
  106. eval_studio_client/api/docs/V1TestCaseLeaderboardItem.md +31 -0
  107. eval_studio_client/api/docs/V1TestCaseRelationshipInfo.md +31 -0
  108. eval_studio_client/api/docs/V1TestCaseResult.md +48 -0
  109. eval_studio_client/api/docs/V1TextSimilarityMetric.md +12 -0
  110. eval_studio_client/api/docs/V1UpdateDashboardTestCaseAnnotationResponse.md +29 -0
  111. eval_studio_client/api/docs/V1UpdateLeaderboardTestCaseAnnotationResponse.md +29 -0
  112. eval_studio_client/api/docs/WorkflowNodeServiceApi.md +5 -5
  113. eval_studio_client/api/docs/WorkflowServiceApi.md +5 -5
  114. eval_studio_client/api/models/__init__.py +62 -0
  115. eval_studio_client/api/models/adversarial_inputs_service_test_adversarial_inputs_robustness_request.py +17 -2
  116. eval_studio_client/api/models/required_the_dashboard_test_case_annotation_to_update.py +108 -0
  117. eval_studio_client/api/models/required_the_leaderboard_test_case_annotation_to_update.py +108 -0
  118. eval_studio_client/api/models/required_the_leaderboard_to_update.py +5 -2
  119. eval_studio_client/api/models/required_the_operation_to_finalize.py +6 -2
  120. eval_studio_client/api/models/required_the_operation_to_update.py +6 -2
  121. eval_studio_client/api/models/test_case_service_append_test_cases_request.py +89 -0
  122. eval_studio_client/api/models/v1_actual_output_meta.py +97 -0
  123. eval_studio_client/api/models/v1_actual_output_meta_diff.py +101 -0
  124. eval_studio_client/api/models/v1_agent_chat_activity_diagram.py +109 -0
  125. eval_studio_client/api/models/v1_agent_chat_activity_diagram_edge.py +97 -0
  126. eval_studio_client/api/models/v1_agent_chat_activity_diagram_node.py +97 -0
  127. eval_studio_client/api/models/v1_agent_chat_activity_diagram_row.py +97 -0
  128. eval_studio_client/api/models/v1_agent_chat_script_usage.py +101 -0
  129. eval_studio_client/api/models/v1_agent_chat_scripts_bar_chart.py +102 -0
  130. eval_studio_client/api/models/v1_agent_chat_tool_usage.py +101 -0
  131. eval_studio_client/api/models/v1_agent_chat_tools_bar_chart.py +102 -0
  132. eval_studio_client/api/models/v1_all_metric_scores.py +87 -0
  133. eval_studio_client/api/models/v1_append_test_cases_response.py +95 -0
  134. eval_studio_client/api/models/v1_batch_create_leaderboards_without_cache_request.py +99 -0
  135. eval_studio_client/api/models/v1_batch_create_leaderboards_without_cache_response.py +91 -0
  136. eval_studio_client/api/models/v1_batch_mark_operation_seen_by_creator_response.py +95 -0
  137. eval_studio_client/api/models/v1_cmp_leaderboard_reports_request.py +96 -0
  138. eval_studio_client/api/models/v1_cmp_leaderboard_reports_response.py +91 -0
  139. eval_studio_client/api/models/v1_comparison_item.py +130 -0
  140. eval_studio_client/api/models/v1_comparison_metric_score.py +89 -0
  141. eval_studio_client/api/models/v1_comparison_result.py +120 -0
  142. eval_studio_client/api/models/v1_comparison_summary.py +91 -0
  143. eval_studio_client/api/models/v1_create_evaluation_request.py +5 -2
  144. eval_studio_client/api/models/v1_create_test_from_test_cases_request.py +93 -0
  145. eval_studio_client/api/models/v1_create_test_from_test_cases_response.py +91 -0
  146. eval_studio_client/api/models/v1_dashboard_report.py +109 -0
  147. eval_studio_client/api/models/v1_dashboard_report_result.py +139 -0
  148. eval_studio_client/api/models/v1_dashboard_test_case_annotation.py +112 -0
  149. eval_studio_client/api/models/v1_data_fragment.py +91 -0
  150. eval_studio_client/api/models/v1_deep_compare_leaderboards_request.py +96 -0
  151. eval_studio_client/api/models/v1_deep_compare_leaderboards_response.py +91 -0
  152. eval_studio_client/api/models/v1_diff_item.py +137 -0
  153. eval_studio_client/api/models/v1_evaluation_type.py +39 -0
  154. eval_studio_client/api/models/v1_flipped_metric.py +91 -0
  155. eval_studio_client/api/models/v1_get_dashboard_report_response.py +91 -0
  156. eval_studio_client/api/models/v1_human_decision.py +38 -0
  157. eval_studio_client/api/models/v1_info.py +4 -2
  158. eval_studio_client/api/models/v1_leaderboard.py +5 -2
  159. eval_studio_client/api/models/v1_leaderboard_cmp_report.py +93 -0
  160. eval_studio_client/api/models/v1_leaderboard_comparison_item.py +91 -0
  161. eval_studio_client/api/models/v1_leaderboard_info.py +97 -0
  162. eval_studio_client/api/models/v1_leaderboard_report_actual_output_meta.py +23 -9
  163. eval_studio_client/api/models/v1_leaderboard_report_result.py +21 -10
  164. eval_studio_client/api/models/v1_leaderboard_report_result_view.py +38 -0
  165. eval_studio_client/api/models/v1_leaderboard_test_case_annotation.py +112 -0
  166. eval_studio_client/api/models/v1_list_dashboard_test_case_annotations_response.py +95 -0
  167. eval_studio_client/api/models/v1_list_leaderboard_test_case_annotations_response.py +95 -0
  168. eval_studio_client/api/models/v1_list_operations_response.py +5 -3
  169. eval_studio_client/api/models/v1_list_unseen_operations_response.py +97 -0
  170. eval_studio_client/api/models/v1_mark_operation_seen_by_creator_response.py +91 -0
  171. eval_studio_client/api/models/v1_metric.py +89 -0
  172. eval_studio_client/api/models/v1_metric_average.py +101 -0
  173. eval_studio_client/api/models/v1_metric_meta.py +109 -0
  174. eval_studio_client/api/models/v1_metric_score.py +6 -1
  175. eval_studio_client/api/models/v1_metric_scores.py +1 -1
  176. eval_studio_client/api/models/v1_model_type.py +2 -1
  177. eval_studio_client/api/models/v1_models_comparisons.py +93 -0
  178. eval_studio_client/api/models/v1_models_comparisons_metrics.py +103 -0
  179. eval_studio_client/api/models/v1_models_overview.py +97 -0
  180. eval_studio_client/api/models/v1_operation.py +6 -2
  181. eval_studio_client/api/models/v1_operation_view.py +38 -0
  182. eval_studio_client/api/models/v1_retrieved_context_diff.py +101 -0
  183. eval_studio_client/api/models/v1_stats.py +16 -2
  184. eval_studio_client/api/models/v1_technical_metrics.py +96 -0
  185. eval_studio_client/api/models/v1_technical_metrics_detail.py +95 -0
  186. eval_studio_client/api/models/v1_test_case_leaderboard_item.py +91 -0
  187. eval_studio_client/api/models/v1_test_case_relationship_info.py +91 -0
  188. eval_studio_client/api/models/v1_test_case_result.py +157 -0
  189. eval_studio_client/api/models/v1_text_similarity_metric.py +39 -0
  190. eval_studio_client/api/models/v1_update_dashboard_test_case_annotation_response.py +91 -0
  191. eval_studio_client/api/models/v1_update_leaderboard_test_case_annotation_response.py +91 -0
  192. eval_studio_client/api/models/v1_workflow_node_type.py +1 -0
  193. eval_studio_client/api/models/v1_workflow_type.py +1 -0
  194. eval_studio_client/api/test/test_adversarial_inputs_service_test_adversarial_inputs_robustness_request.py +6 -0
  195. eval_studio_client/api/test/test_dashboard_report_service_api.py +37 -0
  196. eval_studio_client/api/test/test_dashboard_test_case_annotation_service_api.py +43 -0
  197. eval_studio_client/api/test/test_leaderboard_report_service_api.py +6 -0
  198. eval_studio_client/api/test/test_leaderboard_service_api.py +12 -0
  199. eval_studio_client/api/test/test_leaderboard_test_case_annotation_service_api.py +43 -0
  200. eval_studio_client/api/test/test_operation_service_api.py +18 -0
  201. eval_studio_client/api/test/test_required_the_dashboard_test_case_annotation_to_update.py +57 -0
  202. eval_studio_client/api/test/test_required_the_leaderboard_test_case_annotation_to_update.py +57 -0
  203. eval_studio_client/api/test/test_required_the_leaderboard_to_update.py +2 -1
  204. eval_studio_client/api/test/test_required_the_operation_to_finalize.py +2 -1
  205. eval_studio_client/api/test/test_required_the_operation_to_update.py +2 -1
  206. eval_studio_client/api/test/test_test_case_service_api.py +6 -0
  207. eval_studio_client/api/test/test_test_case_service_append_test_cases_request.py +52 -0
  208. eval_studio_client/api/test/test_test_service_api.py +6 -0
  209. eval_studio_client/api/test/test_v1_abort_operation_response.py +2 -1
  210. eval_studio_client/api/test/test_v1_actual_output_meta.py +61 -0
  211. eval_studio_client/api/test/test_v1_actual_output_meta_diff.py +66 -0
  212. eval_studio_client/api/test/test_v1_agent_chat_activity_diagram.py +65 -0
  213. eval_studio_client/api/test/test_v1_agent_chat_activity_diagram_edge.py +53 -0
  214. eval_studio_client/api/test/test_v1_agent_chat_activity_diagram_node.py +53 -0
  215. eval_studio_client/api/test/test_v1_agent_chat_activity_diagram_row.py +56 -0
  216. eval_studio_client/api/test/test_v1_agent_chat_script_usage.py +54 -0
  217. eval_studio_client/api/test/test_v1_agent_chat_scripts_bar_chart.py +57 -0
  218. eval_studio_client/api/test/test_v1_agent_chat_tool_usage.py +54 -0
  219. eval_studio_client/api/test/test_v1_agent_chat_tools_bar_chart.py +57 -0
  220. eval_studio_client/api/test/test_v1_all_metric_scores.py +53 -0
  221. eval_studio_client/api/test/test_v1_append_test_cases_response.py +74 -0
  222. eval_studio_client/api/test/test_v1_batch_create_leaderboards_request.py +2 -1
  223. eval_studio_client/api/test/test_v1_batch_create_leaderboards_response.py +2 -1
  224. eval_studio_client/api/test/test_v1_batch_create_leaderboards_without_cache_request.py +120 -0
  225. eval_studio_client/api/test/test_v1_batch_create_leaderboards_without_cache_response.py +72 -0
  226. eval_studio_client/api/test/test_v1_batch_delete_leaderboards_response.py +2 -1
  227. eval_studio_client/api/test/test_v1_batch_get_leaderboards_response.py +2 -1
  228. eval_studio_client/api/test/test_v1_batch_get_operations_response.py +2 -1
  229. eval_studio_client/api/test/test_v1_batch_import_leaderboard_response.py +2 -1
  230. eval_studio_client/api/test/test_v1_batch_mark_operation_seen_by_creator_response.py +74 -0
  231. eval_studio_client/api/test/test_v1_cmp_leaderboard_reports_request.py +55 -0
  232. eval_studio_client/api/test/test_v1_cmp_leaderboard_reports_response.py +255 -0
  233. eval_studio_client/api/test/test_v1_comparison_item.py +233 -0
  234. eval_studio_client/api/test/test_v1_comparison_metric_score.py +52 -0
  235. eval_studio_client/api/test/test_v1_comparison_result.py +258 -0
  236. eval_studio_client/api/test/test_v1_comparison_summary.py +53 -0
  237. eval_studio_client/api/test/test_v1_create_evaluation_request.py +2 -1
  238. eval_studio_client/api/test/test_v1_create_leaderboard_request.py +2 -1
  239. eval_studio_client/api/test/test_v1_create_leaderboard_response.py +2 -1
  240. eval_studio_client/api/test/test_v1_create_leaderboard_without_cache_response.py +2 -1
  241. eval_studio_client/api/test/test_v1_create_test_from_test_cases_request.py +54 -0
  242. eval_studio_client/api/test/test_v1_create_test_from_test_cases_response.py +68 -0
  243. eval_studio_client/api/test/test_v1_dashboard_report.py +142 -0
  244. eval_studio_client/api/test/test_v1_dashboard_report_result.py +72 -0
  245. eval_studio_client/api/test/test_v1_dashboard_test_case_annotation.py +58 -0
  246. eval_studio_client/api/test/test_v1_data_fragment.py +57 -0
  247. eval_studio_client/api/test/test_v1_deep_compare_leaderboards_request.py +55 -0
  248. eval_studio_client/api/test/test_v1_deep_compare_leaderboards_response.py +255 -0
  249. eval_studio_client/api/test/test_v1_delete_leaderboard_response.py +2 -1
  250. eval_studio_client/api/test/test_v1_diff_item.py +226 -0
  251. eval_studio_client/api/test/test_v1_evaluation_type.py +33 -0
  252. eval_studio_client/api/test/test_v1_finalize_operation_response.py +2 -1
  253. eval_studio_client/api/test/test_v1_flipped_metric.py +53 -0
  254. eval_studio_client/api/test/test_v1_generate_test_cases_response.py +2 -1
  255. eval_studio_client/api/test/test_v1_get_dashboard_report_response.py +143 -0
  256. eval_studio_client/api/test/test_v1_get_info_response.py +4 -1
  257. eval_studio_client/api/test/test_v1_get_leaderboard_report_response.py +39 -2
  258. eval_studio_client/api/test/test_v1_get_leaderboard_response.py +2 -1
  259. eval_studio_client/api/test/test_v1_get_operation_response.py +2 -1
  260. eval_studio_client/api/test/test_v1_get_stats_response.py +3 -1
  261. eval_studio_client/api/test/test_v1_human_decision.py +33 -0
  262. eval_studio_client/api/test/test_v1_import_leaderboard_response.py +2 -1
  263. eval_studio_client/api/test/test_v1_import_test_cases_from_library_response.py +2 -1
  264. eval_studio_client/api/test/test_v1_info.py +4 -1
  265. eval_studio_client/api/test/test_v1_leaderboard.py +2 -1
  266. eval_studio_client/api/test/test_v1_leaderboard_cmp_report.py +254 -0
  267. eval_studio_client/api/test/test_v1_leaderboard_comparison_item.py +53 -0
  268. eval_studio_client/api/test/test_v1_leaderboard_info.py +57 -0
  269. eval_studio_client/api/test/test_v1_leaderboard_report.py +39 -2
  270. eval_studio_client/api/test/test_v1_leaderboard_report_actual_output_meta.py +33 -1
  271. eval_studio_client/api/test/test_v1_leaderboard_report_result.py +39 -2
  272. eval_studio_client/api/test/test_v1_leaderboard_report_result_view.py +33 -0
  273. eval_studio_client/api/test/test_v1_leaderboard_test_case_annotation.py +58 -0
  274. eval_studio_client/api/test/test_v1_list_dashboard_test_case_annotations_response.py +61 -0
  275. eval_studio_client/api/test/test_v1_list_leaderboard_test_case_annotations_response.py +61 -0
  276. eval_studio_client/api/test/test_v1_list_leaderboards_response.py +2 -1
  277. eval_studio_client/api/test/test_v1_list_most_recent_leaderboards_response.py +2 -1
  278. eval_studio_client/api/test/test_v1_list_operations_response.py +4 -2
  279. eval_studio_client/api/test/test_v1_list_unseen_operations_response.py +75 -0
  280. eval_studio_client/api/test/test_v1_mark_operation_seen_by_creator_response.py +72 -0
  281. eval_studio_client/api/test/test_v1_metric.py +52 -0
  282. eval_studio_client/api/test/test_v1_metric_average.py +58 -0
  283. eval_studio_client/api/test/test_v1_metric_meta.py +66 -0
  284. eval_studio_client/api/test/test_v1_models_comparisons.py +54 -0
  285. eval_studio_client/api/test/test_v1_models_comparisons_metrics.py +65 -0
  286. eval_studio_client/api/test/test_v1_models_overview.py +60 -0
  287. eval_studio_client/api/test/test_v1_operation.py +2 -1
  288. eval_studio_client/api/test/test_v1_operation_view.py +33 -0
  289. eval_studio_client/api/test/test_v1_process_workflow_node_response.py +2 -1
  290. eval_studio_client/api/test/test_v1_retrieved_context_diff.py +66 -0
  291. eval_studio_client/api/test/test_v1_stats.py +3 -1
  292. eval_studio_client/api/test/test_v1_technical_metrics.py +62 -0
  293. eval_studio_client/api/test/test_v1_technical_metrics_detail.py +55 -0
  294. eval_studio_client/api/test/test_v1_test_case_leaderboard_item.py +53 -0
  295. eval_studio_client/api/test/test_v1_test_case_relationship_info.py +53 -0
  296. eval_studio_client/api/test/test_v1_test_case_result.py +106 -0
  297. eval_studio_client/api/test/test_v1_text_similarity_metric.py +33 -0
  298. eval_studio_client/api/test/test_v1_update_dashboard_test_case_annotation_response.py +59 -0
  299. eval_studio_client/api/test/test_v1_update_leaderboard_response.py +2 -1
  300. eval_studio_client/api/test/test_v1_update_leaderboard_test_case_annotation_response.py +59 -0
  301. eval_studio_client/api/test/test_v1_update_operation_response.py +2 -1
  302. eval_studio_client/gen/openapiv2/eval_studio.swagger.json +2340 -210
  303. eval_studio_client/models.py +18 -6
  304. {eval_studio_client-1.2.5.dist-info → eval_studio_client-1.3.0a1.dist-info}/METADATA +2 -2
  305. {eval_studio_client-1.2.5.dist-info → eval_studio_client-1.3.0a1.dist-info}/RECORD +306 -111
  306. {eval_studio_client-1.2.5.dist-info → eval_studio_client-1.3.0a1.dist-info}/WHEEL +0 -0
@@ -0,0 +1,226 @@
1
+ # coding: utf-8
2
+
3
+ """
4
+ ai/h2o/eval_studio/v1/insight.proto
5
+
6
+ No description provided (generated by Openapi Generator https://github.com/openapitools/openapi-generator)
7
+
8
+ The version of the OpenAPI document: version not set
9
+ Generated by OpenAPI Generator (https://openapi-generator.tech)
10
+
11
+ Do not edit the class manually.
12
+ """ # noqa: E501
13
+
14
+
15
+ import unittest
16
+
17
+ from eval_studio_client.api.models.v1_diff_item import V1DiffItem
18
+
19
+ class TestV1DiffItem(unittest.TestCase):
20
+ """V1DiffItem unit test stubs"""
21
+
22
+ def setUp(self):
23
+ pass
24
+
25
+ def tearDown(self):
26
+ pass
27
+
28
+ def make_instance(self, include_optional) -> V1DiffItem:
29
+ """Test V1DiffItem
30
+ include_option is a boolean, when False only required
31
+ params are included, when True both required and
32
+ optional params are included """
33
+ # uncomment below to create an instance of `V1DiffItem`
34
+ """
35
+ model = V1DiffItem()
36
+ if include_optional:
37
+ return V1DiffItem(
38
+ diff_key = '',
39
+ items = [
40
+ eval_studio_client.api.models.a_single_comparison_item_showing_differences_between_baseline_and_current.A single comparison item showing differences between baseline and current(
41
+ question = '',
42
+ diff_flipped_metrics = [
43
+ eval_studio_client.api.models.flipped_metric_information.Flipped metric information(
44
+ metric_name = '',
45
+ baseline_value = 1.337,
46
+ current_value = 1.337, )
47
+ ],
48
+ baseline_test_case_result = eval_studio_client.api.models.test_case_result.Test case result(
49
+ key = '',
50
+ input = '',
51
+ corpus = [
52
+ ''
53
+ ],
54
+ context = [
55
+ ''
56
+ ],
57
+ categories = [
58
+ ''
59
+ ],
60
+ relationships = [
61
+ eval_studio_client.api.models.test_case_relationship_information.Test case relationship information(
62
+ type = '',
63
+ target = '',
64
+ target_type = '', )
65
+ ],
66
+ expected_output = '',
67
+ output_constraints = [
68
+ ''
69
+ ],
70
+ output_condition = '',
71
+ actual_output = '',
72
+ actual_duration = 1.337,
73
+ cost = 1.337,
74
+ model_key = '',
75
+ test_key = '',
76
+ test_case_key = '',
77
+ metrics = [
78
+ eval_studio_client.api.models.metric_information.Metric information(
79
+ key = '',
80
+ value = 1.337, )
81
+ ],
82
+ metrics_meta = {
83
+ 'key' : ''
84
+ },
85
+ actual_output_meta = [
86
+ eval_studio_client.api.models.actual_output_metadata.Actual output metadata(
87
+ tokenization = '',
88
+ data = [
89
+ eval_studio_client.api.models.data_fragment.Data fragment(
90
+ text = '',
91
+ meta = {
92
+ 'key' : ''
93
+ }, )
94
+ ], )
95
+ ],
96
+ metric_scores = [
97
+ eval_studio_client.api.models.comparison_metric_score_information_(specific_to_comparison_reports).Comparison metric score information (specific to comparison reports)(
98
+ metric_name = '',
99
+ metric_score = 1.337, )
100
+ ],
101
+ result_error_message = '', ),
102
+ baseline_diff_actual_output_meta = eval_studio_client.api.models.actual_output_metadata_diff.Actual output metadata diff(
103
+ sentences = [
104
+ ''
105
+ ],
106
+ sentences_count = 56,
107
+ common_sentences = [
108
+ ''
109
+ ],
110
+ common_count = 56,
111
+ unique_sentences = [
112
+ ''
113
+ ],
114
+ unique_count = 56,
115
+ identical = True,
116
+ sentence_similarity = {
117
+ 'key' : 1.337
118
+ }, ),
119
+ baseline_diff_retrieved_context = eval_studio_client.api.models.retrieved_context_diff.Retrieved context diff(
120
+ chunks = [
121
+ ''
122
+ ],
123
+ chunks_count = 56,
124
+ common_chunks = [
125
+ ''
126
+ ],
127
+ common_count = 56,
128
+ unique_chunks = [
129
+ ''
130
+ ],
131
+ unique_count = 56,
132
+ identical = True,
133
+ chunk_similarity = {
134
+ 'key' : 1.337
135
+ }, ),
136
+ current_test_case_result = eval_studio_client.api.models.test_case_result.Test case result(
137
+ key = '',
138
+ input = '',
139
+ expected_output = '',
140
+ output_condition = '',
141
+ actual_output = '',
142
+ actual_duration = 1.337,
143
+ cost = 1.337,
144
+ model_key = '',
145
+ test_key = '',
146
+ test_case_key = '',
147
+ result_error_message = '', ),
148
+ current_diff_actual_output_meta = eval_studio_client.api.models.actual_output_metadata_diff.Actual output metadata diff(
149
+ sentences_count = 56,
150
+ common_count = 56,
151
+ unique_count = 56,
152
+ identical = True, ),
153
+ current_diff_retrieved_context = eval_studio_client.api.models.retrieved_context_diff.Retrieved context diff(
154
+ chunks_count = 56,
155
+ common_count = 56,
156
+ unique_count = 56,
157
+ identical = True, ), )
158
+ ],
159
+ summary = eval_studio_client.api.models.comparison_summary.Comparison summary(
160
+ recommendation_winner = '',
161
+ recommendation = '',
162
+ recommendation_confidence = '', ),
163
+ models_overview = eval_studio_client.api.models.models_overview.Models overview(
164
+ baseline_model_key = '',
165
+ current_model_key = '',
166
+ baseline_model_name = '',
167
+ baseline_collection_id = [
168
+ ''
169
+ ],
170
+ current_model_name = '',
171
+ current_collection_id = [
172
+ ''
173
+ ], ),
174
+ models_comparisons = eval_studio_client.api.models.models_comparison_statistics.Models comparison statistics(
175
+ test_case_ranks_baseline = 56,
176
+ test_case_ranks_current = 56,
177
+ test_case_wins_baseline = 56,
178
+ test_case_wins_current = 56, ),
179
+ models_comparisons_metrics = eval_studio_client.api.models.detailed_metrics_comparisons.Detailed metrics comparisons(
180
+ metrics_ranks_baseline = 1.337,
181
+ metrics_ranks_current = 1.337,
182
+ metrics_wins_baseline = 56,
183
+ metrics_wins_current = 56,
184
+ metrics_averages = [
185
+ eval_studio_client.api.models.metric_average_comparison.Metric average comparison(
186
+ metric_key = '',
187
+ baseline_avg = 1.337,
188
+ current_avg = 1.337,
189
+ diff = 1.337,
190
+ baseline_better_wins = 56,
191
+ current_better_wins = 56,
192
+ baseline_rank_avg = 1.337,
193
+ current_rank_avg = 1.337, )
194
+ ], ),
195
+ technical_metrics = eval_studio_client.api.models.technical_metrics_for_model_performance.Technical metrics for model performance(
196
+ baseline = eval_studio_client.api.models.technical_metrics_detail.Technical metrics detail(
197
+ cost_sum = 1.337,
198
+ duration_sum = 1.337,
199
+ duration_min = 1.337,
200
+ duration_max = 1.337,
201
+ duration_avg = 1.337, ),
202
+ current = eval_studio_client.api.models.technical_metrics_detail.Technical metrics detail(
203
+ cost_sum = 1.337,
204
+ duration_sum = 1.337,
205
+ duration_min = 1.337,
206
+ duration_max = 1.337,
207
+ duration_avg = 1.337, ), ),
208
+ test_cases_leaderboard = [
209
+ eval_studio_client.api.models.test_case_leaderboard_item.Test case leaderboard item(
210
+ wins = 56,
211
+ question = '',
212
+ changed_metrics_count = 56, )
213
+ ]
214
+ )
215
+ else:
216
+ return V1DiffItem(
217
+ )
218
+ """
219
+
220
+ def testV1DiffItem(self):
221
+ """Test V1DiffItem"""
222
+ # inst_req_only = self.make_instance(include_optional=False)
223
+ # inst_req_and_optional = self.make_instance(include_optional=True)
224
+
225
+ if __name__ == '__main__':
226
+ unittest.main()
@@ -0,0 +1,33 @@
1
+ # coding: utf-8
2
+
3
+ """
4
+ ai/h2o/eval_studio/v1/insight.proto
5
+
6
+ No description provided (generated by Openapi Generator https://github.com/openapitools/openapi-generator)
7
+
8
+ The version of the OpenAPI document: version not set
9
+ Generated by OpenAPI Generator (https://openapi-generator.tech)
10
+
11
+ Do not edit the class manually.
12
+ """ # noqa: E501
13
+
14
+
15
+ import unittest
16
+
17
+ from eval_studio_client.api.models.v1_evaluation_type import V1EvaluationType
18
+
19
+ class TestV1EvaluationType(unittest.TestCase):
20
+ """V1EvaluationType unit test stubs"""
21
+
22
+ def setUp(self):
23
+ pass
24
+
25
+ def tearDown(self):
26
+ pass
27
+
28
+ def testV1EvaluationType(self):
29
+ """Test V1EvaluationType"""
30
+ # inst = V1EvaluationType()
31
+
32
+ if __name__ == '__main__':
33
+ unittest.main()
@@ -55,7 +55,8 @@ class TestV1FinalizeOperationResponse(unittest.TestCase):
55
55
  'key' : None
56
56
  }
57
57
  ], ),
58
- response = , )
58
+ response = ,
59
+ seen_by_creator_time = datetime.datetime.strptime('2013-10-20 19:20:30.00', '%Y-%m-%d %H:%M:%S.%f'), )
59
60
  )
60
61
  else:
61
62
  return V1FinalizeOperationResponse(
@@ -0,0 +1,53 @@
1
+ # coding: utf-8
2
+
3
+ """
4
+ ai/h2o/eval_studio/v1/insight.proto
5
+
6
+ No description provided (generated by Openapi Generator https://github.com/openapitools/openapi-generator)
7
+
8
+ The version of the OpenAPI document: version not set
9
+ Generated by OpenAPI Generator (https://openapi-generator.tech)
10
+
11
+ Do not edit the class manually.
12
+ """ # noqa: E501
13
+
14
+
15
+ import unittest
16
+
17
+ from eval_studio_client.api.models.v1_flipped_metric import V1FlippedMetric
18
+
19
+ class TestV1FlippedMetric(unittest.TestCase):
20
+ """V1FlippedMetric unit test stubs"""
21
+
22
+ def setUp(self):
23
+ pass
24
+
25
+ def tearDown(self):
26
+ pass
27
+
28
+ def make_instance(self, include_optional) -> V1FlippedMetric:
29
+ """Test V1FlippedMetric
30
+ include_option is a boolean, when False only required
31
+ params are included, when True both required and
32
+ optional params are included """
33
+ # uncomment below to create an instance of `V1FlippedMetric`
34
+ """
35
+ model = V1FlippedMetric()
36
+ if include_optional:
37
+ return V1FlippedMetric(
38
+ metric_name = '',
39
+ baseline_value = 1.337,
40
+ current_value = 1.337
41
+ )
42
+ else:
43
+ return V1FlippedMetric(
44
+ )
45
+ """
46
+
47
+ def testV1FlippedMetric(self):
48
+ """Test V1FlippedMetric"""
49
+ # inst_req_only = self.make_instance(include_optional=False)
50
+ # inst_req_and_optional = self.make_instance(include_optional=True)
51
+
52
+ if __name__ == '__main__':
53
+ unittest.main()
@@ -55,7 +55,8 @@ class TestV1GenerateTestCasesResponse(unittest.TestCase):
55
55
  'key' : None
56
56
  }
57
57
  ], ),
58
- response = , )
58
+ response = ,
59
+ seen_by_creator_time = datetime.datetime.strptime('2013-10-20 19:20:30.00', '%Y-%m-%d %H:%M:%S.%f'), )
59
60
  )
60
61
  else:
61
62
  return V1GenerateTestCasesResponse(
@@ -0,0 +1,143 @@
1
+ # coding: utf-8
2
+
3
+ """
4
+ ai/h2o/eval_studio/v1/insight.proto
5
+
6
+ No description provided (generated by Openapi Generator https://github.com/openapitools/openapi-generator)
7
+
8
+ The version of the OpenAPI document: version not set
9
+ Generated by OpenAPI Generator (https://openapi-generator.tech)
10
+
11
+ Do not edit the class manually.
12
+ """ # noqa: E501
13
+
14
+
15
+ import unittest
16
+
17
+ from eval_studio_client.api.models.v1_get_dashboard_report_response import V1GetDashboardReportResponse
18
+
19
+ class TestV1GetDashboardReportResponse(unittest.TestCase):
20
+ """V1GetDashboardReportResponse unit test stubs"""
21
+
22
+ def setUp(self):
23
+ pass
24
+
25
+ def tearDown(self):
26
+ pass
27
+
28
+ def make_instance(self, include_optional) -> V1GetDashboardReportResponse:
29
+ """Test V1GetDashboardReportResponse
30
+ include_option is a boolean, when False only required
31
+ params are included, when True both required and
32
+ optional params are included """
33
+ # uncomment below to create an instance of `V1GetDashboardReportResponse`
34
+ """
35
+ model = V1GetDashboardReportResponse()
36
+ if include_optional:
37
+ return V1GetDashboardReportResponse(
38
+ dashboard_report = eval_studio_client.api.models.v1_dashboard_report.v1DashboardReport(
39
+ results = [
40
+ eval_studio_client.api.models.v1_dashboard_report_result.v1DashboardReportResult(
41
+ key = '',
42
+ input = '',
43
+ expected_output = '',
44
+ actual_output = '',
45
+ model_key = '',
46
+ test_case_key = '',
47
+ metrics = {
48
+ 'key' : eval_studio_client.api.models.v1_metric_scores.v1MetricScores(
49
+ scores = [
50
+ eval_studio_client.api.models.v1_metric_score.v1MetricScore(
51
+ key = '',
52
+ value = 1.337, )
53
+ ], )
54
+ },
55
+ result_error_map = {
56
+ 'key' : ''
57
+ },
58
+ human_decision = 'HUMAN_DECISION_UNSPECIFIED',
59
+ comment = '',
60
+ annotations = {
61
+ 'key' : None
62
+ }, )
63
+ ],
64
+ evaluator = [
65
+ eval_studio_client.api.models.v1_leaderboard_report_evaluator.v1LeaderboardReportEvaluator(
66
+ id = '',
67
+ name = '',
68
+ display_name = '',
69
+ tagline = '',
70
+ description = '',
71
+ brief_description = '',
72
+ model_types = [
73
+ ''
74
+ ],
75
+ can_explain = [
76
+ ''
77
+ ],
78
+ explanation_scopes = [
79
+ ''
80
+ ],
81
+ explanations = [
82
+ eval_studio_client.api.models.v1_leaderboard_report_explanation.v1LeaderboardReportExplanation(
83
+ explanation_type = '',
84
+ name = '',
85
+ category = '',
86
+ scope = '',
87
+ has_local = '',
88
+ formats = [
89
+ ''
90
+ ], )
91
+ ],
92
+ parameters = [
93
+ eval_studio_client.api.models.v1_leaderboard_report_evaluator_parameter.v1LeaderboardReportEvaluatorParameter(
94
+ name = '',
95
+ description = '',
96
+ comment = '',
97
+ type = '',
98
+ predefined = [
99
+ None
100
+ ],
101
+ tags = [
102
+ ''
103
+ ],
104
+ min = 1.337,
105
+ max = 1.337,
106
+ category = '', )
107
+ ],
108
+ keywords = [
109
+ ''
110
+ ],
111
+ metrics_meta = [
112
+ eval_studio_client.api.models.v1_leaderboard_report_metrics_meta_entry.v1LeaderboardReportMetricsMetaEntry(
113
+ key = '',
114
+ display_name = '',
115
+ data_type = '',
116
+ display_value = '',
117
+ description = '',
118
+ value_range = [
119
+ 1.337
120
+ ],
121
+ value_enum = [
122
+ ''
123
+ ],
124
+ higher_is_better = True,
125
+ threshold = 1.337,
126
+ is_primary_metric = True,
127
+ parent_metric = '',
128
+ exclude = True, )
129
+ ], )
130
+ ], )
131
+ )
132
+ else:
133
+ return V1GetDashboardReportResponse(
134
+ )
135
+ """
136
+
137
+ def testV1GetDashboardReportResponse(self):
138
+ """Test V1GetDashboardReportResponse"""
139
+ # inst_req_only = self.make_instance(include_optional=False)
140
+ # inst_req_and_optional = self.make_instance(include_optional=True)
141
+
142
+ if __name__ == '__main__':
143
+ unittest.main()
@@ -51,7 +51,10 @@ class TestV1GetInfoResponse(unittest.TestCase):
51
51
  h2o_cloud_url = '',
52
52
  public_instance = True,
53
53
  sharing_enabled = True,
54
- experimental_features_enabled = True, )
54
+ experimental_features_enabled = True,
55
+ model_type_allowlist = [
56
+ ''
57
+ ], )
55
58
  )
56
59
  else:
57
60
  return V1GetInfoResponse(
@@ -77,8 +77,45 @@ class TestV1GetLeaderboardReportResponse(unittest.TestCase):
77
77
  data = [
78
78
  eval_studio_client.api.models.v1_leaderboard_report_actual_output_data.v1LeaderboardReportActualOutputData(
79
79
  text = '', )
80
- ], )
81
- ], )
80
+ ],
81
+ agent_chat_activity_diagram = eval_studio_client.api.models.v1_agent_chat_activity_diagram.v1AgentChatActivityDiagram(
82
+ rows = [
83
+ eval_studio_client.api.models.v1_agent_chat_activity_diagram_row.v1AgentChatActivityDiagramRow(
84
+ nodes = [
85
+ eval_studio_client.api.models.v1_agent_chat_activity_diagram_node.v1AgentChatActivityDiagramNode(
86
+ id = '',
87
+ role = '',
88
+ label = '', )
89
+ ], )
90
+ ],
91
+ edges = [
92
+ eval_studio_client.api.models.v1_agent_chat_activity_diagram_edge.v1AgentChatActivityDiagramEdge(
93
+ from = '',
94
+ to = '',
95
+ label = '', )
96
+ ], ),
97
+ agent_chat_tools_bar_chart = eval_studio_client.api.models.v1_agent_chat_tools_bar_chart.v1AgentChatToolsBarChart(
98
+ tools = {
99
+ 'key' : eval_studio_client.api.models.v1_agent_chat_tool_usage.v1AgentChatToolUsage(
100
+ name = '',
101
+ success_count = 56,
102
+ failure_count = 56,
103
+ total_count = 56, )
104
+ }, ),
105
+ agent_chat_scripts_bar_chart = eval_studio_client.api.models.v1_agent_chat_scripts_bar_chart.v1AgentChatScriptsBarChart(
106
+ scripts = {
107
+ 'key' : eval_studio_client.api.models.v1_agent_chat_script_usage.v1AgentChatScriptUsage(
108
+ name = '',
109
+ success_count = 56,
110
+ failure_count = 56,
111
+ total_count = 56, )
112
+ }, ), )
113
+ ],
114
+ human_decision = 'HUMAN_DECISION_UNSPECIFIED',
115
+ comment = '',
116
+ annotations = {
117
+ 'key' : None
118
+ }, )
82
119
  ],
83
120
  models = [
84
121
  eval_studio_client.api.models.v1_leaderboard_report_model.v1LeaderboardReportModel(
@@ -98,7 +98,8 @@ class TestV1GetLeaderboardResponse(unittest.TestCase):
98
98
  h2ogpte_collection = '',
99
99
  type = 'LEADERBOARD_TYPE_UNSPECIFIED',
100
100
  demo = True,
101
- test_lab = '', )
101
+ test_lab = '',
102
+ evaluation_type = 'EVALUATION_TYPE_UNSPECIFIED', )
102
103
  )
103
104
  else:
104
105
  return V1GetLeaderboardResponse(
@@ -55,7 +55,8 @@ class TestV1GetOperationResponse(unittest.TestCase):
55
55
  'key' : None
56
56
  }
57
57
  ], ),
58
- response = , )
58
+ response = ,
59
+ seen_by_creator_time = datetime.datetime.strptime('2013-10-20 19:20:30.00', '%Y-%m-%d %H:%M:%S.%f'), )
59
60
  )
60
61
  else:
61
62
  return V1GetOperationResponse(
@@ -38,7 +38,9 @@ class TestV1GetStatsResponse(unittest.TestCase):
38
38
  stats = eval_studio_client.api.models.v1_stats.v1Stats(
39
39
  topic_modeling_pending_jobs = '',
40
40
  test_validation_pending_jobs = '',
41
- failure_clustering_pending_jobs = '', )
41
+ failure_clustering_pending_jobs = '',
42
+ test_case_import_pending_jobs = '',
43
+ evaluation_worker_queue_length = '', )
42
44
  )
43
45
  else:
44
46
  return V1GetStatsResponse(
@@ -0,0 +1,33 @@
1
+ # coding: utf-8
2
+
3
+ """
4
+ ai/h2o/eval_studio/v1/insight.proto
5
+
6
+ No description provided (generated by Openapi Generator https://github.com/openapitools/openapi-generator)
7
+
8
+ The version of the OpenAPI document: version not set
9
+ Generated by OpenAPI Generator (https://openapi-generator.tech)
10
+
11
+ Do not edit the class manually.
12
+ """ # noqa: E501
13
+
14
+
15
+ import unittest
16
+
17
+ from eval_studio_client.api.models.v1_human_decision import V1HumanDecision
18
+
19
+ class TestV1HumanDecision(unittest.TestCase):
20
+ """V1HumanDecision unit test stubs"""
21
+
22
+ def setUp(self):
23
+ pass
24
+
25
+ def tearDown(self):
26
+ pass
27
+
28
+ def testV1HumanDecision(self):
29
+ """Test V1HumanDecision"""
30
+ # inst = V1HumanDecision()
31
+
32
+ if __name__ == '__main__':
33
+ unittest.main()
@@ -55,7 +55,8 @@ class TestV1ImportLeaderboardResponse(unittest.TestCase):
55
55
  'key' : None
56
56
  }
57
57
  ], ),
58
- response = , )
58
+ response = ,
59
+ seen_by_creator_time = datetime.datetime.strptime('2013-10-20 19:20:30.00', '%Y-%m-%d %H:%M:%S.%f'), )
59
60
  )
60
61
  else:
61
62
  return V1ImportLeaderboardResponse(
@@ -55,7 +55,8 @@ class TestV1ImportTestCasesFromLibraryResponse(unittest.TestCase):
55
55
  'key' : None
56
56
  }
57
57
  ], ),
58
- response = , )
58
+ response = ,
59
+ seen_by_creator_time = datetime.datetime.strptime('2013-10-20 19:20:30.00', '%Y-%m-%d %H:%M:%S.%f'), )
59
60
  )
60
61
  else:
61
62
  return V1ImportTestCasesFromLibraryResponse(
@@ -50,7 +50,10 @@ class TestV1Info(unittest.TestCase):
50
50
  h2o_cloud_url = '',
51
51
  public_instance = True,
52
52
  sharing_enabled = True,
53
- experimental_features_enabled = True
53
+ experimental_features_enabled = True,
54
+ model_type_allowlist = [
55
+ ''
56
+ ]
54
57
  )
55
58
  else:
56
59
  return V1Info(
@@ -100,7 +100,8 @@ class TestV1Leaderboard(unittest.TestCase):
100
100
  h2ogpte_collection = '',
101
101
  type = 'LEADERBOARD_TYPE_UNSPECIFIED',
102
102
  demo = True,
103
- test_lab = ''
103
+ test_lab = '',
104
+ evaluation_type = 'EVALUATION_TYPE_UNSPECIFIED'
104
105
  )
105
106
  else:
106
107
  return V1Leaderboard(