eval-studio-client 1.2.4a2__py3-none-any.whl → 1.3.0__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (306) hide show
  1. eval_studio_client/api/__init__.py +65 -0
  2. eval_studio_client/api/api/__init__.py +3 -0
  3. eval_studio_client/api/api/dashboard_report_service_api.py +292 -0
  4. eval_studio_client/api/api/dashboard_service_api.py +16 -16
  5. eval_studio_client/api/api/dashboard_test_case_annotation_service_api.py +611 -0
  6. eval_studio_client/api/api/document_service_api.py +16 -16
  7. eval_studio_client/api/api/evaluation_service_api.py +12 -12
  8. eval_studio_client/api/api/evaluator_service_api.py +16 -16
  9. eval_studio_client/api/api/leaderboard_report_service_api.py +304 -17
  10. eval_studio_client/api/api/leaderboard_service_api.py +554 -16
  11. eval_studio_client/api/api/leaderboard_test_case_annotation_service_api.py +611 -0
  12. eval_studio_client/api/api/model_service_api.py +16 -16
  13. eval_studio_client/api/api/operation_service_api.py +821 -17
  14. eval_studio_client/api/api/perturbator_service_api.py +22 -22
  15. eval_studio_client/api/api/test_case_service_api.py +300 -16
  16. eval_studio_client/api/api/test_class_service_api.py +16 -16
  17. eval_studio_client/api/api/test_service_api.py +285 -16
  18. eval_studio_client/api/api/workflow_node_service_api.py +16 -16
  19. eval_studio_client/api/api/workflow_service_api.py +16 -16
  20. eval_studio_client/api/docs/AdversarialInputsServiceTestAdversarialInputsRobustnessRequest.md +2 -1
  21. eval_studio_client/api/docs/DashboardReportServiceApi.md +75 -0
  22. eval_studio_client/api/docs/DashboardServiceApi.md +5 -5
  23. eval_studio_client/api/docs/DashboardTestCaseAnnotationServiceApi.md +149 -0
  24. eval_studio_client/api/docs/DocumentServiceApi.md +5 -5
  25. eval_studio_client/api/docs/EvaluationServiceApi.md +4 -4
  26. eval_studio_client/api/docs/EvaluatorServiceApi.md +5 -5
  27. eval_studio_client/api/docs/LeaderboardReportServiceApi.md +75 -5
  28. eval_studio_client/api/docs/LeaderboardServiceApi.md +141 -5
  29. eval_studio_client/api/docs/LeaderboardTestCaseAnnotationServiceApi.md +149 -0
  30. eval_studio_client/api/docs/ModelServiceApi.md +5 -5
  31. eval_studio_client/api/docs/OperationServiceApi.md +215 -8
  32. eval_studio_client/api/docs/PerturbatorServiceApi.md +7 -7
  33. eval_studio_client/api/docs/RequiredTheDashboardTestCaseAnnotationToUpdate.md +35 -0
  34. eval_studio_client/api/docs/RequiredTheLeaderboardTestCaseAnnotationToUpdate.md +35 -0
  35. eval_studio_client/api/docs/RequiredTheLeaderboardToUpdate.md +1 -0
  36. eval_studio_client/api/docs/RequiredTheOperationToFinalize.md +1 -0
  37. eval_studio_client/api/docs/RequiredTheOperationToUpdate.md +1 -0
  38. eval_studio_client/api/docs/TestCaseServiceApi.md +75 -5
  39. eval_studio_client/api/docs/TestCaseServiceAppendTestCasesRequest.md +30 -0
  40. eval_studio_client/api/docs/TestClassServiceApi.md +5 -5
  41. eval_studio_client/api/docs/TestServiceApi.md +73 -5
  42. eval_studio_client/api/docs/V1ActualOutputMeta.md +30 -0
  43. eval_studio_client/api/docs/V1ActualOutputMetaDiff.md +36 -0
  44. eval_studio_client/api/docs/V1AgentChatActivityDiagram.md +31 -0
  45. eval_studio_client/api/docs/V1AgentChatActivityDiagramEdge.md +32 -0
  46. eval_studio_client/api/docs/V1AgentChatActivityDiagramNode.md +32 -0
  47. eval_studio_client/api/docs/V1AgentChatActivityDiagramRow.md +30 -0
  48. eval_studio_client/api/docs/V1AgentChatScriptUsage.md +33 -0
  49. eval_studio_client/api/docs/V1AgentChatScriptsBarChart.md +30 -0
  50. eval_studio_client/api/docs/V1AgentChatToolUsage.md +33 -0
  51. eval_studio_client/api/docs/V1AgentChatToolsBarChart.md +30 -0
  52. eval_studio_client/api/docs/V1AllMetricScores.md +29 -0
  53. eval_studio_client/api/docs/V1AppendTestCasesResponse.md +29 -0
  54. eval_studio_client/api/docs/V1BatchCreateLeaderboardsWithoutCacheRequest.md +31 -0
  55. eval_studio_client/api/docs/V1BatchCreateLeaderboardsWithoutCacheResponse.md +29 -0
  56. eval_studio_client/api/docs/V1BatchMarkOperationSeenByCreatorResponse.md +29 -0
  57. eval_studio_client/api/docs/V1CmpLeaderboardReportsRequest.md +33 -0
  58. eval_studio_client/api/docs/V1CmpLeaderboardReportsResponse.md +29 -0
  59. eval_studio_client/api/docs/V1ComparisonItem.md +36 -0
  60. eval_studio_client/api/docs/V1ComparisonMetricScore.md +30 -0
  61. eval_studio_client/api/docs/V1ComparisonResult.md +31 -0
  62. eval_studio_client/api/docs/V1ComparisonSummary.md +31 -0
  63. eval_studio_client/api/docs/V1CreateEvaluationRequest.md +1 -0
  64. eval_studio_client/api/docs/V1CreateTestFromTestCasesRequest.md +32 -0
  65. eval_studio_client/api/docs/V1CreateTestFromTestCasesResponse.md +29 -0
  66. eval_studio_client/api/docs/V1DashboardReport.md +31 -0
  67. eval_studio_client/api/docs/V1DashboardReportResult.md +39 -0
  68. eval_studio_client/api/docs/V1DashboardTestCaseAnnotation.md +36 -0
  69. eval_studio_client/api/docs/V1DataFragment.md +31 -0
  70. eval_studio_client/api/docs/V1DeepCompareLeaderboardsRequest.md +33 -0
  71. eval_studio_client/api/docs/V1DeepCompareLeaderboardsResponse.md +29 -0
  72. eval_studio_client/api/docs/V1DiffItem.md +36 -0
  73. eval_studio_client/api/docs/V1EvaluationType.md +12 -0
  74. eval_studio_client/api/docs/V1FlippedMetric.md +31 -0
  75. eval_studio_client/api/docs/V1GetDashboardReportResponse.md +29 -0
  76. eval_studio_client/api/docs/V1HumanDecision.md +12 -0
  77. eval_studio_client/api/docs/V1Info.md +1 -0
  78. eval_studio_client/api/docs/V1Leaderboard.md +1 -0
  79. eval_studio_client/api/docs/V1LeaderboardCmpReport.md +30 -0
  80. eval_studio_client/api/docs/V1LeaderboardComparisonItem.md +31 -0
  81. eval_studio_client/api/docs/V1LeaderboardInfo.md +30 -0
  82. eval_studio_client/api/docs/V1LeaderboardReportActualOutputMeta.md +6 -3
  83. eval_studio_client/api/docs/V1LeaderboardReportResult.md +11 -8
  84. eval_studio_client/api/docs/V1LeaderboardReportResultView.md +12 -0
  85. eval_studio_client/api/docs/V1LeaderboardTestCaseAnnotation.md +36 -0
  86. eval_studio_client/api/docs/V1ListDashboardTestCaseAnnotationsResponse.md +29 -0
  87. eval_studio_client/api/docs/V1ListLeaderboardTestCaseAnnotationsResponse.md +29 -0
  88. eval_studio_client/api/docs/V1ListOperationsResponse.md +1 -0
  89. eval_studio_client/api/docs/V1ListUnseenOperationsResponse.md +30 -0
  90. eval_studio_client/api/docs/V1MarkOperationSeenByCreatorResponse.md +29 -0
  91. eval_studio_client/api/docs/V1Metric.md +30 -0
  92. eval_studio_client/api/docs/V1MetricAverage.md +36 -0
  93. eval_studio_client/api/docs/V1MetricMeta.md +40 -0
  94. eval_studio_client/api/docs/V1MetricScore.md +1 -1
  95. eval_studio_client/api/docs/V1MetricScores.md +1 -1
  96. eval_studio_client/api/docs/V1ModelType.md +1 -1
  97. eval_studio_client/api/docs/V1ModelsComparisons.md +32 -0
  98. eval_studio_client/api/docs/V1ModelsComparisonsMetrics.md +33 -0
  99. eval_studio_client/api/docs/V1ModelsOverview.md +34 -0
  100. eval_studio_client/api/docs/V1Operation.md +1 -0
  101. eval_studio_client/api/docs/V1OperationView.md +12 -0
  102. eval_studio_client/api/docs/V1RetrievedContextDiff.md +36 -0
  103. eval_studio_client/api/docs/V1Stats.md +2 -0
  104. eval_studio_client/api/docs/V1TechnicalMetrics.md +30 -0
  105. eval_studio_client/api/docs/V1TechnicalMetricsDetail.md +33 -0
  106. eval_studio_client/api/docs/V1TestCaseLeaderboardItem.md +31 -0
  107. eval_studio_client/api/docs/V1TestCaseRelationshipInfo.md +31 -0
  108. eval_studio_client/api/docs/V1TestCaseResult.md +48 -0
  109. eval_studio_client/api/docs/V1TextSimilarityMetric.md +12 -0
  110. eval_studio_client/api/docs/V1UpdateDashboardTestCaseAnnotationResponse.md +29 -0
  111. eval_studio_client/api/docs/V1UpdateLeaderboardTestCaseAnnotationResponse.md +29 -0
  112. eval_studio_client/api/docs/WorkflowNodeServiceApi.md +5 -5
  113. eval_studio_client/api/docs/WorkflowServiceApi.md +5 -5
  114. eval_studio_client/api/models/__init__.py +62 -0
  115. eval_studio_client/api/models/adversarial_inputs_service_test_adversarial_inputs_robustness_request.py +17 -2
  116. eval_studio_client/api/models/required_the_dashboard_test_case_annotation_to_update.py +108 -0
  117. eval_studio_client/api/models/required_the_leaderboard_test_case_annotation_to_update.py +108 -0
  118. eval_studio_client/api/models/required_the_leaderboard_to_update.py +5 -2
  119. eval_studio_client/api/models/required_the_operation_to_finalize.py +6 -2
  120. eval_studio_client/api/models/required_the_operation_to_update.py +6 -2
  121. eval_studio_client/api/models/test_case_service_append_test_cases_request.py +89 -0
  122. eval_studio_client/api/models/v1_actual_output_meta.py +97 -0
  123. eval_studio_client/api/models/v1_actual_output_meta_diff.py +101 -0
  124. eval_studio_client/api/models/v1_agent_chat_activity_diagram.py +109 -0
  125. eval_studio_client/api/models/v1_agent_chat_activity_diagram_edge.py +97 -0
  126. eval_studio_client/api/models/v1_agent_chat_activity_diagram_node.py +97 -0
  127. eval_studio_client/api/models/v1_agent_chat_activity_diagram_row.py +97 -0
  128. eval_studio_client/api/models/v1_agent_chat_script_usage.py +101 -0
  129. eval_studio_client/api/models/v1_agent_chat_scripts_bar_chart.py +102 -0
  130. eval_studio_client/api/models/v1_agent_chat_tool_usage.py +101 -0
  131. eval_studio_client/api/models/v1_agent_chat_tools_bar_chart.py +102 -0
  132. eval_studio_client/api/models/v1_all_metric_scores.py +87 -0
  133. eval_studio_client/api/models/v1_append_test_cases_response.py +95 -0
  134. eval_studio_client/api/models/v1_batch_create_leaderboards_without_cache_request.py +99 -0
  135. eval_studio_client/api/models/v1_batch_create_leaderboards_without_cache_response.py +91 -0
  136. eval_studio_client/api/models/v1_batch_mark_operation_seen_by_creator_response.py +95 -0
  137. eval_studio_client/api/models/v1_cmp_leaderboard_reports_request.py +96 -0
  138. eval_studio_client/api/models/v1_cmp_leaderboard_reports_response.py +91 -0
  139. eval_studio_client/api/models/v1_comparison_item.py +130 -0
  140. eval_studio_client/api/models/v1_comparison_metric_score.py +89 -0
  141. eval_studio_client/api/models/v1_comparison_result.py +120 -0
  142. eval_studio_client/api/models/v1_comparison_summary.py +91 -0
  143. eval_studio_client/api/models/v1_create_evaluation_request.py +5 -2
  144. eval_studio_client/api/models/v1_create_test_from_test_cases_request.py +93 -0
  145. eval_studio_client/api/models/v1_create_test_from_test_cases_response.py +91 -0
  146. eval_studio_client/api/models/v1_dashboard_report.py +109 -0
  147. eval_studio_client/api/models/v1_dashboard_report_result.py +139 -0
  148. eval_studio_client/api/models/v1_dashboard_test_case_annotation.py +112 -0
  149. eval_studio_client/api/models/v1_data_fragment.py +91 -0
  150. eval_studio_client/api/models/v1_deep_compare_leaderboards_request.py +96 -0
  151. eval_studio_client/api/models/v1_deep_compare_leaderboards_response.py +91 -0
  152. eval_studio_client/api/models/v1_diff_item.py +137 -0
  153. eval_studio_client/api/models/v1_evaluation_type.py +39 -0
  154. eval_studio_client/api/models/v1_flipped_metric.py +91 -0
  155. eval_studio_client/api/models/v1_get_dashboard_report_response.py +91 -0
  156. eval_studio_client/api/models/v1_human_decision.py +38 -0
  157. eval_studio_client/api/models/v1_info.py +4 -2
  158. eval_studio_client/api/models/v1_leaderboard.py +5 -2
  159. eval_studio_client/api/models/v1_leaderboard_cmp_report.py +93 -0
  160. eval_studio_client/api/models/v1_leaderboard_comparison_item.py +91 -0
  161. eval_studio_client/api/models/v1_leaderboard_info.py +97 -0
  162. eval_studio_client/api/models/v1_leaderboard_report_actual_output_meta.py +23 -9
  163. eval_studio_client/api/models/v1_leaderboard_report_result.py +21 -10
  164. eval_studio_client/api/models/v1_leaderboard_report_result_view.py +38 -0
  165. eval_studio_client/api/models/v1_leaderboard_test_case_annotation.py +112 -0
  166. eval_studio_client/api/models/v1_list_dashboard_test_case_annotations_response.py +95 -0
  167. eval_studio_client/api/models/v1_list_leaderboard_test_case_annotations_response.py +95 -0
  168. eval_studio_client/api/models/v1_list_operations_response.py +5 -3
  169. eval_studio_client/api/models/v1_list_unseen_operations_response.py +97 -0
  170. eval_studio_client/api/models/v1_mark_operation_seen_by_creator_response.py +91 -0
  171. eval_studio_client/api/models/v1_metric.py +89 -0
  172. eval_studio_client/api/models/v1_metric_average.py +101 -0
  173. eval_studio_client/api/models/v1_metric_meta.py +109 -0
  174. eval_studio_client/api/models/v1_metric_score.py +6 -1
  175. eval_studio_client/api/models/v1_metric_scores.py +1 -1
  176. eval_studio_client/api/models/v1_model_type.py +2 -1
  177. eval_studio_client/api/models/v1_models_comparisons.py +93 -0
  178. eval_studio_client/api/models/v1_models_comparisons_metrics.py +103 -0
  179. eval_studio_client/api/models/v1_models_overview.py +97 -0
  180. eval_studio_client/api/models/v1_operation.py +6 -2
  181. eval_studio_client/api/models/v1_operation_view.py +38 -0
  182. eval_studio_client/api/models/v1_retrieved_context_diff.py +101 -0
  183. eval_studio_client/api/models/v1_stats.py +16 -2
  184. eval_studio_client/api/models/v1_technical_metrics.py +96 -0
  185. eval_studio_client/api/models/v1_technical_metrics_detail.py +95 -0
  186. eval_studio_client/api/models/v1_test_case_leaderboard_item.py +91 -0
  187. eval_studio_client/api/models/v1_test_case_relationship_info.py +91 -0
  188. eval_studio_client/api/models/v1_test_case_result.py +157 -0
  189. eval_studio_client/api/models/v1_text_similarity_metric.py +39 -0
  190. eval_studio_client/api/models/v1_update_dashboard_test_case_annotation_response.py +91 -0
  191. eval_studio_client/api/models/v1_update_leaderboard_test_case_annotation_response.py +91 -0
  192. eval_studio_client/api/models/v1_workflow_node_type.py +1 -0
  193. eval_studio_client/api/models/v1_workflow_type.py +1 -0
  194. eval_studio_client/api/test/test_adversarial_inputs_service_test_adversarial_inputs_robustness_request.py +6 -0
  195. eval_studio_client/api/test/test_dashboard_report_service_api.py +37 -0
  196. eval_studio_client/api/test/test_dashboard_test_case_annotation_service_api.py +43 -0
  197. eval_studio_client/api/test/test_leaderboard_report_service_api.py +6 -0
  198. eval_studio_client/api/test/test_leaderboard_service_api.py +12 -0
  199. eval_studio_client/api/test/test_leaderboard_test_case_annotation_service_api.py +43 -0
  200. eval_studio_client/api/test/test_operation_service_api.py +18 -0
  201. eval_studio_client/api/test/test_required_the_dashboard_test_case_annotation_to_update.py +57 -0
  202. eval_studio_client/api/test/test_required_the_leaderboard_test_case_annotation_to_update.py +57 -0
  203. eval_studio_client/api/test/test_required_the_leaderboard_to_update.py +2 -1
  204. eval_studio_client/api/test/test_required_the_operation_to_finalize.py +2 -1
  205. eval_studio_client/api/test/test_required_the_operation_to_update.py +2 -1
  206. eval_studio_client/api/test/test_test_case_service_api.py +6 -0
  207. eval_studio_client/api/test/test_test_case_service_append_test_cases_request.py +52 -0
  208. eval_studio_client/api/test/test_test_service_api.py +6 -0
  209. eval_studio_client/api/test/test_v1_abort_operation_response.py +2 -1
  210. eval_studio_client/api/test/test_v1_actual_output_meta.py +61 -0
  211. eval_studio_client/api/test/test_v1_actual_output_meta_diff.py +66 -0
  212. eval_studio_client/api/test/test_v1_agent_chat_activity_diagram.py +65 -0
  213. eval_studio_client/api/test/test_v1_agent_chat_activity_diagram_edge.py +53 -0
  214. eval_studio_client/api/test/test_v1_agent_chat_activity_diagram_node.py +53 -0
  215. eval_studio_client/api/test/test_v1_agent_chat_activity_diagram_row.py +56 -0
  216. eval_studio_client/api/test/test_v1_agent_chat_script_usage.py +54 -0
  217. eval_studio_client/api/test/test_v1_agent_chat_scripts_bar_chart.py +57 -0
  218. eval_studio_client/api/test/test_v1_agent_chat_tool_usage.py +54 -0
  219. eval_studio_client/api/test/test_v1_agent_chat_tools_bar_chart.py +57 -0
  220. eval_studio_client/api/test/test_v1_all_metric_scores.py +53 -0
  221. eval_studio_client/api/test/test_v1_append_test_cases_response.py +74 -0
  222. eval_studio_client/api/test/test_v1_batch_create_leaderboards_request.py +2 -1
  223. eval_studio_client/api/test/test_v1_batch_create_leaderboards_response.py +2 -1
  224. eval_studio_client/api/test/test_v1_batch_create_leaderboards_without_cache_request.py +120 -0
  225. eval_studio_client/api/test/test_v1_batch_create_leaderboards_without_cache_response.py +72 -0
  226. eval_studio_client/api/test/test_v1_batch_delete_leaderboards_response.py +2 -1
  227. eval_studio_client/api/test/test_v1_batch_get_leaderboards_response.py +2 -1
  228. eval_studio_client/api/test/test_v1_batch_get_operations_response.py +2 -1
  229. eval_studio_client/api/test/test_v1_batch_import_leaderboard_response.py +2 -1
  230. eval_studio_client/api/test/test_v1_batch_mark_operation_seen_by_creator_response.py +74 -0
  231. eval_studio_client/api/test/test_v1_cmp_leaderboard_reports_request.py +55 -0
  232. eval_studio_client/api/test/test_v1_cmp_leaderboard_reports_response.py +255 -0
  233. eval_studio_client/api/test/test_v1_comparison_item.py +233 -0
  234. eval_studio_client/api/test/test_v1_comparison_metric_score.py +52 -0
  235. eval_studio_client/api/test/test_v1_comparison_result.py +258 -0
  236. eval_studio_client/api/test/test_v1_comparison_summary.py +53 -0
  237. eval_studio_client/api/test/test_v1_create_evaluation_request.py +2 -1
  238. eval_studio_client/api/test/test_v1_create_leaderboard_request.py +2 -1
  239. eval_studio_client/api/test/test_v1_create_leaderboard_response.py +2 -1
  240. eval_studio_client/api/test/test_v1_create_leaderboard_without_cache_response.py +2 -1
  241. eval_studio_client/api/test/test_v1_create_test_from_test_cases_request.py +54 -0
  242. eval_studio_client/api/test/test_v1_create_test_from_test_cases_response.py +68 -0
  243. eval_studio_client/api/test/test_v1_dashboard_report.py +142 -0
  244. eval_studio_client/api/test/test_v1_dashboard_report_result.py +72 -0
  245. eval_studio_client/api/test/test_v1_dashboard_test_case_annotation.py +58 -0
  246. eval_studio_client/api/test/test_v1_data_fragment.py +57 -0
  247. eval_studio_client/api/test/test_v1_deep_compare_leaderboards_request.py +55 -0
  248. eval_studio_client/api/test/test_v1_deep_compare_leaderboards_response.py +255 -0
  249. eval_studio_client/api/test/test_v1_delete_leaderboard_response.py +2 -1
  250. eval_studio_client/api/test/test_v1_diff_item.py +226 -0
  251. eval_studio_client/api/test/test_v1_evaluation_type.py +33 -0
  252. eval_studio_client/api/test/test_v1_finalize_operation_response.py +2 -1
  253. eval_studio_client/api/test/test_v1_flipped_metric.py +53 -0
  254. eval_studio_client/api/test/test_v1_generate_test_cases_response.py +2 -1
  255. eval_studio_client/api/test/test_v1_get_dashboard_report_response.py +143 -0
  256. eval_studio_client/api/test/test_v1_get_info_response.py +4 -1
  257. eval_studio_client/api/test/test_v1_get_leaderboard_report_response.py +39 -2
  258. eval_studio_client/api/test/test_v1_get_leaderboard_response.py +2 -1
  259. eval_studio_client/api/test/test_v1_get_operation_response.py +2 -1
  260. eval_studio_client/api/test/test_v1_get_stats_response.py +3 -1
  261. eval_studio_client/api/test/test_v1_human_decision.py +33 -0
  262. eval_studio_client/api/test/test_v1_import_leaderboard_response.py +2 -1
  263. eval_studio_client/api/test/test_v1_import_test_cases_from_library_response.py +2 -1
  264. eval_studio_client/api/test/test_v1_info.py +4 -1
  265. eval_studio_client/api/test/test_v1_leaderboard.py +2 -1
  266. eval_studio_client/api/test/test_v1_leaderboard_cmp_report.py +254 -0
  267. eval_studio_client/api/test/test_v1_leaderboard_comparison_item.py +53 -0
  268. eval_studio_client/api/test/test_v1_leaderboard_info.py +57 -0
  269. eval_studio_client/api/test/test_v1_leaderboard_report.py +39 -2
  270. eval_studio_client/api/test/test_v1_leaderboard_report_actual_output_meta.py +33 -1
  271. eval_studio_client/api/test/test_v1_leaderboard_report_result.py +39 -2
  272. eval_studio_client/api/test/test_v1_leaderboard_report_result_view.py +33 -0
  273. eval_studio_client/api/test/test_v1_leaderboard_test_case_annotation.py +58 -0
  274. eval_studio_client/api/test/test_v1_list_dashboard_test_case_annotations_response.py +61 -0
  275. eval_studio_client/api/test/test_v1_list_leaderboard_test_case_annotations_response.py +61 -0
  276. eval_studio_client/api/test/test_v1_list_leaderboards_response.py +2 -1
  277. eval_studio_client/api/test/test_v1_list_most_recent_leaderboards_response.py +2 -1
  278. eval_studio_client/api/test/test_v1_list_operations_response.py +4 -2
  279. eval_studio_client/api/test/test_v1_list_unseen_operations_response.py +75 -0
  280. eval_studio_client/api/test/test_v1_mark_operation_seen_by_creator_response.py +72 -0
  281. eval_studio_client/api/test/test_v1_metric.py +52 -0
  282. eval_studio_client/api/test/test_v1_metric_average.py +58 -0
  283. eval_studio_client/api/test/test_v1_metric_meta.py +66 -0
  284. eval_studio_client/api/test/test_v1_models_comparisons.py +54 -0
  285. eval_studio_client/api/test/test_v1_models_comparisons_metrics.py +65 -0
  286. eval_studio_client/api/test/test_v1_models_overview.py +60 -0
  287. eval_studio_client/api/test/test_v1_operation.py +2 -1
  288. eval_studio_client/api/test/test_v1_operation_view.py +33 -0
  289. eval_studio_client/api/test/test_v1_process_workflow_node_response.py +2 -1
  290. eval_studio_client/api/test/test_v1_retrieved_context_diff.py +66 -0
  291. eval_studio_client/api/test/test_v1_stats.py +3 -1
  292. eval_studio_client/api/test/test_v1_technical_metrics.py +62 -0
  293. eval_studio_client/api/test/test_v1_technical_metrics_detail.py +55 -0
  294. eval_studio_client/api/test/test_v1_test_case_leaderboard_item.py +53 -0
  295. eval_studio_client/api/test/test_v1_test_case_relationship_info.py +53 -0
  296. eval_studio_client/api/test/test_v1_test_case_result.py +106 -0
  297. eval_studio_client/api/test/test_v1_text_similarity_metric.py +33 -0
  298. eval_studio_client/api/test/test_v1_update_dashboard_test_case_annotation_response.py +59 -0
  299. eval_studio_client/api/test/test_v1_update_leaderboard_response.py +2 -1
  300. eval_studio_client/api/test/test_v1_update_leaderboard_test_case_annotation_response.py +59 -0
  301. eval_studio_client/api/test/test_v1_update_operation_response.py +2 -1
  302. eval_studio_client/gen/openapiv2/eval_studio.swagger.json +2340 -210
  303. eval_studio_client/models.py +18 -6
  304. {eval_studio_client-1.2.4a2.dist-info → eval_studio_client-1.3.0.dist-info}/METADATA +2 -2
  305. {eval_studio_client-1.2.4a2.dist-info → eval_studio_client-1.3.0.dist-info}/RECORD +306 -111
  306. {eval_studio_client-1.2.4a2.dist-info → eval_studio_client-1.3.0.dist-info}/WHEEL +0 -0
@@ -0,0 +1,258 @@
1
+ # coding: utf-8
2
+
3
+ """
4
+ ai/h2o/eval_studio/v1/insight.proto
5
+
6
+ No description provided (generated by Openapi Generator https://github.com/openapitools/openapi-generator)
7
+
8
+ The version of the OpenAPI document: version not set
9
+ Generated by OpenAPI Generator (https://openapi-generator.tech)
10
+
11
+ Do not edit the class manually.
12
+ """ # noqa: E501
13
+
14
+
15
+ import unittest
16
+
17
+ from eval_studio_client.api.models.v1_comparison_result import V1ComparisonResult
18
+
19
+ class TestV1ComparisonResult(unittest.TestCase):
20
+ """V1ComparisonResult unit test stubs"""
21
+
22
+ def setUp(self):
23
+ pass
24
+
25
+ def tearDown(self):
26
+ pass
27
+
28
+ def make_instance(self, include_optional) -> V1ComparisonResult:
29
+ """Test V1ComparisonResult
30
+ include_option is a boolean, when False only required
31
+ params are included, when True both required and
32
+ optional params are included """
33
+ # uncomment below to create an instance of `V1ComparisonResult`
34
+ """
35
+ model = V1ComparisonResult()
36
+ if include_optional:
37
+ return V1ComparisonResult(
38
+ diffs = [
39
+ eval_studio_client.api.models.a_single_diff_item_comparing_two_leaderboards.A single diff item comparing two leaderboards(
40
+ diff_key = '',
41
+ items = [
42
+ eval_studio_client.api.models.a_single_comparison_item_showing_differences_between_baseline_and_current.A single comparison item showing differences between baseline and current(
43
+ question = '',
44
+ diff_flipped_metrics = [
45
+ eval_studio_client.api.models.flipped_metric_information.Flipped metric information(
46
+ metric_name = '',
47
+ baseline_value = 1.337,
48
+ current_value = 1.337, )
49
+ ],
50
+ baseline_test_case_result = eval_studio_client.api.models.test_case_result.Test case result(
51
+ key = '',
52
+ input = '',
53
+ corpus = [
54
+ ''
55
+ ],
56
+ context = [
57
+ ''
58
+ ],
59
+ categories = [
60
+ ''
61
+ ],
62
+ relationships = [
63
+ eval_studio_client.api.models.test_case_relationship_information.Test case relationship information(
64
+ type = '',
65
+ target = '',
66
+ target_type = '', )
67
+ ],
68
+ expected_output = '',
69
+ output_constraints = [
70
+ ''
71
+ ],
72
+ output_condition = '',
73
+ actual_output = '',
74
+ actual_duration = 1.337,
75
+ cost = 1.337,
76
+ model_key = '',
77
+ test_key = '',
78
+ test_case_key = '',
79
+ metrics = [
80
+ eval_studio_client.api.models.metric_information.Metric information(
81
+ key = '',
82
+ value = 1.337, )
83
+ ],
84
+ metrics_meta = {
85
+ 'key' : ''
86
+ },
87
+ actual_output_meta = [
88
+ eval_studio_client.api.models.actual_output_metadata.Actual output metadata(
89
+ tokenization = '',
90
+ data = [
91
+ eval_studio_client.api.models.data_fragment.Data fragment(
92
+ text = '',
93
+ meta = {
94
+ 'key' : ''
95
+ }, )
96
+ ], )
97
+ ],
98
+ metric_scores = [
99
+ eval_studio_client.api.models.comparison_metric_score_information_(specific_to_comparison_reports).Comparison metric score information (specific to comparison reports)(
100
+ metric_name = '',
101
+ metric_score = 1.337, )
102
+ ],
103
+ result_error_message = '', ),
104
+ baseline_diff_actual_output_meta = eval_studio_client.api.models.actual_output_metadata_diff.Actual output metadata diff(
105
+ sentences = [
106
+ ''
107
+ ],
108
+ sentences_count = 56,
109
+ common_sentences = [
110
+ ''
111
+ ],
112
+ common_count = 56,
113
+ unique_sentences = [
114
+ ''
115
+ ],
116
+ unique_count = 56,
117
+ identical = True,
118
+ sentence_similarity = {
119
+ 'key' : 1.337
120
+ }, ),
121
+ baseline_diff_retrieved_context = eval_studio_client.api.models.retrieved_context_diff.Retrieved context diff(
122
+ chunks = [
123
+ ''
124
+ ],
125
+ chunks_count = 56,
126
+ common_chunks = [
127
+ ''
128
+ ],
129
+ common_count = 56,
130
+ unique_chunks = [
131
+ ''
132
+ ],
133
+ unique_count = 56,
134
+ identical = True,
135
+ chunk_similarity = {
136
+ 'key' : 1.337
137
+ }, ),
138
+ current_test_case_result = eval_studio_client.api.models.test_case_result.Test case result(
139
+ key = '',
140
+ input = '',
141
+ expected_output = '',
142
+ output_condition = '',
143
+ actual_output = '',
144
+ actual_duration = 1.337,
145
+ cost = 1.337,
146
+ model_key = '',
147
+ test_key = '',
148
+ test_case_key = '',
149
+ result_error_message = '', ),
150
+ current_diff_actual_output_meta = eval_studio_client.api.models.actual_output_metadata_diff.Actual output metadata diff(
151
+ sentences_count = 56,
152
+ common_count = 56,
153
+ unique_count = 56,
154
+ identical = True, ),
155
+ current_diff_retrieved_context = eval_studio_client.api.models.retrieved_context_diff.Retrieved context diff(
156
+ chunks_count = 56,
157
+ common_count = 56,
158
+ unique_count = 56,
159
+ identical = True, ), )
160
+ ],
161
+ summary = eval_studio_client.api.models.comparison_summary.Comparison summary(
162
+ recommendation_winner = '',
163
+ recommendation = '',
164
+ recommendation_confidence = '', ),
165
+ models_overview = eval_studio_client.api.models.models_overview.Models overview(
166
+ baseline_model_key = '',
167
+ current_model_key = '',
168
+ baseline_model_name = '',
169
+ baseline_collection_id = [
170
+ ''
171
+ ],
172
+ current_model_name = '',
173
+ current_collection_id = [
174
+ ''
175
+ ], ),
176
+ models_comparisons = eval_studio_client.api.models.models_comparison_statistics.Models comparison statistics(
177
+ test_case_ranks_baseline = 56,
178
+ test_case_ranks_current = 56,
179
+ test_case_wins_baseline = 56,
180
+ test_case_wins_current = 56, ),
181
+ models_comparisons_metrics = eval_studio_client.api.models.detailed_metrics_comparisons.Detailed metrics comparisons(
182
+ metrics_ranks_baseline = 1.337,
183
+ metrics_ranks_current = 1.337,
184
+ metrics_wins_baseline = 56,
185
+ metrics_wins_current = 56,
186
+ metrics_averages = [
187
+ eval_studio_client.api.models.metric_average_comparison.Metric average comparison(
188
+ metric_key = '',
189
+ baseline_avg = 1.337,
190
+ current_avg = 1.337,
191
+ diff = 1.337,
192
+ baseline_better_wins = 56,
193
+ current_better_wins = 56,
194
+ baseline_rank_avg = 1.337,
195
+ current_rank_avg = 1.337, )
196
+ ], ),
197
+ technical_metrics = eval_studio_client.api.models.technical_metrics_for_model_performance.Technical metrics for model performance(
198
+ baseline = eval_studio_client.api.models.technical_metrics_detail.Technical metrics detail(
199
+ cost_sum = 1.337,
200
+ duration_sum = 1.337,
201
+ duration_min = 1.337,
202
+ duration_max = 1.337,
203
+ duration_avg = 1.337, ),
204
+ current = eval_studio_client.api.models.technical_metrics_detail.Technical metrics detail(
205
+ cost_sum = 1.337,
206
+ duration_sum = 1.337,
207
+ duration_min = 1.337,
208
+ duration_max = 1.337,
209
+ duration_avg = 1.337, ), ),
210
+ test_cases_leaderboard = [
211
+ eval_studio_client.api.models.test_case_leaderboard_item.Test case leaderboard item(
212
+ wins = 56,
213
+ question = '',
214
+ changed_metrics_count = 56, )
215
+ ], )
216
+ ],
217
+ leaderboards = [
218
+ eval_studio_client.api.models.leaderboard_information.Leaderboard information(
219
+ key = '',
220
+ items = [
221
+ eval_studio_client.api.models.leaderboard_comparison_item.Leaderboard comparison item(
222
+ wins = 56,
223
+ question = '',
224
+ changed_metrics_count = 56, )
225
+ ], )
226
+ ],
227
+ metrics_meta = {
228
+ 'key' : eval_studio_client.api.models.metric_metadata.Metric metadata(
229
+ key = '',
230
+ display_name = '',
231
+ data_type = '',
232
+ display_value = '',
233
+ description = '',
234
+ value_range = [
235
+ 1.337
236
+ ],
237
+ value_enum = [
238
+ ''
239
+ ],
240
+ higher_is_better = True,
241
+ threshold = 1.337,
242
+ is_primary_metric = True,
243
+ parent_metric = '',
244
+ exclude = True, )
245
+ }
246
+ )
247
+ else:
248
+ return V1ComparisonResult(
249
+ )
250
+ """
251
+
252
+ def testV1ComparisonResult(self):
253
+ """Test V1ComparisonResult"""
254
+ # inst_req_only = self.make_instance(include_optional=False)
255
+ # inst_req_and_optional = self.make_instance(include_optional=True)
256
+
257
+ if __name__ == '__main__':
258
+ unittest.main()
@@ -0,0 +1,53 @@
1
+ # coding: utf-8
2
+
3
+ """
4
+ ai/h2o/eval_studio/v1/insight.proto
5
+
6
+ No description provided (generated by Openapi Generator https://github.com/openapitools/openapi-generator)
7
+
8
+ The version of the OpenAPI document: version not set
9
+ Generated by OpenAPI Generator (https://openapi-generator.tech)
10
+
11
+ Do not edit the class manually.
12
+ """ # noqa: E501
13
+
14
+
15
+ import unittest
16
+
17
+ from eval_studio_client.api.models.v1_comparison_summary import V1ComparisonSummary
18
+
19
+ class TestV1ComparisonSummary(unittest.TestCase):
20
+ """V1ComparisonSummary unit test stubs"""
21
+
22
+ def setUp(self):
23
+ pass
24
+
25
+ def tearDown(self):
26
+ pass
27
+
28
+ def make_instance(self, include_optional) -> V1ComparisonSummary:
29
+ """Test V1ComparisonSummary
30
+ include_option is a boolean, when False only required
31
+ params are included, when True both required and
32
+ optional params are included """
33
+ # uncomment below to create an instance of `V1ComparisonSummary`
34
+ """
35
+ model = V1ComparisonSummary()
36
+ if include_optional:
37
+ return V1ComparisonSummary(
38
+ recommendation_winner = '',
39
+ recommendation = '',
40
+ recommendation_confidence = ''
41
+ )
42
+ else:
43
+ return V1ComparisonSummary(
44
+ )
45
+ """
46
+
47
+ def testV1ComparisonSummary(self):
48
+ """Test V1ComparisonSummary"""
49
+ # inst_req_only = self.make_instance(include_optional=False)
50
+ # inst_req_and_optional = self.make_instance(include_optional=True)
51
+
52
+ if __name__ == '__main__':
53
+ unittest.main()
@@ -113,7 +113,8 @@ class TestV1CreateEvaluationRequest(unittest.TestCase):
113
113
  api_key = '',
114
114
  type = 'MODEL_TYPE_UNSPECIFIED',
115
115
  parameters = '',
116
- demo = True, )
116
+ demo = True, ),
117
+ evaluation_type = 'EVALUATION_TYPE_UNSPECIFIED'
117
118
  )
118
119
  else:
119
120
  return V1CreateEvaluationRequest(
@@ -98,7 +98,8 @@ class TestV1CreateLeaderboardRequest(unittest.TestCase):
98
98
  h2ogpte_collection = '',
99
99
  type = 'LEADERBOARD_TYPE_UNSPECIFIED',
100
100
  demo = True,
101
- test_lab = '', )
101
+ test_lab = '',
102
+ evaluation_type = 'EVALUATION_TYPE_UNSPECIFIED', )
102
103
  )
103
104
  else:
104
105
  return V1CreateLeaderboardRequest(
@@ -55,7 +55,8 @@ class TestV1CreateLeaderboardResponse(unittest.TestCase):
55
55
  'key' : None
56
56
  }
57
57
  ], ),
58
- response = , )
58
+ response = ,
59
+ seen_by_creator_time = datetime.datetime.strptime('2013-10-20 19:20:30.00', '%Y-%m-%d %H:%M:%S.%f'), )
59
60
  )
60
61
  else:
61
62
  return V1CreateLeaderboardResponse(
@@ -55,7 +55,8 @@ class TestV1CreateLeaderboardWithoutCacheResponse(unittest.TestCase):
55
55
  'key' : None
56
56
  }
57
57
  ], ),
58
- response = , )
58
+ response = ,
59
+ seen_by_creator_time = datetime.datetime.strptime('2013-10-20 19:20:30.00', '%Y-%m-%d %H:%M:%S.%f'), )
59
60
  )
60
61
  else:
61
62
  return V1CreateLeaderboardWithoutCacheResponse(
@@ -0,0 +1,54 @@
1
+ # coding: utf-8
2
+
3
+ """
4
+ ai/h2o/eval_studio/v1/insight.proto
5
+
6
+ No description provided (generated by Openapi Generator https://github.com/openapitools/openapi-generator)
7
+
8
+ The version of the OpenAPI document: version not set
9
+ Generated by OpenAPI Generator (https://openapi-generator.tech)
10
+
11
+ Do not edit the class manually.
12
+ """ # noqa: E501
13
+
14
+
15
+ import unittest
16
+
17
+ from eval_studio_client.api.models.v1_create_test_from_test_cases_request import V1CreateTestFromTestCasesRequest
18
+
19
+ class TestV1CreateTestFromTestCasesRequest(unittest.TestCase):
20
+ """V1CreateTestFromTestCasesRequest unit test stubs"""
21
+
22
+ def setUp(self):
23
+ pass
24
+
25
+ def tearDown(self):
26
+ pass
27
+
28
+ def make_instance(self, include_optional) -> V1CreateTestFromTestCasesRequest:
29
+ """Test V1CreateTestFromTestCasesRequest
30
+ include_option is a boolean, when False only required
31
+ params are included, when True both required and
32
+ optional params are included """
33
+ # uncomment below to create an instance of `V1CreateTestFromTestCasesRequest`
34
+ """
35
+ model = V1CreateTestFromTestCasesRequest()
36
+ if include_optional:
37
+ return V1CreateTestFromTestCasesRequest(
38
+ tests_json = '',
39
+ url = '',
40
+ test_display_name = '',
41
+ test_description = ''
42
+ )
43
+ else:
44
+ return V1CreateTestFromTestCasesRequest(
45
+ )
46
+ """
47
+
48
+ def testV1CreateTestFromTestCasesRequest(self):
49
+ """Test V1CreateTestFromTestCasesRequest"""
50
+ # inst_req_only = self.make_instance(include_optional=False)
51
+ # inst_req_and_optional = self.make_instance(include_optional=True)
52
+
53
+ if __name__ == '__main__':
54
+ unittest.main()
@@ -0,0 +1,68 @@
1
+ # coding: utf-8
2
+
3
+ """
4
+ ai/h2o/eval_studio/v1/insight.proto
5
+
6
+ No description provided (generated by Openapi Generator https://github.com/openapitools/openapi-generator)
7
+
8
+ The version of the OpenAPI document: version not set
9
+ Generated by OpenAPI Generator (https://openapi-generator.tech)
10
+
11
+ Do not edit the class manually.
12
+ """ # noqa: E501
13
+
14
+
15
+ import unittest
16
+
17
+ from eval_studio_client.api.models.v1_create_test_from_test_cases_response import V1CreateTestFromTestCasesResponse
18
+
19
+ class TestV1CreateTestFromTestCasesResponse(unittest.TestCase):
20
+ """V1CreateTestFromTestCasesResponse unit test stubs"""
21
+
22
+ def setUp(self):
23
+ pass
24
+
25
+ def tearDown(self):
26
+ pass
27
+
28
+ def make_instance(self, include_optional) -> V1CreateTestFromTestCasesResponse:
29
+ """Test V1CreateTestFromTestCasesResponse
30
+ include_option is a boolean, when False only required
31
+ params are included, when True both required and
32
+ optional params are included """
33
+ # uncomment below to create an instance of `V1CreateTestFromTestCasesResponse`
34
+ """
35
+ model = V1CreateTestFromTestCasesResponse()
36
+ if include_optional:
37
+ return V1CreateTestFromTestCasesResponse(
38
+ test = eval_studio_client.api.models.v1_test.v1Test(
39
+ name = '',
40
+ create_time = datetime.datetime.strptime('2013-10-20 19:20:30.00', '%Y-%m-%d %H:%M:%S.%f'),
41
+ creator = '',
42
+ update_time = datetime.datetime.strptime('2013-10-20 19:20:30.00', '%Y-%m-%d %H:%M:%S.%f'),
43
+ updater = '',
44
+ delete_time = datetime.datetime.strptime('2013-10-20 19:20:30.00', '%Y-%m-%d %H:%M:%S.%f'),
45
+ deleter = '',
46
+ display_name = '',
47
+ description = '',
48
+ documents = [
49
+ ''
50
+ ],
51
+ tags = [
52
+ ''
53
+ ],
54
+ demo = True,
55
+ type = 'TEST_TYPE_UNSPECIFIED', )
56
+ )
57
+ else:
58
+ return V1CreateTestFromTestCasesResponse(
59
+ )
60
+ """
61
+
62
+ def testV1CreateTestFromTestCasesResponse(self):
63
+ """Test V1CreateTestFromTestCasesResponse"""
64
+ # inst_req_only = self.make_instance(include_optional=False)
65
+ # inst_req_and_optional = self.make_instance(include_optional=True)
66
+
67
+ if __name__ == '__main__':
68
+ unittest.main()
@@ -0,0 +1,142 @@
1
+ # coding: utf-8
2
+
3
+ """
4
+ ai/h2o/eval_studio/v1/insight.proto
5
+
6
+ No description provided (generated by Openapi Generator https://github.com/openapitools/openapi-generator)
7
+
8
+ The version of the OpenAPI document: version not set
9
+ Generated by OpenAPI Generator (https://openapi-generator.tech)
10
+
11
+ Do not edit the class manually.
12
+ """ # noqa: E501
13
+
14
+
15
+ import unittest
16
+
17
+ from eval_studio_client.api.models.v1_dashboard_report import V1DashboardReport
18
+
19
+ class TestV1DashboardReport(unittest.TestCase):
20
+ """V1DashboardReport unit test stubs"""
21
+
22
+ def setUp(self):
23
+ pass
24
+
25
+ def tearDown(self):
26
+ pass
27
+
28
+ def make_instance(self, include_optional) -> V1DashboardReport:
29
+ """Test V1DashboardReport
30
+ include_option is a boolean, when False only required
31
+ params are included, when True both required and
32
+ optional params are included """
33
+ # uncomment below to create an instance of `V1DashboardReport`
34
+ """
35
+ model = V1DashboardReport()
36
+ if include_optional:
37
+ return V1DashboardReport(
38
+ results = [
39
+ eval_studio_client.api.models.v1_dashboard_report_result.v1DashboardReportResult(
40
+ key = '',
41
+ input = '',
42
+ expected_output = '',
43
+ actual_output = '',
44
+ model_key = '',
45
+ test_case_key = '',
46
+ metrics = {
47
+ 'key' : eval_studio_client.api.models.v1_metric_scores.v1MetricScores(
48
+ scores = [
49
+ eval_studio_client.api.models.v1_metric_score.v1MetricScore(
50
+ key = '',
51
+ value = 1.337, )
52
+ ], )
53
+ },
54
+ result_error_map = {
55
+ 'key' : ''
56
+ },
57
+ human_decision = 'HUMAN_DECISION_UNSPECIFIED',
58
+ comment = '',
59
+ annotations = {
60
+ 'key' : None
61
+ }, )
62
+ ],
63
+ evaluator = [
64
+ eval_studio_client.api.models.v1_leaderboard_report_evaluator.v1LeaderboardReportEvaluator(
65
+ id = '',
66
+ name = '',
67
+ display_name = '',
68
+ tagline = '',
69
+ description = '',
70
+ brief_description = '',
71
+ model_types = [
72
+ ''
73
+ ],
74
+ can_explain = [
75
+ ''
76
+ ],
77
+ explanation_scopes = [
78
+ ''
79
+ ],
80
+ explanations = [
81
+ eval_studio_client.api.models.v1_leaderboard_report_explanation.v1LeaderboardReportExplanation(
82
+ explanation_type = '',
83
+ name = '',
84
+ category = '',
85
+ scope = '',
86
+ has_local = '',
87
+ formats = [
88
+ ''
89
+ ], )
90
+ ],
91
+ parameters = [
92
+ eval_studio_client.api.models.v1_leaderboard_report_evaluator_parameter.v1LeaderboardReportEvaluatorParameter(
93
+ name = '',
94
+ description = '',
95
+ comment = '',
96
+ type = '',
97
+ predefined = [
98
+ None
99
+ ],
100
+ tags = [
101
+ ''
102
+ ],
103
+ min = 1.337,
104
+ max = 1.337,
105
+ category = '', )
106
+ ],
107
+ keywords = [
108
+ ''
109
+ ],
110
+ metrics_meta = [
111
+ eval_studio_client.api.models.v1_leaderboard_report_metrics_meta_entry.v1LeaderboardReportMetricsMetaEntry(
112
+ key = '',
113
+ display_name = '',
114
+ data_type = '',
115
+ display_value = '',
116
+ description = '',
117
+ value_range = [
118
+ 1.337
119
+ ],
120
+ value_enum = [
121
+ ''
122
+ ],
123
+ higher_is_better = True,
124
+ threshold = 1.337,
125
+ is_primary_metric = True,
126
+ parent_metric = '',
127
+ exclude = True, )
128
+ ], )
129
+ ]
130
+ )
131
+ else:
132
+ return V1DashboardReport(
133
+ )
134
+ """
135
+
136
+ def testV1DashboardReport(self):
137
+ """Test V1DashboardReport"""
138
+ # inst_req_only = self.make_instance(include_optional=False)
139
+ # inst_req_and_optional = self.make_instance(include_optional=True)
140
+
141
+ if __name__ == '__main__':
142
+ unittest.main()