eval-studio-client 1.2.5__py3-none-any.whl → 1.3.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- eval_studio_client/api/__init__.py +65 -0
- eval_studio_client/api/api/__init__.py +3 -0
- eval_studio_client/api/api/dashboard_report_service_api.py +292 -0
- eval_studio_client/api/api/dashboard_service_api.py +16 -16
- eval_studio_client/api/api/dashboard_test_case_annotation_service_api.py +611 -0
- eval_studio_client/api/api/document_service_api.py +16 -16
- eval_studio_client/api/api/evaluation_service_api.py +12 -12
- eval_studio_client/api/api/evaluator_service_api.py +16 -16
- eval_studio_client/api/api/leaderboard_report_service_api.py +304 -17
- eval_studio_client/api/api/leaderboard_service_api.py +554 -16
- eval_studio_client/api/api/leaderboard_test_case_annotation_service_api.py +611 -0
- eval_studio_client/api/api/model_service_api.py +16 -16
- eval_studio_client/api/api/operation_service_api.py +821 -17
- eval_studio_client/api/api/perturbator_service_api.py +22 -22
- eval_studio_client/api/api/test_case_service_api.py +300 -16
- eval_studio_client/api/api/test_class_service_api.py +16 -16
- eval_studio_client/api/api/test_service_api.py +285 -16
- eval_studio_client/api/api/workflow_node_service_api.py +16 -16
- eval_studio_client/api/api/workflow_service_api.py +16 -16
- eval_studio_client/api/docs/AdversarialInputsServiceTestAdversarialInputsRobustnessRequest.md +2 -1
- eval_studio_client/api/docs/DashboardReportServiceApi.md +75 -0
- eval_studio_client/api/docs/DashboardServiceApi.md +5 -5
- eval_studio_client/api/docs/DashboardTestCaseAnnotationServiceApi.md +149 -0
- eval_studio_client/api/docs/DocumentServiceApi.md +5 -5
- eval_studio_client/api/docs/EvaluationServiceApi.md +4 -4
- eval_studio_client/api/docs/EvaluatorServiceApi.md +5 -5
- eval_studio_client/api/docs/LeaderboardReportServiceApi.md +75 -5
- eval_studio_client/api/docs/LeaderboardServiceApi.md +141 -5
- eval_studio_client/api/docs/LeaderboardTestCaseAnnotationServiceApi.md +149 -0
- eval_studio_client/api/docs/ModelServiceApi.md +5 -5
- eval_studio_client/api/docs/OperationServiceApi.md +215 -8
- eval_studio_client/api/docs/PerturbatorServiceApi.md +7 -7
- eval_studio_client/api/docs/RequiredTheDashboardTestCaseAnnotationToUpdate.md +35 -0
- eval_studio_client/api/docs/RequiredTheLeaderboardTestCaseAnnotationToUpdate.md +35 -0
- eval_studio_client/api/docs/RequiredTheLeaderboardToUpdate.md +1 -0
- eval_studio_client/api/docs/RequiredTheOperationToFinalize.md +1 -0
- eval_studio_client/api/docs/RequiredTheOperationToUpdate.md +1 -0
- eval_studio_client/api/docs/TestCaseServiceApi.md +75 -5
- eval_studio_client/api/docs/TestCaseServiceAppendTestCasesRequest.md +30 -0
- eval_studio_client/api/docs/TestClassServiceApi.md +5 -5
- eval_studio_client/api/docs/TestServiceApi.md +73 -5
- eval_studio_client/api/docs/V1ActualOutputMeta.md +30 -0
- eval_studio_client/api/docs/V1ActualOutputMetaDiff.md +36 -0
- eval_studio_client/api/docs/V1AgentChatActivityDiagram.md +31 -0
- eval_studio_client/api/docs/V1AgentChatActivityDiagramEdge.md +32 -0
- eval_studio_client/api/docs/V1AgentChatActivityDiagramNode.md +32 -0
- eval_studio_client/api/docs/V1AgentChatActivityDiagramRow.md +30 -0
- eval_studio_client/api/docs/V1AgentChatScriptUsage.md +33 -0
- eval_studio_client/api/docs/V1AgentChatScriptsBarChart.md +30 -0
- eval_studio_client/api/docs/V1AgentChatToolUsage.md +33 -0
- eval_studio_client/api/docs/V1AgentChatToolsBarChart.md +30 -0
- eval_studio_client/api/docs/V1AllMetricScores.md +29 -0
- eval_studio_client/api/docs/V1AppendTestCasesResponse.md +29 -0
- eval_studio_client/api/docs/V1BatchCreateLeaderboardsWithoutCacheRequest.md +31 -0
- eval_studio_client/api/docs/V1BatchCreateLeaderboardsWithoutCacheResponse.md +29 -0
- eval_studio_client/api/docs/V1BatchMarkOperationSeenByCreatorResponse.md +29 -0
- eval_studio_client/api/docs/V1CmpLeaderboardReportsRequest.md +33 -0
- eval_studio_client/api/docs/V1CmpLeaderboardReportsResponse.md +29 -0
- eval_studio_client/api/docs/V1ComparisonItem.md +36 -0
- eval_studio_client/api/docs/V1ComparisonMetricScore.md +30 -0
- eval_studio_client/api/docs/V1ComparisonResult.md +31 -0
- eval_studio_client/api/docs/V1ComparisonSummary.md +31 -0
- eval_studio_client/api/docs/V1CreateEvaluationRequest.md +1 -0
- eval_studio_client/api/docs/V1CreateTestFromTestCasesRequest.md +32 -0
- eval_studio_client/api/docs/V1CreateTestFromTestCasesResponse.md +29 -0
- eval_studio_client/api/docs/V1DashboardReport.md +31 -0
- eval_studio_client/api/docs/V1DashboardReportResult.md +39 -0
- eval_studio_client/api/docs/V1DashboardTestCaseAnnotation.md +36 -0
- eval_studio_client/api/docs/V1DataFragment.md +31 -0
- eval_studio_client/api/docs/V1DeepCompareLeaderboardsRequest.md +33 -0
- eval_studio_client/api/docs/V1DeepCompareLeaderboardsResponse.md +29 -0
- eval_studio_client/api/docs/V1DiffItem.md +36 -0
- eval_studio_client/api/docs/V1EvaluationType.md +12 -0
- eval_studio_client/api/docs/V1FlippedMetric.md +31 -0
- eval_studio_client/api/docs/V1GetDashboardReportResponse.md +29 -0
- eval_studio_client/api/docs/V1HumanDecision.md +12 -0
- eval_studio_client/api/docs/V1Info.md +1 -0
- eval_studio_client/api/docs/V1Leaderboard.md +1 -0
- eval_studio_client/api/docs/V1LeaderboardCmpReport.md +30 -0
- eval_studio_client/api/docs/V1LeaderboardComparisonItem.md +31 -0
- eval_studio_client/api/docs/V1LeaderboardInfo.md +30 -0
- eval_studio_client/api/docs/V1LeaderboardReportActualOutputMeta.md +6 -3
- eval_studio_client/api/docs/V1LeaderboardReportResult.md +11 -8
- eval_studio_client/api/docs/V1LeaderboardReportResultView.md +12 -0
- eval_studio_client/api/docs/V1LeaderboardTestCaseAnnotation.md +36 -0
- eval_studio_client/api/docs/V1ListDashboardTestCaseAnnotationsResponse.md +29 -0
- eval_studio_client/api/docs/V1ListLeaderboardTestCaseAnnotationsResponse.md +29 -0
- eval_studio_client/api/docs/V1ListOperationsResponse.md +1 -0
- eval_studio_client/api/docs/V1ListUnseenOperationsResponse.md +30 -0
- eval_studio_client/api/docs/V1MarkOperationSeenByCreatorResponse.md +29 -0
- eval_studio_client/api/docs/V1Metric.md +30 -0
- eval_studio_client/api/docs/V1MetricAverage.md +36 -0
- eval_studio_client/api/docs/V1MetricMeta.md +40 -0
- eval_studio_client/api/docs/V1MetricScore.md +1 -1
- eval_studio_client/api/docs/V1MetricScores.md +1 -1
- eval_studio_client/api/docs/V1ModelType.md +1 -1
- eval_studio_client/api/docs/V1ModelsComparisons.md +32 -0
- eval_studio_client/api/docs/V1ModelsComparisonsMetrics.md +33 -0
- eval_studio_client/api/docs/V1ModelsOverview.md +34 -0
- eval_studio_client/api/docs/V1Operation.md +1 -0
- eval_studio_client/api/docs/V1OperationView.md +12 -0
- eval_studio_client/api/docs/V1RetrievedContextDiff.md +36 -0
- eval_studio_client/api/docs/V1Stats.md +2 -0
- eval_studio_client/api/docs/V1TechnicalMetrics.md +30 -0
- eval_studio_client/api/docs/V1TechnicalMetricsDetail.md +33 -0
- eval_studio_client/api/docs/V1TestCaseLeaderboardItem.md +31 -0
- eval_studio_client/api/docs/V1TestCaseRelationshipInfo.md +31 -0
- eval_studio_client/api/docs/V1TestCaseResult.md +48 -0
- eval_studio_client/api/docs/V1TextSimilarityMetric.md +12 -0
- eval_studio_client/api/docs/V1UpdateDashboardTestCaseAnnotationResponse.md +29 -0
- eval_studio_client/api/docs/V1UpdateLeaderboardTestCaseAnnotationResponse.md +29 -0
- eval_studio_client/api/docs/WorkflowNodeServiceApi.md +5 -5
- eval_studio_client/api/docs/WorkflowServiceApi.md +5 -5
- eval_studio_client/api/models/__init__.py +62 -0
- eval_studio_client/api/models/adversarial_inputs_service_test_adversarial_inputs_robustness_request.py +17 -2
- eval_studio_client/api/models/required_the_dashboard_test_case_annotation_to_update.py +108 -0
- eval_studio_client/api/models/required_the_leaderboard_test_case_annotation_to_update.py +108 -0
- eval_studio_client/api/models/required_the_leaderboard_to_update.py +5 -2
- eval_studio_client/api/models/required_the_operation_to_finalize.py +6 -2
- eval_studio_client/api/models/required_the_operation_to_update.py +6 -2
- eval_studio_client/api/models/test_case_service_append_test_cases_request.py +89 -0
- eval_studio_client/api/models/v1_actual_output_meta.py +97 -0
- eval_studio_client/api/models/v1_actual_output_meta_diff.py +101 -0
- eval_studio_client/api/models/v1_agent_chat_activity_diagram.py +109 -0
- eval_studio_client/api/models/v1_agent_chat_activity_diagram_edge.py +97 -0
- eval_studio_client/api/models/v1_agent_chat_activity_diagram_node.py +97 -0
- eval_studio_client/api/models/v1_agent_chat_activity_diagram_row.py +97 -0
- eval_studio_client/api/models/v1_agent_chat_script_usage.py +101 -0
- eval_studio_client/api/models/v1_agent_chat_scripts_bar_chart.py +102 -0
- eval_studio_client/api/models/v1_agent_chat_tool_usage.py +101 -0
- eval_studio_client/api/models/v1_agent_chat_tools_bar_chart.py +102 -0
- eval_studio_client/api/models/v1_all_metric_scores.py +87 -0
- eval_studio_client/api/models/v1_append_test_cases_response.py +95 -0
- eval_studio_client/api/models/v1_batch_create_leaderboards_without_cache_request.py +99 -0
- eval_studio_client/api/models/v1_batch_create_leaderboards_without_cache_response.py +91 -0
- eval_studio_client/api/models/v1_batch_mark_operation_seen_by_creator_response.py +95 -0
- eval_studio_client/api/models/v1_cmp_leaderboard_reports_request.py +96 -0
- eval_studio_client/api/models/v1_cmp_leaderboard_reports_response.py +91 -0
- eval_studio_client/api/models/v1_comparison_item.py +130 -0
- eval_studio_client/api/models/v1_comparison_metric_score.py +89 -0
- eval_studio_client/api/models/v1_comparison_result.py +120 -0
- eval_studio_client/api/models/v1_comparison_summary.py +91 -0
- eval_studio_client/api/models/v1_create_evaluation_request.py +5 -2
- eval_studio_client/api/models/v1_create_test_from_test_cases_request.py +93 -0
- eval_studio_client/api/models/v1_create_test_from_test_cases_response.py +91 -0
- eval_studio_client/api/models/v1_dashboard_report.py +109 -0
- eval_studio_client/api/models/v1_dashboard_report_result.py +139 -0
- eval_studio_client/api/models/v1_dashboard_test_case_annotation.py +112 -0
- eval_studio_client/api/models/v1_data_fragment.py +91 -0
- eval_studio_client/api/models/v1_deep_compare_leaderboards_request.py +96 -0
- eval_studio_client/api/models/v1_deep_compare_leaderboards_response.py +91 -0
- eval_studio_client/api/models/v1_diff_item.py +137 -0
- eval_studio_client/api/models/v1_evaluation_type.py +39 -0
- eval_studio_client/api/models/v1_flipped_metric.py +91 -0
- eval_studio_client/api/models/v1_get_dashboard_report_response.py +91 -0
- eval_studio_client/api/models/v1_human_decision.py +38 -0
- eval_studio_client/api/models/v1_info.py +4 -2
- eval_studio_client/api/models/v1_leaderboard.py +5 -2
- eval_studio_client/api/models/v1_leaderboard_cmp_report.py +93 -0
- eval_studio_client/api/models/v1_leaderboard_comparison_item.py +91 -0
- eval_studio_client/api/models/v1_leaderboard_info.py +97 -0
- eval_studio_client/api/models/v1_leaderboard_report_actual_output_meta.py +23 -9
- eval_studio_client/api/models/v1_leaderboard_report_result.py +21 -10
- eval_studio_client/api/models/v1_leaderboard_report_result_view.py +38 -0
- eval_studio_client/api/models/v1_leaderboard_test_case_annotation.py +112 -0
- eval_studio_client/api/models/v1_list_dashboard_test_case_annotations_response.py +95 -0
- eval_studio_client/api/models/v1_list_leaderboard_test_case_annotations_response.py +95 -0
- eval_studio_client/api/models/v1_list_operations_response.py +5 -3
- eval_studio_client/api/models/v1_list_unseen_operations_response.py +97 -0
- eval_studio_client/api/models/v1_mark_operation_seen_by_creator_response.py +91 -0
- eval_studio_client/api/models/v1_metric.py +89 -0
- eval_studio_client/api/models/v1_metric_average.py +101 -0
- eval_studio_client/api/models/v1_metric_meta.py +109 -0
- eval_studio_client/api/models/v1_metric_score.py +6 -1
- eval_studio_client/api/models/v1_metric_scores.py +1 -1
- eval_studio_client/api/models/v1_model_type.py +2 -1
- eval_studio_client/api/models/v1_models_comparisons.py +93 -0
- eval_studio_client/api/models/v1_models_comparisons_metrics.py +103 -0
- eval_studio_client/api/models/v1_models_overview.py +97 -0
- eval_studio_client/api/models/v1_operation.py +6 -2
- eval_studio_client/api/models/v1_operation_view.py +38 -0
- eval_studio_client/api/models/v1_retrieved_context_diff.py +101 -0
- eval_studio_client/api/models/v1_stats.py +16 -2
- eval_studio_client/api/models/v1_technical_metrics.py +96 -0
- eval_studio_client/api/models/v1_technical_metrics_detail.py +95 -0
- eval_studio_client/api/models/v1_test_case_leaderboard_item.py +91 -0
- eval_studio_client/api/models/v1_test_case_relationship_info.py +91 -0
- eval_studio_client/api/models/v1_test_case_result.py +157 -0
- eval_studio_client/api/models/v1_text_similarity_metric.py +39 -0
- eval_studio_client/api/models/v1_update_dashboard_test_case_annotation_response.py +91 -0
- eval_studio_client/api/models/v1_update_leaderboard_test_case_annotation_response.py +91 -0
- eval_studio_client/api/models/v1_workflow_node_type.py +1 -0
- eval_studio_client/api/models/v1_workflow_type.py +1 -0
- eval_studio_client/api/test/test_adversarial_inputs_service_test_adversarial_inputs_robustness_request.py +6 -0
- eval_studio_client/api/test/test_dashboard_report_service_api.py +37 -0
- eval_studio_client/api/test/test_dashboard_test_case_annotation_service_api.py +43 -0
- eval_studio_client/api/test/test_leaderboard_report_service_api.py +6 -0
- eval_studio_client/api/test/test_leaderboard_service_api.py +12 -0
- eval_studio_client/api/test/test_leaderboard_test_case_annotation_service_api.py +43 -0
- eval_studio_client/api/test/test_operation_service_api.py +18 -0
- eval_studio_client/api/test/test_required_the_dashboard_test_case_annotation_to_update.py +57 -0
- eval_studio_client/api/test/test_required_the_leaderboard_test_case_annotation_to_update.py +57 -0
- eval_studio_client/api/test/test_required_the_leaderboard_to_update.py +2 -1
- eval_studio_client/api/test/test_required_the_operation_to_finalize.py +2 -1
- eval_studio_client/api/test/test_required_the_operation_to_update.py +2 -1
- eval_studio_client/api/test/test_test_case_service_api.py +6 -0
- eval_studio_client/api/test/test_test_case_service_append_test_cases_request.py +52 -0
- eval_studio_client/api/test/test_test_service_api.py +6 -0
- eval_studio_client/api/test/test_v1_abort_operation_response.py +2 -1
- eval_studio_client/api/test/test_v1_actual_output_meta.py +61 -0
- eval_studio_client/api/test/test_v1_actual_output_meta_diff.py +66 -0
- eval_studio_client/api/test/test_v1_agent_chat_activity_diagram.py +65 -0
- eval_studio_client/api/test/test_v1_agent_chat_activity_diagram_edge.py +53 -0
- eval_studio_client/api/test/test_v1_agent_chat_activity_diagram_node.py +53 -0
- eval_studio_client/api/test/test_v1_agent_chat_activity_diagram_row.py +56 -0
- eval_studio_client/api/test/test_v1_agent_chat_script_usage.py +54 -0
- eval_studio_client/api/test/test_v1_agent_chat_scripts_bar_chart.py +57 -0
- eval_studio_client/api/test/test_v1_agent_chat_tool_usage.py +54 -0
- eval_studio_client/api/test/test_v1_agent_chat_tools_bar_chart.py +57 -0
- eval_studio_client/api/test/test_v1_all_metric_scores.py +53 -0
- eval_studio_client/api/test/test_v1_append_test_cases_response.py +74 -0
- eval_studio_client/api/test/test_v1_batch_create_leaderboards_request.py +2 -1
- eval_studio_client/api/test/test_v1_batch_create_leaderboards_response.py +2 -1
- eval_studio_client/api/test/test_v1_batch_create_leaderboards_without_cache_request.py +120 -0
- eval_studio_client/api/test/test_v1_batch_create_leaderboards_without_cache_response.py +72 -0
- eval_studio_client/api/test/test_v1_batch_delete_leaderboards_response.py +2 -1
- eval_studio_client/api/test/test_v1_batch_get_leaderboards_response.py +2 -1
- eval_studio_client/api/test/test_v1_batch_get_operations_response.py +2 -1
- eval_studio_client/api/test/test_v1_batch_import_leaderboard_response.py +2 -1
- eval_studio_client/api/test/test_v1_batch_mark_operation_seen_by_creator_response.py +74 -0
- eval_studio_client/api/test/test_v1_cmp_leaderboard_reports_request.py +55 -0
- eval_studio_client/api/test/test_v1_cmp_leaderboard_reports_response.py +255 -0
- eval_studio_client/api/test/test_v1_comparison_item.py +233 -0
- eval_studio_client/api/test/test_v1_comparison_metric_score.py +52 -0
- eval_studio_client/api/test/test_v1_comparison_result.py +258 -0
- eval_studio_client/api/test/test_v1_comparison_summary.py +53 -0
- eval_studio_client/api/test/test_v1_create_evaluation_request.py +2 -1
- eval_studio_client/api/test/test_v1_create_leaderboard_request.py +2 -1
- eval_studio_client/api/test/test_v1_create_leaderboard_response.py +2 -1
- eval_studio_client/api/test/test_v1_create_leaderboard_without_cache_response.py +2 -1
- eval_studio_client/api/test/test_v1_create_test_from_test_cases_request.py +54 -0
- eval_studio_client/api/test/test_v1_create_test_from_test_cases_response.py +68 -0
- eval_studio_client/api/test/test_v1_dashboard_report.py +142 -0
- eval_studio_client/api/test/test_v1_dashboard_report_result.py +72 -0
- eval_studio_client/api/test/test_v1_dashboard_test_case_annotation.py +58 -0
- eval_studio_client/api/test/test_v1_data_fragment.py +57 -0
- eval_studio_client/api/test/test_v1_deep_compare_leaderboards_request.py +55 -0
- eval_studio_client/api/test/test_v1_deep_compare_leaderboards_response.py +255 -0
- eval_studio_client/api/test/test_v1_delete_leaderboard_response.py +2 -1
- eval_studio_client/api/test/test_v1_diff_item.py +226 -0
- eval_studio_client/api/test/test_v1_evaluation_type.py +33 -0
- eval_studio_client/api/test/test_v1_finalize_operation_response.py +2 -1
- eval_studio_client/api/test/test_v1_flipped_metric.py +53 -0
- eval_studio_client/api/test/test_v1_generate_test_cases_response.py +2 -1
- eval_studio_client/api/test/test_v1_get_dashboard_report_response.py +143 -0
- eval_studio_client/api/test/test_v1_get_info_response.py +4 -1
- eval_studio_client/api/test/test_v1_get_leaderboard_report_response.py +39 -2
- eval_studio_client/api/test/test_v1_get_leaderboard_response.py +2 -1
- eval_studio_client/api/test/test_v1_get_operation_response.py +2 -1
- eval_studio_client/api/test/test_v1_get_stats_response.py +3 -1
- eval_studio_client/api/test/test_v1_human_decision.py +33 -0
- eval_studio_client/api/test/test_v1_import_leaderboard_response.py +2 -1
- eval_studio_client/api/test/test_v1_import_test_cases_from_library_response.py +2 -1
- eval_studio_client/api/test/test_v1_info.py +4 -1
- eval_studio_client/api/test/test_v1_leaderboard.py +2 -1
- eval_studio_client/api/test/test_v1_leaderboard_cmp_report.py +254 -0
- eval_studio_client/api/test/test_v1_leaderboard_comparison_item.py +53 -0
- eval_studio_client/api/test/test_v1_leaderboard_info.py +57 -0
- eval_studio_client/api/test/test_v1_leaderboard_report.py +39 -2
- eval_studio_client/api/test/test_v1_leaderboard_report_actual_output_meta.py +33 -1
- eval_studio_client/api/test/test_v1_leaderboard_report_result.py +39 -2
- eval_studio_client/api/test/test_v1_leaderboard_report_result_view.py +33 -0
- eval_studio_client/api/test/test_v1_leaderboard_test_case_annotation.py +58 -0
- eval_studio_client/api/test/test_v1_list_dashboard_test_case_annotations_response.py +61 -0
- eval_studio_client/api/test/test_v1_list_leaderboard_test_case_annotations_response.py +61 -0
- eval_studio_client/api/test/test_v1_list_leaderboards_response.py +2 -1
- eval_studio_client/api/test/test_v1_list_most_recent_leaderboards_response.py +2 -1
- eval_studio_client/api/test/test_v1_list_operations_response.py +4 -2
- eval_studio_client/api/test/test_v1_list_unseen_operations_response.py +75 -0
- eval_studio_client/api/test/test_v1_mark_operation_seen_by_creator_response.py +72 -0
- eval_studio_client/api/test/test_v1_metric.py +52 -0
- eval_studio_client/api/test/test_v1_metric_average.py +58 -0
- eval_studio_client/api/test/test_v1_metric_meta.py +66 -0
- eval_studio_client/api/test/test_v1_models_comparisons.py +54 -0
- eval_studio_client/api/test/test_v1_models_comparisons_metrics.py +65 -0
- eval_studio_client/api/test/test_v1_models_overview.py +60 -0
- eval_studio_client/api/test/test_v1_operation.py +2 -1
- eval_studio_client/api/test/test_v1_operation_view.py +33 -0
- eval_studio_client/api/test/test_v1_process_workflow_node_response.py +2 -1
- eval_studio_client/api/test/test_v1_retrieved_context_diff.py +66 -0
- eval_studio_client/api/test/test_v1_stats.py +3 -1
- eval_studio_client/api/test/test_v1_technical_metrics.py +62 -0
- eval_studio_client/api/test/test_v1_technical_metrics_detail.py +55 -0
- eval_studio_client/api/test/test_v1_test_case_leaderboard_item.py +53 -0
- eval_studio_client/api/test/test_v1_test_case_relationship_info.py +53 -0
- eval_studio_client/api/test/test_v1_test_case_result.py +106 -0
- eval_studio_client/api/test/test_v1_text_similarity_metric.py +33 -0
- eval_studio_client/api/test/test_v1_update_dashboard_test_case_annotation_response.py +59 -0
- eval_studio_client/api/test/test_v1_update_leaderboard_response.py +2 -1
- eval_studio_client/api/test/test_v1_update_leaderboard_test_case_annotation_response.py +59 -0
- eval_studio_client/api/test/test_v1_update_operation_response.py +2 -1
- eval_studio_client/gen/openapiv2/eval_studio.swagger.json +2340 -210
- eval_studio_client/models.py +18 -6
- {eval_studio_client-1.2.5.dist-info → eval_studio_client-1.3.0.dist-info}/METADATA +2 -2
- {eval_studio_client-1.2.5.dist-info → eval_studio_client-1.3.0.dist-info}/RECORD +306 -111
- {eval_studio_client-1.2.5.dist-info → eval_studio_client-1.3.0.dist-info}/WHEEL +0 -0
|
@@ -0,0 +1,258 @@
|
|
|
1
|
+
# coding: utf-8
|
|
2
|
+
|
|
3
|
+
"""
|
|
4
|
+
ai/h2o/eval_studio/v1/insight.proto
|
|
5
|
+
|
|
6
|
+
No description provided (generated by Openapi Generator https://github.com/openapitools/openapi-generator)
|
|
7
|
+
|
|
8
|
+
The version of the OpenAPI document: version not set
|
|
9
|
+
Generated by OpenAPI Generator (https://openapi-generator.tech)
|
|
10
|
+
|
|
11
|
+
Do not edit the class manually.
|
|
12
|
+
""" # noqa: E501
|
|
13
|
+
|
|
14
|
+
|
|
15
|
+
import unittest
|
|
16
|
+
|
|
17
|
+
from eval_studio_client.api.models.v1_comparison_result import V1ComparisonResult
|
|
18
|
+
|
|
19
|
+
class TestV1ComparisonResult(unittest.TestCase):
|
|
20
|
+
"""V1ComparisonResult unit test stubs"""
|
|
21
|
+
|
|
22
|
+
def setUp(self):
|
|
23
|
+
pass
|
|
24
|
+
|
|
25
|
+
def tearDown(self):
|
|
26
|
+
pass
|
|
27
|
+
|
|
28
|
+
def make_instance(self, include_optional) -> V1ComparisonResult:
|
|
29
|
+
"""Test V1ComparisonResult
|
|
30
|
+
include_option is a boolean, when False only required
|
|
31
|
+
params are included, when True both required and
|
|
32
|
+
optional params are included """
|
|
33
|
+
# uncomment below to create an instance of `V1ComparisonResult`
|
|
34
|
+
"""
|
|
35
|
+
model = V1ComparisonResult()
|
|
36
|
+
if include_optional:
|
|
37
|
+
return V1ComparisonResult(
|
|
38
|
+
diffs = [
|
|
39
|
+
eval_studio_client.api.models.a_single_diff_item_comparing_two_leaderboards.A single diff item comparing two leaderboards(
|
|
40
|
+
diff_key = '',
|
|
41
|
+
items = [
|
|
42
|
+
eval_studio_client.api.models.a_single_comparison_item_showing_differences_between_baseline_and_current.A single comparison item showing differences between baseline and current(
|
|
43
|
+
question = '',
|
|
44
|
+
diff_flipped_metrics = [
|
|
45
|
+
eval_studio_client.api.models.flipped_metric_information.Flipped metric information(
|
|
46
|
+
metric_name = '',
|
|
47
|
+
baseline_value = 1.337,
|
|
48
|
+
current_value = 1.337, )
|
|
49
|
+
],
|
|
50
|
+
baseline_test_case_result = eval_studio_client.api.models.test_case_result.Test case result(
|
|
51
|
+
key = '',
|
|
52
|
+
input = '',
|
|
53
|
+
corpus = [
|
|
54
|
+
''
|
|
55
|
+
],
|
|
56
|
+
context = [
|
|
57
|
+
''
|
|
58
|
+
],
|
|
59
|
+
categories = [
|
|
60
|
+
''
|
|
61
|
+
],
|
|
62
|
+
relationships = [
|
|
63
|
+
eval_studio_client.api.models.test_case_relationship_information.Test case relationship information(
|
|
64
|
+
type = '',
|
|
65
|
+
target = '',
|
|
66
|
+
target_type = '', )
|
|
67
|
+
],
|
|
68
|
+
expected_output = '',
|
|
69
|
+
output_constraints = [
|
|
70
|
+
''
|
|
71
|
+
],
|
|
72
|
+
output_condition = '',
|
|
73
|
+
actual_output = '',
|
|
74
|
+
actual_duration = 1.337,
|
|
75
|
+
cost = 1.337,
|
|
76
|
+
model_key = '',
|
|
77
|
+
test_key = '',
|
|
78
|
+
test_case_key = '',
|
|
79
|
+
metrics = [
|
|
80
|
+
eval_studio_client.api.models.metric_information.Metric information(
|
|
81
|
+
key = '',
|
|
82
|
+
value = 1.337, )
|
|
83
|
+
],
|
|
84
|
+
metrics_meta = {
|
|
85
|
+
'key' : ''
|
|
86
|
+
},
|
|
87
|
+
actual_output_meta = [
|
|
88
|
+
eval_studio_client.api.models.actual_output_metadata.Actual output metadata(
|
|
89
|
+
tokenization = '',
|
|
90
|
+
data = [
|
|
91
|
+
eval_studio_client.api.models.data_fragment.Data fragment(
|
|
92
|
+
text = '',
|
|
93
|
+
meta = {
|
|
94
|
+
'key' : ''
|
|
95
|
+
}, )
|
|
96
|
+
], )
|
|
97
|
+
],
|
|
98
|
+
metric_scores = [
|
|
99
|
+
eval_studio_client.api.models.comparison_metric_score_information_(specific_to_comparison_reports).Comparison metric score information (specific to comparison reports)(
|
|
100
|
+
metric_name = '',
|
|
101
|
+
metric_score = 1.337, )
|
|
102
|
+
],
|
|
103
|
+
result_error_message = '', ),
|
|
104
|
+
baseline_diff_actual_output_meta = eval_studio_client.api.models.actual_output_metadata_diff.Actual output metadata diff(
|
|
105
|
+
sentences = [
|
|
106
|
+
''
|
|
107
|
+
],
|
|
108
|
+
sentences_count = 56,
|
|
109
|
+
common_sentences = [
|
|
110
|
+
''
|
|
111
|
+
],
|
|
112
|
+
common_count = 56,
|
|
113
|
+
unique_sentences = [
|
|
114
|
+
''
|
|
115
|
+
],
|
|
116
|
+
unique_count = 56,
|
|
117
|
+
identical = True,
|
|
118
|
+
sentence_similarity = {
|
|
119
|
+
'key' : 1.337
|
|
120
|
+
}, ),
|
|
121
|
+
baseline_diff_retrieved_context = eval_studio_client.api.models.retrieved_context_diff.Retrieved context diff(
|
|
122
|
+
chunks = [
|
|
123
|
+
''
|
|
124
|
+
],
|
|
125
|
+
chunks_count = 56,
|
|
126
|
+
common_chunks = [
|
|
127
|
+
''
|
|
128
|
+
],
|
|
129
|
+
common_count = 56,
|
|
130
|
+
unique_chunks = [
|
|
131
|
+
''
|
|
132
|
+
],
|
|
133
|
+
unique_count = 56,
|
|
134
|
+
identical = True,
|
|
135
|
+
chunk_similarity = {
|
|
136
|
+
'key' : 1.337
|
|
137
|
+
}, ),
|
|
138
|
+
current_test_case_result = eval_studio_client.api.models.test_case_result.Test case result(
|
|
139
|
+
key = '',
|
|
140
|
+
input = '',
|
|
141
|
+
expected_output = '',
|
|
142
|
+
output_condition = '',
|
|
143
|
+
actual_output = '',
|
|
144
|
+
actual_duration = 1.337,
|
|
145
|
+
cost = 1.337,
|
|
146
|
+
model_key = '',
|
|
147
|
+
test_key = '',
|
|
148
|
+
test_case_key = '',
|
|
149
|
+
result_error_message = '', ),
|
|
150
|
+
current_diff_actual_output_meta = eval_studio_client.api.models.actual_output_metadata_diff.Actual output metadata diff(
|
|
151
|
+
sentences_count = 56,
|
|
152
|
+
common_count = 56,
|
|
153
|
+
unique_count = 56,
|
|
154
|
+
identical = True, ),
|
|
155
|
+
current_diff_retrieved_context = eval_studio_client.api.models.retrieved_context_diff.Retrieved context diff(
|
|
156
|
+
chunks_count = 56,
|
|
157
|
+
common_count = 56,
|
|
158
|
+
unique_count = 56,
|
|
159
|
+
identical = True, ), )
|
|
160
|
+
],
|
|
161
|
+
summary = eval_studio_client.api.models.comparison_summary.Comparison summary(
|
|
162
|
+
recommendation_winner = '',
|
|
163
|
+
recommendation = '',
|
|
164
|
+
recommendation_confidence = '', ),
|
|
165
|
+
models_overview = eval_studio_client.api.models.models_overview.Models overview(
|
|
166
|
+
baseline_model_key = '',
|
|
167
|
+
current_model_key = '',
|
|
168
|
+
baseline_model_name = '',
|
|
169
|
+
baseline_collection_id = [
|
|
170
|
+
''
|
|
171
|
+
],
|
|
172
|
+
current_model_name = '',
|
|
173
|
+
current_collection_id = [
|
|
174
|
+
''
|
|
175
|
+
], ),
|
|
176
|
+
models_comparisons = eval_studio_client.api.models.models_comparison_statistics.Models comparison statistics(
|
|
177
|
+
test_case_ranks_baseline = 56,
|
|
178
|
+
test_case_ranks_current = 56,
|
|
179
|
+
test_case_wins_baseline = 56,
|
|
180
|
+
test_case_wins_current = 56, ),
|
|
181
|
+
models_comparisons_metrics = eval_studio_client.api.models.detailed_metrics_comparisons.Detailed metrics comparisons(
|
|
182
|
+
metrics_ranks_baseline = 1.337,
|
|
183
|
+
metrics_ranks_current = 1.337,
|
|
184
|
+
metrics_wins_baseline = 56,
|
|
185
|
+
metrics_wins_current = 56,
|
|
186
|
+
metrics_averages = [
|
|
187
|
+
eval_studio_client.api.models.metric_average_comparison.Metric average comparison(
|
|
188
|
+
metric_key = '',
|
|
189
|
+
baseline_avg = 1.337,
|
|
190
|
+
current_avg = 1.337,
|
|
191
|
+
diff = 1.337,
|
|
192
|
+
baseline_better_wins = 56,
|
|
193
|
+
current_better_wins = 56,
|
|
194
|
+
baseline_rank_avg = 1.337,
|
|
195
|
+
current_rank_avg = 1.337, )
|
|
196
|
+
], ),
|
|
197
|
+
technical_metrics = eval_studio_client.api.models.technical_metrics_for_model_performance.Technical metrics for model performance(
|
|
198
|
+
baseline = eval_studio_client.api.models.technical_metrics_detail.Technical metrics detail(
|
|
199
|
+
cost_sum = 1.337,
|
|
200
|
+
duration_sum = 1.337,
|
|
201
|
+
duration_min = 1.337,
|
|
202
|
+
duration_max = 1.337,
|
|
203
|
+
duration_avg = 1.337, ),
|
|
204
|
+
current = eval_studio_client.api.models.technical_metrics_detail.Technical metrics detail(
|
|
205
|
+
cost_sum = 1.337,
|
|
206
|
+
duration_sum = 1.337,
|
|
207
|
+
duration_min = 1.337,
|
|
208
|
+
duration_max = 1.337,
|
|
209
|
+
duration_avg = 1.337, ), ),
|
|
210
|
+
test_cases_leaderboard = [
|
|
211
|
+
eval_studio_client.api.models.test_case_leaderboard_item.Test case leaderboard item(
|
|
212
|
+
wins = 56,
|
|
213
|
+
question = '',
|
|
214
|
+
changed_metrics_count = 56, )
|
|
215
|
+
], )
|
|
216
|
+
],
|
|
217
|
+
leaderboards = [
|
|
218
|
+
eval_studio_client.api.models.leaderboard_information.Leaderboard information(
|
|
219
|
+
key = '',
|
|
220
|
+
items = [
|
|
221
|
+
eval_studio_client.api.models.leaderboard_comparison_item.Leaderboard comparison item(
|
|
222
|
+
wins = 56,
|
|
223
|
+
question = '',
|
|
224
|
+
changed_metrics_count = 56, )
|
|
225
|
+
], )
|
|
226
|
+
],
|
|
227
|
+
metrics_meta = {
|
|
228
|
+
'key' : eval_studio_client.api.models.metric_metadata.Metric metadata(
|
|
229
|
+
key = '',
|
|
230
|
+
display_name = '',
|
|
231
|
+
data_type = '',
|
|
232
|
+
display_value = '',
|
|
233
|
+
description = '',
|
|
234
|
+
value_range = [
|
|
235
|
+
1.337
|
|
236
|
+
],
|
|
237
|
+
value_enum = [
|
|
238
|
+
''
|
|
239
|
+
],
|
|
240
|
+
higher_is_better = True,
|
|
241
|
+
threshold = 1.337,
|
|
242
|
+
is_primary_metric = True,
|
|
243
|
+
parent_metric = '',
|
|
244
|
+
exclude = True, )
|
|
245
|
+
}
|
|
246
|
+
)
|
|
247
|
+
else:
|
|
248
|
+
return V1ComparisonResult(
|
|
249
|
+
)
|
|
250
|
+
"""
|
|
251
|
+
|
|
252
|
+
def testV1ComparisonResult(self):
|
|
253
|
+
"""Test V1ComparisonResult"""
|
|
254
|
+
# inst_req_only = self.make_instance(include_optional=False)
|
|
255
|
+
# inst_req_and_optional = self.make_instance(include_optional=True)
|
|
256
|
+
|
|
257
|
+
if __name__ == '__main__':
|
|
258
|
+
unittest.main()
|
|
@@ -0,0 +1,53 @@
|
|
|
1
|
+
# coding: utf-8
|
|
2
|
+
|
|
3
|
+
"""
|
|
4
|
+
ai/h2o/eval_studio/v1/insight.proto
|
|
5
|
+
|
|
6
|
+
No description provided (generated by Openapi Generator https://github.com/openapitools/openapi-generator)
|
|
7
|
+
|
|
8
|
+
The version of the OpenAPI document: version not set
|
|
9
|
+
Generated by OpenAPI Generator (https://openapi-generator.tech)
|
|
10
|
+
|
|
11
|
+
Do not edit the class manually.
|
|
12
|
+
""" # noqa: E501
|
|
13
|
+
|
|
14
|
+
|
|
15
|
+
import unittest
|
|
16
|
+
|
|
17
|
+
from eval_studio_client.api.models.v1_comparison_summary import V1ComparisonSummary
|
|
18
|
+
|
|
19
|
+
class TestV1ComparisonSummary(unittest.TestCase):
|
|
20
|
+
"""V1ComparisonSummary unit test stubs"""
|
|
21
|
+
|
|
22
|
+
def setUp(self):
|
|
23
|
+
pass
|
|
24
|
+
|
|
25
|
+
def tearDown(self):
|
|
26
|
+
pass
|
|
27
|
+
|
|
28
|
+
def make_instance(self, include_optional) -> V1ComparisonSummary:
|
|
29
|
+
"""Test V1ComparisonSummary
|
|
30
|
+
include_option is a boolean, when False only required
|
|
31
|
+
params are included, when True both required and
|
|
32
|
+
optional params are included """
|
|
33
|
+
# uncomment below to create an instance of `V1ComparisonSummary`
|
|
34
|
+
"""
|
|
35
|
+
model = V1ComparisonSummary()
|
|
36
|
+
if include_optional:
|
|
37
|
+
return V1ComparisonSummary(
|
|
38
|
+
recommendation_winner = '',
|
|
39
|
+
recommendation = '',
|
|
40
|
+
recommendation_confidence = ''
|
|
41
|
+
)
|
|
42
|
+
else:
|
|
43
|
+
return V1ComparisonSummary(
|
|
44
|
+
)
|
|
45
|
+
"""
|
|
46
|
+
|
|
47
|
+
def testV1ComparisonSummary(self):
|
|
48
|
+
"""Test V1ComparisonSummary"""
|
|
49
|
+
# inst_req_only = self.make_instance(include_optional=False)
|
|
50
|
+
# inst_req_and_optional = self.make_instance(include_optional=True)
|
|
51
|
+
|
|
52
|
+
if __name__ == '__main__':
|
|
53
|
+
unittest.main()
|
|
@@ -113,7 +113,8 @@ class TestV1CreateEvaluationRequest(unittest.TestCase):
|
|
|
113
113
|
api_key = '',
|
|
114
114
|
type = 'MODEL_TYPE_UNSPECIFIED',
|
|
115
115
|
parameters = '',
|
|
116
|
-
demo = True, )
|
|
116
|
+
demo = True, ),
|
|
117
|
+
evaluation_type = 'EVALUATION_TYPE_UNSPECIFIED'
|
|
117
118
|
)
|
|
118
119
|
else:
|
|
119
120
|
return V1CreateEvaluationRequest(
|
|
@@ -98,7 +98,8 @@ class TestV1CreateLeaderboardRequest(unittest.TestCase):
|
|
|
98
98
|
h2ogpte_collection = '',
|
|
99
99
|
type = 'LEADERBOARD_TYPE_UNSPECIFIED',
|
|
100
100
|
demo = True,
|
|
101
|
-
test_lab = '',
|
|
101
|
+
test_lab = '',
|
|
102
|
+
evaluation_type = 'EVALUATION_TYPE_UNSPECIFIED', )
|
|
102
103
|
)
|
|
103
104
|
else:
|
|
104
105
|
return V1CreateLeaderboardRequest(
|
|
@@ -55,7 +55,8 @@ class TestV1CreateLeaderboardResponse(unittest.TestCase):
|
|
|
55
55
|
'key' : None
|
|
56
56
|
}
|
|
57
57
|
], ),
|
|
58
|
-
response = ,
|
|
58
|
+
response = ,
|
|
59
|
+
seen_by_creator_time = datetime.datetime.strptime('2013-10-20 19:20:30.00', '%Y-%m-%d %H:%M:%S.%f'), )
|
|
59
60
|
)
|
|
60
61
|
else:
|
|
61
62
|
return V1CreateLeaderboardResponse(
|
|
@@ -55,7 +55,8 @@ class TestV1CreateLeaderboardWithoutCacheResponse(unittest.TestCase):
|
|
|
55
55
|
'key' : None
|
|
56
56
|
}
|
|
57
57
|
], ),
|
|
58
|
-
response = ,
|
|
58
|
+
response = ,
|
|
59
|
+
seen_by_creator_time = datetime.datetime.strptime('2013-10-20 19:20:30.00', '%Y-%m-%d %H:%M:%S.%f'), )
|
|
59
60
|
)
|
|
60
61
|
else:
|
|
61
62
|
return V1CreateLeaderboardWithoutCacheResponse(
|
|
@@ -0,0 +1,54 @@
|
|
|
1
|
+
# coding: utf-8
|
|
2
|
+
|
|
3
|
+
"""
|
|
4
|
+
ai/h2o/eval_studio/v1/insight.proto
|
|
5
|
+
|
|
6
|
+
No description provided (generated by Openapi Generator https://github.com/openapitools/openapi-generator)
|
|
7
|
+
|
|
8
|
+
The version of the OpenAPI document: version not set
|
|
9
|
+
Generated by OpenAPI Generator (https://openapi-generator.tech)
|
|
10
|
+
|
|
11
|
+
Do not edit the class manually.
|
|
12
|
+
""" # noqa: E501
|
|
13
|
+
|
|
14
|
+
|
|
15
|
+
import unittest
|
|
16
|
+
|
|
17
|
+
from eval_studio_client.api.models.v1_create_test_from_test_cases_request import V1CreateTestFromTestCasesRequest
|
|
18
|
+
|
|
19
|
+
class TestV1CreateTestFromTestCasesRequest(unittest.TestCase):
|
|
20
|
+
"""V1CreateTestFromTestCasesRequest unit test stubs"""
|
|
21
|
+
|
|
22
|
+
def setUp(self):
|
|
23
|
+
pass
|
|
24
|
+
|
|
25
|
+
def tearDown(self):
|
|
26
|
+
pass
|
|
27
|
+
|
|
28
|
+
def make_instance(self, include_optional) -> V1CreateTestFromTestCasesRequest:
|
|
29
|
+
"""Test V1CreateTestFromTestCasesRequest
|
|
30
|
+
include_option is a boolean, when False only required
|
|
31
|
+
params are included, when True both required and
|
|
32
|
+
optional params are included """
|
|
33
|
+
# uncomment below to create an instance of `V1CreateTestFromTestCasesRequest`
|
|
34
|
+
"""
|
|
35
|
+
model = V1CreateTestFromTestCasesRequest()
|
|
36
|
+
if include_optional:
|
|
37
|
+
return V1CreateTestFromTestCasesRequest(
|
|
38
|
+
tests_json = '',
|
|
39
|
+
url = '',
|
|
40
|
+
test_display_name = '',
|
|
41
|
+
test_description = ''
|
|
42
|
+
)
|
|
43
|
+
else:
|
|
44
|
+
return V1CreateTestFromTestCasesRequest(
|
|
45
|
+
)
|
|
46
|
+
"""
|
|
47
|
+
|
|
48
|
+
def testV1CreateTestFromTestCasesRequest(self):
|
|
49
|
+
"""Test V1CreateTestFromTestCasesRequest"""
|
|
50
|
+
# inst_req_only = self.make_instance(include_optional=False)
|
|
51
|
+
# inst_req_and_optional = self.make_instance(include_optional=True)
|
|
52
|
+
|
|
53
|
+
if __name__ == '__main__':
|
|
54
|
+
unittest.main()
|
|
@@ -0,0 +1,68 @@
|
|
|
1
|
+
# coding: utf-8
|
|
2
|
+
|
|
3
|
+
"""
|
|
4
|
+
ai/h2o/eval_studio/v1/insight.proto
|
|
5
|
+
|
|
6
|
+
No description provided (generated by Openapi Generator https://github.com/openapitools/openapi-generator)
|
|
7
|
+
|
|
8
|
+
The version of the OpenAPI document: version not set
|
|
9
|
+
Generated by OpenAPI Generator (https://openapi-generator.tech)
|
|
10
|
+
|
|
11
|
+
Do not edit the class manually.
|
|
12
|
+
""" # noqa: E501
|
|
13
|
+
|
|
14
|
+
|
|
15
|
+
import unittest
|
|
16
|
+
|
|
17
|
+
from eval_studio_client.api.models.v1_create_test_from_test_cases_response import V1CreateTestFromTestCasesResponse
|
|
18
|
+
|
|
19
|
+
class TestV1CreateTestFromTestCasesResponse(unittest.TestCase):
|
|
20
|
+
"""V1CreateTestFromTestCasesResponse unit test stubs"""
|
|
21
|
+
|
|
22
|
+
def setUp(self):
|
|
23
|
+
pass
|
|
24
|
+
|
|
25
|
+
def tearDown(self):
|
|
26
|
+
pass
|
|
27
|
+
|
|
28
|
+
def make_instance(self, include_optional) -> V1CreateTestFromTestCasesResponse:
|
|
29
|
+
"""Test V1CreateTestFromTestCasesResponse
|
|
30
|
+
include_option is a boolean, when False only required
|
|
31
|
+
params are included, when True both required and
|
|
32
|
+
optional params are included """
|
|
33
|
+
# uncomment below to create an instance of `V1CreateTestFromTestCasesResponse`
|
|
34
|
+
"""
|
|
35
|
+
model = V1CreateTestFromTestCasesResponse()
|
|
36
|
+
if include_optional:
|
|
37
|
+
return V1CreateTestFromTestCasesResponse(
|
|
38
|
+
test = eval_studio_client.api.models.v1_test.v1Test(
|
|
39
|
+
name = '',
|
|
40
|
+
create_time = datetime.datetime.strptime('2013-10-20 19:20:30.00', '%Y-%m-%d %H:%M:%S.%f'),
|
|
41
|
+
creator = '',
|
|
42
|
+
update_time = datetime.datetime.strptime('2013-10-20 19:20:30.00', '%Y-%m-%d %H:%M:%S.%f'),
|
|
43
|
+
updater = '',
|
|
44
|
+
delete_time = datetime.datetime.strptime('2013-10-20 19:20:30.00', '%Y-%m-%d %H:%M:%S.%f'),
|
|
45
|
+
deleter = '',
|
|
46
|
+
display_name = '',
|
|
47
|
+
description = '',
|
|
48
|
+
documents = [
|
|
49
|
+
''
|
|
50
|
+
],
|
|
51
|
+
tags = [
|
|
52
|
+
''
|
|
53
|
+
],
|
|
54
|
+
demo = True,
|
|
55
|
+
type = 'TEST_TYPE_UNSPECIFIED', )
|
|
56
|
+
)
|
|
57
|
+
else:
|
|
58
|
+
return V1CreateTestFromTestCasesResponse(
|
|
59
|
+
)
|
|
60
|
+
"""
|
|
61
|
+
|
|
62
|
+
def testV1CreateTestFromTestCasesResponse(self):
|
|
63
|
+
"""Test V1CreateTestFromTestCasesResponse"""
|
|
64
|
+
# inst_req_only = self.make_instance(include_optional=False)
|
|
65
|
+
# inst_req_and_optional = self.make_instance(include_optional=True)
|
|
66
|
+
|
|
67
|
+
if __name__ == '__main__':
|
|
68
|
+
unittest.main()
|
|
@@ -0,0 +1,142 @@
|
|
|
1
|
+
# coding: utf-8
|
|
2
|
+
|
|
3
|
+
"""
|
|
4
|
+
ai/h2o/eval_studio/v1/insight.proto
|
|
5
|
+
|
|
6
|
+
No description provided (generated by Openapi Generator https://github.com/openapitools/openapi-generator)
|
|
7
|
+
|
|
8
|
+
The version of the OpenAPI document: version not set
|
|
9
|
+
Generated by OpenAPI Generator (https://openapi-generator.tech)
|
|
10
|
+
|
|
11
|
+
Do not edit the class manually.
|
|
12
|
+
""" # noqa: E501
|
|
13
|
+
|
|
14
|
+
|
|
15
|
+
import unittest
|
|
16
|
+
|
|
17
|
+
from eval_studio_client.api.models.v1_dashboard_report import V1DashboardReport
|
|
18
|
+
|
|
19
|
+
class TestV1DashboardReport(unittest.TestCase):
|
|
20
|
+
"""V1DashboardReport unit test stubs"""
|
|
21
|
+
|
|
22
|
+
def setUp(self):
|
|
23
|
+
pass
|
|
24
|
+
|
|
25
|
+
def tearDown(self):
|
|
26
|
+
pass
|
|
27
|
+
|
|
28
|
+
def make_instance(self, include_optional) -> V1DashboardReport:
|
|
29
|
+
"""Test V1DashboardReport
|
|
30
|
+
include_option is a boolean, when False only required
|
|
31
|
+
params are included, when True both required and
|
|
32
|
+
optional params are included """
|
|
33
|
+
# uncomment below to create an instance of `V1DashboardReport`
|
|
34
|
+
"""
|
|
35
|
+
model = V1DashboardReport()
|
|
36
|
+
if include_optional:
|
|
37
|
+
return V1DashboardReport(
|
|
38
|
+
results = [
|
|
39
|
+
eval_studio_client.api.models.v1_dashboard_report_result.v1DashboardReportResult(
|
|
40
|
+
key = '',
|
|
41
|
+
input = '',
|
|
42
|
+
expected_output = '',
|
|
43
|
+
actual_output = '',
|
|
44
|
+
model_key = '',
|
|
45
|
+
test_case_key = '',
|
|
46
|
+
metrics = {
|
|
47
|
+
'key' : eval_studio_client.api.models.v1_metric_scores.v1MetricScores(
|
|
48
|
+
scores = [
|
|
49
|
+
eval_studio_client.api.models.v1_metric_score.v1MetricScore(
|
|
50
|
+
key = '',
|
|
51
|
+
value = 1.337, )
|
|
52
|
+
], )
|
|
53
|
+
},
|
|
54
|
+
result_error_map = {
|
|
55
|
+
'key' : ''
|
|
56
|
+
},
|
|
57
|
+
human_decision = 'HUMAN_DECISION_UNSPECIFIED',
|
|
58
|
+
comment = '',
|
|
59
|
+
annotations = {
|
|
60
|
+
'key' : None
|
|
61
|
+
}, )
|
|
62
|
+
],
|
|
63
|
+
evaluator = [
|
|
64
|
+
eval_studio_client.api.models.v1_leaderboard_report_evaluator.v1LeaderboardReportEvaluator(
|
|
65
|
+
id = '',
|
|
66
|
+
name = '',
|
|
67
|
+
display_name = '',
|
|
68
|
+
tagline = '',
|
|
69
|
+
description = '',
|
|
70
|
+
brief_description = '',
|
|
71
|
+
model_types = [
|
|
72
|
+
''
|
|
73
|
+
],
|
|
74
|
+
can_explain = [
|
|
75
|
+
''
|
|
76
|
+
],
|
|
77
|
+
explanation_scopes = [
|
|
78
|
+
''
|
|
79
|
+
],
|
|
80
|
+
explanations = [
|
|
81
|
+
eval_studio_client.api.models.v1_leaderboard_report_explanation.v1LeaderboardReportExplanation(
|
|
82
|
+
explanation_type = '',
|
|
83
|
+
name = '',
|
|
84
|
+
category = '',
|
|
85
|
+
scope = '',
|
|
86
|
+
has_local = '',
|
|
87
|
+
formats = [
|
|
88
|
+
''
|
|
89
|
+
], )
|
|
90
|
+
],
|
|
91
|
+
parameters = [
|
|
92
|
+
eval_studio_client.api.models.v1_leaderboard_report_evaluator_parameter.v1LeaderboardReportEvaluatorParameter(
|
|
93
|
+
name = '',
|
|
94
|
+
description = '',
|
|
95
|
+
comment = '',
|
|
96
|
+
type = '',
|
|
97
|
+
predefined = [
|
|
98
|
+
None
|
|
99
|
+
],
|
|
100
|
+
tags = [
|
|
101
|
+
''
|
|
102
|
+
],
|
|
103
|
+
min = 1.337,
|
|
104
|
+
max = 1.337,
|
|
105
|
+
category = '', )
|
|
106
|
+
],
|
|
107
|
+
keywords = [
|
|
108
|
+
''
|
|
109
|
+
],
|
|
110
|
+
metrics_meta = [
|
|
111
|
+
eval_studio_client.api.models.v1_leaderboard_report_metrics_meta_entry.v1LeaderboardReportMetricsMetaEntry(
|
|
112
|
+
key = '',
|
|
113
|
+
display_name = '',
|
|
114
|
+
data_type = '',
|
|
115
|
+
display_value = '',
|
|
116
|
+
description = '',
|
|
117
|
+
value_range = [
|
|
118
|
+
1.337
|
|
119
|
+
],
|
|
120
|
+
value_enum = [
|
|
121
|
+
''
|
|
122
|
+
],
|
|
123
|
+
higher_is_better = True,
|
|
124
|
+
threshold = 1.337,
|
|
125
|
+
is_primary_metric = True,
|
|
126
|
+
parent_metric = '',
|
|
127
|
+
exclude = True, )
|
|
128
|
+
], )
|
|
129
|
+
]
|
|
130
|
+
)
|
|
131
|
+
else:
|
|
132
|
+
return V1DashboardReport(
|
|
133
|
+
)
|
|
134
|
+
"""
|
|
135
|
+
|
|
136
|
+
def testV1DashboardReport(self):
|
|
137
|
+
"""Test V1DashboardReport"""
|
|
138
|
+
# inst_req_only = self.make_instance(include_optional=False)
|
|
139
|
+
# inst_req_and_optional = self.make_instance(include_optional=True)
|
|
140
|
+
|
|
141
|
+
if __name__ == '__main__':
|
|
142
|
+
unittest.main()
|