eval-studio-client 1.2.4a2__py3-none-any.whl → 1.3.0__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- eval_studio_client/api/__init__.py +65 -0
- eval_studio_client/api/api/__init__.py +3 -0
- eval_studio_client/api/api/dashboard_report_service_api.py +292 -0
- eval_studio_client/api/api/dashboard_service_api.py +16 -16
- eval_studio_client/api/api/dashboard_test_case_annotation_service_api.py +611 -0
- eval_studio_client/api/api/document_service_api.py +16 -16
- eval_studio_client/api/api/evaluation_service_api.py +12 -12
- eval_studio_client/api/api/evaluator_service_api.py +16 -16
- eval_studio_client/api/api/leaderboard_report_service_api.py +304 -17
- eval_studio_client/api/api/leaderboard_service_api.py +554 -16
- eval_studio_client/api/api/leaderboard_test_case_annotation_service_api.py +611 -0
- eval_studio_client/api/api/model_service_api.py +16 -16
- eval_studio_client/api/api/operation_service_api.py +821 -17
- eval_studio_client/api/api/perturbator_service_api.py +22 -22
- eval_studio_client/api/api/test_case_service_api.py +300 -16
- eval_studio_client/api/api/test_class_service_api.py +16 -16
- eval_studio_client/api/api/test_service_api.py +285 -16
- eval_studio_client/api/api/workflow_node_service_api.py +16 -16
- eval_studio_client/api/api/workflow_service_api.py +16 -16
- eval_studio_client/api/docs/AdversarialInputsServiceTestAdversarialInputsRobustnessRequest.md +2 -1
- eval_studio_client/api/docs/DashboardReportServiceApi.md +75 -0
- eval_studio_client/api/docs/DashboardServiceApi.md +5 -5
- eval_studio_client/api/docs/DashboardTestCaseAnnotationServiceApi.md +149 -0
- eval_studio_client/api/docs/DocumentServiceApi.md +5 -5
- eval_studio_client/api/docs/EvaluationServiceApi.md +4 -4
- eval_studio_client/api/docs/EvaluatorServiceApi.md +5 -5
- eval_studio_client/api/docs/LeaderboardReportServiceApi.md +75 -5
- eval_studio_client/api/docs/LeaderboardServiceApi.md +141 -5
- eval_studio_client/api/docs/LeaderboardTestCaseAnnotationServiceApi.md +149 -0
- eval_studio_client/api/docs/ModelServiceApi.md +5 -5
- eval_studio_client/api/docs/OperationServiceApi.md +215 -8
- eval_studio_client/api/docs/PerturbatorServiceApi.md +7 -7
- eval_studio_client/api/docs/RequiredTheDashboardTestCaseAnnotationToUpdate.md +35 -0
- eval_studio_client/api/docs/RequiredTheLeaderboardTestCaseAnnotationToUpdate.md +35 -0
- eval_studio_client/api/docs/RequiredTheLeaderboardToUpdate.md +1 -0
- eval_studio_client/api/docs/RequiredTheOperationToFinalize.md +1 -0
- eval_studio_client/api/docs/RequiredTheOperationToUpdate.md +1 -0
- eval_studio_client/api/docs/TestCaseServiceApi.md +75 -5
- eval_studio_client/api/docs/TestCaseServiceAppendTestCasesRequest.md +30 -0
- eval_studio_client/api/docs/TestClassServiceApi.md +5 -5
- eval_studio_client/api/docs/TestServiceApi.md +73 -5
- eval_studio_client/api/docs/V1ActualOutputMeta.md +30 -0
- eval_studio_client/api/docs/V1ActualOutputMetaDiff.md +36 -0
- eval_studio_client/api/docs/V1AgentChatActivityDiagram.md +31 -0
- eval_studio_client/api/docs/V1AgentChatActivityDiagramEdge.md +32 -0
- eval_studio_client/api/docs/V1AgentChatActivityDiagramNode.md +32 -0
- eval_studio_client/api/docs/V1AgentChatActivityDiagramRow.md +30 -0
- eval_studio_client/api/docs/V1AgentChatScriptUsage.md +33 -0
- eval_studio_client/api/docs/V1AgentChatScriptsBarChart.md +30 -0
- eval_studio_client/api/docs/V1AgentChatToolUsage.md +33 -0
- eval_studio_client/api/docs/V1AgentChatToolsBarChart.md +30 -0
- eval_studio_client/api/docs/V1AllMetricScores.md +29 -0
- eval_studio_client/api/docs/V1AppendTestCasesResponse.md +29 -0
- eval_studio_client/api/docs/V1BatchCreateLeaderboardsWithoutCacheRequest.md +31 -0
- eval_studio_client/api/docs/V1BatchCreateLeaderboardsWithoutCacheResponse.md +29 -0
- eval_studio_client/api/docs/V1BatchMarkOperationSeenByCreatorResponse.md +29 -0
- eval_studio_client/api/docs/V1CmpLeaderboardReportsRequest.md +33 -0
- eval_studio_client/api/docs/V1CmpLeaderboardReportsResponse.md +29 -0
- eval_studio_client/api/docs/V1ComparisonItem.md +36 -0
- eval_studio_client/api/docs/V1ComparisonMetricScore.md +30 -0
- eval_studio_client/api/docs/V1ComparisonResult.md +31 -0
- eval_studio_client/api/docs/V1ComparisonSummary.md +31 -0
- eval_studio_client/api/docs/V1CreateEvaluationRequest.md +1 -0
- eval_studio_client/api/docs/V1CreateTestFromTestCasesRequest.md +32 -0
- eval_studio_client/api/docs/V1CreateTestFromTestCasesResponse.md +29 -0
- eval_studio_client/api/docs/V1DashboardReport.md +31 -0
- eval_studio_client/api/docs/V1DashboardReportResult.md +39 -0
- eval_studio_client/api/docs/V1DashboardTestCaseAnnotation.md +36 -0
- eval_studio_client/api/docs/V1DataFragment.md +31 -0
- eval_studio_client/api/docs/V1DeepCompareLeaderboardsRequest.md +33 -0
- eval_studio_client/api/docs/V1DeepCompareLeaderboardsResponse.md +29 -0
- eval_studio_client/api/docs/V1DiffItem.md +36 -0
- eval_studio_client/api/docs/V1EvaluationType.md +12 -0
- eval_studio_client/api/docs/V1FlippedMetric.md +31 -0
- eval_studio_client/api/docs/V1GetDashboardReportResponse.md +29 -0
- eval_studio_client/api/docs/V1HumanDecision.md +12 -0
- eval_studio_client/api/docs/V1Info.md +1 -0
- eval_studio_client/api/docs/V1Leaderboard.md +1 -0
- eval_studio_client/api/docs/V1LeaderboardCmpReport.md +30 -0
- eval_studio_client/api/docs/V1LeaderboardComparisonItem.md +31 -0
- eval_studio_client/api/docs/V1LeaderboardInfo.md +30 -0
- eval_studio_client/api/docs/V1LeaderboardReportActualOutputMeta.md +6 -3
- eval_studio_client/api/docs/V1LeaderboardReportResult.md +11 -8
- eval_studio_client/api/docs/V1LeaderboardReportResultView.md +12 -0
- eval_studio_client/api/docs/V1LeaderboardTestCaseAnnotation.md +36 -0
- eval_studio_client/api/docs/V1ListDashboardTestCaseAnnotationsResponse.md +29 -0
- eval_studio_client/api/docs/V1ListLeaderboardTestCaseAnnotationsResponse.md +29 -0
- eval_studio_client/api/docs/V1ListOperationsResponse.md +1 -0
- eval_studio_client/api/docs/V1ListUnseenOperationsResponse.md +30 -0
- eval_studio_client/api/docs/V1MarkOperationSeenByCreatorResponse.md +29 -0
- eval_studio_client/api/docs/V1Metric.md +30 -0
- eval_studio_client/api/docs/V1MetricAverage.md +36 -0
- eval_studio_client/api/docs/V1MetricMeta.md +40 -0
- eval_studio_client/api/docs/V1MetricScore.md +1 -1
- eval_studio_client/api/docs/V1MetricScores.md +1 -1
- eval_studio_client/api/docs/V1ModelType.md +1 -1
- eval_studio_client/api/docs/V1ModelsComparisons.md +32 -0
- eval_studio_client/api/docs/V1ModelsComparisonsMetrics.md +33 -0
- eval_studio_client/api/docs/V1ModelsOverview.md +34 -0
- eval_studio_client/api/docs/V1Operation.md +1 -0
- eval_studio_client/api/docs/V1OperationView.md +12 -0
- eval_studio_client/api/docs/V1RetrievedContextDiff.md +36 -0
- eval_studio_client/api/docs/V1Stats.md +2 -0
- eval_studio_client/api/docs/V1TechnicalMetrics.md +30 -0
- eval_studio_client/api/docs/V1TechnicalMetricsDetail.md +33 -0
- eval_studio_client/api/docs/V1TestCaseLeaderboardItem.md +31 -0
- eval_studio_client/api/docs/V1TestCaseRelationshipInfo.md +31 -0
- eval_studio_client/api/docs/V1TestCaseResult.md +48 -0
- eval_studio_client/api/docs/V1TextSimilarityMetric.md +12 -0
- eval_studio_client/api/docs/V1UpdateDashboardTestCaseAnnotationResponse.md +29 -0
- eval_studio_client/api/docs/V1UpdateLeaderboardTestCaseAnnotationResponse.md +29 -0
- eval_studio_client/api/docs/WorkflowNodeServiceApi.md +5 -5
- eval_studio_client/api/docs/WorkflowServiceApi.md +5 -5
- eval_studio_client/api/models/__init__.py +62 -0
- eval_studio_client/api/models/adversarial_inputs_service_test_adversarial_inputs_robustness_request.py +17 -2
- eval_studio_client/api/models/required_the_dashboard_test_case_annotation_to_update.py +108 -0
- eval_studio_client/api/models/required_the_leaderboard_test_case_annotation_to_update.py +108 -0
- eval_studio_client/api/models/required_the_leaderboard_to_update.py +5 -2
- eval_studio_client/api/models/required_the_operation_to_finalize.py +6 -2
- eval_studio_client/api/models/required_the_operation_to_update.py +6 -2
- eval_studio_client/api/models/test_case_service_append_test_cases_request.py +89 -0
- eval_studio_client/api/models/v1_actual_output_meta.py +97 -0
- eval_studio_client/api/models/v1_actual_output_meta_diff.py +101 -0
- eval_studio_client/api/models/v1_agent_chat_activity_diagram.py +109 -0
- eval_studio_client/api/models/v1_agent_chat_activity_diagram_edge.py +97 -0
- eval_studio_client/api/models/v1_agent_chat_activity_diagram_node.py +97 -0
- eval_studio_client/api/models/v1_agent_chat_activity_diagram_row.py +97 -0
- eval_studio_client/api/models/v1_agent_chat_script_usage.py +101 -0
- eval_studio_client/api/models/v1_agent_chat_scripts_bar_chart.py +102 -0
- eval_studio_client/api/models/v1_agent_chat_tool_usage.py +101 -0
- eval_studio_client/api/models/v1_agent_chat_tools_bar_chart.py +102 -0
- eval_studio_client/api/models/v1_all_metric_scores.py +87 -0
- eval_studio_client/api/models/v1_append_test_cases_response.py +95 -0
- eval_studio_client/api/models/v1_batch_create_leaderboards_without_cache_request.py +99 -0
- eval_studio_client/api/models/v1_batch_create_leaderboards_without_cache_response.py +91 -0
- eval_studio_client/api/models/v1_batch_mark_operation_seen_by_creator_response.py +95 -0
- eval_studio_client/api/models/v1_cmp_leaderboard_reports_request.py +96 -0
- eval_studio_client/api/models/v1_cmp_leaderboard_reports_response.py +91 -0
- eval_studio_client/api/models/v1_comparison_item.py +130 -0
- eval_studio_client/api/models/v1_comparison_metric_score.py +89 -0
- eval_studio_client/api/models/v1_comparison_result.py +120 -0
- eval_studio_client/api/models/v1_comparison_summary.py +91 -0
- eval_studio_client/api/models/v1_create_evaluation_request.py +5 -2
- eval_studio_client/api/models/v1_create_test_from_test_cases_request.py +93 -0
- eval_studio_client/api/models/v1_create_test_from_test_cases_response.py +91 -0
- eval_studio_client/api/models/v1_dashboard_report.py +109 -0
- eval_studio_client/api/models/v1_dashboard_report_result.py +139 -0
- eval_studio_client/api/models/v1_dashboard_test_case_annotation.py +112 -0
- eval_studio_client/api/models/v1_data_fragment.py +91 -0
- eval_studio_client/api/models/v1_deep_compare_leaderboards_request.py +96 -0
- eval_studio_client/api/models/v1_deep_compare_leaderboards_response.py +91 -0
- eval_studio_client/api/models/v1_diff_item.py +137 -0
- eval_studio_client/api/models/v1_evaluation_type.py +39 -0
- eval_studio_client/api/models/v1_flipped_metric.py +91 -0
- eval_studio_client/api/models/v1_get_dashboard_report_response.py +91 -0
- eval_studio_client/api/models/v1_human_decision.py +38 -0
- eval_studio_client/api/models/v1_info.py +4 -2
- eval_studio_client/api/models/v1_leaderboard.py +5 -2
- eval_studio_client/api/models/v1_leaderboard_cmp_report.py +93 -0
- eval_studio_client/api/models/v1_leaderboard_comparison_item.py +91 -0
- eval_studio_client/api/models/v1_leaderboard_info.py +97 -0
- eval_studio_client/api/models/v1_leaderboard_report_actual_output_meta.py +23 -9
- eval_studio_client/api/models/v1_leaderboard_report_result.py +21 -10
- eval_studio_client/api/models/v1_leaderboard_report_result_view.py +38 -0
- eval_studio_client/api/models/v1_leaderboard_test_case_annotation.py +112 -0
- eval_studio_client/api/models/v1_list_dashboard_test_case_annotations_response.py +95 -0
- eval_studio_client/api/models/v1_list_leaderboard_test_case_annotations_response.py +95 -0
- eval_studio_client/api/models/v1_list_operations_response.py +5 -3
- eval_studio_client/api/models/v1_list_unseen_operations_response.py +97 -0
- eval_studio_client/api/models/v1_mark_operation_seen_by_creator_response.py +91 -0
- eval_studio_client/api/models/v1_metric.py +89 -0
- eval_studio_client/api/models/v1_metric_average.py +101 -0
- eval_studio_client/api/models/v1_metric_meta.py +109 -0
- eval_studio_client/api/models/v1_metric_score.py +6 -1
- eval_studio_client/api/models/v1_metric_scores.py +1 -1
- eval_studio_client/api/models/v1_model_type.py +2 -1
- eval_studio_client/api/models/v1_models_comparisons.py +93 -0
- eval_studio_client/api/models/v1_models_comparisons_metrics.py +103 -0
- eval_studio_client/api/models/v1_models_overview.py +97 -0
- eval_studio_client/api/models/v1_operation.py +6 -2
- eval_studio_client/api/models/v1_operation_view.py +38 -0
- eval_studio_client/api/models/v1_retrieved_context_diff.py +101 -0
- eval_studio_client/api/models/v1_stats.py +16 -2
- eval_studio_client/api/models/v1_technical_metrics.py +96 -0
- eval_studio_client/api/models/v1_technical_metrics_detail.py +95 -0
- eval_studio_client/api/models/v1_test_case_leaderboard_item.py +91 -0
- eval_studio_client/api/models/v1_test_case_relationship_info.py +91 -0
- eval_studio_client/api/models/v1_test_case_result.py +157 -0
- eval_studio_client/api/models/v1_text_similarity_metric.py +39 -0
- eval_studio_client/api/models/v1_update_dashboard_test_case_annotation_response.py +91 -0
- eval_studio_client/api/models/v1_update_leaderboard_test_case_annotation_response.py +91 -0
- eval_studio_client/api/models/v1_workflow_node_type.py +1 -0
- eval_studio_client/api/models/v1_workflow_type.py +1 -0
- eval_studio_client/api/test/test_adversarial_inputs_service_test_adversarial_inputs_robustness_request.py +6 -0
- eval_studio_client/api/test/test_dashboard_report_service_api.py +37 -0
- eval_studio_client/api/test/test_dashboard_test_case_annotation_service_api.py +43 -0
- eval_studio_client/api/test/test_leaderboard_report_service_api.py +6 -0
- eval_studio_client/api/test/test_leaderboard_service_api.py +12 -0
- eval_studio_client/api/test/test_leaderboard_test_case_annotation_service_api.py +43 -0
- eval_studio_client/api/test/test_operation_service_api.py +18 -0
- eval_studio_client/api/test/test_required_the_dashboard_test_case_annotation_to_update.py +57 -0
- eval_studio_client/api/test/test_required_the_leaderboard_test_case_annotation_to_update.py +57 -0
- eval_studio_client/api/test/test_required_the_leaderboard_to_update.py +2 -1
- eval_studio_client/api/test/test_required_the_operation_to_finalize.py +2 -1
- eval_studio_client/api/test/test_required_the_operation_to_update.py +2 -1
- eval_studio_client/api/test/test_test_case_service_api.py +6 -0
- eval_studio_client/api/test/test_test_case_service_append_test_cases_request.py +52 -0
- eval_studio_client/api/test/test_test_service_api.py +6 -0
- eval_studio_client/api/test/test_v1_abort_operation_response.py +2 -1
- eval_studio_client/api/test/test_v1_actual_output_meta.py +61 -0
- eval_studio_client/api/test/test_v1_actual_output_meta_diff.py +66 -0
- eval_studio_client/api/test/test_v1_agent_chat_activity_diagram.py +65 -0
- eval_studio_client/api/test/test_v1_agent_chat_activity_diagram_edge.py +53 -0
- eval_studio_client/api/test/test_v1_agent_chat_activity_diagram_node.py +53 -0
- eval_studio_client/api/test/test_v1_agent_chat_activity_diagram_row.py +56 -0
- eval_studio_client/api/test/test_v1_agent_chat_script_usage.py +54 -0
- eval_studio_client/api/test/test_v1_agent_chat_scripts_bar_chart.py +57 -0
- eval_studio_client/api/test/test_v1_agent_chat_tool_usage.py +54 -0
- eval_studio_client/api/test/test_v1_agent_chat_tools_bar_chart.py +57 -0
- eval_studio_client/api/test/test_v1_all_metric_scores.py +53 -0
- eval_studio_client/api/test/test_v1_append_test_cases_response.py +74 -0
- eval_studio_client/api/test/test_v1_batch_create_leaderboards_request.py +2 -1
- eval_studio_client/api/test/test_v1_batch_create_leaderboards_response.py +2 -1
- eval_studio_client/api/test/test_v1_batch_create_leaderboards_without_cache_request.py +120 -0
- eval_studio_client/api/test/test_v1_batch_create_leaderboards_without_cache_response.py +72 -0
- eval_studio_client/api/test/test_v1_batch_delete_leaderboards_response.py +2 -1
- eval_studio_client/api/test/test_v1_batch_get_leaderboards_response.py +2 -1
- eval_studio_client/api/test/test_v1_batch_get_operations_response.py +2 -1
- eval_studio_client/api/test/test_v1_batch_import_leaderboard_response.py +2 -1
- eval_studio_client/api/test/test_v1_batch_mark_operation_seen_by_creator_response.py +74 -0
- eval_studio_client/api/test/test_v1_cmp_leaderboard_reports_request.py +55 -0
- eval_studio_client/api/test/test_v1_cmp_leaderboard_reports_response.py +255 -0
- eval_studio_client/api/test/test_v1_comparison_item.py +233 -0
- eval_studio_client/api/test/test_v1_comparison_metric_score.py +52 -0
- eval_studio_client/api/test/test_v1_comparison_result.py +258 -0
- eval_studio_client/api/test/test_v1_comparison_summary.py +53 -0
- eval_studio_client/api/test/test_v1_create_evaluation_request.py +2 -1
- eval_studio_client/api/test/test_v1_create_leaderboard_request.py +2 -1
- eval_studio_client/api/test/test_v1_create_leaderboard_response.py +2 -1
- eval_studio_client/api/test/test_v1_create_leaderboard_without_cache_response.py +2 -1
- eval_studio_client/api/test/test_v1_create_test_from_test_cases_request.py +54 -0
- eval_studio_client/api/test/test_v1_create_test_from_test_cases_response.py +68 -0
- eval_studio_client/api/test/test_v1_dashboard_report.py +142 -0
- eval_studio_client/api/test/test_v1_dashboard_report_result.py +72 -0
- eval_studio_client/api/test/test_v1_dashboard_test_case_annotation.py +58 -0
- eval_studio_client/api/test/test_v1_data_fragment.py +57 -0
- eval_studio_client/api/test/test_v1_deep_compare_leaderboards_request.py +55 -0
- eval_studio_client/api/test/test_v1_deep_compare_leaderboards_response.py +255 -0
- eval_studio_client/api/test/test_v1_delete_leaderboard_response.py +2 -1
- eval_studio_client/api/test/test_v1_diff_item.py +226 -0
- eval_studio_client/api/test/test_v1_evaluation_type.py +33 -0
- eval_studio_client/api/test/test_v1_finalize_operation_response.py +2 -1
- eval_studio_client/api/test/test_v1_flipped_metric.py +53 -0
- eval_studio_client/api/test/test_v1_generate_test_cases_response.py +2 -1
- eval_studio_client/api/test/test_v1_get_dashboard_report_response.py +143 -0
- eval_studio_client/api/test/test_v1_get_info_response.py +4 -1
- eval_studio_client/api/test/test_v1_get_leaderboard_report_response.py +39 -2
- eval_studio_client/api/test/test_v1_get_leaderboard_response.py +2 -1
- eval_studio_client/api/test/test_v1_get_operation_response.py +2 -1
- eval_studio_client/api/test/test_v1_get_stats_response.py +3 -1
- eval_studio_client/api/test/test_v1_human_decision.py +33 -0
- eval_studio_client/api/test/test_v1_import_leaderboard_response.py +2 -1
- eval_studio_client/api/test/test_v1_import_test_cases_from_library_response.py +2 -1
- eval_studio_client/api/test/test_v1_info.py +4 -1
- eval_studio_client/api/test/test_v1_leaderboard.py +2 -1
- eval_studio_client/api/test/test_v1_leaderboard_cmp_report.py +254 -0
- eval_studio_client/api/test/test_v1_leaderboard_comparison_item.py +53 -0
- eval_studio_client/api/test/test_v1_leaderboard_info.py +57 -0
- eval_studio_client/api/test/test_v1_leaderboard_report.py +39 -2
- eval_studio_client/api/test/test_v1_leaderboard_report_actual_output_meta.py +33 -1
- eval_studio_client/api/test/test_v1_leaderboard_report_result.py +39 -2
- eval_studio_client/api/test/test_v1_leaderboard_report_result_view.py +33 -0
- eval_studio_client/api/test/test_v1_leaderboard_test_case_annotation.py +58 -0
- eval_studio_client/api/test/test_v1_list_dashboard_test_case_annotations_response.py +61 -0
- eval_studio_client/api/test/test_v1_list_leaderboard_test_case_annotations_response.py +61 -0
- eval_studio_client/api/test/test_v1_list_leaderboards_response.py +2 -1
- eval_studio_client/api/test/test_v1_list_most_recent_leaderboards_response.py +2 -1
- eval_studio_client/api/test/test_v1_list_operations_response.py +4 -2
- eval_studio_client/api/test/test_v1_list_unseen_operations_response.py +75 -0
- eval_studio_client/api/test/test_v1_mark_operation_seen_by_creator_response.py +72 -0
- eval_studio_client/api/test/test_v1_metric.py +52 -0
- eval_studio_client/api/test/test_v1_metric_average.py +58 -0
- eval_studio_client/api/test/test_v1_metric_meta.py +66 -0
- eval_studio_client/api/test/test_v1_models_comparisons.py +54 -0
- eval_studio_client/api/test/test_v1_models_comparisons_metrics.py +65 -0
- eval_studio_client/api/test/test_v1_models_overview.py +60 -0
- eval_studio_client/api/test/test_v1_operation.py +2 -1
- eval_studio_client/api/test/test_v1_operation_view.py +33 -0
- eval_studio_client/api/test/test_v1_process_workflow_node_response.py +2 -1
- eval_studio_client/api/test/test_v1_retrieved_context_diff.py +66 -0
- eval_studio_client/api/test/test_v1_stats.py +3 -1
- eval_studio_client/api/test/test_v1_technical_metrics.py +62 -0
- eval_studio_client/api/test/test_v1_technical_metrics_detail.py +55 -0
- eval_studio_client/api/test/test_v1_test_case_leaderboard_item.py +53 -0
- eval_studio_client/api/test/test_v1_test_case_relationship_info.py +53 -0
- eval_studio_client/api/test/test_v1_test_case_result.py +106 -0
- eval_studio_client/api/test/test_v1_text_similarity_metric.py +33 -0
- eval_studio_client/api/test/test_v1_update_dashboard_test_case_annotation_response.py +59 -0
- eval_studio_client/api/test/test_v1_update_leaderboard_response.py +2 -1
- eval_studio_client/api/test/test_v1_update_leaderboard_test_case_annotation_response.py +59 -0
- eval_studio_client/api/test/test_v1_update_operation_response.py +2 -1
- eval_studio_client/gen/openapiv2/eval_studio.swagger.json +2340 -210
- eval_studio_client/models.py +18 -6
- {eval_studio_client-1.2.4a2.dist-info → eval_studio_client-1.3.0.dist-info}/METADATA +2 -2
- {eval_studio_client-1.2.4a2.dist-info → eval_studio_client-1.3.0.dist-info}/RECORD +306 -111
- {eval_studio_client-1.2.4a2.dist-info → eval_studio_client-1.3.0.dist-info}/WHEEL +0 -0
|
@@ -19,8 +19,10 @@ from eval_studio_client.api.models.perturbation_service_create_perturbation_requ
|
|
|
19
19
|
from eval_studio_client.api.models.prompt_generation_service_auto_generate_prompts_request import PromptGenerationServiceAutoGeneratePromptsRequest
|
|
20
20
|
from eval_studio_client.api.models.protobuf_any import ProtobufAny
|
|
21
21
|
from eval_studio_client.api.models.protobuf_null_value import ProtobufNullValue
|
|
22
|
+
from eval_studio_client.api.models.required_the_dashboard_test_case_annotation_to_update import RequiredTheDashboardTestCaseAnnotationToUpdate
|
|
22
23
|
from eval_studio_client.api.models.required_the_dashboard_to_update import RequiredTheDashboardToUpdate
|
|
23
24
|
from eval_studio_client.api.models.required_the_document_to_update import RequiredTheDocumentToUpdate
|
|
25
|
+
from eval_studio_client.api.models.required_the_leaderboard_test_case_annotation_to_update import RequiredTheLeaderboardTestCaseAnnotationToUpdate
|
|
24
26
|
from eval_studio_client.api.models.required_the_leaderboard_to_update import RequiredTheLeaderboardToUpdate
|
|
25
27
|
from eval_studio_client.api.models.required_the_model_to_update import RequiredTheModelToUpdate
|
|
26
28
|
from eval_studio_client.api.models.required_the_operation_to_finalize import RequiredTheOperationToFinalize
|
|
@@ -30,6 +32,7 @@ from eval_studio_client.api.models.required_the_test_to_update import RequiredTh
|
|
|
30
32
|
from eval_studio_client.api.models.required_the_updated_workflow import RequiredTheUpdatedWorkflow
|
|
31
33
|
from eval_studio_client.api.models.required_the_updated_workflow_node import RequiredTheUpdatedWorkflowNode
|
|
32
34
|
from eval_studio_client.api.models.rpc_status import RpcStatus
|
|
35
|
+
from eval_studio_client.api.models.test_case_service_append_test_cases_request import TestCaseServiceAppendTestCasesRequest
|
|
33
36
|
from eval_studio_client.api.models.test_case_service_batch_delete_test_cases_request import TestCaseServiceBatchDeleteTestCasesRequest
|
|
34
37
|
from eval_studio_client.api.models.test_service_clone_test_request import TestServiceCloneTestRequest
|
|
35
38
|
from eval_studio_client.api.models.test_service_generate_test_cases_request import TestServiceGenerateTestCasesRequest
|
|
@@ -40,8 +43,22 @@ from eval_studio_client.api.models.test_service_perturb_test_in_place_request im
|
|
|
40
43
|
from eval_studio_client.api.models.test_service_perturb_test_request import TestServicePerturbTestRequest
|
|
41
44
|
from eval_studio_client.api.models.test_service_revoke_test_access_request import TestServiceRevokeTestAccessRequest
|
|
42
45
|
from eval_studio_client.api.models.v1_abort_operation_response import V1AbortOperationResponse
|
|
46
|
+
from eval_studio_client.api.models.v1_actual_output_meta import V1ActualOutputMeta
|
|
47
|
+
from eval_studio_client.api.models.v1_actual_output_meta_diff import V1ActualOutputMetaDiff
|
|
48
|
+
from eval_studio_client.api.models.v1_agent_chat_activity_diagram import V1AgentChatActivityDiagram
|
|
49
|
+
from eval_studio_client.api.models.v1_agent_chat_activity_diagram_edge import V1AgentChatActivityDiagramEdge
|
|
50
|
+
from eval_studio_client.api.models.v1_agent_chat_activity_diagram_node import V1AgentChatActivityDiagramNode
|
|
51
|
+
from eval_studio_client.api.models.v1_agent_chat_activity_diagram_row import V1AgentChatActivityDiagramRow
|
|
52
|
+
from eval_studio_client.api.models.v1_agent_chat_script_usage import V1AgentChatScriptUsage
|
|
53
|
+
from eval_studio_client.api.models.v1_agent_chat_scripts_bar_chart import V1AgentChatScriptsBarChart
|
|
54
|
+
from eval_studio_client.api.models.v1_agent_chat_tool_usage import V1AgentChatToolUsage
|
|
55
|
+
from eval_studio_client.api.models.v1_agent_chat_tools_bar_chart import V1AgentChatToolsBarChart
|
|
56
|
+
from eval_studio_client.api.models.v1_all_metric_scores import V1AllMetricScores
|
|
57
|
+
from eval_studio_client.api.models.v1_append_test_cases_response import V1AppendTestCasesResponse
|
|
43
58
|
from eval_studio_client.api.models.v1_batch_create_leaderboards_request import V1BatchCreateLeaderboardsRequest
|
|
44
59
|
from eval_studio_client.api.models.v1_batch_create_leaderboards_response import V1BatchCreateLeaderboardsResponse
|
|
60
|
+
from eval_studio_client.api.models.v1_batch_create_leaderboards_without_cache_request import V1BatchCreateLeaderboardsWithoutCacheRequest
|
|
61
|
+
from eval_studio_client.api.models.v1_batch_create_leaderboards_without_cache_response import V1BatchCreateLeaderboardsWithoutCacheResponse
|
|
45
62
|
from eval_studio_client.api.models.v1_batch_delete_dashboards_request import V1BatchDeleteDashboardsRequest
|
|
46
63
|
from eval_studio_client.api.models.v1_batch_delete_dashboards_response import V1BatchDeleteDashboardsResponse
|
|
47
64
|
from eval_studio_client.api.models.v1_batch_delete_documents_request import V1BatchDeleteDocumentsRequest
|
|
@@ -69,10 +86,17 @@ from eval_studio_client.api.models.v1_batch_import_leaderboard_request import V1
|
|
|
69
86
|
from eval_studio_client.api.models.v1_batch_import_leaderboard_response import V1BatchImportLeaderboardResponse
|
|
70
87
|
from eval_studio_client.api.models.v1_batch_import_tests_request import V1BatchImportTestsRequest
|
|
71
88
|
from eval_studio_client.api.models.v1_batch_import_tests_response import V1BatchImportTestsResponse
|
|
89
|
+
from eval_studio_client.api.models.v1_batch_mark_operation_seen_by_creator_response import V1BatchMarkOperationSeenByCreatorResponse
|
|
72
90
|
from eval_studio_client.api.models.v1_check_base_models_response import V1CheckBaseModelsResponse
|
|
73
91
|
from eval_studio_client.api.models.v1_clone_test_response import V1CloneTestResponse
|
|
74
92
|
from eval_studio_client.api.models.v1_clone_workflow_response import V1CloneWorkflowResponse
|
|
93
|
+
from eval_studio_client.api.models.v1_cmp_leaderboard_reports_request import V1CmpLeaderboardReportsRequest
|
|
94
|
+
from eval_studio_client.api.models.v1_cmp_leaderboard_reports_response import V1CmpLeaderboardReportsResponse
|
|
75
95
|
from eval_studio_client.api.models.v1_collection_info import V1CollectionInfo
|
|
96
|
+
from eval_studio_client.api.models.v1_comparison_item import V1ComparisonItem
|
|
97
|
+
from eval_studio_client.api.models.v1_comparison_metric_score import V1ComparisonMetricScore
|
|
98
|
+
from eval_studio_client.api.models.v1_comparison_result import V1ComparisonResult
|
|
99
|
+
from eval_studio_client.api.models.v1_comparison_summary import V1ComparisonSummary
|
|
76
100
|
from eval_studio_client.api.models.v1_context import V1Context
|
|
77
101
|
from eval_studio_client.api.models.v1_create_dashboard_response import V1CreateDashboardResponse
|
|
78
102
|
from eval_studio_client.api.models.v1_create_document_response import V1CreateDocumentResponse
|
|
@@ -84,14 +108,22 @@ from eval_studio_client.api.models.v1_create_leaderboard_without_cache_response
|
|
|
84
108
|
from eval_studio_client.api.models.v1_create_model_response import V1CreateModelResponse
|
|
85
109
|
from eval_studio_client.api.models.v1_create_perturbation_response import V1CreatePerturbationResponse
|
|
86
110
|
from eval_studio_client.api.models.v1_create_test_case_response import V1CreateTestCaseResponse
|
|
111
|
+
from eval_studio_client.api.models.v1_create_test_from_test_cases_request import V1CreateTestFromTestCasesRequest
|
|
112
|
+
from eval_studio_client.api.models.v1_create_test_from_test_cases_response import V1CreateTestFromTestCasesResponse
|
|
87
113
|
from eval_studio_client.api.models.v1_create_test_lab_response import V1CreateTestLabResponse
|
|
88
114
|
from eval_studio_client.api.models.v1_create_test_response import V1CreateTestResponse
|
|
89
115
|
from eval_studio_client.api.models.v1_create_workflow_edge_response import V1CreateWorkflowEdgeResponse
|
|
90
116
|
from eval_studio_client.api.models.v1_create_workflow_node_response import V1CreateWorkflowNodeResponse
|
|
91
117
|
from eval_studio_client.api.models.v1_create_workflow_response import V1CreateWorkflowResponse
|
|
92
118
|
from eval_studio_client.api.models.v1_dashboard import V1Dashboard
|
|
119
|
+
from eval_studio_client.api.models.v1_dashboard_report import V1DashboardReport
|
|
120
|
+
from eval_studio_client.api.models.v1_dashboard_report_result import V1DashboardReportResult
|
|
93
121
|
from eval_studio_client.api.models.v1_dashboard_status import V1DashboardStatus
|
|
122
|
+
from eval_studio_client.api.models.v1_dashboard_test_case_annotation import V1DashboardTestCaseAnnotation
|
|
94
123
|
from eval_studio_client.api.models.v1_dashboard_type import V1DashboardType
|
|
124
|
+
from eval_studio_client.api.models.v1_data_fragment import V1DataFragment
|
|
125
|
+
from eval_studio_client.api.models.v1_deep_compare_leaderboards_request import V1DeepCompareLeaderboardsRequest
|
|
126
|
+
from eval_studio_client.api.models.v1_deep_compare_leaderboards_response import V1DeepCompareLeaderboardsResponse
|
|
95
127
|
from eval_studio_client.api.models.v1_delete_dashboard_response import V1DeleteDashboardResponse
|
|
96
128
|
from eval_studio_client.api.models.v1_delete_document_response import V1DeleteDocumentResponse
|
|
97
129
|
from eval_studio_client.api.models.v1_delete_evaluator_response import V1DeleteEvaluatorResponse
|
|
@@ -103,9 +135,11 @@ from eval_studio_client.api.models.v1_delete_workflow_edge_response import V1Del
|
|
|
103
135
|
from eval_studio_client.api.models.v1_delete_workflow_node_response import V1DeleteWorkflowNodeResponse
|
|
104
136
|
from eval_studio_client.api.models.v1_delete_workflow_response import V1DeleteWorkflowResponse
|
|
105
137
|
from eval_studio_client.api.models.v1_dependency_list import V1DependencyList
|
|
138
|
+
from eval_studio_client.api.models.v1_diff_item import V1DiffItem
|
|
106
139
|
from eval_studio_client.api.models.v1_document import V1Document
|
|
107
140
|
from eval_studio_client.api.models.v1_estimate_threshold_request import V1EstimateThresholdRequest
|
|
108
141
|
from eval_studio_client.api.models.v1_evaluation_test import V1EvaluationTest
|
|
142
|
+
from eval_studio_client.api.models.v1_evaluation_type import V1EvaluationType
|
|
109
143
|
from eval_studio_client.api.models.v1_evaluator import V1Evaluator
|
|
110
144
|
from eval_studio_client.api.models.v1_evaluator_param_type import V1EvaluatorParamType
|
|
111
145
|
from eval_studio_client.api.models.v1_evaluator_parameter import V1EvaluatorParameter
|
|
@@ -114,7 +148,9 @@ from eval_studio_client.api.models.v1_finalize_operation_response import V1Final
|
|
|
114
148
|
from eval_studio_client.api.models.v1_find_all_test_cases_by_id_response import V1FindAllTestCasesByIDResponse
|
|
115
149
|
from eval_studio_client.api.models.v1_find_test_lab_response import V1FindTestLabResponse
|
|
116
150
|
from eval_studio_client.api.models.v1_find_workflows_by_collection_id_response import V1FindWorkflowsByCollectionIDResponse
|
|
151
|
+
from eval_studio_client.api.models.v1_flipped_metric import V1FlippedMetric
|
|
117
152
|
from eval_studio_client.api.models.v1_generate_test_cases_response import V1GenerateTestCasesResponse
|
|
153
|
+
from eval_studio_client.api.models.v1_get_dashboard_report_response import V1GetDashboardReportResponse
|
|
118
154
|
from eval_studio_client.api.models.v1_get_dashboard_response import V1GetDashboardResponse
|
|
119
155
|
from eval_studio_client.api.models.v1_get_document_response import V1GetDocumentResponse
|
|
120
156
|
from eval_studio_client.api.models.v1_get_evaluator_response import V1GetEvaluatorResponse
|
|
@@ -137,6 +173,7 @@ from eval_studio_client.api.models.v1_get_workflow_result_corpus_patch_response
|
|
|
137
173
|
from eval_studio_client.api.models.v1_get_workflow_result_report_response import V1GetWorkflowResultReportResponse
|
|
138
174
|
from eval_studio_client.api.models.v1_get_workflow_result_summary_response import V1GetWorkflowResultSummaryResponse
|
|
139
175
|
from eval_studio_client.api.models.v1_get_workflow_result_system_prompt_patch_response import V1GetWorkflowResultSystemPromptPatchResponse
|
|
176
|
+
from eval_studio_client.api.models.v1_human_decision import V1HumanDecision
|
|
140
177
|
from eval_studio_client.api.models.v1_import_evaluation_request import V1ImportEvaluationRequest
|
|
141
178
|
from eval_studio_client.api.models.v1_import_leaderboard_request import V1ImportLeaderboardRequest
|
|
142
179
|
from eval_studio_client.api.models.v1_import_leaderboard_response import V1ImportLeaderboardResponse
|
|
@@ -147,6 +184,9 @@ from eval_studio_client.api.models.v1_init_workflow_node_response import V1InitW
|
|
|
147
184
|
from eval_studio_client.api.models.v1_insight import V1Insight
|
|
148
185
|
from eval_studio_client.api.models.v1_labeled_test_case import V1LabeledTestCase
|
|
149
186
|
from eval_studio_client.api.models.v1_leaderboard import V1Leaderboard
|
|
187
|
+
from eval_studio_client.api.models.v1_leaderboard_cmp_report import V1LeaderboardCmpReport
|
|
188
|
+
from eval_studio_client.api.models.v1_leaderboard_comparison_item import V1LeaderboardComparisonItem
|
|
189
|
+
from eval_studio_client.api.models.v1_leaderboard_info import V1LeaderboardInfo
|
|
150
190
|
from eval_studio_client.api.models.v1_leaderboard_report import V1LeaderboardReport
|
|
151
191
|
from eval_studio_client.api.models.v1_leaderboard_report_actual_output_data import V1LeaderboardReportActualOutputData
|
|
152
192
|
from eval_studio_client.api.models.v1_leaderboard_report_actual_output_meta import V1LeaderboardReportActualOutputMeta
|
|
@@ -157,16 +197,20 @@ from eval_studio_client.api.models.v1_leaderboard_report_metrics_meta_entry impo
|
|
|
157
197
|
from eval_studio_client.api.models.v1_leaderboard_report_model import V1LeaderboardReportModel
|
|
158
198
|
from eval_studio_client.api.models.v1_leaderboard_report_result import V1LeaderboardReportResult
|
|
159
199
|
from eval_studio_client.api.models.v1_leaderboard_report_result_relationship import V1LeaderboardReportResultRelationship
|
|
200
|
+
from eval_studio_client.api.models.v1_leaderboard_report_result_view import V1LeaderboardReportResultView
|
|
160
201
|
from eval_studio_client.api.models.v1_leaderboard_status import V1LeaderboardStatus
|
|
202
|
+
from eval_studio_client.api.models.v1_leaderboard_test_case_annotation import V1LeaderboardTestCaseAnnotation
|
|
161
203
|
from eval_studio_client.api.models.v1_leaderboard_type import V1LeaderboardType
|
|
162
204
|
from eval_studio_client.api.models.v1_leaderboard_view import V1LeaderboardView
|
|
163
205
|
from eval_studio_client.api.models.v1_list_base_models_response import V1ListBaseModelsResponse
|
|
164
206
|
from eval_studio_client.api.models.v1_list_dashboard_access_response import V1ListDashboardAccessResponse
|
|
207
|
+
from eval_studio_client.api.models.v1_list_dashboard_test_case_annotations_response import V1ListDashboardTestCaseAnnotationsResponse
|
|
165
208
|
from eval_studio_client.api.models.v1_list_dashboards_response import V1ListDashboardsResponse
|
|
166
209
|
from eval_studio_client.api.models.v1_list_dashboards_shared_with_me_response import V1ListDashboardsSharedWithMeResponse
|
|
167
210
|
from eval_studio_client.api.models.v1_list_documents_response import V1ListDocumentsResponse
|
|
168
211
|
from eval_studio_client.api.models.v1_list_evaluators_response import V1ListEvaluatorsResponse
|
|
169
212
|
from eval_studio_client.api.models.v1_list_llm_models_response import V1ListLLMModelsResponse
|
|
213
|
+
from eval_studio_client.api.models.v1_list_leaderboard_test_case_annotations_response import V1ListLeaderboardTestCaseAnnotationsResponse
|
|
170
214
|
from eval_studio_client.api.models.v1_list_leaderboards_response import V1ListLeaderboardsResponse
|
|
171
215
|
from eval_studio_client.api.models.v1_list_model_collections_response import V1ListModelCollectionsResponse
|
|
172
216
|
from eval_studio_client.api.models.v1_list_models_response import V1ListModelsResponse
|
|
@@ -185,16 +229,25 @@ from eval_studio_client.api.models.v1_list_test_cases_response import V1ListTest
|
|
|
185
229
|
from eval_studio_client.api.models.v1_list_test_classes_response import V1ListTestClassesResponse
|
|
186
230
|
from eval_studio_client.api.models.v1_list_tests_response import V1ListTestsResponse
|
|
187
231
|
from eval_studio_client.api.models.v1_list_tests_shared_with_me_response import V1ListTestsSharedWithMeResponse
|
|
232
|
+
from eval_studio_client.api.models.v1_list_unseen_operations_response import V1ListUnseenOperationsResponse
|
|
188
233
|
from eval_studio_client.api.models.v1_list_workflow_access_response import V1ListWorkflowAccessResponse
|
|
189
234
|
from eval_studio_client.api.models.v1_list_workflow_dependencies_response import V1ListWorkflowDependenciesResponse
|
|
190
235
|
from eval_studio_client.api.models.v1_list_workflows_response import V1ListWorkflowsResponse
|
|
191
236
|
from eval_studio_client.api.models.v1_list_workflows_shared_with_me_response import V1ListWorkflowsSharedWithMeResponse
|
|
237
|
+
from eval_studio_client.api.models.v1_mark_operation_seen_by_creator_response import V1MarkOperationSeenByCreatorResponse
|
|
238
|
+
from eval_studio_client.api.models.v1_metric import V1Metric
|
|
239
|
+
from eval_studio_client.api.models.v1_metric_average import V1MetricAverage
|
|
240
|
+
from eval_studio_client.api.models.v1_metric_meta import V1MetricMeta
|
|
192
241
|
from eval_studio_client.api.models.v1_metric_score import V1MetricScore
|
|
193
242
|
from eval_studio_client.api.models.v1_metric_scores import V1MetricScores
|
|
194
243
|
from eval_studio_client.api.models.v1_model import V1Model
|
|
195
244
|
from eval_studio_client.api.models.v1_model_type import V1ModelType
|
|
245
|
+
from eval_studio_client.api.models.v1_models_comparisons import V1ModelsComparisons
|
|
246
|
+
from eval_studio_client.api.models.v1_models_comparisons_metrics import V1ModelsComparisonsMetrics
|
|
247
|
+
from eval_studio_client.api.models.v1_models_overview import V1ModelsOverview
|
|
196
248
|
from eval_studio_client.api.models.v1_operation import V1Operation
|
|
197
249
|
from eval_studio_client.api.models.v1_operation_progress import V1OperationProgress
|
|
250
|
+
from eval_studio_client.api.models.v1_operation_view import V1OperationView
|
|
198
251
|
from eval_studio_client.api.models.v1_perturb_test_in_place_response import V1PerturbTestInPlaceResponse
|
|
199
252
|
from eval_studio_client.api.models.v1_perturb_test_response import V1PerturbTestResponse
|
|
200
253
|
from eval_studio_client.api.models.v1_perturbator import V1Perturbator
|
|
@@ -206,21 +259,30 @@ from eval_studio_client.api.models.v1_prompt_library_item import V1PromptLibrary
|
|
|
206
259
|
from eval_studio_client.api.models.v1_repeated_context import V1RepeatedContext
|
|
207
260
|
from eval_studio_client.api.models.v1_repeated_string import V1RepeatedString
|
|
208
261
|
from eval_studio_client.api.models.v1_reset_workflow_node_response import V1ResetWorkflowNodeResponse
|
|
262
|
+
from eval_studio_client.api.models.v1_retrieved_context_diff import V1RetrievedContextDiff
|
|
209
263
|
from eval_studio_client.api.models.v1_role import V1Role
|
|
210
264
|
from eval_studio_client.api.models.v1_role_binding import V1RoleBinding
|
|
211
265
|
from eval_studio_client.api.models.v1_stats import V1Stats
|
|
266
|
+
from eval_studio_client.api.models.v1_technical_metrics import V1TechnicalMetrics
|
|
267
|
+
from eval_studio_client.api.models.v1_technical_metrics_detail import V1TechnicalMetricsDetail
|
|
212
268
|
from eval_studio_client.api.models.v1_test import V1Test
|
|
213
269
|
from eval_studio_client.api.models.v1_test_case import V1TestCase
|
|
270
|
+
from eval_studio_client.api.models.v1_test_case_leaderboard_item import V1TestCaseLeaderboardItem
|
|
214
271
|
from eval_studio_client.api.models.v1_test_case_relationship import V1TestCaseRelationship
|
|
272
|
+
from eval_studio_client.api.models.v1_test_case_relationship_info import V1TestCaseRelationshipInfo
|
|
273
|
+
from eval_studio_client.api.models.v1_test_case_result import V1TestCaseResult
|
|
215
274
|
from eval_studio_client.api.models.v1_test_cases_generator import V1TestCasesGenerator
|
|
216
275
|
from eval_studio_client.api.models.v1_test_class import V1TestClass
|
|
217
276
|
from eval_studio_client.api.models.v1_test_class_type import V1TestClassType
|
|
218
277
|
from eval_studio_client.api.models.v1_test_lab import V1TestLab
|
|
219
278
|
from eval_studio_client.api.models.v1_test_suite_evaluates import V1TestSuiteEvaluates
|
|
220
279
|
from eval_studio_client.api.models.v1_test_type import V1TestType
|
|
280
|
+
from eval_studio_client.api.models.v1_text_similarity_metric import V1TextSimilarityMetric
|
|
221
281
|
from eval_studio_client.api.models.v1_update_dashboard_response import V1UpdateDashboardResponse
|
|
282
|
+
from eval_studio_client.api.models.v1_update_dashboard_test_case_annotation_response import V1UpdateDashboardTestCaseAnnotationResponse
|
|
222
283
|
from eval_studio_client.api.models.v1_update_document_response import V1UpdateDocumentResponse
|
|
223
284
|
from eval_studio_client.api.models.v1_update_leaderboard_response import V1UpdateLeaderboardResponse
|
|
285
|
+
from eval_studio_client.api.models.v1_update_leaderboard_test_case_annotation_response import V1UpdateLeaderboardTestCaseAnnotationResponse
|
|
224
286
|
from eval_studio_client.api.models.v1_update_model_response import V1UpdateModelResponse
|
|
225
287
|
from eval_studio_client.api.models.v1_update_operation_response import V1UpdateOperationResponse
|
|
226
288
|
from eval_studio_client.api.models.v1_update_test_case_response import V1UpdateTestCaseResponse
|
|
@@ -19,6 +19,7 @@ import json
|
|
|
19
19
|
|
|
20
20
|
from pydantic import BaseModel, ConfigDict, Field, StrictInt, StrictStr
|
|
21
21
|
from typing import Any, ClassVar, Dict, List, Optional
|
|
22
|
+
from eval_studio_client.api.models.v1_all_metric_scores import V1AllMetricScores
|
|
22
23
|
from eval_studio_client.api.models.v1_metric_scores import V1MetricScores
|
|
23
24
|
from eval_studio_client.api.models.v1_model import V1Model
|
|
24
25
|
from eval_studio_client.api.models.v1_test_cases_generator import V1TestCasesGenerator
|
|
@@ -45,8 +46,9 @@ class AdversarialInputsServiceTestAdversarialInputsRobustnessRequest(BaseModel):
|
|
|
45
46
|
model_parameters: Optional[StrictStr] = Field(default=None, description="Optional. Parameters overrides for the Model host in JSON format.", alias="modelParameters")
|
|
46
47
|
default_h2ogpte_model: Optional[V1Model] = Field(default=None, alias="defaultH2ogpteModel")
|
|
47
48
|
baseline_eval: Optional[StrictStr] = Field(default=None, description="Required. Baseline evaluation name.", alias="baselineEval")
|
|
48
|
-
baseline_metrics: Optional[Dict[str, V1MetricScores]] = Field(default=None, description="Required. Map of baseline metrics from the evaluator to the metric scores for the evaluator.", alias="baselineMetrics")
|
|
49
|
-
|
|
49
|
+
baseline_metrics: Optional[Dict[str, V1MetricScores]] = Field(default=None, description="Required. Map of baseline metrics from the evaluator to the average metric scores for the evaluator.", alias="baselineMetrics")
|
|
50
|
+
all_baseline_metrics_scores: Optional[Dict[str, V1AllMetricScores]] = Field(default=None, description="Required. Map of baseline metric to all and every test case metric score.", alias="allBaselineMetricsScores")
|
|
51
|
+
__properties: ClassVar[List[str]] = ["operation", "generatorInputTypes", "generatorDocumentUrls", "generatorModel", "generatorBaseLlmModel", "generatorCount", "generatorTopics", "generatorChunks", "generatorH2ogpteCollectionId", "evaluatorIdentifiers", "evaluatorsParameters", "model", "baseLlmModel", "modelParameters", "defaultH2ogpteModel", "baselineEval", "baselineMetrics", "allBaselineMetricsScores"]
|
|
50
52
|
|
|
51
53
|
model_config = ConfigDict(
|
|
52
54
|
populate_by_name=True,
|
|
@@ -103,6 +105,13 @@ class AdversarialInputsServiceTestAdversarialInputsRobustnessRequest(BaseModel):
|
|
|
103
105
|
if self.baseline_metrics[_key_baseline_metrics]:
|
|
104
106
|
_field_dict[_key_baseline_metrics] = self.baseline_metrics[_key_baseline_metrics].to_dict()
|
|
105
107
|
_dict['baselineMetrics'] = _field_dict
|
|
108
|
+
# override the default output from pydantic by calling `to_dict()` of each value in all_baseline_metrics_scores (dict)
|
|
109
|
+
_field_dict = {}
|
|
110
|
+
if self.all_baseline_metrics_scores:
|
|
111
|
+
for _key_all_baseline_metrics_scores in self.all_baseline_metrics_scores:
|
|
112
|
+
if self.all_baseline_metrics_scores[_key_all_baseline_metrics_scores]:
|
|
113
|
+
_field_dict[_key_all_baseline_metrics_scores] = self.all_baseline_metrics_scores[_key_all_baseline_metrics_scores].to_dict()
|
|
114
|
+
_dict['allBaselineMetricsScores'] = _field_dict
|
|
106
115
|
return _dict
|
|
107
116
|
|
|
108
117
|
@classmethod
|
|
@@ -136,6 +145,12 @@ class AdversarialInputsServiceTestAdversarialInputsRobustnessRequest(BaseModel):
|
|
|
136
145
|
for _k, _v in obj["baselineMetrics"].items()
|
|
137
146
|
)
|
|
138
147
|
if obj.get("baselineMetrics") is not None
|
|
148
|
+
else None,
|
|
149
|
+
"allBaselineMetricsScores": dict(
|
|
150
|
+
(_k, V1AllMetricScores.from_dict(_v))
|
|
151
|
+
for _k, _v in obj["allBaselineMetricsScores"].items()
|
|
152
|
+
)
|
|
153
|
+
if obj.get("allBaselineMetricsScores") is not None
|
|
139
154
|
else None
|
|
140
155
|
}, strict=False)
|
|
141
156
|
return _obj
|
|
@@ -0,0 +1,108 @@
|
|
|
1
|
+
# coding: utf-8
|
|
2
|
+
|
|
3
|
+
"""
|
|
4
|
+
ai/h2o/eval_studio/v1/insight.proto
|
|
5
|
+
|
|
6
|
+
No description provided (generated by Openapi Generator https://github.com/openapitools/openapi-generator)
|
|
7
|
+
|
|
8
|
+
The version of the OpenAPI document: version not set
|
|
9
|
+
Generated by OpenAPI Generator (https://openapi-generator.tech)
|
|
10
|
+
|
|
11
|
+
Do not edit the class manually.
|
|
12
|
+
""" # noqa: E501
|
|
13
|
+
|
|
14
|
+
|
|
15
|
+
from __future__ import annotations
|
|
16
|
+
import pprint
|
|
17
|
+
import re # noqa: F401
|
|
18
|
+
import json
|
|
19
|
+
|
|
20
|
+
from datetime import datetime
|
|
21
|
+
from pydantic import BaseModel, ConfigDict, Field, StrictStr
|
|
22
|
+
from typing import Any, ClassVar, Dict, List, Optional
|
|
23
|
+
from typing import Optional, Set
|
|
24
|
+
from typing_extensions import Self
|
|
25
|
+
|
|
26
|
+
class RequiredTheDashboardTestCaseAnnotationToUpdate(BaseModel):
|
|
27
|
+
"""
|
|
28
|
+
RequiredTheDashboardTestCaseAnnotationToUpdate
|
|
29
|
+
""" # noqa: E501
|
|
30
|
+
create_time: Optional[datetime] = Field(default=None, description="Output only. Timestamp when the DashboardTestCaseAnnotation was created.", alias="createTime")
|
|
31
|
+
creator: Optional[StrictStr] = Field(default=None, description="Output only. Name of the user or service that requested creation of the DashboardTestCaseAnnotation.")
|
|
32
|
+
update_time: Optional[datetime] = Field(default=None, description="Output only. Optional. Timestamp when the DashboardTestCaseAnnotation was last updated.", alias="updateTime")
|
|
33
|
+
updater: Optional[StrictStr] = Field(default=None, description="Output only. Optional. Name of the user or service that requested update of the DashboardTestCaseAnnotation.")
|
|
34
|
+
parent: Optional[StrictStr] = Field(default=None, description="Parent Dashboard Test Case resource name. e.g.: \"dashboards/<UUID>/testCases/<UUID>\".")
|
|
35
|
+
key: Optional[StrictStr] = Field(default=None, description="Immutable. Annotation key.")
|
|
36
|
+
value: Optional[Dict[str, Any]] = Field(default=None, description="Annotation value.")
|
|
37
|
+
__properties: ClassVar[List[str]] = ["createTime", "creator", "updateTime", "updater", "parent", "key", "value"]
|
|
38
|
+
|
|
39
|
+
model_config = ConfigDict(
|
|
40
|
+
populate_by_name=True,
|
|
41
|
+
validate_assignment=True,
|
|
42
|
+
protected_namespaces=(),
|
|
43
|
+
)
|
|
44
|
+
|
|
45
|
+
|
|
46
|
+
def to_str(self) -> str:
|
|
47
|
+
"""Returns the string representation of the model using alias"""
|
|
48
|
+
return pprint.pformat(self.model_dump(by_alias=True))
|
|
49
|
+
|
|
50
|
+
def to_json(self) -> str:
|
|
51
|
+
"""Returns the JSON representation of the model using alias"""
|
|
52
|
+
# TODO: pydantic v2: use .model_dump_json(by_alias=True, exclude_unset=True) instead
|
|
53
|
+
return json.dumps(self.to_dict())
|
|
54
|
+
|
|
55
|
+
@classmethod
|
|
56
|
+
def from_json(cls, json_str: str) -> Optional[Self]:
|
|
57
|
+
"""Create an instance of RequiredTheDashboardTestCaseAnnotationToUpdate from a JSON string"""
|
|
58
|
+
return cls.from_dict(json.loads(json_str))
|
|
59
|
+
|
|
60
|
+
def to_dict(self) -> Dict[str, Any]:
|
|
61
|
+
"""Return the dictionary representation of the model using alias.
|
|
62
|
+
|
|
63
|
+
This has the following differences from calling pydantic's
|
|
64
|
+
`self.model_dump(by_alias=True)`:
|
|
65
|
+
|
|
66
|
+
* `None` is only added to the output dict for nullable fields that
|
|
67
|
+
were set at model initialization. Other fields with value `None`
|
|
68
|
+
are ignored.
|
|
69
|
+
* OpenAPI `readOnly` fields are excluded.
|
|
70
|
+
* OpenAPI `readOnly` fields are excluded.
|
|
71
|
+
* OpenAPI `readOnly` fields are excluded.
|
|
72
|
+
* OpenAPI `readOnly` fields are excluded.
|
|
73
|
+
"""
|
|
74
|
+
excluded_fields: Set[str] = set([
|
|
75
|
+
"create_time",
|
|
76
|
+
"creator",
|
|
77
|
+
"update_time",
|
|
78
|
+
"updater",
|
|
79
|
+
])
|
|
80
|
+
|
|
81
|
+
_dict = self.model_dump(
|
|
82
|
+
by_alias=True,
|
|
83
|
+
exclude=excluded_fields,
|
|
84
|
+
exclude_none=True,
|
|
85
|
+
)
|
|
86
|
+
return _dict
|
|
87
|
+
|
|
88
|
+
@classmethod
|
|
89
|
+
def from_dict(cls, obj: Optional[Dict[str, Any]]) -> Optional[Self]:
|
|
90
|
+
"""Create an instance of RequiredTheDashboardTestCaseAnnotationToUpdate from a dict"""
|
|
91
|
+
if obj is None:
|
|
92
|
+
return None
|
|
93
|
+
|
|
94
|
+
if not isinstance(obj, dict):
|
|
95
|
+
return cls.model_validate(obj, strict=False)
|
|
96
|
+
|
|
97
|
+
_obj = cls.model_validate({
|
|
98
|
+
"createTime": obj.get("createTime"),
|
|
99
|
+
"creator": obj.get("creator"),
|
|
100
|
+
"updateTime": obj.get("updateTime"),
|
|
101
|
+
"updater": obj.get("updater"),
|
|
102
|
+
"parent": obj.get("parent"),
|
|
103
|
+
"key": obj.get("key"),
|
|
104
|
+
"value": obj.get("value")
|
|
105
|
+
}, strict=False)
|
|
106
|
+
return _obj
|
|
107
|
+
|
|
108
|
+
|
|
@@ -0,0 +1,108 @@
|
|
|
1
|
+
# coding: utf-8
|
|
2
|
+
|
|
3
|
+
"""
|
|
4
|
+
ai/h2o/eval_studio/v1/insight.proto
|
|
5
|
+
|
|
6
|
+
No description provided (generated by Openapi Generator https://github.com/openapitools/openapi-generator)
|
|
7
|
+
|
|
8
|
+
The version of the OpenAPI document: version not set
|
|
9
|
+
Generated by OpenAPI Generator (https://openapi-generator.tech)
|
|
10
|
+
|
|
11
|
+
Do not edit the class manually.
|
|
12
|
+
""" # noqa: E501
|
|
13
|
+
|
|
14
|
+
|
|
15
|
+
from __future__ import annotations
|
|
16
|
+
import pprint
|
|
17
|
+
import re # noqa: F401
|
|
18
|
+
import json
|
|
19
|
+
|
|
20
|
+
from datetime import datetime
|
|
21
|
+
from pydantic import BaseModel, ConfigDict, Field, StrictStr
|
|
22
|
+
from typing import Any, ClassVar, Dict, List, Optional
|
|
23
|
+
from typing import Optional, Set
|
|
24
|
+
from typing_extensions import Self
|
|
25
|
+
|
|
26
|
+
class RequiredTheLeaderboardTestCaseAnnotationToUpdate(BaseModel):
|
|
27
|
+
"""
|
|
28
|
+
RequiredTheLeaderboardTestCaseAnnotationToUpdate
|
|
29
|
+
""" # noqa: E501
|
|
30
|
+
create_time: Optional[datetime] = Field(default=None, description="Output only. Timestamp when the LeaderboardTestCaseAnnotation was created.", alias="createTime")
|
|
31
|
+
creator: Optional[StrictStr] = Field(default=None, description="Output only. Name of the user or service that requested creation of the LeaderboardTestCaseAnnotation.")
|
|
32
|
+
update_time: Optional[datetime] = Field(default=None, description="Output only. Optional. Timestamp when the LeaderboardTestCaseAnnotation was last updated.", alias="updateTime")
|
|
33
|
+
updater: Optional[StrictStr] = Field(default=None, description="Output only. Optional. Name of the user or service that requested update of the LeaderboardTestCaseAnnotation.")
|
|
34
|
+
parent: Optional[StrictStr] = Field(default=None, description="Parent Leaderboard Test Case resource name. e.g.: \"leaderboards/<UUID>/testCases/<UUID>\".")
|
|
35
|
+
key: Optional[StrictStr] = Field(default=None, description="Immutable. Annotation key.")
|
|
36
|
+
value: Optional[Dict[str, Any]] = Field(default=None, description="Annotation value.")
|
|
37
|
+
__properties: ClassVar[List[str]] = ["createTime", "creator", "updateTime", "updater", "parent", "key", "value"]
|
|
38
|
+
|
|
39
|
+
model_config = ConfigDict(
|
|
40
|
+
populate_by_name=True,
|
|
41
|
+
validate_assignment=True,
|
|
42
|
+
protected_namespaces=(),
|
|
43
|
+
)
|
|
44
|
+
|
|
45
|
+
|
|
46
|
+
def to_str(self) -> str:
|
|
47
|
+
"""Returns the string representation of the model using alias"""
|
|
48
|
+
return pprint.pformat(self.model_dump(by_alias=True))
|
|
49
|
+
|
|
50
|
+
def to_json(self) -> str:
|
|
51
|
+
"""Returns the JSON representation of the model using alias"""
|
|
52
|
+
# TODO: pydantic v2: use .model_dump_json(by_alias=True, exclude_unset=True) instead
|
|
53
|
+
return json.dumps(self.to_dict())
|
|
54
|
+
|
|
55
|
+
@classmethod
|
|
56
|
+
def from_json(cls, json_str: str) -> Optional[Self]:
|
|
57
|
+
"""Create an instance of RequiredTheLeaderboardTestCaseAnnotationToUpdate from a JSON string"""
|
|
58
|
+
return cls.from_dict(json.loads(json_str))
|
|
59
|
+
|
|
60
|
+
def to_dict(self) -> Dict[str, Any]:
|
|
61
|
+
"""Return the dictionary representation of the model using alias.
|
|
62
|
+
|
|
63
|
+
This has the following differences from calling pydantic's
|
|
64
|
+
`self.model_dump(by_alias=True)`:
|
|
65
|
+
|
|
66
|
+
* `None` is only added to the output dict for nullable fields that
|
|
67
|
+
were set at model initialization. Other fields with value `None`
|
|
68
|
+
are ignored.
|
|
69
|
+
* OpenAPI `readOnly` fields are excluded.
|
|
70
|
+
* OpenAPI `readOnly` fields are excluded.
|
|
71
|
+
* OpenAPI `readOnly` fields are excluded.
|
|
72
|
+
* OpenAPI `readOnly` fields are excluded.
|
|
73
|
+
"""
|
|
74
|
+
excluded_fields: Set[str] = set([
|
|
75
|
+
"create_time",
|
|
76
|
+
"creator",
|
|
77
|
+
"update_time",
|
|
78
|
+
"updater",
|
|
79
|
+
])
|
|
80
|
+
|
|
81
|
+
_dict = self.model_dump(
|
|
82
|
+
by_alias=True,
|
|
83
|
+
exclude=excluded_fields,
|
|
84
|
+
exclude_none=True,
|
|
85
|
+
)
|
|
86
|
+
return _dict
|
|
87
|
+
|
|
88
|
+
@classmethod
|
|
89
|
+
def from_dict(cls, obj: Optional[Dict[str, Any]]) -> Optional[Self]:
|
|
90
|
+
"""Create an instance of RequiredTheLeaderboardTestCaseAnnotationToUpdate from a dict"""
|
|
91
|
+
if obj is None:
|
|
92
|
+
return None
|
|
93
|
+
|
|
94
|
+
if not isinstance(obj, dict):
|
|
95
|
+
return cls.model_validate(obj, strict=False)
|
|
96
|
+
|
|
97
|
+
_obj = cls.model_validate({
|
|
98
|
+
"createTime": obj.get("createTime"),
|
|
99
|
+
"creator": obj.get("creator"),
|
|
100
|
+
"updateTime": obj.get("updateTime"),
|
|
101
|
+
"updater": obj.get("updater"),
|
|
102
|
+
"parent": obj.get("parent"),
|
|
103
|
+
"key": obj.get("key"),
|
|
104
|
+
"value": obj.get("value")
|
|
105
|
+
}, strict=False)
|
|
106
|
+
return _obj
|
|
107
|
+
|
|
108
|
+
|
|
@@ -20,6 +20,7 @@ import json
|
|
|
20
20
|
from datetime import datetime
|
|
21
21
|
from pydantic import BaseModel, ConfigDict, Field, StrictBool, StrictStr
|
|
22
22
|
from typing import Any, ClassVar, Dict, List, Optional
|
|
23
|
+
from eval_studio_client.api.models.v1_evaluation_type import V1EvaluationType
|
|
23
24
|
from eval_studio_client.api.models.v1_insight import V1Insight
|
|
24
25
|
from eval_studio_client.api.models.v1_leaderboard_status import V1LeaderboardStatus
|
|
25
26
|
from eval_studio_client.api.models.v1_leaderboard_type import V1LeaderboardType
|
|
@@ -56,7 +57,8 @@ class RequiredTheLeaderboardToUpdate(BaseModel):
|
|
|
56
57
|
type: Optional[V1LeaderboardType] = None
|
|
57
58
|
demo: Optional[StrictBool] = Field(default=None, description="Output only. Whether the Leaderboard is a demo resource or not. Demo resources are read only.")
|
|
58
59
|
test_lab: Optional[StrictStr] = Field(default=None, description="Optional. Resource name of the TestLab if Leaderboard was created from a imported TestLab.", alias="testLab")
|
|
59
|
-
|
|
60
|
+
evaluation_type: Optional[V1EvaluationType] = Field(default=None, alias="evaluationType")
|
|
61
|
+
__properties: ClassVar[List[str]] = ["createTime", "creator", "updateTime", "updater", "deleteTime", "deleter", "displayName", "description", "status", "evaluator", "tests", "model", "createOperation", "leaderboardReport", "leaderboardTable", "leaderboardSummary", "llmModels", "leaderboardProblems", "evaluatorParameters", "insights", "modelParameters", "h2ogpteCollection", "type", "demo", "testLab", "evaluationType"]
|
|
60
62
|
|
|
61
63
|
model_config = ConfigDict(
|
|
62
64
|
populate_by_name=True,
|
|
@@ -171,7 +173,8 @@ class RequiredTheLeaderboardToUpdate(BaseModel):
|
|
|
171
173
|
"h2ogpteCollection": obj.get("h2ogpteCollection"),
|
|
172
174
|
"type": obj.get("type"),
|
|
173
175
|
"demo": obj.get("demo"),
|
|
174
|
-
"testLab": obj.get("testLab")
|
|
176
|
+
"testLab": obj.get("testLab"),
|
|
177
|
+
"evaluationType": obj.get("evaluationType")
|
|
175
178
|
}, strict=False)
|
|
176
179
|
return _obj
|
|
177
180
|
|
|
@@ -39,7 +39,8 @@ class RequiredTheOperationToFinalize(BaseModel):
|
|
|
39
39
|
done: Optional[StrictBool] = Field(default=None, description="If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.")
|
|
40
40
|
error: Optional[RpcStatus] = None
|
|
41
41
|
response: Optional[ProtobufAny] = None
|
|
42
|
-
|
|
42
|
+
seen_by_creator_time: Optional[datetime] = Field(default=None, description="Output only. Optional. Timestamp when the creator marked the Operation as seen. Once set, this field cannot be changed. Set via MarkOperationSeenByCreator method.", alias="seenByCreatorTime")
|
|
43
|
+
__properties: ClassVar[List[str]] = ["createTime", "creator", "updateTime", "updater", "deleteTime", "deleter", "metadata", "done", "error", "response", "seenByCreatorTime"]
|
|
43
44
|
|
|
44
45
|
model_config = ConfigDict(
|
|
45
46
|
populate_by_name=True,
|
|
@@ -77,6 +78,7 @@ class RequiredTheOperationToFinalize(BaseModel):
|
|
|
77
78
|
* OpenAPI `readOnly` fields are excluded.
|
|
78
79
|
* OpenAPI `readOnly` fields are excluded.
|
|
79
80
|
* OpenAPI `readOnly` fields are excluded.
|
|
81
|
+
* OpenAPI `readOnly` fields are excluded.
|
|
80
82
|
"""
|
|
81
83
|
excluded_fields: Set[str] = set([
|
|
82
84
|
"create_time",
|
|
@@ -85,6 +87,7 @@ class RequiredTheOperationToFinalize(BaseModel):
|
|
|
85
87
|
"updater",
|
|
86
88
|
"delete_time",
|
|
87
89
|
"deleter",
|
|
90
|
+
"seen_by_creator_time",
|
|
88
91
|
])
|
|
89
92
|
|
|
90
93
|
_dict = self.model_dump(
|
|
@@ -122,7 +125,8 @@ class RequiredTheOperationToFinalize(BaseModel):
|
|
|
122
125
|
"metadata": ProtobufAny.from_dict(obj["metadata"]) if obj.get("metadata") is not None else None,
|
|
123
126
|
"done": obj.get("done"),
|
|
124
127
|
"error": RpcStatus.from_dict(obj["error"]) if obj.get("error") is not None else None,
|
|
125
|
-
"response": ProtobufAny.from_dict(obj["response"]) if obj.get("response") is not None else None
|
|
128
|
+
"response": ProtobufAny.from_dict(obj["response"]) if obj.get("response") is not None else None,
|
|
129
|
+
"seenByCreatorTime": obj.get("seenByCreatorTime")
|
|
126
130
|
}, strict=False)
|
|
127
131
|
return _obj
|
|
128
132
|
|
|
@@ -39,7 +39,8 @@ class RequiredTheOperationToUpdate(BaseModel):
|
|
|
39
39
|
done: Optional[StrictBool] = Field(default=None, description="If the value is `false`, it means the operation is still in progress. If `true`, the operation is completed, and either `error` or `response` is available.")
|
|
40
40
|
error: Optional[RpcStatus] = None
|
|
41
41
|
response: Optional[ProtobufAny] = None
|
|
42
|
-
|
|
42
|
+
seen_by_creator_time: Optional[datetime] = Field(default=None, description="Output only. Optional. Timestamp when the creator marked the Operation as seen. Once set, this field cannot be changed. Set via MarkOperationSeenByCreator method.", alias="seenByCreatorTime")
|
|
43
|
+
__properties: ClassVar[List[str]] = ["createTime", "creator", "updateTime", "updater", "deleteTime", "deleter", "metadata", "done", "error", "response", "seenByCreatorTime"]
|
|
43
44
|
|
|
44
45
|
model_config = ConfigDict(
|
|
45
46
|
populate_by_name=True,
|
|
@@ -77,6 +78,7 @@ class RequiredTheOperationToUpdate(BaseModel):
|
|
|
77
78
|
* OpenAPI `readOnly` fields are excluded.
|
|
78
79
|
* OpenAPI `readOnly` fields are excluded.
|
|
79
80
|
* OpenAPI `readOnly` fields are excluded.
|
|
81
|
+
* OpenAPI `readOnly` fields are excluded.
|
|
80
82
|
"""
|
|
81
83
|
excluded_fields: Set[str] = set([
|
|
82
84
|
"create_time",
|
|
@@ -85,6 +87,7 @@ class RequiredTheOperationToUpdate(BaseModel):
|
|
|
85
87
|
"updater",
|
|
86
88
|
"delete_time",
|
|
87
89
|
"deleter",
|
|
90
|
+
"seen_by_creator_time",
|
|
88
91
|
])
|
|
89
92
|
|
|
90
93
|
_dict = self.model_dump(
|
|
@@ -122,7 +125,8 @@ class RequiredTheOperationToUpdate(BaseModel):
|
|
|
122
125
|
"metadata": ProtobufAny.from_dict(obj["metadata"]) if obj.get("metadata") is not None else None,
|
|
123
126
|
"done": obj.get("done"),
|
|
124
127
|
"error": RpcStatus.from_dict(obj["error"]) if obj.get("error") is not None else None,
|
|
125
|
-
"response": ProtobufAny.from_dict(obj["response"]) if obj.get("response") is not None else None
|
|
128
|
+
"response": ProtobufAny.from_dict(obj["response"]) if obj.get("response") is not None else None,
|
|
129
|
+
"seenByCreatorTime": obj.get("seenByCreatorTime")
|
|
126
130
|
}, strict=False)
|
|
127
131
|
return _obj
|
|
128
132
|
|