eval-studio-client 1.2.5__py3-none-any.whl → 1.3.0a1__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- eval_studio_client/api/__init__.py +65 -0
- eval_studio_client/api/api/__init__.py +3 -0
- eval_studio_client/api/api/dashboard_report_service_api.py +292 -0
- eval_studio_client/api/api/dashboard_service_api.py +16 -16
- eval_studio_client/api/api/dashboard_test_case_annotation_service_api.py +611 -0
- eval_studio_client/api/api/document_service_api.py +16 -16
- eval_studio_client/api/api/evaluation_service_api.py +12 -12
- eval_studio_client/api/api/evaluator_service_api.py +16 -16
- eval_studio_client/api/api/leaderboard_report_service_api.py +304 -17
- eval_studio_client/api/api/leaderboard_service_api.py +554 -16
- eval_studio_client/api/api/leaderboard_test_case_annotation_service_api.py +611 -0
- eval_studio_client/api/api/model_service_api.py +16 -16
- eval_studio_client/api/api/operation_service_api.py +821 -17
- eval_studio_client/api/api/perturbator_service_api.py +22 -22
- eval_studio_client/api/api/test_case_service_api.py +300 -16
- eval_studio_client/api/api/test_class_service_api.py +16 -16
- eval_studio_client/api/api/test_service_api.py +285 -16
- eval_studio_client/api/api/workflow_node_service_api.py +16 -16
- eval_studio_client/api/api/workflow_service_api.py +16 -16
- eval_studio_client/api/docs/AdversarialInputsServiceTestAdversarialInputsRobustnessRequest.md +2 -1
- eval_studio_client/api/docs/DashboardReportServiceApi.md +75 -0
- eval_studio_client/api/docs/DashboardServiceApi.md +5 -5
- eval_studio_client/api/docs/DashboardTestCaseAnnotationServiceApi.md +149 -0
- eval_studio_client/api/docs/DocumentServiceApi.md +5 -5
- eval_studio_client/api/docs/EvaluationServiceApi.md +4 -4
- eval_studio_client/api/docs/EvaluatorServiceApi.md +5 -5
- eval_studio_client/api/docs/LeaderboardReportServiceApi.md +75 -5
- eval_studio_client/api/docs/LeaderboardServiceApi.md +141 -5
- eval_studio_client/api/docs/LeaderboardTestCaseAnnotationServiceApi.md +149 -0
- eval_studio_client/api/docs/ModelServiceApi.md +5 -5
- eval_studio_client/api/docs/OperationServiceApi.md +215 -8
- eval_studio_client/api/docs/PerturbatorServiceApi.md +7 -7
- eval_studio_client/api/docs/RequiredTheDashboardTestCaseAnnotationToUpdate.md +35 -0
- eval_studio_client/api/docs/RequiredTheLeaderboardTestCaseAnnotationToUpdate.md +35 -0
- eval_studio_client/api/docs/RequiredTheLeaderboardToUpdate.md +1 -0
- eval_studio_client/api/docs/RequiredTheOperationToFinalize.md +1 -0
- eval_studio_client/api/docs/RequiredTheOperationToUpdate.md +1 -0
- eval_studio_client/api/docs/TestCaseServiceApi.md +75 -5
- eval_studio_client/api/docs/TestCaseServiceAppendTestCasesRequest.md +30 -0
- eval_studio_client/api/docs/TestClassServiceApi.md +5 -5
- eval_studio_client/api/docs/TestServiceApi.md +73 -5
- eval_studio_client/api/docs/V1ActualOutputMeta.md +30 -0
- eval_studio_client/api/docs/V1ActualOutputMetaDiff.md +36 -0
- eval_studio_client/api/docs/V1AgentChatActivityDiagram.md +31 -0
- eval_studio_client/api/docs/V1AgentChatActivityDiagramEdge.md +32 -0
- eval_studio_client/api/docs/V1AgentChatActivityDiagramNode.md +32 -0
- eval_studio_client/api/docs/V1AgentChatActivityDiagramRow.md +30 -0
- eval_studio_client/api/docs/V1AgentChatScriptUsage.md +33 -0
- eval_studio_client/api/docs/V1AgentChatScriptsBarChart.md +30 -0
- eval_studio_client/api/docs/V1AgentChatToolUsage.md +33 -0
- eval_studio_client/api/docs/V1AgentChatToolsBarChart.md +30 -0
- eval_studio_client/api/docs/V1AllMetricScores.md +29 -0
- eval_studio_client/api/docs/V1AppendTestCasesResponse.md +29 -0
- eval_studio_client/api/docs/V1BatchCreateLeaderboardsWithoutCacheRequest.md +31 -0
- eval_studio_client/api/docs/V1BatchCreateLeaderboardsWithoutCacheResponse.md +29 -0
- eval_studio_client/api/docs/V1BatchMarkOperationSeenByCreatorResponse.md +29 -0
- eval_studio_client/api/docs/V1CmpLeaderboardReportsRequest.md +33 -0
- eval_studio_client/api/docs/V1CmpLeaderboardReportsResponse.md +29 -0
- eval_studio_client/api/docs/V1ComparisonItem.md +36 -0
- eval_studio_client/api/docs/V1ComparisonMetricScore.md +30 -0
- eval_studio_client/api/docs/V1ComparisonResult.md +31 -0
- eval_studio_client/api/docs/V1ComparisonSummary.md +31 -0
- eval_studio_client/api/docs/V1CreateEvaluationRequest.md +1 -0
- eval_studio_client/api/docs/V1CreateTestFromTestCasesRequest.md +32 -0
- eval_studio_client/api/docs/V1CreateTestFromTestCasesResponse.md +29 -0
- eval_studio_client/api/docs/V1DashboardReport.md +31 -0
- eval_studio_client/api/docs/V1DashboardReportResult.md +39 -0
- eval_studio_client/api/docs/V1DashboardTestCaseAnnotation.md +36 -0
- eval_studio_client/api/docs/V1DataFragment.md +31 -0
- eval_studio_client/api/docs/V1DeepCompareLeaderboardsRequest.md +33 -0
- eval_studio_client/api/docs/V1DeepCompareLeaderboardsResponse.md +29 -0
- eval_studio_client/api/docs/V1DiffItem.md +36 -0
- eval_studio_client/api/docs/V1EvaluationType.md +12 -0
- eval_studio_client/api/docs/V1FlippedMetric.md +31 -0
- eval_studio_client/api/docs/V1GetDashboardReportResponse.md +29 -0
- eval_studio_client/api/docs/V1HumanDecision.md +12 -0
- eval_studio_client/api/docs/V1Info.md +1 -0
- eval_studio_client/api/docs/V1Leaderboard.md +1 -0
- eval_studio_client/api/docs/V1LeaderboardCmpReport.md +30 -0
- eval_studio_client/api/docs/V1LeaderboardComparisonItem.md +31 -0
- eval_studio_client/api/docs/V1LeaderboardInfo.md +30 -0
- eval_studio_client/api/docs/V1LeaderboardReportActualOutputMeta.md +6 -3
- eval_studio_client/api/docs/V1LeaderboardReportResult.md +11 -8
- eval_studio_client/api/docs/V1LeaderboardReportResultView.md +12 -0
- eval_studio_client/api/docs/V1LeaderboardTestCaseAnnotation.md +36 -0
- eval_studio_client/api/docs/V1ListDashboardTestCaseAnnotationsResponse.md +29 -0
- eval_studio_client/api/docs/V1ListLeaderboardTestCaseAnnotationsResponse.md +29 -0
- eval_studio_client/api/docs/V1ListOperationsResponse.md +1 -0
- eval_studio_client/api/docs/V1ListUnseenOperationsResponse.md +30 -0
- eval_studio_client/api/docs/V1MarkOperationSeenByCreatorResponse.md +29 -0
- eval_studio_client/api/docs/V1Metric.md +30 -0
- eval_studio_client/api/docs/V1MetricAverage.md +36 -0
- eval_studio_client/api/docs/V1MetricMeta.md +40 -0
- eval_studio_client/api/docs/V1MetricScore.md +1 -1
- eval_studio_client/api/docs/V1MetricScores.md +1 -1
- eval_studio_client/api/docs/V1ModelType.md +1 -1
- eval_studio_client/api/docs/V1ModelsComparisons.md +32 -0
- eval_studio_client/api/docs/V1ModelsComparisonsMetrics.md +33 -0
- eval_studio_client/api/docs/V1ModelsOverview.md +34 -0
- eval_studio_client/api/docs/V1Operation.md +1 -0
- eval_studio_client/api/docs/V1OperationView.md +12 -0
- eval_studio_client/api/docs/V1RetrievedContextDiff.md +36 -0
- eval_studio_client/api/docs/V1Stats.md +2 -0
- eval_studio_client/api/docs/V1TechnicalMetrics.md +30 -0
- eval_studio_client/api/docs/V1TechnicalMetricsDetail.md +33 -0
- eval_studio_client/api/docs/V1TestCaseLeaderboardItem.md +31 -0
- eval_studio_client/api/docs/V1TestCaseRelationshipInfo.md +31 -0
- eval_studio_client/api/docs/V1TestCaseResult.md +48 -0
- eval_studio_client/api/docs/V1TextSimilarityMetric.md +12 -0
- eval_studio_client/api/docs/V1UpdateDashboardTestCaseAnnotationResponse.md +29 -0
- eval_studio_client/api/docs/V1UpdateLeaderboardTestCaseAnnotationResponse.md +29 -0
- eval_studio_client/api/docs/WorkflowNodeServiceApi.md +5 -5
- eval_studio_client/api/docs/WorkflowServiceApi.md +5 -5
- eval_studio_client/api/models/__init__.py +62 -0
- eval_studio_client/api/models/adversarial_inputs_service_test_adversarial_inputs_robustness_request.py +17 -2
- eval_studio_client/api/models/required_the_dashboard_test_case_annotation_to_update.py +108 -0
- eval_studio_client/api/models/required_the_leaderboard_test_case_annotation_to_update.py +108 -0
- eval_studio_client/api/models/required_the_leaderboard_to_update.py +5 -2
- eval_studio_client/api/models/required_the_operation_to_finalize.py +6 -2
- eval_studio_client/api/models/required_the_operation_to_update.py +6 -2
- eval_studio_client/api/models/test_case_service_append_test_cases_request.py +89 -0
- eval_studio_client/api/models/v1_actual_output_meta.py +97 -0
- eval_studio_client/api/models/v1_actual_output_meta_diff.py +101 -0
- eval_studio_client/api/models/v1_agent_chat_activity_diagram.py +109 -0
- eval_studio_client/api/models/v1_agent_chat_activity_diagram_edge.py +97 -0
- eval_studio_client/api/models/v1_agent_chat_activity_diagram_node.py +97 -0
- eval_studio_client/api/models/v1_agent_chat_activity_diagram_row.py +97 -0
- eval_studio_client/api/models/v1_agent_chat_script_usage.py +101 -0
- eval_studio_client/api/models/v1_agent_chat_scripts_bar_chart.py +102 -0
- eval_studio_client/api/models/v1_agent_chat_tool_usage.py +101 -0
- eval_studio_client/api/models/v1_agent_chat_tools_bar_chart.py +102 -0
- eval_studio_client/api/models/v1_all_metric_scores.py +87 -0
- eval_studio_client/api/models/v1_append_test_cases_response.py +95 -0
- eval_studio_client/api/models/v1_batch_create_leaderboards_without_cache_request.py +99 -0
- eval_studio_client/api/models/v1_batch_create_leaderboards_without_cache_response.py +91 -0
- eval_studio_client/api/models/v1_batch_mark_operation_seen_by_creator_response.py +95 -0
- eval_studio_client/api/models/v1_cmp_leaderboard_reports_request.py +96 -0
- eval_studio_client/api/models/v1_cmp_leaderboard_reports_response.py +91 -0
- eval_studio_client/api/models/v1_comparison_item.py +130 -0
- eval_studio_client/api/models/v1_comparison_metric_score.py +89 -0
- eval_studio_client/api/models/v1_comparison_result.py +120 -0
- eval_studio_client/api/models/v1_comparison_summary.py +91 -0
- eval_studio_client/api/models/v1_create_evaluation_request.py +5 -2
- eval_studio_client/api/models/v1_create_test_from_test_cases_request.py +93 -0
- eval_studio_client/api/models/v1_create_test_from_test_cases_response.py +91 -0
- eval_studio_client/api/models/v1_dashboard_report.py +109 -0
- eval_studio_client/api/models/v1_dashboard_report_result.py +139 -0
- eval_studio_client/api/models/v1_dashboard_test_case_annotation.py +112 -0
- eval_studio_client/api/models/v1_data_fragment.py +91 -0
- eval_studio_client/api/models/v1_deep_compare_leaderboards_request.py +96 -0
- eval_studio_client/api/models/v1_deep_compare_leaderboards_response.py +91 -0
- eval_studio_client/api/models/v1_diff_item.py +137 -0
- eval_studio_client/api/models/v1_evaluation_type.py +39 -0
- eval_studio_client/api/models/v1_flipped_metric.py +91 -0
- eval_studio_client/api/models/v1_get_dashboard_report_response.py +91 -0
- eval_studio_client/api/models/v1_human_decision.py +38 -0
- eval_studio_client/api/models/v1_info.py +4 -2
- eval_studio_client/api/models/v1_leaderboard.py +5 -2
- eval_studio_client/api/models/v1_leaderboard_cmp_report.py +93 -0
- eval_studio_client/api/models/v1_leaderboard_comparison_item.py +91 -0
- eval_studio_client/api/models/v1_leaderboard_info.py +97 -0
- eval_studio_client/api/models/v1_leaderboard_report_actual_output_meta.py +23 -9
- eval_studio_client/api/models/v1_leaderboard_report_result.py +21 -10
- eval_studio_client/api/models/v1_leaderboard_report_result_view.py +38 -0
- eval_studio_client/api/models/v1_leaderboard_test_case_annotation.py +112 -0
- eval_studio_client/api/models/v1_list_dashboard_test_case_annotations_response.py +95 -0
- eval_studio_client/api/models/v1_list_leaderboard_test_case_annotations_response.py +95 -0
- eval_studio_client/api/models/v1_list_operations_response.py +5 -3
- eval_studio_client/api/models/v1_list_unseen_operations_response.py +97 -0
- eval_studio_client/api/models/v1_mark_operation_seen_by_creator_response.py +91 -0
- eval_studio_client/api/models/v1_metric.py +89 -0
- eval_studio_client/api/models/v1_metric_average.py +101 -0
- eval_studio_client/api/models/v1_metric_meta.py +109 -0
- eval_studio_client/api/models/v1_metric_score.py +6 -1
- eval_studio_client/api/models/v1_metric_scores.py +1 -1
- eval_studio_client/api/models/v1_model_type.py +2 -1
- eval_studio_client/api/models/v1_models_comparisons.py +93 -0
- eval_studio_client/api/models/v1_models_comparisons_metrics.py +103 -0
- eval_studio_client/api/models/v1_models_overview.py +97 -0
- eval_studio_client/api/models/v1_operation.py +6 -2
- eval_studio_client/api/models/v1_operation_view.py +38 -0
- eval_studio_client/api/models/v1_retrieved_context_diff.py +101 -0
- eval_studio_client/api/models/v1_stats.py +16 -2
- eval_studio_client/api/models/v1_technical_metrics.py +96 -0
- eval_studio_client/api/models/v1_technical_metrics_detail.py +95 -0
- eval_studio_client/api/models/v1_test_case_leaderboard_item.py +91 -0
- eval_studio_client/api/models/v1_test_case_relationship_info.py +91 -0
- eval_studio_client/api/models/v1_test_case_result.py +157 -0
- eval_studio_client/api/models/v1_text_similarity_metric.py +39 -0
- eval_studio_client/api/models/v1_update_dashboard_test_case_annotation_response.py +91 -0
- eval_studio_client/api/models/v1_update_leaderboard_test_case_annotation_response.py +91 -0
- eval_studio_client/api/models/v1_workflow_node_type.py +1 -0
- eval_studio_client/api/models/v1_workflow_type.py +1 -0
- eval_studio_client/api/test/test_adversarial_inputs_service_test_adversarial_inputs_robustness_request.py +6 -0
- eval_studio_client/api/test/test_dashboard_report_service_api.py +37 -0
- eval_studio_client/api/test/test_dashboard_test_case_annotation_service_api.py +43 -0
- eval_studio_client/api/test/test_leaderboard_report_service_api.py +6 -0
- eval_studio_client/api/test/test_leaderboard_service_api.py +12 -0
- eval_studio_client/api/test/test_leaderboard_test_case_annotation_service_api.py +43 -0
- eval_studio_client/api/test/test_operation_service_api.py +18 -0
- eval_studio_client/api/test/test_required_the_dashboard_test_case_annotation_to_update.py +57 -0
- eval_studio_client/api/test/test_required_the_leaderboard_test_case_annotation_to_update.py +57 -0
- eval_studio_client/api/test/test_required_the_leaderboard_to_update.py +2 -1
- eval_studio_client/api/test/test_required_the_operation_to_finalize.py +2 -1
- eval_studio_client/api/test/test_required_the_operation_to_update.py +2 -1
- eval_studio_client/api/test/test_test_case_service_api.py +6 -0
- eval_studio_client/api/test/test_test_case_service_append_test_cases_request.py +52 -0
- eval_studio_client/api/test/test_test_service_api.py +6 -0
- eval_studio_client/api/test/test_v1_abort_operation_response.py +2 -1
- eval_studio_client/api/test/test_v1_actual_output_meta.py +61 -0
- eval_studio_client/api/test/test_v1_actual_output_meta_diff.py +66 -0
- eval_studio_client/api/test/test_v1_agent_chat_activity_diagram.py +65 -0
- eval_studio_client/api/test/test_v1_agent_chat_activity_diagram_edge.py +53 -0
- eval_studio_client/api/test/test_v1_agent_chat_activity_diagram_node.py +53 -0
- eval_studio_client/api/test/test_v1_agent_chat_activity_diagram_row.py +56 -0
- eval_studio_client/api/test/test_v1_agent_chat_script_usage.py +54 -0
- eval_studio_client/api/test/test_v1_agent_chat_scripts_bar_chart.py +57 -0
- eval_studio_client/api/test/test_v1_agent_chat_tool_usage.py +54 -0
- eval_studio_client/api/test/test_v1_agent_chat_tools_bar_chart.py +57 -0
- eval_studio_client/api/test/test_v1_all_metric_scores.py +53 -0
- eval_studio_client/api/test/test_v1_append_test_cases_response.py +74 -0
- eval_studio_client/api/test/test_v1_batch_create_leaderboards_request.py +2 -1
- eval_studio_client/api/test/test_v1_batch_create_leaderboards_response.py +2 -1
- eval_studio_client/api/test/test_v1_batch_create_leaderboards_without_cache_request.py +120 -0
- eval_studio_client/api/test/test_v1_batch_create_leaderboards_without_cache_response.py +72 -0
- eval_studio_client/api/test/test_v1_batch_delete_leaderboards_response.py +2 -1
- eval_studio_client/api/test/test_v1_batch_get_leaderboards_response.py +2 -1
- eval_studio_client/api/test/test_v1_batch_get_operations_response.py +2 -1
- eval_studio_client/api/test/test_v1_batch_import_leaderboard_response.py +2 -1
- eval_studio_client/api/test/test_v1_batch_mark_operation_seen_by_creator_response.py +74 -0
- eval_studio_client/api/test/test_v1_cmp_leaderboard_reports_request.py +55 -0
- eval_studio_client/api/test/test_v1_cmp_leaderboard_reports_response.py +255 -0
- eval_studio_client/api/test/test_v1_comparison_item.py +233 -0
- eval_studio_client/api/test/test_v1_comparison_metric_score.py +52 -0
- eval_studio_client/api/test/test_v1_comparison_result.py +258 -0
- eval_studio_client/api/test/test_v1_comparison_summary.py +53 -0
- eval_studio_client/api/test/test_v1_create_evaluation_request.py +2 -1
- eval_studio_client/api/test/test_v1_create_leaderboard_request.py +2 -1
- eval_studio_client/api/test/test_v1_create_leaderboard_response.py +2 -1
- eval_studio_client/api/test/test_v1_create_leaderboard_without_cache_response.py +2 -1
- eval_studio_client/api/test/test_v1_create_test_from_test_cases_request.py +54 -0
- eval_studio_client/api/test/test_v1_create_test_from_test_cases_response.py +68 -0
- eval_studio_client/api/test/test_v1_dashboard_report.py +142 -0
- eval_studio_client/api/test/test_v1_dashboard_report_result.py +72 -0
- eval_studio_client/api/test/test_v1_dashboard_test_case_annotation.py +58 -0
- eval_studio_client/api/test/test_v1_data_fragment.py +57 -0
- eval_studio_client/api/test/test_v1_deep_compare_leaderboards_request.py +55 -0
- eval_studio_client/api/test/test_v1_deep_compare_leaderboards_response.py +255 -0
- eval_studio_client/api/test/test_v1_delete_leaderboard_response.py +2 -1
- eval_studio_client/api/test/test_v1_diff_item.py +226 -0
- eval_studio_client/api/test/test_v1_evaluation_type.py +33 -0
- eval_studio_client/api/test/test_v1_finalize_operation_response.py +2 -1
- eval_studio_client/api/test/test_v1_flipped_metric.py +53 -0
- eval_studio_client/api/test/test_v1_generate_test_cases_response.py +2 -1
- eval_studio_client/api/test/test_v1_get_dashboard_report_response.py +143 -0
- eval_studio_client/api/test/test_v1_get_info_response.py +4 -1
- eval_studio_client/api/test/test_v1_get_leaderboard_report_response.py +39 -2
- eval_studio_client/api/test/test_v1_get_leaderboard_response.py +2 -1
- eval_studio_client/api/test/test_v1_get_operation_response.py +2 -1
- eval_studio_client/api/test/test_v1_get_stats_response.py +3 -1
- eval_studio_client/api/test/test_v1_human_decision.py +33 -0
- eval_studio_client/api/test/test_v1_import_leaderboard_response.py +2 -1
- eval_studio_client/api/test/test_v1_import_test_cases_from_library_response.py +2 -1
- eval_studio_client/api/test/test_v1_info.py +4 -1
- eval_studio_client/api/test/test_v1_leaderboard.py +2 -1
- eval_studio_client/api/test/test_v1_leaderboard_cmp_report.py +254 -0
- eval_studio_client/api/test/test_v1_leaderboard_comparison_item.py +53 -0
- eval_studio_client/api/test/test_v1_leaderboard_info.py +57 -0
- eval_studio_client/api/test/test_v1_leaderboard_report.py +39 -2
- eval_studio_client/api/test/test_v1_leaderboard_report_actual_output_meta.py +33 -1
- eval_studio_client/api/test/test_v1_leaderboard_report_result.py +39 -2
- eval_studio_client/api/test/test_v1_leaderboard_report_result_view.py +33 -0
- eval_studio_client/api/test/test_v1_leaderboard_test_case_annotation.py +58 -0
- eval_studio_client/api/test/test_v1_list_dashboard_test_case_annotations_response.py +61 -0
- eval_studio_client/api/test/test_v1_list_leaderboard_test_case_annotations_response.py +61 -0
- eval_studio_client/api/test/test_v1_list_leaderboards_response.py +2 -1
- eval_studio_client/api/test/test_v1_list_most_recent_leaderboards_response.py +2 -1
- eval_studio_client/api/test/test_v1_list_operations_response.py +4 -2
- eval_studio_client/api/test/test_v1_list_unseen_operations_response.py +75 -0
- eval_studio_client/api/test/test_v1_mark_operation_seen_by_creator_response.py +72 -0
- eval_studio_client/api/test/test_v1_metric.py +52 -0
- eval_studio_client/api/test/test_v1_metric_average.py +58 -0
- eval_studio_client/api/test/test_v1_metric_meta.py +66 -0
- eval_studio_client/api/test/test_v1_models_comparisons.py +54 -0
- eval_studio_client/api/test/test_v1_models_comparisons_metrics.py +65 -0
- eval_studio_client/api/test/test_v1_models_overview.py +60 -0
- eval_studio_client/api/test/test_v1_operation.py +2 -1
- eval_studio_client/api/test/test_v1_operation_view.py +33 -0
- eval_studio_client/api/test/test_v1_process_workflow_node_response.py +2 -1
- eval_studio_client/api/test/test_v1_retrieved_context_diff.py +66 -0
- eval_studio_client/api/test/test_v1_stats.py +3 -1
- eval_studio_client/api/test/test_v1_technical_metrics.py +62 -0
- eval_studio_client/api/test/test_v1_technical_metrics_detail.py +55 -0
- eval_studio_client/api/test/test_v1_test_case_leaderboard_item.py +53 -0
- eval_studio_client/api/test/test_v1_test_case_relationship_info.py +53 -0
- eval_studio_client/api/test/test_v1_test_case_result.py +106 -0
- eval_studio_client/api/test/test_v1_text_similarity_metric.py +33 -0
- eval_studio_client/api/test/test_v1_update_dashboard_test_case_annotation_response.py +59 -0
- eval_studio_client/api/test/test_v1_update_leaderboard_response.py +2 -1
- eval_studio_client/api/test/test_v1_update_leaderboard_test_case_annotation_response.py +59 -0
- eval_studio_client/api/test/test_v1_update_operation_response.py +2 -1
- eval_studio_client/gen/openapiv2/eval_studio.swagger.json +2340 -210
- eval_studio_client/models.py +18 -6
- {eval_studio_client-1.2.5.dist-info → eval_studio_client-1.3.0a1.dist-info}/METADATA +2 -2
- {eval_studio_client-1.2.5.dist-info → eval_studio_client-1.3.0a1.dist-info}/RECORD +306 -111
- {eval_studio_client-1.2.5.dist-info → eval_studio_client-1.3.0a1.dist-info}/WHEEL +0 -0
|
@@ -18,7 +18,9 @@ __version__ = "1.0.0"
|
|
|
18
18
|
|
|
19
19
|
# import apis into sdk package
|
|
20
20
|
from eval_studio_client.api.api.adversarial_inputs_service_api import AdversarialInputsServiceApi
|
|
21
|
+
from eval_studio_client.api.api.dashboard_report_service_api import DashboardReportServiceApi
|
|
21
22
|
from eval_studio_client.api.api.dashboard_service_api import DashboardServiceApi
|
|
23
|
+
from eval_studio_client.api.api.dashboard_test_case_annotation_service_api import DashboardTestCaseAnnotationServiceApi
|
|
22
24
|
from eval_studio_client.api.api.document_service_api import DocumentServiceApi
|
|
23
25
|
from eval_studio_client.api.api.evaluation_service_api import EvaluationServiceApi
|
|
24
26
|
from eval_studio_client.api.api.evaluator_service_api import EvaluatorServiceApi
|
|
@@ -26,6 +28,7 @@ from eval_studio_client.api.api.human_calibration_service_api import HumanCalibr
|
|
|
26
28
|
from eval_studio_client.api.api.info_service_api import InfoServiceApi
|
|
27
29
|
from eval_studio_client.api.api.leaderboard_report_service_api import LeaderboardReportServiceApi
|
|
28
30
|
from eval_studio_client.api.api.leaderboard_service_api import LeaderboardServiceApi
|
|
31
|
+
from eval_studio_client.api.api.leaderboard_test_case_annotation_service_api import LeaderboardTestCaseAnnotationServiceApi
|
|
29
32
|
from eval_studio_client.api.api.model_service_api import ModelServiceApi
|
|
30
33
|
from eval_studio_client.api.api.operation_progress_service_api import OperationProgressServiceApi
|
|
31
34
|
from eval_studio_client.api.api.operation_service_api import OperationServiceApi
|
|
@@ -61,8 +64,10 @@ from eval_studio_client.api.models.perturbation_service_create_perturbation_requ
|
|
|
61
64
|
from eval_studio_client.api.models.prompt_generation_service_auto_generate_prompts_request import PromptGenerationServiceAutoGeneratePromptsRequest
|
|
62
65
|
from eval_studio_client.api.models.protobuf_any import ProtobufAny
|
|
63
66
|
from eval_studio_client.api.models.protobuf_null_value import ProtobufNullValue
|
|
67
|
+
from eval_studio_client.api.models.required_the_dashboard_test_case_annotation_to_update import RequiredTheDashboardTestCaseAnnotationToUpdate
|
|
64
68
|
from eval_studio_client.api.models.required_the_dashboard_to_update import RequiredTheDashboardToUpdate
|
|
65
69
|
from eval_studio_client.api.models.required_the_document_to_update import RequiredTheDocumentToUpdate
|
|
70
|
+
from eval_studio_client.api.models.required_the_leaderboard_test_case_annotation_to_update import RequiredTheLeaderboardTestCaseAnnotationToUpdate
|
|
66
71
|
from eval_studio_client.api.models.required_the_leaderboard_to_update import RequiredTheLeaderboardToUpdate
|
|
67
72
|
from eval_studio_client.api.models.required_the_model_to_update import RequiredTheModelToUpdate
|
|
68
73
|
from eval_studio_client.api.models.required_the_operation_to_finalize import RequiredTheOperationToFinalize
|
|
@@ -72,6 +77,7 @@ from eval_studio_client.api.models.required_the_test_to_update import RequiredTh
|
|
|
72
77
|
from eval_studio_client.api.models.required_the_updated_workflow import RequiredTheUpdatedWorkflow
|
|
73
78
|
from eval_studio_client.api.models.required_the_updated_workflow_node import RequiredTheUpdatedWorkflowNode
|
|
74
79
|
from eval_studio_client.api.models.rpc_status import RpcStatus
|
|
80
|
+
from eval_studio_client.api.models.test_case_service_append_test_cases_request import TestCaseServiceAppendTestCasesRequest
|
|
75
81
|
from eval_studio_client.api.models.test_case_service_batch_delete_test_cases_request import TestCaseServiceBatchDeleteTestCasesRequest
|
|
76
82
|
from eval_studio_client.api.models.test_service_clone_test_request import TestServiceCloneTestRequest
|
|
77
83
|
from eval_studio_client.api.models.test_service_generate_test_cases_request import TestServiceGenerateTestCasesRequest
|
|
@@ -82,8 +88,22 @@ from eval_studio_client.api.models.test_service_perturb_test_in_place_request im
|
|
|
82
88
|
from eval_studio_client.api.models.test_service_perturb_test_request import TestServicePerturbTestRequest
|
|
83
89
|
from eval_studio_client.api.models.test_service_revoke_test_access_request import TestServiceRevokeTestAccessRequest
|
|
84
90
|
from eval_studio_client.api.models.v1_abort_operation_response import V1AbortOperationResponse
|
|
91
|
+
from eval_studio_client.api.models.v1_actual_output_meta import V1ActualOutputMeta
|
|
92
|
+
from eval_studio_client.api.models.v1_actual_output_meta_diff import V1ActualOutputMetaDiff
|
|
93
|
+
from eval_studio_client.api.models.v1_agent_chat_activity_diagram import V1AgentChatActivityDiagram
|
|
94
|
+
from eval_studio_client.api.models.v1_agent_chat_activity_diagram_edge import V1AgentChatActivityDiagramEdge
|
|
95
|
+
from eval_studio_client.api.models.v1_agent_chat_activity_diagram_node import V1AgentChatActivityDiagramNode
|
|
96
|
+
from eval_studio_client.api.models.v1_agent_chat_activity_diagram_row import V1AgentChatActivityDiagramRow
|
|
97
|
+
from eval_studio_client.api.models.v1_agent_chat_script_usage import V1AgentChatScriptUsage
|
|
98
|
+
from eval_studio_client.api.models.v1_agent_chat_scripts_bar_chart import V1AgentChatScriptsBarChart
|
|
99
|
+
from eval_studio_client.api.models.v1_agent_chat_tool_usage import V1AgentChatToolUsage
|
|
100
|
+
from eval_studio_client.api.models.v1_agent_chat_tools_bar_chart import V1AgentChatToolsBarChart
|
|
101
|
+
from eval_studio_client.api.models.v1_all_metric_scores import V1AllMetricScores
|
|
102
|
+
from eval_studio_client.api.models.v1_append_test_cases_response import V1AppendTestCasesResponse
|
|
85
103
|
from eval_studio_client.api.models.v1_batch_create_leaderboards_request import V1BatchCreateLeaderboardsRequest
|
|
86
104
|
from eval_studio_client.api.models.v1_batch_create_leaderboards_response import V1BatchCreateLeaderboardsResponse
|
|
105
|
+
from eval_studio_client.api.models.v1_batch_create_leaderboards_without_cache_request import V1BatchCreateLeaderboardsWithoutCacheRequest
|
|
106
|
+
from eval_studio_client.api.models.v1_batch_create_leaderboards_without_cache_response import V1BatchCreateLeaderboardsWithoutCacheResponse
|
|
87
107
|
from eval_studio_client.api.models.v1_batch_delete_dashboards_request import V1BatchDeleteDashboardsRequest
|
|
88
108
|
from eval_studio_client.api.models.v1_batch_delete_dashboards_response import V1BatchDeleteDashboardsResponse
|
|
89
109
|
from eval_studio_client.api.models.v1_batch_delete_documents_request import V1BatchDeleteDocumentsRequest
|
|
@@ -111,10 +131,17 @@ from eval_studio_client.api.models.v1_batch_import_leaderboard_request import V1
|
|
|
111
131
|
from eval_studio_client.api.models.v1_batch_import_leaderboard_response import V1BatchImportLeaderboardResponse
|
|
112
132
|
from eval_studio_client.api.models.v1_batch_import_tests_request import V1BatchImportTestsRequest
|
|
113
133
|
from eval_studio_client.api.models.v1_batch_import_tests_response import V1BatchImportTestsResponse
|
|
134
|
+
from eval_studio_client.api.models.v1_batch_mark_operation_seen_by_creator_response import V1BatchMarkOperationSeenByCreatorResponse
|
|
114
135
|
from eval_studio_client.api.models.v1_check_base_models_response import V1CheckBaseModelsResponse
|
|
115
136
|
from eval_studio_client.api.models.v1_clone_test_response import V1CloneTestResponse
|
|
116
137
|
from eval_studio_client.api.models.v1_clone_workflow_response import V1CloneWorkflowResponse
|
|
138
|
+
from eval_studio_client.api.models.v1_cmp_leaderboard_reports_request import V1CmpLeaderboardReportsRequest
|
|
139
|
+
from eval_studio_client.api.models.v1_cmp_leaderboard_reports_response import V1CmpLeaderboardReportsResponse
|
|
117
140
|
from eval_studio_client.api.models.v1_collection_info import V1CollectionInfo
|
|
141
|
+
from eval_studio_client.api.models.v1_comparison_item import V1ComparisonItem
|
|
142
|
+
from eval_studio_client.api.models.v1_comparison_metric_score import V1ComparisonMetricScore
|
|
143
|
+
from eval_studio_client.api.models.v1_comparison_result import V1ComparisonResult
|
|
144
|
+
from eval_studio_client.api.models.v1_comparison_summary import V1ComparisonSummary
|
|
118
145
|
from eval_studio_client.api.models.v1_context import V1Context
|
|
119
146
|
from eval_studio_client.api.models.v1_create_dashboard_response import V1CreateDashboardResponse
|
|
120
147
|
from eval_studio_client.api.models.v1_create_document_response import V1CreateDocumentResponse
|
|
@@ -126,14 +153,22 @@ from eval_studio_client.api.models.v1_create_leaderboard_without_cache_response
|
|
|
126
153
|
from eval_studio_client.api.models.v1_create_model_response import V1CreateModelResponse
|
|
127
154
|
from eval_studio_client.api.models.v1_create_perturbation_response import V1CreatePerturbationResponse
|
|
128
155
|
from eval_studio_client.api.models.v1_create_test_case_response import V1CreateTestCaseResponse
|
|
156
|
+
from eval_studio_client.api.models.v1_create_test_from_test_cases_request import V1CreateTestFromTestCasesRequest
|
|
157
|
+
from eval_studio_client.api.models.v1_create_test_from_test_cases_response import V1CreateTestFromTestCasesResponse
|
|
129
158
|
from eval_studio_client.api.models.v1_create_test_lab_response import V1CreateTestLabResponse
|
|
130
159
|
from eval_studio_client.api.models.v1_create_test_response import V1CreateTestResponse
|
|
131
160
|
from eval_studio_client.api.models.v1_create_workflow_edge_response import V1CreateWorkflowEdgeResponse
|
|
132
161
|
from eval_studio_client.api.models.v1_create_workflow_node_response import V1CreateWorkflowNodeResponse
|
|
133
162
|
from eval_studio_client.api.models.v1_create_workflow_response import V1CreateWorkflowResponse
|
|
134
163
|
from eval_studio_client.api.models.v1_dashboard import V1Dashboard
|
|
164
|
+
from eval_studio_client.api.models.v1_dashboard_report import V1DashboardReport
|
|
165
|
+
from eval_studio_client.api.models.v1_dashboard_report_result import V1DashboardReportResult
|
|
135
166
|
from eval_studio_client.api.models.v1_dashboard_status import V1DashboardStatus
|
|
167
|
+
from eval_studio_client.api.models.v1_dashboard_test_case_annotation import V1DashboardTestCaseAnnotation
|
|
136
168
|
from eval_studio_client.api.models.v1_dashboard_type import V1DashboardType
|
|
169
|
+
from eval_studio_client.api.models.v1_data_fragment import V1DataFragment
|
|
170
|
+
from eval_studio_client.api.models.v1_deep_compare_leaderboards_request import V1DeepCompareLeaderboardsRequest
|
|
171
|
+
from eval_studio_client.api.models.v1_deep_compare_leaderboards_response import V1DeepCompareLeaderboardsResponse
|
|
137
172
|
from eval_studio_client.api.models.v1_delete_dashboard_response import V1DeleteDashboardResponse
|
|
138
173
|
from eval_studio_client.api.models.v1_delete_document_response import V1DeleteDocumentResponse
|
|
139
174
|
from eval_studio_client.api.models.v1_delete_evaluator_response import V1DeleteEvaluatorResponse
|
|
@@ -145,9 +180,11 @@ from eval_studio_client.api.models.v1_delete_workflow_edge_response import V1Del
|
|
|
145
180
|
from eval_studio_client.api.models.v1_delete_workflow_node_response import V1DeleteWorkflowNodeResponse
|
|
146
181
|
from eval_studio_client.api.models.v1_delete_workflow_response import V1DeleteWorkflowResponse
|
|
147
182
|
from eval_studio_client.api.models.v1_dependency_list import V1DependencyList
|
|
183
|
+
from eval_studio_client.api.models.v1_diff_item import V1DiffItem
|
|
148
184
|
from eval_studio_client.api.models.v1_document import V1Document
|
|
149
185
|
from eval_studio_client.api.models.v1_estimate_threshold_request import V1EstimateThresholdRequest
|
|
150
186
|
from eval_studio_client.api.models.v1_evaluation_test import V1EvaluationTest
|
|
187
|
+
from eval_studio_client.api.models.v1_evaluation_type import V1EvaluationType
|
|
151
188
|
from eval_studio_client.api.models.v1_evaluator import V1Evaluator
|
|
152
189
|
from eval_studio_client.api.models.v1_evaluator_param_type import V1EvaluatorParamType
|
|
153
190
|
from eval_studio_client.api.models.v1_evaluator_parameter import V1EvaluatorParameter
|
|
@@ -156,7 +193,9 @@ from eval_studio_client.api.models.v1_finalize_operation_response import V1Final
|
|
|
156
193
|
from eval_studio_client.api.models.v1_find_all_test_cases_by_id_response import V1FindAllTestCasesByIDResponse
|
|
157
194
|
from eval_studio_client.api.models.v1_find_test_lab_response import V1FindTestLabResponse
|
|
158
195
|
from eval_studio_client.api.models.v1_find_workflows_by_collection_id_response import V1FindWorkflowsByCollectionIDResponse
|
|
196
|
+
from eval_studio_client.api.models.v1_flipped_metric import V1FlippedMetric
|
|
159
197
|
from eval_studio_client.api.models.v1_generate_test_cases_response import V1GenerateTestCasesResponse
|
|
198
|
+
from eval_studio_client.api.models.v1_get_dashboard_report_response import V1GetDashboardReportResponse
|
|
160
199
|
from eval_studio_client.api.models.v1_get_dashboard_response import V1GetDashboardResponse
|
|
161
200
|
from eval_studio_client.api.models.v1_get_document_response import V1GetDocumentResponse
|
|
162
201
|
from eval_studio_client.api.models.v1_get_evaluator_response import V1GetEvaluatorResponse
|
|
@@ -179,6 +218,7 @@ from eval_studio_client.api.models.v1_get_workflow_result_corpus_patch_response
|
|
|
179
218
|
from eval_studio_client.api.models.v1_get_workflow_result_report_response import V1GetWorkflowResultReportResponse
|
|
180
219
|
from eval_studio_client.api.models.v1_get_workflow_result_summary_response import V1GetWorkflowResultSummaryResponse
|
|
181
220
|
from eval_studio_client.api.models.v1_get_workflow_result_system_prompt_patch_response import V1GetWorkflowResultSystemPromptPatchResponse
|
|
221
|
+
from eval_studio_client.api.models.v1_human_decision import V1HumanDecision
|
|
182
222
|
from eval_studio_client.api.models.v1_import_evaluation_request import V1ImportEvaluationRequest
|
|
183
223
|
from eval_studio_client.api.models.v1_import_leaderboard_request import V1ImportLeaderboardRequest
|
|
184
224
|
from eval_studio_client.api.models.v1_import_leaderboard_response import V1ImportLeaderboardResponse
|
|
@@ -189,6 +229,9 @@ from eval_studio_client.api.models.v1_init_workflow_node_response import V1InitW
|
|
|
189
229
|
from eval_studio_client.api.models.v1_insight import V1Insight
|
|
190
230
|
from eval_studio_client.api.models.v1_labeled_test_case import V1LabeledTestCase
|
|
191
231
|
from eval_studio_client.api.models.v1_leaderboard import V1Leaderboard
|
|
232
|
+
from eval_studio_client.api.models.v1_leaderboard_cmp_report import V1LeaderboardCmpReport
|
|
233
|
+
from eval_studio_client.api.models.v1_leaderboard_comparison_item import V1LeaderboardComparisonItem
|
|
234
|
+
from eval_studio_client.api.models.v1_leaderboard_info import V1LeaderboardInfo
|
|
192
235
|
from eval_studio_client.api.models.v1_leaderboard_report import V1LeaderboardReport
|
|
193
236
|
from eval_studio_client.api.models.v1_leaderboard_report_actual_output_data import V1LeaderboardReportActualOutputData
|
|
194
237
|
from eval_studio_client.api.models.v1_leaderboard_report_actual_output_meta import V1LeaderboardReportActualOutputMeta
|
|
@@ -199,16 +242,20 @@ from eval_studio_client.api.models.v1_leaderboard_report_metrics_meta_entry impo
|
|
|
199
242
|
from eval_studio_client.api.models.v1_leaderboard_report_model import V1LeaderboardReportModel
|
|
200
243
|
from eval_studio_client.api.models.v1_leaderboard_report_result import V1LeaderboardReportResult
|
|
201
244
|
from eval_studio_client.api.models.v1_leaderboard_report_result_relationship import V1LeaderboardReportResultRelationship
|
|
245
|
+
from eval_studio_client.api.models.v1_leaderboard_report_result_view import V1LeaderboardReportResultView
|
|
202
246
|
from eval_studio_client.api.models.v1_leaderboard_status import V1LeaderboardStatus
|
|
247
|
+
from eval_studio_client.api.models.v1_leaderboard_test_case_annotation import V1LeaderboardTestCaseAnnotation
|
|
203
248
|
from eval_studio_client.api.models.v1_leaderboard_type import V1LeaderboardType
|
|
204
249
|
from eval_studio_client.api.models.v1_leaderboard_view import V1LeaderboardView
|
|
205
250
|
from eval_studio_client.api.models.v1_list_base_models_response import V1ListBaseModelsResponse
|
|
206
251
|
from eval_studio_client.api.models.v1_list_dashboard_access_response import V1ListDashboardAccessResponse
|
|
252
|
+
from eval_studio_client.api.models.v1_list_dashboard_test_case_annotations_response import V1ListDashboardTestCaseAnnotationsResponse
|
|
207
253
|
from eval_studio_client.api.models.v1_list_dashboards_response import V1ListDashboardsResponse
|
|
208
254
|
from eval_studio_client.api.models.v1_list_dashboards_shared_with_me_response import V1ListDashboardsSharedWithMeResponse
|
|
209
255
|
from eval_studio_client.api.models.v1_list_documents_response import V1ListDocumentsResponse
|
|
210
256
|
from eval_studio_client.api.models.v1_list_evaluators_response import V1ListEvaluatorsResponse
|
|
211
257
|
from eval_studio_client.api.models.v1_list_llm_models_response import V1ListLLMModelsResponse
|
|
258
|
+
from eval_studio_client.api.models.v1_list_leaderboard_test_case_annotations_response import V1ListLeaderboardTestCaseAnnotationsResponse
|
|
212
259
|
from eval_studio_client.api.models.v1_list_leaderboards_response import V1ListLeaderboardsResponse
|
|
213
260
|
from eval_studio_client.api.models.v1_list_model_collections_response import V1ListModelCollectionsResponse
|
|
214
261
|
from eval_studio_client.api.models.v1_list_models_response import V1ListModelsResponse
|
|
@@ -227,16 +274,25 @@ from eval_studio_client.api.models.v1_list_test_cases_response import V1ListTest
|
|
|
227
274
|
from eval_studio_client.api.models.v1_list_test_classes_response import V1ListTestClassesResponse
|
|
228
275
|
from eval_studio_client.api.models.v1_list_tests_response import V1ListTestsResponse
|
|
229
276
|
from eval_studio_client.api.models.v1_list_tests_shared_with_me_response import V1ListTestsSharedWithMeResponse
|
|
277
|
+
from eval_studio_client.api.models.v1_list_unseen_operations_response import V1ListUnseenOperationsResponse
|
|
230
278
|
from eval_studio_client.api.models.v1_list_workflow_access_response import V1ListWorkflowAccessResponse
|
|
231
279
|
from eval_studio_client.api.models.v1_list_workflow_dependencies_response import V1ListWorkflowDependenciesResponse
|
|
232
280
|
from eval_studio_client.api.models.v1_list_workflows_response import V1ListWorkflowsResponse
|
|
233
281
|
from eval_studio_client.api.models.v1_list_workflows_shared_with_me_response import V1ListWorkflowsSharedWithMeResponse
|
|
282
|
+
from eval_studio_client.api.models.v1_mark_operation_seen_by_creator_response import V1MarkOperationSeenByCreatorResponse
|
|
283
|
+
from eval_studio_client.api.models.v1_metric import V1Metric
|
|
284
|
+
from eval_studio_client.api.models.v1_metric_average import V1MetricAverage
|
|
285
|
+
from eval_studio_client.api.models.v1_metric_meta import V1MetricMeta
|
|
234
286
|
from eval_studio_client.api.models.v1_metric_score import V1MetricScore
|
|
235
287
|
from eval_studio_client.api.models.v1_metric_scores import V1MetricScores
|
|
236
288
|
from eval_studio_client.api.models.v1_model import V1Model
|
|
237
289
|
from eval_studio_client.api.models.v1_model_type import V1ModelType
|
|
290
|
+
from eval_studio_client.api.models.v1_models_comparisons import V1ModelsComparisons
|
|
291
|
+
from eval_studio_client.api.models.v1_models_comparisons_metrics import V1ModelsComparisonsMetrics
|
|
292
|
+
from eval_studio_client.api.models.v1_models_overview import V1ModelsOverview
|
|
238
293
|
from eval_studio_client.api.models.v1_operation import V1Operation
|
|
239
294
|
from eval_studio_client.api.models.v1_operation_progress import V1OperationProgress
|
|
295
|
+
from eval_studio_client.api.models.v1_operation_view import V1OperationView
|
|
240
296
|
from eval_studio_client.api.models.v1_perturb_test_in_place_response import V1PerturbTestInPlaceResponse
|
|
241
297
|
from eval_studio_client.api.models.v1_perturb_test_response import V1PerturbTestResponse
|
|
242
298
|
from eval_studio_client.api.models.v1_perturbator import V1Perturbator
|
|
@@ -248,21 +304,30 @@ from eval_studio_client.api.models.v1_prompt_library_item import V1PromptLibrary
|
|
|
248
304
|
from eval_studio_client.api.models.v1_repeated_context import V1RepeatedContext
|
|
249
305
|
from eval_studio_client.api.models.v1_repeated_string import V1RepeatedString
|
|
250
306
|
from eval_studio_client.api.models.v1_reset_workflow_node_response import V1ResetWorkflowNodeResponse
|
|
307
|
+
from eval_studio_client.api.models.v1_retrieved_context_diff import V1RetrievedContextDiff
|
|
251
308
|
from eval_studio_client.api.models.v1_role import V1Role
|
|
252
309
|
from eval_studio_client.api.models.v1_role_binding import V1RoleBinding
|
|
253
310
|
from eval_studio_client.api.models.v1_stats import V1Stats
|
|
311
|
+
from eval_studio_client.api.models.v1_technical_metrics import V1TechnicalMetrics
|
|
312
|
+
from eval_studio_client.api.models.v1_technical_metrics_detail import V1TechnicalMetricsDetail
|
|
254
313
|
from eval_studio_client.api.models.v1_test import V1Test
|
|
255
314
|
from eval_studio_client.api.models.v1_test_case import V1TestCase
|
|
315
|
+
from eval_studio_client.api.models.v1_test_case_leaderboard_item import V1TestCaseLeaderboardItem
|
|
256
316
|
from eval_studio_client.api.models.v1_test_case_relationship import V1TestCaseRelationship
|
|
317
|
+
from eval_studio_client.api.models.v1_test_case_relationship_info import V1TestCaseRelationshipInfo
|
|
318
|
+
from eval_studio_client.api.models.v1_test_case_result import V1TestCaseResult
|
|
257
319
|
from eval_studio_client.api.models.v1_test_cases_generator import V1TestCasesGenerator
|
|
258
320
|
from eval_studio_client.api.models.v1_test_class import V1TestClass
|
|
259
321
|
from eval_studio_client.api.models.v1_test_class_type import V1TestClassType
|
|
260
322
|
from eval_studio_client.api.models.v1_test_lab import V1TestLab
|
|
261
323
|
from eval_studio_client.api.models.v1_test_suite_evaluates import V1TestSuiteEvaluates
|
|
262
324
|
from eval_studio_client.api.models.v1_test_type import V1TestType
|
|
325
|
+
from eval_studio_client.api.models.v1_text_similarity_metric import V1TextSimilarityMetric
|
|
263
326
|
from eval_studio_client.api.models.v1_update_dashboard_response import V1UpdateDashboardResponse
|
|
327
|
+
from eval_studio_client.api.models.v1_update_dashboard_test_case_annotation_response import V1UpdateDashboardTestCaseAnnotationResponse
|
|
264
328
|
from eval_studio_client.api.models.v1_update_document_response import V1UpdateDocumentResponse
|
|
265
329
|
from eval_studio_client.api.models.v1_update_leaderboard_response import V1UpdateLeaderboardResponse
|
|
330
|
+
from eval_studio_client.api.models.v1_update_leaderboard_test_case_annotation_response import V1UpdateLeaderboardTestCaseAnnotationResponse
|
|
266
331
|
from eval_studio_client.api.models.v1_update_model_response import V1UpdateModelResponse
|
|
267
332
|
from eval_studio_client.api.models.v1_update_operation_response import V1UpdateOperationResponse
|
|
268
333
|
from eval_studio_client.api.models.v1_update_test_case_response import V1UpdateTestCaseResponse
|
|
@@ -2,7 +2,9 @@
|
|
|
2
2
|
|
|
3
3
|
# import apis into api package
|
|
4
4
|
from eval_studio_client.api.api.adversarial_inputs_service_api import AdversarialInputsServiceApi
|
|
5
|
+
from eval_studio_client.api.api.dashboard_report_service_api import DashboardReportServiceApi
|
|
5
6
|
from eval_studio_client.api.api.dashboard_service_api import DashboardServiceApi
|
|
7
|
+
from eval_studio_client.api.api.dashboard_test_case_annotation_service_api import DashboardTestCaseAnnotationServiceApi
|
|
6
8
|
from eval_studio_client.api.api.document_service_api import DocumentServiceApi
|
|
7
9
|
from eval_studio_client.api.api.evaluation_service_api import EvaluationServiceApi
|
|
8
10
|
from eval_studio_client.api.api.evaluator_service_api import EvaluatorServiceApi
|
|
@@ -10,6 +12,7 @@ from eval_studio_client.api.api.human_calibration_service_api import HumanCalibr
|
|
|
10
12
|
from eval_studio_client.api.api.info_service_api import InfoServiceApi
|
|
11
13
|
from eval_studio_client.api.api.leaderboard_report_service_api import LeaderboardReportServiceApi
|
|
12
14
|
from eval_studio_client.api.api.leaderboard_service_api import LeaderboardServiceApi
|
|
15
|
+
from eval_studio_client.api.api.leaderboard_test_case_annotation_service_api import LeaderboardTestCaseAnnotationServiceApi
|
|
13
16
|
from eval_studio_client.api.api.model_service_api import ModelServiceApi
|
|
14
17
|
from eval_studio_client.api.api.operation_progress_service_api import OperationProgressServiceApi
|
|
15
18
|
from eval_studio_client.api.api.operation_service_api import OperationServiceApi
|
|
@@ -0,0 +1,292 @@
|
|
|
1
|
+
# coding: utf-8
|
|
2
|
+
|
|
3
|
+
"""
|
|
4
|
+
ai/h2o/eval_studio/v1/insight.proto
|
|
5
|
+
|
|
6
|
+
No description provided (generated by Openapi Generator https://github.com/openapitools/openapi-generator)
|
|
7
|
+
|
|
8
|
+
The version of the OpenAPI document: version not set
|
|
9
|
+
Generated by OpenAPI Generator (https://openapi-generator.tech)
|
|
10
|
+
|
|
11
|
+
Do not edit the class manually.
|
|
12
|
+
""" # noqa: E501
|
|
13
|
+
|
|
14
|
+
import warnings
|
|
15
|
+
from pydantic import validate_call, Field, StrictFloat, StrictStr, StrictInt
|
|
16
|
+
from typing import Any, Dict, List, Optional, Tuple, Union
|
|
17
|
+
from typing_extensions import Annotated
|
|
18
|
+
|
|
19
|
+
from pydantic import Field, field_validator
|
|
20
|
+
from typing_extensions import Annotated
|
|
21
|
+
from eval_studio_client.api.models.v1_get_dashboard_report_response import V1GetDashboardReportResponse
|
|
22
|
+
|
|
23
|
+
from eval_studio_client.api.api_client import ApiClient, RequestSerialized
|
|
24
|
+
from eval_studio_client.api.api_response import ApiResponse
|
|
25
|
+
from eval_studio_client.api.rest import RESTResponseType
|
|
26
|
+
|
|
27
|
+
|
|
28
|
+
class DashboardReportServiceApi:
|
|
29
|
+
"""NOTE: This class is auto generated by OpenAPI Generator
|
|
30
|
+
Ref: https://openapi-generator.tech
|
|
31
|
+
|
|
32
|
+
Do not edit the class manually.
|
|
33
|
+
"""
|
|
34
|
+
|
|
35
|
+
def __init__(self, api_client=None) -> None:
|
|
36
|
+
if api_client is None:
|
|
37
|
+
api_client = ApiClient.get_default()
|
|
38
|
+
self.api_client = api_client
|
|
39
|
+
|
|
40
|
+
|
|
41
|
+
@validate_call
|
|
42
|
+
def dashboard_report_service_get_dashboard_report(
|
|
43
|
+
self,
|
|
44
|
+
name: Annotated[str, Field(strict=True, description="Required. The name of the Dashboard to retrieve.")],
|
|
45
|
+
_request_timeout: Union[
|
|
46
|
+
None,
|
|
47
|
+
Annotated[StrictFloat, Field(gt=0)],
|
|
48
|
+
Tuple[
|
|
49
|
+
Annotated[StrictFloat, Field(gt=0)],
|
|
50
|
+
Annotated[StrictFloat, Field(gt=0)]
|
|
51
|
+
]
|
|
52
|
+
] = None,
|
|
53
|
+
_request_auth: Optional[Dict[StrictStr, Any]] = None,
|
|
54
|
+
_content_type: Optional[StrictStr] = None,
|
|
55
|
+
_headers: Optional[Dict[StrictStr, Any]] = None,
|
|
56
|
+
_host_index: Annotated[StrictInt, Field(ge=0, le=0)] = 0,
|
|
57
|
+
) -> V1GetDashboardReportResponse:
|
|
58
|
+
"""dashboard_report_service_get_dashboard_report
|
|
59
|
+
|
|
60
|
+
|
|
61
|
+
:param name: Required. The name of the Dashboard to retrieve. (required)
|
|
62
|
+
:type name: str
|
|
63
|
+
:param _request_timeout: timeout setting for this request. If one
|
|
64
|
+
number provided, it will be total request
|
|
65
|
+
timeout. It can also be a pair (tuple) of
|
|
66
|
+
(connection, read) timeouts.
|
|
67
|
+
:type _request_timeout: int, tuple(int, int), optional
|
|
68
|
+
:param _request_auth: set to override the auth_settings for an a single
|
|
69
|
+
request; this effectively ignores the
|
|
70
|
+
authentication in the spec for a single request.
|
|
71
|
+
:type _request_auth: dict, optional
|
|
72
|
+
:param _content_type: force content-type for the request.
|
|
73
|
+
:type _content_type: str, Optional
|
|
74
|
+
:param _headers: set to override the headers for a single
|
|
75
|
+
request; this effectively ignores the headers
|
|
76
|
+
in the spec for a single request.
|
|
77
|
+
:type _headers: dict, optional
|
|
78
|
+
:param _host_index: set to override the host_index for a single
|
|
79
|
+
request; this effectively ignores the host_index
|
|
80
|
+
in the spec for a single request.
|
|
81
|
+
:type _host_index: int, optional
|
|
82
|
+
:return: Returns the result object.
|
|
83
|
+
""" # noqa: E501
|
|
84
|
+
|
|
85
|
+
_param = self._dashboard_report_service_get_dashboard_report_serialize(
|
|
86
|
+
name=name,
|
|
87
|
+
_request_auth=_request_auth,
|
|
88
|
+
_content_type=_content_type,
|
|
89
|
+
_headers=_headers,
|
|
90
|
+
_host_index=_host_index
|
|
91
|
+
)
|
|
92
|
+
|
|
93
|
+
_response_types_map: Dict[str, Optional[str]] = {
|
|
94
|
+
'200': "V1GetDashboardReportResponse",
|
|
95
|
+
}
|
|
96
|
+
response_data = self.api_client.call_api(
|
|
97
|
+
*_param,
|
|
98
|
+
_request_timeout=_request_timeout
|
|
99
|
+
)
|
|
100
|
+
response_data.read()
|
|
101
|
+
return self.api_client.response_deserialize(
|
|
102
|
+
response_data=response_data,
|
|
103
|
+
response_types_map=_response_types_map,
|
|
104
|
+
).data
|
|
105
|
+
|
|
106
|
+
|
|
107
|
+
@validate_call
|
|
108
|
+
def dashboard_report_service_get_dashboard_report_with_http_info(
|
|
109
|
+
self,
|
|
110
|
+
name: Annotated[str, Field(strict=True, description="Required. The name of the Dashboard to retrieve.")],
|
|
111
|
+
_request_timeout: Union[
|
|
112
|
+
None,
|
|
113
|
+
Annotated[StrictFloat, Field(gt=0)],
|
|
114
|
+
Tuple[
|
|
115
|
+
Annotated[StrictFloat, Field(gt=0)],
|
|
116
|
+
Annotated[StrictFloat, Field(gt=0)]
|
|
117
|
+
]
|
|
118
|
+
] = None,
|
|
119
|
+
_request_auth: Optional[Dict[StrictStr, Any]] = None,
|
|
120
|
+
_content_type: Optional[StrictStr] = None,
|
|
121
|
+
_headers: Optional[Dict[StrictStr, Any]] = None,
|
|
122
|
+
_host_index: Annotated[StrictInt, Field(ge=0, le=0)] = 0,
|
|
123
|
+
) -> ApiResponse[V1GetDashboardReportResponse]:
|
|
124
|
+
"""dashboard_report_service_get_dashboard_report
|
|
125
|
+
|
|
126
|
+
|
|
127
|
+
:param name: Required. The name of the Dashboard to retrieve. (required)
|
|
128
|
+
:type name: str
|
|
129
|
+
:param _request_timeout: timeout setting for this request. If one
|
|
130
|
+
number provided, it will be total request
|
|
131
|
+
timeout. It can also be a pair (tuple) of
|
|
132
|
+
(connection, read) timeouts.
|
|
133
|
+
:type _request_timeout: int, tuple(int, int), optional
|
|
134
|
+
:param _request_auth: set to override the auth_settings for an a single
|
|
135
|
+
request; this effectively ignores the
|
|
136
|
+
authentication in the spec for a single request.
|
|
137
|
+
:type _request_auth: dict, optional
|
|
138
|
+
:param _content_type: force content-type for the request.
|
|
139
|
+
:type _content_type: str, Optional
|
|
140
|
+
:param _headers: set to override the headers for a single
|
|
141
|
+
request; this effectively ignores the headers
|
|
142
|
+
in the spec for a single request.
|
|
143
|
+
:type _headers: dict, optional
|
|
144
|
+
:param _host_index: set to override the host_index for a single
|
|
145
|
+
request; this effectively ignores the host_index
|
|
146
|
+
in the spec for a single request.
|
|
147
|
+
:type _host_index: int, optional
|
|
148
|
+
:return: Returns the result object.
|
|
149
|
+
""" # noqa: E501
|
|
150
|
+
|
|
151
|
+
_param = self._dashboard_report_service_get_dashboard_report_serialize(
|
|
152
|
+
name=name,
|
|
153
|
+
_request_auth=_request_auth,
|
|
154
|
+
_content_type=_content_type,
|
|
155
|
+
_headers=_headers,
|
|
156
|
+
_host_index=_host_index
|
|
157
|
+
)
|
|
158
|
+
|
|
159
|
+
_response_types_map: Dict[str, Optional[str]] = {
|
|
160
|
+
'200': "V1GetDashboardReportResponse",
|
|
161
|
+
}
|
|
162
|
+
response_data = self.api_client.call_api(
|
|
163
|
+
*_param,
|
|
164
|
+
_request_timeout=_request_timeout
|
|
165
|
+
)
|
|
166
|
+
response_data.read()
|
|
167
|
+
return self.api_client.response_deserialize(
|
|
168
|
+
response_data=response_data,
|
|
169
|
+
response_types_map=_response_types_map,
|
|
170
|
+
)
|
|
171
|
+
|
|
172
|
+
|
|
173
|
+
@validate_call
|
|
174
|
+
def dashboard_report_service_get_dashboard_report_without_preload_content(
|
|
175
|
+
self,
|
|
176
|
+
name: Annotated[str, Field(strict=True, description="Required. The name of the Dashboard to retrieve.")],
|
|
177
|
+
_request_timeout: Union[
|
|
178
|
+
None,
|
|
179
|
+
Annotated[StrictFloat, Field(gt=0)],
|
|
180
|
+
Tuple[
|
|
181
|
+
Annotated[StrictFloat, Field(gt=0)],
|
|
182
|
+
Annotated[StrictFloat, Field(gt=0)]
|
|
183
|
+
]
|
|
184
|
+
] = None,
|
|
185
|
+
_request_auth: Optional[Dict[StrictStr, Any]] = None,
|
|
186
|
+
_content_type: Optional[StrictStr] = None,
|
|
187
|
+
_headers: Optional[Dict[StrictStr, Any]] = None,
|
|
188
|
+
_host_index: Annotated[StrictInt, Field(ge=0, le=0)] = 0,
|
|
189
|
+
) -> RESTResponseType:
|
|
190
|
+
"""dashboard_report_service_get_dashboard_report
|
|
191
|
+
|
|
192
|
+
|
|
193
|
+
:param name: Required. The name of the Dashboard to retrieve. (required)
|
|
194
|
+
:type name: str
|
|
195
|
+
:param _request_timeout: timeout setting for this request. If one
|
|
196
|
+
number provided, it will be total request
|
|
197
|
+
timeout. It can also be a pair (tuple) of
|
|
198
|
+
(connection, read) timeouts.
|
|
199
|
+
:type _request_timeout: int, tuple(int, int), optional
|
|
200
|
+
:param _request_auth: set to override the auth_settings for an a single
|
|
201
|
+
request; this effectively ignores the
|
|
202
|
+
authentication in the spec for a single request.
|
|
203
|
+
:type _request_auth: dict, optional
|
|
204
|
+
:param _content_type: force content-type for the request.
|
|
205
|
+
:type _content_type: str, Optional
|
|
206
|
+
:param _headers: set to override the headers for a single
|
|
207
|
+
request; this effectively ignores the headers
|
|
208
|
+
in the spec for a single request.
|
|
209
|
+
:type _headers: dict, optional
|
|
210
|
+
:param _host_index: set to override the host_index for a single
|
|
211
|
+
request; this effectively ignores the host_index
|
|
212
|
+
in the spec for a single request.
|
|
213
|
+
:type _host_index: int, optional
|
|
214
|
+
:return: Returns the result object.
|
|
215
|
+
""" # noqa: E501
|
|
216
|
+
|
|
217
|
+
_param = self._dashboard_report_service_get_dashboard_report_serialize(
|
|
218
|
+
name=name,
|
|
219
|
+
_request_auth=_request_auth,
|
|
220
|
+
_content_type=_content_type,
|
|
221
|
+
_headers=_headers,
|
|
222
|
+
_host_index=_host_index
|
|
223
|
+
)
|
|
224
|
+
|
|
225
|
+
_response_types_map: Dict[str, Optional[str]] = {
|
|
226
|
+
'200': "V1GetDashboardReportResponse",
|
|
227
|
+
}
|
|
228
|
+
response_data = self.api_client.call_api(
|
|
229
|
+
*_param,
|
|
230
|
+
_request_timeout=_request_timeout
|
|
231
|
+
)
|
|
232
|
+
return response_data.response
|
|
233
|
+
|
|
234
|
+
|
|
235
|
+
def _dashboard_report_service_get_dashboard_report_serialize(
|
|
236
|
+
self,
|
|
237
|
+
name,
|
|
238
|
+
_request_auth,
|
|
239
|
+
_content_type,
|
|
240
|
+
_headers,
|
|
241
|
+
_host_index,
|
|
242
|
+
) -> RequestSerialized:
|
|
243
|
+
|
|
244
|
+
_host = None
|
|
245
|
+
|
|
246
|
+
_collection_formats: Dict[str, str] = {
|
|
247
|
+
}
|
|
248
|
+
|
|
249
|
+
_path_params: Dict[str, str] = {}
|
|
250
|
+
_query_params: List[Tuple[str, str]] = []
|
|
251
|
+
_header_params: Dict[str, Optional[str]] = _headers or {}
|
|
252
|
+
_form_params: List[Tuple[str, str]] = []
|
|
253
|
+
_files: Dict[str, Union[str, bytes]] = {}
|
|
254
|
+
_body_params: Optional[bytes] = None
|
|
255
|
+
|
|
256
|
+
# process the path parameters
|
|
257
|
+
if name is not None:
|
|
258
|
+
_path_params['name'] = name
|
|
259
|
+
# process the query parameters
|
|
260
|
+
# process the header parameters
|
|
261
|
+
# process the form parameters
|
|
262
|
+
# process the body parameter
|
|
263
|
+
|
|
264
|
+
|
|
265
|
+
# set the HTTP header `Accept`
|
|
266
|
+
_header_params['Accept'] = self.api_client.select_header_accept(
|
|
267
|
+
[
|
|
268
|
+
'application/json'
|
|
269
|
+
]
|
|
270
|
+
)
|
|
271
|
+
|
|
272
|
+
|
|
273
|
+
# authentication setting
|
|
274
|
+
_auth_settings: List[str] = [
|
|
275
|
+
]
|
|
276
|
+
|
|
277
|
+
return self.api_client.param_serialize(
|
|
278
|
+
method='GET',
|
|
279
|
+
resource_path='/v1/{name}',
|
|
280
|
+
path_params=_path_params,
|
|
281
|
+
query_params=_query_params,
|
|
282
|
+
header_params=_header_params,
|
|
283
|
+
body=_body_params,
|
|
284
|
+
post_params=_form_params,
|
|
285
|
+
files=_files,
|
|
286
|
+
auth_settings=_auth_settings,
|
|
287
|
+
collection_formats=_collection_formats,
|
|
288
|
+
_host=_host,
|
|
289
|
+
_request_auth=_request_auth
|
|
290
|
+
)
|
|
291
|
+
|
|
292
|
+
|
|
@@ -1101,7 +1101,7 @@ class DashboardServiceApi:
|
|
|
1101
1101
|
@validate_call
|
|
1102
1102
|
def dashboard_service_get_dashboard(
|
|
1103
1103
|
self,
|
|
1104
|
-
|
|
1104
|
+
name_1: Annotated[str, Field(strict=True, description="Required. The name of the Dashboard to retrieve.")],
|
|
1105
1105
|
_request_timeout: Union[
|
|
1106
1106
|
None,
|
|
1107
1107
|
Annotated[StrictFloat, Field(gt=0)],
|
|
@@ -1118,8 +1118,8 @@ class DashboardServiceApi:
|
|
|
1118
1118
|
"""dashboard_service_get_dashboard
|
|
1119
1119
|
|
|
1120
1120
|
|
|
1121
|
-
:param
|
|
1122
|
-
:type
|
|
1121
|
+
:param name_1: Required. The name of the Dashboard to retrieve. (required)
|
|
1122
|
+
:type name_1: str
|
|
1123
1123
|
:param _request_timeout: timeout setting for this request. If one
|
|
1124
1124
|
number provided, it will be total request
|
|
1125
1125
|
timeout. It can also be a pair (tuple) of
|
|
@@ -1143,7 +1143,7 @@ class DashboardServiceApi:
|
|
|
1143
1143
|
""" # noqa: E501
|
|
1144
1144
|
|
|
1145
1145
|
_param = self._dashboard_service_get_dashboard_serialize(
|
|
1146
|
-
|
|
1146
|
+
name_1=name_1,
|
|
1147
1147
|
_request_auth=_request_auth,
|
|
1148
1148
|
_content_type=_content_type,
|
|
1149
1149
|
_headers=_headers,
|
|
@@ -1167,7 +1167,7 @@ class DashboardServiceApi:
|
|
|
1167
1167
|
@validate_call
|
|
1168
1168
|
def dashboard_service_get_dashboard_with_http_info(
|
|
1169
1169
|
self,
|
|
1170
|
-
|
|
1170
|
+
name_1: Annotated[str, Field(strict=True, description="Required. The name of the Dashboard to retrieve.")],
|
|
1171
1171
|
_request_timeout: Union[
|
|
1172
1172
|
None,
|
|
1173
1173
|
Annotated[StrictFloat, Field(gt=0)],
|
|
@@ -1184,8 +1184,8 @@ class DashboardServiceApi:
|
|
|
1184
1184
|
"""dashboard_service_get_dashboard
|
|
1185
1185
|
|
|
1186
1186
|
|
|
1187
|
-
:param
|
|
1188
|
-
:type
|
|
1187
|
+
:param name_1: Required. The name of the Dashboard to retrieve. (required)
|
|
1188
|
+
:type name_1: str
|
|
1189
1189
|
:param _request_timeout: timeout setting for this request. If one
|
|
1190
1190
|
number provided, it will be total request
|
|
1191
1191
|
timeout. It can also be a pair (tuple) of
|
|
@@ -1209,7 +1209,7 @@ class DashboardServiceApi:
|
|
|
1209
1209
|
""" # noqa: E501
|
|
1210
1210
|
|
|
1211
1211
|
_param = self._dashboard_service_get_dashboard_serialize(
|
|
1212
|
-
|
|
1212
|
+
name_1=name_1,
|
|
1213
1213
|
_request_auth=_request_auth,
|
|
1214
1214
|
_content_type=_content_type,
|
|
1215
1215
|
_headers=_headers,
|
|
@@ -1233,7 +1233,7 @@ class DashboardServiceApi:
|
|
|
1233
1233
|
@validate_call
|
|
1234
1234
|
def dashboard_service_get_dashboard_without_preload_content(
|
|
1235
1235
|
self,
|
|
1236
|
-
|
|
1236
|
+
name_1: Annotated[str, Field(strict=True, description="Required. The name of the Dashboard to retrieve.")],
|
|
1237
1237
|
_request_timeout: Union[
|
|
1238
1238
|
None,
|
|
1239
1239
|
Annotated[StrictFloat, Field(gt=0)],
|
|
@@ -1250,8 +1250,8 @@ class DashboardServiceApi:
|
|
|
1250
1250
|
"""dashboard_service_get_dashboard
|
|
1251
1251
|
|
|
1252
1252
|
|
|
1253
|
-
:param
|
|
1254
|
-
:type
|
|
1253
|
+
:param name_1: Required. The name of the Dashboard to retrieve. (required)
|
|
1254
|
+
:type name_1: str
|
|
1255
1255
|
:param _request_timeout: timeout setting for this request. If one
|
|
1256
1256
|
number provided, it will be total request
|
|
1257
1257
|
timeout. It can also be a pair (tuple) of
|
|
@@ -1275,7 +1275,7 @@ class DashboardServiceApi:
|
|
|
1275
1275
|
""" # noqa: E501
|
|
1276
1276
|
|
|
1277
1277
|
_param = self._dashboard_service_get_dashboard_serialize(
|
|
1278
|
-
|
|
1278
|
+
name_1=name_1,
|
|
1279
1279
|
_request_auth=_request_auth,
|
|
1280
1280
|
_content_type=_content_type,
|
|
1281
1281
|
_headers=_headers,
|
|
@@ -1294,7 +1294,7 @@ class DashboardServiceApi:
|
|
|
1294
1294
|
|
|
1295
1295
|
def _dashboard_service_get_dashboard_serialize(
|
|
1296
1296
|
self,
|
|
1297
|
-
|
|
1297
|
+
name_1,
|
|
1298
1298
|
_request_auth,
|
|
1299
1299
|
_content_type,
|
|
1300
1300
|
_headers,
|
|
@@ -1314,8 +1314,8 @@ class DashboardServiceApi:
|
|
|
1314
1314
|
_body_params: Optional[bytes] = None
|
|
1315
1315
|
|
|
1316
1316
|
# process the path parameters
|
|
1317
|
-
if
|
|
1318
|
-
_path_params['
|
|
1317
|
+
if name_1 is not None:
|
|
1318
|
+
_path_params['name_1'] = name_1
|
|
1319
1319
|
# process the query parameters
|
|
1320
1320
|
# process the header parameters
|
|
1321
1321
|
# process the form parameters
|
|
@@ -1336,7 +1336,7 @@ class DashboardServiceApi:
|
|
|
1336
1336
|
|
|
1337
1337
|
return self.api_client.param_serialize(
|
|
1338
1338
|
method='GET',
|
|
1339
|
-
resource_path='/v1/{
|
|
1339
|
+
resource_path='/v1/{name_1}',
|
|
1340
1340
|
path_params=_path_params,
|
|
1341
1341
|
query_params=_query_params,
|
|
1342
1342
|
header_params=_header_params,
|