eval-studio-client 1.2.2a1__py3-none-any.whl → 1.2.3__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- eval_studio_client/api/models/adversarial_inputs_service_test_adversarial_inputs_robustness_request.py +5 -5
- eval_studio_client/api/models/perturbation_service_create_perturbation_request.py +11 -11
- eval_studio_client/api/models/prompt_generation_service_auto_generate_prompts_request.py +2 -2
- eval_studio_client/api/models/protobuf_any.py +2 -2
- eval_studio_client/api/models/required_the_dashboard_to_update.py +2 -2
- eval_studio_client/api/models/required_the_document_to_update.py +2 -2
- eval_studio_client/api/models/required_the_leaderboard_to_update.py +8 -8
- eval_studio_client/api/models/required_the_model_to_update.py +2 -2
- eval_studio_client/api/models/required_the_operation_to_finalize.py +2 -2
- eval_studio_client/api/models/required_the_operation_to_update.py +2 -2
- eval_studio_client/api/models/required_the_test_case_to_update.py +2 -2
- eval_studio_client/api/models/required_the_test_to_update.py +2 -2
- eval_studio_client/api/models/required_the_updated_workflow.py +5 -5
- eval_studio_client/api/models/required_the_updated_workflow_node.py +5 -5
- eval_studio_client/api/models/rpc_status.py +5 -5
- eval_studio_client/api/models/test_case_service_batch_delete_test_cases_request.py +2 -2
- eval_studio_client/api/models/test_service_clone_test_request.py +2 -2
- eval_studio_client/api/models/test_service_generate_test_cases_request.py +5 -5
- eval_studio_client/api/models/test_service_grant_test_access_request.py +2 -2
- eval_studio_client/api/models/test_service_import_test_cases_from_library_request.py +2 -2
- eval_studio_client/api/models/test_service_list_test_case_library_items_request.py +2 -2
- eval_studio_client/api/models/test_service_perturb_test_in_place_request.py +5 -5
- eval_studio_client/api/models/test_service_perturb_test_request.py +5 -5
- eval_studio_client/api/models/test_service_revoke_test_access_request.py +2 -2
- eval_studio_client/api/models/v1_abort_operation_response.py +2 -2
- eval_studio_client/api/models/v1_batch_create_leaderboards_request.py +5 -5
- eval_studio_client/api/models/v1_batch_create_leaderboards_response.py +2 -2
- eval_studio_client/api/models/v1_batch_delete_dashboards_request.py +2 -2
- eval_studio_client/api/models/v1_batch_delete_dashboards_response.py +5 -5
- eval_studio_client/api/models/v1_batch_delete_documents_request.py +2 -2
- eval_studio_client/api/models/v1_batch_delete_documents_response.py +5 -5
- eval_studio_client/api/models/v1_batch_delete_evaluators_request.py +2 -2
- eval_studio_client/api/models/v1_batch_delete_evaluators_response.py +5 -5
- eval_studio_client/api/models/v1_batch_delete_leaderboards_request.py +2 -2
- eval_studio_client/api/models/v1_batch_delete_leaderboards_response.py +5 -5
- eval_studio_client/api/models/v1_batch_delete_models_request.py +2 -2
- eval_studio_client/api/models/v1_batch_delete_models_response.py +5 -5
- eval_studio_client/api/models/v1_batch_delete_test_cases_response.py +5 -5
- eval_studio_client/api/models/v1_batch_delete_tests_request.py +2 -2
- eval_studio_client/api/models/v1_batch_delete_tests_response.py +5 -5
- eval_studio_client/api/models/v1_batch_delete_workflows_request.py +2 -2
- eval_studio_client/api/models/v1_batch_delete_workflows_response.py +5 -5
- eval_studio_client/api/models/v1_batch_get_dashboards_response.py +5 -5
- eval_studio_client/api/models/v1_batch_get_documents_response.py +5 -5
- eval_studio_client/api/models/v1_batch_get_leaderboards_response.py +5 -5
- eval_studio_client/api/models/v1_batch_get_models_response.py +5 -5
- eval_studio_client/api/models/v1_batch_get_operations_response.py +5 -5
- eval_studio_client/api/models/v1_batch_get_tests_response.py +5 -5
- eval_studio_client/api/models/v1_batch_get_workflow_edges_response.py +5 -5
- eval_studio_client/api/models/v1_batch_get_workflow_nodes_response.py +5 -5
- eval_studio_client/api/models/v1_batch_import_leaderboard_request.py +2 -2
- eval_studio_client/api/models/v1_batch_import_leaderboard_response.py +2 -2
- eval_studio_client/api/models/v1_batch_import_tests_request.py +2 -2
- eval_studio_client/api/models/v1_batch_import_tests_response.py +5 -5
- eval_studio_client/api/models/v1_check_base_models_response.py +2 -2
- eval_studio_client/api/models/v1_clone_test_response.py +2 -2
- eval_studio_client/api/models/v1_clone_workflow_response.py +2 -2
- eval_studio_client/api/models/v1_collection_info.py +2 -2
- eval_studio_client/api/models/v1_context.py +2 -2
- eval_studio_client/api/models/v1_create_dashboard_response.py +2 -2
- eval_studio_client/api/models/v1_create_document_response.py +2 -2
- eval_studio_client/api/models/v1_create_evaluation_request.py +5 -5
- eval_studio_client/api/models/v1_create_evaluator_response.py +2 -2
- eval_studio_client/api/models/v1_create_leaderboard_request.py +2 -2
- eval_studio_client/api/models/v1_create_leaderboard_response.py +2 -2
- eval_studio_client/api/models/v1_create_leaderboard_without_cache_response.py +2 -2
- eval_studio_client/api/models/v1_create_model_response.py +2 -2
- eval_studio_client/api/models/v1_create_perturbation_response.py +2 -2
- eval_studio_client/api/models/v1_create_test_case_response.py +2 -2
- eval_studio_client/api/models/v1_create_test_lab_response.py +2 -2
- eval_studio_client/api/models/v1_create_test_response.py +2 -2
- eval_studio_client/api/models/v1_create_workflow_edge_response.py +2 -2
- eval_studio_client/api/models/v1_create_workflow_node_response.py +2 -2
- eval_studio_client/api/models/v1_create_workflow_response.py +2 -2
- eval_studio_client/api/models/v1_dashboard.py +2 -2
- eval_studio_client/api/models/v1_delete_dashboard_response.py +2 -2
- eval_studio_client/api/models/v1_delete_document_response.py +2 -2
- eval_studio_client/api/models/v1_delete_evaluator_response.py +2 -2
- eval_studio_client/api/models/v1_delete_leaderboard_response.py +2 -2
- eval_studio_client/api/models/v1_delete_model_response.py +2 -2
- eval_studio_client/api/models/v1_delete_test_case_response.py +2 -2
- eval_studio_client/api/models/v1_delete_test_response.py +2 -2
- eval_studio_client/api/models/v1_delete_workflow_edge_response.py +2 -2
- eval_studio_client/api/models/v1_delete_workflow_node_response.py +2 -2
- eval_studio_client/api/models/v1_delete_workflow_response.py +2 -2
- eval_studio_client/api/models/v1_dependency_list.py +5 -5
- eval_studio_client/api/models/v1_document.py +2 -2
- eval_studio_client/api/models/v1_estimate_threshold_request.py +5 -5
- eval_studio_client/api/models/v1_evaluation_test.py +8 -8
- eval_studio_client/api/models/v1_evaluator.py +5 -5
- eval_studio_client/api/models/v1_evaluator_parameter.py +2 -2
- eval_studio_client/api/models/v1_finalize_operation_response.py +2 -2
- eval_studio_client/api/models/v1_find_all_test_cases_by_id_response.py +5 -5
- eval_studio_client/api/models/v1_find_test_lab_response.py +2 -2
- eval_studio_client/api/models/v1_find_workflows_by_collection_id_response.py +5 -5
- eval_studio_client/api/models/v1_generate_test_cases_response.py +2 -2
- eval_studio_client/api/models/v1_get_dashboard_response.py +2 -2
- eval_studio_client/api/models/v1_get_document_response.py +2 -2
- eval_studio_client/api/models/v1_get_evaluator_response.py +2 -2
- eval_studio_client/api/models/v1_get_guardrails_configuration_response.py +2 -2
- eval_studio_client/api/models/v1_get_info_response.py +2 -2
- eval_studio_client/api/models/v1_get_leaderboard_report_response.py +2 -2
- eval_studio_client/api/models/v1_get_leaderboard_response.py +2 -2
- eval_studio_client/api/models/v1_get_model_response.py +2 -2
- eval_studio_client/api/models/v1_get_operation_progress_by_parent_response.py +2 -2
- eval_studio_client/api/models/v1_get_operation_response.py +2 -2
- eval_studio_client/api/models/v1_get_perturbator_response.py +2 -2
- eval_studio_client/api/models/v1_get_stats_response.py +2 -2
- eval_studio_client/api/models/v1_get_test_case_response.py +2 -2
- eval_studio_client/api/models/v1_get_test_class_response.py +2 -2
- eval_studio_client/api/models/v1_get_test_response.py +2 -2
- eval_studio_client/api/models/v1_get_workflow_node_prerequisites_response.py +2 -2
- eval_studio_client/api/models/v1_get_workflow_node_response.py +2 -2
- eval_studio_client/api/models/v1_get_workflow_response.py +2 -2
- eval_studio_client/api/models/v1_get_workflow_result_corpus_patch_response.py +2 -2
- eval_studio_client/api/models/v1_get_workflow_result_report_response.py +2 -2
- eval_studio_client/api/models/v1_get_workflow_result_summary_response.py +2 -2
- eval_studio_client/api/models/v1_get_workflow_result_system_prompt_patch_response.py +2 -2
- eval_studio_client/api/models/v1_import_evaluation_request.py +2 -2
- eval_studio_client/api/models/v1_import_leaderboard_request.py +2 -2
- eval_studio_client/api/models/v1_import_leaderboard_response.py +2 -2
- eval_studio_client/api/models/v1_import_test_cases_from_library_response.py +2 -2
- eval_studio_client/api/models/v1_import_test_cases_request.py +2 -2
- eval_studio_client/api/models/v1_info.py +2 -2
- eval_studio_client/api/models/v1_init_workflow_node_response.py +2 -2
- eval_studio_client/api/models/v1_insight.py +2 -2
- eval_studio_client/api/models/v1_labeled_test_case.py +2 -2
- eval_studio_client/api/models/v1_leaderboard.py +8 -8
- eval_studio_client/api/models/v1_leaderboard_report.py +8 -8
- eval_studio_client/api/models/v1_leaderboard_report_actual_output_data.py +2 -2
- eval_studio_client/api/models/v1_leaderboard_report_actual_output_meta.py +5 -5
- eval_studio_client/api/models/v1_leaderboard_report_evaluator.py +11 -11
- eval_studio_client/api/models/v1_leaderboard_report_evaluator_parameter.py +2 -2
- eval_studio_client/api/models/v1_leaderboard_report_explanation.py +2 -2
- eval_studio_client/api/models/v1_leaderboard_report_metrics_meta_entry.py +2 -2
- eval_studio_client/api/models/v1_leaderboard_report_model.py +2 -2
- eval_studio_client/api/models/v1_leaderboard_report_result.py +11 -11
- eval_studio_client/api/models/v1_leaderboard_report_result_relationship.py +2 -2
- eval_studio_client/api/models/v1_list_base_models_response.py +2 -2
- eval_studio_client/api/models/v1_list_dashboard_access_response.py +5 -5
- eval_studio_client/api/models/v1_list_dashboards_response.py +5 -5
- eval_studio_client/api/models/v1_list_dashboards_shared_with_me_response.py +5 -5
- eval_studio_client/api/models/v1_list_documents_response.py +5 -5
- eval_studio_client/api/models/v1_list_evaluators_response.py +5 -5
- eval_studio_client/api/models/v1_list_leaderboards_response.py +5 -5
- eval_studio_client/api/models/v1_list_llm_models_response.py +2 -2
- eval_studio_client/api/models/v1_list_model_collections_response.py +5 -5
- eval_studio_client/api/models/v1_list_models_response.py +5 -5
- eval_studio_client/api/models/v1_list_most_recent_dashboards_response.py +5 -5
- eval_studio_client/api/models/v1_list_most_recent_leaderboards_response.py +5 -5
- eval_studio_client/api/models/v1_list_most_recent_models_response.py +5 -5
- eval_studio_client/api/models/v1_list_most_recent_tests_response.py +5 -5
- eval_studio_client/api/models/v1_list_operations_response.py +5 -5
- eval_studio_client/api/models/v1_list_perturbators_response.py +5 -5
- eval_studio_client/api/models/v1_list_prompt_library_items_response.py +5 -5
- eval_studio_client/api/models/v1_list_rag_collections_response.py +5 -5
- eval_studio_client/api/models/v1_list_test_access_response.py +5 -5
- eval_studio_client/api/models/v1_list_test_case_library_items_response.py +5 -5
- eval_studio_client/api/models/v1_list_test_case_relationships_response.py +5 -5
- eval_studio_client/api/models/v1_list_test_cases_response.py +5 -5
- eval_studio_client/api/models/v1_list_test_classes_response.py +5 -5
- eval_studio_client/api/models/v1_list_tests_response.py +5 -5
- eval_studio_client/api/models/v1_list_tests_shared_with_me_response.py +5 -5
- eval_studio_client/api/models/v1_list_workflow_access_response.py +5 -5
- eval_studio_client/api/models/v1_list_workflow_dependencies_response.py +8 -8
- eval_studio_client/api/models/v1_list_workflows_response.py +5 -5
- eval_studio_client/api/models/v1_list_workflows_shared_with_me_response.py +5 -5
- eval_studio_client/api/models/v1_metric_score.py +2 -2
- eval_studio_client/api/models/v1_metric_scores.py +5 -5
- eval_studio_client/api/models/v1_model.py +2 -2
- eval_studio_client/api/models/v1_operation.py +2 -2
- eval_studio_client/api/models/v1_operation_progress.py +2 -2
- eval_studio_client/api/models/v1_perturb_test_in_place_response.py +2 -2
- eval_studio_client/api/models/v1_perturb_test_response.py +2 -2
- eval_studio_client/api/models/v1_perturbator.py +2 -2
- eval_studio_client/api/models/v1_perturbator_configuration.py +2 -2
- eval_studio_client/api/models/v1_problem_and_action.py +2 -2
- eval_studio_client/api/models/v1_process_workflow_node_response.py +2 -2
- eval_studio_client/api/models/v1_prompt_library_item.py +2 -2
- eval_studio_client/api/models/v1_repeated_context.py +5 -5
- eval_studio_client/api/models/v1_repeated_string.py +2 -2
- eval_studio_client/api/models/v1_reset_workflow_node_response.py +2 -2
- eval_studio_client/api/models/v1_role_binding.py +2 -2
- eval_studio_client/api/models/v1_stats.py +2 -2
- eval_studio_client/api/models/v1_test.py +2 -2
- eval_studio_client/api/models/v1_test_case.py +2 -2
- eval_studio_client/api/models/v1_test_case_relationship.py +2 -2
- eval_studio_client/api/models/v1_test_class.py +2 -2
- eval_studio_client/api/models/v1_test_lab.py +2 -2
- eval_studio_client/api/models/v1_update_dashboard_response.py +2 -2
- eval_studio_client/api/models/v1_update_document_response.py +2 -2
- eval_studio_client/api/models/v1_update_leaderboard_response.py +2 -2
- eval_studio_client/api/models/v1_update_model_response.py +2 -2
- eval_studio_client/api/models/v1_update_operation_response.py +2 -2
- eval_studio_client/api/models/v1_update_test_case_response.py +2 -2
- eval_studio_client/api/models/v1_update_test_response.py +2 -2
- eval_studio_client/api/models/v1_update_workflow_node_response.py +2 -2
- eval_studio_client/api/models/v1_update_workflow_response.py +2 -2
- eval_studio_client/api/models/v1_who_am_i_response.py +2 -2
- eval_studio_client/api/models/v1_workflow.py +5 -5
- eval_studio_client/api/models/v1_workflow_dependency.py +2 -2
- eval_studio_client/api/models/v1_workflow_edge.py +2 -2
- eval_studio_client/api/models/v1_workflow_node.py +5 -5
- eval_studio_client/api/models/v1_workflow_node_artifact.py +2 -2
- eval_studio_client/api/models/v1_workflow_node_artifacts.py +5 -5
- eval_studio_client/api/models/v1_workflow_node_attributes.py +2 -2
- eval_studio_client/api/models/workflow_service_clone_workflow_request.py +2 -2
- eval_studio_client/api/models/workflow_service_revoke_workflow_access_request.py +2 -2
- eval_studio_client/leaderboards.py +17 -0
- eval_studio_client/models.py +153 -8
- {eval_studio_client-1.2.2a1.dist-info → eval_studio_client-1.2.3.dist-info}/METADATA +1 -1
- {eval_studio_client-1.2.2a1.dist-info → eval_studio_client-1.2.3.dist-info}/RECORD +213 -213
- {eval_studio_client-1.2.2a1.dist-info → eval_studio_client-1.2.3.dist-info}/WHEEL +0 -0
|
@@ -72,9 +72,9 @@ class V1BatchDeleteDashboardsResponse(BaseModel):
|
|
|
72
72
|
# override the default output from pydantic by calling `to_dict()` of each item in dashboards (list)
|
|
73
73
|
_items = []
|
|
74
74
|
if self.dashboards:
|
|
75
|
-
for
|
|
76
|
-
if
|
|
77
|
-
_items.append(
|
|
75
|
+
for _item_dashboards in self.dashboards:
|
|
76
|
+
if _item_dashboards:
|
|
77
|
+
_items.append(_item_dashboards.to_dict())
|
|
78
78
|
_dict['dashboards'] = _items
|
|
79
79
|
return _dict
|
|
80
80
|
|
|
@@ -85,11 +85,11 @@ class V1BatchDeleteDashboardsResponse(BaseModel):
|
|
|
85
85
|
return None
|
|
86
86
|
|
|
87
87
|
if not isinstance(obj, dict):
|
|
88
|
-
return cls.model_validate(obj)
|
|
88
|
+
return cls.model_validate(obj, strict=False)
|
|
89
89
|
|
|
90
90
|
_obj = cls.model_validate({
|
|
91
91
|
"dashboards": [V1Dashboard.from_dict(_item) for _item in obj["dashboards"]] if obj.get("dashboards") is not None else None
|
|
92
|
-
})
|
|
92
|
+
}, strict=False)
|
|
93
93
|
return _obj
|
|
94
94
|
|
|
95
95
|
|
|
@@ -77,11 +77,11 @@ class V1BatchDeleteDocumentsRequest(BaseModel):
|
|
|
77
77
|
return None
|
|
78
78
|
|
|
79
79
|
if not isinstance(obj, dict):
|
|
80
|
-
return cls.model_validate(obj)
|
|
80
|
+
return cls.model_validate(obj, strict=False)
|
|
81
81
|
|
|
82
82
|
_obj = cls.model_validate({
|
|
83
83
|
"names": obj.get("names")
|
|
84
|
-
})
|
|
84
|
+
}, strict=False)
|
|
85
85
|
return _obj
|
|
86
86
|
|
|
87
87
|
|
|
@@ -72,9 +72,9 @@ class V1BatchDeleteDocumentsResponse(BaseModel):
|
|
|
72
72
|
# override the default output from pydantic by calling `to_dict()` of each item in documents (list)
|
|
73
73
|
_items = []
|
|
74
74
|
if self.documents:
|
|
75
|
-
for
|
|
76
|
-
if
|
|
77
|
-
_items.append(
|
|
75
|
+
for _item_documents in self.documents:
|
|
76
|
+
if _item_documents:
|
|
77
|
+
_items.append(_item_documents.to_dict())
|
|
78
78
|
_dict['documents'] = _items
|
|
79
79
|
return _dict
|
|
80
80
|
|
|
@@ -85,11 +85,11 @@ class V1BatchDeleteDocumentsResponse(BaseModel):
|
|
|
85
85
|
return None
|
|
86
86
|
|
|
87
87
|
if not isinstance(obj, dict):
|
|
88
|
-
return cls.model_validate(obj)
|
|
88
|
+
return cls.model_validate(obj, strict=False)
|
|
89
89
|
|
|
90
90
|
_obj = cls.model_validate({
|
|
91
91
|
"documents": [V1Document.from_dict(_item) for _item in obj["documents"]] if obj.get("documents") is not None else None
|
|
92
|
-
})
|
|
92
|
+
}, strict=False)
|
|
93
93
|
return _obj
|
|
94
94
|
|
|
95
95
|
|
|
@@ -77,11 +77,11 @@ class V1BatchDeleteEvaluatorsRequest(BaseModel):
|
|
|
77
77
|
return None
|
|
78
78
|
|
|
79
79
|
if not isinstance(obj, dict):
|
|
80
|
-
return cls.model_validate(obj)
|
|
80
|
+
return cls.model_validate(obj, strict=False)
|
|
81
81
|
|
|
82
82
|
_obj = cls.model_validate({
|
|
83
83
|
"names": obj.get("names")
|
|
84
|
-
})
|
|
84
|
+
}, strict=False)
|
|
85
85
|
return _obj
|
|
86
86
|
|
|
87
87
|
|
|
@@ -72,9 +72,9 @@ class V1BatchDeleteEvaluatorsResponse(BaseModel):
|
|
|
72
72
|
# override the default output from pydantic by calling `to_dict()` of each item in evaluators (list)
|
|
73
73
|
_items = []
|
|
74
74
|
if self.evaluators:
|
|
75
|
-
for
|
|
76
|
-
if
|
|
77
|
-
_items.append(
|
|
75
|
+
for _item_evaluators in self.evaluators:
|
|
76
|
+
if _item_evaluators:
|
|
77
|
+
_items.append(_item_evaluators.to_dict())
|
|
78
78
|
_dict['evaluators'] = _items
|
|
79
79
|
return _dict
|
|
80
80
|
|
|
@@ -85,11 +85,11 @@ class V1BatchDeleteEvaluatorsResponse(BaseModel):
|
|
|
85
85
|
return None
|
|
86
86
|
|
|
87
87
|
if not isinstance(obj, dict):
|
|
88
|
-
return cls.model_validate(obj)
|
|
88
|
+
return cls.model_validate(obj, strict=False)
|
|
89
89
|
|
|
90
90
|
_obj = cls.model_validate({
|
|
91
91
|
"evaluators": [V1Evaluator.from_dict(_item) for _item in obj["evaluators"]] if obj.get("evaluators") is not None else None
|
|
92
|
-
})
|
|
92
|
+
}, strict=False)
|
|
93
93
|
return _obj
|
|
94
94
|
|
|
95
95
|
|
|
@@ -79,12 +79,12 @@ class V1BatchDeleteLeaderboardsRequest(BaseModel):
|
|
|
79
79
|
return None
|
|
80
80
|
|
|
81
81
|
if not isinstance(obj, dict):
|
|
82
|
-
return cls.model_validate(obj)
|
|
82
|
+
return cls.model_validate(obj, strict=False)
|
|
83
83
|
|
|
84
84
|
_obj = cls.model_validate({
|
|
85
85
|
"names": obj.get("names"),
|
|
86
86
|
"view": obj.get("view")
|
|
87
|
-
})
|
|
87
|
+
}, strict=False)
|
|
88
88
|
return _obj
|
|
89
89
|
|
|
90
90
|
|
|
@@ -72,9 +72,9 @@ class V1BatchDeleteLeaderboardsResponse(BaseModel):
|
|
|
72
72
|
# override the default output from pydantic by calling `to_dict()` of each item in leaderboards (list)
|
|
73
73
|
_items = []
|
|
74
74
|
if self.leaderboards:
|
|
75
|
-
for
|
|
76
|
-
if
|
|
77
|
-
_items.append(
|
|
75
|
+
for _item_leaderboards in self.leaderboards:
|
|
76
|
+
if _item_leaderboards:
|
|
77
|
+
_items.append(_item_leaderboards.to_dict())
|
|
78
78
|
_dict['leaderboards'] = _items
|
|
79
79
|
return _dict
|
|
80
80
|
|
|
@@ -85,11 +85,11 @@ class V1BatchDeleteLeaderboardsResponse(BaseModel):
|
|
|
85
85
|
return None
|
|
86
86
|
|
|
87
87
|
if not isinstance(obj, dict):
|
|
88
|
-
return cls.model_validate(obj)
|
|
88
|
+
return cls.model_validate(obj, strict=False)
|
|
89
89
|
|
|
90
90
|
_obj = cls.model_validate({
|
|
91
91
|
"leaderboards": [V1Leaderboard.from_dict(_item) for _item in obj["leaderboards"]] if obj.get("leaderboards") is not None else None
|
|
92
|
-
})
|
|
92
|
+
}, strict=False)
|
|
93
93
|
return _obj
|
|
94
94
|
|
|
95
95
|
|
|
@@ -77,11 +77,11 @@ class V1BatchDeleteModelsRequest(BaseModel):
|
|
|
77
77
|
return None
|
|
78
78
|
|
|
79
79
|
if not isinstance(obj, dict):
|
|
80
|
-
return cls.model_validate(obj)
|
|
80
|
+
return cls.model_validate(obj, strict=False)
|
|
81
81
|
|
|
82
82
|
_obj = cls.model_validate({
|
|
83
83
|
"names": obj.get("names")
|
|
84
|
-
})
|
|
84
|
+
}, strict=False)
|
|
85
85
|
return _obj
|
|
86
86
|
|
|
87
87
|
|
|
@@ -72,9 +72,9 @@ class V1BatchDeleteModelsResponse(BaseModel):
|
|
|
72
72
|
# override the default output from pydantic by calling `to_dict()` of each item in models (list)
|
|
73
73
|
_items = []
|
|
74
74
|
if self.models:
|
|
75
|
-
for
|
|
76
|
-
if
|
|
77
|
-
_items.append(
|
|
75
|
+
for _item_models in self.models:
|
|
76
|
+
if _item_models:
|
|
77
|
+
_items.append(_item_models.to_dict())
|
|
78
78
|
_dict['models'] = _items
|
|
79
79
|
return _dict
|
|
80
80
|
|
|
@@ -85,11 +85,11 @@ class V1BatchDeleteModelsResponse(BaseModel):
|
|
|
85
85
|
return None
|
|
86
86
|
|
|
87
87
|
if not isinstance(obj, dict):
|
|
88
|
-
return cls.model_validate(obj)
|
|
88
|
+
return cls.model_validate(obj, strict=False)
|
|
89
89
|
|
|
90
90
|
_obj = cls.model_validate({
|
|
91
91
|
"models": [V1Model.from_dict(_item) for _item in obj["models"]] if obj.get("models") is not None else None
|
|
92
|
-
})
|
|
92
|
+
}, strict=False)
|
|
93
93
|
return _obj
|
|
94
94
|
|
|
95
95
|
|
|
@@ -72,9 +72,9 @@ class V1BatchDeleteTestCasesResponse(BaseModel):
|
|
|
72
72
|
# override the default output from pydantic by calling `to_dict()` of each item in test_cases (list)
|
|
73
73
|
_items = []
|
|
74
74
|
if self.test_cases:
|
|
75
|
-
for
|
|
76
|
-
if
|
|
77
|
-
_items.append(
|
|
75
|
+
for _item_test_cases in self.test_cases:
|
|
76
|
+
if _item_test_cases:
|
|
77
|
+
_items.append(_item_test_cases.to_dict())
|
|
78
78
|
_dict['testCases'] = _items
|
|
79
79
|
return _dict
|
|
80
80
|
|
|
@@ -85,11 +85,11 @@ class V1BatchDeleteTestCasesResponse(BaseModel):
|
|
|
85
85
|
return None
|
|
86
86
|
|
|
87
87
|
if not isinstance(obj, dict):
|
|
88
|
-
return cls.model_validate(obj)
|
|
88
|
+
return cls.model_validate(obj, strict=False)
|
|
89
89
|
|
|
90
90
|
_obj = cls.model_validate({
|
|
91
91
|
"testCases": [V1TestCase.from_dict(_item) for _item in obj["testCases"]] if obj.get("testCases") is not None else None
|
|
92
|
-
})
|
|
92
|
+
}, strict=False)
|
|
93
93
|
return _obj
|
|
94
94
|
|
|
95
95
|
|
|
@@ -78,12 +78,12 @@ class V1BatchDeleteTestsRequest(BaseModel):
|
|
|
78
78
|
return None
|
|
79
79
|
|
|
80
80
|
if not isinstance(obj, dict):
|
|
81
|
-
return cls.model_validate(obj)
|
|
81
|
+
return cls.model_validate(obj, strict=False)
|
|
82
82
|
|
|
83
83
|
_obj = cls.model_validate({
|
|
84
84
|
"names": obj.get("names"),
|
|
85
85
|
"force": obj.get("force")
|
|
86
|
-
})
|
|
86
|
+
}, strict=False)
|
|
87
87
|
return _obj
|
|
88
88
|
|
|
89
89
|
|
|
@@ -72,9 +72,9 @@ class V1BatchDeleteTestsResponse(BaseModel):
|
|
|
72
72
|
# override the default output from pydantic by calling `to_dict()` of each item in tests (list)
|
|
73
73
|
_items = []
|
|
74
74
|
if self.tests:
|
|
75
|
-
for
|
|
76
|
-
if
|
|
77
|
-
_items.append(
|
|
75
|
+
for _item_tests in self.tests:
|
|
76
|
+
if _item_tests:
|
|
77
|
+
_items.append(_item_tests.to_dict())
|
|
78
78
|
_dict['tests'] = _items
|
|
79
79
|
return _dict
|
|
80
80
|
|
|
@@ -85,11 +85,11 @@ class V1BatchDeleteTestsResponse(BaseModel):
|
|
|
85
85
|
return None
|
|
86
86
|
|
|
87
87
|
if not isinstance(obj, dict):
|
|
88
|
-
return cls.model_validate(obj)
|
|
88
|
+
return cls.model_validate(obj, strict=False)
|
|
89
89
|
|
|
90
90
|
_obj = cls.model_validate({
|
|
91
91
|
"tests": [V1Test.from_dict(_item) for _item in obj["tests"]] if obj.get("tests") is not None else None
|
|
92
|
-
})
|
|
92
|
+
}, strict=False)
|
|
93
93
|
return _obj
|
|
94
94
|
|
|
95
95
|
|
|
@@ -77,11 +77,11 @@ class V1BatchDeleteWorkflowsRequest(BaseModel):
|
|
|
77
77
|
return None
|
|
78
78
|
|
|
79
79
|
if not isinstance(obj, dict):
|
|
80
|
-
return cls.model_validate(obj)
|
|
80
|
+
return cls.model_validate(obj, strict=False)
|
|
81
81
|
|
|
82
82
|
_obj = cls.model_validate({
|
|
83
83
|
"names": obj.get("names")
|
|
84
|
-
})
|
|
84
|
+
}, strict=False)
|
|
85
85
|
return _obj
|
|
86
86
|
|
|
87
87
|
|
|
@@ -72,9 +72,9 @@ class V1BatchDeleteWorkflowsResponse(BaseModel):
|
|
|
72
72
|
# override the default output from pydantic by calling `to_dict()` of each item in workflows (list)
|
|
73
73
|
_items = []
|
|
74
74
|
if self.workflows:
|
|
75
|
-
for
|
|
76
|
-
if
|
|
77
|
-
_items.append(
|
|
75
|
+
for _item_workflows in self.workflows:
|
|
76
|
+
if _item_workflows:
|
|
77
|
+
_items.append(_item_workflows.to_dict())
|
|
78
78
|
_dict['workflows'] = _items
|
|
79
79
|
return _dict
|
|
80
80
|
|
|
@@ -85,11 +85,11 @@ class V1BatchDeleteWorkflowsResponse(BaseModel):
|
|
|
85
85
|
return None
|
|
86
86
|
|
|
87
87
|
if not isinstance(obj, dict):
|
|
88
|
-
return cls.model_validate(obj)
|
|
88
|
+
return cls.model_validate(obj, strict=False)
|
|
89
89
|
|
|
90
90
|
_obj = cls.model_validate({
|
|
91
91
|
"workflows": [V1Workflow.from_dict(_item) for _item in obj["workflows"]] if obj.get("workflows") is not None else None
|
|
92
|
-
})
|
|
92
|
+
}, strict=False)
|
|
93
93
|
return _obj
|
|
94
94
|
|
|
95
95
|
|
|
@@ -72,9 +72,9 @@ class V1BatchGetDashboardsResponse(BaseModel):
|
|
|
72
72
|
# override the default output from pydantic by calling `to_dict()` of each item in dashboards (list)
|
|
73
73
|
_items = []
|
|
74
74
|
if self.dashboards:
|
|
75
|
-
for
|
|
76
|
-
if
|
|
77
|
-
_items.append(
|
|
75
|
+
for _item_dashboards in self.dashboards:
|
|
76
|
+
if _item_dashboards:
|
|
77
|
+
_items.append(_item_dashboards.to_dict())
|
|
78
78
|
_dict['dashboards'] = _items
|
|
79
79
|
return _dict
|
|
80
80
|
|
|
@@ -85,11 +85,11 @@ class V1BatchGetDashboardsResponse(BaseModel):
|
|
|
85
85
|
return None
|
|
86
86
|
|
|
87
87
|
if not isinstance(obj, dict):
|
|
88
|
-
return cls.model_validate(obj)
|
|
88
|
+
return cls.model_validate(obj, strict=False)
|
|
89
89
|
|
|
90
90
|
_obj = cls.model_validate({
|
|
91
91
|
"dashboards": [V1Dashboard.from_dict(_item) for _item in obj["dashboards"]] if obj.get("dashboards") is not None else None
|
|
92
|
-
})
|
|
92
|
+
}, strict=False)
|
|
93
93
|
return _obj
|
|
94
94
|
|
|
95
95
|
|
|
@@ -72,9 +72,9 @@ class V1BatchGetDocumentsResponse(BaseModel):
|
|
|
72
72
|
# override the default output from pydantic by calling `to_dict()` of each item in documents (list)
|
|
73
73
|
_items = []
|
|
74
74
|
if self.documents:
|
|
75
|
-
for
|
|
76
|
-
if
|
|
77
|
-
_items.append(
|
|
75
|
+
for _item_documents in self.documents:
|
|
76
|
+
if _item_documents:
|
|
77
|
+
_items.append(_item_documents.to_dict())
|
|
78
78
|
_dict['documents'] = _items
|
|
79
79
|
return _dict
|
|
80
80
|
|
|
@@ -85,11 +85,11 @@ class V1BatchGetDocumentsResponse(BaseModel):
|
|
|
85
85
|
return None
|
|
86
86
|
|
|
87
87
|
if not isinstance(obj, dict):
|
|
88
|
-
return cls.model_validate(obj)
|
|
88
|
+
return cls.model_validate(obj, strict=False)
|
|
89
89
|
|
|
90
90
|
_obj = cls.model_validate({
|
|
91
91
|
"documents": [V1Document.from_dict(_item) for _item in obj["documents"]] if obj.get("documents") is not None else None
|
|
92
|
-
})
|
|
92
|
+
}, strict=False)
|
|
93
93
|
return _obj
|
|
94
94
|
|
|
95
95
|
|
|
@@ -72,9 +72,9 @@ class V1BatchGetLeaderboardsResponse(BaseModel):
|
|
|
72
72
|
# override the default output from pydantic by calling `to_dict()` of each item in leaderboards (list)
|
|
73
73
|
_items = []
|
|
74
74
|
if self.leaderboards:
|
|
75
|
-
for
|
|
76
|
-
if
|
|
77
|
-
_items.append(
|
|
75
|
+
for _item_leaderboards in self.leaderboards:
|
|
76
|
+
if _item_leaderboards:
|
|
77
|
+
_items.append(_item_leaderboards.to_dict())
|
|
78
78
|
_dict['leaderboards'] = _items
|
|
79
79
|
return _dict
|
|
80
80
|
|
|
@@ -85,11 +85,11 @@ class V1BatchGetLeaderboardsResponse(BaseModel):
|
|
|
85
85
|
return None
|
|
86
86
|
|
|
87
87
|
if not isinstance(obj, dict):
|
|
88
|
-
return cls.model_validate(obj)
|
|
88
|
+
return cls.model_validate(obj, strict=False)
|
|
89
89
|
|
|
90
90
|
_obj = cls.model_validate({
|
|
91
91
|
"leaderboards": [V1Leaderboard.from_dict(_item) for _item in obj["leaderboards"]] if obj.get("leaderboards") is not None else None
|
|
92
|
-
})
|
|
92
|
+
}, strict=False)
|
|
93
93
|
return _obj
|
|
94
94
|
|
|
95
95
|
|
|
@@ -72,9 +72,9 @@ class V1BatchGetModelsResponse(BaseModel):
|
|
|
72
72
|
# override the default output from pydantic by calling `to_dict()` of each item in models (list)
|
|
73
73
|
_items = []
|
|
74
74
|
if self.models:
|
|
75
|
-
for
|
|
76
|
-
if
|
|
77
|
-
_items.append(
|
|
75
|
+
for _item_models in self.models:
|
|
76
|
+
if _item_models:
|
|
77
|
+
_items.append(_item_models.to_dict())
|
|
78
78
|
_dict['models'] = _items
|
|
79
79
|
return _dict
|
|
80
80
|
|
|
@@ -85,11 +85,11 @@ class V1BatchGetModelsResponse(BaseModel):
|
|
|
85
85
|
return None
|
|
86
86
|
|
|
87
87
|
if not isinstance(obj, dict):
|
|
88
|
-
return cls.model_validate(obj)
|
|
88
|
+
return cls.model_validate(obj, strict=False)
|
|
89
89
|
|
|
90
90
|
_obj = cls.model_validate({
|
|
91
91
|
"models": [V1Model.from_dict(_item) for _item in obj["models"]] if obj.get("models") is not None else None
|
|
92
|
-
})
|
|
92
|
+
}, strict=False)
|
|
93
93
|
return _obj
|
|
94
94
|
|
|
95
95
|
|
|
@@ -72,9 +72,9 @@ class V1BatchGetOperationsResponse(BaseModel):
|
|
|
72
72
|
# override the default output from pydantic by calling `to_dict()` of each item in operations (list)
|
|
73
73
|
_items = []
|
|
74
74
|
if self.operations:
|
|
75
|
-
for
|
|
76
|
-
if
|
|
77
|
-
_items.append(
|
|
75
|
+
for _item_operations in self.operations:
|
|
76
|
+
if _item_operations:
|
|
77
|
+
_items.append(_item_operations.to_dict())
|
|
78
78
|
_dict['operations'] = _items
|
|
79
79
|
return _dict
|
|
80
80
|
|
|
@@ -85,11 +85,11 @@ class V1BatchGetOperationsResponse(BaseModel):
|
|
|
85
85
|
return None
|
|
86
86
|
|
|
87
87
|
if not isinstance(obj, dict):
|
|
88
|
-
return cls.model_validate(obj)
|
|
88
|
+
return cls.model_validate(obj, strict=False)
|
|
89
89
|
|
|
90
90
|
_obj = cls.model_validate({
|
|
91
91
|
"operations": [V1Operation.from_dict(_item) for _item in obj["operations"]] if obj.get("operations") is not None else None
|
|
92
|
-
})
|
|
92
|
+
}, strict=False)
|
|
93
93
|
return _obj
|
|
94
94
|
|
|
95
95
|
|
|
@@ -72,9 +72,9 @@ class V1BatchGetTestsResponse(BaseModel):
|
|
|
72
72
|
# override the default output from pydantic by calling `to_dict()` of each item in tests (list)
|
|
73
73
|
_items = []
|
|
74
74
|
if self.tests:
|
|
75
|
-
for
|
|
76
|
-
if
|
|
77
|
-
_items.append(
|
|
75
|
+
for _item_tests in self.tests:
|
|
76
|
+
if _item_tests:
|
|
77
|
+
_items.append(_item_tests.to_dict())
|
|
78
78
|
_dict['tests'] = _items
|
|
79
79
|
return _dict
|
|
80
80
|
|
|
@@ -85,11 +85,11 @@ class V1BatchGetTestsResponse(BaseModel):
|
|
|
85
85
|
return None
|
|
86
86
|
|
|
87
87
|
if not isinstance(obj, dict):
|
|
88
|
-
return cls.model_validate(obj)
|
|
88
|
+
return cls.model_validate(obj, strict=False)
|
|
89
89
|
|
|
90
90
|
_obj = cls.model_validate({
|
|
91
91
|
"tests": [V1Test.from_dict(_item) for _item in obj["tests"]] if obj.get("tests") is not None else None
|
|
92
|
-
})
|
|
92
|
+
}, strict=False)
|
|
93
93
|
return _obj
|
|
94
94
|
|
|
95
95
|
|
|
@@ -72,9 +72,9 @@ class V1BatchGetWorkflowEdgesResponse(BaseModel):
|
|
|
72
72
|
# override the default output from pydantic by calling `to_dict()` of each item in edges (list)
|
|
73
73
|
_items = []
|
|
74
74
|
if self.edges:
|
|
75
|
-
for
|
|
76
|
-
if
|
|
77
|
-
_items.append(
|
|
75
|
+
for _item_edges in self.edges:
|
|
76
|
+
if _item_edges:
|
|
77
|
+
_items.append(_item_edges.to_dict())
|
|
78
78
|
_dict['edges'] = _items
|
|
79
79
|
return _dict
|
|
80
80
|
|
|
@@ -85,11 +85,11 @@ class V1BatchGetWorkflowEdgesResponse(BaseModel):
|
|
|
85
85
|
return None
|
|
86
86
|
|
|
87
87
|
if not isinstance(obj, dict):
|
|
88
|
-
return cls.model_validate(obj)
|
|
88
|
+
return cls.model_validate(obj, strict=False)
|
|
89
89
|
|
|
90
90
|
_obj = cls.model_validate({
|
|
91
91
|
"edges": [V1WorkflowEdge.from_dict(_item) for _item in obj["edges"]] if obj.get("edges") is not None else None
|
|
92
|
-
})
|
|
92
|
+
}, strict=False)
|
|
93
93
|
return _obj
|
|
94
94
|
|
|
95
95
|
|
|
@@ -72,9 +72,9 @@ class V1BatchGetWorkflowNodesResponse(BaseModel):
|
|
|
72
72
|
# override the default output from pydantic by calling `to_dict()` of each item in nodes (list)
|
|
73
73
|
_items = []
|
|
74
74
|
if self.nodes:
|
|
75
|
-
for
|
|
76
|
-
if
|
|
77
|
-
_items.append(
|
|
75
|
+
for _item_nodes in self.nodes:
|
|
76
|
+
if _item_nodes:
|
|
77
|
+
_items.append(_item_nodes.to_dict())
|
|
78
78
|
_dict['nodes'] = _items
|
|
79
79
|
return _dict
|
|
80
80
|
|
|
@@ -85,11 +85,11 @@ class V1BatchGetWorkflowNodesResponse(BaseModel):
|
|
|
85
85
|
return None
|
|
86
86
|
|
|
87
87
|
if not isinstance(obj, dict):
|
|
88
|
-
return cls.model_validate(obj)
|
|
88
|
+
return cls.model_validate(obj, strict=False)
|
|
89
89
|
|
|
90
90
|
_obj = cls.model_validate({
|
|
91
91
|
"nodes": [V1WorkflowNode.from_dict(_item) for _item in obj["nodes"]] if obj.get("nodes") is not None else None
|
|
92
|
-
})
|
|
92
|
+
}, strict=False)
|
|
93
93
|
return _obj
|
|
94
94
|
|
|
95
95
|
|
|
@@ -86,7 +86,7 @@ class V1BatchImportLeaderboardRequest(BaseModel):
|
|
|
86
86
|
return None
|
|
87
87
|
|
|
88
88
|
if not isinstance(obj, dict):
|
|
89
|
-
return cls.model_validate(obj)
|
|
89
|
+
return cls.model_validate(obj, strict=False)
|
|
90
90
|
|
|
91
91
|
_obj = cls.model_validate({
|
|
92
92
|
"testLabJson": obj.get("testLabJson"),
|
|
@@ -98,7 +98,7 @@ class V1BatchImportLeaderboardRequest(BaseModel):
|
|
|
98
98
|
"leaderboardType": obj.get("leaderboardType"),
|
|
99
99
|
"dashboardDisplayName": obj.get("dashboardDisplayName"),
|
|
100
100
|
"dashboardDescription": obj.get("dashboardDescription")
|
|
101
|
-
})
|
|
101
|
+
}, strict=False)
|
|
102
102
|
return _obj
|
|
103
103
|
|
|
104
104
|
|
|
@@ -81,11 +81,11 @@ class V1BatchImportLeaderboardResponse(BaseModel):
|
|
|
81
81
|
return None
|
|
82
82
|
|
|
83
83
|
if not isinstance(obj, dict):
|
|
84
|
-
return cls.model_validate(obj)
|
|
84
|
+
return cls.model_validate(obj, strict=False)
|
|
85
85
|
|
|
86
86
|
_obj = cls.model_validate({
|
|
87
87
|
"operation": V1Operation.from_dict(obj["operation"]) if obj.get("operation") is not None else None
|
|
88
|
-
})
|
|
88
|
+
}, strict=False)
|
|
89
89
|
return _obj
|
|
90
90
|
|
|
91
91
|
|
|
@@ -80,14 +80,14 @@ class V1BatchImportTestsRequest(BaseModel):
|
|
|
80
80
|
return None
|
|
81
81
|
|
|
82
82
|
if not isinstance(obj, dict):
|
|
83
|
-
return cls.model_validate(obj)
|
|
83
|
+
return cls.model_validate(obj, strict=False)
|
|
84
84
|
|
|
85
85
|
_obj = cls.model_validate({
|
|
86
86
|
"testsJson": obj.get("testsJson"),
|
|
87
87
|
"url": obj.get("url"),
|
|
88
88
|
"testDisplayNamePrefix": obj.get("testDisplayNamePrefix"),
|
|
89
89
|
"testDescription": obj.get("testDescription")
|
|
90
|
-
})
|
|
90
|
+
}, strict=False)
|
|
91
91
|
return _obj
|
|
92
92
|
|
|
93
93
|
|
|
@@ -72,9 +72,9 @@ class V1BatchImportTestsResponse(BaseModel):
|
|
|
72
72
|
# override the default output from pydantic by calling `to_dict()` of each item in tests (list)
|
|
73
73
|
_items = []
|
|
74
74
|
if self.tests:
|
|
75
|
-
for
|
|
76
|
-
if
|
|
77
|
-
_items.append(
|
|
75
|
+
for _item_tests in self.tests:
|
|
76
|
+
if _item_tests:
|
|
77
|
+
_items.append(_item_tests.to_dict())
|
|
78
78
|
_dict['tests'] = _items
|
|
79
79
|
return _dict
|
|
80
80
|
|
|
@@ -85,11 +85,11 @@ class V1BatchImportTestsResponse(BaseModel):
|
|
|
85
85
|
return None
|
|
86
86
|
|
|
87
87
|
if not isinstance(obj, dict):
|
|
88
|
-
return cls.model_validate(obj)
|
|
88
|
+
return cls.model_validate(obj, strict=False)
|
|
89
89
|
|
|
90
90
|
_obj = cls.model_validate({
|
|
91
91
|
"tests": [V1Test.from_dict(_item) for _item in obj["tests"]] if obj.get("tests") is not None else None
|
|
92
|
-
})
|
|
92
|
+
}, strict=False)
|
|
93
93
|
return _obj
|
|
94
94
|
|
|
95
95
|
|
|
@@ -78,12 +78,12 @@ class V1CheckBaseModelsResponse(BaseModel):
|
|
|
78
78
|
return None
|
|
79
79
|
|
|
80
80
|
if not isinstance(obj, dict):
|
|
81
|
-
return cls.model_validate(obj)
|
|
81
|
+
return cls.model_validate(obj, strict=False)
|
|
82
82
|
|
|
83
83
|
_obj = cls.model_validate({
|
|
84
84
|
"modelAvailability": obj.get("modelAvailability"),
|
|
85
85
|
"reason": obj.get("reason")
|
|
86
|
-
})
|
|
86
|
+
}, strict=False)
|
|
87
87
|
return _obj
|
|
88
88
|
|
|
89
89
|
|
|
@@ -81,11 +81,11 @@ class V1CloneTestResponse(BaseModel):
|
|
|
81
81
|
return None
|
|
82
82
|
|
|
83
83
|
if not isinstance(obj, dict):
|
|
84
|
-
return cls.model_validate(obj)
|
|
84
|
+
return cls.model_validate(obj, strict=False)
|
|
85
85
|
|
|
86
86
|
_obj = cls.model_validate({
|
|
87
87
|
"test": V1Test.from_dict(obj["test"]) if obj.get("test") is not None else None
|
|
88
|
-
})
|
|
88
|
+
}, strict=False)
|
|
89
89
|
return _obj
|
|
90
90
|
|
|
91
91
|
|