eval-studio-client 1.2.2a1__py3-none-any.whl → 1.2.3__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- eval_studio_client/api/models/adversarial_inputs_service_test_adversarial_inputs_robustness_request.py +5 -5
- eval_studio_client/api/models/perturbation_service_create_perturbation_request.py +11 -11
- eval_studio_client/api/models/prompt_generation_service_auto_generate_prompts_request.py +2 -2
- eval_studio_client/api/models/protobuf_any.py +2 -2
- eval_studio_client/api/models/required_the_dashboard_to_update.py +2 -2
- eval_studio_client/api/models/required_the_document_to_update.py +2 -2
- eval_studio_client/api/models/required_the_leaderboard_to_update.py +8 -8
- eval_studio_client/api/models/required_the_model_to_update.py +2 -2
- eval_studio_client/api/models/required_the_operation_to_finalize.py +2 -2
- eval_studio_client/api/models/required_the_operation_to_update.py +2 -2
- eval_studio_client/api/models/required_the_test_case_to_update.py +2 -2
- eval_studio_client/api/models/required_the_test_to_update.py +2 -2
- eval_studio_client/api/models/required_the_updated_workflow.py +5 -5
- eval_studio_client/api/models/required_the_updated_workflow_node.py +5 -5
- eval_studio_client/api/models/rpc_status.py +5 -5
- eval_studio_client/api/models/test_case_service_batch_delete_test_cases_request.py +2 -2
- eval_studio_client/api/models/test_service_clone_test_request.py +2 -2
- eval_studio_client/api/models/test_service_generate_test_cases_request.py +5 -5
- eval_studio_client/api/models/test_service_grant_test_access_request.py +2 -2
- eval_studio_client/api/models/test_service_import_test_cases_from_library_request.py +2 -2
- eval_studio_client/api/models/test_service_list_test_case_library_items_request.py +2 -2
- eval_studio_client/api/models/test_service_perturb_test_in_place_request.py +5 -5
- eval_studio_client/api/models/test_service_perturb_test_request.py +5 -5
- eval_studio_client/api/models/test_service_revoke_test_access_request.py +2 -2
- eval_studio_client/api/models/v1_abort_operation_response.py +2 -2
- eval_studio_client/api/models/v1_batch_create_leaderboards_request.py +5 -5
- eval_studio_client/api/models/v1_batch_create_leaderboards_response.py +2 -2
- eval_studio_client/api/models/v1_batch_delete_dashboards_request.py +2 -2
- eval_studio_client/api/models/v1_batch_delete_dashboards_response.py +5 -5
- eval_studio_client/api/models/v1_batch_delete_documents_request.py +2 -2
- eval_studio_client/api/models/v1_batch_delete_documents_response.py +5 -5
- eval_studio_client/api/models/v1_batch_delete_evaluators_request.py +2 -2
- eval_studio_client/api/models/v1_batch_delete_evaluators_response.py +5 -5
- eval_studio_client/api/models/v1_batch_delete_leaderboards_request.py +2 -2
- eval_studio_client/api/models/v1_batch_delete_leaderboards_response.py +5 -5
- eval_studio_client/api/models/v1_batch_delete_models_request.py +2 -2
- eval_studio_client/api/models/v1_batch_delete_models_response.py +5 -5
- eval_studio_client/api/models/v1_batch_delete_test_cases_response.py +5 -5
- eval_studio_client/api/models/v1_batch_delete_tests_request.py +2 -2
- eval_studio_client/api/models/v1_batch_delete_tests_response.py +5 -5
- eval_studio_client/api/models/v1_batch_delete_workflows_request.py +2 -2
- eval_studio_client/api/models/v1_batch_delete_workflows_response.py +5 -5
- eval_studio_client/api/models/v1_batch_get_dashboards_response.py +5 -5
- eval_studio_client/api/models/v1_batch_get_documents_response.py +5 -5
- eval_studio_client/api/models/v1_batch_get_leaderboards_response.py +5 -5
- eval_studio_client/api/models/v1_batch_get_models_response.py +5 -5
- eval_studio_client/api/models/v1_batch_get_operations_response.py +5 -5
- eval_studio_client/api/models/v1_batch_get_tests_response.py +5 -5
- eval_studio_client/api/models/v1_batch_get_workflow_edges_response.py +5 -5
- eval_studio_client/api/models/v1_batch_get_workflow_nodes_response.py +5 -5
- eval_studio_client/api/models/v1_batch_import_leaderboard_request.py +2 -2
- eval_studio_client/api/models/v1_batch_import_leaderboard_response.py +2 -2
- eval_studio_client/api/models/v1_batch_import_tests_request.py +2 -2
- eval_studio_client/api/models/v1_batch_import_tests_response.py +5 -5
- eval_studio_client/api/models/v1_check_base_models_response.py +2 -2
- eval_studio_client/api/models/v1_clone_test_response.py +2 -2
- eval_studio_client/api/models/v1_clone_workflow_response.py +2 -2
- eval_studio_client/api/models/v1_collection_info.py +2 -2
- eval_studio_client/api/models/v1_context.py +2 -2
- eval_studio_client/api/models/v1_create_dashboard_response.py +2 -2
- eval_studio_client/api/models/v1_create_document_response.py +2 -2
- eval_studio_client/api/models/v1_create_evaluation_request.py +5 -5
- eval_studio_client/api/models/v1_create_evaluator_response.py +2 -2
- eval_studio_client/api/models/v1_create_leaderboard_request.py +2 -2
- eval_studio_client/api/models/v1_create_leaderboard_response.py +2 -2
- eval_studio_client/api/models/v1_create_leaderboard_without_cache_response.py +2 -2
- eval_studio_client/api/models/v1_create_model_response.py +2 -2
- eval_studio_client/api/models/v1_create_perturbation_response.py +2 -2
- eval_studio_client/api/models/v1_create_test_case_response.py +2 -2
- eval_studio_client/api/models/v1_create_test_lab_response.py +2 -2
- eval_studio_client/api/models/v1_create_test_response.py +2 -2
- eval_studio_client/api/models/v1_create_workflow_edge_response.py +2 -2
- eval_studio_client/api/models/v1_create_workflow_node_response.py +2 -2
- eval_studio_client/api/models/v1_create_workflow_response.py +2 -2
- eval_studio_client/api/models/v1_dashboard.py +2 -2
- eval_studio_client/api/models/v1_delete_dashboard_response.py +2 -2
- eval_studio_client/api/models/v1_delete_document_response.py +2 -2
- eval_studio_client/api/models/v1_delete_evaluator_response.py +2 -2
- eval_studio_client/api/models/v1_delete_leaderboard_response.py +2 -2
- eval_studio_client/api/models/v1_delete_model_response.py +2 -2
- eval_studio_client/api/models/v1_delete_test_case_response.py +2 -2
- eval_studio_client/api/models/v1_delete_test_response.py +2 -2
- eval_studio_client/api/models/v1_delete_workflow_edge_response.py +2 -2
- eval_studio_client/api/models/v1_delete_workflow_node_response.py +2 -2
- eval_studio_client/api/models/v1_delete_workflow_response.py +2 -2
- eval_studio_client/api/models/v1_dependency_list.py +5 -5
- eval_studio_client/api/models/v1_document.py +2 -2
- eval_studio_client/api/models/v1_estimate_threshold_request.py +5 -5
- eval_studio_client/api/models/v1_evaluation_test.py +8 -8
- eval_studio_client/api/models/v1_evaluator.py +5 -5
- eval_studio_client/api/models/v1_evaluator_parameter.py +2 -2
- eval_studio_client/api/models/v1_finalize_operation_response.py +2 -2
- eval_studio_client/api/models/v1_find_all_test_cases_by_id_response.py +5 -5
- eval_studio_client/api/models/v1_find_test_lab_response.py +2 -2
- eval_studio_client/api/models/v1_find_workflows_by_collection_id_response.py +5 -5
- eval_studio_client/api/models/v1_generate_test_cases_response.py +2 -2
- eval_studio_client/api/models/v1_get_dashboard_response.py +2 -2
- eval_studio_client/api/models/v1_get_document_response.py +2 -2
- eval_studio_client/api/models/v1_get_evaluator_response.py +2 -2
- eval_studio_client/api/models/v1_get_guardrails_configuration_response.py +2 -2
- eval_studio_client/api/models/v1_get_info_response.py +2 -2
- eval_studio_client/api/models/v1_get_leaderboard_report_response.py +2 -2
- eval_studio_client/api/models/v1_get_leaderboard_response.py +2 -2
- eval_studio_client/api/models/v1_get_model_response.py +2 -2
- eval_studio_client/api/models/v1_get_operation_progress_by_parent_response.py +2 -2
- eval_studio_client/api/models/v1_get_operation_response.py +2 -2
- eval_studio_client/api/models/v1_get_perturbator_response.py +2 -2
- eval_studio_client/api/models/v1_get_stats_response.py +2 -2
- eval_studio_client/api/models/v1_get_test_case_response.py +2 -2
- eval_studio_client/api/models/v1_get_test_class_response.py +2 -2
- eval_studio_client/api/models/v1_get_test_response.py +2 -2
- eval_studio_client/api/models/v1_get_workflow_node_prerequisites_response.py +2 -2
- eval_studio_client/api/models/v1_get_workflow_node_response.py +2 -2
- eval_studio_client/api/models/v1_get_workflow_response.py +2 -2
- eval_studio_client/api/models/v1_get_workflow_result_corpus_patch_response.py +2 -2
- eval_studio_client/api/models/v1_get_workflow_result_report_response.py +2 -2
- eval_studio_client/api/models/v1_get_workflow_result_summary_response.py +2 -2
- eval_studio_client/api/models/v1_get_workflow_result_system_prompt_patch_response.py +2 -2
- eval_studio_client/api/models/v1_import_evaluation_request.py +2 -2
- eval_studio_client/api/models/v1_import_leaderboard_request.py +2 -2
- eval_studio_client/api/models/v1_import_leaderboard_response.py +2 -2
- eval_studio_client/api/models/v1_import_test_cases_from_library_response.py +2 -2
- eval_studio_client/api/models/v1_import_test_cases_request.py +2 -2
- eval_studio_client/api/models/v1_info.py +2 -2
- eval_studio_client/api/models/v1_init_workflow_node_response.py +2 -2
- eval_studio_client/api/models/v1_insight.py +2 -2
- eval_studio_client/api/models/v1_labeled_test_case.py +2 -2
- eval_studio_client/api/models/v1_leaderboard.py +8 -8
- eval_studio_client/api/models/v1_leaderboard_report.py +8 -8
- eval_studio_client/api/models/v1_leaderboard_report_actual_output_data.py +2 -2
- eval_studio_client/api/models/v1_leaderboard_report_actual_output_meta.py +5 -5
- eval_studio_client/api/models/v1_leaderboard_report_evaluator.py +11 -11
- eval_studio_client/api/models/v1_leaderboard_report_evaluator_parameter.py +2 -2
- eval_studio_client/api/models/v1_leaderboard_report_explanation.py +2 -2
- eval_studio_client/api/models/v1_leaderboard_report_metrics_meta_entry.py +2 -2
- eval_studio_client/api/models/v1_leaderboard_report_model.py +2 -2
- eval_studio_client/api/models/v1_leaderboard_report_result.py +11 -11
- eval_studio_client/api/models/v1_leaderboard_report_result_relationship.py +2 -2
- eval_studio_client/api/models/v1_list_base_models_response.py +2 -2
- eval_studio_client/api/models/v1_list_dashboard_access_response.py +5 -5
- eval_studio_client/api/models/v1_list_dashboards_response.py +5 -5
- eval_studio_client/api/models/v1_list_dashboards_shared_with_me_response.py +5 -5
- eval_studio_client/api/models/v1_list_documents_response.py +5 -5
- eval_studio_client/api/models/v1_list_evaluators_response.py +5 -5
- eval_studio_client/api/models/v1_list_leaderboards_response.py +5 -5
- eval_studio_client/api/models/v1_list_llm_models_response.py +2 -2
- eval_studio_client/api/models/v1_list_model_collections_response.py +5 -5
- eval_studio_client/api/models/v1_list_models_response.py +5 -5
- eval_studio_client/api/models/v1_list_most_recent_dashboards_response.py +5 -5
- eval_studio_client/api/models/v1_list_most_recent_leaderboards_response.py +5 -5
- eval_studio_client/api/models/v1_list_most_recent_models_response.py +5 -5
- eval_studio_client/api/models/v1_list_most_recent_tests_response.py +5 -5
- eval_studio_client/api/models/v1_list_operations_response.py +5 -5
- eval_studio_client/api/models/v1_list_perturbators_response.py +5 -5
- eval_studio_client/api/models/v1_list_prompt_library_items_response.py +5 -5
- eval_studio_client/api/models/v1_list_rag_collections_response.py +5 -5
- eval_studio_client/api/models/v1_list_test_access_response.py +5 -5
- eval_studio_client/api/models/v1_list_test_case_library_items_response.py +5 -5
- eval_studio_client/api/models/v1_list_test_case_relationships_response.py +5 -5
- eval_studio_client/api/models/v1_list_test_cases_response.py +5 -5
- eval_studio_client/api/models/v1_list_test_classes_response.py +5 -5
- eval_studio_client/api/models/v1_list_tests_response.py +5 -5
- eval_studio_client/api/models/v1_list_tests_shared_with_me_response.py +5 -5
- eval_studio_client/api/models/v1_list_workflow_access_response.py +5 -5
- eval_studio_client/api/models/v1_list_workflow_dependencies_response.py +8 -8
- eval_studio_client/api/models/v1_list_workflows_response.py +5 -5
- eval_studio_client/api/models/v1_list_workflows_shared_with_me_response.py +5 -5
- eval_studio_client/api/models/v1_metric_score.py +2 -2
- eval_studio_client/api/models/v1_metric_scores.py +5 -5
- eval_studio_client/api/models/v1_model.py +2 -2
- eval_studio_client/api/models/v1_operation.py +2 -2
- eval_studio_client/api/models/v1_operation_progress.py +2 -2
- eval_studio_client/api/models/v1_perturb_test_in_place_response.py +2 -2
- eval_studio_client/api/models/v1_perturb_test_response.py +2 -2
- eval_studio_client/api/models/v1_perturbator.py +2 -2
- eval_studio_client/api/models/v1_perturbator_configuration.py +2 -2
- eval_studio_client/api/models/v1_problem_and_action.py +2 -2
- eval_studio_client/api/models/v1_process_workflow_node_response.py +2 -2
- eval_studio_client/api/models/v1_prompt_library_item.py +2 -2
- eval_studio_client/api/models/v1_repeated_context.py +5 -5
- eval_studio_client/api/models/v1_repeated_string.py +2 -2
- eval_studio_client/api/models/v1_reset_workflow_node_response.py +2 -2
- eval_studio_client/api/models/v1_role_binding.py +2 -2
- eval_studio_client/api/models/v1_stats.py +2 -2
- eval_studio_client/api/models/v1_test.py +2 -2
- eval_studio_client/api/models/v1_test_case.py +2 -2
- eval_studio_client/api/models/v1_test_case_relationship.py +2 -2
- eval_studio_client/api/models/v1_test_class.py +2 -2
- eval_studio_client/api/models/v1_test_lab.py +2 -2
- eval_studio_client/api/models/v1_update_dashboard_response.py +2 -2
- eval_studio_client/api/models/v1_update_document_response.py +2 -2
- eval_studio_client/api/models/v1_update_leaderboard_response.py +2 -2
- eval_studio_client/api/models/v1_update_model_response.py +2 -2
- eval_studio_client/api/models/v1_update_operation_response.py +2 -2
- eval_studio_client/api/models/v1_update_test_case_response.py +2 -2
- eval_studio_client/api/models/v1_update_test_response.py +2 -2
- eval_studio_client/api/models/v1_update_workflow_node_response.py +2 -2
- eval_studio_client/api/models/v1_update_workflow_response.py +2 -2
- eval_studio_client/api/models/v1_who_am_i_response.py +2 -2
- eval_studio_client/api/models/v1_workflow.py +5 -5
- eval_studio_client/api/models/v1_workflow_dependency.py +2 -2
- eval_studio_client/api/models/v1_workflow_edge.py +2 -2
- eval_studio_client/api/models/v1_workflow_node.py +5 -5
- eval_studio_client/api/models/v1_workflow_node_artifact.py +2 -2
- eval_studio_client/api/models/v1_workflow_node_artifacts.py +5 -5
- eval_studio_client/api/models/v1_workflow_node_attributes.py +2 -2
- eval_studio_client/api/models/workflow_service_clone_workflow_request.py +2 -2
- eval_studio_client/api/models/workflow_service_revoke_workflow_access_request.py +2 -2
- eval_studio_client/leaderboards.py +17 -0
- eval_studio_client/models.py +153 -8
- {eval_studio_client-1.2.2a1.dist-info → eval_studio_client-1.2.3.dist-info}/METADATA +1 -1
- {eval_studio_client-1.2.2a1.dist-info → eval_studio_client-1.2.3.dist-info}/RECORD +213 -213
- {eval_studio_client-1.2.2a1.dist-info → eval_studio_client-1.2.3.dist-info}/WHEEL +0 -0
|
@@ -80,16 +80,16 @@ class V1LeaderboardReport(BaseModel):
|
|
|
80
80
|
# override the default output from pydantic by calling `to_dict()` of each item in results (list)
|
|
81
81
|
_items = []
|
|
82
82
|
if self.results:
|
|
83
|
-
for
|
|
84
|
-
if
|
|
85
|
-
_items.append(
|
|
83
|
+
for _item_results in self.results:
|
|
84
|
+
if _item_results:
|
|
85
|
+
_items.append(_item_results.to_dict())
|
|
86
86
|
_dict['results'] = _items
|
|
87
87
|
# override the default output from pydantic by calling `to_dict()` of each item in models (list)
|
|
88
88
|
_items = []
|
|
89
89
|
if self.models:
|
|
90
|
-
for
|
|
91
|
-
if
|
|
92
|
-
_items.append(
|
|
90
|
+
for _item_models in self.models:
|
|
91
|
+
if _item_models:
|
|
92
|
+
_items.append(_item_models.to_dict())
|
|
93
93
|
_dict['models'] = _items
|
|
94
94
|
# override the default output from pydantic by calling `to_dict()` of evaluator
|
|
95
95
|
if self.evaluator:
|
|
@@ -103,13 +103,13 @@ class V1LeaderboardReport(BaseModel):
|
|
|
103
103
|
return None
|
|
104
104
|
|
|
105
105
|
if not isinstance(obj, dict):
|
|
106
|
-
return cls.model_validate(obj)
|
|
106
|
+
return cls.model_validate(obj, strict=False)
|
|
107
107
|
|
|
108
108
|
_obj = cls.model_validate({
|
|
109
109
|
"results": [V1LeaderboardReportResult.from_dict(_item) for _item in obj["results"]] if obj.get("results") is not None else None,
|
|
110
110
|
"models": [V1LeaderboardReportModel.from_dict(_item) for _item in obj["models"]] if obj.get("models") is not None else None,
|
|
111
111
|
"evaluator": V1LeaderboardReportEvaluator.from_dict(obj["evaluator"]) if obj.get("evaluator") is not None else None
|
|
112
|
-
})
|
|
112
|
+
}, strict=False)
|
|
113
113
|
return _obj
|
|
114
114
|
|
|
115
115
|
|
|
@@ -82,12 +82,12 @@ class V1LeaderboardReportActualOutputData(BaseModel):
|
|
|
82
82
|
return None
|
|
83
83
|
|
|
84
84
|
if not isinstance(obj, dict):
|
|
85
|
-
return cls.model_validate(obj)
|
|
85
|
+
return cls.model_validate(obj, strict=False)
|
|
86
86
|
|
|
87
87
|
_obj = cls.model_validate({
|
|
88
88
|
"text": obj.get("text"),
|
|
89
89
|
"metrics": obj.get("metrics")
|
|
90
|
-
})
|
|
90
|
+
}, strict=False)
|
|
91
91
|
return _obj
|
|
92
92
|
|
|
93
93
|
|
|
@@ -77,9 +77,9 @@ class V1LeaderboardReportActualOutputMeta(BaseModel):
|
|
|
77
77
|
# override the default output from pydantic by calling `to_dict()` of each item in data (list)
|
|
78
78
|
_items = []
|
|
79
79
|
if self.data:
|
|
80
|
-
for
|
|
81
|
-
if
|
|
82
|
-
_items.append(
|
|
80
|
+
for _item_data in self.data:
|
|
81
|
+
if _item_data:
|
|
82
|
+
_items.append(_item_data.to_dict())
|
|
83
83
|
_dict['data'] = _items
|
|
84
84
|
return _dict
|
|
85
85
|
|
|
@@ -90,12 +90,12 @@ class V1LeaderboardReportActualOutputMeta(BaseModel):
|
|
|
90
90
|
return None
|
|
91
91
|
|
|
92
92
|
if not isinstance(obj, dict):
|
|
93
|
-
return cls.model_validate(obj)
|
|
93
|
+
return cls.model_validate(obj, strict=False)
|
|
94
94
|
|
|
95
95
|
_obj = cls.model_validate({
|
|
96
96
|
"tokenization": obj.get("tokenization"),
|
|
97
97
|
"data": [V1LeaderboardReportActualOutputData.from_dict(_item) for _item in obj["data"]] if obj.get("data") is not None else None
|
|
98
|
-
})
|
|
98
|
+
}, strict=False)
|
|
99
99
|
return _obj
|
|
100
100
|
|
|
101
101
|
|
|
@@ -106,23 +106,23 @@ class V1LeaderboardReportEvaluator(BaseModel):
|
|
|
106
106
|
# override the default output from pydantic by calling `to_dict()` of each item in explanations (list)
|
|
107
107
|
_items = []
|
|
108
108
|
if self.explanations:
|
|
109
|
-
for
|
|
110
|
-
if
|
|
111
|
-
_items.append(
|
|
109
|
+
for _item_explanations in self.explanations:
|
|
110
|
+
if _item_explanations:
|
|
111
|
+
_items.append(_item_explanations.to_dict())
|
|
112
112
|
_dict['explanations'] = _items
|
|
113
113
|
# override the default output from pydantic by calling `to_dict()` of each item in parameters (list)
|
|
114
114
|
_items = []
|
|
115
115
|
if self.parameters:
|
|
116
|
-
for
|
|
117
|
-
if
|
|
118
|
-
_items.append(
|
|
116
|
+
for _item_parameters in self.parameters:
|
|
117
|
+
if _item_parameters:
|
|
118
|
+
_items.append(_item_parameters.to_dict())
|
|
119
119
|
_dict['parameters'] = _items
|
|
120
120
|
# override the default output from pydantic by calling `to_dict()` of each item in metrics_meta (list)
|
|
121
121
|
_items = []
|
|
122
122
|
if self.metrics_meta:
|
|
123
|
-
for
|
|
124
|
-
if
|
|
125
|
-
_items.append(
|
|
123
|
+
for _item_metrics_meta in self.metrics_meta:
|
|
124
|
+
if _item_metrics_meta:
|
|
125
|
+
_items.append(_item_metrics_meta.to_dict())
|
|
126
126
|
_dict['metricsMeta'] = _items
|
|
127
127
|
return _dict
|
|
128
128
|
|
|
@@ -133,7 +133,7 @@ class V1LeaderboardReportEvaluator(BaseModel):
|
|
|
133
133
|
return None
|
|
134
134
|
|
|
135
135
|
if not isinstance(obj, dict):
|
|
136
|
-
return cls.model_validate(obj)
|
|
136
|
+
return cls.model_validate(obj, strict=False)
|
|
137
137
|
|
|
138
138
|
_obj = cls.model_validate({
|
|
139
139
|
"id": obj.get("id"),
|
|
@@ -149,7 +149,7 @@ class V1LeaderboardReportEvaluator(BaseModel):
|
|
|
149
149
|
"parameters": [V1LeaderboardReportEvaluatorParameter.from_dict(_item) for _item in obj["parameters"]] if obj.get("parameters") is not None else None,
|
|
150
150
|
"keywords": obj.get("keywords"),
|
|
151
151
|
"metricsMeta": [V1LeaderboardReportMetricsMetaEntry.from_dict(_item) for _item in obj["metricsMeta"]] if obj.get("metricsMeta") is not None else None
|
|
152
|
-
})
|
|
152
|
+
}, strict=False)
|
|
153
153
|
return _obj
|
|
154
154
|
|
|
155
155
|
|
|
@@ -91,7 +91,7 @@ class V1LeaderboardReportEvaluatorParameter(BaseModel):
|
|
|
91
91
|
return None
|
|
92
92
|
|
|
93
93
|
if not isinstance(obj, dict):
|
|
94
|
-
return cls.model_validate(obj)
|
|
94
|
+
return cls.model_validate(obj, strict=False)
|
|
95
95
|
|
|
96
96
|
_obj = cls.model_validate({
|
|
97
97
|
"name": obj.get("name"),
|
|
@@ -103,7 +103,7 @@ class V1LeaderboardReportEvaluatorParameter(BaseModel):
|
|
|
103
103
|
"min": obj.get("min"),
|
|
104
104
|
"max": obj.get("max"),
|
|
105
105
|
"category": obj.get("category")
|
|
106
|
-
})
|
|
106
|
+
}, strict=False)
|
|
107
107
|
return _obj
|
|
108
108
|
|
|
109
109
|
|
|
@@ -88,7 +88,7 @@ class V1LeaderboardReportExplanation(BaseModel):
|
|
|
88
88
|
return None
|
|
89
89
|
|
|
90
90
|
if not isinstance(obj, dict):
|
|
91
|
-
return cls.model_validate(obj)
|
|
91
|
+
return cls.model_validate(obj, strict=False)
|
|
92
92
|
|
|
93
93
|
_obj = cls.model_validate({
|
|
94
94
|
"explanationType": obj.get("explanationType"),
|
|
@@ -97,7 +97,7 @@ class V1LeaderboardReportExplanation(BaseModel):
|
|
|
97
97
|
"scope": obj.get("scope"),
|
|
98
98
|
"hasLocal": obj.get("hasLocal"),
|
|
99
99
|
"formats": obj.get("formats")
|
|
100
|
-
})
|
|
100
|
+
}, strict=False)
|
|
101
101
|
return _obj
|
|
102
102
|
|
|
103
103
|
|
|
@@ -108,7 +108,7 @@ class V1LeaderboardReportMetricsMetaEntry(BaseModel):
|
|
|
108
108
|
return None
|
|
109
109
|
|
|
110
110
|
if not isinstance(obj, dict):
|
|
111
|
-
return cls.model_validate(obj)
|
|
111
|
+
return cls.model_validate(obj, strict=False)
|
|
112
112
|
|
|
113
113
|
_obj = cls.model_validate({
|
|
114
114
|
"key": obj.get("key"),
|
|
@@ -123,7 +123,7 @@ class V1LeaderboardReportMetricsMetaEntry(BaseModel):
|
|
|
123
123
|
"isPrimaryMetric": obj.get("isPrimaryMetric"),
|
|
124
124
|
"parentMetric": obj.get("parentMetric"),
|
|
125
125
|
"exclude": obj.get("exclude")
|
|
126
|
-
})
|
|
126
|
+
}, strict=False)
|
|
127
127
|
return _obj
|
|
128
128
|
|
|
129
129
|
|
|
@@ -96,7 +96,7 @@ class V1LeaderboardReportModel(BaseModel):
|
|
|
96
96
|
return None
|
|
97
97
|
|
|
98
98
|
if not isinstance(obj, dict):
|
|
99
|
-
return cls.model_validate(obj)
|
|
99
|
+
return cls.model_validate(obj, strict=False)
|
|
100
100
|
|
|
101
101
|
_obj = cls.model_validate({
|
|
102
102
|
"connection": obj.get("connection"),
|
|
@@ -107,7 +107,7 @@ class V1LeaderboardReportModel(BaseModel):
|
|
|
107
107
|
"llmModelName": obj.get("llmModelName"),
|
|
108
108
|
"documents": obj.get("documents"),
|
|
109
109
|
"key": obj.get("key")
|
|
110
|
-
})
|
|
110
|
+
}, strict=False)
|
|
111
111
|
return _obj
|
|
112
112
|
|
|
113
113
|
|
|
@@ -122,23 +122,23 @@ class V1LeaderboardReportResult(BaseModel):
|
|
|
122
122
|
# override the default output from pydantic by calling `to_dict()` of each item in relationships (list)
|
|
123
123
|
_items = []
|
|
124
124
|
if self.relationships:
|
|
125
|
-
for
|
|
126
|
-
if
|
|
127
|
-
_items.append(
|
|
125
|
+
for _item_relationships in self.relationships:
|
|
126
|
+
if _item_relationships:
|
|
127
|
+
_items.append(_item_relationships.to_dict())
|
|
128
128
|
_dict['relationships'] = _items
|
|
129
129
|
# override the default output from pydantic by calling `to_dict()` of each item in metrics (list)
|
|
130
130
|
_items = []
|
|
131
131
|
if self.metrics:
|
|
132
|
-
for
|
|
133
|
-
if
|
|
134
|
-
_items.append(
|
|
132
|
+
for _item_metrics in self.metrics:
|
|
133
|
+
if _item_metrics:
|
|
134
|
+
_items.append(_item_metrics.to_dict())
|
|
135
135
|
_dict['metrics'] = _items
|
|
136
136
|
# override the default output from pydantic by calling `to_dict()` of each item in actual_output_meta (list)
|
|
137
137
|
_items = []
|
|
138
138
|
if self.actual_output_meta:
|
|
139
|
-
for
|
|
140
|
-
if
|
|
141
|
-
_items.append(
|
|
139
|
+
for _item_actual_output_meta in self.actual_output_meta:
|
|
140
|
+
if _item_actual_output_meta:
|
|
141
|
+
_items.append(_item_actual_output_meta.to_dict())
|
|
142
142
|
_dict['actualOutputMeta'] = _items
|
|
143
143
|
return _dict
|
|
144
144
|
|
|
@@ -149,7 +149,7 @@ class V1LeaderboardReportResult(BaseModel):
|
|
|
149
149
|
return None
|
|
150
150
|
|
|
151
151
|
if not isinstance(obj, dict):
|
|
152
|
-
return cls.model_validate(obj)
|
|
152
|
+
return cls.model_validate(obj, strict=False)
|
|
153
153
|
|
|
154
154
|
_obj = cls.model_validate({
|
|
155
155
|
"key": obj.get("key"),
|
|
@@ -169,7 +169,7 @@ class V1LeaderboardReportResult(BaseModel):
|
|
|
169
169
|
"metrics": [V1MetricScore.from_dict(_item) for _item in obj["metrics"]] if obj.get("metrics") is not None else None,
|
|
170
170
|
"resultErrorMessage": obj.get("resultErrorMessage"),
|
|
171
171
|
"actualOutputMeta": [V1LeaderboardReportActualOutputMeta.from_dict(_item) for _item in obj["actualOutputMeta"]] if obj.get("actualOutputMeta") is not None else None
|
|
172
|
-
})
|
|
172
|
+
}, strict=False)
|
|
173
173
|
return _obj
|
|
174
174
|
|
|
175
175
|
|
|
@@ -85,13 +85,13 @@ class V1LeaderboardReportResultRelationship(BaseModel):
|
|
|
85
85
|
return None
|
|
86
86
|
|
|
87
87
|
if not isinstance(obj, dict):
|
|
88
|
-
return cls.model_validate(obj)
|
|
88
|
+
return cls.model_validate(obj, strict=False)
|
|
89
89
|
|
|
90
90
|
_obj = cls.model_validate({
|
|
91
91
|
"type": obj.get("type"),
|
|
92
92
|
"target": obj.get("target"),
|
|
93
93
|
"targetType": obj.get("targetType")
|
|
94
|
-
})
|
|
94
|
+
}, strict=False)
|
|
95
95
|
return _obj
|
|
96
96
|
|
|
97
97
|
|
|
@@ -77,11 +77,11 @@ class V1ListBaseModelsResponse(BaseModel):
|
|
|
77
77
|
return None
|
|
78
78
|
|
|
79
79
|
if not isinstance(obj, dict):
|
|
80
|
-
return cls.model_validate(obj)
|
|
80
|
+
return cls.model_validate(obj, strict=False)
|
|
81
81
|
|
|
82
82
|
_obj = cls.model_validate({
|
|
83
83
|
"baseModels": obj.get("baseModels")
|
|
84
|
-
})
|
|
84
|
+
}, strict=False)
|
|
85
85
|
return _obj
|
|
86
86
|
|
|
87
87
|
|
|
@@ -72,9 +72,9 @@ class V1ListDashboardAccessResponse(BaseModel):
|
|
|
72
72
|
# override the default output from pydantic by calling `to_dict()` of each item in role_bindings (list)
|
|
73
73
|
_items = []
|
|
74
74
|
if self.role_bindings:
|
|
75
|
-
for
|
|
76
|
-
if
|
|
77
|
-
_items.append(
|
|
75
|
+
for _item_role_bindings in self.role_bindings:
|
|
76
|
+
if _item_role_bindings:
|
|
77
|
+
_items.append(_item_role_bindings.to_dict())
|
|
78
78
|
_dict['roleBindings'] = _items
|
|
79
79
|
return _dict
|
|
80
80
|
|
|
@@ -85,11 +85,11 @@ class V1ListDashboardAccessResponse(BaseModel):
|
|
|
85
85
|
return None
|
|
86
86
|
|
|
87
87
|
if not isinstance(obj, dict):
|
|
88
|
-
return cls.model_validate(obj)
|
|
88
|
+
return cls.model_validate(obj, strict=False)
|
|
89
89
|
|
|
90
90
|
_obj = cls.model_validate({
|
|
91
91
|
"roleBindings": [V1RoleBinding.from_dict(_item) for _item in obj["roleBindings"]] if obj.get("roleBindings") is not None else None
|
|
92
|
-
})
|
|
92
|
+
}, strict=False)
|
|
93
93
|
return _obj
|
|
94
94
|
|
|
95
95
|
|
|
@@ -72,9 +72,9 @@ class V1ListDashboardsResponse(BaseModel):
|
|
|
72
72
|
# override the default output from pydantic by calling `to_dict()` of each item in dashboards (list)
|
|
73
73
|
_items = []
|
|
74
74
|
if self.dashboards:
|
|
75
|
-
for
|
|
76
|
-
if
|
|
77
|
-
_items.append(
|
|
75
|
+
for _item_dashboards in self.dashboards:
|
|
76
|
+
if _item_dashboards:
|
|
77
|
+
_items.append(_item_dashboards.to_dict())
|
|
78
78
|
_dict['dashboards'] = _items
|
|
79
79
|
return _dict
|
|
80
80
|
|
|
@@ -85,11 +85,11 @@ class V1ListDashboardsResponse(BaseModel):
|
|
|
85
85
|
return None
|
|
86
86
|
|
|
87
87
|
if not isinstance(obj, dict):
|
|
88
|
-
return cls.model_validate(obj)
|
|
88
|
+
return cls.model_validate(obj, strict=False)
|
|
89
89
|
|
|
90
90
|
_obj = cls.model_validate({
|
|
91
91
|
"dashboards": [V1Dashboard.from_dict(_item) for _item in obj["dashboards"]] if obj.get("dashboards") is not None else None
|
|
92
|
-
})
|
|
92
|
+
}, strict=False)
|
|
93
93
|
return _obj
|
|
94
94
|
|
|
95
95
|
|
|
@@ -72,9 +72,9 @@ class V1ListDashboardsSharedWithMeResponse(BaseModel):
|
|
|
72
72
|
# override the default output from pydantic by calling `to_dict()` of each item in dashboards (list)
|
|
73
73
|
_items = []
|
|
74
74
|
if self.dashboards:
|
|
75
|
-
for
|
|
76
|
-
if
|
|
77
|
-
_items.append(
|
|
75
|
+
for _item_dashboards in self.dashboards:
|
|
76
|
+
if _item_dashboards:
|
|
77
|
+
_items.append(_item_dashboards.to_dict())
|
|
78
78
|
_dict['dashboards'] = _items
|
|
79
79
|
return _dict
|
|
80
80
|
|
|
@@ -85,11 +85,11 @@ class V1ListDashboardsSharedWithMeResponse(BaseModel):
|
|
|
85
85
|
return None
|
|
86
86
|
|
|
87
87
|
if not isinstance(obj, dict):
|
|
88
|
-
return cls.model_validate(obj)
|
|
88
|
+
return cls.model_validate(obj, strict=False)
|
|
89
89
|
|
|
90
90
|
_obj = cls.model_validate({
|
|
91
91
|
"dashboards": [V1Dashboard.from_dict(_item) for _item in obj["dashboards"]] if obj.get("dashboards") is not None else None
|
|
92
|
-
})
|
|
92
|
+
}, strict=False)
|
|
93
93
|
return _obj
|
|
94
94
|
|
|
95
95
|
|
|
@@ -72,9 +72,9 @@ class V1ListDocumentsResponse(BaseModel):
|
|
|
72
72
|
# override the default output from pydantic by calling `to_dict()` of each item in documents (list)
|
|
73
73
|
_items = []
|
|
74
74
|
if self.documents:
|
|
75
|
-
for
|
|
76
|
-
if
|
|
77
|
-
_items.append(
|
|
75
|
+
for _item_documents in self.documents:
|
|
76
|
+
if _item_documents:
|
|
77
|
+
_items.append(_item_documents.to_dict())
|
|
78
78
|
_dict['documents'] = _items
|
|
79
79
|
return _dict
|
|
80
80
|
|
|
@@ -85,11 +85,11 @@ class V1ListDocumentsResponse(BaseModel):
|
|
|
85
85
|
return None
|
|
86
86
|
|
|
87
87
|
if not isinstance(obj, dict):
|
|
88
|
-
return cls.model_validate(obj)
|
|
88
|
+
return cls.model_validate(obj, strict=False)
|
|
89
89
|
|
|
90
90
|
_obj = cls.model_validate({
|
|
91
91
|
"documents": [V1Document.from_dict(_item) for _item in obj["documents"]] if obj.get("documents") is not None else None
|
|
92
|
-
})
|
|
92
|
+
}, strict=False)
|
|
93
93
|
return _obj
|
|
94
94
|
|
|
95
95
|
|
|
@@ -72,9 +72,9 @@ class V1ListEvaluatorsResponse(BaseModel):
|
|
|
72
72
|
# override the default output from pydantic by calling `to_dict()` of each item in evaluators (list)
|
|
73
73
|
_items = []
|
|
74
74
|
if self.evaluators:
|
|
75
|
-
for
|
|
76
|
-
if
|
|
77
|
-
_items.append(
|
|
75
|
+
for _item_evaluators in self.evaluators:
|
|
76
|
+
if _item_evaluators:
|
|
77
|
+
_items.append(_item_evaluators.to_dict())
|
|
78
78
|
_dict['evaluators'] = _items
|
|
79
79
|
return _dict
|
|
80
80
|
|
|
@@ -85,11 +85,11 @@ class V1ListEvaluatorsResponse(BaseModel):
|
|
|
85
85
|
return None
|
|
86
86
|
|
|
87
87
|
if not isinstance(obj, dict):
|
|
88
|
-
return cls.model_validate(obj)
|
|
88
|
+
return cls.model_validate(obj, strict=False)
|
|
89
89
|
|
|
90
90
|
_obj = cls.model_validate({
|
|
91
91
|
"evaluators": [V1Evaluator.from_dict(_item) for _item in obj["evaluators"]] if obj.get("evaluators") is not None else None
|
|
92
|
-
})
|
|
92
|
+
}, strict=False)
|
|
93
93
|
return _obj
|
|
94
94
|
|
|
95
95
|
|
|
@@ -73,9 +73,9 @@ class V1ListLeaderboardsResponse(BaseModel):
|
|
|
73
73
|
# override the default output from pydantic by calling `to_dict()` of each item in leaderboards (list)
|
|
74
74
|
_items = []
|
|
75
75
|
if self.leaderboards:
|
|
76
|
-
for
|
|
77
|
-
if
|
|
78
|
-
_items.append(
|
|
76
|
+
for _item_leaderboards in self.leaderboards:
|
|
77
|
+
if _item_leaderboards:
|
|
78
|
+
_items.append(_item_leaderboards.to_dict())
|
|
79
79
|
_dict['leaderboards'] = _items
|
|
80
80
|
return _dict
|
|
81
81
|
|
|
@@ -86,12 +86,12 @@ class V1ListLeaderboardsResponse(BaseModel):
|
|
|
86
86
|
return None
|
|
87
87
|
|
|
88
88
|
if not isinstance(obj, dict):
|
|
89
|
-
return cls.model_validate(obj)
|
|
89
|
+
return cls.model_validate(obj, strict=False)
|
|
90
90
|
|
|
91
91
|
_obj = cls.model_validate({
|
|
92
92
|
"leaderboards": [V1Leaderboard.from_dict(_item) for _item in obj["leaderboards"]] if obj.get("leaderboards") is not None else None,
|
|
93
93
|
"nextPageToken": obj.get("nextPageToken")
|
|
94
|
-
})
|
|
94
|
+
}, strict=False)
|
|
95
95
|
return _obj
|
|
96
96
|
|
|
97
97
|
|
|
@@ -77,11 +77,11 @@ class V1ListLLMModelsResponse(BaseModel):
|
|
|
77
77
|
return None
|
|
78
78
|
|
|
79
79
|
if not isinstance(obj, dict):
|
|
80
|
-
return cls.model_validate(obj)
|
|
80
|
+
return cls.model_validate(obj, strict=False)
|
|
81
81
|
|
|
82
82
|
_obj = cls.model_validate({
|
|
83
83
|
"models": obj.get("models")
|
|
84
|
-
})
|
|
84
|
+
}, strict=False)
|
|
85
85
|
return _obj
|
|
86
86
|
|
|
87
87
|
|
|
@@ -72,9 +72,9 @@ class V1ListModelCollectionsResponse(BaseModel):
|
|
|
72
72
|
# override the default output from pydantic by calling `to_dict()` of each item in collections (list)
|
|
73
73
|
_items = []
|
|
74
74
|
if self.collections:
|
|
75
|
-
for
|
|
76
|
-
if
|
|
77
|
-
_items.append(
|
|
75
|
+
for _item_collections in self.collections:
|
|
76
|
+
if _item_collections:
|
|
77
|
+
_items.append(_item_collections.to_dict())
|
|
78
78
|
_dict['collections'] = _items
|
|
79
79
|
return _dict
|
|
80
80
|
|
|
@@ -85,11 +85,11 @@ class V1ListModelCollectionsResponse(BaseModel):
|
|
|
85
85
|
return None
|
|
86
86
|
|
|
87
87
|
if not isinstance(obj, dict):
|
|
88
|
-
return cls.model_validate(obj)
|
|
88
|
+
return cls.model_validate(obj, strict=False)
|
|
89
89
|
|
|
90
90
|
_obj = cls.model_validate({
|
|
91
91
|
"collections": [V1CollectionInfo.from_dict(_item) for _item in obj["collections"]] if obj.get("collections") is not None else None
|
|
92
|
-
})
|
|
92
|
+
}, strict=False)
|
|
93
93
|
return _obj
|
|
94
94
|
|
|
95
95
|
|
|
@@ -72,9 +72,9 @@ class V1ListModelsResponse(BaseModel):
|
|
|
72
72
|
# override the default output from pydantic by calling `to_dict()` of each item in models (list)
|
|
73
73
|
_items = []
|
|
74
74
|
if self.models:
|
|
75
|
-
for
|
|
76
|
-
if
|
|
77
|
-
_items.append(
|
|
75
|
+
for _item_models in self.models:
|
|
76
|
+
if _item_models:
|
|
77
|
+
_items.append(_item_models.to_dict())
|
|
78
78
|
_dict['models'] = _items
|
|
79
79
|
return _dict
|
|
80
80
|
|
|
@@ -85,11 +85,11 @@ class V1ListModelsResponse(BaseModel):
|
|
|
85
85
|
return None
|
|
86
86
|
|
|
87
87
|
if not isinstance(obj, dict):
|
|
88
|
-
return cls.model_validate(obj)
|
|
88
|
+
return cls.model_validate(obj, strict=False)
|
|
89
89
|
|
|
90
90
|
_obj = cls.model_validate({
|
|
91
91
|
"models": [V1Model.from_dict(_item) for _item in obj["models"]] if obj.get("models") is not None else None
|
|
92
|
-
})
|
|
92
|
+
}, strict=False)
|
|
93
93
|
return _obj
|
|
94
94
|
|
|
95
95
|
|
|
@@ -72,9 +72,9 @@ class V1ListMostRecentDashboardsResponse(BaseModel):
|
|
|
72
72
|
# override the default output from pydantic by calling `to_dict()` of each item in dashboards (list)
|
|
73
73
|
_items = []
|
|
74
74
|
if self.dashboards:
|
|
75
|
-
for
|
|
76
|
-
if
|
|
77
|
-
_items.append(
|
|
75
|
+
for _item_dashboards in self.dashboards:
|
|
76
|
+
if _item_dashboards:
|
|
77
|
+
_items.append(_item_dashboards.to_dict())
|
|
78
78
|
_dict['dashboards'] = _items
|
|
79
79
|
return _dict
|
|
80
80
|
|
|
@@ -85,11 +85,11 @@ class V1ListMostRecentDashboardsResponse(BaseModel):
|
|
|
85
85
|
return None
|
|
86
86
|
|
|
87
87
|
if not isinstance(obj, dict):
|
|
88
|
-
return cls.model_validate(obj)
|
|
88
|
+
return cls.model_validate(obj, strict=False)
|
|
89
89
|
|
|
90
90
|
_obj = cls.model_validate({
|
|
91
91
|
"dashboards": [V1Dashboard.from_dict(_item) for _item in obj["dashboards"]] if obj.get("dashboards") is not None else None
|
|
92
|
-
})
|
|
92
|
+
}, strict=False)
|
|
93
93
|
return _obj
|
|
94
94
|
|
|
95
95
|
|
|
@@ -72,9 +72,9 @@ class V1ListMostRecentLeaderboardsResponse(BaseModel):
|
|
|
72
72
|
# override the default output from pydantic by calling `to_dict()` of each item in leaderboards (list)
|
|
73
73
|
_items = []
|
|
74
74
|
if self.leaderboards:
|
|
75
|
-
for
|
|
76
|
-
if
|
|
77
|
-
_items.append(
|
|
75
|
+
for _item_leaderboards in self.leaderboards:
|
|
76
|
+
if _item_leaderboards:
|
|
77
|
+
_items.append(_item_leaderboards.to_dict())
|
|
78
78
|
_dict['leaderboards'] = _items
|
|
79
79
|
return _dict
|
|
80
80
|
|
|
@@ -85,11 +85,11 @@ class V1ListMostRecentLeaderboardsResponse(BaseModel):
|
|
|
85
85
|
return None
|
|
86
86
|
|
|
87
87
|
if not isinstance(obj, dict):
|
|
88
|
-
return cls.model_validate(obj)
|
|
88
|
+
return cls.model_validate(obj, strict=False)
|
|
89
89
|
|
|
90
90
|
_obj = cls.model_validate({
|
|
91
91
|
"leaderboards": [V1Leaderboard.from_dict(_item) for _item in obj["leaderboards"]] if obj.get("leaderboards") is not None else None
|
|
92
|
-
})
|
|
92
|
+
}, strict=False)
|
|
93
93
|
return _obj
|
|
94
94
|
|
|
95
95
|
|
|
@@ -72,9 +72,9 @@ class V1ListMostRecentModelsResponse(BaseModel):
|
|
|
72
72
|
# override the default output from pydantic by calling `to_dict()` of each item in models (list)
|
|
73
73
|
_items = []
|
|
74
74
|
if self.models:
|
|
75
|
-
for
|
|
76
|
-
if
|
|
77
|
-
_items.append(
|
|
75
|
+
for _item_models in self.models:
|
|
76
|
+
if _item_models:
|
|
77
|
+
_items.append(_item_models.to_dict())
|
|
78
78
|
_dict['models'] = _items
|
|
79
79
|
return _dict
|
|
80
80
|
|
|
@@ -85,11 +85,11 @@ class V1ListMostRecentModelsResponse(BaseModel):
|
|
|
85
85
|
return None
|
|
86
86
|
|
|
87
87
|
if not isinstance(obj, dict):
|
|
88
|
-
return cls.model_validate(obj)
|
|
88
|
+
return cls.model_validate(obj, strict=False)
|
|
89
89
|
|
|
90
90
|
_obj = cls.model_validate({
|
|
91
91
|
"models": [V1Model.from_dict(_item) for _item in obj["models"]] if obj.get("models") is not None else None
|
|
92
|
-
})
|
|
92
|
+
}, strict=False)
|
|
93
93
|
return _obj
|
|
94
94
|
|
|
95
95
|
|
|
@@ -72,9 +72,9 @@ class V1ListMostRecentTestsResponse(BaseModel):
|
|
|
72
72
|
# override the default output from pydantic by calling `to_dict()` of each item in tests (list)
|
|
73
73
|
_items = []
|
|
74
74
|
if self.tests:
|
|
75
|
-
for
|
|
76
|
-
if
|
|
77
|
-
_items.append(
|
|
75
|
+
for _item_tests in self.tests:
|
|
76
|
+
if _item_tests:
|
|
77
|
+
_items.append(_item_tests.to_dict())
|
|
78
78
|
_dict['tests'] = _items
|
|
79
79
|
return _dict
|
|
80
80
|
|
|
@@ -85,11 +85,11 @@ class V1ListMostRecentTestsResponse(BaseModel):
|
|
|
85
85
|
return None
|
|
86
86
|
|
|
87
87
|
if not isinstance(obj, dict):
|
|
88
|
-
return cls.model_validate(obj)
|
|
88
|
+
return cls.model_validate(obj, strict=False)
|
|
89
89
|
|
|
90
90
|
_obj = cls.model_validate({
|
|
91
91
|
"tests": [V1Test.from_dict(_item) for _item in obj["tests"]] if obj.get("tests") is not None else None
|
|
92
|
-
})
|
|
92
|
+
}, strict=False)
|
|
93
93
|
return _obj
|
|
94
94
|
|
|
95
95
|
|