eval-studio-client 1.2.2a1__py3-none-any.whl → 1.2.2a2__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (213) hide show
  1. eval_studio_client/api/models/adversarial_inputs_service_test_adversarial_inputs_robustness_request.py +5 -5
  2. eval_studio_client/api/models/perturbation_service_create_perturbation_request.py +11 -11
  3. eval_studio_client/api/models/prompt_generation_service_auto_generate_prompts_request.py +2 -2
  4. eval_studio_client/api/models/protobuf_any.py +2 -2
  5. eval_studio_client/api/models/required_the_dashboard_to_update.py +2 -2
  6. eval_studio_client/api/models/required_the_document_to_update.py +2 -2
  7. eval_studio_client/api/models/required_the_leaderboard_to_update.py +8 -8
  8. eval_studio_client/api/models/required_the_model_to_update.py +2 -2
  9. eval_studio_client/api/models/required_the_operation_to_finalize.py +2 -2
  10. eval_studio_client/api/models/required_the_operation_to_update.py +2 -2
  11. eval_studio_client/api/models/required_the_test_case_to_update.py +2 -2
  12. eval_studio_client/api/models/required_the_test_to_update.py +2 -2
  13. eval_studio_client/api/models/required_the_updated_workflow.py +5 -5
  14. eval_studio_client/api/models/required_the_updated_workflow_node.py +5 -5
  15. eval_studio_client/api/models/rpc_status.py +5 -5
  16. eval_studio_client/api/models/test_case_service_batch_delete_test_cases_request.py +2 -2
  17. eval_studio_client/api/models/test_service_clone_test_request.py +2 -2
  18. eval_studio_client/api/models/test_service_generate_test_cases_request.py +5 -5
  19. eval_studio_client/api/models/test_service_grant_test_access_request.py +2 -2
  20. eval_studio_client/api/models/test_service_import_test_cases_from_library_request.py +2 -2
  21. eval_studio_client/api/models/test_service_list_test_case_library_items_request.py +2 -2
  22. eval_studio_client/api/models/test_service_perturb_test_in_place_request.py +5 -5
  23. eval_studio_client/api/models/test_service_perturb_test_request.py +5 -5
  24. eval_studio_client/api/models/test_service_revoke_test_access_request.py +2 -2
  25. eval_studio_client/api/models/v1_abort_operation_response.py +2 -2
  26. eval_studio_client/api/models/v1_batch_create_leaderboards_request.py +5 -5
  27. eval_studio_client/api/models/v1_batch_create_leaderboards_response.py +2 -2
  28. eval_studio_client/api/models/v1_batch_delete_dashboards_request.py +2 -2
  29. eval_studio_client/api/models/v1_batch_delete_dashboards_response.py +5 -5
  30. eval_studio_client/api/models/v1_batch_delete_documents_request.py +2 -2
  31. eval_studio_client/api/models/v1_batch_delete_documents_response.py +5 -5
  32. eval_studio_client/api/models/v1_batch_delete_evaluators_request.py +2 -2
  33. eval_studio_client/api/models/v1_batch_delete_evaluators_response.py +5 -5
  34. eval_studio_client/api/models/v1_batch_delete_leaderboards_request.py +2 -2
  35. eval_studio_client/api/models/v1_batch_delete_leaderboards_response.py +5 -5
  36. eval_studio_client/api/models/v1_batch_delete_models_request.py +2 -2
  37. eval_studio_client/api/models/v1_batch_delete_models_response.py +5 -5
  38. eval_studio_client/api/models/v1_batch_delete_test_cases_response.py +5 -5
  39. eval_studio_client/api/models/v1_batch_delete_tests_request.py +2 -2
  40. eval_studio_client/api/models/v1_batch_delete_tests_response.py +5 -5
  41. eval_studio_client/api/models/v1_batch_delete_workflows_request.py +2 -2
  42. eval_studio_client/api/models/v1_batch_delete_workflows_response.py +5 -5
  43. eval_studio_client/api/models/v1_batch_get_dashboards_response.py +5 -5
  44. eval_studio_client/api/models/v1_batch_get_documents_response.py +5 -5
  45. eval_studio_client/api/models/v1_batch_get_leaderboards_response.py +5 -5
  46. eval_studio_client/api/models/v1_batch_get_models_response.py +5 -5
  47. eval_studio_client/api/models/v1_batch_get_operations_response.py +5 -5
  48. eval_studio_client/api/models/v1_batch_get_tests_response.py +5 -5
  49. eval_studio_client/api/models/v1_batch_get_workflow_edges_response.py +5 -5
  50. eval_studio_client/api/models/v1_batch_get_workflow_nodes_response.py +5 -5
  51. eval_studio_client/api/models/v1_batch_import_leaderboard_request.py +2 -2
  52. eval_studio_client/api/models/v1_batch_import_leaderboard_response.py +2 -2
  53. eval_studio_client/api/models/v1_batch_import_tests_request.py +2 -2
  54. eval_studio_client/api/models/v1_batch_import_tests_response.py +5 -5
  55. eval_studio_client/api/models/v1_check_base_models_response.py +2 -2
  56. eval_studio_client/api/models/v1_clone_test_response.py +2 -2
  57. eval_studio_client/api/models/v1_clone_workflow_response.py +2 -2
  58. eval_studio_client/api/models/v1_collection_info.py +2 -2
  59. eval_studio_client/api/models/v1_context.py +2 -2
  60. eval_studio_client/api/models/v1_create_dashboard_response.py +2 -2
  61. eval_studio_client/api/models/v1_create_document_response.py +2 -2
  62. eval_studio_client/api/models/v1_create_evaluation_request.py +5 -5
  63. eval_studio_client/api/models/v1_create_evaluator_response.py +2 -2
  64. eval_studio_client/api/models/v1_create_leaderboard_request.py +2 -2
  65. eval_studio_client/api/models/v1_create_leaderboard_response.py +2 -2
  66. eval_studio_client/api/models/v1_create_leaderboard_without_cache_response.py +2 -2
  67. eval_studio_client/api/models/v1_create_model_response.py +2 -2
  68. eval_studio_client/api/models/v1_create_perturbation_response.py +2 -2
  69. eval_studio_client/api/models/v1_create_test_case_response.py +2 -2
  70. eval_studio_client/api/models/v1_create_test_lab_response.py +2 -2
  71. eval_studio_client/api/models/v1_create_test_response.py +2 -2
  72. eval_studio_client/api/models/v1_create_workflow_edge_response.py +2 -2
  73. eval_studio_client/api/models/v1_create_workflow_node_response.py +2 -2
  74. eval_studio_client/api/models/v1_create_workflow_response.py +2 -2
  75. eval_studio_client/api/models/v1_dashboard.py +2 -2
  76. eval_studio_client/api/models/v1_delete_dashboard_response.py +2 -2
  77. eval_studio_client/api/models/v1_delete_document_response.py +2 -2
  78. eval_studio_client/api/models/v1_delete_evaluator_response.py +2 -2
  79. eval_studio_client/api/models/v1_delete_leaderboard_response.py +2 -2
  80. eval_studio_client/api/models/v1_delete_model_response.py +2 -2
  81. eval_studio_client/api/models/v1_delete_test_case_response.py +2 -2
  82. eval_studio_client/api/models/v1_delete_test_response.py +2 -2
  83. eval_studio_client/api/models/v1_delete_workflow_edge_response.py +2 -2
  84. eval_studio_client/api/models/v1_delete_workflow_node_response.py +2 -2
  85. eval_studio_client/api/models/v1_delete_workflow_response.py +2 -2
  86. eval_studio_client/api/models/v1_dependency_list.py +5 -5
  87. eval_studio_client/api/models/v1_document.py +2 -2
  88. eval_studio_client/api/models/v1_estimate_threshold_request.py +5 -5
  89. eval_studio_client/api/models/v1_evaluation_test.py +8 -8
  90. eval_studio_client/api/models/v1_evaluator.py +5 -5
  91. eval_studio_client/api/models/v1_evaluator_parameter.py +2 -2
  92. eval_studio_client/api/models/v1_finalize_operation_response.py +2 -2
  93. eval_studio_client/api/models/v1_find_all_test_cases_by_id_response.py +5 -5
  94. eval_studio_client/api/models/v1_find_test_lab_response.py +2 -2
  95. eval_studio_client/api/models/v1_find_workflows_by_collection_id_response.py +5 -5
  96. eval_studio_client/api/models/v1_generate_test_cases_response.py +2 -2
  97. eval_studio_client/api/models/v1_get_dashboard_response.py +2 -2
  98. eval_studio_client/api/models/v1_get_document_response.py +2 -2
  99. eval_studio_client/api/models/v1_get_evaluator_response.py +2 -2
  100. eval_studio_client/api/models/v1_get_guardrails_configuration_response.py +2 -2
  101. eval_studio_client/api/models/v1_get_info_response.py +2 -2
  102. eval_studio_client/api/models/v1_get_leaderboard_report_response.py +2 -2
  103. eval_studio_client/api/models/v1_get_leaderboard_response.py +2 -2
  104. eval_studio_client/api/models/v1_get_model_response.py +2 -2
  105. eval_studio_client/api/models/v1_get_operation_progress_by_parent_response.py +2 -2
  106. eval_studio_client/api/models/v1_get_operation_response.py +2 -2
  107. eval_studio_client/api/models/v1_get_perturbator_response.py +2 -2
  108. eval_studio_client/api/models/v1_get_stats_response.py +2 -2
  109. eval_studio_client/api/models/v1_get_test_case_response.py +2 -2
  110. eval_studio_client/api/models/v1_get_test_class_response.py +2 -2
  111. eval_studio_client/api/models/v1_get_test_response.py +2 -2
  112. eval_studio_client/api/models/v1_get_workflow_node_prerequisites_response.py +2 -2
  113. eval_studio_client/api/models/v1_get_workflow_node_response.py +2 -2
  114. eval_studio_client/api/models/v1_get_workflow_response.py +2 -2
  115. eval_studio_client/api/models/v1_get_workflow_result_corpus_patch_response.py +2 -2
  116. eval_studio_client/api/models/v1_get_workflow_result_report_response.py +2 -2
  117. eval_studio_client/api/models/v1_get_workflow_result_summary_response.py +2 -2
  118. eval_studio_client/api/models/v1_get_workflow_result_system_prompt_patch_response.py +2 -2
  119. eval_studio_client/api/models/v1_import_evaluation_request.py +2 -2
  120. eval_studio_client/api/models/v1_import_leaderboard_request.py +2 -2
  121. eval_studio_client/api/models/v1_import_leaderboard_response.py +2 -2
  122. eval_studio_client/api/models/v1_import_test_cases_from_library_response.py +2 -2
  123. eval_studio_client/api/models/v1_import_test_cases_request.py +2 -2
  124. eval_studio_client/api/models/v1_info.py +2 -2
  125. eval_studio_client/api/models/v1_init_workflow_node_response.py +2 -2
  126. eval_studio_client/api/models/v1_insight.py +2 -2
  127. eval_studio_client/api/models/v1_labeled_test_case.py +2 -2
  128. eval_studio_client/api/models/v1_leaderboard.py +8 -8
  129. eval_studio_client/api/models/v1_leaderboard_report.py +8 -8
  130. eval_studio_client/api/models/v1_leaderboard_report_actual_output_data.py +2 -2
  131. eval_studio_client/api/models/v1_leaderboard_report_actual_output_meta.py +5 -5
  132. eval_studio_client/api/models/v1_leaderboard_report_evaluator.py +11 -11
  133. eval_studio_client/api/models/v1_leaderboard_report_evaluator_parameter.py +2 -2
  134. eval_studio_client/api/models/v1_leaderboard_report_explanation.py +2 -2
  135. eval_studio_client/api/models/v1_leaderboard_report_metrics_meta_entry.py +2 -2
  136. eval_studio_client/api/models/v1_leaderboard_report_model.py +2 -2
  137. eval_studio_client/api/models/v1_leaderboard_report_result.py +11 -11
  138. eval_studio_client/api/models/v1_leaderboard_report_result_relationship.py +2 -2
  139. eval_studio_client/api/models/v1_list_base_models_response.py +2 -2
  140. eval_studio_client/api/models/v1_list_dashboard_access_response.py +5 -5
  141. eval_studio_client/api/models/v1_list_dashboards_response.py +5 -5
  142. eval_studio_client/api/models/v1_list_dashboards_shared_with_me_response.py +5 -5
  143. eval_studio_client/api/models/v1_list_documents_response.py +5 -5
  144. eval_studio_client/api/models/v1_list_evaluators_response.py +5 -5
  145. eval_studio_client/api/models/v1_list_leaderboards_response.py +5 -5
  146. eval_studio_client/api/models/v1_list_llm_models_response.py +2 -2
  147. eval_studio_client/api/models/v1_list_model_collections_response.py +5 -5
  148. eval_studio_client/api/models/v1_list_models_response.py +5 -5
  149. eval_studio_client/api/models/v1_list_most_recent_dashboards_response.py +5 -5
  150. eval_studio_client/api/models/v1_list_most_recent_leaderboards_response.py +5 -5
  151. eval_studio_client/api/models/v1_list_most_recent_models_response.py +5 -5
  152. eval_studio_client/api/models/v1_list_most_recent_tests_response.py +5 -5
  153. eval_studio_client/api/models/v1_list_operations_response.py +5 -5
  154. eval_studio_client/api/models/v1_list_perturbators_response.py +5 -5
  155. eval_studio_client/api/models/v1_list_prompt_library_items_response.py +5 -5
  156. eval_studio_client/api/models/v1_list_rag_collections_response.py +5 -5
  157. eval_studio_client/api/models/v1_list_test_access_response.py +5 -5
  158. eval_studio_client/api/models/v1_list_test_case_library_items_response.py +5 -5
  159. eval_studio_client/api/models/v1_list_test_case_relationships_response.py +5 -5
  160. eval_studio_client/api/models/v1_list_test_cases_response.py +5 -5
  161. eval_studio_client/api/models/v1_list_test_classes_response.py +5 -5
  162. eval_studio_client/api/models/v1_list_tests_response.py +5 -5
  163. eval_studio_client/api/models/v1_list_tests_shared_with_me_response.py +5 -5
  164. eval_studio_client/api/models/v1_list_workflow_access_response.py +5 -5
  165. eval_studio_client/api/models/v1_list_workflow_dependencies_response.py +8 -8
  166. eval_studio_client/api/models/v1_list_workflows_response.py +5 -5
  167. eval_studio_client/api/models/v1_list_workflows_shared_with_me_response.py +5 -5
  168. eval_studio_client/api/models/v1_metric_score.py +2 -2
  169. eval_studio_client/api/models/v1_metric_scores.py +5 -5
  170. eval_studio_client/api/models/v1_model.py +2 -2
  171. eval_studio_client/api/models/v1_operation.py +2 -2
  172. eval_studio_client/api/models/v1_operation_progress.py +2 -2
  173. eval_studio_client/api/models/v1_perturb_test_in_place_response.py +2 -2
  174. eval_studio_client/api/models/v1_perturb_test_response.py +2 -2
  175. eval_studio_client/api/models/v1_perturbator.py +2 -2
  176. eval_studio_client/api/models/v1_perturbator_configuration.py +2 -2
  177. eval_studio_client/api/models/v1_problem_and_action.py +2 -2
  178. eval_studio_client/api/models/v1_process_workflow_node_response.py +2 -2
  179. eval_studio_client/api/models/v1_prompt_library_item.py +2 -2
  180. eval_studio_client/api/models/v1_repeated_context.py +5 -5
  181. eval_studio_client/api/models/v1_repeated_string.py +2 -2
  182. eval_studio_client/api/models/v1_reset_workflow_node_response.py +2 -2
  183. eval_studio_client/api/models/v1_role_binding.py +2 -2
  184. eval_studio_client/api/models/v1_stats.py +2 -2
  185. eval_studio_client/api/models/v1_test.py +2 -2
  186. eval_studio_client/api/models/v1_test_case.py +2 -2
  187. eval_studio_client/api/models/v1_test_case_relationship.py +2 -2
  188. eval_studio_client/api/models/v1_test_class.py +2 -2
  189. eval_studio_client/api/models/v1_test_lab.py +2 -2
  190. eval_studio_client/api/models/v1_update_dashboard_response.py +2 -2
  191. eval_studio_client/api/models/v1_update_document_response.py +2 -2
  192. eval_studio_client/api/models/v1_update_leaderboard_response.py +2 -2
  193. eval_studio_client/api/models/v1_update_model_response.py +2 -2
  194. eval_studio_client/api/models/v1_update_operation_response.py +2 -2
  195. eval_studio_client/api/models/v1_update_test_case_response.py +2 -2
  196. eval_studio_client/api/models/v1_update_test_response.py +2 -2
  197. eval_studio_client/api/models/v1_update_workflow_node_response.py +2 -2
  198. eval_studio_client/api/models/v1_update_workflow_response.py +2 -2
  199. eval_studio_client/api/models/v1_who_am_i_response.py +2 -2
  200. eval_studio_client/api/models/v1_workflow.py +5 -5
  201. eval_studio_client/api/models/v1_workflow_dependency.py +2 -2
  202. eval_studio_client/api/models/v1_workflow_edge.py +2 -2
  203. eval_studio_client/api/models/v1_workflow_node.py +5 -5
  204. eval_studio_client/api/models/v1_workflow_node_artifact.py +2 -2
  205. eval_studio_client/api/models/v1_workflow_node_artifacts.py +5 -5
  206. eval_studio_client/api/models/v1_workflow_node_attributes.py +2 -2
  207. eval_studio_client/api/models/workflow_service_clone_workflow_request.py +2 -2
  208. eval_studio_client/api/models/workflow_service_revoke_workflow_access_request.py +2 -2
  209. eval_studio_client/leaderboards.py +17 -0
  210. eval_studio_client/models.py +153 -8
  211. {eval_studio_client-1.2.2a1.dist-info → eval_studio_client-1.2.2a2.dist-info}/METADATA +1 -1
  212. {eval_studio_client-1.2.2a1.dist-info → eval_studio_client-1.2.2a2.dist-info}/RECORD +213 -213
  213. {eval_studio_client-1.2.2a1.dist-info → eval_studio_client-1.2.2a2.dist-info}/WHEEL +0 -0
@@ -80,16 +80,16 @@ class V1LeaderboardReport(BaseModel):
80
80
  # override the default output from pydantic by calling `to_dict()` of each item in results (list)
81
81
  _items = []
82
82
  if self.results:
83
- for _item in self.results:
84
- if _item:
85
- _items.append(_item.to_dict())
83
+ for _item_results in self.results:
84
+ if _item_results:
85
+ _items.append(_item_results.to_dict())
86
86
  _dict['results'] = _items
87
87
  # override the default output from pydantic by calling `to_dict()` of each item in models (list)
88
88
  _items = []
89
89
  if self.models:
90
- for _item in self.models:
91
- if _item:
92
- _items.append(_item.to_dict())
90
+ for _item_models in self.models:
91
+ if _item_models:
92
+ _items.append(_item_models.to_dict())
93
93
  _dict['models'] = _items
94
94
  # override the default output from pydantic by calling `to_dict()` of evaluator
95
95
  if self.evaluator:
@@ -103,13 +103,13 @@ class V1LeaderboardReport(BaseModel):
103
103
  return None
104
104
 
105
105
  if not isinstance(obj, dict):
106
- return cls.model_validate(obj)
106
+ return cls.model_validate(obj, strict=False)
107
107
 
108
108
  _obj = cls.model_validate({
109
109
  "results": [V1LeaderboardReportResult.from_dict(_item) for _item in obj["results"]] if obj.get("results") is not None else None,
110
110
  "models": [V1LeaderboardReportModel.from_dict(_item) for _item in obj["models"]] if obj.get("models") is not None else None,
111
111
  "evaluator": V1LeaderboardReportEvaluator.from_dict(obj["evaluator"]) if obj.get("evaluator") is not None else None
112
- })
112
+ }, strict=False)
113
113
  return _obj
114
114
 
115
115
 
@@ -82,12 +82,12 @@ class V1LeaderboardReportActualOutputData(BaseModel):
82
82
  return None
83
83
 
84
84
  if not isinstance(obj, dict):
85
- return cls.model_validate(obj)
85
+ return cls.model_validate(obj, strict=False)
86
86
 
87
87
  _obj = cls.model_validate({
88
88
  "text": obj.get("text"),
89
89
  "metrics": obj.get("metrics")
90
- })
90
+ }, strict=False)
91
91
  return _obj
92
92
 
93
93
 
@@ -77,9 +77,9 @@ class V1LeaderboardReportActualOutputMeta(BaseModel):
77
77
  # override the default output from pydantic by calling `to_dict()` of each item in data (list)
78
78
  _items = []
79
79
  if self.data:
80
- for _item in self.data:
81
- if _item:
82
- _items.append(_item.to_dict())
80
+ for _item_data in self.data:
81
+ if _item_data:
82
+ _items.append(_item_data.to_dict())
83
83
  _dict['data'] = _items
84
84
  return _dict
85
85
 
@@ -90,12 +90,12 @@ class V1LeaderboardReportActualOutputMeta(BaseModel):
90
90
  return None
91
91
 
92
92
  if not isinstance(obj, dict):
93
- return cls.model_validate(obj)
93
+ return cls.model_validate(obj, strict=False)
94
94
 
95
95
  _obj = cls.model_validate({
96
96
  "tokenization": obj.get("tokenization"),
97
97
  "data": [V1LeaderboardReportActualOutputData.from_dict(_item) for _item in obj["data"]] if obj.get("data") is not None else None
98
- })
98
+ }, strict=False)
99
99
  return _obj
100
100
 
101
101
 
@@ -106,23 +106,23 @@ class V1LeaderboardReportEvaluator(BaseModel):
106
106
  # override the default output from pydantic by calling `to_dict()` of each item in explanations (list)
107
107
  _items = []
108
108
  if self.explanations:
109
- for _item in self.explanations:
110
- if _item:
111
- _items.append(_item.to_dict())
109
+ for _item_explanations in self.explanations:
110
+ if _item_explanations:
111
+ _items.append(_item_explanations.to_dict())
112
112
  _dict['explanations'] = _items
113
113
  # override the default output from pydantic by calling `to_dict()` of each item in parameters (list)
114
114
  _items = []
115
115
  if self.parameters:
116
- for _item in self.parameters:
117
- if _item:
118
- _items.append(_item.to_dict())
116
+ for _item_parameters in self.parameters:
117
+ if _item_parameters:
118
+ _items.append(_item_parameters.to_dict())
119
119
  _dict['parameters'] = _items
120
120
  # override the default output from pydantic by calling `to_dict()` of each item in metrics_meta (list)
121
121
  _items = []
122
122
  if self.metrics_meta:
123
- for _item in self.metrics_meta:
124
- if _item:
125
- _items.append(_item.to_dict())
123
+ for _item_metrics_meta in self.metrics_meta:
124
+ if _item_metrics_meta:
125
+ _items.append(_item_metrics_meta.to_dict())
126
126
  _dict['metricsMeta'] = _items
127
127
  return _dict
128
128
 
@@ -133,7 +133,7 @@ class V1LeaderboardReportEvaluator(BaseModel):
133
133
  return None
134
134
 
135
135
  if not isinstance(obj, dict):
136
- return cls.model_validate(obj)
136
+ return cls.model_validate(obj, strict=False)
137
137
 
138
138
  _obj = cls.model_validate({
139
139
  "id": obj.get("id"),
@@ -149,7 +149,7 @@ class V1LeaderboardReportEvaluator(BaseModel):
149
149
  "parameters": [V1LeaderboardReportEvaluatorParameter.from_dict(_item) for _item in obj["parameters"]] if obj.get("parameters") is not None else None,
150
150
  "keywords": obj.get("keywords"),
151
151
  "metricsMeta": [V1LeaderboardReportMetricsMetaEntry.from_dict(_item) for _item in obj["metricsMeta"]] if obj.get("metricsMeta") is not None else None
152
- })
152
+ }, strict=False)
153
153
  return _obj
154
154
 
155
155
 
@@ -91,7 +91,7 @@ class V1LeaderboardReportEvaluatorParameter(BaseModel):
91
91
  return None
92
92
 
93
93
  if not isinstance(obj, dict):
94
- return cls.model_validate(obj)
94
+ return cls.model_validate(obj, strict=False)
95
95
 
96
96
  _obj = cls.model_validate({
97
97
  "name": obj.get("name"),
@@ -103,7 +103,7 @@ class V1LeaderboardReportEvaluatorParameter(BaseModel):
103
103
  "min": obj.get("min"),
104
104
  "max": obj.get("max"),
105
105
  "category": obj.get("category")
106
- })
106
+ }, strict=False)
107
107
  return _obj
108
108
 
109
109
 
@@ -88,7 +88,7 @@ class V1LeaderboardReportExplanation(BaseModel):
88
88
  return None
89
89
 
90
90
  if not isinstance(obj, dict):
91
- return cls.model_validate(obj)
91
+ return cls.model_validate(obj, strict=False)
92
92
 
93
93
  _obj = cls.model_validate({
94
94
  "explanationType": obj.get("explanationType"),
@@ -97,7 +97,7 @@ class V1LeaderboardReportExplanation(BaseModel):
97
97
  "scope": obj.get("scope"),
98
98
  "hasLocal": obj.get("hasLocal"),
99
99
  "formats": obj.get("formats")
100
- })
100
+ }, strict=False)
101
101
  return _obj
102
102
 
103
103
 
@@ -108,7 +108,7 @@ class V1LeaderboardReportMetricsMetaEntry(BaseModel):
108
108
  return None
109
109
 
110
110
  if not isinstance(obj, dict):
111
- return cls.model_validate(obj)
111
+ return cls.model_validate(obj, strict=False)
112
112
 
113
113
  _obj = cls.model_validate({
114
114
  "key": obj.get("key"),
@@ -123,7 +123,7 @@ class V1LeaderboardReportMetricsMetaEntry(BaseModel):
123
123
  "isPrimaryMetric": obj.get("isPrimaryMetric"),
124
124
  "parentMetric": obj.get("parentMetric"),
125
125
  "exclude": obj.get("exclude")
126
- })
126
+ }, strict=False)
127
127
  return _obj
128
128
 
129
129
 
@@ -96,7 +96,7 @@ class V1LeaderboardReportModel(BaseModel):
96
96
  return None
97
97
 
98
98
  if not isinstance(obj, dict):
99
- return cls.model_validate(obj)
99
+ return cls.model_validate(obj, strict=False)
100
100
 
101
101
  _obj = cls.model_validate({
102
102
  "connection": obj.get("connection"),
@@ -107,7 +107,7 @@ class V1LeaderboardReportModel(BaseModel):
107
107
  "llmModelName": obj.get("llmModelName"),
108
108
  "documents": obj.get("documents"),
109
109
  "key": obj.get("key")
110
- })
110
+ }, strict=False)
111
111
  return _obj
112
112
 
113
113
 
@@ -122,23 +122,23 @@ class V1LeaderboardReportResult(BaseModel):
122
122
  # override the default output from pydantic by calling `to_dict()` of each item in relationships (list)
123
123
  _items = []
124
124
  if self.relationships:
125
- for _item in self.relationships:
126
- if _item:
127
- _items.append(_item.to_dict())
125
+ for _item_relationships in self.relationships:
126
+ if _item_relationships:
127
+ _items.append(_item_relationships.to_dict())
128
128
  _dict['relationships'] = _items
129
129
  # override the default output from pydantic by calling `to_dict()` of each item in metrics (list)
130
130
  _items = []
131
131
  if self.metrics:
132
- for _item in self.metrics:
133
- if _item:
134
- _items.append(_item.to_dict())
132
+ for _item_metrics in self.metrics:
133
+ if _item_metrics:
134
+ _items.append(_item_metrics.to_dict())
135
135
  _dict['metrics'] = _items
136
136
  # override the default output from pydantic by calling `to_dict()` of each item in actual_output_meta (list)
137
137
  _items = []
138
138
  if self.actual_output_meta:
139
- for _item in self.actual_output_meta:
140
- if _item:
141
- _items.append(_item.to_dict())
139
+ for _item_actual_output_meta in self.actual_output_meta:
140
+ if _item_actual_output_meta:
141
+ _items.append(_item_actual_output_meta.to_dict())
142
142
  _dict['actualOutputMeta'] = _items
143
143
  return _dict
144
144
 
@@ -149,7 +149,7 @@ class V1LeaderboardReportResult(BaseModel):
149
149
  return None
150
150
 
151
151
  if not isinstance(obj, dict):
152
- return cls.model_validate(obj)
152
+ return cls.model_validate(obj, strict=False)
153
153
 
154
154
  _obj = cls.model_validate({
155
155
  "key": obj.get("key"),
@@ -169,7 +169,7 @@ class V1LeaderboardReportResult(BaseModel):
169
169
  "metrics": [V1MetricScore.from_dict(_item) for _item in obj["metrics"]] if obj.get("metrics") is not None else None,
170
170
  "resultErrorMessage": obj.get("resultErrorMessage"),
171
171
  "actualOutputMeta": [V1LeaderboardReportActualOutputMeta.from_dict(_item) for _item in obj["actualOutputMeta"]] if obj.get("actualOutputMeta") is not None else None
172
- })
172
+ }, strict=False)
173
173
  return _obj
174
174
 
175
175
 
@@ -85,13 +85,13 @@ class V1LeaderboardReportResultRelationship(BaseModel):
85
85
  return None
86
86
 
87
87
  if not isinstance(obj, dict):
88
- return cls.model_validate(obj)
88
+ return cls.model_validate(obj, strict=False)
89
89
 
90
90
  _obj = cls.model_validate({
91
91
  "type": obj.get("type"),
92
92
  "target": obj.get("target"),
93
93
  "targetType": obj.get("targetType")
94
- })
94
+ }, strict=False)
95
95
  return _obj
96
96
 
97
97
 
@@ -77,11 +77,11 @@ class V1ListBaseModelsResponse(BaseModel):
77
77
  return None
78
78
 
79
79
  if not isinstance(obj, dict):
80
- return cls.model_validate(obj)
80
+ return cls.model_validate(obj, strict=False)
81
81
 
82
82
  _obj = cls.model_validate({
83
83
  "baseModels": obj.get("baseModels")
84
- })
84
+ }, strict=False)
85
85
  return _obj
86
86
 
87
87
 
@@ -72,9 +72,9 @@ class V1ListDashboardAccessResponse(BaseModel):
72
72
  # override the default output from pydantic by calling `to_dict()` of each item in role_bindings (list)
73
73
  _items = []
74
74
  if self.role_bindings:
75
- for _item in self.role_bindings:
76
- if _item:
77
- _items.append(_item.to_dict())
75
+ for _item_role_bindings in self.role_bindings:
76
+ if _item_role_bindings:
77
+ _items.append(_item_role_bindings.to_dict())
78
78
  _dict['roleBindings'] = _items
79
79
  return _dict
80
80
 
@@ -85,11 +85,11 @@ class V1ListDashboardAccessResponse(BaseModel):
85
85
  return None
86
86
 
87
87
  if not isinstance(obj, dict):
88
- return cls.model_validate(obj)
88
+ return cls.model_validate(obj, strict=False)
89
89
 
90
90
  _obj = cls.model_validate({
91
91
  "roleBindings": [V1RoleBinding.from_dict(_item) for _item in obj["roleBindings"]] if obj.get("roleBindings") is not None else None
92
- })
92
+ }, strict=False)
93
93
  return _obj
94
94
 
95
95
 
@@ -72,9 +72,9 @@ class V1ListDashboardsResponse(BaseModel):
72
72
  # override the default output from pydantic by calling `to_dict()` of each item in dashboards (list)
73
73
  _items = []
74
74
  if self.dashboards:
75
- for _item in self.dashboards:
76
- if _item:
77
- _items.append(_item.to_dict())
75
+ for _item_dashboards in self.dashboards:
76
+ if _item_dashboards:
77
+ _items.append(_item_dashboards.to_dict())
78
78
  _dict['dashboards'] = _items
79
79
  return _dict
80
80
 
@@ -85,11 +85,11 @@ class V1ListDashboardsResponse(BaseModel):
85
85
  return None
86
86
 
87
87
  if not isinstance(obj, dict):
88
- return cls.model_validate(obj)
88
+ return cls.model_validate(obj, strict=False)
89
89
 
90
90
  _obj = cls.model_validate({
91
91
  "dashboards": [V1Dashboard.from_dict(_item) for _item in obj["dashboards"]] if obj.get("dashboards") is not None else None
92
- })
92
+ }, strict=False)
93
93
  return _obj
94
94
 
95
95
 
@@ -72,9 +72,9 @@ class V1ListDashboardsSharedWithMeResponse(BaseModel):
72
72
  # override the default output from pydantic by calling `to_dict()` of each item in dashboards (list)
73
73
  _items = []
74
74
  if self.dashboards:
75
- for _item in self.dashboards:
76
- if _item:
77
- _items.append(_item.to_dict())
75
+ for _item_dashboards in self.dashboards:
76
+ if _item_dashboards:
77
+ _items.append(_item_dashboards.to_dict())
78
78
  _dict['dashboards'] = _items
79
79
  return _dict
80
80
 
@@ -85,11 +85,11 @@ class V1ListDashboardsSharedWithMeResponse(BaseModel):
85
85
  return None
86
86
 
87
87
  if not isinstance(obj, dict):
88
- return cls.model_validate(obj)
88
+ return cls.model_validate(obj, strict=False)
89
89
 
90
90
  _obj = cls.model_validate({
91
91
  "dashboards": [V1Dashboard.from_dict(_item) for _item in obj["dashboards"]] if obj.get("dashboards") is not None else None
92
- })
92
+ }, strict=False)
93
93
  return _obj
94
94
 
95
95
 
@@ -72,9 +72,9 @@ class V1ListDocumentsResponse(BaseModel):
72
72
  # override the default output from pydantic by calling `to_dict()` of each item in documents (list)
73
73
  _items = []
74
74
  if self.documents:
75
- for _item in self.documents:
76
- if _item:
77
- _items.append(_item.to_dict())
75
+ for _item_documents in self.documents:
76
+ if _item_documents:
77
+ _items.append(_item_documents.to_dict())
78
78
  _dict['documents'] = _items
79
79
  return _dict
80
80
 
@@ -85,11 +85,11 @@ class V1ListDocumentsResponse(BaseModel):
85
85
  return None
86
86
 
87
87
  if not isinstance(obj, dict):
88
- return cls.model_validate(obj)
88
+ return cls.model_validate(obj, strict=False)
89
89
 
90
90
  _obj = cls.model_validate({
91
91
  "documents": [V1Document.from_dict(_item) for _item in obj["documents"]] if obj.get("documents") is not None else None
92
- })
92
+ }, strict=False)
93
93
  return _obj
94
94
 
95
95
 
@@ -72,9 +72,9 @@ class V1ListEvaluatorsResponse(BaseModel):
72
72
  # override the default output from pydantic by calling `to_dict()` of each item in evaluators (list)
73
73
  _items = []
74
74
  if self.evaluators:
75
- for _item in self.evaluators:
76
- if _item:
77
- _items.append(_item.to_dict())
75
+ for _item_evaluators in self.evaluators:
76
+ if _item_evaluators:
77
+ _items.append(_item_evaluators.to_dict())
78
78
  _dict['evaluators'] = _items
79
79
  return _dict
80
80
 
@@ -85,11 +85,11 @@ class V1ListEvaluatorsResponse(BaseModel):
85
85
  return None
86
86
 
87
87
  if not isinstance(obj, dict):
88
- return cls.model_validate(obj)
88
+ return cls.model_validate(obj, strict=False)
89
89
 
90
90
  _obj = cls.model_validate({
91
91
  "evaluators": [V1Evaluator.from_dict(_item) for _item in obj["evaluators"]] if obj.get("evaluators") is not None else None
92
- })
92
+ }, strict=False)
93
93
  return _obj
94
94
 
95
95
 
@@ -73,9 +73,9 @@ class V1ListLeaderboardsResponse(BaseModel):
73
73
  # override the default output from pydantic by calling `to_dict()` of each item in leaderboards (list)
74
74
  _items = []
75
75
  if self.leaderboards:
76
- for _item in self.leaderboards:
77
- if _item:
78
- _items.append(_item.to_dict())
76
+ for _item_leaderboards in self.leaderboards:
77
+ if _item_leaderboards:
78
+ _items.append(_item_leaderboards.to_dict())
79
79
  _dict['leaderboards'] = _items
80
80
  return _dict
81
81
 
@@ -86,12 +86,12 @@ class V1ListLeaderboardsResponse(BaseModel):
86
86
  return None
87
87
 
88
88
  if not isinstance(obj, dict):
89
- return cls.model_validate(obj)
89
+ return cls.model_validate(obj, strict=False)
90
90
 
91
91
  _obj = cls.model_validate({
92
92
  "leaderboards": [V1Leaderboard.from_dict(_item) for _item in obj["leaderboards"]] if obj.get("leaderboards") is not None else None,
93
93
  "nextPageToken": obj.get("nextPageToken")
94
- })
94
+ }, strict=False)
95
95
  return _obj
96
96
 
97
97
 
@@ -77,11 +77,11 @@ class V1ListLLMModelsResponse(BaseModel):
77
77
  return None
78
78
 
79
79
  if not isinstance(obj, dict):
80
- return cls.model_validate(obj)
80
+ return cls.model_validate(obj, strict=False)
81
81
 
82
82
  _obj = cls.model_validate({
83
83
  "models": obj.get("models")
84
- })
84
+ }, strict=False)
85
85
  return _obj
86
86
 
87
87
 
@@ -72,9 +72,9 @@ class V1ListModelCollectionsResponse(BaseModel):
72
72
  # override the default output from pydantic by calling `to_dict()` of each item in collections (list)
73
73
  _items = []
74
74
  if self.collections:
75
- for _item in self.collections:
76
- if _item:
77
- _items.append(_item.to_dict())
75
+ for _item_collections in self.collections:
76
+ if _item_collections:
77
+ _items.append(_item_collections.to_dict())
78
78
  _dict['collections'] = _items
79
79
  return _dict
80
80
 
@@ -85,11 +85,11 @@ class V1ListModelCollectionsResponse(BaseModel):
85
85
  return None
86
86
 
87
87
  if not isinstance(obj, dict):
88
- return cls.model_validate(obj)
88
+ return cls.model_validate(obj, strict=False)
89
89
 
90
90
  _obj = cls.model_validate({
91
91
  "collections": [V1CollectionInfo.from_dict(_item) for _item in obj["collections"]] if obj.get("collections") is not None else None
92
- })
92
+ }, strict=False)
93
93
  return _obj
94
94
 
95
95
 
@@ -72,9 +72,9 @@ class V1ListModelsResponse(BaseModel):
72
72
  # override the default output from pydantic by calling `to_dict()` of each item in models (list)
73
73
  _items = []
74
74
  if self.models:
75
- for _item in self.models:
76
- if _item:
77
- _items.append(_item.to_dict())
75
+ for _item_models in self.models:
76
+ if _item_models:
77
+ _items.append(_item_models.to_dict())
78
78
  _dict['models'] = _items
79
79
  return _dict
80
80
 
@@ -85,11 +85,11 @@ class V1ListModelsResponse(BaseModel):
85
85
  return None
86
86
 
87
87
  if not isinstance(obj, dict):
88
- return cls.model_validate(obj)
88
+ return cls.model_validate(obj, strict=False)
89
89
 
90
90
  _obj = cls.model_validate({
91
91
  "models": [V1Model.from_dict(_item) for _item in obj["models"]] if obj.get("models") is not None else None
92
- })
92
+ }, strict=False)
93
93
  return _obj
94
94
 
95
95
 
@@ -72,9 +72,9 @@ class V1ListMostRecentDashboardsResponse(BaseModel):
72
72
  # override the default output from pydantic by calling `to_dict()` of each item in dashboards (list)
73
73
  _items = []
74
74
  if self.dashboards:
75
- for _item in self.dashboards:
76
- if _item:
77
- _items.append(_item.to_dict())
75
+ for _item_dashboards in self.dashboards:
76
+ if _item_dashboards:
77
+ _items.append(_item_dashboards.to_dict())
78
78
  _dict['dashboards'] = _items
79
79
  return _dict
80
80
 
@@ -85,11 +85,11 @@ class V1ListMostRecentDashboardsResponse(BaseModel):
85
85
  return None
86
86
 
87
87
  if not isinstance(obj, dict):
88
- return cls.model_validate(obj)
88
+ return cls.model_validate(obj, strict=False)
89
89
 
90
90
  _obj = cls.model_validate({
91
91
  "dashboards": [V1Dashboard.from_dict(_item) for _item in obj["dashboards"]] if obj.get("dashboards") is not None else None
92
- })
92
+ }, strict=False)
93
93
  return _obj
94
94
 
95
95
 
@@ -72,9 +72,9 @@ class V1ListMostRecentLeaderboardsResponse(BaseModel):
72
72
  # override the default output from pydantic by calling `to_dict()` of each item in leaderboards (list)
73
73
  _items = []
74
74
  if self.leaderboards:
75
- for _item in self.leaderboards:
76
- if _item:
77
- _items.append(_item.to_dict())
75
+ for _item_leaderboards in self.leaderboards:
76
+ if _item_leaderboards:
77
+ _items.append(_item_leaderboards.to_dict())
78
78
  _dict['leaderboards'] = _items
79
79
  return _dict
80
80
 
@@ -85,11 +85,11 @@ class V1ListMostRecentLeaderboardsResponse(BaseModel):
85
85
  return None
86
86
 
87
87
  if not isinstance(obj, dict):
88
- return cls.model_validate(obj)
88
+ return cls.model_validate(obj, strict=False)
89
89
 
90
90
  _obj = cls.model_validate({
91
91
  "leaderboards": [V1Leaderboard.from_dict(_item) for _item in obj["leaderboards"]] if obj.get("leaderboards") is not None else None
92
- })
92
+ }, strict=False)
93
93
  return _obj
94
94
 
95
95
 
@@ -72,9 +72,9 @@ class V1ListMostRecentModelsResponse(BaseModel):
72
72
  # override the default output from pydantic by calling `to_dict()` of each item in models (list)
73
73
  _items = []
74
74
  if self.models:
75
- for _item in self.models:
76
- if _item:
77
- _items.append(_item.to_dict())
75
+ for _item_models in self.models:
76
+ if _item_models:
77
+ _items.append(_item_models.to_dict())
78
78
  _dict['models'] = _items
79
79
  return _dict
80
80
 
@@ -85,11 +85,11 @@ class V1ListMostRecentModelsResponse(BaseModel):
85
85
  return None
86
86
 
87
87
  if not isinstance(obj, dict):
88
- return cls.model_validate(obj)
88
+ return cls.model_validate(obj, strict=False)
89
89
 
90
90
  _obj = cls.model_validate({
91
91
  "models": [V1Model.from_dict(_item) for _item in obj["models"]] if obj.get("models") is not None else None
92
- })
92
+ }, strict=False)
93
93
  return _obj
94
94
 
95
95
 
@@ -72,9 +72,9 @@ class V1ListMostRecentTestsResponse(BaseModel):
72
72
  # override the default output from pydantic by calling `to_dict()` of each item in tests (list)
73
73
  _items = []
74
74
  if self.tests:
75
- for _item in self.tests:
76
- if _item:
77
- _items.append(_item.to_dict())
75
+ for _item_tests in self.tests:
76
+ if _item_tests:
77
+ _items.append(_item_tests.to_dict())
78
78
  _dict['tests'] = _items
79
79
  return _dict
80
80
 
@@ -85,11 +85,11 @@ class V1ListMostRecentTestsResponse(BaseModel):
85
85
  return None
86
86
 
87
87
  if not isinstance(obj, dict):
88
- return cls.model_validate(obj)
88
+ return cls.model_validate(obj, strict=False)
89
89
 
90
90
  _obj = cls.model_validate({
91
91
  "tests": [V1Test.from_dict(_item) for _item in obj["tests"]] if obj.get("tests") is not None else None
92
- })
92
+ }, strict=False)
93
93
  return _obj
94
94
 
95
95