scale-gp-beta 0.1.0a32__py3-none-any.whl → 0.1.0a34__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (38) hide show
  1. scale_gp_beta/_client.py +20 -0
  2. scale_gp_beta/_version.py +1 -1
  3. scale_gp_beta/resources/__init__.py +28 -0
  4. scale_gp_beta/resources/credentials.py +853 -0
  5. scale_gp_beta/resources/evaluations.py +104 -9
  6. scale_gp_beta/resources/span_assessments.py +675 -0
  7. scale_gp_beta/resources/spans.py +24 -0
  8. scale_gp_beta/types/__init__.py +13 -4
  9. scale_gp_beta/types/approval_status.py +7 -0
  10. scale_gp_beta/types/assessment_type.py +7 -0
  11. scale_gp_beta/types/credential.py +30 -0
  12. scale_gp_beta/types/credential_create_params.py +25 -0
  13. scale_gp_beta/types/credential_delete_response.py +13 -0
  14. scale_gp_beta/types/credential_list_params.py +20 -0
  15. scale_gp_beta/types/credential_secret.py +13 -0
  16. scale_gp_beta/types/credential_update_params.py +25 -0
  17. scale_gp_beta/types/dataset_delete_response.py +1 -4
  18. scale_gp_beta/types/dataset_item_delete_response.py +1 -4
  19. scale_gp_beta/types/evaluation_create_params.py +1 -1
  20. scale_gp_beta/types/evaluation_task.py +182 -0
  21. scale_gp_beta/types/evaluation_task_param.py +192 -0
  22. scale_gp_beta/types/evaluation_update_params.py +12 -3
  23. scale_gp_beta/types/file_delete_response.py +1 -4
  24. scale_gp_beta/types/model_delete_response.py +1 -4
  25. scale_gp_beta/types/span_assessment.py +72 -0
  26. scale_gp_beta/types/span_assessment_create_params.py +40 -0
  27. scale_gp_beta/types/span_assessment_delete_response.py +13 -0
  28. scale_gp_beta/types/span_assessment_list_params.py +26 -0
  29. scale_gp_beta/types/span_assessment_update_params.py +34 -0
  30. scale_gp_beta/types/span_search_params.py +9 -0
  31. {scale_gp_beta-0.1.0a32.dist-info → scale_gp_beta-0.1.0a34.dist-info}/METADATA +1 -1
  32. {scale_gp_beta-0.1.0a32.dist-info → scale_gp_beta-0.1.0a34.dist-info}/RECORD +34 -23
  33. scale_gp_beta/types/evaluation_delete_response.py +0 -16
  34. scale_gp_beta/types/file_list.py +0 -27
  35. scale_gp_beta/types/inference_model_list.py +0 -27
  36. scale_gp_beta/types/question_list.py +0 -27
  37. {scale_gp_beta-0.1.0a32.dist-info → scale_gp_beta-0.1.0a34.dist-info}/WHEEL +0 -0
  38. {scale_gp_beta-0.1.0a32.dist-info → scale_gp_beta-0.1.0a34.dist-info}/licenses/LICENSE +0 -0
@@ -256,10 +256,13 @@ class SpansResource(SyncAPIResource):
256
256
  starting_after: str | Omit = omit,
257
257
  to_ts: Union[str, datetime] | Omit = omit,
258
258
  application_variant_ids: SequenceNotStr[str] | Omit = omit,
259
+ assessment_types: SequenceNotStr[str] | Omit = omit,
259
260
  excluded_span_ids: SequenceNotStr[str] | Omit = omit,
260
261
  excluded_trace_ids: SequenceNotStr[str] | Omit = omit,
261
262
  extra_metadata: Dict[str, object] | Omit = omit,
262
263
  group_id: str | Omit = omit,
264
+ max_duration_ms: int | Omit = omit,
265
+ min_duration_ms: int | Omit = omit,
263
266
  names: SequenceNotStr[str] | Omit = omit,
264
267
  parents_only: bool | Omit = omit,
265
268
  search_texts: SequenceNotStr[str] | Omit = omit,
@@ -284,6 +287,8 @@ class SpansResource(SyncAPIResource):
284
287
 
285
288
  application_variant_ids: Filter by application variant IDs
286
289
 
290
+ assessment_types: Filter spans by traces that have assessments of these types
291
+
287
292
  excluded_span_ids: List of span IDs to exclude from results
288
293
 
289
294
  excluded_trace_ids: List of trace IDs to exclude from results
@@ -292,6 +297,10 @@ class SpansResource(SyncAPIResource):
292
297
 
293
298
  group_id: Filter by group ID
294
299
 
300
+ max_duration_ms: Maximum span duration in milliseconds (inclusive)
301
+
302
+ min_duration_ms: Minimum span duration in milliseconds (inclusive)
303
+
295
304
  names: Filter by trace/span name
296
305
 
297
306
  parents_only: Only fetch spans that are the top-level (ie. have no parent_id)
@@ -318,10 +327,13 @@ class SpansResource(SyncAPIResource):
318
327
  body=maybe_transform(
319
328
  {
320
329
  "application_variant_ids": application_variant_ids,
330
+ "assessment_types": assessment_types,
321
331
  "excluded_span_ids": excluded_span_ids,
322
332
  "excluded_trace_ids": excluded_trace_ids,
323
333
  "extra_metadata": extra_metadata,
324
334
  "group_id": group_id,
335
+ "max_duration_ms": max_duration_ms,
336
+ "min_duration_ms": min_duration_ms,
325
337
  "names": names,
326
338
  "parents_only": parents_only,
327
339
  "search_texts": search_texts,
@@ -604,10 +616,13 @@ class AsyncSpansResource(AsyncAPIResource):
604
616
  starting_after: str | Omit = omit,
605
617
  to_ts: Union[str, datetime] | Omit = omit,
606
618
  application_variant_ids: SequenceNotStr[str] | Omit = omit,
619
+ assessment_types: SequenceNotStr[str] | Omit = omit,
607
620
  excluded_span_ids: SequenceNotStr[str] | Omit = omit,
608
621
  excluded_trace_ids: SequenceNotStr[str] | Omit = omit,
609
622
  extra_metadata: Dict[str, object] | Omit = omit,
610
623
  group_id: str | Omit = omit,
624
+ max_duration_ms: int | Omit = omit,
625
+ min_duration_ms: int | Omit = omit,
611
626
  names: SequenceNotStr[str] | Omit = omit,
612
627
  parents_only: bool | Omit = omit,
613
628
  search_texts: SequenceNotStr[str] | Omit = omit,
@@ -632,6 +647,8 @@ class AsyncSpansResource(AsyncAPIResource):
632
647
 
633
648
  application_variant_ids: Filter by application variant IDs
634
649
 
650
+ assessment_types: Filter spans by traces that have assessments of these types
651
+
635
652
  excluded_span_ids: List of span IDs to exclude from results
636
653
 
637
654
  excluded_trace_ids: List of trace IDs to exclude from results
@@ -640,6 +657,10 @@ class AsyncSpansResource(AsyncAPIResource):
640
657
 
641
658
  group_id: Filter by group ID
642
659
 
660
+ max_duration_ms: Maximum span duration in milliseconds (inclusive)
661
+
662
+ min_duration_ms: Minimum span duration in milliseconds (inclusive)
663
+
643
664
  names: Filter by trace/span name
644
665
 
645
666
  parents_only: Only fetch spans that are the top-level (ie. have no parent_id)
@@ -666,10 +687,13 @@ class AsyncSpansResource(AsyncAPIResource):
666
687
  body=maybe_transform(
667
688
  {
668
689
  "application_variant_ids": application_variant_ids,
690
+ "assessment_types": assessment_types,
669
691
  "excluded_span_ids": excluded_span_ids,
670
692
  "excluded_trace_ids": excluded_trace_ids,
671
693
  "extra_metadata": extra_metadata,
672
694
  "group_id": group_id,
695
+ "max_duration_ms": max_duration_ms,
696
+ "min_duration_ms": min_duration_ms,
673
697
  "names": names,
674
698
  "parents_only": parents_only,
675
699
  "search_texts": search_texts,
@@ -12,20 +12,23 @@ from .question import Question as Question
12
12
  from .response import Response as Response
13
13
  from .component import Component as Component
14
14
  from .container import Container as Container
15
- from .file_list import FileList as FileList
16
15
  from .span_type import SpanType as SpanType
17
16
  from .completion import Completion as Completion
17
+ from .credential import Credential as Credential
18
18
  from .evaluation import Evaluation as Evaluation
19
19
  from .span_status import SpanStatus as SpanStatus
20
20
  from .dataset_item import DatasetItem as DatasetItem
21
21
  from .item_locator import ItemLocator as ItemLocator
22
- from .question_list import QuestionList as QuestionList
22
+ from .approval_status import ApprovalStatus as ApprovalStatus
23
+ from .assessment_type import AssessmentType as AssessmentType
23
24
  from .component_param import ComponentParam as ComponentParam
24
25
  from .container_param import ContainerParam as ContainerParam
25
26
  from .evaluation_item import EvaluationItem as EvaluationItem
26
27
  from .evaluation_task import EvaluationTask as EvaluationTask
27
28
  from .inference_model import InferenceModel as InferenceModel
29
+ from .span_assessment import SpanAssessment as SpanAssessment
28
30
  from .file_list_params import FileListParams as FileListParams
31
+ from .credential_secret import CredentialSecret as CredentialSecret
29
32
  from .model_list_params import ModelListParams as ModelListParams
30
33
  from .span_batch_params import SpanBatchParams as SpanBatchParams
31
34
  from .file_create_params import FileCreateParams as FileCreateParams
@@ -39,13 +42,13 @@ from .model_create_params import ModelCreateParams as ModelCreateParams
39
42
  from .model_update_params import ModelUpdateParams as ModelUpdateParams
40
43
  from .span_batch_response import SpanBatchResponse as SpanBatchResponse
41
44
  from .file_delete_response import FileDeleteResponse as FileDeleteResponse
42
- from .inference_model_list import InferenceModelList as InferenceModelList
43
45
  from .question_list_params import QuestionListParams as QuestionListParams
44
46
  from .dataset_create_params import DatasetCreateParams as DatasetCreateParams
45
47
  from .dataset_update_params import DatasetUpdateParams as DatasetUpdateParams
46
48
  from .evaluation_task_param import EvaluationTaskParam as EvaluationTaskParam
47
49
  from .item_locator_template import ItemLocatorTemplate as ItemLocatorTemplate
48
50
  from .model_delete_response import ModelDeleteResponse as ModelDeleteResponse
51
+ from .credential_list_params import CredentialListParams as CredentialListParams
49
52
  from .evaluation_list_params import EvaluationListParams as EvaluationListParams
50
53
  from .question_create_params import QuestionCreateParams as QuestionCreateParams
51
54
  from .response_create_params import ResponseCreateParams as ResponseCreateParams
@@ -53,6 +56,8 @@ from .dataset_delete_response import DatasetDeleteResponse as DatasetDeleteRespo
53
56
  from .dataset_retrieve_params import DatasetRetrieveParams as DatasetRetrieveParams
54
57
  from .inference_create_params import InferenceCreateParams as InferenceCreateParams
55
58
  from .completion_create_params import CompletionCreateParams as CompletionCreateParams
59
+ from .credential_create_params import CredentialCreateParams as CredentialCreateParams
60
+ from .credential_update_params import CredentialUpdateParams as CredentialUpdateParams
56
61
  from .dataset_item_list_params import DatasetItemListParams as DatasetItemListParams
57
62
  from .evaluation_create_params import EvaluationCreateParams as EvaluationCreateParams
58
63
  from .evaluation_update_params import EvaluationUpdateParams as EvaluationUpdateParams
@@ -60,14 +65,18 @@ from .inference_response_chunk import InferenceResponseChunk as InferenceRespons
60
65
  from .response_create_response import ResponseCreateResponse as ResponseCreateResponse
61
66
  from .span_upsert_batch_params import SpanUpsertBatchParams as SpanUpsertBatchParams
62
67
  from .inference_create_response import InferenceCreateResponse as InferenceCreateResponse
68
+ from .credential_delete_response import CredentialDeleteResponse as CredentialDeleteResponse
63
69
  from .dataset_item_update_params import DatasetItemUpdateParams as DatasetItemUpdateParams
64
- from .evaluation_delete_response import EvaluationDeleteResponse as EvaluationDeleteResponse
65
70
  from .evaluation_retrieve_params import EvaluationRetrieveParams as EvaluationRetrieveParams
66
71
  from .span_upsert_batch_response import SpanUpsertBatchResponse as SpanUpsertBatchResponse
67
72
  from .evaluation_item_list_params import EvaluationItemListParams as EvaluationItemListParams
73
+ from .span_assessment_list_params import SpanAssessmentListParams as SpanAssessmentListParams
68
74
  from .dataset_item_delete_response import DatasetItemDeleteResponse as DatasetItemDeleteResponse
69
75
  from .dataset_item_retrieve_params import DatasetItemRetrieveParams as DatasetItemRetrieveParams
76
+ from .span_assessment_create_params import SpanAssessmentCreateParams as SpanAssessmentCreateParams
77
+ from .span_assessment_update_params import SpanAssessmentUpdateParams as SpanAssessmentUpdateParams
70
78
  from .evaluation_item_retrieve_params import EvaluationItemRetrieveParams as EvaluationItemRetrieveParams
79
+ from .span_assessment_delete_response import SpanAssessmentDeleteResponse as SpanAssessmentDeleteResponse
71
80
  from .dataset_item_batch_create_params import DatasetItemBatchCreateParams as DatasetItemBatchCreateParams
72
81
  from .dataset_item_batch_create_response import DatasetItemBatchCreateResponse as DatasetItemBatchCreateResponse
73
82
 
@@ -0,0 +1,7 @@
1
+ # File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.
2
+
3
+ from typing_extensions import Literal, TypeAlias
4
+
5
+ __all__ = ["ApprovalStatus"]
6
+
7
+ ApprovalStatus: TypeAlias = Literal["approved", "rejected"]
@@ -0,0 +1,7 @@
1
+ # File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.
2
+
3
+ from typing_extensions import Literal, TypeAlias
4
+
5
+ __all__ = ["AssessmentType"]
6
+
7
+ AssessmentType: TypeAlias = Literal["comment", "rating", "approval", "rubric", "metadata", "overwrite"]
@@ -0,0 +1,30 @@
1
+ # File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.
2
+
3
+ from typing import Dict, Optional
4
+ from datetime import datetime
5
+
6
+ from .._models import BaseModel
7
+
8
+ __all__ = ["Credential"]
9
+
10
+
11
+ class Credential(BaseModel):
12
+ id: str
13
+
14
+ created_at: datetime
15
+
16
+ created_by_identity_type: str
17
+
18
+ created_by_user_id: str
19
+
20
+ credential_metadata: Dict[str, object]
21
+
22
+ description: str
23
+
24
+ name: str
25
+
26
+ type: str
27
+
28
+ updated_at: datetime
29
+
30
+ object: Optional[str] = None
@@ -0,0 +1,25 @@
1
+ # File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.
2
+
3
+ from __future__ import annotations
4
+
5
+ from typing import Dict
6
+ from typing_extensions import Required, TypedDict
7
+
8
+ __all__ = ["CredentialCreateParams"]
9
+
10
+
11
+ class CredentialCreateParams(TypedDict, total=False):
12
+ name: Required[str]
13
+ """User-friendly name for the credential"""
14
+
15
+ payload: Required[str]
16
+ """The credential payload to be encrypted"""
17
+
18
+ type: Required[str]
19
+ """Type of credential: key or json"""
20
+
21
+ credential_metadata: Dict[str, object]
22
+ """Optional unencrypted credential_metadata"""
23
+
24
+ description: str
25
+ """Optional description"""
@@ -0,0 +1,13 @@
1
+ # File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.
2
+
3
+ from .._models import BaseModel
4
+
5
+ __all__ = ["CredentialDeleteResponse"]
6
+
7
+
8
+ class CredentialDeleteResponse(BaseModel):
9
+ id: str
10
+
11
+ deleted: bool
12
+
13
+ object: str
@@ -0,0 +1,20 @@
1
+ # File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.
2
+
3
+ from __future__ import annotations
4
+
5
+ from typing_extensions import Literal, TypedDict
6
+
7
+ __all__ = ["CredentialListParams"]
8
+
9
+
10
+ class CredentialListParams(TypedDict, total=False):
11
+ ending_before: str
12
+
13
+ limit: int
14
+
15
+ name: str
16
+ """Filter credentials by name"""
17
+
18
+ sort_order: Literal["asc", "desc"]
19
+
20
+ starting_after: str
@@ -0,0 +1,13 @@
1
+ # File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.
2
+
3
+ from typing import Optional
4
+
5
+ from .._models import BaseModel
6
+
7
+ __all__ = ["CredentialSecret"]
8
+
9
+
10
+ class CredentialSecret(BaseModel):
11
+ result: str
12
+
13
+ object: Optional[str] = None
@@ -0,0 +1,25 @@
1
+ # File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.
2
+
3
+ from __future__ import annotations
4
+
5
+ from typing import Dict
6
+ from typing_extensions import TypedDict
7
+
8
+ __all__ = ["CredentialUpdateParams"]
9
+
10
+
11
+ class CredentialUpdateParams(TypedDict, total=False):
12
+ credential_metadata: Dict[str, object]
13
+ """Optional unencrypted credential_metadata"""
14
+
15
+ description: str
16
+ """Optional description"""
17
+
18
+ name: str
19
+ """User-friendly name for the credential"""
20
+
21
+ payload: str
22
+ """The credential payload to be encrypted"""
23
+
24
+ type: str
25
+ """Type of credential: key or json"""
@@ -1,8 +1,5 @@
1
1
  # File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.
2
2
 
3
- from typing import Optional
4
- from typing_extensions import Literal
5
-
6
3
  from .._models import BaseModel
7
4
 
8
5
  __all__ = ["DatasetDeleteResponse"]
@@ -13,4 +10,4 @@ class DatasetDeleteResponse(BaseModel):
13
10
 
14
11
  deleted: bool
15
12
 
16
- object: Optional[Literal["dataset"]] = None
13
+ object: str
@@ -1,8 +1,5 @@
1
1
  # File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.
2
2
 
3
- from typing import Optional
4
- from typing_extensions import Literal
5
-
6
3
  from .._models import BaseModel
7
4
 
8
5
  __all__ = ["DatasetItemDeleteResponse"]
@@ -13,4 +10,4 @@ class DatasetItemDeleteResponse(BaseModel):
13
10
 
14
11
  deleted: bool
15
12
 
16
- object: Optional[Literal["dataset.item"]] = None
13
+ object: str
@@ -42,7 +42,7 @@ class EvaluationFromDatasetCreateRequest(TypedDict, total=False):
42
42
  name: Required[str]
43
43
 
44
44
  data: Iterable[EvaluationFromDatasetCreateRequestData]
45
- """Items to be evaluated, including references to the input dataset items"""
45
+ """Items to be evaluated, including references to the input dataset"""
46
46
 
47
47
  description: str
48
48
 
@@ -41,6 +41,27 @@ __all__ = [
41
41
  "AutoEvaluationGuidedDecodingEvaluationTaskConfiguration",
42
42
  "AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationStructuredOutputTaskRequestWithItemLocator",
43
43
  "AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationGuidedDecodingTaskRequestWithItemLocator",
44
+ "AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationAgentTaskRequestWithItemLocator",
45
+ "AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationAgentTaskRequestWithItemLocatorDesignatedTo",
46
+ "AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationAgentTaskRequestWithItemLocatorDesignatedToApeAgent",
47
+ "AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationAgentTaskRequestWithItemLocatorDesignatedToApeAgentConfig",
48
+ "AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationAgentTaskRequestWithItemLocatorDesignatedToIfAgent",
49
+ "AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationAgentTaskRequestWithItemLocatorDesignatedToIfAgentConfig",
50
+ "AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationAgentTaskRequestWithItemLocatorDesignatedToTruthfulnessAgent",
51
+ "AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationAgentTaskRequestWithItemLocatorDesignatedToTruthfulnessAgentConfig",
52
+ "AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationAgentTaskRequestWithItemLocatorDesignatedToBaseAgent",
53
+ "AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationAgentTaskRequestWithItemLocatorDesignatedToBaseAgentConfig",
54
+ "AutoEvaluationAgentEvaluationTask",
55
+ "AutoEvaluationAgentEvaluationTaskConfiguration",
56
+ "AutoEvaluationAgentEvaluationTaskConfigurationDesignatedTo",
57
+ "AutoEvaluationAgentEvaluationTaskConfigurationDesignatedToApeAgent",
58
+ "AutoEvaluationAgentEvaluationTaskConfigurationDesignatedToApeAgentConfig",
59
+ "AutoEvaluationAgentEvaluationTaskConfigurationDesignatedToIfAgent",
60
+ "AutoEvaluationAgentEvaluationTaskConfigurationDesignatedToIfAgentConfig",
61
+ "AutoEvaluationAgentEvaluationTaskConfigurationDesignatedToTruthfulnessAgent",
62
+ "AutoEvaluationAgentEvaluationTaskConfigurationDesignatedToTruthfulnessAgentConfig",
63
+ "AutoEvaluationAgentEvaluationTaskConfigurationDesignatedToBaseAgent",
64
+ "AutoEvaluationAgentEvaluationTaskConfigurationDesignatedToBaseAgentConfig",
44
65
  "ContributorEvaluationQuestionTask",
45
66
  "ContributorEvaluationQuestionTaskConfiguration",
46
67
  ]
@@ -368,9 +389,94 @@ class AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationGuide
368
389
  system_prompt: Optional[str] = None
369
390
 
370
391
 
392
+ class AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationAgentTaskRequestWithItemLocatorDesignatedToApeAgentConfig(
393
+ BaseModel
394
+ ):
395
+ model: Optional[str] = None
396
+
397
+ temperature: Optional[float] = None
398
+
399
+
400
+ class AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationAgentTaskRequestWithItemLocatorDesignatedToApeAgent(
401
+ BaseModel
402
+ ):
403
+ config: AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationAgentTaskRequestWithItemLocatorDesignatedToApeAgentConfig
404
+
405
+ agent_name: Optional[Literal["APEAgent"]] = None
406
+
407
+
408
+ class AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationAgentTaskRequestWithItemLocatorDesignatedToIfAgentConfig(
409
+ BaseModel
410
+ ):
411
+ model: Optional[str] = None
412
+
413
+
414
+ class AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationAgentTaskRequestWithItemLocatorDesignatedToIfAgent(
415
+ BaseModel
416
+ ):
417
+ config: AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationAgentTaskRequestWithItemLocatorDesignatedToIfAgentConfig
418
+
419
+ agent_name: Optional[Literal["IFAgent"]] = None
420
+
421
+
422
+ class AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationAgentTaskRequestWithItemLocatorDesignatedToTruthfulnessAgentConfig(
423
+ BaseModel
424
+ ):
425
+ model: Optional[str] = None
426
+
427
+
428
+ class AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationAgentTaskRequestWithItemLocatorDesignatedToTruthfulnessAgent(
429
+ BaseModel
430
+ ):
431
+ config: AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationAgentTaskRequestWithItemLocatorDesignatedToTruthfulnessAgentConfig
432
+
433
+ agent_name: Optional[Literal["TruthfulnessAgent"]] = None
434
+
435
+
436
+ class AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationAgentTaskRequestWithItemLocatorDesignatedToBaseAgentConfig(
437
+ BaseModel
438
+ ):
439
+ model: Optional[str] = None
440
+
441
+
442
+ class AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationAgentTaskRequestWithItemLocatorDesignatedToBaseAgent(
443
+ BaseModel
444
+ ):
445
+ config: AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationAgentTaskRequestWithItemLocatorDesignatedToBaseAgentConfig
446
+
447
+ agent_name: Optional[Literal["BaseAgent"]] = None
448
+
449
+
450
+ AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationAgentTaskRequestWithItemLocatorDesignatedTo: TypeAlias = Union[
451
+ AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationAgentTaskRequestWithItemLocatorDesignatedToApeAgent,
452
+ AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationAgentTaskRequestWithItemLocatorDesignatedToIfAgent,
453
+ AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationAgentTaskRequestWithItemLocatorDesignatedToTruthfulnessAgent,
454
+ AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationAgentTaskRequestWithItemLocatorDesignatedToBaseAgent,
455
+ ]
456
+
457
+
458
+ class AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationAgentTaskRequestWithItemLocator(BaseModel):
459
+ definition: str
460
+
461
+ name: str
462
+
463
+ output_rules: List[str]
464
+
465
+ data_fields: Optional[List[str]] = None
466
+
467
+ designated_to: Optional[
468
+ AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationAgentTaskRequestWithItemLocatorDesignatedTo
469
+ ] = None
470
+
471
+ output_type: Optional[Literal["text", "integer", "float", "boolean"]] = None
472
+
473
+ output_values: Optional[List[Union[str, float, bool]]] = None
474
+
475
+
371
476
  AutoEvaluationGuidedDecodingEvaluationTaskConfiguration: TypeAlias = Union[
372
477
  AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationStructuredOutputTaskRequestWithItemLocator,
373
478
  AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationGuidedDecodingTaskRequestWithItemLocator,
479
+ AutoEvaluationGuidedDecodingEvaluationTaskConfigurationAutoEvaluationAgentTaskRequestWithItemLocator,
374
480
  ]
375
481
 
376
482
 
@@ -386,6 +492,81 @@ class AutoEvaluationGuidedDecodingEvaluationTask(BaseModel):
386
492
  task_type: Optional[Literal["auto_evaluation.guided_decoding"]] = None
387
493
 
388
494
 
495
+ class AutoEvaluationAgentEvaluationTaskConfigurationDesignatedToApeAgentConfig(BaseModel):
496
+ model: Optional[str] = None
497
+
498
+ temperature: Optional[float] = None
499
+
500
+
501
+ class AutoEvaluationAgentEvaluationTaskConfigurationDesignatedToApeAgent(BaseModel):
502
+ config: AutoEvaluationAgentEvaluationTaskConfigurationDesignatedToApeAgentConfig
503
+
504
+ agent_name: Optional[Literal["APEAgent"]] = None
505
+
506
+
507
+ class AutoEvaluationAgentEvaluationTaskConfigurationDesignatedToIfAgentConfig(BaseModel):
508
+ model: Optional[str] = None
509
+
510
+
511
+ class AutoEvaluationAgentEvaluationTaskConfigurationDesignatedToIfAgent(BaseModel):
512
+ config: AutoEvaluationAgentEvaluationTaskConfigurationDesignatedToIfAgentConfig
513
+
514
+ agent_name: Optional[Literal["IFAgent"]] = None
515
+
516
+
517
+ class AutoEvaluationAgentEvaluationTaskConfigurationDesignatedToTruthfulnessAgentConfig(BaseModel):
518
+ model: Optional[str] = None
519
+
520
+
521
+ class AutoEvaluationAgentEvaluationTaskConfigurationDesignatedToTruthfulnessAgent(BaseModel):
522
+ config: AutoEvaluationAgentEvaluationTaskConfigurationDesignatedToTruthfulnessAgentConfig
523
+
524
+ agent_name: Optional[Literal["TruthfulnessAgent"]] = None
525
+
526
+
527
+ class AutoEvaluationAgentEvaluationTaskConfigurationDesignatedToBaseAgentConfig(BaseModel):
528
+ model: Optional[str] = None
529
+
530
+
531
+ class AutoEvaluationAgentEvaluationTaskConfigurationDesignatedToBaseAgent(BaseModel):
532
+ config: AutoEvaluationAgentEvaluationTaskConfigurationDesignatedToBaseAgentConfig
533
+
534
+ agent_name: Optional[Literal["BaseAgent"]] = None
535
+
536
+
537
+ AutoEvaluationAgentEvaluationTaskConfigurationDesignatedTo: TypeAlias = Union[
538
+ AutoEvaluationAgentEvaluationTaskConfigurationDesignatedToApeAgent,
539
+ AutoEvaluationAgentEvaluationTaskConfigurationDesignatedToIfAgent,
540
+ AutoEvaluationAgentEvaluationTaskConfigurationDesignatedToTruthfulnessAgent,
541
+ AutoEvaluationAgentEvaluationTaskConfigurationDesignatedToBaseAgent,
542
+ ]
543
+
544
+
545
+ class AutoEvaluationAgentEvaluationTaskConfiguration(BaseModel):
546
+ definition: str
547
+
548
+ name: str
549
+
550
+ output_rules: List[str]
551
+
552
+ data_fields: Optional[List[str]] = None
553
+
554
+ designated_to: Optional[AutoEvaluationAgentEvaluationTaskConfigurationDesignatedTo] = None
555
+
556
+ output_type: Optional[Literal["text", "integer", "float", "boolean"]] = None
557
+
558
+ output_values: Optional[List[Union[str, float, bool]]] = None
559
+
560
+
561
+ class AutoEvaluationAgentEvaluationTask(BaseModel):
562
+ configuration: AutoEvaluationAgentEvaluationTaskConfiguration
563
+
564
+ alias: Optional[str] = None
565
+ """Alias to title the results column. Defaults to the `auto_evaluation_agent`"""
566
+
567
+ task_type: Optional[Literal["auto_evaluation.agent"]] = None
568
+
569
+
389
570
  class ContributorEvaluationQuestionTaskConfiguration(BaseModel):
390
571
  layout: "Container"
391
572
 
@@ -415,6 +596,7 @@ EvaluationTask: TypeAlias = Annotated[
415
596
  MetricEvaluationTask,
416
597
  AutoEvaluationQuestionTask,
417
598
  AutoEvaluationGuidedDecodingEvaluationTask,
599
+ AutoEvaluationAgentEvaluationTask,
418
600
  ContributorEvaluationQuestionTask,
419
601
  ],
420
602
  PropertyInfo(discriminator="task_type"),