scale-gp-beta 0.1.0a11__py3-none-any.whl → 0.1.0a12__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
scale_gp_beta/_version.py CHANGED
@@ -1,4 +1,4 @@
1
1
  # File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.
2
2
 
3
3
  __title__ = "scale_gp_beta"
4
- __version__ = "0.1.0-alpha.11" # x-release-please-version
4
+ __version__ = "0.1.0-alpha.12" # x-release-please-version
@@ -21,6 +21,15 @@ __all__ = [
21
21
  "ApplicationVariantV1EvaluationTaskConfigurationOverridesAgenticApplicationOverrides",
22
22
  "ApplicationVariantV1EvaluationTaskConfigurationOverridesAgenticApplicationOverridesInitialState",
23
23
  "ApplicationVariantV1EvaluationTaskConfigurationOverridesAgenticApplicationOverridesPartialTrace",
24
+ "MetricEvaluationTask",
25
+ "MetricEvaluationTaskConfiguration",
26
+ "MetricEvaluationTaskConfigurationBleuScorerConfigWithItemLocator",
27
+ "MetricEvaluationTaskConfigurationMeteorScorerConfigWithItemLocator",
28
+ "MetricEvaluationTaskConfigurationCosineSimilarityScorerConfigWithItemLocator",
29
+ "MetricEvaluationTaskConfigurationF1ScorerConfigWithItemLocator",
30
+ "MetricEvaluationTaskConfigurationRougeScorer1ConfigWithItemLocator",
31
+ "MetricEvaluationTaskConfigurationRougeScorer2ConfigWithItemLocator",
32
+ "MetricEvaluationTaskConfigurationRougeScorerLConfigWithItemLocator",
24
33
  ]
25
34
 
26
35
 
@@ -199,7 +208,91 @@ class ApplicationVariantV1EvaluationTask(BaseModel):
199
208
  task_type: Optional[Literal["application_variant"]] = None
200
209
 
201
210
 
211
+ class MetricEvaluationTaskConfigurationBleuScorerConfigWithItemLocator(BaseModel):
212
+ candidate: str
213
+
214
+ reference: str
215
+
216
+ type: Literal["bleu"]
217
+
218
+
219
+ class MetricEvaluationTaskConfigurationMeteorScorerConfigWithItemLocator(BaseModel):
220
+ candidate: str
221
+
222
+ reference: str
223
+
224
+ type: Literal["meteor"]
225
+
226
+
227
+ class MetricEvaluationTaskConfigurationCosineSimilarityScorerConfigWithItemLocator(BaseModel):
228
+ candidate: str
229
+
230
+ reference: str
231
+
232
+ type: Literal["cosine_similarity"]
233
+
234
+
235
+ class MetricEvaluationTaskConfigurationF1ScorerConfigWithItemLocator(BaseModel):
236
+ candidate: str
237
+
238
+ reference: str
239
+
240
+ type: Literal["f1"]
241
+
242
+
243
+ class MetricEvaluationTaskConfigurationRougeScorer1ConfigWithItemLocator(BaseModel):
244
+ candidate: str
245
+
246
+ reference: str
247
+
248
+ type: Literal["rouge1"]
249
+
250
+
251
+ class MetricEvaluationTaskConfigurationRougeScorer2ConfigWithItemLocator(BaseModel):
252
+ candidate: str
253
+
254
+ reference: str
255
+
256
+ type: Literal["rouge2"]
257
+
258
+
259
+ class MetricEvaluationTaskConfigurationRougeScorerLConfigWithItemLocator(BaseModel):
260
+ candidate: str
261
+
262
+ reference: str
263
+
264
+ type: Literal["rougeL"]
265
+
266
+
267
+ MetricEvaluationTaskConfiguration: TypeAlias = Annotated[
268
+ Union[
269
+ MetricEvaluationTaskConfigurationBleuScorerConfigWithItemLocator,
270
+ MetricEvaluationTaskConfigurationMeteorScorerConfigWithItemLocator,
271
+ MetricEvaluationTaskConfigurationCosineSimilarityScorerConfigWithItemLocator,
272
+ MetricEvaluationTaskConfigurationF1ScorerConfigWithItemLocator,
273
+ MetricEvaluationTaskConfigurationRougeScorer1ConfigWithItemLocator,
274
+ MetricEvaluationTaskConfigurationRougeScorer2ConfigWithItemLocator,
275
+ MetricEvaluationTaskConfigurationRougeScorerLConfigWithItemLocator,
276
+ ],
277
+ PropertyInfo(discriminator="type"),
278
+ ]
279
+
280
+
281
+ class MetricEvaluationTask(BaseModel):
282
+ configuration: MetricEvaluationTaskConfiguration
283
+
284
+ alias: Optional[str] = None
285
+ """Alias to title the results column. Defaults to the `task_type`"""
286
+
287
+ task_type: Optional[Literal["metric"]] = None
288
+
289
+
202
290
  EvaluationTask: TypeAlias = Annotated[
203
- Union[ChatCompletionEvaluationTask, GenericInferenceEvaluationTask, ApplicationVariantV1EvaluationTask],
291
+ Union[
292
+ ChatCompletionEvaluationTask,
293
+ GenericInferenceEvaluationTask,
294
+ ApplicationVariantV1EvaluationTask,
295
+ MetricEvaluationTask,
296
+ ],
204
297
  PropertyInfo(discriminator="task_type"),
205
298
  ]
@@ -20,6 +20,15 @@ __all__ = [
20
20
  "ApplicationVariantV1EvaluationTaskConfigurationOverridesAgenticApplicationOverrides",
21
21
  "ApplicationVariantV1EvaluationTaskConfigurationOverridesAgenticApplicationOverridesInitialState",
22
22
  "ApplicationVariantV1EvaluationTaskConfigurationOverridesAgenticApplicationOverridesPartialTrace",
23
+ "MetricEvaluationTask",
24
+ "MetricEvaluationTaskConfiguration",
25
+ "MetricEvaluationTaskConfigurationBleuScorerConfigWithItemLocator",
26
+ "MetricEvaluationTaskConfigurationMeteorScorerConfigWithItemLocator",
27
+ "MetricEvaluationTaskConfigurationCosineSimilarityScorerConfigWithItemLocator",
28
+ "MetricEvaluationTaskConfigurationF1ScorerConfigWithItemLocator",
29
+ "MetricEvaluationTaskConfigurationRougeScorer1ConfigWithItemLocator",
30
+ "MetricEvaluationTaskConfigurationRougeScorer2ConfigWithItemLocator",
31
+ "MetricEvaluationTaskConfigurationRougeScorerLConfigWithItemLocator",
23
32
  ]
24
33
 
25
34
 
@@ -201,6 +210,85 @@ class ApplicationVariantV1EvaluationTask(TypedDict, total=False):
201
210
  task_type: Literal["application_variant"]
202
211
 
203
212
 
213
+ class MetricEvaluationTaskConfigurationBleuScorerConfigWithItemLocator(TypedDict, total=False):
214
+ candidate: Required[str]
215
+
216
+ reference: Required[str]
217
+
218
+ type: Required[Literal["bleu"]]
219
+
220
+
221
+ class MetricEvaluationTaskConfigurationMeteorScorerConfigWithItemLocator(TypedDict, total=False):
222
+ candidate: Required[str]
223
+
224
+ reference: Required[str]
225
+
226
+ type: Required[Literal["meteor"]]
227
+
228
+
229
+ class MetricEvaluationTaskConfigurationCosineSimilarityScorerConfigWithItemLocator(TypedDict, total=False):
230
+ candidate: Required[str]
231
+
232
+ reference: Required[str]
233
+
234
+ type: Required[Literal["cosine_similarity"]]
235
+
236
+
237
+ class MetricEvaluationTaskConfigurationF1ScorerConfigWithItemLocator(TypedDict, total=False):
238
+ candidate: Required[str]
239
+
240
+ reference: Required[str]
241
+
242
+ type: Required[Literal["f1"]]
243
+
244
+
245
+ class MetricEvaluationTaskConfigurationRougeScorer1ConfigWithItemLocator(TypedDict, total=False):
246
+ candidate: Required[str]
247
+
248
+ reference: Required[str]
249
+
250
+ type: Required[Literal["rouge1"]]
251
+
252
+
253
+ class MetricEvaluationTaskConfigurationRougeScorer2ConfigWithItemLocator(TypedDict, total=False):
254
+ candidate: Required[str]
255
+
256
+ reference: Required[str]
257
+
258
+ type: Required[Literal["rouge2"]]
259
+
260
+
261
+ class MetricEvaluationTaskConfigurationRougeScorerLConfigWithItemLocator(TypedDict, total=False):
262
+ candidate: Required[str]
263
+
264
+ reference: Required[str]
265
+
266
+ type: Required[Literal["rougeL"]]
267
+
268
+
269
+ MetricEvaluationTaskConfiguration: TypeAlias = Union[
270
+ MetricEvaluationTaskConfigurationBleuScorerConfigWithItemLocator,
271
+ MetricEvaluationTaskConfigurationMeteorScorerConfigWithItemLocator,
272
+ MetricEvaluationTaskConfigurationCosineSimilarityScorerConfigWithItemLocator,
273
+ MetricEvaluationTaskConfigurationF1ScorerConfigWithItemLocator,
274
+ MetricEvaluationTaskConfigurationRougeScorer1ConfigWithItemLocator,
275
+ MetricEvaluationTaskConfigurationRougeScorer2ConfigWithItemLocator,
276
+ MetricEvaluationTaskConfigurationRougeScorerLConfigWithItemLocator,
277
+ ]
278
+
279
+
280
+ class MetricEvaluationTask(TypedDict, total=False):
281
+ configuration: Required[MetricEvaluationTaskConfiguration]
282
+
283
+ alias: str
284
+ """Alias to title the results column. Defaults to the `task_type`"""
285
+
286
+ task_type: Literal["metric"]
287
+
288
+
204
289
  EvaluationTaskParam: TypeAlias = Union[
205
- ChatCompletionEvaluationTask, GenericInferenceEvaluationTask, ApplicationVariantV1EvaluationTask
290
+ ChatCompletionEvaluationTask,
291
+ GenericInferenceEvaluationTask,
292
+ ApplicationVariantV1EvaluationTask,
293
+ MetricEvaluationTask,
206
294
  ]
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.3
2
2
  Name: scale-gp-beta
3
- Version: 0.1.0a11
3
+ Version: 0.1.0a12
4
4
  Summary: The official Python library for the Scale GP API
5
5
  Project-URL: Homepage, https://github.com/scaleapi/sgp-python-beta
6
6
  Project-URL: Repository, https://github.com/scaleapi/sgp-python-beta
@@ -11,7 +11,7 @@ scale_gp_beta/_resource.py,sha256=siZly_U6D0AOVLAzaOsqUdEFFzVMbWRj-ml30nvRp7E,11
11
11
  scale_gp_beta/_response.py,sha256=GemuybPk0uemovTlGHyHkj-ScYTTDJA0jqH5FQqIPwQ,28852
12
12
  scale_gp_beta/_streaming.py,sha256=fcCSGXslmi2SmmkM05g2SACXHk2Mj7k1X5uMBu6U5s8,10112
13
13
  scale_gp_beta/_types.py,sha256=ScQhVBaKbtJrER3NkXbjokWE9DqSqREMIw9LE0NrFfA,6150
14
- scale_gp_beta/_version.py,sha256=mODcLQ7YXuAUUS-DJtjwA5JDqLi0iX-NsPL3W6BhJLY,174
14
+ scale_gp_beta/_version.py,sha256=8Knp4MLFQyXpPmUI1ikQFJejcnJ6hlWsrVpiUf6jLlA,174
15
15
  scale_gp_beta/pagination.py,sha256=6AAa8_V0wARlMd1MIXijugYbG1mILGc2tHVKbUQbZyQ,2595
16
16
  scale_gp_beta/py.typed,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
17
17
  scale_gp_beta/_utils/__init__.py,sha256=PNZ_QJuzZEgyYXqkO1HVhGkj5IU9bglVUcw7H-Knjzw,2062
@@ -63,8 +63,8 @@ scale_gp_beta/types/evaluation_item_list_params.py,sha256=LquF3dWIU6b7O_Sy_b0R2F
63
63
  scale_gp_beta/types/evaluation_item_retrieve_params.py,sha256=UYEKIAQ4dy92ZOSV1tWDZcvXG7_0BSpOND5Ehzs7QM4,296
64
64
  scale_gp_beta/types/evaluation_list_params.py,sha256=bAYktX3x-rEqNKbBocDyJ0Iqwwz2rZLPK9bHxlU4wYQ,443
65
65
  scale_gp_beta/types/evaluation_retrieve_params.py,sha256=_YuT-E2VO-f_SvHaIe24KBbhTNoK8T-3tVB6Ov6cqfg,356
66
- scale_gp_beta/types/evaluation_task.py,sha256=2ariEI9Hg63xvMo7AlVv74vKFnxxQkuRgm6ODcP_hew,6461
67
- scale_gp_beta/types/evaluation_task_param.py,sha256=8C43Kt2077v4XzU-QAKJe1qigmdsW1_L8PwB2t21QBo,5958
66
+ scale_gp_beta/types/evaluation_task.py,sha256=B6W-7Q9LXJlWcMimCA-FQAFfa6txg7LmGqx8Hterhms,9114
67
+ scale_gp_beta/types/evaluation_task_param.py,sha256=cJz1MyC4CSBuu1R6rGpHNZGrZaEjGdUPqhrGb5eRxDw,8785
68
68
  scale_gp_beta/types/file.py,sha256=Xkha0eSr1q6hkwjE9e2XNgk8kuHNoTEe1LXNhz6o-1k,528
69
69
  scale_gp_beta/types/file_create_params.py,sha256=KpXv6JCbd8BlgceTmBTewxOky2JTJaTW3mcGiVVU7wE,317
70
70
  scale_gp_beta/types/file_delete_response.py,sha256=lOsiaw8qrUOnH7smxb27-n7M4D1chfXlAUaMTRmdldY,336
@@ -91,7 +91,7 @@ scale_gp_beta/types/chat/chat_completion_chunk.py,sha256=57-i6LyOk6IX2HZvXsoUC26
91
91
  scale_gp_beta/types/chat/completion_create_params.py,sha256=Y7vJNvNM4Sov77l55aS5YtyRnrf7isediu3nKr6YE-A,4505
92
92
  scale_gp_beta/types/chat/completion_create_response.py,sha256=0OhfoJW8azVRrZdXRRMuiJ7kEEeMDnKScxrr3sayzDo,374
93
93
  scale_gp_beta/types/files/__init__.py,sha256=OKfJYcKb4NObdiRObqJV_dOyDQ8feXekDUge2o_4pXQ,122
94
- scale_gp_beta-0.1.0a11.dist-info/METADATA,sha256=SdvTUqYYe0ZNeAhUEiyFLYpEGXUmiq5lZ7mHgEjkPts,16942
95
- scale_gp_beta-0.1.0a11.dist-info/WHEEL,sha256=C2FUgwZgiLbznR-k0b_5k3Ai_1aASOXDss3lzCUsUug,87
96
- scale_gp_beta-0.1.0a11.dist-info/licenses/LICENSE,sha256=x49Bj8r_ZpqfzThbmfHyZ_bE88XvHdIMI_ANyLHFFRE,11338
97
- scale_gp_beta-0.1.0a11.dist-info/RECORD,,
94
+ scale_gp_beta-0.1.0a12.dist-info/METADATA,sha256=GBvVuBH4tEVKba9R-Drxp0nXTkMQlYRUYxQrLL7A5Mk,16942
95
+ scale_gp_beta-0.1.0a12.dist-info/WHEEL,sha256=C2FUgwZgiLbznR-k0b_5k3Ai_1aASOXDss3lzCUsUug,87
96
+ scale_gp_beta-0.1.0a12.dist-info/licenses/LICENSE,sha256=x49Bj8r_ZpqfzThbmfHyZ_bE88XvHdIMI_ANyLHFFRE,11338
97
+ scale_gp_beta-0.1.0a12.dist-info/RECORD,,