scale-gp-beta 0.1.0a6__tar.gz → 0.1.0a8__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- scale_gp_beta-0.1.0a8/.release-please-manifest.json +3 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/CHANGELOG.md +23 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/PKG-INFO +1 -1
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/api.md +1 -1
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/pyproject.toml +1 -1
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/_models.py +1 -1
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/_utils/_transform.py +1 -1
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/_version.py +1 -1
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/resources/evaluations.py +23 -18
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/__init__.py +1 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/chat/chat_completion.py +32 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/chat/chat_completion_chunk.py +1 -1
- scale_gp_beta-0.1.0a8/src/scale_gp_beta/types/evaluation_create_params.py +85 -0
- scale_gp_beta-0.1.0a8/src/scale_gp_beta/types/evaluation_task_param.py +212 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/tests/api_resources/test_evaluations.py +2 -8
- scale_gp_beta-0.1.0a6/.release-please-manifest.json +0 -3
- scale_gp_beta-0.1.0a6/src/scale_gp_beta/types/evaluation_create_params.py +0 -734
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/.gitignore +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/CONTRIBUTING.md +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/LICENSE +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/README.md +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/SECURITY.md +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/bin/check-release-environment +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/bin/publish-pypi +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/examples/.keep +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/mypy.ini +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/noxfile.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/release-please-config.json +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/requirements-dev.lock +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/requirements.lock +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp/lib/.keep +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/__init__.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/_base_client.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/_client.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/_compat.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/_constants.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/_exceptions.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/_files.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/_qs.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/_resource.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/_response.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/_streaming.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/_types.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/_utils/__init__.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/_utils/_logs.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/_utils/_proxy.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/_utils/_reflection.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/_utils/_streams.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/_utils/_sync.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/_utils/_typing.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/_utils/_utils.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/lib/.keep +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/pagination.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/py.typed +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/resources/__init__.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/resources/chat/__init__.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/resources/chat/chat.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/resources/chat/completions.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/resources/completions.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/resources/dataset_items.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/resources/datasets.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/resources/evaluation_items.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/resources/files/__init__.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/resources/files/content.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/resources/files/files.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/resources/inference.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/resources/models.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/chat/__init__.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/chat/completion_create_params.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/chat/completion_create_response.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/completion.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/completion_create_params.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/dataset.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/dataset_create_params.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/dataset_delete_response.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/dataset_item.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/dataset_item_batch_create_params.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/dataset_item_batch_create_response.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/dataset_item_delete_response.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/dataset_item_list_params.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/dataset_item_retrieve_params.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/dataset_item_update_params.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/dataset_list_params.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/dataset_retrieve_params.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/dataset_update_params.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/evaluation.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/evaluation_archive_response.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/evaluation_item.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/evaluation_item_list_params.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/evaluation_item_retrieve_params.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/evaluation_list_params.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/evaluation_retrieve_params.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/file.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/file_create_params.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/file_delete_response.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/file_list.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/file_list_params.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/file_update_params.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/files/__init__.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/inference_create_params.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/inference_create_response.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/inference_model.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/inference_model_list.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/inference_response.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/inference_response_chunk.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/model_create_params.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/model_delete_response.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/model_list_params.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/model_update_params.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/sgp_dev/lib/.keep +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/tests/__init__.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/tests/api_resources/__init__.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/tests/api_resources/chat/__init__.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/tests/api_resources/chat/test_completions.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/tests/api_resources/files/__init__.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/tests/api_resources/files/test_content.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/tests/api_resources/test_completions.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/tests/api_resources/test_dataset_items.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/tests/api_resources/test_datasets.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/tests/api_resources/test_evaluation_items.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/tests/api_resources/test_files.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/tests/api_resources/test_inference.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/tests/api_resources/test_models.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/tests/conftest.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/tests/sample_file.txt +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/tests/test_client.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/tests/test_deepcopy.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/tests/test_extract_files.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/tests/test_files.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/tests/test_models.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/tests/test_qs.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/tests/test_required_args.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/tests/test_response.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/tests/test_streaming.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/tests/test_transform.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/tests/test_utils/test_proxy.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/tests/test_utils/test_typing.py +0 -0
- {scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/tests/utils.py +0 -0
|
@@ -1,5 +1,28 @@
|
|
|
1
1
|
# Changelog
|
|
2
2
|
|
|
3
|
+
## 0.1.0-alpha.8 (2025-03-26)
|
|
4
|
+
|
|
5
|
+
Full Changelog: [v0.1.0-alpha.7...v0.1.0-alpha.8](https://github.com/scaleapi/sgp-python-beta/compare/v0.1.0-alpha.7...v0.1.0-alpha.8)
|
|
6
|
+
|
|
7
|
+
### Features
|
|
8
|
+
|
|
9
|
+
* **api:** manual updates ([#83](https://github.com/scaleapi/sgp-python-beta/issues/83)) ([a742789](https://github.com/scaleapi/sgp-python-beta/commit/a74278982def1c9ce87d1568cc3e6f65c992ec8b))
|
|
10
|
+
* **api:** manual updates ([#85](https://github.com/scaleapi/sgp-python-beta/issues/85)) ([de471f9](https://github.com/scaleapi/sgp-python-beta/commit/de471f973d6dfd512841ec7af8b762da4c1c9ade))
|
|
11
|
+
|
|
12
|
+
## 0.1.0-alpha.7 (2025-03-26)
|
|
13
|
+
|
|
14
|
+
Full Changelog: [v0.1.0-alpha.6...v0.1.0-alpha.7](https://github.com/scaleapi/sgp-python-beta/compare/v0.1.0-alpha.6...v0.1.0-alpha.7)
|
|
15
|
+
|
|
16
|
+
### Features
|
|
17
|
+
|
|
18
|
+
* **api:** api update ([#78](https://github.com/scaleapi/sgp-python-beta/issues/78)) ([29cfd2f](https://github.com/scaleapi/sgp-python-beta/commit/29cfd2f2eb0cdc8526ea45c382433f52734c393c))
|
|
19
|
+
* **api:** api update ([#81](https://github.com/scaleapi/sgp-python-beta/issues/81)) ([10144d1](https://github.com/scaleapi/sgp-python-beta/commit/10144d14f6fcbafef4725bbcce06b90fc9c74152))
|
|
20
|
+
|
|
21
|
+
|
|
22
|
+
### Chores
|
|
23
|
+
|
|
24
|
+
* fix typos ([#80](https://github.com/scaleapi/sgp-python-beta/issues/80)) ([89f1f30](https://github.com/scaleapi/sgp-python-beta/commit/89f1f30ab410cc5af7f55dfb7e5bfc4a050324a7))
|
|
25
|
+
|
|
3
26
|
## 0.1.0-alpha.6 (2025-03-25)
|
|
4
27
|
|
|
5
28
|
Full Changelog: [v0.1.0-alpha.5...v0.1.0-alpha.6](https://github.com/scaleapi/sgp-python-beta/compare/v0.1.0-alpha.5...v0.1.0-alpha.6)
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.3
|
|
2
2
|
Name: scale-gp-beta
|
|
3
|
-
Version: 0.1.
|
|
3
|
+
Version: 0.1.0a8
|
|
4
4
|
Summary: The official Python library for the Scale GP API
|
|
5
5
|
Project-URL: Homepage, https://github.com/scaleapi/sgp-python-beta
|
|
6
6
|
Project-URL: Repository, https://github.com/scaleapi/sgp-python-beta
|
|
@@ -681,7 +681,7 @@ def set_pydantic_config(typ: Any, config: pydantic.ConfigDict) -> None:
|
|
|
681
681
|
setattr(typ, "__pydantic_config__", config) # noqa: B010
|
|
682
682
|
|
|
683
683
|
|
|
684
|
-
# our use of
|
|
684
|
+
# our use of subclassing here causes weirdness for type checkers,
|
|
685
685
|
# so we just pretend that we don't subclass
|
|
686
686
|
if TYPE_CHECKING:
|
|
687
687
|
GenericModel = BaseModel
|
|
@@ -126,7 +126,7 @@ def _get_annotated_type(type_: type) -> type | None:
|
|
|
126
126
|
def _maybe_transform_key(key: str, type_: type) -> str:
|
|
127
127
|
"""Transform the given `data` based on the annotations provided in `type_`.
|
|
128
128
|
|
|
129
|
-
Note: this function only looks at `Annotated` types that contain `
|
|
129
|
+
Note: this function only looks at `Annotated` types that contain `PropertyInfo` metadata.
|
|
130
130
|
"""
|
|
131
131
|
annotated_type = _get_annotated_type(type_)
|
|
132
132
|
if annotated_type is None:
|
|
@@ -25,6 +25,7 @@ from .._response import (
|
|
|
25
25
|
from ..pagination import SyncCursorPage, AsyncCursorPage
|
|
26
26
|
from .._base_client import AsyncPaginator, make_request_options
|
|
27
27
|
from ..types.evaluation import Evaluation
|
|
28
|
+
from ..types.evaluation_task_param import EvaluationTaskParam
|
|
28
29
|
from ..types.evaluation_archive_response import EvaluationArchiveResponse
|
|
29
30
|
|
|
30
31
|
__all__ = ["EvaluationsResource", "AsyncEvaluationsResource"]
|
|
@@ -57,7 +58,7 @@ class EvaluationsResource(SyncAPIResource):
|
|
|
57
58
|
data: Iterable[Dict[str, object]],
|
|
58
59
|
name: str,
|
|
59
60
|
description: str | NotGiven = NOT_GIVEN,
|
|
60
|
-
tasks: Iterable[
|
|
61
|
+
tasks: Iterable[EvaluationTaskParam] | NotGiven = NOT_GIVEN,
|
|
61
62
|
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
|
|
62
63
|
# The extra values given here take precedence over values defined on the client or passed to this method.
|
|
63
64
|
extra_headers: Headers | None = None,
|
|
@@ -87,11 +88,11 @@ class EvaluationsResource(SyncAPIResource):
|
|
|
87
88
|
def create(
|
|
88
89
|
self,
|
|
89
90
|
*,
|
|
90
|
-
data: Iterable[evaluation_create_params.EvaluationFromDatasetCreateRequestData],
|
|
91
91
|
dataset_id: str,
|
|
92
92
|
name: str,
|
|
93
|
+
data: Iterable[evaluation_create_params.EvaluationFromDatasetCreateRequestData] | NotGiven = NOT_GIVEN,
|
|
93
94
|
description: str | NotGiven = NOT_GIVEN,
|
|
94
|
-
tasks: Iterable[
|
|
95
|
+
tasks: Iterable[EvaluationTaskParam] | NotGiven = NOT_GIVEN,
|
|
95
96
|
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
|
|
96
97
|
# The extra values given here take precedence over values defined on the client or passed to this method.
|
|
97
98
|
extra_headers: Headers | None = None,
|
|
@@ -103,10 +104,10 @@ class EvaluationsResource(SyncAPIResource):
|
|
|
103
104
|
Create Evaluation
|
|
104
105
|
|
|
105
106
|
Args:
|
|
106
|
-
data: Items to be evaluated, including references to the input dataset items
|
|
107
|
-
|
|
108
107
|
dataset_id: The ID of the dataset containing the items referenced by the `data` field
|
|
109
108
|
|
|
109
|
+
data: Items to be evaluated, including references to the input dataset items
|
|
110
|
+
|
|
110
111
|
tasks: Tasks allow you to augment and evaluate your data
|
|
111
112
|
|
|
112
113
|
extra_headers: Send extra headers
|
|
@@ -127,7 +128,7 @@ class EvaluationsResource(SyncAPIResource):
|
|
|
127
128
|
dataset: evaluation_create_params.EvaluationWithDatasetCreateRequestDataset,
|
|
128
129
|
name: str,
|
|
129
130
|
description: str | NotGiven = NOT_GIVEN,
|
|
130
|
-
tasks: Iterable[
|
|
131
|
+
tasks: Iterable[EvaluationTaskParam] | NotGiven = NOT_GIVEN,
|
|
131
132
|
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
|
|
132
133
|
# The extra values given here take precedence over values defined on the client or passed to this method.
|
|
133
134
|
extra_headers: Headers | None = None,
|
|
@@ -155,14 +156,16 @@ class EvaluationsResource(SyncAPIResource):
|
|
|
155
156
|
"""
|
|
156
157
|
...
|
|
157
158
|
|
|
158
|
-
@required_args(["data", "name"], ["
|
|
159
|
+
@required_args(["data", "name"], ["dataset_id", "name"], ["data", "dataset", "name"])
|
|
159
160
|
def create(
|
|
160
161
|
self,
|
|
161
162
|
*,
|
|
162
|
-
data: Iterable[Dict[str, object]]
|
|
163
|
+
data: Iterable[Dict[str, object]]
|
|
164
|
+
| Iterable[evaluation_create_params.EvaluationFromDatasetCreateRequestData]
|
|
165
|
+
| NotGiven = NOT_GIVEN,
|
|
163
166
|
name: str,
|
|
164
167
|
description: str | NotGiven = NOT_GIVEN,
|
|
165
|
-
tasks: Iterable[
|
|
168
|
+
tasks: Iterable[EvaluationTaskParam] | NotGiven = NOT_GIVEN,
|
|
166
169
|
dataset_id: str | NotGiven = NOT_GIVEN,
|
|
167
170
|
dataset: evaluation_create_params.EvaluationWithDatasetCreateRequestDataset | NotGiven = NOT_GIVEN,
|
|
168
171
|
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
|
|
@@ -339,7 +342,7 @@ class AsyncEvaluationsResource(AsyncAPIResource):
|
|
|
339
342
|
data: Iterable[Dict[str, object]],
|
|
340
343
|
name: str,
|
|
341
344
|
description: str | NotGiven = NOT_GIVEN,
|
|
342
|
-
tasks: Iterable[
|
|
345
|
+
tasks: Iterable[EvaluationTaskParam] | NotGiven = NOT_GIVEN,
|
|
343
346
|
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
|
|
344
347
|
# The extra values given here take precedence over values defined on the client or passed to this method.
|
|
345
348
|
extra_headers: Headers | None = None,
|
|
@@ -369,11 +372,11 @@ class AsyncEvaluationsResource(AsyncAPIResource):
|
|
|
369
372
|
async def create(
|
|
370
373
|
self,
|
|
371
374
|
*,
|
|
372
|
-
data: Iterable[evaluation_create_params.EvaluationFromDatasetCreateRequestData],
|
|
373
375
|
dataset_id: str,
|
|
374
376
|
name: str,
|
|
377
|
+
data: Iterable[evaluation_create_params.EvaluationFromDatasetCreateRequestData] | NotGiven = NOT_GIVEN,
|
|
375
378
|
description: str | NotGiven = NOT_GIVEN,
|
|
376
|
-
tasks: Iterable[
|
|
379
|
+
tasks: Iterable[EvaluationTaskParam] | NotGiven = NOT_GIVEN,
|
|
377
380
|
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
|
|
378
381
|
# The extra values given here take precedence over values defined on the client or passed to this method.
|
|
379
382
|
extra_headers: Headers | None = None,
|
|
@@ -385,10 +388,10 @@ class AsyncEvaluationsResource(AsyncAPIResource):
|
|
|
385
388
|
Create Evaluation
|
|
386
389
|
|
|
387
390
|
Args:
|
|
388
|
-
data: Items to be evaluated, including references to the input dataset items
|
|
389
|
-
|
|
390
391
|
dataset_id: The ID of the dataset containing the items referenced by the `data` field
|
|
391
392
|
|
|
393
|
+
data: Items to be evaluated, including references to the input dataset items
|
|
394
|
+
|
|
392
395
|
tasks: Tasks allow you to augment and evaluate your data
|
|
393
396
|
|
|
394
397
|
extra_headers: Send extra headers
|
|
@@ -409,7 +412,7 @@ class AsyncEvaluationsResource(AsyncAPIResource):
|
|
|
409
412
|
dataset: evaluation_create_params.EvaluationWithDatasetCreateRequestDataset,
|
|
410
413
|
name: str,
|
|
411
414
|
description: str | NotGiven = NOT_GIVEN,
|
|
412
|
-
tasks: Iterable[
|
|
415
|
+
tasks: Iterable[EvaluationTaskParam] | NotGiven = NOT_GIVEN,
|
|
413
416
|
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
|
|
414
417
|
# The extra values given here take precedence over values defined on the client or passed to this method.
|
|
415
418
|
extra_headers: Headers | None = None,
|
|
@@ -437,14 +440,16 @@ class AsyncEvaluationsResource(AsyncAPIResource):
|
|
|
437
440
|
"""
|
|
438
441
|
...
|
|
439
442
|
|
|
440
|
-
@required_args(["data", "name"], ["
|
|
443
|
+
@required_args(["data", "name"], ["dataset_id", "name"], ["data", "dataset", "name"])
|
|
441
444
|
async def create(
|
|
442
445
|
self,
|
|
443
446
|
*,
|
|
444
|
-
data: Iterable[Dict[str, object]]
|
|
447
|
+
data: Iterable[Dict[str, object]]
|
|
448
|
+
| Iterable[evaluation_create_params.EvaluationFromDatasetCreateRequestData]
|
|
449
|
+
| NotGiven = NOT_GIVEN,
|
|
445
450
|
name: str,
|
|
446
451
|
description: str | NotGiven = NOT_GIVEN,
|
|
447
|
-
tasks: Iterable[
|
|
452
|
+
tasks: Iterable[EvaluationTaskParam] | NotGiven = NOT_GIVEN,
|
|
448
453
|
dataset_id: str | NotGiven = NOT_GIVEN,
|
|
449
454
|
dataset: evaluation_create_params.EvaluationWithDatasetCreateRequestDataset | NotGiven = NOT_GIVEN,
|
|
450
455
|
# Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs.
|
|
@@ -22,6 +22,7 @@ from .file_delete_response import FileDeleteResponse as FileDeleteResponse
|
|
|
22
22
|
from .inference_model_list import InferenceModelList as InferenceModelList
|
|
23
23
|
from .dataset_create_params import DatasetCreateParams as DatasetCreateParams
|
|
24
24
|
from .dataset_update_params import DatasetUpdateParams as DatasetUpdateParams
|
|
25
|
+
from .evaluation_task_param import EvaluationTaskParam as EvaluationTaskParam
|
|
25
26
|
from .model_delete_response import ModelDeleteResponse as ModelDeleteResponse
|
|
26
27
|
from .evaluation_list_params import EvaluationListParams as EvaluationListParams
|
|
27
28
|
from .dataset_delete_response import DatasetDeleteResponse as DatasetDeleteResponse
|
{scale_gp_beta-0.1.0a6 → scale_gp_beta-0.1.0a8}/src/scale_gp_beta/types/chat/chat_completion.py
RENAMED
|
@@ -10,6 +10,8 @@ __all__ = [
|
|
|
10
10
|
"ChatCompletion",
|
|
11
11
|
"Choice",
|
|
12
12
|
"ChoiceMessage",
|
|
13
|
+
"ChoiceMessageAnnotation",
|
|
14
|
+
"ChoiceMessageAnnotationURLCitation",
|
|
13
15
|
"ChoiceMessageAudio",
|
|
14
16
|
"ChoiceMessageFunctionCall",
|
|
15
17
|
"ChoiceMessageToolCall",
|
|
@@ -25,6 +27,34 @@ __all__ = [
|
|
|
25
27
|
]
|
|
26
28
|
|
|
27
29
|
|
|
30
|
+
class ChoiceMessageAnnotationURLCitation(BaseModel):
|
|
31
|
+
end_index: int
|
|
32
|
+
|
|
33
|
+
start_index: int
|
|
34
|
+
|
|
35
|
+
title: str
|
|
36
|
+
|
|
37
|
+
url: str
|
|
38
|
+
|
|
39
|
+
if TYPE_CHECKING:
|
|
40
|
+
# Stub to indicate that arbitrary properties are accepted.
|
|
41
|
+
# To access properties that are not valid identifiers you can use `getattr`, e.g.
|
|
42
|
+
# `getattr(obj, '$type')`
|
|
43
|
+
def __getattr__(self, attr: str) -> object: ...
|
|
44
|
+
|
|
45
|
+
|
|
46
|
+
class ChoiceMessageAnnotation(BaseModel):
|
|
47
|
+
type: Literal["url_citation"]
|
|
48
|
+
|
|
49
|
+
url_citation: ChoiceMessageAnnotationURLCitation
|
|
50
|
+
|
|
51
|
+
if TYPE_CHECKING:
|
|
52
|
+
# Stub to indicate that arbitrary properties are accepted.
|
|
53
|
+
# To access properties that are not valid identifiers you can use `getattr`, e.g.
|
|
54
|
+
# `getattr(obj, '$type')`
|
|
55
|
+
def __getattr__(self, attr: str) -> object: ...
|
|
56
|
+
|
|
57
|
+
|
|
28
58
|
class ChoiceMessageAudio(BaseModel):
|
|
29
59
|
id: str
|
|
30
60
|
|
|
@@ -82,6 +112,8 @@ class ChoiceMessageToolCall(BaseModel):
|
|
|
82
112
|
class ChoiceMessage(BaseModel):
|
|
83
113
|
role: Literal["assistant"]
|
|
84
114
|
|
|
115
|
+
annotations: Optional[List[ChoiceMessageAnnotation]] = None
|
|
116
|
+
|
|
85
117
|
audio: Optional[ChoiceMessageAudio] = None
|
|
86
118
|
|
|
87
119
|
content: Optional[str] = None
|
|
@@ -71,7 +71,7 @@ class ChoiceDelta(BaseModel):
|
|
|
71
71
|
|
|
72
72
|
refusal: Optional[str] = None
|
|
73
73
|
|
|
74
|
-
role: Optional[Literal["system", "user", "assistant", "tool"]] = None
|
|
74
|
+
role: Optional[Literal["developer", "system", "user", "assistant", "tool"]] = None
|
|
75
75
|
|
|
76
76
|
tool_calls: Optional[List[ChoiceDeltaToolCall]] = None
|
|
77
77
|
|
|
@@ -0,0 +1,85 @@
|
|
|
1
|
+
# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.
|
|
2
|
+
|
|
3
|
+
from __future__ import annotations
|
|
4
|
+
|
|
5
|
+
from typing import Dict, List, Union, Iterable
|
|
6
|
+
from typing_extensions import Required, TypeAlias, TypedDict
|
|
7
|
+
|
|
8
|
+
from .evaluation_task_param import EvaluationTaskParam
|
|
9
|
+
|
|
10
|
+
__all__ = [
|
|
11
|
+
"EvaluationCreateParams",
|
|
12
|
+
"EvaluationStandaloneCreateRequest",
|
|
13
|
+
"EvaluationFromDatasetCreateRequest",
|
|
14
|
+
"EvaluationFromDatasetCreateRequestData",
|
|
15
|
+
"EvaluationWithDatasetCreateRequest",
|
|
16
|
+
"EvaluationWithDatasetCreateRequestDataset",
|
|
17
|
+
]
|
|
18
|
+
|
|
19
|
+
|
|
20
|
+
class EvaluationStandaloneCreateRequest(TypedDict, total=False):
|
|
21
|
+
data: Required[Iterable[Dict[str, object]]]
|
|
22
|
+
"""Items to be evaluated"""
|
|
23
|
+
|
|
24
|
+
name: Required[str]
|
|
25
|
+
|
|
26
|
+
description: str
|
|
27
|
+
|
|
28
|
+
tasks: Iterable[EvaluationTaskParam]
|
|
29
|
+
"""Tasks allow you to augment and evaluate your data"""
|
|
30
|
+
|
|
31
|
+
|
|
32
|
+
class EvaluationFromDatasetCreateRequest(TypedDict, total=False):
|
|
33
|
+
dataset_id: Required[str]
|
|
34
|
+
"""The ID of the dataset containing the items referenced by the `data` field"""
|
|
35
|
+
|
|
36
|
+
name: Required[str]
|
|
37
|
+
|
|
38
|
+
data: Iterable[EvaluationFromDatasetCreateRequestData]
|
|
39
|
+
"""Items to be evaluated, including references to the input dataset items"""
|
|
40
|
+
|
|
41
|
+
description: str
|
|
42
|
+
|
|
43
|
+
tasks: Iterable[EvaluationTaskParam]
|
|
44
|
+
"""Tasks allow you to augment and evaluate your data"""
|
|
45
|
+
|
|
46
|
+
|
|
47
|
+
class EvaluationFromDatasetCreateRequestDataTyped(TypedDict, total=False):
|
|
48
|
+
dataset_item_id: Required[str]
|
|
49
|
+
|
|
50
|
+
|
|
51
|
+
EvaluationFromDatasetCreateRequestData: TypeAlias = Union[
|
|
52
|
+
EvaluationFromDatasetCreateRequestDataTyped, Dict[str, object]
|
|
53
|
+
]
|
|
54
|
+
|
|
55
|
+
|
|
56
|
+
class EvaluationWithDatasetCreateRequest(TypedDict, total=False):
|
|
57
|
+
data: Required[Iterable[Dict[str, object]]]
|
|
58
|
+
"""Items to be evaluated"""
|
|
59
|
+
|
|
60
|
+
dataset: Required[EvaluationWithDatasetCreateRequestDataset]
|
|
61
|
+
"""Create a reusable dataset from items in the `data` field"""
|
|
62
|
+
|
|
63
|
+
name: Required[str]
|
|
64
|
+
|
|
65
|
+
description: str
|
|
66
|
+
|
|
67
|
+
tasks: Iterable[EvaluationTaskParam]
|
|
68
|
+
"""Tasks allow you to augment and evaluate your data"""
|
|
69
|
+
|
|
70
|
+
|
|
71
|
+
class EvaluationWithDatasetCreateRequestDataset(TypedDict, total=False):
|
|
72
|
+
name: Required[str]
|
|
73
|
+
|
|
74
|
+
description: str
|
|
75
|
+
|
|
76
|
+
keys: List[str]
|
|
77
|
+
"""Keys from items in the `data` field that should be included in the dataset.
|
|
78
|
+
|
|
79
|
+
If not provided, all keys will be included.
|
|
80
|
+
"""
|
|
81
|
+
|
|
82
|
+
|
|
83
|
+
EvaluationCreateParams: TypeAlias = Union[
|
|
84
|
+
EvaluationStandaloneCreateRequest, EvaluationFromDatasetCreateRequest, EvaluationWithDatasetCreateRequest
|
|
85
|
+
]
|
|
@@ -0,0 +1,212 @@
|
|
|
1
|
+
# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details.
|
|
2
|
+
|
|
3
|
+
from __future__ import annotations
|
|
4
|
+
|
|
5
|
+
from typing import Dict, List, Union, Iterable
|
|
6
|
+
from typing_extensions import Literal, Required, TypeAlias, TypedDict
|
|
7
|
+
|
|
8
|
+
__all__ = [
|
|
9
|
+
"EvaluationTaskParam",
|
|
10
|
+
"ChatCompletionEvaluationTaskRequest",
|
|
11
|
+
"ChatCompletionEvaluationTaskRequestConfiguration",
|
|
12
|
+
"GenericInferenceEvaluationTaskRequest",
|
|
13
|
+
"GenericInferenceEvaluationTaskRequestConfiguration",
|
|
14
|
+
"GenericInferenceEvaluationTaskRequestConfigurationInferenceConfiguration",
|
|
15
|
+
"GenericInferenceEvaluationTaskRequestConfigurationInferenceConfigurationLaunchInferenceConfiguration",
|
|
16
|
+
"ApplicationVariantV1EvaluationTaskRequest",
|
|
17
|
+
"ApplicationVariantV1EvaluationTaskRequestConfiguration",
|
|
18
|
+
"ApplicationVariantV1EvaluationTaskRequestConfigurationHistoryUnionMember0",
|
|
19
|
+
"ApplicationVariantV1EvaluationTaskRequestConfigurationOverrides",
|
|
20
|
+
"ApplicationVariantV1EvaluationTaskRequestConfigurationOverridesAgenticApplicationOverrides",
|
|
21
|
+
"ApplicationVariantV1EvaluationTaskRequestConfigurationOverridesAgenticApplicationOverridesInitialState",
|
|
22
|
+
"ApplicationVariantV1EvaluationTaskRequestConfigurationOverridesAgenticApplicationOverridesPartialTrace",
|
|
23
|
+
]
|
|
24
|
+
|
|
25
|
+
|
|
26
|
+
class ChatCompletionEvaluationTaskRequestConfigurationTyped(TypedDict, total=False):
|
|
27
|
+
messages: Required[Union[Iterable[Dict[str, object]], str]]
|
|
28
|
+
|
|
29
|
+
model: Required[str]
|
|
30
|
+
|
|
31
|
+
audio: Union[Dict[str, object], str]
|
|
32
|
+
|
|
33
|
+
frequency_penalty: Union[float, str]
|
|
34
|
+
|
|
35
|
+
function_call: Union[Dict[str, object], str]
|
|
36
|
+
|
|
37
|
+
functions: Union[Iterable[Dict[str, object]], str]
|
|
38
|
+
|
|
39
|
+
logit_bias: Union[Dict[str, int], str]
|
|
40
|
+
|
|
41
|
+
logprobs: Union[bool, str]
|
|
42
|
+
|
|
43
|
+
max_completion_tokens: Union[int, str]
|
|
44
|
+
|
|
45
|
+
max_tokens: Union[int, str]
|
|
46
|
+
|
|
47
|
+
metadata: Union[Dict[str, str], str]
|
|
48
|
+
|
|
49
|
+
modalities: Union[List[str], str]
|
|
50
|
+
|
|
51
|
+
n: Union[int, str]
|
|
52
|
+
|
|
53
|
+
parallel_tool_calls: Union[bool, str]
|
|
54
|
+
|
|
55
|
+
prediction: Union[Dict[str, object], str]
|
|
56
|
+
|
|
57
|
+
presence_penalty: Union[float, str]
|
|
58
|
+
|
|
59
|
+
reasoning_effort: str
|
|
60
|
+
|
|
61
|
+
response_format: Union[Dict[str, object], str]
|
|
62
|
+
|
|
63
|
+
seed: Union[int, str]
|
|
64
|
+
|
|
65
|
+
stop: str
|
|
66
|
+
|
|
67
|
+
store: Union[bool, str]
|
|
68
|
+
|
|
69
|
+
temperature: Union[float, str]
|
|
70
|
+
|
|
71
|
+
tool_choice: str
|
|
72
|
+
|
|
73
|
+
tools: Union[Iterable[Dict[str, object]], str]
|
|
74
|
+
|
|
75
|
+
top_k: Union[int, str]
|
|
76
|
+
|
|
77
|
+
top_logprobs: Union[int, str]
|
|
78
|
+
|
|
79
|
+
top_p: Union[float, str]
|
|
80
|
+
|
|
81
|
+
|
|
82
|
+
ChatCompletionEvaluationTaskRequestConfiguration: TypeAlias = Union[
|
|
83
|
+
ChatCompletionEvaluationTaskRequestConfigurationTyped, Dict[str, object]
|
|
84
|
+
]
|
|
85
|
+
|
|
86
|
+
|
|
87
|
+
class ChatCompletionEvaluationTaskRequest(TypedDict, total=False):
|
|
88
|
+
configuration: Required[ChatCompletionEvaluationTaskRequestConfiguration]
|
|
89
|
+
|
|
90
|
+
alias: str
|
|
91
|
+
"""Alias to title the results column. Defaults to the `task_type`"""
|
|
92
|
+
|
|
93
|
+
task_type: Literal["chat_completion"]
|
|
94
|
+
|
|
95
|
+
|
|
96
|
+
class GenericInferenceEvaluationTaskRequestConfigurationInferenceConfigurationLaunchInferenceConfiguration(
|
|
97
|
+
TypedDict, total=False
|
|
98
|
+
):
|
|
99
|
+
num_retries: int
|
|
100
|
+
|
|
101
|
+
timeout_seconds: int
|
|
102
|
+
|
|
103
|
+
|
|
104
|
+
GenericInferenceEvaluationTaskRequestConfigurationInferenceConfiguration: TypeAlias = Union[
|
|
105
|
+
GenericInferenceEvaluationTaskRequestConfigurationInferenceConfigurationLaunchInferenceConfiguration, str
|
|
106
|
+
]
|
|
107
|
+
|
|
108
|
+
|
|
109
|
+
class GenericInferenceEvaluationTaskRequestConfiguration(TypedDict, total=False):
|
|
110
|
+
model: Required[str]
|
|
111
|
+
|
|
112
|
+
args: Union[Dict[str, object], str]
|
|
113
|
+
|
|
114
|
+
inference_configuration: GenericInferenceEvaluationTaskRequestConfigurationInferenceConfiguration
|
|
115
|
+
|
|
116
|
+
|
|
117
|
+
class GenericInferenceEvaluationTaskRequest(TypedDict, total=False):
|
|
118
|
+
configuration: Required[GenericInferenceEvaluationTaskRequestConfiguration]
|
|
119
|
+
|
|
120
|
+
alias: str
|
|
121
|
+
"""Alias to title the results column. Defaults to the `task_type`"""
|
|
122
|
+
|
|
123
|
+
task_type: Literal["inference"]
|
|
124
|
+
|
|
125
|
+
|
|
126
|
+
class ApplicationVariantV1EvaluationTaskRequestConfigurationHistoryUnionMember0(TypedDict, total=False):
|
|
127
|
+
request: Required[str]
|
|
128
|
+
"""Request inputs"""
|
|
129
|
+
|
|
130
|
+
response: Required[str]
|
|
131
|
+
"""Response outputs"""
|
|
132
|
+
|
|
133
|
+
session_data: Dict[str, object]
|
|
134
|
+
"""Session data corresponding to the request response pair"""
|
|
135
|
+
|
|
136
|
+
|
|
137
|
+
class ApplicationVariantV1EvaluationTaskRequestConfigurationOverridesAgenticApplicationOverridesInitialState(
|
|
138
|
+
TypedDict, total=False
|
|
139
|
+
):
|
|
140
|
+
current_node: Required[str]
|
|
141
|
+
|
|
142
|
+
state: Required[Dict[str, object]]
|
|
143
|
+
|
|
144
|
+
|
|
145
|
+
class ApplicationVariantV1EvaluationTaskRequestConfigurationOverridesAgenticApplicationOverridesPartialTrace(
|
|
146
|
+
TypedDict, total=False
|
|
147
|
+
):
|
|
148
|
+
duration_ms: Required[int]
|
|
149
|
+
|
|
150
|
+
node_id: Required[str]
|
|
151
|
+
|
|
152
|
+
operation_input: Required[str]
|
|
153
|
+
|
|
154
|
+
operation_output: Required[str]
|
|
155
|
+
|
|
156
|
+
operation_type: Required[str]
|
|
157
|
+
|
|
158
|
+
start_timestamp: Required[str]
|
|
159
|
+
|
|
160
|
+
workflow_id: Required[str]
|
|
161
|
+
|
|
162
|
+
operation_metadata: Dict[str, object]
|
|
163
|
+
|
|
164
|
+
|
|
165
|
+
class ApplicationVariantV1EvaluationTaskRequestConfigurationOverridesAgenticApplicationOverrides(
|
|
166
|
+
TypedDict, total=False
|
|
167
|
+
):
|
|
168
|
+
concurrent: bool
|
|
169
|
+
|
|
170
|
+
initial_state: (
|
|
171
|
+
ApplicationVariantV1EvaluationTaskRequestConfigurationOverridesAgenticApplicationOverridesInitialState
|
|
172
|
+
)
|
|
173
|
+
|
|
174
|
+
partial_trace: Iterable[
|
|
175
|
+
ApplicationVariantV1EvaluationTaskRequestConfigurationOverridesAgenticApplicationOverridesPartialTrace
|
|
176
|
+
]
|
|
177
|
+
|
|
178
|
+
use_channels: bool
|
|
179
|
+
|
|
180
|
+
|
|
181
|
+
ApplicationVariantV1EvaluationTaskRequestConfigurationOverrides: TypeAlias = Union[
|
|
182
|
+
ApplicationVariantV1EvaluationTaskRequestConfigurationOverridesAgenticApplicationOverrides, str
|
|
183
|
+
]
|
|
184
|
+
|
|
185
|
+
|
|
186
|
+
class ApplicationVariantV1EvaluationTaskRequestConfiguration(TypedDict, total=False):
|
|
187
|
+
application_variant_id: Required[str]
|
|
188
|
+
|
|
189
|
+
inputs: Required[Union[Dict[str, object], str]]
|
|
190
|
+
|
|
191
|
+
history: Union[Iterable[ApplicationVariantV1EvaluationTaskRequestConfigurationHistoryUnionMember0], str]
|
|
192
|
+
|
|
193
|
+
operation_metadata: Union[Dict[str, object], str]
|
|
194
|
+
|
|
195
|
+
overrides: ApplicationVariantV1EvaluationTaskRequestConfigurationOverrides
|
|
196
|
+
"""Execution override options for agentic applications"""
|
|
197
|
+
|
|
198
|
+
|
|
199
|
+
class ApplicationVariantV1EvaluationTaskRequest(TypedDict, total=False):
|
|
200
|
+
configuration: Required[ApplicationVariantV1EvaluationTaskRequestConfiguration]
|
|
201
|
+
|
|
202
|
+
alias: str
|
|
203
|
+
"""Alias to title the results column. Defaults to the `task_type`"""
|
|
204
|
+
|
|
205
|
+
task_type: Literal["application_variant"]
|
|
206
|
+
|
|
207
|
+
|
|
208
|
+
EvaluationTaskParam: TypeAlias = Union[
|
|
209
|
+
ChatCompletionEvaluationTaskRequest,
|
|
210
|
+
GenericInferenceEvaluationTaskRequest,
|
|
211
|
+
ApplicationVariantV1EvaluationTaskRequest,
|
|
212
|
+
]
|
|
@@ -102,7 +102,6 @@ class TestEvaluations:
|
|
|
102
102
|
@parametrize
|
|
103
103
|
def test_method_create_overload_2(self, client: SGPClient) -> None:
|
|
104
104
|
evaluation = client.evaluations.create(
|
|
105
|
-
data=[{"dataset_item_id": "dataset_item_id"}],
|
|
106
105
|
dataset_id="dataset_id",
|
|
107
106
|
name="name",
|
|
108
107
|
)
|
|
@@ -111,9 +110,9 @@ class TestEvaluations:
|
|
|
111
110
|
@parametrize
|
|
112
111
|
def test_method_create_with_all_params_overload_2(self, client: SGPClient) -> None:
|
|
113
112
|
evaluation = client.evaluations.create(
|
|
114
|
-
data=[{"dataset_item_id": "dataset_item_id"}],
|
|
115
113
|
dataset_id="dataset_id",
|
|
116
114
|
name="name",
|
|
115
|
+
data=[{"dataset_item_id": "dataset_item_id"}],
|
|
117
116
|
description="description",
|
|
118
117
|
tasks=[
|
|
119
118
|
{
|
|
@@ -156,7 +155,6 @@ class TestEvaluations:
|
|
|
156
155
|
@parametrize
|
|
157
156
|
def test_raw_response_create_overload_2(self, client: SGPClient) -> None:
|
|
158
157
|
response = client.evaluations.with_raw_response.create(
|
|
159
|
-
data=[{"dataset_item_id": "dataset_item_id"}],
|
|
160
158
|
dataset_id="dataset_id",
|
|
161
159
|
name="name",
|
|
162
160
|
)
|
|
@@ -169,7 +167,6 @@ class TestEvaluations:
|
|
|
169
167
|
@parametrize
|
|
170
168
|
def test_streaming_response_create_overload_2(self, client: SGPClient) -> None:
|
|
171
169
|
with client.evaluations.with_streaming_response.create(
|
|
172
|
-
data=[{"dataset_item_id": "dataset_item_id"}],
|
|
173
170
|
dataset_id="dataset_id",
|
|
174
171
|
name="name",
|
|
175
172
|
) as response:
|
|
@@ -471,7 +468,6 @@ class TestAsyncEvaluations:
|
|
|
471
468
|
@parametrize
|
|
472
469
|
async def test_method_create_overload_2(self, async_client: AsyncSGPClient) -> None:
|
|
473
470
|
evaluation = await async_client.evaluations.create(
|
|
474
|
-
data=[{"dataset_item_id": "dataset_item_id"}],
|
|
475
471
|
dataset_id="dataset_id",
|
|
476
472
|
name="name",
|
|
477
473
|
)
|
|
@@ -480,9 +476,9 @@ class TestAsyncEvaluations:
|
|
|
480
476
|
@parametrize
|
|
481
477
|
async def test_method_create_with_all_params_overload_2(self, async_client: AsyncSGPClient) -> None:
|
|
482
478
|
evaluation = await async_client.evaluations.create(
|
|
483
|
-
data=[{"dataset_item_id": "dataset_item_id"}],
|
|
484
479
|
dataset_id="dataset_id",
|
|
485
480
|
name="name",
|
|
481
|
+
data=[{"dataset_item_id": "dataset_item_id"}],
|
|
486
482
|
description="description",
|
|
487
483
|
tasks=[
|
|
488
484
|
{
|
|
@@ -525,7 +521,6 @@ class TestAsyncEvaluations:
|
|
|
525
521
|
@parametrize
|
|
526
522
|
async def test_raw_response_create_overload_2(self, async_client: AsyncSGPClient) -> None:
|
|
527
523
|
response = await async_client.evaluations.with_raw_response.create(
|
|
528
|
-
data=[{"dataset_item_id": "dataset_item_id"}],
|
|
529
524
|
dataset_id="dataset_id",
|
|
530
525
|
name="name",
|
|
531
526
|
)
|
|
@@ -538,7 +533,6 @@ class TestAsyncEvaluations:
|
|
|
538
533
|
@parametrize
|
|
539
534
|
async def test_streaming_response_create_overload_2(self, async_client: AsyncSGPClient) -> None:
|
|
540
535
|
async with async_client.evaluations.with_streaming_response.create(
|
|
541
|
-
data=[{"dataset_item_id": "dataset_item_id"}],
|
|
542
536
|
dataset_id="dataset_id",
|
|
543
537
|
name="name",
|
|
544
538
|
) as response:
|