dkist-processing-common 11.8.2rc1__tar.gz → 11.9.0__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/CHANGELOG.rst +21 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/PKG-INFO +17 -13
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/README.rst +10 -6
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/config.py +28 -6
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/graphql.py +33 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/l1_output_data.py +28 -37
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/mixin/globus.py +23 -26
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/mixin/metadata_store.py +49 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/mixin/object_store.py +0 -21
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/mixin/quality/_metrics.py +4 -6
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/trial_catalog.py +1 -5
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/trial_output_data.py +1 -1
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/mock_metadata_store.py +4 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_publish_catalog_messages.py +21 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_quality_mixin.py +11 -3
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_submit_dataset_metadata.py +5 -1
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_trial_output_data.py +2 -1
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common.egg-info/PKG-INFO +17 -13
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common.egg-info/SOURCES.txt +0 -1
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common.egg-info/requires.txt +6 -6
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/pyproject.toml +6 -6
- dkist_processing_common-11.8.2rc1/changelog/276.feature.rst +0 -1
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/.gitignore +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/.pre-commit-config.yaml +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/.readthedocs.yml +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/.snyk +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/bitbucket-pipelines.yml +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/changelog/.gitempty +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/__init__.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/_util/__init__.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/_util/constants.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/_util/graphql.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/_util/scratch.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/_util/tags.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/codecs/__init__.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/codecs/array.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/codecs/asdf.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/codecs/basemodel.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/codecs/bytes.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/codecs/fits.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/codecs/iobase.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/codecs/json.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/codecs/path.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/codecs/quality.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/codecs/str.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/fonts/Lato-Regular.ttf +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/fonts/__init__.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/manual.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/__init__.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/constants.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/dkist_location.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/fits_access.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/flower_pot.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/fried_parameter.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/input_dataset.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/message.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/message_queue_binding.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/metric_code.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/parameters.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/quality.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/tags.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/task_name.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/telemetry.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/wavelength.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/__init__.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/average_bud.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/cs_step.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/dsps_repeat.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/experiment_id_bud.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/id_bud.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/l0_fits_access.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/l1_fits_access.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/near_bud.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/observing_program_id_bud.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/proposal_id_bud.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/quality.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/retarder.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/single_value_single_key_flower.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/task.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/time.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/unique_bud.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/wavelength.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/__init__.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/assemble_movie.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/base.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/mixin/__init__.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/mixin/interservice_bus.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/mixin/quality/__init__.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/mixin/quality/_base.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/output_data_base.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/parse_l0_input_data.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/quality_metrics.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/teardown.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/transfer_input_data.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/write_l1.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/__init__.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/conftest.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_assemble_movie.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_assemble_quality.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_base.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_codecs.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_constants.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_cs_step.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_dkist_location.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_fits_access.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_flower_pot.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_fried_parameter.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_input_dataset.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_interservice_bus.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_interservice_bus_mixin.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_manual_processing.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_output_data_base.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_parameters.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_parse_l0_input_data.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_quality.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_scratch.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_stems.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_tags.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_task_name.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_task_parsing.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_teardown.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_transfer_input_data.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_transfer_l1_output_data.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_trial_catalog.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_workflow_task_base.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_write_l1.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common.egg-info/dependency_links.txt +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common.egg-info/top_level.txt +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/docs/Makefile +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/docs/changelog.rst +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/docs/conf.py +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/docs/index.rst +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/docs/landing_page.rst +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/docs/make.bat +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/docs/requirements.txt +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/licenses/LICENSE.rst +0 -0
- {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/setup.cfg +0 -0
|
@@ -1,3 +1,24 @@
|
|
|
1
|
+
v11.9.0 (2025-12-03)
|
|
2
|
+
====================
|
|
3
|
+
|
|
4
|
+
Misc
|
|
5
|
+
----
|
|
6
|
+
|
|
7
|
+
- Upgrade to the globus 4.x SDK. (`#274 <https://bitbucket.org/dkistdc/dkist-processing-common/pull-requests/274>`__)
|
|
8
|
+
- Use a pool of direction aware globus accounts for transfers to mitigate contention between simultaneous inbound (TransferL0Data) and/or outbound (TransferL1Data) transfers. (`#274 <https://bitbucket.org/dkistdc/dkist-processing-common/pull-requests/274>`__)
|
|
9
|
+
- Integrate dkist-processing-core 6.0.1 which adds additional attributes to metrics and tracing to facilitate discoverability and analysis. (`#275 <https://bitbucket.org/dkistdc/dkist-processing-common/pull-requests/275>`__)
|
|
10
|
+
|
|
11
|
+
|
|
12
|
+
v11.8.1 (2025-12-02)
|
|
13
|
+
====================
|
|
14
|
+
|
|
15
|
+
Misc
|
|
16
|
+
----
|
|
17
|
+
|
|
18
|
+
- Bump minimum version of `pydantic` to 2.7.2 to avoid install failures for python versions >= 3.12. (`#271 <https://bitbucket.org/dkistdc/dkist-processing-common/pull-requests/271>`__)
|
|
19
|
+
- Move `solar-wavelength-calibration` dep to 2.0.0 and make use of new helper properties in that release. (`#271 <https://bitbucket.org/dkistdc/dkist-processing-common/pull-requests/271>`__)
|
|
20
|
+
|
|
21
|
+
|
|
1
22
|
v11.8.0 (2025-11-03)
|
|
2
23
|
====================
|
|
3
24
|
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.4
|
|
2
2
|
Name: dkist-processing-common
|
|
3
|
-
Version: 11.
|
|
3
|
+
Version: 11.9.0
|
|
4
4
|
Summary: Common task classes used by the DKIST science data processing pipelines
|
|
5
5
|
Author-email: NSO / AURA <dkistdc@nso.edu>
|
|
6
6
|
License: BSD-3-Clause
|
|
@@ -17,12 +17,12 @@ Requires-Dist: asdf<4.0.0,>=3.5.0
|
|
|
17
17
|
Requires-Dist: astropy>=7.0.0
|
|
18
18
|
Requires-Dist: dkist-fits-specifications<5.0,>=4.0.0
|
|
19
19
|
Requires-Dist: dkist-header-validator<6.0,>=5.0.0
|
|
20
|
-
Requires-Dist: dkist-processing-core==6.0.
|
|
20
|
+
Requires-Dist: dkist-processing-core==6.0.1
|
|
21
21
|
Requires-Dist: dkist-processing-pac<4.0,>=3.1
|
|
22
22
|
Requires-Dist: dkist-service-configuration<5.0,>=4.1.7
|
|
23
23
|
Requires-Dist: dkist-spectral-lines<4.0,>=3.0.0
|
|
24
|
-
Requires-Dist: solar-wavelength-calibration<
|
|
25
|
-
Requires-Dist: globus-sdk<
|
|
24
|
+
Requires-Dist: solar-wavelength-calibration<3.0,>=2.0.0
|
|
25
|
+
Requires-Dist: globus-sdk<5.0.0,>=4.0.0
|
|
26
26
|
Requires-Dist: gqlclient[pydantic]==1.2.3
|
|
27
27
|
Requires-Dist: sqids==0.5.1
|
|
28
28
|
Requires-Dist: matplotlib>=3.4
|
|
@@ -31,7 +31,7 @@ Requires-Dist: numpy>=1.26.4
|
|
|
31
31
|
Requires-Dist: object-clerk==1.0.0
|
|
32
32
|
Requires-Dist: pandas>=1.4.2
|
|
33
33
|
Requires-Dist: pillow>=10.2.0
|
|
34
|
-
Requires-Dist: pydantic>=2.
|
|
34
|
+
Requires-Dist: pydantic>=2.7.2
|
|
35
35
|
Requires-Dist: redis==6.4.0
|
|
36
36
|
Requires-Dist: requests>=2.23
|
|
37
37
|
Requires-Dist: scipy>=1.15.1
|
|
@@ -57,9 +57,9 @@ Requires-Dist: pytest; extra == "docs"
|
|
|
57
57
|
Requires-Dist: towncrier<22.12.0; extra == "docs"
|
|
58
58
|
Requires-Dist: dkist-sphinx-theme; extra == "docs"
|
|
59
59
|
Provides-Extra: inventory
|
|
60
|
-
Requires-Dist: dkist-inventory
|
|
60
|
+
Requires-Dist: dkist-inventory<2.0,>=1.11.1; extra == "inventory"
|
|
61
61
|
Provides-Extra: asdf
|
|
62
|
-
Requires-Dist: dkist-inventory[asdf]
|
|
62
|
+
Requires-Dist: dkist-inventory[asdf]<2.0,>=1.11.1; extra == "asdf"
|
|
63
63
|
Provides-Extra: quality
|
|
64
64
|
Requires-Dist: dkist-quality<3.0,>=2.0.0; extra == "quality"
|
|
65
65
|
|
|
@@ -118,6 +118,10 @@ Environment Variables
|
|
|
118
118
|
- annotation=str required=False default='unknown-service-version' alias_priority=2 validation_alias='DKIST_SERVICE_VERSION' description='Service version for OpenTelemetry'
|
|
119
119
|
* - NOMAD_ALLOC_ID
|
|
120
120
|
- annotation=str required=False default='unknown-allocation-id' alias_priority=2 validation_alias='NOMAD_ALLOC_ID' description='Nomad allocation ID for OpenTelemetry'
|
|
121
|
+
* - NOMAD_ALLOC_NAME
|
|
122
|
+
- annotation=str required=False default='unknown-allocation-name' alias='NOMAD_ALLOC_NAME' alias_priority=2 description='Allocation name for the deployed container the task is running on.'
|
|
123
|
+
* - NOMAD_GROUP_NAME
|
|
124
|
+
- annotation=str required=False default='unknown-allocation-group' alias='NOMAD_GROUP_NAME' alias_priority=2 description='Allocation group for the deployed container the task is running on'
|
|
121
125
|
* - OTEL_EXPORTER_OTLP_TRACES_INSECURE
|
|
122
126
|
- annotation=bool required=False default=True description='Use insecure connection for OTLP traces'
|
|
123
127
|
* - OTEL_EXPORTER_OTLP_METRICS_INSECURE
|
|
@@ -158,12 +162,12 @@ Environment Variables
|
|
|
158
162
|
- annotation=Union[dict, NoneType] required=False default=None description='S3 upload configuration for the object store.'
|
|
159
163
|
* - S3_DOWNLOAD_CONFIG
|
|
160
164
|
- annotation=Union[dict, NoneType] required=False default=None description='S3 download configuration for the object store.'
|
|
161
|
-
* -
|
|
162
|
-
- annotation=
|
|
163
|
-
* -
|
|
164
|
-
- annotation=
|
|
165
|
-
* -
|
|
166
|
-
- annotation=
|
|
165
|
+
* - GLOBUS_MAX_RETRIES
|
|
166
|
+
- annotation=int required=False default=5 description='Max retries for transient errors on calls to the globus api.'
|
|
167
|
+
* - GLOBUS_INBOUND_CLIENT_CREDENTIALS
|
|
168
|
+
- annotation=list[GlobusClientCredential] required=False default_factory=list description='Globus client credentials for inbound transfers.' examples=[[{'client_id': 'id1', 'client_secret': 'secret1'}, {'client_id': 'id2', 'client_secret': 'secret2'}]]
|
|
169
|
+
* - GLOBUS_OUTBOUND_CLIENT_CREDENTIALS
|
|
170
|
+
- annotation=list[GlobusClientCredential] required=False default_factory=list description='Globus client credentials for outbound transfers.' examples=[[{'client_id': 'id3', 'client_secret': 'secret3'}, {'client_id': 'id4', 'client_secret': 'secret4'}]]
|
|
167
171
|
* - OBJECT_STORE_ENDPOINT
|
|
168
172
|
- annotation=Union[str, NoneType] required=False default=None description='Object store Globus Endpoint ID.'
|
|
169
173
|
* - SCRATCH_ENDPOINT
|
|
@@ -53,6 +53,10 @@ Environment Variables
|
|
|
53
53
|
- annotation=str required=False default='unknown-service-version' alias_priority=2 validation_alias='DKIST_SERVICE_VERSION' description='Service version for OpenTelemetry'
|
|
54
54
|
* - NOMAD_ALLOC_ID
|
|
55
55
|
- annotation=str required=False default='unknown-allocation-id' alias_priority=2 validation_alias='NOMAD_ALLOC_ID' description='Nomad allocation ID for OpenTelemetry'
|
|
56
|
+
* - NOMAD_ALLOC_NAME
|
|
57
|
+
- annotation=str required=False default='unknown-allocation-name' alias='NOMAD_ALLOC_NAME' alias_priority=2 description='Allocation name for the deployed container the task is running on.'
|
|
58
|
+
* - NOMAD_GROUP_NAME
|
|
59
|
+
- annotation=str required=False default='unknown-allocation-group' alias='NOMAD_GROUP_NAME' alias_priority=2 description='Allocation group for the deployed container the task is running on'
|
|
56
60
|
* - OTEL_EXPORTER_OTLP_TRACES_INSECURE
|
|
57
61
|
- annotation=bool required=False default=True description='Use insecure connection for OTLP traces'
|
|
58
62
|
* - OTEL_EXPORTER_OTLP_METRICS_INSECURE
|
|
@@ -93,12 +97,12 @@ Environment Variables
|
|
|
93
97
|
- annotation=Union[dict, NoneType] required=False default=None description='S3 upload configuration for the object store.'
|
|
94
98
|
* - S3_DOWNLOAD_CONFIG
|
|
95
99
|
- annotation=Union[dict, NoneType] required=False default=None description='S3 download configuration for the object store.'
|
|
96
|
-
* -
|
|
97
|
-
- annotation=
|
|
98
|
-
* -
|
|
99
|
-
- annotation=
|
|
100
|
-
* -
|
|
101
|
-
- annotation=
|
|
100
|
+
* - GLOBUS_MAX_RETRIES
|
|
101
|
+
- annotation=int required=False default=5 description='Max retries for transient errors on calls to the globus api.'
|
|
102
|
+
* - GLOBUS_INBOUND_CLIENT_CREDENTIALS
|
|
103
|
+
- annotation=list[GlobusClientCredential] required=False default_factory=list description='Globus client credentials for inbound transfers.' examples=[[{'client_id': 'id1', 'client_secret': 'secret1'}, {'client_id': 'id2', 'client_secret': 'secret2'}]]
|
|
104
|
+
* - GLOBUS_OUTBOUND_CLIENT_CREDENTIALS
|
|
105
|
+
- annotation=list[GlobusClientCredential] required=False default_factory=list description='Globus client credentials for outbound transfers.' examples=[[{'client_id': 'id3', 'client_secret': 'secret3'}, {'client_id': 'id4', 'client_secret': 'secret4'}]]
|
|
102
106
|
* - OBJECT_STORE_ENDPOINT
|
|
103
107
|
- annotation=Union[str, NoneType] required=False default=None description='Object store Globus Endpoint ID.'
|
|
104
108
|
* - SCRATCH_ENDPOINT
|
|
@@ -3,12 +3,20 @@
|
|
|
3
3
|
from dkist_processing_core.config import DKISTProcessingCoreConfiguration
|
|
4
4
|
from dkist_service_configuration.settings import DEFAULT_MESH_SERVICE
|
|
5
5
|
from dkist_service_configuration.settings import MeshService
|
|
6
|
+
from pydantic import BaseModel
|
|
6
7
|
from pydantic import Field
|
|
7
8
|
from talus import ConnectionRetryerFactory
|
|
8
9
|
from talus import ConsumerConnectionParameterFactory
|
|
9
10
|
from talus import ProducerConnectionParameterFactory
|
|
10
11
|
|
|
11
12
|
|
|
13
|
+
class GlobusClientCredential(BaseModel):
|
|
14
|
+
"""Globus client credential."""
|
|
15
|
+
|
|
16
|
+
client_id: str = Field(..., description="Globus client ID for transfers.")
|
|
17
|
+
client_secret: str = Field(..., description="Globus client secret for transfers.")
|
|
18
|
+
|
|
19
|
+
|
|
12
20
|
class DKISTProcessingCommonConfiguration(DKISTProcessingCoreConfiguration):
|
|
13
21
|
"""Common configurations."""
|
|
14
22
|
|
|
@@ -40,14 +48,28 @@ class DKISTProcessingCommonConfiguration(DKISTProcessingCoreConfiguration):
|
|
|
40
48
|
default=None, description="S3 download configuration for the object store."
|
|
41
49
|
)
|
|
42
50
|
# globus
|
|
43
|
-
|
|
44
|
-
|
|
51
|
+
globus_max_retries: int = Field(
|
|
52
|
+
default=5, description="Max retries for transient errors on calls to the globus api."
|
|
45
53
|
)
|
|
46
|
-
|
|
47
|
-
|
|
54
|
+
globus_inbound_client_credentials: list[GlobusClientCredential] = Field(
|
|
55
|
+
default_factory=list,
|
|
56
|
+
description="Globus client credentials for inbound transfers.",
|
|
57
|
+
examples=[
|
|
58
|
+
[
|
|
59
|
+
{"client_id": "id1", "client_secret": "secret1"},
|
|
60
|
+
{"client_id": "id2", "client_secret": "secret2"},
|
|
61
|
+
],
|
|
62
|
+
],
|
|
48
63
|
)
|
|
49
|
-
|
|
50
|
-
|
|
64
|
+
globus_outbound_client_credentials: list[GlobusClientCredential] = Field(
|
|
65
|
+
default_factory=list,
|
|
66
|
+
description="Globus client credentials for outbound transfers.",
|
|
67
|
+
examples=[
|
|
68
|
+
[
|
|
69
|
+
{"client_id": "id3", "client_secret": "secret3"},
|
|
70
|
+
{"client_id": "id4", "client_secret": "secret4"},
|
|
71
|
+
],
|
|
72
|
+
],
|
|
51
73
|
)
|
|
52
74
|
object_store_endpoint: str | None = Field(
|
|
53
75
|
default=None, description="Object store Globus Endpoint ID."
|
|
@@ -176,3 +176,36 @@ class RecipeRunProvenanceMutation(GraphqlBaseModel):
|
|
|
176
176
|
libraryVersions: str
|
|
177
177
|
workflowVersion: str
|
|
178
178
|
codeVersion: str | None = None
|
|
179
|
+
|
|
180
|
+
|
|
181
|
+
class QualityCreation(GraphqlBaseModel):
|
|
182
|
+
"""Quality data creation record."""
|
|
183
|
+
|
|
184
|
+
datasetId: str
|
|
185
|
+
metricCode: str
|
|
186
|
+
facet: str | None = None
|
|
187
|
+
name: str | None = None
|
|
188
|
+
description: str | None = None
|
|
189
|
+
statement: str | None = None
|
|
190
|
+
# JSON array
|
|
191
|
+
warnings: str | None = None
|
|
192
|
+
# JSON objects
|
|
193
|
+
plotData: str | None = None
|
|
194
|
+
multiPlotData: str | None = None
|
|
195
|
+
tableData: str | None = None
|
|
196
|
+
histogramData: str | None = None
|
|
197
|
+
modmatData: str | None = None
|
|
198
|
+
raincloudData: str | None = None
|
|
199
|
+
efficiencyData: str | None = None
|
|
200
|
+
|
|
201
|
+
|
|
202
|
+
class QualitiesRequest(GraphqlBaseModel):
|
|
203
|
+
"""Query parameters for quality data."""
|
|
204
|
+
|
|
205
|
+
datasetId: str
|
|
206
|
+
|
|
207
|
+
|
|
208
|
+
class QualityResponse(GraphqlBaseModel):
|
|
209
|
+
"""Query Response for quality data."""
|
|
210
|
+
|
|
211
|
+
qualityId: int
|
|
@@ -38,9 +38,8 @@ class L1OutputDataBase(OutputDataBase, ABC):
|
|
|
38
38
|
|
|
39
39
|
@property
|
|
40
40
|
def dataset_has_quality_data(self) -> bool:
|
|
41
|
-
"""Return True if
|
|
42
|
-
|
|
43
|
-
return len(paths) > 0
|
|
41
|
+
"""Return True if quality data has been persisted to the metadata-store."""
|
|
42
|
+
return self.metadata_store_quality_data_exists(dataset_id=self.constants.dataset_id)
|
|
44
43
|
|
|
45
44
|
def rollback(self):
|
|
46
45
|
"""Warn that the metadata-store and the interservice bus retain the effect of this tasks execution. Rolling back this task may not be achievable without other action."""
|
|
@@ -59,9 +58,6 @@ class TransferL1Data(TransferDataBase, GlobusMixin):
|
|
|
59
58
|
# Movie needs to be transferred separately as the movie headers need to go with it
|
|
60
59
|
self.transfer_movie()
|
|
61
60
|
|
|
62
|
-
with self.telemetry_span("Upload quality data"):
|
|
63
|
-
self.transfer_quality_data()
|
|
64
|
-
|
|
65
61
|
with self.telemetry_span("Upload science frames"):
|
|
66
62
|
self.transfer_output_frames()
|
|
67
63
|
|
|
@@ -105,33 +101,6 @@ class TransferL1Data(TransferDataBase, GlobusMixin):
|
|
|
105
101
|
content_type="video/mp4",
|
|
106
102
|
)
|
|
107
103
|
|
|
108
|
-
def transfer_quality_data(self):
|
|
109
|
-
"""Transfer quality data to the object store."""
|
|
110
|
-
paths = list(self.read(tags=[Tag.output(), Tag.quality_data()]))
|
|
111
|
-
if len(paths) == 0:
|
|
112
|
-
logger.warning(
|
|
113
|
-
f"No quality data found to upload for dataset. recipe_run_id={self.recipe_run_id}"
|
|
114
|
-
)
|
|
115
|
-
return
|
|
116
|
-
|
|
117
|
-
if count := len(paths) > 1:
|
|
118
|
-
# dataset inventory does not support multiple quality data object keys
|
|
119
|
-
raise RuntimeError(
|
|
120
|
-
f"Found multiple quality data files to upload. Not supported."
|
|
121
|
-
f"{count=}, recipe_run_id={self.recipe_run_id}"
|
|
122
|
-
)
|
|
123
|
-
|
|
124
|
-
with self.telemetry_span(f"Uploading the trial quality data"):
|
|
125
|
-
path = paths[0]
|
|
126
|
-
logger.info(f"Uploading quality data: recipe_run_id={self.recipe_run_id}, {path=}")
|
|
127
|
-
quality_data_object_key = self.format_object_key(path)
|
|
128
|
-
self.object_store_upload_quality_data(
|
|
129
|
-
quality_data=path,
|
|
130
|
-
bucket=self.destination_bucket,
|
|
131
|
-
object_key=quality_data_object_key,
|
|
132
|
-
content_type="application/json",
|
|
133
|
-
)
|
|
134
|
-
|
|
135
104
|
|
|
136
105
|
class AssembleQualityData(L1OutputDataBase, QualityMixin):
|
|
137
106
|
"""
|
|
@@ -159,7 +128,7 @@ class AssembleQualityData(L1OutputDataBase, QualityMixin):
|
|
|
159
128
|
):
|
|
160
129
|
self.write(
|
|
161
130
|
quality_data,
|
|
162
|
-
tags=
|
|
131
|
+
tags=Tag.quality_data(),
|
|
163
132
|
encoder=quality_data_encoder,
|
|
164
133
|
relative_path=f"{self.constants.dataset_id}_quality_data.json",
|
|
165
134
|
)
|
|
@@ -177,10 +146,20 @@ class SubmitDatasetMetadata(L1OutputDataBase):
|
|
|
177
146
|
|
|
178
147
|
def run(self) -> None:
|
|
179
148
|
"""Run method for this task."""
|
|
149
|
+
with self.telemetry_span(f"Storing quality data to metadata store"):
|
|
150
|
+
# each quality_data file is a list - this will combine the elements of multiple lists into a single list
|
|
151
|
+
quality_data = list(
|
|
152
|
+
chain.from_iterable(
|
|
153
|
+
self.read(tags=Tag.quality_data(), decoder=quality_data_decoder)
|
|
154
|
+
)
|
|
155
|
+
)
|
|
156
|
+
self.metadata_store_add_quality_data(
|
|
157
|
+
dataset_id=self.constants.dataset_id, quality_data=quality_data
|
|
158
|
+
)
|
|
180
159
|
with self.telemetry_span("Count Expected Outputs"):
|
|
181
160
|
dataset_id = self.constants.dataset_id
|
|
182
161
|
expected_object_count = self.count(tags=Tag.output())
|
|
183
|
-
if
|
|
162
|
+
if quality_data:
|
|
184
163
|
expected_object_count += 1
|
|
185
164
|
logger.info(
|
|
186
165
|
f"Adding Dataset Receipt Account: "
|
|
@@ -251,12 +230,24 @@ class PublishCatalogAndQualityMessages(L1OutputDataBase, InterserviceBusMixin):
|
|
|
251
230
|
messages = [CatalogObjectMessage(body=body) for body in message_bodies]
|
|
252
231
|
return messages
|
|
253
232
|
|
|
233
|
+
@property
|
|
234
|
+
def quality_report_message(self) -> CreateQualityReportMessage:
|
|
235
|
+
"""Create the Quality Report Message."""
|
|
236
|
+
file_name = Path(f"{self.constants.dataset_id}_quality_report.pdf")
|
|
237
|
+
body = CreateQualityReportMessageBody(
|
|
238
|
+
bucket=self.destination_bucket,
|
|
239
|
+
objectName=self.format_object_key(file_name),
|
|
240
|
+
conversationId=str(self.recipe_run_id),
|
|
241
|
+
datasetId=self.constants.dataset_id,
|
|
242
|
+
incrementDatasetCatalogReceiptCount=True,
|
|
243
|
+
)
|
|
244
|
+
return CreateQualityReportMessage(body=body)
|
|
245
|
+
|
|
254
246
|
def run(self) -> None:
|
|
255
247
|
"""Run method for this task."""
|
|
256
248
|
with self.telemetry_span("Gather output data"):
|
|
257
249
|
frames = self.read(tags=self.output_frame_tags)
|
|
258
250
|
movies = self.read(tags=[Tag.output(), Tag.movie()])
|
|
259
|
-
quality_data = self.read(tags=[Tag.output(), Tag.quality_data()])
|
|
260
251
|
with self.telemetry_span("Create message objects"):
|
|
261
252
|
messages = []
|
|
262
253
|
messages += self.frame_messages(paths=frames)
|
|
@@ -265,7 +256,7 @@ class PublishCatalogAndQualityMessages(L1OutputDataBase, InterserviceBusMixin):
|
|
|
265
256
|
object_message_count = len(messages) - frame_message_count
|
|
266
257
|
dataset_has_quality_data = self.dataset_has_quality_data
|
|
267
258
|
if dataset_has_quality_data:
|
|
268
|
-
messages
|
|
259
|
+
messages.append(self.quality_report_message)
|
|
269
260
|
with self.telemetry_span(
|
|
270
261
|
f"Publish messages: {frame_message_count = }, {object_message_count = }, {dataset_has_quality_data = }"
|
|
271
262
|
):
|
|
@@ -9,6 +9,8 @@ from globus_sdk import ConfidentialAppAuthClient
|
|
|
9
9
|
from globus_sdk import GlobusError
|
|
10
10
|
from globus_sdk import TransferClient
|
|
11
11
|
from globus_sdk import TransferData
|
|
12
|
+
from globus_sdk.scopes import TransferScopes
|
|
13
|
+
from globus_sdk.transport import RetryConfig
|
|
12
14
|
|
|
13
15
|
from dkist_processing_common.config import common_configurations
|
|
14
16
|
|
|
@@ -31,27 +33,32 @@ class GlobusTransferItem:
|
|
|
31
33
|
class GlobusMixin:
|
|
32
34
|
"""Mixin to add methods to a Task to support globus transfers."""
|
|
33
35
|
|
|
34
|
-
|
|
35
|
-
|
|
36
|
-
|
|
37
|
-
|
|
38
|
-
|
|
36
|
+
def globus_transfer_client_factory(self, transfer_data: TransferData) -> TransferClient:
|
|
37
|
+
"""Create a globus transfer client based on the direction of transfer and round-robin the available application credentials."""
|
|
38
|
+
if (
|
|
39
|
+
transfer_data["source_endpoint"] == common_configurations.object_store_endpoint
|
|
40
|
+
): # inbound
|
|
41
|
+
client_credentials = common_configurations.globus_inbound_client_credentials
|
|
42
|
+
else: # outbound
|
|
43
|
+
client_credentials = common_configurations.globus_outbound_client_credentials
|
|
44
|
+
|
|
45
|
+
# Round-robin the client credentials based on the recipe run id
|
|
46
|
+
index = self.recipe_run_id % len(client_credentials)
|
|
47
|
+
selected_credential = client_credentials[index]
|
|
48
|
+
|
|
39
49
|
confidential_client = ConfidentialAppAuthClient(
|
|
40
|
-
client_id=
|
|
41
|
-
client_secret=
|
|
42
|
-
transport_params=common_configurations.globus_transport_params,
|
|
50
|
+
client_id=selected_credential.client_id,
|
|
51
|
+
client_secret=selected_credential.client_secret,
|
|
43
52
|
)
|
|
44
|
-
authorizer = ClientCredentialsAuthorizer(
|
|
45
|
-
|
|
46
|
-
|
|
47
|
-
|
|
48
|
-
return self._globus_transfer_client
|
|
53
|
+
authorizer = ClientCredentialsAuthorizer(confidential_client, scopes=TransferScopes)
|
|
54
|
+
retry_config = RetryConfig(max_retries=common_configurations.globus_max_retries)
|
|
55
|
+
|
|
56
|
+
return TransferClient(authorizer=authorizer, retry_config=retry_config)
|
|
49
57
|
|
|
50
58
|
def globus_transfer_scratch_to_object_store(
|
|
51
59
|
self,
|
|
52
60
|
transfer_items: list[GlobusTransferItem],
|
|
53
61
|
label: str = None,
|
|
54
|
-
sync_level: str = None,
|
|
55
62
|
verify_checksum: bool = True,
|
|
56
63
|
) -> None:
|
|
57
64
|
"""Transfer data from scratch to the object store."""
|
|
@@ -60,7 +67,6 @@ class GlobusMixin:
|
|
|
60
67
|
destination_endpoint=common_configurations.object_store_endpoint,
|
|
61
68
|
transfer_items=transfer_items,
|
|
62
69
|
label=label,
|
|
63
|
-
sync_level=sync_level,
|
|
64
70
|
verify_checksum=verify_checksum,
|
|
65
71
|
)
|
|
66
72
|
|
|
@@ -68,7 +74,6 @@ class GlobusMixin:
|
|
|
68
74
|
self,
|
|
69
75
|
transfer_items: list[GlobusTransferItem],
|
|
70
76
|
label: str = None,
|
|
71
|
-
sync_level: str = None,
|
|
72
77
|
verify_checksum: bool = True,
|
|
73
78
|
) -> None:
|
|
74
79
|
"""Transfer data from the object store to scratch."""
|
|
@@ -77,7 +82,6 @@ class GlobusMixin:
|
|
|
77
82
|
destination_endpoint=common_configurations.scratch_endpoint,
|
|
78
83
|
transfer_items=transfer_items,
|
|
79
84
|
label=label,
|
|
80
|
-
sync_level=sync_level,
|
|
81
85
|
verify_checksum=verify_checksum,
|
|
82
86
|
)
|
|
83
87
|
|
|
@@ -87,7 +91,6 @@ class GlobusMixin:
|
|
|
87
91
|
destination_endpoint: str,
|
|
88
92
|
transfer_items: list[GlobusTransferItem],
|
|
89
93
|
label: str = None,
|
|
90
|
-
sync_level: str = None,
|
|
91
94
|
verify_checksum: bool = True,
|
|
92
95
|
) -> TransferData:
|
|
93
96
|
"""Format a globus TransferData instance."""
|
|
@@ -95,7 +98,6 @@ class GlobusMixin:
|
|
|
95
98
|
source_endpoint=source_endpoint,
|
|
96
99
|
destination_endpoint=destination_endpoint,
|
|
97
100
|
label=label,
|
|
98
|
-
sync_level=sync_level,
|
|
99
101
|
verify_checksum=verify_checksum,
|
|
100
102
|
)
|
|
101
103
|
for item in transfer_items:
|
|
@@ -112,7 +114,6 @@ class GlobusMixin:
|
|
|
112
114
|
destination_endpoint: str,
|
|
113
115
|
transfer_items: list[GlobusTransferItem],
|
|
114
116
|
label: str = None,
|
|
115
|
-
sync_level: str = None,
|
|
116
117
|
verify_checksum: bool = True,
|
|
117
118
|
) -> None:
|
|
118
119
|
"""Perform a transfer of data using globus."""
|
|
@@ -121,7 +122,6 @@ class GlobusMixin:
|
|
|
121
122
|
destination_endpoint=destination_endpoint,
|
|
122
123
|
transfer_items=transfer_items,
|
|
123
124
|
label=label,
|
|
124
|
-
sync_level=sync_level,
|
|
125
125
|
verify_checksum=verify_checksum,
|
|
126
126
|
)
|
|
127
127
|
self._blocking_globus_transfer(transfer_data=transfer_data)
|
|
@@ -131,24 +131,21 @@ class GlobusMixin:
|
|
|
131
131
|
source_endpoint: str,
|
|
132
132
|
destination_endpoint: str,
|
|
133
133
|
label: str = None,
|
|
134
|
-
sync_level: str = None,
|
|
135
134
|
verify_checksum: bool = True,
|
|
136
135
|
) -> TransferData:
|
|
137
136
|
label = label or "Data Processing Transfer"
|
|
138
137
|
return TransferData(
|
|
139
|
-
transfer_client=self.globus_transfer_client,
|
|
140
138
|
source_endpoint=source_endpoint,
|
|
141
139
|
destination_endpoint=destination_endpoint,
|
|
142
140
|
label=label,
|
|
143
|
-
sync_level=sync_level,
|
|
144
141
|
verify_checksum=verify_checksum,
|
|
145
142
|
)
|
|
146
143
|
|
|
147
144
|
def _blocking_globus_transfer(self, transfer_data: TransferData) -> None:
|
|
148
|
-
tc = self.
|
|
149
|
-
logger.info(f"Starting globus transfer: label={transfer_data.get('label')}")
|
|
145
|
+
tc = self.globus_transfer_client_factory(transfer_data=transfer_data)
|
|
150
146
|
transfer_result = tc.submit_transfer(transfer_data)
|
|
151
147
|
task_id = transfer_result["task_id"]
|
|
148
|
+
logger.info(f"Starting globus transfer: label={transfer_data.get('label')}, {task_id=}, ")
|
|
152
149
|
polling_interval = 60
|
|
153
150
|
while not tc.task_wait(
|
|
154
151
|
task_id=task_id, timeout=polling_interval, polling_interval=polling_interval
|
|
@@ -1,5 +1,6 @@
|
|
|
1
1
|
"""Mixin for a WorkflowDataTaskBase subclass which implements Metadata Store data access functionality."""
|
|
2
2
|
|
|
3
|
+
import json
|
|
3
4
|
import logging
|
|
4
5
|
from functools import cached_property
|
|
5
6
|
from typing import Literal
|
|
@@ -7,11 +8,15 @@ from typing import Literal
|
|
|
7
8
|
from pydantic import validate_call
|
|
8
9
|
|
|
9
10
|
from dkist_processing_common._util.graphql import GraphQLClient
|
|
11
|
+
from dkist_processing_common.codecs.quality import QualityDataEncoder
|
|
10
12
|
from dkist_processing_common.config import common_configurations
|
|
11
13
|
from dkist_processing_common.models.graphql import DatasetCatalogReceiptAccountMutation
|
|
12
14
|
from dkist_processing_common.models.graphql import DatasetCatalogReceiptAccountResponse
|
|
13
15
|
from dkist_processing_common.models.graphql import InputDatasetPartResponse
|
|
14
16
|
from dkist_processing_common.models.graphql import InputDatasetRecipeRunResponse
|
|
17
|
+
from dkist_processing_common.models.graphql import QualitiesRequest
|
|
18
|
+
from dkist_processing_common.models.graphql import QualityCreation
|
|
19
|
+
from dkist_processing_common.models.graphql import QualityResponse
|
|
15
20
|
from dkist_processing_common.models.graphql import RecipeRunMutation
|
|
16
21
|
from dkist_processing_common.models.graphql import RecipeRunMutationResponse
|
|
17
22
|
from dkist_processing_common.models.graphql import RecipeRunProvenanceMutation
|
|
@@ -145,6 +150,50 @@ class MetadataStoreMixin:
|
|
|
145
150
|
mutation_response_cls=RecipeRunProvenanceResponse,
|
|
146
151
|
)
|
|
147
152
|
|
|
153
|
+
# QUALITY
|
|
154
|
+
|
|
155
|
+
def metadata_store_add_quality_data(self, dataset_id: str, quality_data: list[dict]):
|
|
156
|
+
"""Add the quality data to the metadata-store."""
|
|
157
|
+
if self.metadata_store_quality_data_exists(dataset_id):
|
|
158
|
+
raise RuntimeError(f"Quality data already persisted for dataset {dataset_id!r}")
|
|
159
|
+
for metric in quality_data:
|
|
160
|
+
if (metric_code := metric.get("metric_code")) is None:
|
|
161
|
+
name = metric.get("name")
|
|
162
|
+
raise ValueError(f"No metric_code for {name!r} in dataset {dataset_id!r}")
|
|
163
|
+
params = QualityCreation(
|
|
164
|
+
datasetId=dataset_id,
|
|
165
|
+
metricCode=metric_code,
|
|
166
|
+
facet=metric.get("facet"),
|
|
167
|
+
name=metric.get("name"),
|
|
168
|
+
description=metric.get("description"),
|
|
169
|
+
statement=metric.get("statement"),
|
|
170
|
+
# JSON array
|
|
171
|
+
warnings=json.dumps(metric.get("warnings")),
|
|
172
|
+
# JSON objects
|
|
173
|
+
plotData=json.dumps(metric.get("plot_data"), cls=QualityDataEncoder),
|
|
174
|
+
multiPlotData=json.dumps(metric.get("multi_plot_data"), cls=QualityDataEncoder),
|
|
175
|
+
tableData=json.dumps(metric.get("table_data"), cls=QualityDataEncoder),
|
|
176
|
+
histogramData=json.dumps(metric.get("histogram_data"), cls=QualityDataEncoder),
|
|
177
|
+
modmatData=json.dumps(metric.get("modmat_data"), cls=QualityDataEncoder),
|
|
178
|
+
raincloudData=json.dumps(metric.get("raincloud_data"), cls=QualityDataEncoder),
|
|
179
|
+
efficiencyData=json.dumps(metric.get("efficiency_data"), cls=QualityDataEncoder),
|
|
180
|
+
)
|
|
181
|
+
self.metadata_store_client.execute_gql_mutation(
|
|
182
|
+
mutation_base="createQuality",
|
|
183
|
+
mutation_parameters=params,
|
|
184
|
+
mutation_response_cls=QualityResponse,
|
|
185
|
+
)
|
|
186
|
+
|
|
187
|
+
def metadata_store_quality_data_exists(self, dataset_id: str) -> bool:
|
|
188
|
+
"""Return True if quality data exists in the metadata-store for the given dataset id."""
|
|
189
|
+
params = QualitiesRequest(datasetId=dataset_id)
|
|
190
|
+
response = self.metadata_store_client.execute_gql_query(
|
|
191
|
+
query_base="qualities",
|
|
192
|
+
query_response_cls=QualityResponse,
|
|
193
|
+
query_parameters=params,
|
|
194
|
+
)
|
|
195
|
+
return bool(response)
|
|
196
|
+
|
|
148
197
|
# INPUT DATASET RECIPE RUN
|
|
149
198
|
|
|
150
199
|
@cached_property
|
|
@@ -55,27 +55,6 @@ class ObjectStoreMixin:
|
|
|
55
55
|
},
|
|
56
56
|
)
|
|
57
57
|
|
|
58
|
-
def object_store_upload_quality_data(
|
|
59
|
-
self,
|
|
60
|
-
quality_data: Path | bytes,
|
|
61
|
-
bucket: str,
|
|
62
|
-
object_key: str,
|
|
63
|
-
content_type: str = "application/json",
|
|
64
|
-
):
|
|
65
|
-
"""Upload quality data to the object store."""
|
|
66
|
-
self.object_store_client.upload_object(
|
|
67
|
-
object_data=quality_data,
|
|
68
|
-
bucket=bucket,
|
|
69
|
-
object_key=object_key,
|
|
70
|
-
verify_checksum=True,
|
|
71
|
-
content_type=content_type,
|
|
72
|
-
metadata={
|
|
73
|
-
"groupname": "DATASET",
|
|
74
|
-
"groupid": self.constants.dataset_id,
|
|
75
|
-
"objecttype": "QDATA",
|
|
76
|
-
},
|
|
77
|
-
)
|
|
78
|
-
|
|
79
58
|
def object_store_remove_folder_objects(self, bucket: str, path: Path | str) -> list[str]:
|
|
80
59
|
"""
|
|
81
60
|
Remove folder objects (end with /) in the specified bucket and path.
|
|
@@ -1356,15 +1356,13 @@ class _WavecalQualityMixin:
|
|
|
1356
1356
|
Note that the residuals are the *unweighed* residuals.
|
|
1357
1357
|
"""
|
|
1358
1358
|
weight_data = np.ones(input_wavelength.size) if weights is None else weights
|
|
1359
|
-
prepared_weights =
|
|
1359
|
+
prepared_weights = fit_result.prepared_weights
|
|
1360
1360
|
residuals = fit_result.minimizer_result.residual / prepared_weights
|
|
1361
1361
|
residuals[~np.isfinite(residuals)] = 0.0
|
|
1362
|
-
best_fit_atlas = input_spectrum - residuals
|
|
1363
1362
|
normalized_residuals = residuals / input_spectrum
|
|
1364
1363
|
|
|
1365
|
-
|
|
1366
|
-
|
|
1367
|
-
best_fit_wavelength = wcs.spectral.pixel_to_world(np.arange(input_spectrum.size))
|
|
1364
|
+
best_fit_atlas = fit_result.best_fit_atlas
|
|
1365
|
+
best_fit_wavelength = fit_result.best_fit_wavelength_vector
|
|
1368
1366
|
|
|
1369
1367
|
finite_idx = (
|
|
1370
1368
|
np.isfinite(input_wavelength)
|
|
@@ -1378,7 +1376,7 @@ class _WavecalQualityMixin:
|
|
|
1378
1376
|
data = {
|
|
1379
1377
|
"input_wavelength_nm": input_wavelength.to_value(u.nm)[finite_idx].tolist(),
|
|
1380
1378
|
"input_spectrum": input_spectrum[finite_idx].tolist(),
|
|
1381
|
-
"best_fit_wavelength_nm": best_fit_wavelength
|
|
1379
|
+
"best_fit_wavelength_nm": best_fit_wavelength[finite_idx].tolist(),
|
|
1382
1380
|
"best_fit_atlas": best_fit_atlas[finite_idx].tolist(),
|
|
1383
1381
|
"normalized_residuals": normalized_residuals[finite_idx].tolist(),
|
|
1384
1382
|
"weights": None if weights is None else weight_data[finite_idx].tolist(),
|
|
@@ -9,8 +9,6 @@ from typing import Any
|
|
|
9
9
|
from typing import Generator
|
|
10
10
|
from uuid import uuid4
|
|
11
11
|
|
|
12
|
-
from dkist_inventory.inventory import generate_quality_report_filename
|
|
13
|
-
|
|
14
12
|
from dkist_processing_common.codecs.asdf import asdf_fileobj_encoder
|
|
15
13
|
from dkist_processing_common.codecs.basemodel import basemodel_decoder
|
|
16
14
|
from dkist_processing_common.codecs.fits import fits_access_decoder
|
|
@@ -247,7 +245,5 @@ class CreateTrialQualityReport(OutputDataBase):
|
|
|
247
245
|
self.write(
|
|
248
246
|
quality_report,
|
|
249
247
|
tags=[Tag.output(), Tag.quality_report()],
|
|
250
|
-
relative_path=
|
|
251
|
-
dataset_id=self.constants.dataset_id
|
|
252
|
-
),
|
|
248
|
+
relative_path=f"{self.constants.dataset_id}_quality_report.pdf",
|
|
253
249
|
)
|
|
@@ -100,7 +100,7 @@ class TransferTrialData(TransferDataBase, GlobusMixin):
|
|
|
100
100
|
tag_list = []
|
|
101
101
|
tag_list += [[Tag.output(), Tag.dataset_inventory()]]
|
|
102
102
|
tag_list += [[Tag.output(), Tag.asdf()]]
|
|
103
|
-
tag_list += [[Tag.
|
|
103
|
+
tag_list += [[Tag.quality_data()]] # quality data is not tagged as OUTPUT
|
|
104
104
|
tag_list += [[Tag.output(), Tag.quality_report()]]
|
|
105
105
|
tag_list += [[Tag.output(), Tag.movie()]]
|
|
106
106
|
return tag_list
|
|
@@ -135,6 +135,10 @@ class InputDatasetRecipeRunResponseMapping(ResponseMapping):
|
|
|
135
135
|
return Unset
|
|
136
136
|
|
|
137
137
|
|
|
138
|
+
class QualityResponseMapping(ResponseMapping):
|
|
139
|
+
pass # TODO
|
|
140
|
+
|
|
141
|
+
|
|
138
142
|
def make_default_recipe_run_status_response() -> RecipeRunStatusResponse:
|
|
139
143
|
return RecipeRunStatusResponse(recipeRunStatusId=1)
|
|
140
144
|
|