dkist-processing-common 11.8.2rc1__tar.gz → 11.9.0__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (137) hide show
  1. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/CHANGELOG.rst +21 -0
  2. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/PKG-INFO +17 -13
  3. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/README.rst +10 -6
  4. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/config.py +28 -6
  5. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/graphql.py +33 -0
  6. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/l1_output_data.py +28 -37
  7. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/mixin/globus.py +23 -26
  8. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/mixin/metadata_store.py +49 -0
  9. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/mixin/object_store.py +0 -21
  10. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/mixin/quality/_metrics.py +4 -6
  11. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/trial_catalog.py +1 -5
  12. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/trial_output_data.py +1 -1
  13. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/mock_metadata_store.py +4 -0
  14. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_publish_catalog_messages.py +21 -0
  15. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_quality_mixin.py +11 -3
  16. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_submit_dataset_metadata.py +5 -1
  17. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_trial_output_data.py +2 -1
  18. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common.egg-info/PKG-INFO +17 -13
  19. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common.egg-info/SOURCES.txt +0 -1
  20. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common.egg-info/requires.txt +6 -6
  21. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/pyproject.toml +6 -6
  22. dkist_processing_common-11.8.2rc1/changelog/276.feature.rst +0 -1
  23. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/.gitignore +0 -0
  24. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/.pre-commit-config.yaml +0 -0
  25. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/.readthedocs.yml +0 -0
  26. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/.snyk +0 -0
  27. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/bitbucket-pipelines.yml +0 -0
  28. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/changelog/.gitempty +0 -0
  29. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/__init__.py +0 -0
  30. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/_util/__init__.py +0 -0
  31. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/_util/constants.py +0 -0
  32. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/_util/graphql.py +0 -0
  33. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/_util/scratch.py +0 -0
  34. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/_util/tags.py +0 -0
  35. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/codecs/__init__.py +0 -0
  36. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/codecs/array.py +0 -0
  37. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/codecs/asdf.py +0 -0
  38. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/codecs/basemodel.py +0 -0
  39. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/codecs/bytes.py +0 -0
  40. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/codecs/fits.py +0 -0
  41. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/codecs/iobase.py +0 -0
  42. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/codecs/json.py +0 -0
  43. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/codecs/path.py +0 -0
  44. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/codecs/quality.py +0 -0
  45. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/codecs/str.py +0 -0
  46. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/fonts/Lato-Regular.ttf +0 -0
  47. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/fonts/__init__.py +0 -0
  48. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/manual.py +0 -0
  49. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/__init__.py +0 -0
  50. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/constants.py +0 -0
  51. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/dkist_location.py +0 -0
  52. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/fits_access.py +0 -0
  53. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/flower_pot.py +0 -0
  54. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/fried_parameter.py +0 -0
  55. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/input_dataset.py +0 -0
  56. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/message.py +0 -0
  57. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/message_queue_binding.py +0 -0
  58. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/metric_code.py +0 -0
  59. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/parameters.py +0 -0
  60. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/quality.py +0 -0
  61. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/tags.py +0 -0
  62. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/task_name.py +0 -0
  63. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/telemetry.py +0 -0
  64. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/models/wavelength.py +0 -0
  65. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/__init__.py +0 -0
  66. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/average_bud.py +0 -0
  67. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/cs_step.py +0 -0
  68. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/dsps_repeat.py +0 -0
  69. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/experiment_id_bud.py +0 -0
  70. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/id_bud.py +0 -0
  71. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/l0_fits_access.py +0 -0
  72. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/l1_fits_access.py +0 -0
  73. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/near_bud.py +0 -0
  74. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/observing_program_id_bud.py +0 -0
  75. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/proposal_id_bud.py +0 -0
  76. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/quality.py +0 -0
  77. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/retarder.py +0 -0
  78. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/single_value_single_key_flower.py +0 -0
  79. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/task.py +0 -0
  80. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/time.py +0 -0
  81. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/unique_bud.py +0 -0
  82. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/parsers/wavelength.py +0 -0
  83. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/__init__.py +0 -0
  84. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/assemble_movie.py +0 -0
  85. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/base.py +0 -0
  86. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/mixin/__init__.py +0 -0
  87. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/mixin/interservice_bus.py +0 -0
  88. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/mixin/quality/__init__.py +0 -0
  89. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/mixin/quality/_base.py +0 -0
  90. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/output_data_base.py +0 -0
  91. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/parse_l0_input_data.py +0 -0
  92. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/quality_metrics.py +0 -0
  93. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/teardown.py +0 -0
  94. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/transfer_input_data.py +0 -0
  95. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tasks/write_l1.py +0 -0
  96. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/__init__.py +0 -0
  97. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/conftest.py +0 -0
  98. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_assemble_movie.py +0 -0
  99. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_assemble_quality.py +0 -0
  100. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_base.py +0 -0
  101. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_codecs.py +0 -0
  102. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_constants.py +0 -0
  103. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_cs_step.py +0 -0
  104. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_dkist_location.py +0 -0
  105. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_fits_access.py +0 -0
  106. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_flower_pot.py +0 -0
  107. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_fried_parameter.py +0 -0
  108. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_input_dataset.py +0 -0
  109. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_interservice_bus.py +0 -0
  110. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_interservice_bus_mixin.py +0 -0
  111. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_manual_processing.py +0 -0
  112. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_output_data_base.py +0 -0
  113. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_parameters.py +0 -0
  114. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_parse_l0_input_data.py +0 -0
  115. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_quality.py +0 -0
  116. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_scratch.py +0 -0
  117. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_stems.py +0 -0
  118. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_tags.py +0 -0
  119. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_task_name.py +0 -0
  120. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_task_parsing.py +0 -0
  121. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_teardown.py +0 -0
  122. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_transfer_input_data.py +0 -0
  123. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_transfer_l1_output_data.py +0 -0
  124. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_trial_catalog.py +0 -0
  125. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_workflow_task_base.py +0 -0
  126. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common/tests/test_write_l1.py +0 -0
  127. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common.egg-info/dependency_links.txt +0 -0
  128. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/dkist_processing_common.egg-info/top_level.txt +0 -0
  129. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/docs/Makefile +0 -0
  130. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/docs/changelog.rst +0 -0
  131. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/docs/conf.py +0 -0
  132. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/docs/index.rst +0 -0
  133. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/docs/landing_page.rst +0 -0
  134. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/docs/make.bat +0 -0
  135. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/docs/requirements.txt +0 -0
  136. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/licenses/LICENSE.rst +0 -0
  137. {dkist_processing_common-11.8.2rc1 → dkist_processing_common-11.9.0}/setup.cfg +0 -0
@@ -1,3 +1,24 @@
1
+ v11.9.0 (2025-12-03)
2
+ ====================
3
+
4
+ Misc
5
+ ----
6
+
7
+ - Upgrade to the globus 4.x SDK. (`#274 <https://bitbucket.org/dkistdc/dkist-processing-common/pull-requests/274>`__)
8
+ - Use a pool of direction aware globus accounts for transfers to mitigate contention between simultaneous inbound (TransferL0Data) and/or outbound (TransferL1Data) transfers. (`#274 <https://bitbucket.org/dkistdc/dkist-processing-common/pull-requests/274>`__)
9
+ - Integrate dkist-processing-core 6.0.1 which adds additional attributes to metrics and tracing to facilitate discoverability and analysis. (`#275 <https://bitbucket.org/dkistdc/dkist-processing-common/pull-requests/275>`__)
10
+
11
+
12
+ v11.8.1 (2025-12-02)
13
+ ====================
14
+
15
+ Misc
16
+ ----
17
+
18
+ - Bump minimum version of `pydantic` to 2.7.2 to avoid install failures for python versions >= 3.12. (`#271 <https://bitbucket.org/dkistdc/dkist-processing-common/pull-requests/271>`__)
19
+ - Move `solar-wavelength-calibration` dep to 2.0.0 and make use of new helper properties in that release. (`#271 <https://bitbucket.org/dkistdc/dkist-processing-common/pull-requests/271>`__)
20
+
21
+
1
22
  v11.8.0 (2025-11-03)
2
23
  ====================
3
24
 
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: dkist-processing-common
3
- Version: 11.8.2rc1
3
+ Version: 11.9.0
4
4
  Summary: Common task classes used by the DKIST science data processing pipelines
5
5
  Author-email: NSO / AURA <dkistdc@nso.edu>
6
6
  License: BSD-3-Clause
@@ -17,12 +17,12 @@ Requires-Dist: asdf<4.0.0,>=3.5.0
17
17
  Requires-Dist: astropy>=7.0.0
18
18
  Requires-Dist: dkist-fits-specifications<5.0,>=4.0.0
19
19
  Requires-Dist: dkist-header-validator<6.0,>=5.0.0
20
- Requires-Dist: dkist-processing-core==6.0.0
20
+ Requires-Dist: dkist-processing-core==6.0.1
21
21
  Requires-Dist: dkist-processing-pac<4.0,>=3.1
22
22
  Requires-Dist: dkist-service-configuration<5.0,>=4.1.7
23
23
  Requires-Dist: dkist-spectral-lines<4.0,>=3.0.0
24
- Requires-Dist: solar-wavelength-calibration<2.0,>=1.0
25
- Requires-Dist: globus-sdk<4.0.0,>=3.12.0
24
+ Requires-Dist: solar-wavelength-calibration<3.0,>=2.0.0
25
+ Requires-Dist: globus-sdk<5.0.0,>=4.0.0
26
26
  Requires-Dist: gqlclient[pydantic]==1.2.3
27
27
  Requires-Dist: sqids==0.5.1
28
28
  Requires-Dist: matplotlib>=3.4
@@ -31,7 +31,7 @@ Requires-Dist: numpy>=1.26.4
31
31
  Requires-Dist: object-clerk==1.0.0
32
32
  Requires-Dist: pandas>=1.4.2
33
33
  Requires-Dist: pillow>=10.2.0
34
- Requires-Dist: pydantic>=2.0
34
+ Requires-Dist: pydantic>=2.7.2
35
35
  Requires-Dist: redis==6.4.0
36
36
  Requires-Dist: requests>=2.23
37
37
  Requires-Dist: scipy>=1.15.1
@@ -57,9 +57,9 @@ Requires-Dist: pytest; extra == "docs"
57
57
  Requires-Dist: towncrier<22.12.0; extra == "docs"
58
58
  Requires-Dist: dkist-sphinx-theme; extra == "docs"
59
59
  Provides-Extra: inventory
60
- Requires-Dist: dkist-inventory==1.11.3rc2; extra == "inventory"
60
+ Requires-Dist: dkist-inventory<2.0,>=1.11.1; extra == "inventory"
61
61
  Provides-Extra: asdf
62
- Requires-Dist: dkist-inventory[asdf]==1.11.3rc2; extra == "asdf"
62
+ Requires-Dist: dkist-inventory[asdf]<2.0,>=1.11.1; extra == "asdf"
63
63
  Provides-Extra: quality
64
64
  Requires-Dist: dkist-quality<3.0,>=2.0.0; extra == "quality"
65
65
 
@@ -118,6 +118,10 @@ Environment Variables
118
118
  - annotation=str required=False default='unknown-service-version' alias_priority=2 validation_alias='DKIST_SERVICE_VERSION' description='Service version for OpenTelemetry'
119
119
  * - NOMAD_ALLOC_ID
120
120
  - annotation=str required=False default='unknown-allocation-id' alias_priority=2 validation_alias='NOMAD_ALLOC_ID' description='Nomad allocation ID for OpenTelemetry'
121
+ * - NOMAD_ALLOC_NAME
122
+ - annotation=str required=False default='unknown-allocation-name' alias='NOMAD_ALLOC_NAME' alias_priority=2 description='Allocation name for the deployed container the task is running on.'
123
+ * - NOMAD_GROUP_NAME
124
+ - annotation=str required=False default='unknown-allocation-group' alias='NOMAD_GROUP_NAME' alias_priority=2 description='Allocation group for the deployed container the task is running on'
121
125
  * - OTEL_EXPORTER_OTLP_TRACES_INSECURE
122
126
  - annotation=bool required=False default=True description='Use insecure connection for OTLP traces'
123
127
  * - OTEL_EXPORTER_OTLP_METRICS_INSECURE
@@ -158,12 +162,12 @@ Environment Variables
158
162
  - annotation=Union[dict, NoneType] required=False default=None description='S3 upload configuration for the object store.'
159
163
  * - S3_DOWNLOAD_CONFIG
160
164
  - annotation=Union[dict, NoneType] required=False default=None description='S3 download configuration for the object store.'
161
- * - GLOBUS_TRANSPORT_PARAMS
162
- - annotation=dict required=False default_factory=dict description='Globus transfer parameters.'
163
- * - GLOBUS_CLIENT_ID
164
- - annotation=Union[str, NoneType] required=False default=None description='Globus client ID for inbound/outbound transfers.'
165
- * - GLOBUS_CLIENT_SECRET
166
- - annotation=Union[str, NoneType] required=False default=None description='Globus client secret for inbound/outbound transfers.'
165
+ * - GLOBUS_MAX_RETRIES
166
+ - annotation=int required=False default=5 description='Max retries for transient errors on calls to the globus api.'
167
+ * - GLOBUS_INBOUND_CLIENT_CREDENTIALS
168
+ - annotation=list[GlobusClientCredential] required=False default_factory=list description='Globus client credentials for inbound transfers.' examples=[[{'client_id': 'id1', 'client_secret': 'secret1'}, {'client_id': 'id2', 'client_secret': 'secret2'}]]
169
+ * - GLOBUS_OUTBOUND_CLIENT_CREDENTIALS
170
+ - annotation=list[GlobusClientCredential] required=False default_factory=list description='Globus client credentials for outbound transfers.' examples=[[{'client_id': 'id3', 'client_secret': 'secret3'}, {'client_id': 'id4', 'client_secret': 'secret4'}]]
167
171
  * - OBJECT_STORE_ENDPOINT
168
172
  - annotation=Union[str, NoneType] required=False default=None description='Object store Globus Endpoint ID.'
169
173
  * - SCRATCH_ENDPOINT
@@ -53,6 +53,10 @@ Environment Variables
53
53
  - annotation=str required=False default='unknown-service-version' alias_priority=2 validation_alias='DKIST_SERVICE_VERSION' description='Service version for OpenTelemetry'
54
54
  * - NOMAD_ALLOC_ID
55
55
  - annotation=str required=False default='unknown-allocation-id' alias_priority=2 validation_alias='NOMAD_ALLOC_ID' description='Nomad allocation ID for OpenTelemetry'
56
+ * - NOMAD_ALLOC_NAME
57
+ - annotation=str required=False default='unknown-allocation-name' alias='NOMAD_ALLOC_NAME' alias_priority=2 description='Allocation name for the deployed container the task is running on.'
58
+ * - NOMAD_GROUP_NAME
59
+ - annotation=str required=False default='unknown-allocation-group' alias='NOMAD_GROUP_NAME' alias_priority=2 description='Allocation group for the deployed container the task is running on'
56
60
  * - OTEL_EXPORTER_OTLP_TRACES_INSECURE
57
61
  - annotation=bool required=False default=True description='Use insecure connection for OTLP traces'
58
62
  * - OTEL_EXPORTER_OTLP_METRICS_INSECURE
@@ -93,12 +97,12 @@ Environment Variables
93
97
  - annotation=Union[dict, NoneType] required=False default=None description='S3 upload configuration for the object store.'
94
98
  * - S3_DOWNLOAD_CONFIG
95
99
  - annotation=Union[dict, NoneType] required=False default=None description='S3 download configuration for the object store.'
96
- * - GLOBUS_TRANSPORT_PARAMS
97
- - annotation=dict required=False default_factory=dict description='Globus transfer parameters.'
98
- * - GLOBUS_CLIENT_ID
99
- - annotation=Union[str, NoneType] required=False default=None description='Globus client ID for inbound/outbound transfers.'
100
- * - GLOBUS_CLIENT_SECRET
101
- - annotation=Union[str, NoneType] required=False default=None description='Globus client secret for inbound/outbound transfers.'
100
+ * - GLOBUS_MAX_RETRIES
101
+ - annotation=int required=False default=5 description='Max retries for transient errors on calls to the globus api.'
102
+ * - GLOBUS_INBOUND_CLIENT_CREDENTIALS
103
+ - annotation=list[GlobusClientCredential] required=False default_factory=list description='Globus client credentials for inbound transfers.' examples=[[{'client_id': 'id1', 'client_secret': 'secret1'}, {'client_id': 'id2', 'client_secret': 'secret2'}]]
104
+ * - GLOBUS_OUTBOUND_CLIENT_CREDENTIALS
105
+ - annotation=list[GlobusClientCredential] required=False default_factory=list description='Globus client credentials for outbound transfers.' examples=[[{'client_id': 'id3', 'client_secret': 'secret3'}, {'client_id': 'id4', 'client_secret': 'secret4'}]]
102
106
  * - OBJECT_STORE_ENDPOINT
103
107
  - annotation=Union[str, NoneType] required=False default=None description='Object store Globus Endpoint ID.'
104
108
  * - SCRATCH_ENDPOINT
@@ -3,12 +3,20 @@
3
3
  from dkist_processing_core.config import DKISTProcessingCoreConfiguration
4
4
  from dkist_service_configuration.settings import DEFAULT_MESH_SERVICE
5
5
  from dkist_service_configuration.settings import MeshService
6
+ from pydantic import BaseModel
6
7
  from pydantic import Field
7
8
  from talus import ConnectionRetryerFactory
8
9
  from talus import ConsumerConnectionParameterFactory
9
10
  from talus import ProducerConnectionParameterFactory
10
11
 
11
12
 
13
+ class GlobusClientCredential(BaseModel):
14
+ """Globus client credential."""
15
+
16
+ client_id: str = Field(..., description="Globus client ID for transfers.")
17
+ client_secret: str = Field(..., description="Globus client secret for transfers.")
18
+
19
+
12
20
  class DKISTProcessingCommonConfiguration(DKISTProcessingCoreConfiguration):
13
21
  """Common configurations."""
14
22
 
@@ -40,14 +48,28 @@ class DKISTProcessingCommonConfiguration(DKISTProcessingCoreConfiguration):
40
48
  default=None, description="S3 download configuration for the object store."
41
49
  )
42
50
  # globus
43
- globus_transport_params: dict = Field(
44
- default_factory=dict, description="Globus transfer parameters."
51
+ globus_max_retries: int = Field(
52
+ default=5, description="Max retries for transient errors on calls to the globus api."
45
53
  )
46
- globus_client_id: str | None = Field(
47
- default=None, description="Globus client ID for inbound/outbound transfers."
54
+ globus_inbound_client_credentials: list[GlobusClientCredential] = Field(
55
+ default_factory=list,
56
+ description="Globus client credentials for inbound transfers.",
57
+ examples=[
58
+ [
59
+ {"client_id": "id1", "client_secret": "secret1"},
60
+ {"client_id": "id2", "client_secret": "secret2"},
61
+ ],
62
+ ],
48
63
  )
49
- globus_client_secret: str | None = Field(
50
- default=None, description="Globus client secret for inbound/outbound transfers."
64
+ globus_outbound_client_credentials: list[GlobusClientCredential] = Field(
65
+ default_factory=list,
66
+ description="Globus client credentials for outbound transfers.",
67
+ examples=[
68
+ [
69
+ {"client_id": "id3", "client_secret": "secret3"},
70
+ {"client_id": "id4", "client_secret": "secret4"},
71
+ ],
72
+ ],
51
73
  )
52
74
  object_store_endpoint: str | None = Field(
53
75
  default=None, description="Object store Globus Endpoint ID."
@@ -176,3 +176,36 @@ class RecipeRunProvenanceMutation(GraphqlBaseModel):
176
176
  libraryVersions: str
177
177
  workflowVersion: str
178
178
  codeVersion: str | None = None
179
+
180
+
181
+ class QualityCreation(GraphqlBaseModel):
182
+ """Quality data creation record."""
183
+
184
+ datasetId: str
185
+ metricCode: str
186
+ facet: str | None = None
187
+ name: str | None = None
188
+ description: str | None = None
189
+ statement: str | None = None
190
+ # JSON array
191
+ warnings: str | None = None
192
+ # JSON objects
193
+ plotData: str | None = None
194
+ multiPlotData: str | None = None
195
+ tableData: str | None = None
196
+ histogramData: str | None = None
197
+ modmatData: str | None = None
198
+ raincloudData: str | None = None
199
+ efficiencyData: str | None = None
200
+
201
+
202
+ class QualitiesRequest(GraphqlBaseModel):
203
+ """Query parameters for quality data."""
204
+
205
+ datasetId: str
206
+
207
+
208
+ class QualityResponse(GraphqlBaseModel):
209
+ """Query Response for quality data."""
210
+
211
+ qualityId: int
@@ -38,9 +38,8 @@ class L1OutputDataBase(OutputDataBase, ABC):
38
38
 
39
39
  @property
40
40
  def dataset_has_quality_data(self) -> bool:
41
- """Return True if the dataset has quality data."""
42
- paths = list(self.read(tags=[Tag.output(), Tag.quality_data()]))
43
- return len(paths) > 0
41
+ """Return True if quality data has been persisted to the metadata-store."""
42
+ return self.metadata_store_quality_data_exists(dataset_id=self.constants.dataset_id)
44
43
 
45
44
  def rollback(self):
46
45
  """Warn that the metadata-store and the interservice bus retain the effect of this tasks execution. Rolling back this task may not be achievable without other action."""
@@ -59,9 +58,6 @@ class TransferL1Data(TransferDataBase, GlobusMixin):
59
58
  # Movie needs to be transferred separately as the movie headers need to go with it
60
59
  self.transfer_movie()
61
60
 
62
- with self.telemetry_span("Upload quality data"):
63
- self.transfer_quality_data()
64
-
65
61
  with self.telemetry_span("Upload science frames"):
66
62
  self.transfer_output_frames()
67
63
 
@@ -105,33 +101,6 @@ class TransferL1Data(TransferDataBase, GlobusMixin):
105
101
  content_type="video/mp4",
106
102
  )
107
103
 
108
- def transfer_quality_data(self):
109
- """Transfer quality data to the object store."""
110
- paths = list(self.read(tags=[Tag.output(), Tag.quality_data()]))
111
- if len(paths) == 0:
112
- logger.warning(
113
- f"No quality data found to upload for dataset. recipe_run_id={self.recipe_run_id}"
114
- )
115
- return
116
-
117
- if count := len(paths) > 1:
118
- # dataset inventory does not support multiple quality data object keys
119
- raise RuntimeError(
120
- f"Found multiple quality data files to upload. Not supported."
121
- f"{count=}, recipe_run_id={self.recipe_run_id}"
122
- )
123
-
124
- with self.telemetry_span(f"Uploading the trial quality data"):
125
- path = paths[0]
126
- logger.info(f"Uploading quality data: recipe_run_id={self.recipe_run_id}, {path=}")
127
- quality_data_object_key = self.format_object_key(path)
128
- self.object_store_upload_quality_data(
129
- quality_data=path,
130
- bucket=self.destination_bucket,
131
- object_key=quality_data_object_key,
132
- content_type="application/json",
133
- )
134
-
135
104
 
136
105
  class AssembleQualityData(L1OutputDataBase, QualityMixin):
137
106
  """
@@ -159,7 +128,7 @@ class AssembleQualityData(L1OutputDataBase, QualityMixin):
159
128
  ):
160
129
  self.write(
161
130
  quality_data,
162
- tags=[Tag.output(), Tag.quality_data()],
131
+ tags=Tag.quality_data(),
163
132
  encoder=quality_data_encoder,
164
133
  relative_path=f"{self.constants.dataset_id}_quality_data.json",
165
134
  )
@@ -177,10 +146,20 @@ class SubmitDatasetMetadata(L1OutputDataBase):
177
146
 
178
147
  def run(self) -> None:
179
148
  """Run method for this task."""
149
+ with self.telemetry_span(f"Storing quality data to metadata store"):
150
+ # each quality_data file is a list - this will combine the elements of multiple lists into a single list
151
+ quality_data = list(
152
+ chain.from_iterable(
153
+ self.read(tags=Tag.quality_data(), decoder=quality_data_decoder)
154
+ )
155
+ )
156
+ self.metadata_store_add_quality_data(
157
+ dataset_id=self.constants.dataset_id, quality_data=quality_data
158
+ )
180
159
  with self.telemetry_span("Count Expected Outputs"):
181
160
  dataset_id = self.constants.dataset_id
182
161
  expected_object_count = self.count(tags=Tag.output())
183
- if self.dataset_has_quality_data:
162
+ if quality_data:
184
163
  expected_object_count += 1
185
164
  logger.info(
186
165
  f"Adding Dataset Receipt Account: "
@@ -251,12 +230,24 @@ class PublishCatalogAndQualityMessages(L1OutputDataBase, InterserviceBusMixin):
251
230
  messages = [CatalogObjectMessage(body=body) for body in message_bodies]
252
231
  return messages
253
232
 
233
+ @property
234
+ def quality_report_message(self) -> CreateQualityReportMessage:
235
+ """Create the Quality Report Message."""
236
+ file_name = Path(f"{self.constants.dataset_id}_quality_report.pdf")
237
+ body = CreateQualityReportMessageBody(
238
+ bucket=self.destination_bucket,
239
+ objectName=self.format_object_key(file_name),
240
+ conversationId=str(self.recipe_run_id),
241
+ datasetId=self.constants.dataset_id,
242
+ incrementDatasetCatalogReceiptCount=True,
243
+ )
244
+ return CreateQualityReportMessage(body=body)
245
+
254
246
  def run(self) -> None:
255
247
  """Run method for this task."""
256
248
  with self.telemetry_span("Gather output data"):
257
249
  frames = self.read(tags=self.output_frame_tags)
258
250
  movies = self.read(tags=[Tag.output(), Tag.movie()])
259
- quality_data = self.read(tags=[Tag.output(), Tag.quality_data()])
260
251
  with self.telemetry_span("Create message objects"):
261
252
  messages = []
262
253
  messages += self.frame_messages(paths=frames)
@@ -265,7 +256,7 @@ class PublishCatalogAndQualityMessages(L1OutputDataBase, InterserviceBusMixin):
265
256
  object_message_count = len(messages) - frame_message_count
266
257
  dataset_has_quality_data = self.dataset_has_quality_data
267
258
  if dataset_has_quality_data:
268
- messages += self.object_messages(paths=quality_data, object_type="QDATA")
259
+ messages.append(self.quality_report_message)
269
260
  with self.telemetry_span(
270
261
  f"Publish messages: {frame_message_count = }, {object_message_count = }, {dataset_has_quality_data = }"
271
262
  ):
@@ -9,6 +9,8 @@ from globus_sdk import ConfidentialAppAuthClient
9
9
  from globus_sdk import GlobusError
10
10
  from globus_sdk import TransferClient
11
11
  from globus_sdk import TransferData
12
+ from globus_sdk.scopes import TransferScopes
13
+ from globus_sdk.transport import RetryConfig
12
14
 
13
15
  from dkist_processing_common.config import common_configurations
14
16
 
@@ -31,27 +33,32 @@ class GlobusTransferItem:
31
33
  class GlobusMixin:
32
34
  """Mixin to add methods to a Task to support globus transfers."""
33
35
 
34
- @property
35
- def globus_transfer_client(self) -> TransferClient:
36
- """Get the globus transfer client, creating it if it doesn't exist."""
37
- if getattr(self, "_globus_transfer_client", False):
38
- return self._globus_transfer_client
36
+ def globus_transfer_client_factory(self, transfer_data: TransferData) -> TransferClient:
37
+ """Create a globus transfer client based on the direction of transfer and round-robin the available application credentials."""
38
+ if (
39
+ transfer_data["source_endpoint"] == common_configurations.object_store_endpoint
40
+ ): # inbound
41
+ client_credentials = common_configurations.globus_inbound_client_credentials
42
+ else: # outbound
43
+ client_credentials = common_configurations.globus_outbound_client_credentials
44
+
45
+ # Round-robin the client credentials based on the recipe run id
46
+ index = self.recipe_run_id % len(client_credentials)
47
+ selected_credential = client_credentials[index]
48
+
39
49
  confidential_client = ConfidentialAppAuthClient(
40
- client_id=common_configurations.globus_client_id,
41
- client_secret=common_configurations.globus_client_secret,
42
- transport_params=common_configurations.globus_transport_params,
50
+ client_id=selected_credential.client_id,
51
+ client_secret=selected_credential.client_secret,
43
52
  )
44
- authorizer = ClientCredentialsAuthorizer(
45
- confidential_client, scopes="urn:globus:auth:scope:transfer.api.globus.org:all"
46
- )
47
- self._globus_transfer_client = TransferClient(authorizer=authorizer)
48
- return self._globus_transfer_client
53
+ authorizer = ClientCredentialsAuthorizer(confidential_client, scopes=TransferScopes)
54
+ retry_config = RetryConfig(max_retries=common_configurations.globus_max_retries)
55
+
56
+ return TransferClient(authorizer=authorizer, retry_config=retry_config)
49
57
 
50
58
  def globus_transfer_scratch_to_object_store(
51
59
  self,
52
60
  transfer_items: list[GlobusTransferItem],
53
61
  label: str = None,
54
- sync_level: str = None,
55
62
  verify_checksum: bool = True,
56
63
  ) -> None:
57
64
  """Transfer data from scratch to the object store."""
@@ -60,7 +67,6 @@ class GlobusMixin:
60
67
  destination_endpoint=common_configurations.object_store_endpoint,
61
68
  transfer_items=transfer_items,
62
69
  label=label,
63
- sync_level=sync_level,
64
70
  verify_checksum=verify_checksum,
65
71
  )
66
72
 
@@ -68,7 +74,6 @@ class GlobusMixin:
68
74
  self,
69
75
  transfer_items: list[GlobusTransferItem],
70
76
  label: str = None,
71
- sync_level: str = None,
72
77
  verify_checksum: bool = True,
73
78
  ) -> None:
74
79
  """Transfer data from the object store to scratch."""
@@ -77,7 +82,6 @@ class GlobusMixin:
77
82
  destination_endpoint=common_configurations.scratch_endpoint,
78
83
  transfer_items=transfer_items,
79
84
  label=label,
80
- sync_level=sync_level,
81
85
  verify_checksum=verify_checksum,
82
86
  )
83
87
 
@@ -87,7 +91,6 @@ class GlobusMixin:
87
91
  destination_endpoint: str,
88
92
  transfer_items: list[GlobusTransferItem],
89
93
  label: str = None,
90
- sync_level: str = None,
91
94
  verify_checksum: bool = True,
92
95
  ) -> TransferData:
93
96
  """Format a globus TransferData instance."""
@@ -95,7 +98,6 @@ class GlobusMixin:
95
98
  source_endpoint=source_endpoint,
96
99
  destination_endpoint=destination_endpoint,
97
100
  label=label,
98
- sync_level=sync_level,
99
101
  verify_checksum=verify_checksum,
100
102
  )
101
103
  for item in transfer_items:
@@ -112,7 +114,6 @@ class GlobusMixin:
112
114
  destination_endpoint: str,
113
115
  transfer_items: list[GlobusTransferItem],
114
116
  label: str = None,
115
- sync_level: str = None,
116
117
  verify_checksum: bool = True,
117
118
  ) -> None:
118
119
  """Perform a transfer of data using globus."""
@@ -121,7 +122,6 @@ class GlobusMixin:
121
122
  destination_endpoint=destination_endpoint,
122
123
  transfer_items=transfer_items,
123
124
  label=label,
124
- sync_level=sync_level,
125
125
  verify_checksum=verify_checksum,
126
126
  )
127
127
  self._blocking_globus_transfer(transfer_data=transfer_data)
@@ -131,24 +131,21 @@ class GlobusMixin:
131
131
  source_endpoint: str,
132
132
  destination_endpoint: str,
133
133
  label: str = None,
134
- sync_level: str = None,
135
134
  verify_checksum: bool = True,
136
135
  ) -> TransferData:
137
136
  label = label or "Data Processing Transfer"
138
137
  return TransferData(
139
- transfer_client=self.globus_transfer_client,
140
138
  source_endpoint=source_endpoint,
141
139
  destination_endpoint=destination_endpoint,
142
140
  label=label,
143
- sync_level=sync_level,
144
141
  verify_checksum=verify_checksum,
145
142
  )
146
143
 
147
144
  def _blocking_globus_transfer(self, transfer_data: TransferData) -> None:
148
- tc = self.globus_transfer_client
149
- logger.info(f"Starting globus transfer: label={transfer_data.get('label')}")
145
+ tc = self.globus_transfer_client_factory(transfer_data=transfer_data)
150
146
  transfer_result = tc.submit_transfer(transfer_data)
151
147
  task_id = transfer_result["task_id"]
148
+ logger.info(f"Starting globus transfer: label={transfer_data.get('label')}, {task_id=}, ")
152
149
  polling_interval = 60
153
150
  while not tc.task_wait(
154
151
  task_id=task_id, timeout=polling_interval, polling_interval=polling_interval
@@ -1,5 +1,6 @@
1
1
  """Mixin for a WorkflowDataTaskBase subclass which implements Metadata Store data access functionality."""
2
2
 
3
+ import json
3
4
  import logging
4
5
  from functools import cached_property
5
6
  from typing import Literal
@@ -7,11 +8,15 @@ from typing import Literal
7
8
  from pydantic import validate_call
8
9
 
9
10
  from dkist_processing_common._util.graphql import GraphQLClient
11
+ from dkist_processing_common.codecs.quality import QualityDataEncoder
10
12
  from dkist_processing_common.config import common_configurations
11
13
  from dkist_processing_common.models.graphql import DatasetCatalogReceiptAccountMutation
12
14
  from dkist_processing_common.models.graphql import DatasetCatalogReceiptAccountResponse
13
15
  from dkist_processing_common.models.graphql import InputDatasetPartResponse
14
16
  from dkist_processing_common.models.graphql import InputDatasetRecipeRunResponse
17
+ from dkist_processing_common.models.graphql import QualitiesRequest
18
+ from dkist_processing_common.models.graphql import QualityCreation
19
+ from dkist_processing_common.models.graphql import QualityResponse
15
20
  from dkist_processing_common.models.graphql import RecipeRunMutation
16
21
  from dkist_processing_common.models.graphql import RecipeRunMutationResponse
17
22
  from dkist_processing_common.models.graphql import RecipeRunProvenanceMutation
@@ -145,6 +150,50 @@ class MetadataStoreMixin:
145
150
  mutation_response_cls=RecipeRunProvenanceResponse,
146
151
  )
147
152
 
153
+ # QUALITY
154
+
155
+ def metadata_store_add_quality_data(self, dataset_id: str, quality_data: list[dict]):
156
+ """Add the quality data to the metadata-store."""
157
+ if self.metadata_store_quality_data_exists(dataset_id):
158
+ raise RuntimeError(f"Quality data already persisted for dataset {dataset_id!r}")
159
+ for metric in quality_data:
160
+ if (metric_code := metric.get("metric_code")) is None:
161
+ name = metric.get("name")
162
+ raise ValueError(f"No metric_code for {name!r} in dataset {dataset_id!r}")
163
+ params = QualityCreation(
164
+ datasetId=dataset_id,
165
+ metricCode=metric_code,
166
+ facet=metric.get("facet"),
167
+ name=metric.get("name"),
168
+ description=metric.get("description"),
169
+ statement=metric.get("statement"),
170
+ # JSON array
171
+ warnings=json.dumps(metric.get("warnings")),
172
+ # JSON objects
173
+ plotData=json.dumps(metric.get("plot_data"), cls=QualityDataEncoder),
174
+ multiPlotData=json.dumps(metric.get("multi_plot_data"), cls=QualityDataEncoder),
175
+ tableData=json.dumps(metric.get("table_data"), cls=QualityDataEncoder),
176
+ histogramData=json.dumps(metric.get("histogram_data"), cls=QualityDataEncoder),
177
+ modmatData=json.dumps(metric.get("modmat_data"), cls=QualityDataEncoder),
178
+ raincloudData=json.dumps(metric.get("raincloud_data"), cls=QualityDataEncoder),
179
+ efficiencyData=json.dumps(metric.get("efficiency_data"), cls=QualityDataEncoder),
180
+ )
181
+ self.metadata_store_client.execute_gql_mutation(
182
+ mutation_base="createQuality",
183
+ mutation_parameters=params,
184
+ mutation_response_cls=QualityResponse,
185
+ )
186
+
187
+ def metadata_store_quality_data_exists(self, dataset_id: str) -> bool:
188
+ """Return True if quality data exists in the metadata-store for the given dataset id."""
189
+ params = QualitiesRequest(datasetId=dataset_id)
190
+ response = self.metadata_store_client.execute_gql_query(
191
+ query_base="qualities",
192
+ query_response_cls=QualityResponse,
193
+ query_parameters=params,
194
+ )
195
+ return bool(response)
196
+
148
197
  # INPUT DATASET RECIPE RUN
149
198
 
150
199
  @cached_property
@@ -55,27 +55,6 @@ class ObjectStoreMixin:
55
55
  },
56
56
  )
57
57
 
58
- def object_store_upload_quality_data(
59
- self,
60
- quality_data: Path | bytes,
61
- bucket: str,
62
- object_key: str,
63
- content_type: str = "application/json",
64
- ):
65
- """Upload quality data to the object store."""
66
- self.object_store_client.upload_object(
67
- object_data=quality_data,
68
- bucket=bucket,
69
- object_key=object_key,
70
- verify_checksum=True,
71
- content_type=content_type,
72
- metadata={
73
- "groupname": "DATASET",
74
- "groupid": self.constants.dataset_id,
75
- "objecttype": "QDATA",
76
- },
77
- )
78
-
79
58
  def object_store_remove_folder_objects(self, bucket: str, path: Path | str) -> list[str]:
80
59
  """
81
60
  Remove folder objects (end with /) in the specified bucket and path.
@@ -1356,15 +1356,13 @@ class _WavecalQualityMixin:
1356
1356
  Note that the residuals are the *unweighed* residuals.
1357
1357
  """
1358
1358
  weight_data = np.ones(input_wavelength.size) if weights is None else weights
1359
- prepared_weights = np.sqrt(weight_data / np.sum(weight_data))
1359
+ prepared_weights = fit_result.prepared_weights
1360
1360
  residuals = fit_result.minimizer_result.residual / prepared_weights
1361
1361
  residuals[~np.isfinite(residuals)] = 0.0
1362
- best_fit_atlas = input_spectrum - residuals
1363
1362
  normalized_residuals = residuals / input_spectrum
1364
1363
 
1365
- best_fit_header = fit_result.wavelength_parameters.to_header(axis_num=1)
1366
- wcs = WCS(best_fit_header)
1367
- best_fit_wavelength = wcs.spectral.pixel_to_world(np.arange(input_spectrum.size))
1364
+ best_fit_atlas = fit_result.best_fit_atlas
1365
+ best_fit_wavelength = fit_result.best_fit_wavelength_vector
1368
1366
 
1369
1367
  finite_idx = (
1370
1368
  np.isfinite(input_wavelength)
@@ -1378,7 +1376,7 @@ class _WavecalQualityMixin:
1378
1376
  data = {
1379
1377
  "input_wavelength_nm": input_wavelength.to_value(u.nm)[finite_idx].tolist(),
1380
1378
  "input_spectrum": input_spectrum[finite_idx].tolist(),
1381
- "best_fit_wavelength_nm": best_fit_wavelength.to_value(u.nm)[finite_idx].tolist(),
1379
+ "best_fit_wavelength_nm": best_fit_wavelength[finite_idx].tolist(),
1382
1380
  "best_fit_atlas": best_fit_atlas[finite_idx].tolist(),
1383
1381
  "normalized_residuals": normalized_residuals[finite_idx].tolist(),
1384
1382
  "weights": None if weights is None else weight_data[finite_idx].tolist(),
@@ -9,8 +9,6 @@ from typing import Any
9
9
  from typing import Generator
10
10
  from uuid import uuid4
11
11
 
12
- from dkist_inventory.inventory import generate_quality_report_filename
13
-
14
12
  from dkist_processing_common.codecs.asdf import asdf_fileobj_encoder
15
13
  from dkist_processing_common.codecs.basemodel import basemodel_decoder
16
14
  from dkist_processing_common.codecs.fits import fits_access_decoder
@@ -247,7 +245,5 @@ class CreateTrialQualityReport(OutputDataBase):
247
245
  self.write(
248
246
  quality_report,
249
247
  tags=[Tag.output(), Tag.quality_report()],
250
- relative_path=generate_quality_report_filename(
251
- dataset_id=self.constants.dataset_id
252
- ),
248
+ relative_path=f"{self.constants.dataset_id}_quality_report.pdf",
253
249
  )
@@ -100,7 +100,7 @@ class TransferTrialData(TransferDataBase, GlobusMixin):
100
100
  tag_list = []
101
101
  tag_list += [[Tag.output(), Tag.dataset_inventory()]]
102
102
  tag_list += [[Tag.output(), Tag.asdf()]]
103
- tag_list += [[Tag.output(), Tag.quality_data()]]
103
+ tag_list += [[Tag.quality_data()]] # quality data is not tagged as OUTPUT
104
104
  tag_list += [[Tag.output(), Tag.quality_report()]]
105
105
  tag_list += [[Tag.output(), Tag.movie()]]
106
106
  return tag_list
@@ -135,6 +135,10 @@ class InputDatasetRecipeRunResponseMapping(ResponseMapping):
135
135
  return Unset
136
136
 
137
137
 
138
+ class QualityResponseMapping(ResponseMapping):
139
+ pass # TODO
140
+
141
+
138
142
  def make_default_recipe_run_status_response() -> RecipeRunStatusResponse:
139
143
  return RecipeRunStatusResponse(recipeRunStatusId=1)
140
144