orchestrator-core 4.3.0rc2__py3-none-any.whl → 4.4.0rc1__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
orchestrator/__init__.py CHANGED
@@ -13,7 +13,7 @@
13
13
 
14
14
  """This is the orchestrator workflow engine."""
15
15
 
16
- __version__ = "4.3.0rc2"
16
+ __version__ = "4.4.0rc1"
17
17
 
18
18
  from orchestrator.app import OrchestratorCore
19
19
  from orchestrator.settings import app_settings
@@ -25,7 +25,7 @@ from fastapi.param_functions import Body, Depends, Header
25
25
  from fastapi.routing import APIRouter
26
26
  from fastapi.websockets import WebSocket
27
27
  from fastapi_etag.dependency import CacheHit
28
- from more_itertools import chunked, last
28
+ from more_itertools import chunked, first, last
29
29
  from sentry_sdk.tracing import trace
30
30
  from sqlalchemy import CompoundSelect, Select, select
31
31
  from sqlalchemy.orm import defer, joinedload
@@ -88,11 +88,17 @@ def check_global_lock() -> None:
88
88
  )
89
89
 
90
90
 
91
- def get_current_steps(pstat: ProcessStat) -> StepList:
92
- """Extract past and current steps from the ProcessStat."""
93
- remaining_steps = pstat.log
91
+ def get_steps_to_evaluate_for_rbac(pstat: ProcessStat) -> StepList:
92
+ """Extract all steps from the ProcessStat for a process that should be evaluated for a RBAC callback.
93
+
94
+ For a suspended process this includes all previously completed steps as well as the current step.
95
+ For a completed process this includes all steps.
96
+ """
97
+ if not (remaining_steps := pstat.log):
98
+ return pstat.workflow.steps
99
+
94
100
  past_steps = pstat.workflow.steps[: -len(remaining_steps)]
95
- return StepList(past_steps + [pstat.log[0]])
101
+ return StepList(past_steps >> first(remaining_steps))
96
102
 
97
103
 
98
104
  def get_auth_callbacks(steps: StepList, workflow: Workflow) -> tuple[Authorizer | None, Authorizer | None]:
@@ -200,7 +206,7 @@ def resume_process_endpoint(
200
206
  raise_status(HTTPStatus.CONFLICT, f"Resuming a {process.last_status.lower()} workflow is not possible")
201
207
 
202
208
  pstat = load_process(process)
203
- auth_resume, auth_retry = get_auth_callbacks(get_current_steps(pstat), pstat.workflow)
209
+ auth_resume, auth_retry = get_auth_callbacks(get_steps_to_evaluate_for_rbac(pstat), pstat.workflow)
204
210
  if process.last_status == ProcessStatus.SUSPENDED:
205
211
  if auth_resume is not None and not auth_resume(user_model):
206
212
  raise_status(HTTPStatus.FORBIDDEN, "User is not authorized to resume step")
@@ -13,12 +13,11 @@
13
13
 
14
14
 
15
15
  import logging
16
- from time import sleep
17
16
 
18
- import schedule
19
17
  import typer
18
+ from apscheduler.schedulers.blocking import BlockingScheduler
20
19
 
21
- from orchestrator.schedules import ALL_SCHEDULERS
20
+ from orchestrator.schedules.scheduler import jobstores, scheduler, scheduler_dispose_db_connections
22
21
 
23
22
  log = logging.getLogger(__name__)
24
23
 
@@ -27,36 +26,47 @@ app: typer.Typer = typer.Typer()
27
26
 
28
27
  @app.command()
29
28
  def run() -> None:
30
- """Loop eternally and run schedulers at configured times."""
31
- for s in ALL_SCHEDULERS:
32
- job = getattr(schedule.every(s.period), s.time_unit)
33
- if s.at:
34
- job = job.at(s.at)
35
- job.do(s).tag(s.name)
36
- log.info("Starting Schedule")
37
- for j in schedule.jobs:
38
- log.info("%s: %s", ", ".join(j.tags), j)
39
- while True:
40
- schedule.run_pending()
41
- idle = schedule.idle_seconds()
42
- if idle < 0:
43
- log.info("Next job in queue is scheduled in the past, run it now.")
44
- else:
45
- log.info("Sleeping for %d seconds", idle)
46
- sleep(idle)
29
+ """Start scheduler and loop eternally to keep thread alive."""
30
+ blocking_scheduler = BlockingScheduler(jobstores=jobstores)
31
+
32
+ try:
33
+ blocking_scheduler.start()
34
+ except (KeyboardInterrupt, SystemExit):
35
+ scheduler.shutdown()
36
+ scheduler_dispose_db_connections()
47
37
 
48
38
 
49
39
  @app.command()
50
40
  def show_schedule() -> None:
51
- """Show the currently configured schedule."""
52
- for s in ALL_SCHEDULERS:
53
- at_str = f"@ {s.at} " if s.at else ""
54
- typer.echo(f"{s.name}: {s.__name__} {at_str}every {s.period} {s.time_unit}")
41
+ """Show the currently configured schedule.
42
+
43
+ in cli underscore is replaced by a dash `show-schedule`
44
+ """
45
+ scheduler.start(paused=True) # paused: avoid triggering jobs during CLI
46
+ jobs = scheduler.get_jobs()
47
+ scheduler.shutdown(wait=False)
48
+ scheduler_dispose_db_connections()
49
+
50
+ for job in jobs:
51
+ typer.echo(f"[{job.id}] Next run: {job.next_run_time} | Trigger: {job.trigger}")
55
52
 
56
53
 
57
54
  @app.command()
58
- def force(keyword: str) -> None:
59
- """Force the execution of (a) scheduler(s) based on a keyword."""
60
- for s in ALL_SCHEDULERS:
61
- if keyword in s.name or keyword in s.__name__:
62
- s()
55
+ def force(job_id: str) -> None:
56
+ """Force the execution of (a) scheduler(s) based on a job_id."""
57
+ scheduler.start(paused=True) # paused: avoid triggering jobs during CLI
58
+ job = scheduler.get_job(job_id)
59
+ scheduler.shutdown(wait=False)
60
+ scheduler_dispose_db_connections()
61
+
62
+ if not job:
63
+ typer.echo(f"Job '{job_id}' not found.")
64
+ raise typer.Exit(code=1)
65
+
66
+ typer.echo(f"Running job [{job.id}] now...")
67
+ try:
68
+ job.func(*job.args or (), **job.kwargs or {})
69
+ typer.echo("Job executed successfully.")
70
+ except Exception as e:
71
+ typer.echo(f"Job execution failed: {e}")
72
+ raise typer.Exit(code=1)
@@ -0,0 +1,36 @@
1
+ import structlog
2
+
3
+ from orchestrator.db.filters import Filter
4
+ from orchestrator.db.sorting import Sort
5
+ from orchestrator.graphql.pagination import Connection
6
+ from orchestrator.graphql.schemas.scheduled_task import ScheduledTaskGraphql
7
+ from orchestrator.graphql.types import GraphqlFilter, GraphqlSort, OrchestratorInfo
8
+ from orchestrator.graphql.utils import create_resolver_error_handler, to_graphql_result_page
9
+ from orchestrator.graphql.utils.is_query_detailed import is_querying_page_data
10
+ from orchestrator.schedules.scheduler import get_scheduler_tasks, scheduled_task_filter_keys, scheduled_task_sort_keys
11
+
12
+ logger = structlog.get_logger(__name__)
13
+
14
+
15
+ async def resolve_scheduled_tasks(
16
+ info: OrchestratorInfo,
17
+ filter_by: list[GraphqlFilter] | None = None,
18
+ sort_by: list[GraphqlSort] | None = None,
19
+ first: int = 10,
20
+ after: int = 0,
21
+ ) -> Connection[ScheduledTaskGraphql]:
22
+ _error_handler = create_resolver_error_handler(info)
23
+
24
+ pydantic_filter_by: list[Filter] = [item.to_pydantic() for item in filter_by] if filter_by else []
25
+ pydantic_sort_by: list[Sort] = [item.to_pydantic() for item in sort_by] if sort_by else []
26
+ scheduled_tasks, total = get_scheduler_tasks(
27
+ first=first, after=after, filter_by=pydantic_filter_by, sort_by=pydantic_sort_by, error_handler=_error_handler
28
+ )
29
+
30
+ graphql_scheduled_tasks = []
31
+ if is_querying_page_data(info):
32
+ graphql_scheduled_tasks = [ScheduledTaskGraphql.from_pydantic(p) for p in scheduled_tasks]
33
+
34
+ return to_graphql_result_page(
35
+ graphql_scheduled_tasks, first, after, total, scheduled_task_filter_keys, scheduled_task_sort_keys
36
+ )
@@ -51,12 +51,14 @@ from orchestrator.graphql.resolvers import (
51
51
  resolve_version,
52
52
  resolve_workflows,
53
53
  )
54
+ from orchestrator.graphql.resolvers.scheduled_tasks import resolve_scheduled_tasks
54
55
  from orchestrator.graphql.schemas import DEFAULT_GRAPHQL_MODELS
55
56
  from orchestrator.graphql.schemas.customer import CustomerType
56
57
  from orchestrator.graphql.schemas.process import ProcessType
57
58
  from orchestrator.graphql.schemas.product import ProductType
58
59
  from orchestrator.graphql.schemas.product_block import ProductBlock
59
60
  from orchestrator.graphql.schemas.resource_type import ResourceType
61
+ from orchestrator.graphql.schemas.scheduled_task import ScheduledTaskGraphql
60
62
  from orchestrator.graphql.schemas.settings import StatusType
61
63
  from orchestrator.graphql.schemas.subscription import SubscriptionInterface
62
64
  from orchestrator.graphql.schemas.version import VersionType
@@ -99,6 +101,9 @@ class OrchestratorQuery:
99
101
  description="Returns information about cache, workers, and global engine settings",
100
102
  )
101
103
  version: VersionType = authenticated_field(resolver=resolve_version, description="Returns version information")
104
+ scheduled_tasks: Connection[ScheduledTaskGraphql] = authenticated_field(
105
+ resolver=resolve_scheduled_tasks, description="Returns scheduled job information"
106
+ )
102
107
 
103
108
 
104
109
  @strawberry.federation.type(description="Orchestrator customer Query")
@@ -6,7 +6,7 @@ from strawberry.federation.schema_directives import Key
6
6
  from strawberry.scalars import JSON
7
7
 
8
8
  from oauth2_lib.strawberry import authenticated_field
9
- from orchestrator.api.api_v1.endpoints.processes import get_auth_callbacks, get_current_steps
9
+ from orchestrator.api.api_v1.endpoints.processes import get_auth_callbacks, get_steps_to_evaluate_for_rbac
10
10
  from orchestrator.db import ProcessTable, ProductTable, db
11
11
  from orchestrator.graphql.pagination import EMPTY_PAGE, Connection
12
12
  from orchestrator.graphql.schemas.customer import CustomerType
@@ -86,7 +86,7 @@ class ProcessType:
86
86
  oidc_user = info.context.get_current_user
87
87
  workflow = get_workflow(self.workflow_name)
88
88
  process = load_process(db.session.get(ProcessTable, self.process_id)) # type: ignore[arg-type]
89
- auth_resume, auth_retry = get_auth_callbacks(get_current_steps(process), workflow) # type: ignore[arg-type]
89
+ auth_resume, auth_retry = get_auth_callbacks(get_steps_to_evaluate_for_rbac(process), workflow) # type: ignore[arg-type]
90
90
 
91
91
  return FormUserPermissionsType(
92
92
  retryAllowed=auth_retry and auth_retry(oidc_user), # type: ignore[arg-type]
@@ -0,0 +1,8 @@
1
+ import strawberry
2
+
3
+ from orchestrator.schedules.scheduler import ScheduledTask
4
+
5
+
6
+ @strawberry.experimental.pydantic.type(model=ScheduledTask, all_fields=True)
7
+ class ScheduledTaskGraphql:
8
+ pass
@@ -25,6 +25,6 @@ def _format_context(context: dict) -> str:
25
25
 
26
26
  def create_resolver_error_handler(info: OrchestratorInfo) -> CallableErrorHandler:
27
27
  def handle_error(message: str, **context) -> None: # type: ignore
28
- return register_error(" ".join([message, _format_context(context)]), info, error_type=ErrorType.BAD_REQUEST)
28
+ return register_error(f"{message} {_format_context(context)}", info, error_type=ErrorType.BAD_REQUEST)
29
29
 
30
30
  return handle_error
File without changes
@@ -13,12 +13,11 @@
13
13
 
14
14
 
15
15
  from orchestrator.schedules.resume_workflows import run_resume_workflows
16
- from orchestrator.schedules.scheduling import SchedulingFunction
17
16
  from orchestrator.schedules.task_vacuum import vacuum_tasks
18
17
  from orchestrator.schedules.validate_products import validate_products
19
18
  from orchestrator.schedules.validate_subscriptions import validate_subscriptions
20
19
 
21
- ALL_SCHEDULERS: list[SchedulingFunction] = [
20
+ ALL_SCHEDULERS: list = [
22
21
  run_resume_workflows,
23
22
  vacuum_tasks,
24
23
  validate_subscriptions,
@@ -12,10 +12,10 @@
12
12
  # limitations under the License.
13
13
 
14
14
 
15
- from orchestrator.schedules.scheduling import scheduler
15
+ from orchestrator.schedules.scheduler import scheduler
16
16
  from orchestrator.services.processes import start_process
17
17
 
18
18
 
19
- @scheduler(name="Resume workflows", time_unit="hour", period=1)
19
+ @scheduler.scheduled_job(id="resume-workflows", name="Resume workflows", trigger="interval", hours=1) # type: ignore[misc]
20
20
  def run_resume_workflows() -> None:
21
21
  start_process("task_resume_workflows")
@@ -0,0 +1,153 @@
1
+ # Copyright 2019-2020 SURF.
2
+ # Licensed under the Apache License, Version 2.0 (the "License");
3
+ # you may not use this file except in compliance with the License.
4
+ # You may obtain a copy of the License at
5
+ #
6
+ # http://www.apache.org/licenses/LICENSE-2.0
7
+ #
8
+ # Unless required by applicable law or agreed to in writing, software
9
+ # distributed under the License is distributed on an "AS IS" BASIS,
10
+ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
11
+ # See the License for the specific language governing permissions and
12
+ # limitations under the License.
13
+
14
+
15
+ from datetime import datetime
16
+ from typing import Any
17
+
18
+ from apscheduler.jobstores.sqlalchemy import SQLAlchemyJobStore
19
+ from apscheduler.schedulers.background import BackgroundScheduler
20
+ from more_itertools import partition
21
+ from pydantic import BaseModel
22
+
23
+ from orchestrator.db.filters import Filter
24
+ from orchestrator.db.filters.filters import CallableErrorHandler
25
+ from orchestrator.db.sorting import Sort
26
+ from orchestrator.db.sorting.sorting import SortOrder
27
+ from orchestrator.settings import app_settings
28
+ from orchestrator.utils.helpers import camel_to_snake, to_camel
29
+
30
+ jobstores = {"default": SQLAlchemyJobStore(url=str(app_settings.DATABASE_URI))}
31
+
32
+ scheduler = BackgroundScheduler(jobstores=jobstores)
33
+
34
+
35
+ def scheduler_dispose_db_connections() -> None:
36
+ jobstores["default"].engine.dispose()
37
+
38
+
39
+ class ScheduledTask(BaseModel):
40
+ id: str
41
+ name: str | None = None
42
+ next_run_time: datetime | None = None
43
+ trigger: str
44
+
45
+
46
+ scheduled_task_keys = set(ScheduledTask.model_fields.keys())
47
+ scheduled_task_filter_keys = sorted(scheduled_task_keys | {to_camel(key) for key in scheduled_task_keys})
48
+ scheduled_task_sort_keys = scheduled_task_filter_keys
49
+
50
+
51
+ def scheduled_task_in_filter(job: ScheduledTask, filter_by: list[Filter]) -> bool:
52
+ return any(f.value.lower() in getattr(job, camel_to_snake(f.field), "").lower() for f in filter_by)
53
+
54
+
55
+ def filter_scheduled_tasks(
56
+ scheduled_tasks: list[ScheduledTask],
57
+ handle_filter_error: CallableErrorHandler,
58
+ filter_by: list[Filter] | None = None,
59
+ ) -> list[ScheduledTask]:
60
+ if not filter_by:
61
+ return scheduled_tasks
62
+
63
+ try:
64
+ invalid_filters, valid_filters = partition(lambda x: x.field in scheduled_task_filter_keys, filter_by)
65
+
66
+ if invalid_list := [item.field for item in invalid_filters]:
67
+ handle_filter_error(
68
+ "Invalid filter arguments", invalid_filters=invalid_list, valid_filter_keys=scheduled_task_filter_keys
69
+ )
70
+
71
+ valid_filter_list = list(valid_filters)
72
+ return [task for task in scheduled_tasks if scheduled_task_in_filter(task, valid_filter_list)]
73
+ except Exception as e:
74
+ handle_filter_error(str(e))
75
+ return []
76
+
77
+
78
+ def _invert(value: Any) -> Any:
79
+ """Invert value for descending order."""
80
+ if isinstance(value, (int, float)):
81
+ return -value
82
+ if isinstance(value, str):
83
+ return tuple(-ord(c) for c in value)
84
+ if isinstance(value, datetime):
85
+ return -value.timestamp()
86
+ return value
87
+
88
+
89
+ def sort_key(sort_field: str, sort_order: SortOrder) -> Any:
90
+ def _sort_key(task: Any) -> Any:
91
+ value = getattr(task, camel_to_snake(sort_field), None)
92
+ if sort_field == "next_run_time" and value is None:
93
+ return float("inf") if sort_order == SortOrder.ASC else float("-inf")
94
+ return value if sort_order == SortOrder.ASC else _invert(value)
95
+
96
+ return _sort_key
97
+
98
+
99
+ def sort_scheduled_tasks(
100
+ scheduled_tasks: list[ScheduledTask], handle_sort_error: CallableErrorHandler, sort_by: list[Sort] | None = None
101
+ ) -> list[ScheduledTask]:
102
+ if not sort_by:
103
+ return scheduled_tasks
104
+
105
+ try:
106
+ invalid_sorting, valid_sorting = partition(lambda x: x.field in scheduled_task_sort_keys, sort_by)
107
+ if invalid_list := [item.field for item in invalid_sorting]:
108
+ handle_sort_error(
109
+ "Invalid sort arguments", invalid_sorting=invalid_list, valid_sort_keys=scheduled_task_sort_keys
110
+ )
111
+
112
+ valid_sort_list = list(valid_sorting)
113
+ return sorted(
114
+ scheduled_tasks, key=lambda task: tuple(sort_key(sort.field, sort.order)(task) for sort in valid_sort_list)
115
+ )
116
+ except Exception as e:
117
+ handle_sort_error(str(e))
118
+ return []
119
+
120
+
121
+ def default_error_handler(message: str, **context) -> None: # type: ignore
122
+ from orchestrator.graphql.utils.create_resolver_error_handler import _format_context
123
+
124
+ raise ValueError(f"{message} {_format_context(context)}")
125
+
126
+
127
+ def get_scheduler_tasks(
128
+ first: int = 10,
129
+ after: int = 0,
130
+ filter_by: list[Filter] | None = None,
131
+ sort_by: list[Sort] | None = None,
132
+ error_handler: CallableErrorHandler = default_error_handler,
133
+ ) -> tuple[list[ScheduledTask], int]:
134
+ scheduler.start(paused=True)
135
+ scheduled_tasks = scheduler.get_jobs()
136
+ scheduler.shutdown(wait=False)
137
+ scheduler_dispose_db_connections()
138
+
139
+ scheduled_tasks = filter_scheduled_tasks(scheduled_tasks, error_handler, filter_by)
140
+ scheduled_tasks = sort_scheduled_tasks(scheduled_tasks, error_handler, sort_by)
141
+
142
+ total = len(scheduled_tasks)
143
+ paginated_tasks = scheduled_tasks[after : after + first + 1]
144
+
145
+ return [
146
+ ScheduledTask(
147
+ id=task.id,
148
+ name=task.name,
149
+ next_run_time=task.next_run_time,
150
+ trigger=str(task.trigger),
151
+ )
152
+ for task in paginated_tasks
153
+ ], total
@@ -12,37 +12,77 @@
12
12
  # limitations under the License.
13
13
 
14
14
  from collections.abc import Callable
15
- from typing import Protocol, cast
15
+ from typing import TypeVar
16
16
 
17
- from schedule import CancelJob
17
+ from apscheduler.schedulers.base import BaseScheduler
18
+ from deprecated import deprecated
18
19
 
20
+ from orchestrator.schedules.scheduler import scheduler as default_scheduler # your global scheduler instance
19
21
 
20
- class SchedulingFunction(Protocol):
21
- __name__: str
22
- name: str
23
- time_unit: str
24
- period: int | None
25
- at: str | None
26
-
27
- def __call__(self) -> CancelJob | None: ...
22
+ F = TypeVar("F", bound=Callable[..., object])
28
23
 
29
24
 
25
+ @deprecated(
26
+ reason="We changed from scheduler to apscheduler which has its own decoractor, use `@scheduler.scheduled_job()` from `from orchestrator.scheduling.scheduler import scheduler`"
27
+ )
30
28
  def scheduler(
31
- name: str, time_unit: str, period: int = 1, at: str | None = None
32
- ) -> Callable[[Callable[[], CancelJob | None]], SchedulingFunction]:
33
- """Create schedule.
29
+ name: str,
30
+ time_unit: str,
31
+ period: int = 1,
32
+ at: str | None = None,
33
+ *,
34
+ id: str | None = None,
35
+ scheduler: BaseScheduler = default_scheduler,
36
+ ) -> Callable[[F], F]:
37
+ """APScheduler-compatible decorator to schedule a function.
38
+
39
+ id is necessary with apscheduler, if left empty it takes the function name.
34
40
 
35
- Either specify the period or the at. Examples:
36
- time_unit = "hours", period = 12 -> will run every 12 hours
37
- time_unit = "day", at="01:00" -> will run every day at 1 o'clock
41
+ - `time_unit = "hours", period = 12` → every 12 hours
42
+ - `time_unit = "day", at = "01:00"` → every day at 1 AM
38
43
  """
39
44
 
40
- def _scheduler(f: Callable[[], CancelJob | None]) -> SchedulingFunction:
41
- schedule = cast(SchedulingFunction, f)
42
- schedule.name = name
43
- schedule.time_unit = time_unit
44
- schedule.period = period
45
- schedule.at = at
46
- return schedule
45
+ def decorator(func: F) -> F:
46
+ job_id = id or func.__name__
47
+
48
+ trigger = "interval"
49
+ kwargs: dict[str, int] = {}
50
+ if time_unit == "day" and at:
51
+ trigger = "cron"
52
+ try:
53
+ hour, minute = map(int, at.split(":"))
54
+ except ValueError:
55
+ raise ValueError(f"Invalid time format for 'at': {at}, expected 'HH:MM'")
56
+
57
+ kwargs = {
58
+ "hour": hour,
59
+ "minute": minute,
60
+ }
61
+ else:
62
+ # Map string units to timedelta kwargs for IntervalTrigger
63
+ unit_map = {
64
+ "seconds": "seconds",
65
+ "second": "seconds",
66
+ "minutes": "minutes",
67
+ "minute": "minutes",
68
+ "hours": "hours",
69
+ "hour": "hours",
70
+ "days": "days",
71
+ "day": "days",
72
+ }
73
+
74
+ interval_arg = unit_map.get(time_unit.lower(), time_unit.lower())
75
+ kwargs = {interval_arg: period}
76
+
77
+ scheduler.add_job(
78
+ func,
79
+ trigger=trigger,
80
+ id=job_id,
81
+ name=name,
82
+ replace_existing=True,
83
+ **kwargs,
84
+ )
85
+
86
+ return func
47
87
 
48
- return _scheduler
88
+ return decorator
@@ -12,10 +12,10 @@
12
12
  # limitations under the License.
13
13
 
14
14
 
15
- from orchestrator.schedules.scheduling import scheduler
15
+ from orchestrator.schedules.scheduler import scheduler
16
16
  from orchestrator.services.processes import start_process
17
17
 
18
18
 
19
- @scheduler(name="Clean up tasks", time_unit="hours", period=6)
19
+ @scheduler.scheduled_job(id="clean-tasks", name="Clean up tasks", trigger="interval", hours=6) # type: ignore[misc]
20
20
  def vacuum_tasks() -> None:
21
21
  start_process("task_clean_up_tasks")
@@ -14,11 +14,17 @@ from sqlalchemy import func, select
14
14
 
15
15
  from orchestrator.db import db
16
16
  from orchestrator.db.models import ProcessTable
17
- from orchestrator.schedules.scheduling import scheduler
17
+ from orchestrator.schedules.scheduler import scheduler
18
18
  from orchestrator.services.processes import start_process
19
19
 
20
20
 
21
- @scheduler(name="Validate Products and inactive subscriptions", time_unit="day", at="02:30")
21
+ @scheduler.scheduled_job( # type: ignore[misc]
22
+ id="validate-products",
23
+ name="Validate Products and inactive subscriptions",
24
+ trigger="cron",
25
+ hour=2,
26
+ minute=30,
27
+ )
22
28
  def validate_products() -> None:
23
29
  uncompleted_products = db.session.scalar(
24
30
  select(func.count())
@@ -16,7 +16,7 @@ from threading import BoundedSemaphore
16
16
 
17
17
  import structlog
18
18
 
19
- from orchestrator.schedules.scheduling import scheduler
19
+ from orchestrator.schedules.scheduler import scheduler
20
20
  from orchestrator.services.subscriptions import (
21
21
  get_subscriptions_on_product_table,
22
22
  get_subscriptions_on_product_table_in_sync,
@@ -33,7 +33,7 @@ logger = structlog.get_logger(__name__)
33
33
  task_semaphore = BoundedSemaphore(value=2)
34
34
 
35
35
 
36
- @scheduler(name="Subscriptions Validator", time_unit="day", at="00:10")
36
+ @scheduler.scheduled_job(id="subscriptions-validator", name="Subscriptions Validator", trigger="cron", hour=0, minute=10) # type: ignore[misc]
37
37
  def validate_subscriptions() -> None:
38
38
  if app_settings.VALIDATE_OUT_OF_SYNC_SUBSCRIPTIONS:
39
39
  # Automatically re-validate out-of-sync subscriptions. This is not recommended for production.
@@ -41,7 +41,7 @@ def retrieve_input_state(process_id: UUID, input_type: InputType, raise_exceptio
41
41
  select(InputStateTable)
42
42
  .filter(InputStateTable.process_id == process_id)
43
43
  .filter(InputStateTable.input_type == input_type)
44
- .order_by(InputStateTable.input_time.asc())
44
+ .order_by(InputStateTable.input_time.desc())
45
45
  ).first()
46
46
 
47
47
  if res:
@@ -733,11 +733,17 @@ def abort_process(process: ProcessTable, user: str, broadcast_func: Callable | N
733
733
 
734
734
 
735
735
  def _recoverwf(wf: Workflow, log: list[WFProcess]) -> tuple[WFProcess, StepList]:
736
- # Remove all extra steps (Failed, Suspended and (A)waiting steps in db). Only keep cleared steps.
736
+ """Recover workflow state and remaining steps from the given 'Process step' objects.
737
737
 
738
- persistent = list(
739
- filter(lambda p: not (p.isfailed() or p.issuspend() or p.iswaiting() or p.isawaitingcallback()), log)
740
- )
738
+ Returns:
739
+ - The state accumulated up until the last cleared (completed) step\
740
+ - The remaining steps to execute (including Failed, Suspended and (A)waiting steps)
741
+ """
742
+
743
+ def is_cleared(p: WFProcess) -> bool:
744
+ return not (p.isfailed() or p.issuspend() or p.iswaiting() or p.isawaitingcallback())
745
+
746
+ persistent = [p for p in log if is_cleared(p)]
741
747
  stepcount = len(persistent)
742
748
 
743
749
  if log and (log[-1].issuspend() or log[-1].isawaitingcallback()):
@@ -760,15 +766,14 @@ def _recoverwf(wf: Workflow, log: list[WFProcess]) -> tuple[WFProcess, StepList]
760
766
 
761
767
 
762
768
  def _restore_log(steps: list[ProcessStepTable]) -> list[WFProcess]:
763
- result = []
764
- for step in steps:
765
- process = WFProcess.from_status(step.status, step.state)
769
+ """Deserialize ProcessStepTable objects into foldable 'Process step' objects."""
766
770
 
767
- if not process:
768
- raise ValueError(step.status)
771
+ def deserialize(step: ProcessStepTable) -> WFProcess:
772
+ if not (wf_process := WFProcess.from_status(step.status, step.state)):
773
+ raise ValueError(f"Unable to deserialize step from it's status {step.status}")
774
+ return wf_process
769
775
 
770
- result.append(process)
771
- return result
776
+ return [deserialize(step) for step in steps]
772
777
 
773
778
 
774
779
  def load_process(process: ProcessTable) -> ProcessStat:
orchestrator/workflow.py CHANGED
@@ -552,7 +552,7 @@ class ProcessStat:
552
552
  process_id: UUID
553
553
  workflow: Workflow
554
554
  state: Process
555
- log: StepList
555
+ log: StepList # Remaining steps to execute
556
556
  current_user: str
557
557
  user_model: OIDCUserModel | None = None
558
558
 
@@ -597,6 +597,13 @@ class StepStatus(strEnum):
597
597
 
598
598
 
599
599
  class Process(Generic[S]):
600
+ """ADT base class.
601
+
602
+ This class defines an Algebraic Data Type - specifically a "sum type" - that defines the possible
603
+ variants of a Process. It encapsulates the state and allows to fold _instances_ of a process into
604
+ a single value. These instances correspond to subsequent steps of the process.
605
+ """
606
+
600
607
  def __init__(self, s: S):
601
608
  self.s = s
602
609
 
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: orchestrator-core
3
- Version: 4.3.0rc2
3
+ Version: 4.4.0rc1
4
4
  Summary: This is the orchestrator workflow engine.
5
5
  Author-email: SURF <automation-beheer@surf.nl>
6
6
  Requires-Python: >=3.11,<3.14
@@ -32,6 +32,7 @@ Classifier: Typing :: Typed
32
32
  License-File: LICENSE
33
33
  Requires-Dist: alembic==1.16.1
34
34
  Requires-Dist: anyio>=3.7.0
35
+ Requires-Dist: apscheduler>=3.11.0
35
36
  Requires-Dist: click==8.*
36
37
  Requires-Dist: deepmerge==2.0
37
38
  Requires-Dist: deprecated>=1.2.18
@@ -52,7 +53,6 @@ Requires-Dist: python-dateutil==2.8.2
52
53
  Requires-Dist: python-rapidjson>=1.18,<1.21
53
54
  Requires-Dist: pytz==2025.2
54
55
  Requires-Dist: redis==5.1.1
55
- Requires-Dist: schedule==1.1.0
56
56
  Requires-Dist: semver==3.0.4
57
57
  Requires-Dist: sentry-sdk[fastapi]~=2.29.1
58
58
  Requires-Dist: sqlalchemy==2.0.41
@@ -1,4 +1,4 @@
1
- orchestrator/__init__.py,sha256=k7E0AiB5QFgGg_GlFGkYHF-DTfhWCYkziVwvIXSWeAc,1066
1
+ orchestrator/__init__.py,sha256=2VCmBMCQvlduTSR0AOWiSAcfGIGfm7kGJxYEKzqiXew,1066
2
2
  orchestrator/app.py,sha256=7UrXKjBKNSEaSSXAd5ww_RdMFhFqE4yvfj8faS2MzAA,12089
3
3
  orchestrator/exception_handlers.py,sha256=UsW3dw8q0QQlNLcV359bIotah8DYjMsj2Ts1LfX4ClY,1268
4
4
  orchestrator/log_config.py,sha256=1tPRX5q65e57a6a_zEii_PFK8SzWT0mnA5w2sKg4hh8,1853
@@ -8,7 +8,7 @@ orchestrator/settings.py,sha256=2Kgc6m3qUCcSM3Z_IVUeehfgO0QphMFkLrS0RC3sU-U,4365
8
8
  orchestrator/targets.py,sha256=WizBgnp8hWX9YLFUIju7ewSubiwQqinCvyiYNcXHbHI,802
9
9
  orchestrator/types.py,sha256=qzs7xx5AYRmKbpYRyJJP3wuDb0W0bcAzefCN0RWLAco,15459
10
10
  orchestrator/version.py,sha256=b58e08lxs47wUNXv0jXFO_ykpksmytuzEXD4La4W-NQ,1366
11
- orchestrator/workflow.py,sha256=PVHe6vnnkswzqw2UoY-j6NMSEhL6rLHXRnO7yLOyDC8,45551
11
+ orchestrator/workflow.py,sha256=meDCPnyyX_n5PsMUaFy2wWb5EKNm1_ff7zRDBYrbcDg,45901
12
12
  orchestrator/api/__init__.py,sha256=GyHNfEFCGKQwRiN6rQmvSRH2iYX7npjMZn97n8XzmLU,571
13
13
  orchestrator/api/error_handling.py,sha256=YrPCxSa-DSa9KwqIMlXI-KGBGnbGIW5ukOPiikUH9E4,1502
14
14
  orchestrator/api/helpers.py,sha256=s0QRHYw8AvEmlkmRhuEzz9xixaZKUF3YuPzUVHkcoXk,6933
@@ -17,7 +17,7 @@ orchestrator/api/api_v1/__init__.py,sha256=GyHNfEFCGKQwRiN6rQmvSRH2iYX7npjMZn97n
17
17
  orchestrator/api/api_v1/api.py,sha256=m4iDktsSpzxUDaudkdgXeZ83a6B4wfc3pczQsa-Pb-8,2866
18
18
  orchestrator/api/api_v1/endpoints/__init__.py,sha256=GyHNfEFCGKQwRiN6rQmvSRH2iYX7npjMZn97n8XzmLU,571
19
19
  orchestrator/api/api_v1/endpoints/health.py,sha256=iaxs1XX1_250_gKNsspuULCV2GEMBjbtjsmfQTOvMAI,1284
20
- orchestrator/api/api_v1/endpoints/processes.py,sha256=kWz_jL8_sTNwl44tU17VwkwZGjBIw1IIW5pYCCSHwgs,15891
20
+ orchestrator/api/api_v1/endpoints/processes.py,sha256=238Bydgj4ILNyMU_7j_Q7a0WGlfIvKv5ypP7lESU32w,16188
21
21
  orchestrator/api/api_v1/endpoints/product_blocks.py,sha256=kZ6ywIOsS_S2qGq7RvZ4KzjvaS1LmwbGWR37AKRvWOw,2146
22
22
  orchestrator/api/api_v1/endpoints/products.py,sha256=BfFtwu9dZXEQbtKxYj9icc73GKGvAGMR5ytyf41nQlQ,3081
23
23
  orchestrator/api/api_v1/endpoints/resource_types.py,sha256=gGyuaDyOD0TAVoeFGaGmjDGnQ8eQQArOxKrrk4MaDzA,2145
@@ -36,7 +36,7 @@ orchestrator/cli/migrate_domain_models.py,sha256=WRXy_1OnziQwpsCFZXvjB30nDJtjj0i
36
36
  orchestrator/cli/migrate_tasks.py,sha256=bju8XColjSZD0v3rS4kl-24dLr8En_H4-6enBmqd494,7255
37
37
  orchestrator/cli/migrate_workflows.py,sha256=nxUpx0vgEIc_8aJrjAyrw3E9Dt8JmaamTts8oiQ4vHY,8923
38
38
  orchestrator/cli/migration_helpers.py,sha256=C5tpkP5WEBr7G9S-1k1hgSI8ili6xd9Z5ygc9notaK0,4110
39
- orchestrator/cli/scheduler.py,sha256=iCKBWYUwQIYTDqKQ9rMVvs2sNiAzE-J2SkV170TPP2g,1896
39
+ orchestrator/cli/scheduler.py,sha256=jdBbgqE7bNUOevFe6gZS5C-SyDTOow8baNJKEBTkS0A,2293
40
40
  orchestrator/cli/domain_gen_helpers/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
41
41
  orchestrator/cli/domain_gen_helpers/fixed_input_helpers.py,sha256=uzpwsaau81hHSxNMOS9-o7kF-9_78R0f_UE0AvWooZQ,6775
42
42
  orchestrator/cli/domain_gen_helpers/helpers.py,sha256=tIPxn8ezED_xYZxH7ZAtQLwkDc6RNmLZVxWAoJ3a9lw,4203
@@ -158,7 +158,7 @@ orchestrator/forms/validators/product_id.py,sha256=u5mURLT0pOhbFLdwvYcy2_2fXMt35
158
158
  orchestrator/graphql/__init__.py,sha256=avq8Yg3Jr_9pJqh7ClyIAOX7YSg1eM_AWmt5C3FRYUY,1440
159
159
  orchestrator/graphql/autoregistration.py,sha256=pF2jbMKG26MvYoMSa6ZpqpHjVks7_NvSRFymHTgmfjs,6342
160
160
  orchestrator/graphql/pagination.py,sha256=iqVDn3GPZpiQhEydfwkBJLURY-X8wwUphS8Lkeg0BOc,2413
161
- orchestrator/graphql/schema.py,sha256=gwZ3nAgKL0zlpc-aK58hSUAGPVD11Tb3aRSSK9hC39I,9204
161
+ orchestrator/graphql/schema.py,sha256=dw4m4sM1ek2DscB8vINN6L8vVDE0h5GXclHGa8CiUJo,9537
162
162
  orchestrator/graphql/types.py,sha256=_kHKMusrRPuRtF4wm42NsBzoFZ4egbu3ibMmhd2D6Fs,5432
163
163
  orchestrator/graphql/extensions/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
164
164
  orchestrator/graphql/extensions/model_cache.py,sha256=1uhMRjBs9eK7zJ1Y6P6BopX06822w2Yh9jliwYvG6yQ,1085
@@ -174,6 +174,7 @@ orchestrator/graphql/resolvers/process.py,sha256=Hqs1F7-gw0yO_ioHjh2eLAyxrK2WSuL
174
174
  orchestrator/graphql/resolvers/product.py,sha256=uPBmYwMdau-zUqNjoDl-LDn927u3aCFW5JQ4A_it8q0,2772
175
175
  orchestrator/graphql/resolvers/product_block.py,sha256=Ker1CpxGab5h2BZujOHHwRUj8W4uphRr3WSpQGk2PnI,2939
176
176
  orchestrator/graphql/resolvers/resource_type.py,sha256=SREZXjkLYpuo4nCM8DqVeImIrZcP3xDiWr_gq4wWaxQ,2956
177
+ orchestrator/graphql/resolvers/scheduled_tasks.py,sha256=QsnesRrj8ESuS9vPKG9DXYcG2Wfj9m5LWGeZgmc6hu8,1640
177
178
  orchestrator/graphql/resolvers/settings.py,sha256=xVYqxo-EWQ24F4hUHm9OZeN9vsqQXJzIJ1_HF4Ci9Cs,3777
178
179
  orchestrator/graphql/resolvers/subscription.py,sha256=57niFv-JCro_wm0peJ5Ne04F2WIPuJ-Lx2h8yd9qubA,6541
179
180
  orchestrator/graphql/resolvers/version.py,sha256=qgwe1msPOexeg3RHCscJ8s45vNfMhYh9ZKyCZ3MNw30,809
@@ -184,17 +185,18 @@ orchestrator/graphql/schemas/customer_description.py,sha256=fize71IMpkvk_rTzcqCY
184
185
  orchestrator/graphql/schemas/errors.py,sha256=VRl-Zd1FHMnscyozhfxzqeEUZ0ERAWum_Y8YwjGxwmA,203
185
186
  orchestrator/graphql/schemas/fixed_input.py,sha256=1yqYHADQRgHz8OIP7ObYsPFS-gmzfkCvEO0a-KKf7zI,513
186
187
  orchestrator/graphql/schemas/helpers.py,sha256=Kpj4kIbmoKKN35bdgUSwQvGUIbeg7VJAVMEq65YS_ik,346
187
- orchestrator/graphql/schemas/process.py,sha256=nvD6Rvr0hnrMINdXF_rQuLF8szKJ7E-SywCFMuZsnlg,4940
188
+ orchestrator/graphql/schemas/process.py,sha256=g3noYh_USfnaK59fnoX2DI5tAf1PhdLMJGI_lA2xX1M,4966
188
189
  orchestrator/graphql/schemas/product.py,sha256=vUCqcjrKBJj-VKSrMYPKzjmmxLMXL7alKTJ8UdUkhTg,4342
189
190
  orchestrator/graphql/schemas/product_block.py,sha256=Qk9cbA6vm7ZPrhdgPHatKRuy6TytBmxSr97McEOxAu8,2860
190
191
  orchestrator/graphql/schemas/resource_type.py,sha256=s5d_FwQXL2-Sc-IDUxTJun5qFQ4zOP4-XcHF9ql-t1g,898
192
+ orchestrator/graphql/schemas/scheduled_task.py,sha256=22Kb7r2pUoefz9AeiTYQ1t6YBYCmAs_zVp7FqNdN5HQ,194
191
193
  orchestrator/graphql/schemas/settings.py,sha256=drhm5VcLmUbiYAk6WUSJcyJqjNM96E6GvpxVdPAobnA,999
192
194
  orchestrator/graphql/schemas/strawberry_pydantic_patch.py,sha256=CjNUhTKdYmLiaem-WY_mzw4HASIeaZitxGF8pPocqVw,1602
193
195
  orchestrator/graphql/schemas/subscription.py,sha256=hTA34C27kgLguH9V53173CxMKIWiQKh3vFzyJ2yBfE0,9918
194
196
  orchestrator/graphql/schemas/version.py,sha256=HSzVg_y4Sjd5_H5rRUtu3FJKOG_8ifhvBNt_qjOtC-E,92
195
197
  orchestrator/graphql/schemas/workflow.py,sha256=WLbegRNxOfvXg4kPYrO5KPBwtHmUofAr2pvZT2JsW1c,1761
196
198
  orchestrator/graphql/utils/__init__.py,sha256=1JvenzEVW1CBa1sGVI9I8IWnnoXIkb1hneDqph9EEZY,524
197
- orchestrator/graphql/utils/create_resolver_error_handler.py,sha256=PpQMVwGrE9t0nZ12TwoxPxksXxEwQM7lSNPeh7qW3vk,1233
199
+ orchestrator/graphql/utils/create_resolver_error_handler.py,sha256=XzCnL482M4wz3fg5fUdGUwCAuzSZQ9Ufu1mscLyeoWU,1227
198
200
  orchestrator/graphql/utils/get_query_loaders.py,sha256=abS_HJ7K9een78gMiGq3IhwGwxQXHvZygExe0h_t9ns,815
199
201
  orchestrator/graphql/utils/get_selected_fields.py,sha256=0hBcQkU-7TNVO_KG-MmLItKm0O3gmbqoxXNkLHO-wHo,1002
200
202
  orchestrator/graphql/utils/get_selected_paths.py,sha256=H0btESeOr3_VB7zy5Cx25OS0uzBcg2Y1I-arAmSOnsQ,1382
@@ -208,6 +210,7 @@ orchestrator/metrics/init.py,sha256=xBITvDjbNf-iabbBg0tAW8TPj6-wzr_MerOOqgDsoS4,
208
210
  orchestrator/metrics/processes.py,sha256=SyogN5NSuhYoRv2CSUE1So9e8Gkrwa71J6oGLOdODQU,5333
209
211
  orchestrator/metrics/subscriptions.py,sha256=vC1O8VmTq5oJxNrn5CU99Rf8cxzdyhc7tXbZBSAU-O8,3036
210
212
  orchestrator/migrations/README,sha256=heMzebYwlGhnE8_4CWJ4LS74WoEZjBy-S-mIJRxAEKI,39
213
+ orchestrator/migrations/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
211
214
  orchestrator/migrations/alembic.ini,sha256=kMoADqhGeubU8xanILNaqm4oixLy9m4ngYtdGpZcc7I,873
212
215
  orchestrator/migrations/env.py,sha256=M_cPoAL2axuuup5fvMy8I_WTPHEw0RbPEHkhZ3QEGoE,3740
213
216
  orchestrator/migrations/helpers.py,sha256=CAGGKhxpmhyKGfYcO-SUCPfMTOCZPfEpkJrcm2MYfcE,47979
@@ -244,12 +247,13 @@ orchestrator/migrations/versions/schema/2025-05-08_161918133bec_add_is_task_to_w
244
247
  orchestrator/migrations/versions/schema/2025-07-01_93fc5834c7e5_changed_timestamping_fields_in_process_steps.py,sha256=Oezd8b2qaI1Kyq-sZFVFmdzd4d9NjXrf6HtJGk11fy0,1914
245
248
  orchestrator/migrations/versions/schema/2025-07-04_4b58e336d1bf_deprecating_workflow_target_in_.py,sha256=xnD6w-97R4ClS7rbmXQEXc36K3fdcXKhCy7ZZNy_FX4,742
246
249
  orchestrator/migrations/versions/schema/2025-07-28_850dccac3b02_update_description_of_resume_workflows_.py,sha256=R6Qoga83DJ1IL0WYPu0u5u2ZvAmqGlDmUMv_KtJyOhQ,812
247
- orchestrator/schedules/__init__.py,sha256=JnnaglfK1qYUBKI6Dd9taV-tCZIPlAdAkHtnkJDMXxY,1066
248
- orchestrator/schedules/resume_workflows.py,sha256=kSotzTAXjX7p9fpSYiGOpuxuTQfv54eRFAe0YSG0DHc,832
249
- orchestrator/schedules/scheduling.py,sha256=ehtwgpbvMOk1jhn-hHgVzg_9wLJkI6l3mRY3DcO9ZVY,1526
250
- orchestrator/schedules/task_vacuum.py,sha256=eovnuKimU8SFRw1IF62MsAVFSdgeeV1u57kapUbz8As,821
251
- orchestrator/schedules/validate_products.py,sha256=YMr7ASSqdXM6pd6oZu0kr8mfmH8If16MzprrsHdN_ZU,1234
252
- orchestrator/schedules/validate_subscriptions.py,sha256=9SYvsn4BJ5yo_1nu555hWjl5XffTx7QMaRhH5oOjM9E,2042
250
+ orchestrator/schedules/__init__.py,sha256=Zy0fTOBMGIRFoh5iVFDLF9_PRAFaONYDThGK9EsysWo,981
251
+ orchestrator/schedules/resume_workflows.py,sha256=jRnVRWDy687pQu-gtk80ecwiLSdrvtL15tG3U2zWA6I,891
252
+ orchestrator/schedules/scheduler.py,sha256=vze3xaZhUL5maKQB6a1gCvc9AcGw3jX-BHT3d5xvy6A,5430
253
+ orchestrator/schedules/scheduling.py,sha256=_mbpHMhijey8Y56ebtJ4wVkrp_kPVRm8hoByzlQF4SE,2821
254
+ orchestrator/schedules/task_vacuum.py,sha256=mxb7fsy1GphRwvUWi_lvwNaj51YAXUdIDlkOJd90AFI,874
255
+ orchestrator/schedules/validate_products.py,sha256=zWFQeVn3F8LP3joExLiKdmHs008pZsO-RolcIXHjFyE,1322
256
+ orchestrator/schedules/validate_subscriptions.py,sha256=bUBV45aEuqVdtqYBAXh1lX4O5vuNTeTfds4J_zq35dI,2113
253
257
  orchestrator/schemas/__init__.py,sha256=YDyZ0YBvzB4ML9oDBCBPGnBvf680zFFgUzg7X0tYBRY,2326
254
258
  orchestrator/schemas/base.py,sha256=Vc444LetsINLRhG2SxW9Bq01hOzChPOhQWCImQTr-As,930
255
259
  orchestrator/schemas/engine_settings.py,sha256=LF8al7tJssiilb5A4emPtUYo0tVDSaT1Lvo_DN_ttrY,1296
@@ -264,9 +268,9 @@ orchestrator/schemas/subscription_descriptions.py,sha256=Ft_jw1U0bf9Z0U8O4OWfLlc
264
268
  orchestrator/schemas/workflow.py,sha256=VqQ9XfV4fVd6MjY0LRRQzWBJHmlPsAamWfTwDx1cZkg,2102
265
269
  orchestrator/services/__init__.py,sha256=GyHNfEFCGKQwRiN6rQmvSRH2iYX7npjMZn97n8XzmLU,571
266
270
  orchestrator/services/fixed_inputs.py,sha256=kyz7s2HLzyDulvcq-ZqefTw1om86COvyvTjz0_5CmgI,876
267
- orchestrator/services/input_state.py,sha256=WfiQrvatPtE1jR8xArE0XamONt1nemTv5iaROgSiJpM,2389
271
+ orchestrator/services/input_state.py,sha256=6BZOpb3cHpO18K-XG-3QUIV9pIM25_ufdODrp5CmXG4,2390
268
272
  orchestrator/services/process_broadcast_thread.py,sha256=D44YbjF8mRqGuznkRUV4SoRn1J0lfy_x1H508GnSVlU,4649
269
- orchestrator/services/processes.py,sha256=JGM9vWbUjvEpy-IpTIgaYaqcTBKMI-CWTY8SJKBf3eI,30153
273
+ orchestrator/services/processes.py,sha256=NfzdtH4eZK_wYuSmFtUX69qDvoeI8J7sJ2fFyY_VYaM,30544
270
274
  orchestrator/services/products.py,sha256=BP4KyE8zO-8z7Trrs5T6zKBOw53S9BfBJnHWI3p6u5Y,1943
271
275
  orchestrator/services/resource_types.py,sha256=_QBy_JOW_X3aSTqH0CuLrq4zBJL0p7Q-UDJUcuK2_qc,884
272
276
  orchestrator/services/settings.py,sha256=HEWfFulgoEDwgfxGEO__QTr5fDiwNBEj1UhAeTAdbLQ,3159
@@ -315,7 +319,7 @@ orchestrator/workflows/tasks/resume_workflows.py,sha256=T3iobSJjVgiupe0rClD34kUZ
315
319
  orchestrator/workflows/tasks/validate_product_type.py,sha256=paG-NAY1bdde3Adt8zItkcBKf5Pxw6f5ngGW6an6dYU,3192
316
320
  orchestrator/workflows/tasks/validate_products.py,sha256=GZJBoFF-WMphS7ghMs2-gqvV2iL1F0POhk0uSNt93n0,8510
317
321
  orchestrator/workflows/translations/en-GB.json,sha256=ST53HxkphFLTMjFHonykDBOZ7-P_KxksktZU3GbxLt0,846
318
- orchestrator_core-4.3.0rc2.dist-info/licenses/LICENSE,sha256=b-aA5OZQuuBATmLKo_mln8CQrDPPhg3ghLzjPjLn4Tg,11409
319
- orchestrator_core-4.3.0rc2.dist-info/WHEEL,sha256=G2gURzTEtmeR8nrdXUJfNiB3VYVxigPQ-bEQujpNiNs,82
320
- orchestrator_core-4.3.0rc2.dist-info/METADATA,sha256=aJsbvoK-FPIjx_tH3UJy3KWgj6aLhFwUb7tJvUxq9gQ,5963
321
- orchestrator_core-4.3.0rc2.dist-info/RECORD,,
322
+ orchestrator_core-4.4.0rc1.dist-info/licenses/LICENSE,sha256=b-aA5OZQuuBATmLKo_mln8CQrDPPhg3ghLzjPjLn4Tg,11409
323
+ orchestrator_core-4.4.0rc1.dist-info/WHEEL,sha256=G2gURzTEtmeR8nrdXUJfNiB3VYVxigPQ-bEQujpNiNs,82
324
+ orchestrator_core-4.4.0rc1.dist-info/METADATA,sha256=as5RXg0Y5DUGPcpT-6raY0HmvEeDJfEMsg1dIQfqoqU,5967
325
+ orchestrator_core-4.4.0rc1.dist-info/RECORD,,