orchestrator-core 4.3.0rc2__py3-none-any.whl → 4.4.0rc1__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- orchestrator/__init__.py +1 -1
- orchestrator/api/api_v1/endpoints/processes.py +12 -6
- orchestrator/cli/scheduler.py +39 -29
- orchestrator/graphql/resolvers/scheduled_tasks.py +36 -0
- orchestrator/graphql/schema.py +5 -0
- orchestrator/graphql/schemas/process.py +2 -2
- orchestrator/graphql/schemas/scheduled_task.py +8 -0
- orchestrator/graphql/utils/create_resolver_error_handler.py +1 -1
- orchestrator/migrations/__init__.py +0 -0
- orchestrator/schedules/__init__.py +1 -2
- orchestrator/schedules/resume_workflows.py +2 -2
- orchestrator/schedules/scheduler.py +153 -0
- orchestrator/schedules/scheduling.py +64 -24
- orchestrator/schedules/task_vacuum.py +2 -2
- orchestrator/schedules/validate_products.py +8 -2
- orchestrator/schedules/validate_subscriptions.py +2 -2
- orchestrator/services/input_state.py +1 -1
- orchestrator/services/processes.py +16 -11
- orchestrator/workflow.py +8 -1
- {orchestrator_core-4.3.0rc2.dist-info → orchestrator_core-4.4.0rc1.dist-info}/METADATA +2 -2
- {orchestrator_core-4.3.0rc2.dist-info → orchestrator_core-4.4.0rc1.dist-info}/RECORD +23 -19
- {orchestrator_core-4.3.0rc2.dist-info → orchestrator_core-4.4.0rc1.dist-info}/WHEEL +0 -0
- {orchestrator_core-4.3.0rc2.dist-info → orchestrator_core-4.4.0rc1.dist-info}/licenses/LICENSE +0 -0
orchestrator/__init__.py
CHANGED
|
@@ -25,7 +25,7 @@ from fastapi.param_functions import Body, Depends, Header
|
|
|
25
25
|
from fastapi.routing import APIRouter
|
|
26
26
|
from fastapi.websockets import WebSocket
|
|
27
27
|
from fastapi_etag.dependency import CacheHit
|
|
28
|
-
from more_itertools import chunked, last
|
|
28
|
+
from more_itertools import chunked, first, last
|
|
29
29
|
from sentry_sdk.tracing import trace
|
|
30
30
|
from sqlalchemy import CompoundSelect, Select, select
|
|
31
31
|
from sqlalchemy.orm import defer, joinedload
|
|
@@ -88,11 +88,17 @@ def check_global_lock() -> None:
|
|
|
88
88
|
)
|
|
89
89
|
|
|
90
90
|
|
|
91
|
-
def
|
|
92
|
-
"""Extract
|
|
93
|
-
|
|
91
|
+
def get_steps_to_evaluate_for_rbac(pstat: ProcessStat) -> StepList:
|
|
92
|
+
"""Extract all steps from the ProcessStat for a process that should be evaluated for a RBAC callback.
|
|
93
|
+
|
|
94
|
+
For a suspended process this includes all previously completed steps as well as the current step.
|
|
95
|
+
For a completed process this includes all steps.
|
|
96
|
+
"""
|
|
97
|
+
if not (remaining_steps := pstat.log):
|
|
98
|
+
return pstat.workflow.steps
|
|
99
|
+
|
|
94
100
|
past_steps = pstat.workflow.steps[: -len(remaining_steps)]
|
|
95
|
-
return StepList(past_steps
|
|
101
|
+
return StepList(past_steps >> first(remaining_steps))
|
|
96
102
|
|
|
97
103
|
|
|
98
104
|
def get_auth_callbacks(steps: StepList, workflow: Workflow) -> tuple[Authorizer | None, Authorizer | None]:
|
|
@@ -200,7 +206,7 @@ def resume_process_endpoint(
|
|
|
200
206
|
raise_status(HTTPStatus.CONFLICT, f"Resuming a {process.last_status.lower()} workflow is not possible")
|
|
201
207
|
|
|
202
208
|
pstat = load_process(process)
|
|
203
|
-
auth_resume, auth_retry = get_auth_callbacks(
|
|
209
|
+
auth_resume, auth_retry = get_auth_callbacks(get_steps_to_evaluate_for_rbac(pstat), pstat.workflow)
|
|
204
210
|
if process.last_status == ProcessStatus.SUSPENDED:
|
|
205
211
|
if auth_resume is not None and not auth_resume(user_model):
|
|
206
212
|
raise_status(HTTPStatus.FORBIDDEN, "User is not authorized to resume step")
|
orchestrator/cli/scheduler.py
CHANGED
|
@@ -13,12 +13,11 @@
|
|
|
13
13
|
|
|
14
14
|
|
|
15
15
|
import logging
|
|
16
|
-
from time import sleep
|
|
17
16
|
|
|
18
|
-
import schedule
|
|
19
17
|
import typer
|
|
18
|
+
from apscheduler.schedulers.blocking import BlockingScheduler
|
|
20
19
|
|
|
21
|
-
from orchestrator.schedules import
|
|
20
|
+
from orchestrator.schedules.scheduler import jobstores, scheduler, scheduler_dispose_db_connections
|
|
22
21
|
|
|
23
22
|
log = logging.getLogger(__name__)
|
|
24
23
|
|
|
@@ -27,36 +26,47 @@ app: typer.Typer = typer.Typer()
|
|
|
27
26
|
|
|
28
27
|
@app.command()
|
|
29
28
|
def run() -> None:
|
|
30
|
-
"""
|
|
31
|
-
|
|
32
|
-
|
|
33
|
-
|
|
34
|
-
|
|
35
|
-
|
|
36
|
-
|
|
37
|
-
|
|
38
|
-
log.info("%s: %s", ", ".join(j.tags), j)
|
|
39
|
-
while True:
|
|
40
|
-
schedule.run_pending()
|
|
41
|
-
idle = schedule.idle_seconds()
|
|
42
|
-
if idle < 0:
|
|
43
|
-
log.info("Next job in queue is scheduled in the past, run it now.")
|
|
44
|
-
else:
|
|
45
|
-
log.info("Sleeping for %d seconds", idle)
|
|
46
|
-
sleep(idle)
|
|
29
|
+
"""Start scheduler and loop eternally to keep thread alive."""
|
|
30
|
+
blocking_scheduler = BlockingScheduler(jobstores=jobstores)
|
|
31
|
+
|
|
32
|
+
try:
|
|
33
|
+
blocking_scheduler.start()
|
|
34
|
+
except (KeyboardInterrupt, SystemExit):
|
|
35
|
+
scheduler.shutdown()
|
|
36
|
+
scheduler_dispose_db_connections()
|
|
47
37
|
|
|
48
38
|
|
|
49
39
|
@app.command()
|
|
50
40
|
def show_schedule() -> None:
|
|
51
|
-
"""Show the currently configured schedule.
|
|
52
|
-
|
|
53
|
-
|
|
54
|
-
|
|
41
|
+
"""Show the currently configured schedule.
|
|
42
|
+
|
|
43
|
+
in cli underscore is replaced by a dash `show-schedule`
|
|
44
|
+
"""
|
|
45
|
+
scheduler.start(paused=True) # paused: avoid triggering jobs during CLI
|
|
46
|
+
jobs = scheduler.get_jobs()
|
|
47
|
+
scheduler.shutdown(wait=False)
|
|
48
|
+
scheduler_dispose_db_connections()
|
|
49
|
+
|
|
50
|
+
for job in jobs:
|
|
51
|
+
typer.echo(f"[{job.id}] Next run: {job.next_run_time} | Trigger: {job.trigger}")
|
|
55
52
|
|
|
56
53
|
|
|
57
54
|
@app.command()
|
|
58
|
-
def force(
|
|
59
|
-
"""Force the execution of (a) scheduler(s) based on a
|
|
60
|
-
|
|
61
|
-
|
|
62
|
-
|
|
55
|
+
def force(job_id: str) -> None:
|
|
56
|
+
"""Force the execution of (a) scheduler(s) based on a job_id."""
|
|
57
|
+
scheduler.start(paused=True) # paused: avoid triggering jobs during CLI
|
|
58
|
+
job = scheduler.get_job(job_id)
|
|
59
|
+
scheduler.shutdown(wait=False)
|
|
60
|
+
scheduler_dispose_db_connections()
|
|
61
|
+
|
|
62
|
+
if not job:
|
|
63
|
+
typer.echo(f"Job '{job_id}' not found.")
|
|
64
|
+
raise typer.Exit(code=1)
|
|
65
|
+
|
|
66
|
+
typer.echo(f"Running job [{job.id}] now...")
|
|
67
|
+
try:
|
|
68
|
+
job.func(*job.args or (), **job.kwargs or {})
|
|
69
|
+
typer.echo("Job executed successfully.")
|
|
70
|
+
except Exception as e:
|
|
71
|
+
typer.echo(f"Job execution failed: {e}")
|
|
72
|
+
raise typer.Exit(code=1)
|
|
@@ -0,0 +1,36 @@
|
|
|
1
|
+
import structlog
|
|
2
|
+
|
|
3
|
+
from orchestrator.db.filters import Filter
|
|
4
|
+
from orchestrator.db.sorting import Sort
|
|
5
|
+
from orchestrator.graphql.pagination import Connection
|
|
6
|
+
from orchestrator.graphql.schemas.scheduled_task import ScheduledTaskGraphql
|
|
7
|
+
from orchestrator.graphql.types import GraphqlFilter, GraphqlSort, OrchestratorInfo
|
|
8
|
+
from orchestrator.graphql.utils import create_resolver_error_handler, to_graphql_result_page
|
|
9
|
+
from orchestrator.graphql.utils.is_query_detailed import is_querying_page_data
|
|
10
|
+
from orchestrator.schedules.scheduler import get_scheduler_tasks, scheduled_task_filter_keys, scheduled_task_sort_keys
|
|
11
|
+
|
|
12
|
+
logger = structlog.get_logger(__name__)
|
|
13
|
+
|
|
14
|
+
|
|
15
|
+
async def resolve_scheduled_tasks(
|
|
16
|
+
info: OrchestratorInfo,
|
|
17
|
+
filter_by: list[GraphqlFilter] | None = None,
|
|
18
|
+
sort_by: list[GraphqlSort] | None = None,
|
|
19
|
+
first: int = 10,
|
|
20
|
+
after: int = 0,
|
|
21
|
+
) -> Connection[ScheduledTaskGraphql]:
|
|
22
|
+
_error_handler = create_resolver_error_handler(info)
|
|
23
|
+
|
|
24
|
+
pydantic_filter_by: list[Filter] = [item.to_pydantic() for item in filter_by] if filter_by else []
|
|
25
|
+
pydantic_sort_by: list[Sort] = [item.to_pydantic() for item in sort_by] if sort_by else []
|
|
26
|
+
scheduled_tasks, total = get_scheduler_tasks(
|
|
27
|
+
first=first, after=after, filter_by=pydantic_filter_by, sort_by=pydantic_sort_by, error_handler=_error_handler
|
|
28
|
+
)
|
|
29
|
+
|
|
30
|
+
graphql_scheduled_tasks = []
|
|
31
|
+
if is_querying_page_data(info):
|
|
32
|
+
graphql_scheduled_tasks = [ScheduledTaskGraphql.from_pydantic(p) for p in scheduled_tasks]
|
|
33
|
+
|
|
34
|
+
return to_graphql_result_page(
|
|
35
|
+
graphql_scheduled_tasks, first, after, total, scheduled_task_filter_keys, scheduled_task_sort_keys
|
|
36
|
+
)
|
orchestrator/graphql/schema.py
CHANGED
|
@@ -51,12 +51,14 @@ from orchestrator.graphql.resolvers import (
|
|
|
51
51
|
resolve_version,
|
|
52
52
|
resolve_workflows,
|
|
53
53
|
)
|
|
54
|
+
from orchestrator.graphql.resolvers.scheduled_tasks import resolve_scheduled_tasks
|
|
54
55
|
from orchestrator.graphql.schemas import DEFAULT_GRAPHQL_MODELS
|
|
55
56
|
from orchestrator.graphql.schemas.customer import CustomerType
|
|
56
57
|
from orchestrator.graphql.schemas.process import ProcessType
|
|
57
58
|
from orchestrator.graphql.schemas.product import ProductType
|
|
58
59
|
from orchestrator.graphql.schemas.product_block import ProductBlock
|
|
59
60
|
from orchestrator.graphql.schemas.resource_type import ResourceType
|
|
61
|
+
from orchestrator.graphql.schemas.scheduled_task import ScheduledTaskGraphql
|
|
60
62
|
from orchestrator.graphql.schemas.settings import StatusType
|
|
61
63
|
from orchestrator.graphql.schemas.subscription import SubscriptionInterface
|
|
62
64
|
from orchestrator.graphql.schemas.version import VersionType
|
|
@@ -99,6 +101,9 @@ class OrchestratorQuery:
|
|
|
99
101
|
description="Returns information about cache, workers, and global engine settings",
|
|
100
102
|
)
|
|
101
103
|
version: VersionType = authenticated_field(resolver=resolve_version, description="Returns version information")
|
|
104
|
+
scheduled_tasks: Connection[ScheduledTaskGraphql] = authenticated_field(
|
|
105
|
+
resolver=resolve_scheduled_tasks, description="Returns scheduled job information"
|
|
106
|
+
)
|
|
102
107
|
|
|
103
108
|
|
|
104
109
|
@strawberry.federation.type(description="Orchestrator customer Query")
|
|
@@ -6,7 +6,7 @@ from strawberry.federation.schema_directives import Key
|
|
|
6
6
|
from strawberry.scalars import JSON
|
|
7
7
|
|
|
8
8
|
from oauth2_lib.strawberry import authenticated_field
|
|
9
|
-
from orchestrator.api.api_v1.endpoints.processes import get_auth_callbacks,
|
|
9
|
+
from orchestrator.api.api_v1.endpoints.processes import get_auth_callbacks, get_steps_to_evaluate_for_rbac
|
|
10
10
|
from orchestrator.db import ProcessTable, ProductTable, db
|
|
11
11
|
from orchestrator.graphql.pagination import EMPTY_PAGE, Connection
|
|
12
12
|
from orchestrator.graphql.schemas.customer import CustomerType
|
|
@@ -86,7 +86,7 @@ class ProcessType:
|
|
|
86
86
|
oidc_user = info.context.get_current_user
|
|
87
87
|
workflow = get_workflow(self.workflow_name)
|
|
88
88
|
process = load_process(db.session.get(ProcessTable, self.process_id)) # type: ignore[arg-type]
|
|
89
|
-
auth_resume, auth_retry = get_auth_callbacks(
|
|
89
|
+
auth_resume, auth_retry = get_auth_callbacks(get_steps_to_evaluate_for_rbac(process), workflow) # type: ignore[arg-type]
|
|
90
90
|
|
|
91
91
|
return FormUserPermissionsType(
|
|
92
92
|
retryAllowed=auth_retry and auth_retry(oidc_user), # type: ignore[arg-type]
|
|
@@ -25,6 +25,6 @@ def _format_context(context: dict) -> str:
|
|
|
25
25
|
|
|
26
26
|
def create_resolver_error_handler(info: OrchestratorInfo) -> CallableErrorHandler:
|
|
27
27
|
def handle_error(message: str, **context) -> None: # type: ignore
|
|
28
|
-
return register_error("
|
|
28
|
+
return register_error(f"{message} {_format_context(context)}", info, error_type=ErrorType.BAD_REQUEST)
|
|
29
29
|
|
|
30
30
|
return handle_error
|
|
File without changes
|
|
@@ -13,12 +13,11 @@
|
|
|
13
13
|
|
|
14
14
|
|
|
15
15
|
from orchestrator.schedules.resume_workflows import run_resume_workflows
|
|
16
|
-
from orchestrator.schedules.scheduling import SchedulingFunction
|
|
17
16
|
from orchestrator.schedules.task_vacuum import vacuum_tasks
|
|
18
17
|
from orchestrator.schedules.validate_products import validate_products
|
|
19
18
|
from orchestrator.schedules.validate_subscriptions import validate_subscriptions
|
|
20
19
|
|
|
21
|
-
ALL_SCHEDULERS: list
|
|
20
|
+
ALL_SCHEDULERS: list = [
|
|
22
21
|
run_resume_workflows,
|
|
23
22
|
vacuum_tasks,
|
|
24
23
|
validate_subscriptions,
|
|
@@ -12,10 +12,10 @@
|
|
|
12
12
|
# limitations under the License.
|
|
13
13
|
|
|
14
14
|
|
|
15
|
-
from orchestrator.schedules.
|
|
15
|
+
from orchestrator.schedules.scheduler import scheduler
|
|
16
16
|
from orchestrator.services.processes import start_process
|
|
17
17
|
|
|
18
18
|
|
|
19
|
-
@scheduler(name="Resume workflows",
|
|
19
|
+
@scheduler.scheduled_job(id="resume-workflows", name="Resume workflows", trigger="interval", hours=1) # type: ignore[misc]
|
|
20
20
|
def run_resume_workflows() -> None:
|
|
21
21
|
start_process("task_resume_workflows")
|
|
@@ -0,0 +1,153 @@
|
|
|
1
|
+
# Copyright 2019-2020 SURF.
|
|
2
|
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
3
|
+
# you may not use this file except in compliance with the License.
|
|
4
|
+
# You may obtain a copy of the License at
|
|
5
|
+
#
|
|
6
|
+
# http://www.apache.org/licenses/LICENSE-2.0
|
|
7
|
+
#
|
|
8
|
+
# Unless required by applicable law or agreed to in writing, software
|
|
9
|
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
10
|
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
11
|
+
# See the License for the specific language governing permissions and
|
|
12
|
+
# limitations under the License.
|
|
13
|
+
|
|
14
|
+
|
|
15
|
+
from datetime import datetime
|
|
16
|
+
from typing import Any
|
|
17
|
+
|
|
18
|
+
from apscheduler.jobstores.sqlalchemy import SQLAlchemyJobStore
|
|
19
|
+
from apscheduler.schedulers.background import BackgroundScheduler
|
|
20
|
+
from more_itertools import partition
|
|
21
|
+
from pydantic import BaseModel
|
|
22
|
+
|
|
23
|
+
from orchestrator.db.filters import Filter
|
|
24
|
+
from orchestrator.db.filters.filters import CallableErrorHandler
|
|
25
|
+
from orchestrator.db.sorting import Sort
|
|
26
|
+
from orchestrator.db.sorting.sorting import SortOrder
|
|
27
|
+
from orchestrator.settings import app_settings
|
|
28
|
+
from orchestrator.utils.helpers import camel_to_snake, to_camel
|
|
29
|
+
|
|
30
|
+
jobstores = {"default": SQLAlchemyJobStore(url=str(app_settings.DATABASE_URI))}
|
|
31
|
+
|
|
32
|
+
scheduler = BackgroundScheduler(jobstores=jobstores)
|
|
33
|
+
|
|
34
|
+
|
|
35
|
+
def scheduler_dispose_db_connections() -> None:
|
|
36
|
+
jobstores["default"].engine.dispose()
|
|
37
|
+
|
|
38
|
+
|
|
39
|
+
class ScheduledTask(BaseModel):
|
|
40
|
+
id: str
|
|
41
|
+
name: str | None = None
|
|
42
|
+
next_run_time: datetime | None = None
|
|
43
|
+
trigger: str
|
|
44
|
+
|
|
45
|
+
|
|
46
|
+
scheduled_task_keys = set(ScheduledTask.model_fields.keys())
|
|
47
|
+
scheduled_task_filter_keys = sorted(scheduled_task_keys | {to_camel(key) for key in scheduled_task_keys})
|
|
48
|
+
scheduled_task_sort_keys = scheduled_task_filter_keys
|
|
49
|
+
|
|
50
|
+
|
|
51
|
+
def scheduled_task_in_filter(job: ScheduledTask, filter_by: list[Filter]) -> bool:
|
|
52
|
+
return any(f.value.lower() in getattr(job, camel_to_snake(f.field), "").lower() for f in filter_by)
|
|
53
|
+
|
|
54
|
+
|
|
55
|
+
def filter_scheduled_tasks(
|
|
56
|
+
scheduled_tasks: list[ScheduledTask],
|
|
57
|
+
handle_filter_error: CallableErrorHandler,
|
|
58
|
+
filter_by: list[Filter] | None = None,
|
|
59
|
+
) -> list[ScheduledTask]:
|
|
60
|
+
if not filter_by:
|
|
61
|
+
return scheduled_tasks
|
|
62
|
+
|
|
63
|
+
try:
|
|
64
|
+
invalid_filters, valid_filters = partition(lambda x: x.field in scheduled_task_filter_keys, filter_by)
|
|
65
|
+
|
|
66
|
+
if invalid_list := [item.field for item in invalid_filters]:
|
|
67
|
+
handle_filter_error(
|
|
68
|
+
"Invalid filter arguments", invalid_filters=invalid_list, valid_filter_keys=scheduled_task_filter_keys
|
|
69
|
+
)
|
|
70
|
+
|
|
71
|
+
valid_filter_list = list(valid_filters)
|
|
72
|
+
return [task for task in scheduled_tasks if scheduled_task_in_filter(task, valid_filter_list)]
|
|
73
|
+
except Exception as e:
|
|
74
|
+
handle_filter_error(str(e))
|
|
75
|
+
return []
|
|
76
|
+
|
|
77
|
+
|
|
78
|
+
def _invert(value: Any) -> Any:
|
|
79
|
+
"""Invert value for descending order."""
|
|
80
|
+
if isinstance(value, (int, float)):
|
|
81
|
+
return -value
|
|
82
|
+
if isinstance(value, str):
|
|
83
|
+
return tuple(-ord(c) for c in value)
|
|
84
|
+
if isinstance(value, datetime):
|
|
85
|
+
return -value.timestamp()
|
|
86
|
+
return value
|
|
87
|
+
|
|
88
|
+
|
|
89
|
+
def sort_key(sort_field: str, sort_order: SortOrder) -> Any:
|
|
90
|
+
def _sort_key(task: Any) -> Any:
|
|
91
|
+
value = getattr(task, camel_to_snake(sort_field), None)
|
|
92
|
+
if sort_field == "next_run_time" and value is None:
|
|
93
|
+
return float("inf") if sort_order == SortOrder.ASC else float("-inf")
|
|
94
|
+
return value if sort_order == SortOrder.ASC else _invert(value)
|
|
95
|
+
|
|
96
|
+
return _sort_key
|
|
97
|
+
|
|
98
|
+
|
|
99
|
+
def sort_scheduled_tasks(
|
|
100
|
+
scheduled_tasks: list[ScheduledTask], handle_sort_error: CallableErrorHandler, sort_by: list[Sort] | None = None
|
|
101
|
+
) -> list[ScheduledTask]:
|
|
102
|
+
if not sort_by:
|
|
103
|
+
return scheduled_tasks
|
|
104
|
+
|
|
105
|
+
try:
|
|
106
|
+
invalid_sorting, valid_sorting = partition(lambda x: x.field in scheduled_task_sort_keys, sort_by)
|
|
107
|
+
if invalid_list := [item.field for item in invalid_sorting]:
|
|
108
|
+
handle_sort_error(
|
|
109
|
+
"Invalid sort arguments", invalid_sorting=invalid_list, valid_sort_keys=scheduled_task_sort_keys
|
|
110
|
+
)
|
|
111
|
+
|
|
112
|
+
valid_sort_list = list(valid_sorting)
|
|
113
|
+
return sorted(
|
|
114
|
+
scheduled_tasks, key=lambda task: tuple(sort_key(sort.field, sort.order)(task) for sort in valid_sort_list)
|
|
115
|
+
)
|
|
116
|
+
except Exception as e:
|
|
117
|
+
handle_sort_error(str(e))
|
|
118
|
+
return []
|
|
119
|
+
|
|
120
|
+
|
|
121
|
+
def default_error_handler(message: str, **context) -> None: # type: ignore
|
|
122
|
+
from orchestrator.graphql.utils.create_resolver_error_handler import _format_context
|
|
123
|
+
|
|
124
|
+
raise ValueError(f"{message} {_format_context(context)}")
|
|
125
|
+
|
|
126
|
+
|
|
127
|
+
def get_scheduler_tasks(
|
|
128
|
+
first: int = 10,
|
|
129
|
+
after: int = 0,
|
|
130
|
+
filter_by: list[Filter] | None = None,
|
|
131
|
+
sort_by: list[Sort] | None = None,
|
|
132
|
+
error_handler: CallableErrorHandler = default_error_handler,
|
|
133
|
+
) -> tuple[list[ScheduledTask], int]:
|
|
134
|
+
scheduler.start(paused=True)
|
|
135
|
+
scheduled_tasks = scheduler.get_jobs()
|
|
136
|
+
scheduler.shutdown(wait=False)
|
|
137
|
+
scheduler_dispose_db_connections()
|
|
138
|
+
|
|
139
|
+
scheduled_tasks = filter_scheduled_tasks(scheduled_tasks, error_handler, filter_by)
|
|
140
|
+
scheduled_tasks = sort_scheduled_tasks(scheduled_tasks, error_handler, sort_by)
|
|
141
|
+
|
|
142
|
+
total = len(scheduled_tasks)
|
|
143
|
+
paginated_tasks = scheduled_tasks[after : after + first + 1]
|
|
144
|
+
|
|
145
|
+
return [
|
|
146
|
+
ScheduledTask(
|
|
147
|
+
id=task.id,
|
|
148
|
+
name=task.name,
|
|
149
|
+
next_run_time=task.next_run_time,
|
|
150
|
+
trigger=str(task.trigger),
|
|
151
|
+
)
|
|
152
|
+
for task in paginated_tasks
|
|
153
|
+
], total
|
|
@@ -12,37 +12,77 @@
|
|
|
12
12
|
# limitations under the License.
|
|
13
13
|
|
|
14
14
|
from collections.abc import Callable
|
|
15
|
-
from typing import
|
|
15
|
+
from typing import TypeVar
|
|
16
16
|
|
|
17
|
-
from
|
|
17
|
+
from apscheduler.schedulers.base import BaseScheduler
|
|
18
|
+
from deprecated import deprecated
|
|
18
19
|
|
|
20
|
+
from orchestrator.schedules.scheduler import scheduler as default_scheduler # your global scheduler instance
|
|
19
21
|
|
|
20
|
-
|
|
21
|
-
__name__: str
|
|
22
|
-
name: str
|
|
23
|
-
time_unit: str
|
|
24
|
-
period: int | None
|
|
25
|
-
at: str | None
|
|
26
|
-
|
|
27
|
-
def __call__(self) -> CancelJob | None: ...
|
|
22
|
+
F = TypeVar("F", bound=Callable[..., object])
|
|
28
23
|
|
|
29
24
|
|
|
25
|
+
@deprecated(
|
|
26
|
+
reason="We changed from scheduler to apscheduler which has its own decoractor, use `@scheduler.scheduled_job()` from `from orchestrator.scheduling.scheduler import scheduler`"
|
|
27
|
+
)
|
|
30
28
|
def scheduler(
|
|
31
|
-
name: str,
|
|
32
|
-
|
|
33
|
-
|
|
29
|
+
name: str,
|
|
30
|
+
time_unit: str,
|
|
31
|
+
period: int = 1,
|
|
32
|
+
at: str | None = None,
|
|
33
|
+
*,
|
|
34
|
+
id: str | None = None,
|
|
35
|
+
scheduler: BaseScheduler = default_scheduler,
|
|
36
|
+
) -> Callable[[F], F]:
|
|
37
|
+
"""APScheduler-compatible decorator to schedule a function.
|
|
38
|
+
|
|
39
|
+
id is necessary with apscheduler, if left empty it takes the function name.
|
|
34
40
|
|
|
35
|
-
|
|
36
|
-
time_unit = "
|
|
37
|
-
time_unit = "day", at="01:00" -> will run every day at 1 o'clock
|
|
41
|
+
- `time_unit = "hours", period = 12` → every 12 hours
|
|
42
|
+
- `time_unit = "day", at = "01:00"` → every day at 1 AM
|
|
38
43
|
"""
|
|
39
44
|
|
|
40
|
-
def
|
|
41
|
-
|
|
42
|
-
|
|
43
|
-
|
|
44
|
-
|
|
45
|
-
|
|
46
|
-
|
|
45
|
+
def decorator(func: F) -> F:
|
|
46
|
+
job_id = id or func.__name__
|
|
47
|
+
|
|
48
|
+
trigger = "interval"
|
|
49
|
+
kwargs: dict[str, int] = {}
|
|
50
|
+
if time_unit == "day" and at:
|
|
51
|
+
trigger = "cron"
|
|
52
|
+
try:
|
|
53
|
+
hour, minute = map(int, at.split(":"))
|
|
54
|
+
except ValueError:
|
|
55
|
+
raise ValueError(f"Invalid time format for 'at': {at}, expected 'HH:MM'")
|
|
56
|
+
|
|
57
|
+
kwargs = {
|
|
58
|
+
"hour": hour,
|
|
59
|
+
"minute": minute,
|
|
60
|
+
}
|
|
61
|
+
else:
|
|
62
|
+
# Map string units to timedelta kwargs for IntervalTrigger
|
|
63
|
+
unit_map = {
|
|
64
|
+
"seconds": "seconds",
|
|
65
|
+
"second": "seconds",
|
|
66
|
+
"minutes": "minutes",
|
|
67
|
+
"minute": "minutes",
|
|
68
|
+
"hours": "hours",
|
|
69
|
+
"hour": "hours",
|
|
70
|
+
"days": "days",
|
|
71
|
+
"day": "days",
|
|
72
|
+
}
|
|
73
|
+
|
|
74
|
+
interval_arg = unit_map.get(time_unit.lower(), time_unit.lower())
|
|
75
|
+
kwargs = {interval_arg: period}
|
|
76
|
+
|
|
77
|
+
scheduler.add_job(
|
|
78
|
+
func,
|
|
79
|
+
trigger=trigger,
|
|
80
|
+
id=job_id,
|
|
81
|
+
name=name,
|
|
82
|
+
replace_existing=True,
|
|
83
|
+
**kwargs,
|
|
84
|
+
)
|
|
85
|
+
|
|
86
|
+
return func
|
|
47
87
|
|
|
48
|
-
return
|
|
88
|
+
return decorator
|
|
@@ -12,10 +12,10 @@
|
|
|
12
12
|
# limitations under the License.
|
|
13
13
|
|
|
14
14
|
|
|
15
|
-
from orchestrator.schedules.
|
|
15
|
+
from orchestrator.schedules.scheduler import scheduler
|
|
16
16
|
from orchestrator.services.processes import start_process
|
|
17
17
|
|
|
18
18
|
|
|
19
|
-
@scheduler(name="Clean up tasks",
|
|
19
|
+
@scheduler.scheduled_job(id="clean-tasks", name="Clean up tasks", trigger="interval", hours=6) # type: ignore[misc]
|
|
20
20
|
def vacuum_tasks() -> None:
|
|
21
21
|
start_process("task_clean_up_tasks")
|
|
@@ -14,11 +14,17 @@ from sqlalchemy import func, select
|
|
|
14
14
|
|
|
15
15
|
from orchestrator.db import db
|
|
16
16
|
from orchestrator.db.models import ProcessTable
|
|
17
|
-
from orchestrator.schedules.
|
|
17
|
+
from orchestrator.schedules.scheduler import scheduler
|
|
18
18
|
from orchestrator.services.processes import start_process
|
|
19
19
|
|
|
20
20
|
|
|
21
|
-
@scheduler(
|
|
21
|
+
@scheduler.scheduled_job( # type: ignore[misc]
|
|
22
|
+
id="validate-products",
|
|
23
|
+
name="Validate Products and inactive subscriptions",
|
|
24
|
+
trigger="cron",
|
|
25
|
+
hour=2,
|
|
26
|
+
minute=30,
|
|
27
|
+
)
|
|
22
28
|
def validate_products() -> None:
|
|
23
29
|
uncompleted_products = db.session.scalar(
|
|
24
30
|
select(func.count())
|
|
@@ -16,7 +16,7 @@ from threading import BoundedSemaphore
|
|
|
16
16
|
|
|
17
17
|
import structlog
|
|
18
18
|
|
|
19
|
-
from orchestrator.schedules.
|
|
19
|
+
from orchestrator.schedules.scheduler import scheduler
|
|
20
20
|
from orchestrator.services.subscriptions import (
|
|
21
21
|
get_subscriptions_on_product_table,
|
|
22
22
|
get_subscriptions_on_product_table_in_sync,
|
|
@@ -33,7 +33,7 @@ logger = structlog.get_logger(__name__)
|
|
|
33
33
|
task_semaphore = BoundedSemaphore(value=2)
|
|
34
34
|
|
|
35
35
|
|
|
36
|
-
@scheduler(name="Subscriptions Validator",
|
|
36
|
+
@scheduler.scheduled_job(id="subscriptions-validator", name="Subscriptions Validator", trigger="cron", hour=0, minute=10) # type: ignore[misc]
|
|
37
37
|
def validate_subscriptions() -> None:
|
|
38
38
|
if app_settings.VALIDATE_OUT_OF_SYNC_SUBSCRIPTIONS:
|
|
39
39
|
# Automatically re-validate out-of-sync subscriptions. This is not recommended for production.
|
|
@@ -41,7 +41,7 @@ def retrieve_input_state(process_id: UUID, input_type: InputType, raise_exceptio
|
|
|
41
41
|
select(InputStateTable)
|
|
42
42
|
.filter(InputStateTable.process_id == process_id)
|
|
43
43
|
.filter(InputStateTable.input_type == input_type)
|
|
44
|
-
.order_by(InputStateTable.input_time.
|
|
44
|
+
.order_by(InputStateTable.input_time.desc())
|
|
45
45
|
).first()
|
|
46
46
|
|
|
47
47
|
if res:
|
|
@@ -733,11 +733,17 @@ def abort_process(process: ProcessTable, user: str, broadcast_func: Callable | N
|
|
|
733
733
|
|
|
734
734
|
|
|
735
735
|
def _recoverwf(wf: Workflow, log: list[WFProcess]) -> tuple[WFProcess, StepList]:
|
|
736
|
-
|
|
736
|
+
"""Recover workflow state and remaining steps from the given 'Process step' objects.
|
|
737
737
|
|
|
738
|
-
|
|
739
|
-
|
|
740
|
-
|
|
738
|
+
Returns:
|
|
739
|
+
- The state accumulated up until the last cleared (completed) step\
|
|
740
|
+
- The remaining steps to execute (including Failed, Suspended and (A)waiting steps)
|
|
741
|
+
"""
|
|
742
|
+
|
|
743
|
+
def is_cleared(p: WFProcess) -> bool:
|
|
744
|
+
return not (p.isfailed() or p.issuspend() or p.iswaiting() or p.isawaitingcallback())
|
|
745
|
+
|
|
746
|
+
persistent = [p for p in log if is_cleared(p)]
|
|
741
747
|
stepcount = len(persistent)
|
|
742
748
|
|
|
743
749
|
if log and (log[-1].issuspend() or log[-1].isawaitingcallback()):
|
|
@@ -760,15 +766,14 @@ def _recoverwf(wf: Workflow, log: list[WFProcess]) -> tuple[WFProcess, StepList]
|
|
|
760
766
|
|
|
761
767
|
|
|
762
768
|
def _restore_log(steps: list[ProcessStepTable]) -> list[WFProcess]:
|
|
763
|
-
|
|
764
|
-
for step in steps:
|
|
765
|
-
process = WFProcess.from_status(step.status, step.state)
|
|
769
|
+
"""Deserialize ProcessStepTable objects into foldable 'Process step' objects."""
|
|
766
770
|
|
|
767
|
-
|
|
768
|
-
|
|
771
|
+
def deserialize(step: ProcessStepTable) -> WFProcess:
|
|
772
|
+
if not (wf_process := WFProcess.from_status(step.status, step.state)):
|
|
773
|
+
raise ValueError(f"Unable to deserialize step from it's status {step.status}")
|
|
774
|
+
return wf_process
|
|
769
775
|
|
|
770
|
-
|
|
771
|
-
return result
|
|
776
|
+
return [deserialize(step) for step in steps]
|
|
772
777
|
|
|
773
778
|
|
|
774
779
|
def load_process(process: ProcessTable) -> ProcessStat:
|
orchestrator/workflow.py
CHANGED
|
@@ -552,7 +552,7 @@ class ProcessStat:
|
|
|
552
552
|
process_id: UUID
|
|
553
553
|
workflow: Workflow
|
|
554
554
|
state: Process
|
|
555
|
-
log: StepList
|
|
555
|
+
log: StepList # Remaining steps to execute
|
|
556
556
|
current_user: str
|
|
557
557
|
user_model: OIDCUserModel | None = None
|
|
558
558
|
|
|
@@ -597,6 +597,13 @@ class StepStatus(strEnum):
|
|
|
597
597
|
|
|
598
598
|
|
|
599
599
|
class Process(Generic[S]):
|
|
600
|
+
"""ADT base class.
|
|
601
|
+
|
|
602
|
+
This class defines an Algebraic Data Type - specifically a "sum type" - that defines the possible
|
|
603
|
+
variants of a Process. It encapsulates the state and allows to fold _instances_ of a process into
|
|
604
|
+
a single value. These instances correspond to subsequent steps of the process.
|
|
605
|
+
"""
|
|
606
|
+
|
|
600
607
|
def __init__(self, s: S):
|
|
601
608
|
self.s = s
|
|
602
609
|
|
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
Metadata-Version: 2.4
|
|
2
2
|
Name: orchestrator-core
|
|
3
|
-
Version: 4.
|
|
3
|
+
Version: 4.4.0rc1
|
|
4
4
|
Summary: This is the orchestrator workflow engine.
|
|
5
5
|
Author-email: SURF <automation-beheer@surf.nl>
|
|
6
6
|
Requires-Python: >=3.11,<3.14
|
|
@@ -32,6 +32,7 @@ Classifier: Typing :: Typed
|
|
|
32
32
|
License-File: LICENSE
|
|
33
33
|
Requires-Dist: alembic==1.16.1
|
|
34
34
|
Requires-Dist: anyio>=3.7.0
|
|
35
|
+
Requires-Dist: apscheduler>=3.11.0
|
|
35
36
|
Requires-Dist: click==8.*
|
|
36
37
|
Requires-Dist: deepmerge==2.0
|
|
37
38
|
Requires-Dist: deprecated>=1.2.18
|
|
@@ -52,7 +53,6 @@ Requires-Dist: python-dateutil==2.8.2
|
|
|
52
53
|
Requires-Dist: python-rapidjson>=1.18,<1.21
|
|
53
54
|
Requires-Dist: pytz==2025.2
|
|
54
55
|
Requires-Dist: redis==5.1.1
|
|
55
|
-
Requires-Dist: schedule==1.1.0
|
|
56
56
|
Requires-Dist: semver==3.0.4
|
|
57
57
|
Requires-Dist: sentry-sdk[fastapi]~=2.29.1
|
|
58
58
|
Requires-Dist: sqlalchemy==2.0.41
|
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
orchestrator/__init__.py,sha256=
|
|
1
|
+
orchestrator/__init__.py,sha256=2VCmBMCQvlduTSR0AOWiSAcfGIGfm7kGJxYEKzqiXew,1066
|
|
2
2
|
orchestrator/app.py,sha256=7UrXKjBKNSEaSSXAd5ww_RdMFhFqE4yvfj8faS2MzAA,12089
|
|
3
3
|
orchestrator/exception_handlers.py,sha256=UsW3dw8q0QQlNLcV359bIotah8DYjMsj2Ts1LfX4ClY,1268
|
|
4
4
|
orchestrator/log_config.py,sha256=1tPRX5q65e57a6a_zEii_PFK8SzWT0mnA5w2sKg4hh8,1853
|
|
@@ -8,7 +8,7 @@ orchestrator/settings.py,sha256=2Kgc6m3qUCcSM3Z_IVUeehfgO0QphMFkLrS0RC3sU-U,4365
|
|
|
8
8
|
orchestrator/targets.py,sha256=WizBgnp8hWX9YLFUIju7ewSubiwQqinCvyiYNcXHbHI,802
|
|
9
9
|
orchestrator/types.py,sha256=qzs7xx5AYRmKbpYRyJJP3wuDb0W0bcAzefCN0RWLAco,15459
|
|
10
10
|
orchestrator/version.py,sha256=b58e08lxs47wUNXv0jXFO_ykpksmytuzEXD4La4W-NQ,1366
|
|
11
|
-
orchestrator/workflow.py,sha256=
|
|
11
|
+
orchestrator/workflow.py,sha256=meDCPnyyX_n5PsMUaFy2wWb5EKNm1_ff7zRDBYrbcDg,45901
|
|
12
12
|
orchestrator/api/__init__.py,sha256=GyHNfEFCGKQwRiN6rQmvSRH2iYX7npjMZn97n8XzmLU,571
|
|
13
13
|
orchestrator/api/error_handling.py,sha256=YrPCxSa-DSa9KwqIMlXI-KGBGnbGIW5ukOPiikUH9E4,1502
|
|
14
14
|
orchestrator/api/helpers.py,sha256=s0QRHYw8AvEmlkmRhuEzz9xixaZKUF3YuPzUVHkcoXk,6933
|
|
@@ -17,7 +17,7 @@ orchestrator/api/api_v1/__init__.py,sha256=GyHNfEFCGKQwRiN6rQmvSRH2iYX7npjMZn97n
|
|
|
17
17
|
orchestrator/api/api_v1/api.py,sha256=m4iDktsSpzxUDaudkdgXeZ83a6B4wfc3pczQsa-Pb-8,2866
|
|
18
18
|
orchestrator/api/api_v1/endpoints/__init__.py,sha256=GyHNfEFCGKQwRiN6rQmvSRH2iYX7npjMZn97n8XzmLU,571
|
|
19
19
|
orchestrator/api/api_v1/endpoints/health.py,sha256=iaxs1XX1_250_gKNsspuULCV2GEMBjbtjsmfQTOvMAI,1284
|
|
20
|
-
orchestrator/api/api_v1/endpoints/processes.py,sha256=
|
|
20
|
+
orchestrator/api/api_v1/endpoints/processes.py,sha256=238Bydgj4ILNyMU_7j_Q7a0WGlfIvKv5ypP7lESU32w,16188
|
|
21
21
|
orchestrator/api/api_v1/endpoints/product_blocks.py,sha256=kZ6ywIOsS_S2qGq7RvZ4KzjvaS1LmwbGWR37AKRvWOw,2146
|
|
22
22
|
orchestrator/api/api_v1/endpoints/products.py,sha256=BfFtwu9dZXEQbtKxYj9icc73GKGvAGMR5ytyf41nQlQ,3081
|
|
23
23
|
orchestrator/api/api_v1/endpoints/resource_types.py,sha256=gGyuaDyOD0TAVoeFGaGmjDGnQ8eQQArOxKrrk4MaDzA,2145
|
|
@@ -36,7 +36,7 @@ orchestrator/cli/migrate_domain_models.py,sha256=WRXy_1OnziQwpsCFZXvjB30nDJtjj0i
|
|
|
36
36
|
orchestrator/cli/migrate_tasks.py,sha256=bju8XColjSZD0v3rS4kl-24dLr8En_H4-6enBmqd494,7255
|
|
37
37
|
orchestrator/cli/migrate_workflows.py,sha256=nxUpx0vgEIc_8aJrjAyrw3E9Dt8JmaamTts8oiQ4vHY,8923
|
|
38
38
|
orchestrator/cli/migration_helpers.py,sha256=C5tpkP5WEBr7G9S-1k1hgSI8ili6xd9Z5ygc9notaK0,4110
|
|
39
|
-
orchestrator/cli/scheduler.py,sha256=
|
|
39
|
+
orchestrator/cli/scheduler.py,sha256=jdBbgqE7bNUOevFe6gZS5C-SyDTOow8baNJKEBTkS0A,2293
|
|
40
40
|
orchestrator/cli/domain_gen_helpers/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
41
41
|
orchestrator/cli/domain_gen_helpers/fixed_input_helpers.py,sha256=uzpwsaau81hHSxNMOS9-o7kF-9_78R0f_UE0AvWooZQ,6775
|
|
42
42
|
orchestrator/cli/domain_gen_helpers/helpers.py,sha256=tIPxn8ezED_xYZxH7ZAtQLwkDc6RNmLZVxWAoJ3a9lw,4203
|
|
@@ -158,7 +158,7 @@ orchestrator/forms/validators/product_id.py,sha256=u5mURLT0pOhbFLdwvYcy2_2fXMt35
|
|
|
158
158
|
orchestrator/graphql/__init__.py,sha256=avq8Yg3Jr_9pJqh7ClyIAOX7YSg1eM_AWmt5C3FRYUY,1440
|
|
159
159
|
orchestrator/graphql/autoregistration.py,sha256=pF2jbMKG26MvYoMSa6ZpqpHjVks7_NvSRFymHTgmfjs,6342
|
|
160
160
|
orchestrator/graphql/pagination.py,sha256=iqVDn3GPZpiQhEydfwkBJLURY-X8wwUphS8Lkeg0BOc,2413
|
|
161
|
-
orchestrator/graphql/schema.py,sha256=
|
|
161
|
+
orchestrator/graphql/schema.py,sha256=dw4m4sM1ek2DscB8vINN6L8vVDE0h5GXclHGa8CiUJo,9537
|
|
162
162
|
orchestrator/graphql/types.py,sha256=_kHKMusrRPuRtF4wm42NsBzoFZ4egbu3ibMmhd2D6Fs,5432
|
|
163
163
|
orchestrator/graphql/extensions/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
164
164
|
orchestrator/graphql/extensions/model_cache.py,sha256=1uhMRjBs9eK7zJ1Y6P6BopX06822w2Yh9jliwYvG6yQ,1085
|
|
@@ -174,6 +174,7 @@ orchestrator/graphql/resolvers/process.py,sha256=Hqs1F7-gw0yO_ioHjh2eLAyxrK2WSuL
|
|
|
174
174
|
orchestrator/graphql/resolvers/product.py,sha256=uPBmYwMdau-zUqNjoDl-LDn927u3aCFW5JQ4A_it8q0,2772
|
|
175
175
|
orchestrator/graphql/resolvers/product_block.py,sha256=Ker1CpxGab5h2BZujOHHwRUj8W4uphRr3WSpQGk2PnI,2939
|
|
176
176
|
orchestrator/graphql/resolvers/resource_type.py,sha256=SREZXjkLYpuo4nCM8DqVeImIrZcP3xDiWr_gq4wWaxQ,2956
|
|
177
|
+
orchestrator/graphql/resolvers/scheduled_tasks.py,sha256=QsnesRrj8ESuS9vPKG9DXYcG2Wfj9m5LWGeZgmc6hu8,1640
|
|
177
178
|
orchestrator/graphql/resolvers/settings.py,sha256=xVYqxo-EWQ24F4hUHm9OZeN9vsqQXJzIJ1_HF4Ci9Cs,3777
|
|
178
179
|
orchestrator/graphql/resolvers/subscription.py,sha256=57niFv-JCro_wm0peJ5Ne04F2WIPuJ-Lx2h8yd9qubA,6541
|
|
179
180
|
orchestrator/graphql/resolvers/version.py,sha256=qgwe1msPOexeg3RHCscJ8s45vNfMhYh9ZKyCZ3MNw30,809
|
|
@@ -184,17 +185,18 @@ orchestrator/graphql/schemas/customer_description.py,sha256=fize71IMpkvk_rTzcqCY
|
|
|
184
185
|
orchestrator/graphql/schemas/errors.py,sha256=VRl-Zd1FHMnscyozhfxzqeEUZ0ERAWum_Y8YwjGxwmA,203
|
|
185
186
|
orchestrator/graphql/schemas/fixed_input.py,sha256=1yqYHADQRgHz8OIP7ObYsPFS-gmzfkCvEO0a-KKf7zI,513
|
|
186
187
|
orchestrator/graphql/schemas/helpers.py,sha256=Kpj4kIbmoKKN35bdgUSwQvGUIbeg7VJAVMEq65YS_ik,346
|
|
187
|
-
orchestrator/graphql/schemas/process.py,sha256=
|
|
188
|
+
orchestrator/graphql/schemas/process.py,sha256=g3noYh_USfnaK59fnoX2DI5tAf1PhdLMJGI_lA2xX1M,4966
|
|
188
189
|
orchestrator/graphql/schemas/product.py,sha256=vUCqcjrKBJj-VKSrMYPKzjmmxLMXL7alKTJ8UdUkhTg,4342
|
|
189
190
|
orchestrator/graphql/schemas/product_block.py,sha256=Qk9cbA6vm7ZPrhdgPHatKRuy6TytBmxSr97McEOxAu8,2860
|
|
190
191
|
orchestrator/graphql/schemas/resource_type.py,sha256=s5d_FwQXL2-Sc-IDUxTJun5qFQ4zOP4-XcHF9ql-t1g,898
|
|
192
|
+
orchestrator/graphql/schemas/scheduled_task.py,sha256=22Kb7r2pUoefz9AeiTYQ1t6YBYCmAs_zVp7FqNdN5HQ,194
|
|
191
193
|
orchestrator/graphql/schemas/settings.py,sha256=drhm5VcLmUbiYAk6WUSJcyJqjNM96E6GvpxVdPAobnA,999
|
|
192
194
|
orchestrator/graphql/schemas/strawberry_pydantic_patch.py,sha256=CjNUhTKdYmLiaem-WY_mzw4HASIeaZitxGF8pPocqVw,1602
|
|
193
195
|
orchestrator/graphql/schemas/subscription.py,sha256=hTA34C27kgLguH9V53173CxMKIWiQKh3vFzyJ2yBfE0,9918
|
|
194
196
|
orchestrator/graphql/schemas/version.py,sha256=HSzVg_y4Sjd5_H5rRUtu3FJKOG_8ifhvBNt_qjOtC-E,92
|
|
195
197
|
orchestrator/graphql/schemas/workflow.py,sha256=WLbegRNxOfvXg4kPYrO5KPBwtHmUofAr2pvZT2JsW1c,1761
|
|
196
198
|
orchestrator/graphql/utils/__init__.py,sha256=1JvenzEVW1CBa1sGVI9I8IWnnoXIkb1hneDqph9EEZY,524
|
|
197
|
-
orchestrator/graphql/utils/create_resolver_error_handler.py,sha256=
|
|
199
|
+
orchestrator/graphql/utils/create_resolver_error_handler.py,sha256=XzCnL482M4wz3fg5fUdGUwCAuzSZQ9Ufu1mscLyeoWU,1227
|
|
198
200
|
orchestrator/graphql/utils/get_query_loaders.py,sha256=abS_HJ7K9een78gMiGq3IhwGwxQXHvZygExe0h_t9ns,815
|
|
199
201
|
orchestrator/graphql/utils/get_selected_fields.py,sha256=0hBcQkU-7TNVO_KG-MmLItKm0O3gmbqoxXNkLHO-wHo,1002
|
|
200
202
|
orchestrator/graphql/utils/get_selected_paths.py,sha256=H0btESeOr3_VB7zy5Cx25OS0uzBcg2Y1I-arAmSOnsQ,1382
|
|
@@ -208,6 +210,7 @@ orchestrator/metrics/init.py,sha256=xBITvDjbNf-iabbBg0tAW8TPj6-wzr_MerOOqgDsoS4,
|
|
|
208
210
|
orchestrator/metrics/processes.py,sha256=SyogN5NSuhYoRv2CSUE1So9e8Gkrwa71J6oGLOdODQU,5333
|
|
209
211
|
orchestrator/metrics/subscriptions.py,sha256=vC1O8VmTq5oJxNrn5CU99Rf8cxzdyhc7tXbZBSAU-O8,3036
|
|
210
212
|
orchestrator/migrations/README,sha256=heMzebYwlGhnE8_4CWJ4LS74WoEZjBy-S-mIJRxAEKI,39
|
|
213
|
+
orchestrator/migrations/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
211
214
|
orchestrator/migrations/alembic.ini,sha256=kMoADqhGeubU8xanILNaqm4oixLy9m4ngYtdGpZcc7I,873
|
|
212
215
|
orchestrator/migrations/env.py,sha256=M_cPoAL2axuuup5fvMy8I_WTPHEw0RbPEHkhZ3QEGoE,3740
|
|
213
216
|
orchestrator/migrations/helpers.py,sha256=CAGGKhxpmhyKGfYcO-SUCPfMTOCZPfEpkJrcm2MYfcE,47979
|
|
@@ -244,12 +247,13 @@ orchestrator/migrations/versions/schema/2025-05-08_161918133bec_add_is_task_to_w
|
|
|
244
247
|
orchestrator/migrations/versions/schema/2025-07-01_93fc5834c7e5_changed_timestamping_fields_in_process_steps.py,sha256=Oezd8b2qaI1Kyq-sZFVFmdzd4d9NjXrf6HtJGk11fy0,1914
|
|
245
248
|
orchestrator/migrations/versions/schema/2025-07-04_4b58e336d1bf_deprecating_workflow_target_in_.py,sha256=xnD6w-97R4ClS7rbmXQEXc36K3fdcXKhCy7ZZNy_FX4,742
|
|
246
249
|
orchestrator/migrations/versions/schema/2025-07-28_850dccac3b02_update_description_of_resume_workflows_.py,sha256=R6Qoga83DJ1IL0WYPu0u5u2ZvAmqGlDmUMv_KtJyOhQ,812
|
|
247
|
-
orchestrator/schedules/__init__.py,sha256=
|
|
248
|
-
orchestrator/schedules/resume_workflows.py,sha256=
|
|
249
|
-
orchestrator/schedules/
|
|
250
|
-
orchestrator/schedules/
|
|
251
|
-
orchestrator/schedules/
|
|
252
|
-
orchestrator/schedules/
|
|
250
|
+
orchestrator/schedules/__init__.py,sha256=Zy0fTOBMGIRFoh5iVFDLF9_PRAFaONYDThGK9EsysWo,981
|
|
251
|
+
orchestrator/schedules/resume_workflows.py,sha256=jRnVRWDy687pQu-gtk80ecwiLSdrvtL15tG3U2zWA6I,891
|
|
252
|
+
orchestrator/schedules/scheduler.py,sha256=vze3xaZhUL5maKQB6a1gCvc9AcGw3jX-BHT3d5xvy6A,5430
|
|
253
|
+
orchestrator/schedules/scheduling.py,sha256=_mbpHMhijey8Y56ebtJ4wVkrp_kPVRm8hoByzlQF4SE,2821
|
|
254
|
+
orchestrator/schedules/task_vacuum.py,sha256=mxb7fsy1GphRwvUWi_lvwNaj51YAXUdIDlkOJd90AFI,874
|
|
255
|
+
orchestrator/schedules/validate_products.py,sha256=zWFQeVn3F8LP3joExLiKdmHs008pZsO-RolcIXHjFyE,1322
|
|
256
|
+
orchestrator/schedules/validate_subscriptions.py,sha256=bUBV45aEuqVdtqYBAXh1lX4O5vuNTeTfds4J_zq35dI,2113
|
|
253
257
|
orchestrator/schemas/__init__.py,sha256=YDyZ0YBvzB4ML9oDBCBPGnBvf680zFFgUzg7X0tYBRY,2326
|
|
254
258
|
orchestrator/schemas/base.py,sha256=Vc444LetsINLRhG2SxW9Bq01hOzChPOhQWCImQTr-As,930
|
|
255
259
|
orchestrator/schemas/engine_settings.py,sha256=LF8al7tJssiilb5A4emPtUYo0tVDSaT1Lvo_DN_ttrY,1296
|
|
@@ -264,9 +268,9 @@ orchestrator/schemas/subscription_descriptions.py,sha256=Ft_jw1U0bf9Z0U8O4OWfLlc
|
|
|
264
268
|
orchestrator/schemas/workflow.py,sha256=VqQ9XfV4fVd6MjY0LRRQzWBJHmlPsAamWfTwDx1cZkg,2102
|
|
265
269
|
orchestrator/services/__init__.py,sha256=GyHNfEFCGKQwRiN6rQmvSRH2iYX7npjMZn97n8XzmLU,571
|
|
266
270
|
orchestrator/services/fixed_inputs.py,sha256=kyz7s2HLzyDulvcq-ZqefTw1om86COvyvTjz0_5CmgI,876
|
|
267
|
-
orchestrator/services/input_state.py,sha256=
|
|
271
|
+
orchestrator/services/input_state.py,sha256=6BZOpb3cHpO18K-XG-3QUIV9pIM25_ufdODrp5CmXG4,2390
|
|
268
272
|
orchestrator/services/process_broadcast_thread.py,sha256=D44YbjF8mRqGuznkRUV4SoRn1J0lfy_x1H508GnSVlU,4649
|
|
269
|
-
orchestrator/services/processes.py,sha256=
|
|
273
|
+
orchestrator/services/processes.py,sha256=NfzdtH4eZK_wYuSmFtUX69qDvoeI8J7sJ2fFyY_VYaM,30544
|
|
270
274
|
orchestrator/services/products.py,sha256=BP4KyE8zO-8z7Trrs5T6zKBOw53S9BfBJnHWI3p6u5Y,1943
|
|
271
275
|
orchestrator/services/resource_types.py,sha256=_QBy_JOW_X3aSTqH0CuLrq4zBJL0p7Q-UDJUcuK2_qc,884
|
|
272
276
|
orchestrator/services/settings.py,sha256=HEWfFulgoEDwgfxGEO__QTr5fDiwNBEj1UhAeTAdbLQ,3159
|
|
@@ -315,7 +319,7 @@ orchestrator/workflows/tasks/resume_workflows.py,sha256=T3iobSJjVgiupe0rClD34kUZ
|
|
|
315
319
|
orchestrator/workflows/tasks/validate_product_type.py,sha256=paG-NAY1bdde3Adt8zItkcBKf5Pxw6f5ngGW6an6dYU,3192
|
|
316
320
|
orchestrator/workflows/tasks/validate_products.py,sha256=GZJBoFF-WMphS7ghMs2-gqvV2iL1F0POhk0uSNt93n0,8510
|
|
317
321
|
orchestrator/workflows/translations/en-GB.json,sha256=ST53HxkphFLTMjFHonykDBOZ7-P_KxksktZU3GbxLt0,846
|
|
318
|
-
orchestrator_core-4.
|
|
319
|
-
orchestrator_core-4.
|
|
320
|
-
orchestrator_core-4.
|
|
321
|
-
orchestrator_core-4.
|
|
322
|
+
orchestrator_core-4.4.0rc1.dist-info/licenses/LICENSE,sha256=b-aA5OZQuuBATmLKo_mln8CQrDPPhg3ghLzjPjLn4Tg,11409
|
|
323
|
+
orchestrator_core-4.4.0rc1.dist-info/WHEEL,sha256=G2gURzTEtmeR8nrdXUJfNiB3VYVxigPQ-bEQujpNiNs,82
|
|
324
|
+
orchestrator_core-4.4.0rc1.dist-info/METADATA,sha256=as5RXg0Y5DUGPcpT-6raY0HmvEeDJfEMsg1dIQfqoqU,5967
|
|
325
|
+
orchestrator_core-4.4.0rc1.dist-info/RECORD,,
|
|
File without changes
|
{orchestrator_core-4.3.0rc2.dist-info → orchestrator_core-4.4.0rc1.dist-info}/licenses/LICENSE
RENAMED
|
File without changes
|