fractal-server 2.11.1__py3-none-any.whl → 2.12.0a1__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- fractal_server/__init__.py +1 -1
- fractal_server/app/models/__init__.py +0 -2
- fractal_server/app/models/linkuserproject.py +0 -9
- fractal_server/app/routes/aux/_job.py +1 -3
- fractal_server/app/runner/executors/slurm/ssh/executor.py +9 -6
- fractal_server/app/runner/executors/slurm/sudo/executor.py +1 -5
- fractal_server/app/runner/filenames.py +0 -2
- fractal_server/app/runner/shutdown.py +3 -27
- fractal_server/app/schemas/_validators.py +0 -19
- fractal_server/config.py +1 -15
- fractal_server/main.py +1 -12
- fractal_server/migrations/versions/1eac13a26c83_drop_v1_tables.py +67 -0
- fractal_server/string_tools.py +0 -21
- fractal_server/tasks/utils.py +0 -28
- {fractal_server-2.11.1.dist-info → fractal_server-2.12.0a1.dist-info}/METADATA +1 -1
- {fractal_server-2.11.1.dist-info → fractal_server-2.12.0a1.dist-info}/RECORD +19 -63
- fractal_server/app/models/v1/__init__.py +0 -13
- fractal_server/app/models/v1/dataset.py +0 -71
- fractal_server/app/models/v1/job.py +0 -101
- fractal_server/app/models/v1/project.py +0 -29
- fractal_server/app/models/v1/state.py +0 -34
- fractal_server/app/models/v1/task.py +0 -85
- fractal_server/app/models/v1/workflow.py +0 -133
- fractal_server/app/routes/admin/v1.py +0 -377
- fractal_server/app/routes/api/v1/__init__.py +0 -26
- fractal_server/app/routes/api/v1/_aux_functions.py +0 -478
- fractal_server/app/routes/api/v1/dataset.py +0 -554
- fractal_server/app/routes/api/v1/job.py +0 -195
- fractal_server/app/routes/api/v1/project.py +0 -475
- fractal_server/app/routes/api/v1/task.py +0 -203
- fractal_server/app/routes/api/v1/task_collection.py +0 -239
- fractal_server/app/routes/api/v1/workflow.py +0 -355
- fractal_server/app/routes/api/v1/workflowtask.py +0 -187
- fractal_server/app/runner/async_wrap_v1.py +0 -27
- fractal_server/app/runner/v1/__init__.py +0 -415
- fractal_server/app/runner/v1/_common.py +0 -620
- fractal_server/app/runner/v1/_local/__init__.py +0 -186
- fractal_server/app/runner/v1/_local/_local_config.py +0 -105
- fractal_server/app/runner/v1/_local/_submit_setup.py +0 -48
- fractal_server/app/runner/v1/_local/executor.py +0 -100
- fractal_server/app/runner/v1/_slurm/__init__.py +0 -312
- fractal_server/app/runner/v1/_slurm/_submit_setup.py +0 -81
- fractal_server/app/runner/v1/_slurm/get_slurm_config.py +0 -163
- fractal_server/app/runner/v1/common.py +0 -117
- fractal_server/app/runner/v1/handle_failed_job.py +0 -141
- fractal_server/app/schemas/v1/__init__.py +0 -37
- fractal_server/app/schemas/v1/applyworkflow.py +0 -161
- fractal_server/app/schemas/v1/dataset.py +0 -165
- fractal_server/app/schemas/v1/dumps.py +0 -64
- fractal_server/app/schemas/v1/manifest.py +0 -126
- fractal_server/app/schemas/v1/project.py +0 -66
- fractal_server/app/schemas/v1/state.py +0 -18
- fractal_server/app/schemas/v1/task.py +0 -167
- fractal_server/app/schemas/v1/task_collection.py +0 -110
- fractal_server/app/schemas/v1/workflow.py +0 -212
- fractal_server/tasks/v1/_TaskCollectPip.py +0 -103
- fractal_server/tasks/v1/__init__.py +0 -0
- fractal_server/tasks/v1/background_operations.py +0 -352
- fractal_server/tasks/v1/endpoint_operations.py +0 -156
- fractal_server/tasks/v1/get_collection_data.py +0 -14
- fractal_server/tasks/v1/utils.py +0 -67
- {fractal_server-2.11.1.dist-info → fractal_server-2.12.0a1.dist-info}/LICENSE +0 -0
- {fractal_server-2.11.1.dist-info → fractal_server-2.12.0a1.dist-info}/WHEEL +0 -0
- {fractal_server-2.11.1.dist-info → fractal_server-2.12.0a1.dist-info}/entry_points.txt +0 -0
@@ -1,355 +0,0 @@
|
|
1
|
-
# Copyright 2022 (C) Friedrich Miescher Institute for Biomedical Research and
|
2
|
-
# University of Zurich
|
3
|
-
#
|
4
|
-
# Original author(s):
|
5
|
-
# Jacopo Nespolo <jacopo.nespolo@exact-lab.it>
|
6
|
-
# Marco Franzon <marco.franzon@exact-lab.it>
|
7
|
-
# Tommaso Comparin <tommaso.comparin@exact-lab.it>
|
8
|
-
#
|
9
|
-
# This file is part of Fractal and was originally developed by eXact lab S.r.l.
|
10
|
-
# <exact-lab.it> under contract with Liberali Lab from the Friedrich Miescher
|
11
|
-
# Institute for Biomedical Research and Pelkmans Lab from the University of
|
12
|
-
# Zurich.
|
13
|
-
from typing import Optional
|
14
|
-
|
15
|
-
from fastapi import APIRouter
|
16
|
-
from fastapi import Depends
|
17
|
-
from fastapi import HTTPException
|
18
|
-
from fastapi import Response
|
19
|
-
from fastapi import status
|
20
|
-
from sqlmodel import select
|
21
|
-
|
22
|
-
from .....logger import close_logger
|
23
|
-
from .....logger import set_logger
|
24
|
-
from ....db import AsyncSession
|
25
|
-
from ....db import get_async_db
|
26
|
-
from ....models.v1 import ApplyWorkflow
|
27
|
-
from ....models.v1 import Project
|
28
|
-
from ....models.v1 import Task
|
29
|
-
from ....models.v1 import Workflow
|
30
|
-
from ....schemas.v1 import WorkflowCreateV1
|
31
|
-
from ....schemas.v1 import WorkflowExportV1
|
32
|
-
from ....schemas.v1 import WorkflowImportV1
|
33
|
-
from ....schemas.v1 import WorkflowReadV1
|
34
|
-
from ....schemas.v1 import WorkflowTaskCreateV1
|
35
|
-
from ....schemas.v1 import WorkflowUpdateV1
|
36
|
-
from ._aux_functions import _check_workflow_exists
|
37
|
-
from ._aux_functions import _get_project_check_owner
|
38
|
-
from ._aux_functions import _get_submitted_jobs_statement
|
39
|
-
from ._aux_functions import _get_workflow_check_owner
|
40
|
-
from ._aux_functions import _raise_if_v1_is_read_only
|
41
|
-
from ._aux_functions import _workflow_insert_task
|
42
|
-
from fractal_server.app.models import UserOAuth
|
43
|
-
from fractal_server.app.routes.auth import current_active_user
|
44
|
-
|
45
|
-
|
46
|
-
router = APIRouter()
|
47
|
-
|
48
|
-
|
49
|
-
@router.get(
|
50
|
-
"/project/{project_id}/workflow/",
|
51
|
-
response_model=list[WorkflowReadV1],
|
52
|
-
)
|
53
|
-
async def get_workflow_list(
|
54
|
-
project_id: int,
|
55
|
-
user: UserOAuth = Depends(current_active_user),
|
56
|
-
db: AsyncSession = Depends(get_async_db),
|
57
|
-
) -> Optional[list[WorkflowReadV1]]:
|
58
|
-
"""
|
59
|
-
Get workflow list for given project
|
60
|
-
"""
|
61
|
-
# Access control
|
62
|
-
project = await _get_project_check_owner(
|
63
|
-
project_id=project_id, user_id=user.id, db=db
|
64
|
-
)
|
65
|
-
# Find workflows of the current project. Note: this select/where approach
|
66
|
-
# has much better scaling than refreshing all elements of
|
67
|
-
# `project.workflow_list` - ref
|
68
|
-
# https://github.com/fractal-analytics-platform/fractal-server/pull/1082#issuecomment-1856676097.
|
69
|
-
stm = select(Workflow).where(Workflow.project_id == project.id)
|
70
|
-
workflow_list = (await db.execute(stm)).scalars().all()
|
71
|
-
return workflow_list
|
72
|
-
|
73
|
-
|
74
|
-
@router.post(
|
75
|
-
"/project/{project_id}/workflow/",
|
76
|
-
response_model=WorkflowReadV1,
|
77
|
-
status_code=status.HTTP_201_CREATED,
|
78
|
-
)
|
79
|
-
async def create_workflow(
|
80
|
-
project_id: int,
|
81
|
-
workflow: WorkflowCreateV1,
|
82
|
-
user: UserOAuth = Depends(current_active_user),
|
83
|
-
db: AsyncSession = Depends(get_async_db),
|
84
|
-
) -> Optional[WorkflowReadV1]:
|
85
|
-
"""
|
86
|
-
Create a workflow, associate to a project
|
87
|
-
"""
|
88
|
-
_raise_if_v1_is_read_only()
|
89
|
-
await _get_project_check_owner(
|
90
|
-
project_id=project_id,
|
91
|
-
user_id=user.id,
|
92
|
-
db=db,
|
93
|
-
)
|
94
|
-
await _check_workflow_exists(
|
95
|
-
name=workflow.name, project_id=project_id, db=db
|
96
|
-
)
|
97
|
-
|
98
|
-
db_workflow = Workflow(project_id=project_id, **workflow.dict())
|
99
|
-
db.add(db_workflow)
|
100
|
-
await db.commit()
|
101
|
-
await db.refresh(db_workflow)
|
102
|
-
await db.close()
|
103
|
-
return db_workflow
|
104
|
-
|
105
|
-
|
106
|
-
@router.get(
|
107
|
-
"/project/{project_id}/workflow/{workflow_id}/",
|
108
|
-
response_model=WorkflowReadV1,
|
109
|
-
)
|
110
|
-
async def read_workflow(
|
111
|
-
project_id: int,
|
112
|
-
workflow_id: int,
|
113
|
-
user: UserOAuth = Depends(current_active_user),
|
114
|
-
db: AsyncSession = Depends(get_async_db),
|
115
|
-
) -> Optional[WorkflowReadV1]:
|
116
|
-
"""
|
117
|
-
Get info on an existing workflow
|
118
|
-
"""
|
119
|
-
|
120
|
-
workflow = await _get_workflow_check_owner(
|
121
|
-
project_id=project_id, workflow_id=workflow_id, user_id=user.id, db=db
|
122
|
-
)
|
123
|
-
|
124
|
-
return workflow
|
125
|
-
|
126
|
-
|
127
|
-
@router.patch(
|
128
|
-
"/project/{project_id}/workflow/{workflow_id}/",
|
129
|
-
response_model=WorkflowReadV1,
|
130
|
-
)
|
131
|
-
async def update_workflow(
|
132
|
-
project_id: int,
|
133
|
-
workflow_id: int,
|
134
|
-
patch: WorkflowUpdateV1,
|
135
|
-
user: UserOAuth = Depends(current_active_user),
|
136
|
-
db: AsyncSession = Depends(get_async_db),
|
137
|
-
) -> Optional[WorkflowReadV1]:
|
138
|
-
"""
|
139
|
-
Edit a workflow
|
140
|
-
"""
|
141
|
-
_raise_if_v1_is_read_only()
|
142
|
-
workflow = await _get_workflow_check_owner(
|
143
|
-
project_id=project_id, workflow_id=workflow_id, user_id=user.id, db=db
|
144
|
-
)
|
145
|
-
|
146
|
-
if patch.name:
|
147
|
-
await _check_workflow_exists(
|
148
|
-
name=patch.name, project_id=project_id, db=db
|
149
|
-
)
|
150
|
-
|
151
|
-
for key, value in patch.dict(exclude_unset=True).items():
|
152
|
-
if key == "reordered_workflowtask_ids":
|
153
|
-
current_workflowtask_ids = [
|
154
|
-
wftask.id for wftask in workflow.task_list
|
155
|
-
]
|
156
|
-
num_tasks = len(workflow.task_list)
|
157
|
-
if len(value) != num_tasks or set(value) != set(
|
158
|
-
current_workflowtask_ids
|
159
|
-
):
|
160
|
-
raise HTTPException(
|
161
|
-
status_code=status.HTTP_422_UNPROCESSABLE_ENTITY,
|
162
|
-
detail=(
|
163
|
-
"`reordered_workflowtask_ids` must be a permutation of"
|
164
|
-
f" {current_workflowtask_ids} (given {value})"
|
165
|
-
),
|
166
|
-
)
|
167
|
-
for ind_wftask in range(num_tasks):
|
168
|
-
new_order = value.index(workflow.task_list[ind_wftask].id)
|
169
|
-
workflow.task_list[ind_wftask].order = new_order
|
170
|
-
else:
|
171
|
-
setattr(workflow, key, value)
|
172
|
-
|
173
|
-
await db.commit()
|
174
|
-
await db.refresh(workflow)
|
175
|
-
await db.close()
|
176
|
-
|
177
|
-
return workflow
|
178
|
-
|
179
|
-
|
180
|
-
@router.delete(
|
181
|
-
"/project/{project_id}/workflow/{workflow_id}/",
|
182
|
-
status_code=status.HTTP_204_NO_CONTENT,
|
183
|
-
)
|
184
|
-
async def delete_workflow(
|
185
|
-
project_id: int,
|
186
|
-
workflow_id: int,
|
187
|
-
user: UserOAuth = Depends(current_active_user),
|
188
|
-
db: AsyncSession = Depends(get_async_db),
|
189
|
-
) -> Response:
|
190
|
-
"""
|
191
|
-
Delete a workflow
|
192
|
-
"""
|
193
|
-
_raise_if_v1_is_read_only()
|
194
|
-
workflow = await _get_workflow_check_owner(
|
195
|
-
project_id=project_id, workflow_id=workflow_id, user_id=user.id, db=db
|
196
|
-
)
|
197
|
-
|
198
|
-
# Fail if there exist jobs that are submitted and in relation with the
|
199
|
-
# current workflow.
|
200
|
-
stm = _get_submitted_jobs_statement().where(
|
201
|
-
ApplyWorkflow.workflow_id == workflow.id
|
202
|
-
)
|
203
|
-
res = await db.execute(stm)
|
204
|
-
jobs = res.scalars().all()
|
205
|
-
if jobs:
|
206
|
-
string_ids = str([job.id for job in jobs])[1:-1]
|
207
|
-
raise HTTPException(
|
208
|
-
status_code=status.HTTP_422_UNPROCESSABLE_ENTITY,
|
209
|
-
detail=(
|
210
|
-
f"Cannot delete workflow {workflow.id} because it "
|
211
|
-
f"is linked to active job(s) {string_ids}."
|
212
|
-
),
|
213
|
-
)
|
214
|
-
|
215
|
-
# Cascade operations: set foreign-keys to null for jobs which are in
|
216
|
-
# relationship with the current workflow
|
217
|
-
stm = select(ApplyWorkflow).where(ApplyWorkflow.workflow_id == workflow_id)
|
218
|
-
res = await db.execute(stm)
|
219
|
-
jobs = res.scalars().all()
|
220
|
-
for job in jobs:
|
221
|
-
job.workflow_id = None
|
222
|
-
await db.merge(job)
|
223
|
-
await db.commit()
|
224
|
-
|
225
|
-
# Delete workflow
|
226
|
-
await db.delete(workflow)
|
227
|
-
await db.commit()
|
228
|
-
|
229
|
-
return Response(status_code=status.HTTP_204_NO_CONTENT)
|
230
|
-
|
231
|
-
|
232
|
-
@router.get(
|
233
|
-
"/project/{project_id}/workflow/{workflow_id}/export/",
|
234
|
-
response_model=WorkflowExportV1,
|
235
|
-
)
|
236
|
-
async def export_worfklow(
|
237
|
-
project_id: int,
|
238
|
-
workflow_id: int,
|
239
|
-
user: UserOAuth = Depends(current_active_user),
|
240
|
-
db: AsyncSession = Depends(get_async_db),
|
241
|
-
) -> Optional[WorkflowExportV1]:
|
242
|
-
"""
|
243
|
-
Export an existing workflow, after stripping all IDs
|
244
|
-
"""
|
245
|
-
workflow = await _get_workflow_check_owner(
|
246
|
-
project_id=project_id, workflow_id=workflow_id, user_id=user.id, db=db
|
247
|
-
)
|
248
|
-
# Emit a warning when exporting a workflow with custom tasks
|
249
|
-
logger = set_logger(None)
|
250
|
-
for wftask in workflow.task_list:
|
251
|
-
if wftask.task.owner is not None:
|
252
|
-
logger.warning(
|
253
|
-
f"Custom tasks (like the one with id={wftask.task.id} and "
|
254
|
-
f'source="{wftask.task.source}") are not meant to be '
|
255
|
-
"portable; re-importing this workflow may not work as "
|
256
|
-
"expected."
|
257
|
-
)
|
258
|
-
close_logger(logger)
|
259
|
-
|
260
|
-
await db.close()
|
261
|
-
return workflow
|
262
|
-
|
263
|
-
|
264
|
-
@router.post(
|
265
|
-
"/project/{project_id}/workflow/import/",
|
266
|
-
response_model=WorkflowReadV1,
|
267
|
-
status_code=status.HTTP_201_CREATED,
|
268
|
-
)
|
269
|
-
async def import_workflow(
|
270
|
-
project_id: int,
|
271
|
-
workflow: WorkflowImportV1,
|
272
|
-
user: UserOAuth = Depends(current_active_user),
|
273
|
-
db: AsyncSession = Depends(get_async_db),
|
274
|
-
) -> Optional[WorkflowReadV1]:
|
275
|
-
"""
|
276
|
-
Import an existing workflow into a project
|
277
|
-
|
278
|
-
Also create all required objects (i.e. Workflow and WorkflowTask's) along
|
279
|
-
the way.
|
280
|
-
"""
|
281
|
-
_raise_if_v1_is_read_only()
|
282
|
-
# Preliminary checks
|
283
|
-
await _get_project_check_owner(
|
284
|
-
project_id=project_id,
|
285
|
-
user_id=user.id,
|
286
|
-
db=db,
|
287
|
-
)
|
288
|
-
|
289
|
-
await _check_workflow_exists(
|
290
|
-
name=workflow.name, project_id=project_id, db=db
|
291
|
-
)
|
292
|
-
|
293
|
-
# Check that all required tasks are available
|
294
|
-
tasks = [wf_task.task for wf_task in workflow.task_list]
|
295
|
-
source_to_id = {}
|
296
|
-
for task in tasks:
|
297
|
-
source = task.source
|
298
|
-
if source not in source_to_id.keys():
|
299
|
-
stm = select(Task).where(Task.source == source)
|
300
|
-
tasks_by_source = (await db.execute(stm)).scalars().all()
|
301
|
-
if len(tasks_by_source) != 1:
|
302
|
-
raise HTTPException(
|
303
|
-
status_code=status.HTTP_422_UNPROCESSABLE_ENTITY,
|
304
|
-
detail=(
|
305
|
-
f"Found {len(tasks_by_source)} tasks with {source=}."
|
306
|
-
),
|
307
|
-
)
|
308
|
-
source_to_id[source] = tasks_by_source[0].id
|
309
|
-
|
310
|
-
# Create new Workflow (with empty task_list)
|
311
|
-
db_workflow = Workflow(
|
312
|
-
project_id=project_id,
|
313
|
-
**workflow.dict(exclude_none=True, exclude={"task_list"}),
|
314
|
-
)
|
315
|
-
db.add(db_workflow)
|
316
|
-
await db.commit()
|
317
|
-
await db.refresh(db_workflow)
|
318
|
-
|
319
|
-
# Insert tasks
|
320
|
-
async with db:
|
321
|
-
for _, wf_task in enumerate(workflow.task_list):
|
322
|
-
# Identify task_id
|
323
|
-
source = wf_task.task.source
|
324
|
-
task_id = source_to_id[source]
|
325
|
-
# Prepare new_wf_task
|
326
|
-
new_wf_task = WorkflowTaskCreateV1(
|
327
|
-
**wf_task.dict(exclude_none=True),
|
328
|
-
)
|
329
|
-
# Insert task
|
330
|
-
await _workflow_insert_task(
|
331
|
-
**new_wf_task.dict(),
|
332
|
-
workflow_id=db_workflow.id,
|
333
|
-
task_id=task_id,
|
334
|
-
db=db,
|
335
|
-
)
|
336
|
-
|
337
|
-
await db.close()
|
338
|
-
return db_workflow
|
339
|
-
|
340
|
-
|
341
|
-
@router.get("/workflow/", response_model=list[WorkflowReadV1])
|
342
|
-
async def get_user_workflows(
|
343
|
-
user: UserOAuth = Depends(current_active_user),
|
344
|
-
db: AsyncSession = Depends(get_async_db),
|
345
|
-
) -> list[WorkflowReadV1]:
|
346
|
-
"""
|
347
|
-
Returns all the workflows of the current user
|
348
|
-
"""
|
349
|
-
stm = select(Workflow)
|
350
|
-
stm = stm.join(Project).where(
|
351
|
-
Project.user_list.any(UserOAuth.id == user.id)
|
352
|
-
)
|
353
|
-
res = await db.execute(stm)
|
354
|
-
workflow_list = res.scalars().all()
|
355
|
-
return workflow_list
|
@@ -1,187 +0,0 @@
|
|
1
|
-
# Copyright 2022 (C) Friedrich Miescher Institute for Biomedical Research and
|
2
|
-
# University of Zurich
|
3
|
-
#
|
4
|
-
# Original author(s):
|
5
|
-
# Jacopo Nespolo <jacopo.nespolo@exact-lab.it>
|
6
|
-
# Marco Franzon <marco.franzon@exact-lab.it>
|
7
|
-
# Tommaso Comparin <tommaso.comparin@exact-lab.it>
|
8
|
-
# Yuri Chiucconi <yuri.chiucconi@exact-lab.it>
|
9
|
-
#
|
10
|
-
# This file is part of Fractal and was originally developed by eXact lab S.r.l.
|
11
|
-
# <exact-lab.it> under contract with Liberali Lab from the Friedrich Miescher
|
12
|
-
# Institute for Biomedical Research and Pelkmans Lab from the University of
|
13
|
-
# Zurich.
|
14
|
-
from copy import deepcopy
|
15
|
-
from typing import Optional
|
16
|
-
|
17
|
-
from fastapi import APIRouter
|
18
|
-
from fastapi import Depends
|
19
|
-
from fastapi import HTTPException
|
20
|
-
from fastapi import Response
|
21
|
-
from fastapi import status
|
22
|
-
|
23
|
-
from ....db import AsyncSession
|
24
|
-
from ....db import get_async_db
|
25
|
-
from ....models.v1 import Task
|
26
|
-
from ....schemas.v1 import WorkflowTaskCreateV1
|
27
|
-
from ....schemas.v1 import WorkflowTaskReadV1
|
28
|
-
from ....schemas.v1 import WorkflowTaskUpdateV1
|
29
|
-
from ._aux_functions import _get_workflow_check_owner
|
30
|
-
from ._aux_functions import _get_workflow_task_check_owner
|
31
|
-
from ._aux_functions import _raise_if_v1_is_read_only
|
32
|
-
from ._aux_functions import _workflow_insert_task
|
33
|
-
from fractal_server.app.models import UserOAuth
|
34
|
-
from fractal_server.app.routes.auth import current_active_user
|
35
|
-
|
36
|
-
router = APIRouter()
|
37
|
-
|
38
|
-
|
39
|
-
@router.post(
|
40
|
-
"/project/{project_id}/workflow/{workflow_id}/wftask/",
|
41
|
-
response_model=WorkflowTaskReadV1,
|
42
|
-
status_code=status.HTTP_201_CREATED,
|
43
|
-
)
|
44
|
-
async def create_workflowtask(
|
45
|
-
project_id: int,
|
46
|
-
workflow_id: int,
|
47
|
-
task_id: int,
|
48
|
-
new_task: WorkflowTaskCreateV1,
|
49
|
-
user: UserOAuth = Depends(current_active_user),
|
50
|
-
db: AsyncSession = Depends(get_async_db),
|
51
|
-
) -> Optional[WorkflowTaskReadV1]:
|
52
|
-
"""
|
53
|
-
Add a WorkflowTask to a Workflow
|
54
|
-
"""
|
55
|
-
_raise_if_v1_is_read_only()
|
56
|
-
workflow = await _get_workflow_check_owner(
|
57
|
-
project_id=project_id, workflow_id=workflow_id, user_id=user.id, db=db
|
58
|
-
)
|
59
|
-
|
60
|
-
# Check that task exists
|
61
|
-
task = await db.get(Task, task_id)
|
62
|
-
if not task:
|
63
|
-
raise HTTPException(
|
64
|
-
status_code=status.HTTP_404_NOT_FOUND,
|
65
|
-
detail=f"Task {task_id} not found.",
|
66
|
-
)
|
67
|
-
|
68
|
-
async with db:
|
69
|
-
workflow_task = await _workflow_insert_task(
|
70
|
-
**new_task.dict(),
|
71
|
-
workflow_id=workflow.id,
|
72
|
-
task_id=task_id,
|
73
|
-
db=db,
|
74
|
-
)
|
75
|
-
|
76
|
-
await db.close()
|
77
|
-
return workflow_task
|
78
|
-
|
79
|
-
|
80
|
-
@router.get(
|
81
|
-
"/project/{project_id}/workflow/{workflow_id}/wftask/{workflow_task_id}/",
|
82
|
-
response_model=WorkflowTaskReadV1,
|
83
|
-
)
|
84
|
-
async def read_workflowtask(
|
85
|
-
project_id: int,
|
86
|
-
workflow_id: int,
|
87
|
-
workflow_task_id: int,
|
88
|
-
user: UserOAuth = Depends(current_active_user),
|
89
|
-
db: AsyncSession = Depends(get_async_db),
|
90
|
-
):
|
91
|
-
workflow_task, _ = await _get_workflow_task_check_owner(
|
92
|
-
project_id=project_id,
|
93
|
-
workflow_task_id=workflow_task_id,
|
94
|
-
workflow_id=workflow_id,
|
95
|
-
user_id=user.id,
|
96
|
-
db=db,
|
97
|
-
)
|
98
|
-
return workflow_task
|
99
|
-
|
100
|
-
|
101
|
-
@router.patch(
|
102
|
-
"/project/{project_id}/workflow/{workflow_id}/wftask/{workflow_task_id}/",
|
103
|
-
response_model=WorkflowTaskReadV1,
|
104
|
-
)
|
105
|
-
async def update_workflowtask(
|
106
|
-
project_id: int,
|
107
|
-
workflow_id: int,
|
108
|
-
workflow_task_id: int,
|
109
|
-
workflow_task_update: WorkflowTaskUpdateV1,
|
110
|
-
user: UserOAuth = Depends(current_active_user),
|
111
|
-
db: AsyncSession = Depends(get_async_db),
|
112
|
-
) -> Optional[WorkflowTaskReadV1]:
|
113
|
-
"""
|
114
|
-
Edit a WorkflowTask of a Workflow
|
115
|
-
"""
|
116
|
-
_raise_if_v1_is_read_only()
|
117
|
-
db_workflow_task, db_workflow = await _get_workflow_task_check_owner(
|
118
|
-
project_id=project_id,
|
119
|
-
workflow_task_id=workflow_task_id,
|
120
|
-
workflow_id=workflow_id,
|
121
|
-
user_id=user.id,
|
122
|
-
db=db,
|
123
|
-
)
|
124
|
-
|
125
|
-
for key, value in workflow_task_update.dict(exclude_unset=True).items():
|
126
|
-
if key == "args":
|
127
|
-
|
128
|
-
# Get default arguments via a Task property method
|
129
|
-
default_args = deepcopy(
|
130
|
-
db_workflow_task.task.default_args_from_args_schema
|
131
|
-
)
|
132
|
-
# Override default_args with args value items
|
133
|
-
actual_args = default_args.copy()
|
134
|
-
if value is not None:
|
135
|
-
for k, v in value.items():
|
136
|
-
actual_args[k] = v
|
137
|
-
if not actual_args:
|
138
|
-
actual_args = None
|
139
|
-
setattr(db_workflow_task, key, actual_args)
|
140
|
-
elif key == "meta":
|
141
|
-
current_meta = deepcopy(db_workflow_task.meta) or {}
|
142
|
-
current_meta.update(value)
|
143
|
-
setattr(db_workflow_task, key, current_meta)
|
144
|
-
else:
|
145
|
-
raise HTTPException(
|
146
|
-
status_code=status.HTTP_422_UNPROCESSABLE_ENTITY,
|
147
|
-
detail=f"patch_workflow_task endpoint cannot set {key=}",
|
148
|
-
)
|
149
|
-
|
150
|
-
await db.commit()
|
151
|
-
await db.refresh(db_workflow_task)
|
152
|
-
await db.close()
|
153
|
-
|
154
|
-
return db_workflow_task
|
155
|
-
|
156
|
-
|
157
|
-
@router.delete(
|
158
|
-
"/project/{project_id}/workflow/{workflow_id}/wftask/{workflow_task_id}/",
|
159
|
-
status_code=status.HTTP_204_NO_CONTENT,
|
160
|
-
)
|
161
|
-
async def delete_workflowtask(
|
162
|
-
project_id: int,
|
163
|
-
workflow_id: int,
|
164
|
-
workflow_task_id: int,
|
165
|
-
user: UserOAuth = Depends(current_active_user),
|
166
|
-
db: AsyncSession = Depends(get_async_db),
|
167
|
-
) -> Response:
|
168
|
-
"""
|
169
|
-
Delete a WorkflowTask of a Workflow
|
170
|
-
"""
|
171
|
-
_raise_if_v1_is_read_only()
|
172
|
-
db_workflow_task, db_workflow = await _get_workflow_task_check_owner(
|
173
|
-
project_id=project_id,
|
174
|
-
workflow_task_id=workflow_task_id,
|
175
|
-
workflow_id=workflow_id,
|
176
|
-
user_id=user.id,
|
177
|
-
db=db,
|
178
|
-
)
|
179
|
-
|
180
|
-
await db.delete(db_workflow_task)
|
181
|
-
await db.commit()
|
182
|
-
|
183
|
-
await db.refresh(db_workflow)
|
184
|
-
db_workflow.task_list.reorder()
|
185
|
-
await db.commit()
|
186
|
-
|
187
|
-
return Response(status_code=status.HTTP_204_NO_CONTENT)
|
@@ -1,27 +0,0 @@
|
|
1
|
-
import asyncio
|
2
|
-
from functools import partial
|
3
|
-
from functools import wraps
|
4
|
-
from typing import Callable
|
5
|
-
|
6
|
-
|
7
|
-
def async_wrap_v1(func: Callable) -> Callable:
|
8
|
-
"""
|
9
|
-
Wrap a synchronous callable in an async task
|
10
|
-
|
11
|
-
Ref: [issue #140](https://github.com/fractal-analytics-platform/fractal-server/issues/140)
|
12
|
-
and [this StackOverflow answer](https://stackoverflow.com/q/43241221/19085332).
|
13
|
-
|
14
|
-
Returns:
|
15
|
-
async_wrapper:
|
16
|
-
A factory that allows wrapping a blocking callable within a
|
17
|
-
coroutine.
|
18
|
-
""" # noqa: E501
|
19
|
-
|
20
|
-
@wraps(func)
|
21
|
-
async def async_wrapper(*args, loop=None, executor=None, **kwargs):
|
22
|
-
if loop is None:
|
23
|
-
loop = asyncio.get_event_loop()
|
24
|
-
pfunc = partial(func, *args, **kwargs)
|
25
|
-
return await loop.run_in_executor(executor, pfunc)
|
26
|
-
|
27
|
-
return async_wrapper
|