fractal-server 2.17.0a11__py3-none-any.whl → 2.17.1__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- fractal_server/__init__.py +1 -1
- fractal_server/__main__.py +1 -1
- fractal_server/app/models/__init__.py +0 -1
- fractal_server/app/models/security.py +1 -14
- fractal_server/app/models/v2/project.py +1 -4
- fractal_server/app/models/v2/task_group.py +1 -4
- fractal_server/app/routes/admin/v2/job.py +45 -9
- fractal_server/app/routes/admin/v2/task.py +49 -32
- fractal_server/app/schemas/v2/task_collection.py +11 -3
- fractal_server/app/shutdown.py +23 -19
- fractal_server/config/_database.py +3 -2
- fractal_server/config/_main.py +1 -1
- fractal_server/migrations/versions/45fbb391d7af_make_resource_id_fk_non_nullable.py +46 -0
- fractal_server/migrations/versions/49d0856e9569_drop_table.py +63 -0
- fractal_server/migrations/versions/7673fe18c05d_remove_project_dir_server_default.py +29 -0
- fractal_server/migrations/versions/caba9fb1ea5e_drop_useroauth_user_settings_id.py +49 -0
- fractal_server/runner/v2/runner.py +24 -0
- fractal_server/tasks/config/_python.py +1 -1
- fractal_server/tasks/v2/local/delete.py +1 -1
- {fractal_server-2.17.0a11.dist-info → fractal_server-2.17.1.dist-info}/METADATA +1 -1
- {fractal_server-2.17.0a11.dist-info → fractal_server-2.17.1.dist-info}/RECORD +24 -22
- fractal_server/app/models/user_settings.py +0 -37
- fractal_server/data_migrations/2_17_0.py +0 -321
- {fractal_server-2.17.0a11.dist-info → fractal_server-2.17.1.dist-info}/WHEEL +0 -0
- {fractal_server-2.17.0a11.dist-info → fractal_server-2.17.1.dist-info}/entry_points.txt +0 -0
- {fractal_server-2.17.0a11.dist-info → fractal_server-2.17.1.dist-info}/licenses/LICENSE +0 -0
fractal_server/__init__.py
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
__VERSION__ = "2.17.
|
|
1
|
+
__VERSION__ = "2.17.1"
|
fractal_server/__main__.py
CHANGED
|
@@ -7,7 +7,6 @@ from pathlib import Path
|
|
|
7
7
|
import uvicorn
|
|
8
8
|
from pydantic import ValidationError
|
|
9
9
|
|
|
10
|
-
from fractal_server.app.schemas.v2 import ResourceType
|
|
11
10
|
|
|
12
11
|
parser = ap.ArgumentParser(description="fractal-server commands")
|
|
13
12
|
|
|
@@ -149,6 +148,7 @@ def init_db_data(
|
|
|
149
148
|
from fractal_server.app.models import Resource, Profile
|
|
150
149
|
from fractal_server.app.schemas.v2.resource import cast_serialize_resource
|
|
151
150
|
from fractal_server.app.schemas.v2.profile import cast_serialize_profile
|
|
151
|
+
from fractal_server.app.schemas.v2 import ResourceType
|
|
152
152
|
|
|
153
153
|
# Create default group and user
|
|
154
154
|
print()
|
|
@@ -95,24 +95,11 @@ class UserOAuth(SQLModel, table=True):
|
|
|
95
95
|
ondelete="RESTRICT",
|
|
96
96
|
)
|
|
97
97
|
|
|
98
|
-
|
|
99
|
-
project_dir: str = Field(
|
|
100
|
-
sa_column=Column(
|
|
101
|
-
String,
|
|
102
|
-
server_default="/PLACEHOLDER",
|
|
103
|
-
nullable=False,
|
|
104
|
-
)
|
|
105
|
-
)
|
|
98
|
+
project_dir: str
|
|
106
99
|
slurm_accounts: list[str] = Field(
|
|
107
100
|
sa_column=Column(ARRAY(String), server_default="{}"),
|
|
108
101
|
)
|
|
109
102
|
|
|
110
|
-
# TODO-2.17.1: remove
|
|
111
|
-
user_settings_id: int | None = Field(
|
|
112
|
-
foreign_key="user_settings.id",
|
|
113
|
-
default=None,
|
|
114
|
-
)
|
|
115
|
-
|
|
116
103
|
|
|
117
104
|
class UserGroup(SQLModel, table=True):
|
|
118
105
|
"""
|
|
@@ -15,10 +15,7 @@ class ProjectV2(SQLModel, table=True):
|
|
|
15
15
|
id: int | None = Field(default=None, primary_key=True)
|
|
16
16
|
name: str
|
|
17
17
|
|
|
18
|
-
|
|
19
|
-
resource_id: int | None = Field(
|
|
20
|
-
foreign_key="resource.id", default=None, ondelete="RESTRICT"
|
|
21
|
-
)
|
|
18
|
+
resource_id: int = Field(foreign_key="resource.id", ondelete="RESTRICT")
|
|
22
19
|
timestamp_created: datetime = Field(
|
|
23
20
|
default_factory=get_timestamp,
|
|
24
21
|
sa_column=Column(DateTime(timezone=True), nullable=False),
|
|
@@ -42,10 +42,7 @@ class TaskGroupV2(SQLModel, table=True):
|
|
|
42
42
|
user_group_id: int | None = Field(
|
|
43
43
|
foreign_key="usergroup.id", default=None, ondelete="SET NULL"
|
|
44
44
|
)
|
|
45
|
-
|
|
46
|
-
resource_id: int | None = Field(
|
|
47
|
-
foreign_key="resource.id", default=None, ondelete="RESTRICT"
|
|
48
|
-
)
|
|
45
|
+
resource_id: int = Field(foreign_key="resource.id", ondelete="RESTRICT")
|
|
49
46
|
|
|
50
47
|
origin: str
|
|
51
48
|
pkg_name: str
|
|
@@ -7,6 +7,7 @@ from fastapi import Response
|
|
|
7
7
|
from fastapi import status
|
|
8
8
|
from fastapi.responses import StreamingResponse
|
|
9
9
|
from pydantic.types import AwareDatetime
|
|
10
|
+
from sqlalchemy import func
|
|
10
11
|
from sqlmodel import select
|
|
11
12
|
|
|
12
13
|
from fractal_server.app.db import AsyncSession
|
|
@@ -19,6 +20,9 @@ from fractal_server.app.models.v2 import ProjectV2
|
|
|
19
20
|
from fractal_server.app.routes.auth import current_superuser_act
|
|
20
21
|
from fractal_server.app.routes.aux._job import _write_shutdown_file
|
|
21
22
|
from fractal_server.app.routes.aux._runner import _check_shutdown_is_supported
|
|
23
|
+
from fractal_server.app.routes.pagination import get_pagination_params
|
|
24
|
+
from fractal_server.app.routes.pagination import PaginationRequest
|
|
25
|
+
from fractal_server.app.routes.pagination import PaginationResponse
|
|
22
26
|
from fractal_server.app.schemas.v2 import HistoryUnitStatus
|
|
23
27
|
from fractal_server.app.schemas.v2 import JobReadV2
|
|
24
28
|
from fractal_server.app.schemas.v2 import JobStatusTypeV2
|
|
@@ -30,7 +34,7 @@ from fractal_server.zip_tools import _zip_folder_to_byte_stream_iterator
|
|
|
30
34
|
router = APIRouter()
|
|
31
35
|
|
|
32
36
|
|
|
33
|
-
@router.get("/", response_model=
|
|
37
|
+
@router.get("/", response_model=PaginationResponse[JobReadV2])
|
|
34
38
|
async def view_job(
|
|
35
39
|
id: int | None = None,
|
|
36
40
|
user_id: int | None = None,
|
|
@@ -43,9 +47,10 @@ async def view_job(
|
|
|
43
47
|
end_timestamp_min: AwareDatetime | None = None,
|
|
44
48
|
end_timestamp_max: AwareDatetime | None = None,
|
|
45
49
|
log: bool = True,
|
|
50
|
+
pagination: PaginationRequest = Depends(get_pagination_params),
|
|
46
51
|
user: UserOAuth = Depends(current_superuser_act),
|
|
47
52
|
db: AsyncSession = Depends(get_async_db),
|
|
48
|
-
) ->
|
|
53
|
+
) -> PaginationResponse[JobReadV2]:
|
|
49
54
|
"""
|
|
50
55
|
Query `ApplyWorkflow` table.
|
|
51
56
|
|
|
@@ -68,43 +73,74 @@ async def view_job(
|
|
|
68
73
|
`job.log` is set to `None`.
|
|
69
74
|
"""
|
|
70
75
|
|
|
71
|
-
|
|
76
|
+
# Assign pagination parameters
|
|
77
|
+
page = pagination.page
|
|
78
|
+
page_size = pagination.page_size
|
|
72
79
|
|
|
80
|
+
# Prepare statements
|
|
81
|
+
stm = select(JobV2).order_by(JobV2.start_timestamp.desc())
|
|
82
|
+
stm_count = select(func.count(JobV2.id))
|
|
73
83
|
if id is not None:
|
|
74
84
|
stm = stm.where(JobV2.id == id)
|
|
85
|
+
stm_count = stm_count.where(JobV2.id == id)
|
|
75
86
|
if user_id is not None:
|
|
76
87
|
stm = stm.join(ProjectV2).where(
|
|
77
88
|
ProjectV2.user_list.any(UserOAuth.id == user_id)
|
|
78
89
|
)
|
|
90
|
+
stm_count = stm_count.join(ProjectV2).where(
|
|
91
|
+
ProjectV2.user_list.any(UserOAuth.id == user_id)
|
|
92
|
+
)
|
|
79
93
|
if project_id is not None:
|
|
80
94
|
stm = stm.where(JobV2.project_id == project_id)
|
|
95
|
+
stm_count = stm_count.where(JobV2.project_id == project_id)
|
|
81
96
|
if dataset_id is not None:
|
|
82
97
|
stm = stm.where(JobV2.dataset_id == dataset_id)
|
|
98
|
+
stm_count = stm_count.where(JobV2.dataset_id == dataset_id)
|
|
83
99
|
if workflow_id is not None:
|
|
84
100
|
stm = stm.where(JobV2.workflow_id == workflow_id)
|
|
101
|
+
stm_count = stm_count.where(JobV2.workflow_id == workflow_id)
|
|
85
102
|
if status is not None:
|
|
86
103
|
stm = stm.where(JobV2.status == status)
|
|
104
|
+
stm_count = stm_count.where(JobV2.status == status)
|
|
87
105
|
if start_timestamp_min is not None:
|
|
88
|
-
start_timestamp_min = start_timestamp_min
|
|
89
106
|
stm = stm.where(JobV2.start_timestamp >= start_timestamp_min)
|
|
107
|
+
stm_count = stm_count.where(
|
|
108
|
+
JobV2.start_timestamp >= start_timestamp_min
|
|
109
|
+
)
|
|
90
110
|
if start_timestamp_max is not None:
|
|
91
|
-
start_timestamp_max = start_timestamp_max
|
|
92
111
|
stm = stm.where(JobV2.start_timestamp <= start_timestamp_max)
|
|
112
|
+
stm_count = stm_count.where(
|
|
113
|
+
JobV2.start_timestamp <= start_timestamp_max
|
|
114
|
+
)
|
|
93
115
|
if end_timestamp_min is not None:
|
|
94
|
-
end_timestamp_min = end_timestamp_min
|
|
95
116
|
stm = stm.where(JobV2.end_timestamp >= end_timestamp_min)
|
|
117
|
+
stm_count = stm_count.where(JobV2.end_timestamp >= end_timestamp_min)
|
|
96
118
|
if end_timestamp_max is not None:
|
|
97
|
-
end_timestamp_max = end_timestamp_max
|
|
98
119
|
stm = stm.where(JobV2.end_timestamp <= end_timestamp_max)
|
|
120
|
+
stm_count = stm_count.where(JobV2.end_timestamp <= end_timestamp_max)
|
|
99
121
|
|
|
122
|
+
# Find total number of elements
|
|
123
|
+
res_total_count = await db.execute(stm_count)
|
|
124
|
+
total_count = res_total_count.scalar()
|
|
125
|
+
if page_size is None:
|
|
126
|
+
page_size = total_count
|
|
127
|
+
else:
|
|
128
|
+
stm = stm.offset((page - 1) * page_size).limit(page_size)
|
|
129
|
+
|
|
130
|
+
# Get `page_size` rows
|
|
100
131
|
res = await db.execute(stm)
|
|
101
132
|
job_list = res.scalars().all()
|
|
102
|
-
|
|
133
|
+
|
|
103
134
|
if not log:
|
|
104
135
|
for job in job_list:
|
|
105
136
|
setattr(job, "log", None)
|
|
106
137
|
|
|
107
|
-
return
|
|
138
|
+
return PaginationResponse[JobReadV2](
|
|
139
|
+
total_count=total_count,
|
|
140
|
+
page_size=page_size,
|
|
141
|
+
current_page=page,
|
|
142
|
+
items=[job.model_dump() for job in job_list],
|
|
143
|
+
)
|
|
108
144
|
|
|
109
145
|
|
|
110
146
|
@router.get("/{job_id}/", response_model=JobReadV2)
|
|
@@ -1,7 +1,5 @@
|
|
|
1
1
|
from fastapi import APIRouter
|
|
2
2
|
from fastapi import Depends
|
|
3
|
-
from fastapi import HTTPException
|
|
4
|
-
from fastapi import status
|
|
5
3
|
from pydantic import BaseModel
|
|
6
4
|
from pydantic import EmailStr
|
|
7
5
|
from pydantic import Field
|
|
@@ -16,6 +14,10 @@ from fractal_server.app.models.v2 import TaskV2
|
|
|
16
14
|
from fractal_server.app.models.v2 import WorkflowTaskV2
|
|
17
15
|
from fractal_server.app.models.v2 import WorkflowV2
|
|
18
16
|
from fractal_server.app.routes.auth import current_superuser_act
|
|
17
|
+
from fractal_server.app.routes.pagination import get_pagination_params
|
|
18
|
+
from fractal_server.app.routes.pagination import PaginationRequest
|
|
19
|
+
from fractal_server.app.routes.pagination import PaginationResponse
|
|
20
|
+
from fractal_server.app.schemas.v2.task import TaskType
|
|
19
21
|
|
|
20
22
|
router = APIRouter()
|
|
21
23
|
|
|
@@ -49,75 +51,85 @@ class TaskV2Info(BaseModel):
|
|
|
49
51
|
relationships: list[TaskV2Relationship]
|
|
50
52
|
|
|
51
53
|
|
|
52
|
-
@router.get("/", response_model=
|
|
54
|
+
@router.get("/", response_model=PaginationResponse[TaskV2Info])
|
|
53
55
|
async def query_tasks(
|
|
54
56
|
id: int | None = None,
|
|
55
57
|
source: str | None = None,
|
|
56
58
|
version: str | None = None,
|
|
57
59
|
name: str | None = None,
|
|
58
|
-
|
|
60
|
+
task_type: TaskType | None = None,
|
|
59
61
|
category: str | None = None,
|
|
60
62
|
modality: str | None = None,
|
|
61
63
|
author: str | None = None,
|
|
62
64
|
resource_id: int | None = None,
|
|
65
|
+
pagination: PaginationRequest = Depends(get_pagination_params),
|
|
63
66
|
user: UserOAuth = Depends(current_superuser_act),
|
|
64
67
|
db: AsyncSession = Depends(get_async_db),
|
|
65
|
-
) ->
|
|
68
|
+
) -> PaginationResponse[TaskV2Info]:
|
|
66
69
|
"""
|
|
67
|
-
Query `TaskV2`
|
|
68
|
-
(WorkflowV2s and ProjectV2s)
|
|
69
|
-
|
|
70
|
-
Args:
|
|
71
|
-
id: If not `None`, query for matching `task.id`.
|
|
72
|
-
source: If not `None`, query for contained case insensitive
|
|
73
|
-
`task.source`.
|
|
74
|
-
version: If not `None`, query for matching `task.version`.
|
|
75
|
-
name: If not `None`, query for contained case insensitive `task.name`.
|
|
76
|
-
max_number_of_results: The maximum length of the response.
|
|
77
|
-
category:
|
|
78
|
-
modality:
|
|
79
|
-
author:
|
|
80
|
-
resource_id:
|
|
70
|
+
Query `TaskV2` and get information about related workflows and projects.
|
|
81
71
|
"""
|
|
82
72
|
|
|
83
|
-
|
|
73
|
+
# Assign pagination parameters
|
|
74
|
+
page = pagination.page
|
|
75
|
+
page_size = pagination.page_size
|
|
84
76
|
|
|
77
|
+
# Prepare statements
|
|
78
|
+
stm = select(TaskV2).order_by(TaskV2.id)
|
|
79
|
+
stm_count = select(func.count(TaskV2.id))
|
|
85
80
|
if id is not None:
|
|
86
81
|
stm = stm.where(TaskV2.id == id)
|
|
82
|
+
stm_count = stm_count.where(TaskV2.id == id)
|
|
87
83
|
if source is not None:
|
|
88
84
|
stm = stm.where(TaskV2.source.icontains(source))
|
|
85
|
+
stm_count = stm_count.where(TaskV2.source.icontains(source))
|
|
89
86
|
if version is not None:
|
|
90
87
|
stm = stm.where(TaskV2.version == version)
|
|
88
|
+
stm_count = stm_count.where(TaskV2.version == version)
|
|
91
89
|
if name is not None:
|
|
92
90
|
stm = stm.where(TaskV2.name.icontains(name))
|
|
91
|
+
stm_count = stm_count.where(TaskV2.name.icontains(name))
|
|
92
|
+
if task_type is not None:
|
|
93
|
+
stm = stm.where(TaskV2.type == task_type)
|
|
94
|
+
stm_count = stm_count.where(TaskV2.type == task_type)
|
|
93
95
|
if category is not None:
|
|
94
96
|
stm = stm.where(func.lower(TaskV2.category) == category.lower())
|
|
97
|
+
stm_count = stm_count.where(
|
|
98
|
+
func.lower(TaskV2.category) == category.lower()
|
|
99
|
+
)
|
|
95
100
|
if modality is not None:
|
|
96
101
|
stm = stm.where(func.lower(TaskV2.modality) == modality.lower())
|
|
102
|
+
stm_count = stm_count.where(
|
|
103
|
+
func.lower(TaskV2.modality) == modality.lower()
|
|
104
|
+
)
|
|
97
105
|
if author is not None:
|
|
98
106
|
stm = stm.where(TaskV2.authors.icontains(author))
|
|
107
|
+
stm_count = stm_count.where(TaskV2.authors.icontains(author))
|
|
99
108
|
if resource_id is not None:
|
|
100
109
|
stm = (
|
|
101
110
|
stm.join(TaskGroupV2)
|
|
102
111
|
.where(TaskGroupV2.id == TaskV2.taskgroupv2_id)
|
|
103
112
|
.where(TaskGroupV2.resource_id == resource_id)
|
|
104
113
|
)
|
|
114
|
+
stm_count = (
|
|
115
|
+
stm_count.join(TaskGroupV2)
|
|
116
|
+
.where(TaskGroupV2.id == TaskV2.taskgroupv2_id)
|
|
117
|
+
.where(TaskGroupV2.resource_id == resource_id)
|
|
118
|
+
)
|
|
119
|
+
|
|
120
|
+
# Find total number of elements
|
|
121
|
+
res_total_count = await db.execute(stm_count)
|
|
122
|
+
total_count = res_total_count.scalar()
|
|
123
|
+
if page_size is None:
|
|
124
|
+
page_size = total_count
|
|
125
|
+
else:
|
|
126
|
+
stm = stm.offset((page - 1) * page_size).limit(page_size)
|
|
105
127
|
|
|
106
|
-
|
|
128
|
+
# Get `page_size` rows
|
|
107
129
|
res = await db.execute(stm)
|
|
108
130
|
task_list = res.scalars().all()
|
|
109
|
-
if len(task_list) > max_number_of_results:
|
|
110
|
-
await db.close()
|
|
111
|
-
raise HTTPException(
|
|
112
|
-
status_code=status.HTTP_422_UNPROCESSABLE_CONTENT,
|
|
113
|
-
detail=(
|
|
114
|
-
f"Too many Tasks ({len(task_list)} > {max_number_of_results})."
|
|
115
|
-
" Please add more query filters."
|
|
116
|
-
),
|
|
117
|
-
)
|
|
118
131
|
|
|
119
132
|
task_info_list = []
|
|
120
|
-
|
|
121
133
|
for task in task_list:
|
|
122
134
|
stm = (
|
|
123
135
|
select(WorkflowV2)
|
|
@@ -147,4 +159,9 @@ async def query_tasks(
|
|
|
147
159
|
)
|
|
148
160
|
)
|
|
149
161
|
|
|
150
|
-
return
|
|
162
|
+
return PaginationResponse[TaskV2Info](
|
|
163
|
+
total_count=total_count,
|
|
164
|
+
page_size=page_size,
|
|
165
|
+
current_page=page,
|
|
166
|
+
items=task_info_list,
|
|
167
|
+
)
|
|
@@ -51,9 +51,17 @@ class TaskCollectPipV2(BaseModel):
|
|
|
51
51
|
package: NonEmptyStr | None = None
|
|
52
52
|
package_version: NonEmptyStr | None = None
|
|
53
53
|
package_extras: NonEmptyStr | None = None
|
|
54
|
-
python_version:
|
|
55
|
-
|
|
56
|
-
|
|
54
|
+
python_version: (
|
|
55
|
+
Literal[
|
|
56
|
+
"3.9",
|
|
57
|
+
"3.10",
|
|
58
|
+
"3.11",
|
|
59
|
+
"3.12",
|
|
60
|
+
"3.13",
|
|
61
|
+
"3.14",
|
|
62
|
+
]
|
|
63
|
+
| None
|
|
64
|
+
) = None
|
|
57
65
|
pinned_package_versions_pre: DictStrStr | None = None
|
|
58
66
|
pinned_package_versions_post: DictStrStr | None = None
|
|
59
67
|
|
fractal_server/app/shutdown.py
CHANGED
|
@@ -12,44 +12,48 @@ from fractal_server.syringe import Inject
|
|
|
12
12
|
|
|
13
13
|
|
|
14
14
|
async def cleanup_after_shutdown(*, jobsV2: list[int], logger_name: str):
|
|
15
|
+
settings = Inject(get_settings)
|
|
15
16
|
logger = get_logger(logger_name)
|
|
16
17
|
logger.info("Cleanup function after shutdown")
|
|
17
|
-
|
|
18
|
+
stm_objects = (
|
|
18
19
|
select(JobV2)
|
|
19
20
|
.where(JobV2.id.in_(jobsV2))
|
|
20
21
|
.where(JobV2.status == JobStatusTypeV2.SUBMITTED)
|
|
21
22
|
)
|
|
23
|
+
stm_ids = (
|
|
24
|
+
select(JobV2.id)
|
|
25
|
+
.where(JobV2.id.in_(jobsV2))
|
|
26
|
+
.where(JobV2.status == JobStatusTypeV2.SUBMITTED)
|
|
27
|
+
)
|
|
22
28
|
|
|
23
29
|
async for session in get_async_db():
|
|
24
|
-
|
|
25
|
-
|
|
26
|
-
for job in
|
|
30
|
+
# Write shutdown file for all jobs
|
|
31
|
+
jobs = (await session.execute(stm_objects)).scalars().all()
|
|
32
|
+
for job in jobs:
|
|
27
33
|
_write_shutdown_file(job=job)
|
|
28
34
|
|
|
29
|
-
|
|
30
|
-
|
|
35
|
+
# Wait for completion of all job - with a timeout
|
|
36
|
+
interval = settings.FRACTAL_GRACEFUL_SHUTDOWN_TIME / 20
|
|
31
37
|
t_start = time.perf_counter()
|
|
32
38
|
while (
|
|
33
39
|
time.perf_counter() - t_start
|
|
34
|
-
)
|
|
35
|
-
|
|
36
|
-
|
|
37
|
-
|
|
38
|
-
|
|
39
|
-
if len(jobsV2_db) == 0:
|
|
40
|
-
logger.info(
|
|
41
|
-
"All jobs associated to this app are "
|
|
42
|
-
"either done or failed. Exit."
|
|
43
|
-
)
|
|
40
|
+
) <= settings.FRACTAL_GRACEFUL_SHUTDOWN_TIME:
|
|
41
|
+
job_ids = (await session.execute(stm_ids)).scalars().all()
|
|
42
|
+
if len(job_ids) == 0:
|
|
43
|
+
logger.info("All jobs are either done or failed. Exit.")
|
|
44
44
|
return
|
|
45
45
|
else:
|
|
46
|
-
logger.info(f"Some jobs are still 'submitted' {
|
|
46
|
+
logger.info(f"Some jobs are still 'submitted': {job_ids=}")
|
|
47
|
+
logger.info(f"Wait {interval:.4f} seconds before next check.")
|
|
48
|
+
time.sleep(interval)
|
|
47
49
|
logger.info(
|
|
48
50
|
"Graceful shutdown reached its maximum time, "
|
|
49
|
-
"but some jobs are still submitted"
|
|
51
|
+
"but some jobs are still submitted."
|
|
50
52
|
)
|
|
51
53
|
|
|
52
|
-
|
|
54
|
+
# Mark jobs as failed and update their logs.
|
|
55
|
+
jobs = (await session.execute(stm_objects)).scalars().all()
|
|
56
|
+
for job in jobs:
|
|
53
57
|
job.status = "failed"
|
|
54
58
|
job.log = (job.log or "") + "\nJob stopped due to app shutdown\n"
|
|
55
59
|
session.add(job)
|
|
@@ -1,4 +1,5 @@
|
|
|
1
1
|
from pydantic import SecretStr
|
|
2
|
+
from pydantic.types import NonNegativeInt
|
|
2
3
|
from pydantic_settings import BaseSettings
|
|
3
4
|
from pydantic_settings import SettingsConfigDict
|
|
4
5
|
from sqlalchemy.engine import URL
|
|
@@ -27,11 +28,11 @@ class DatabaseSettings(BaseSettings):
|
|
|
27
28
|
"""
|
|
28
29
|
Password to use when connecting to the PostgreSQL database.
|
|
29
30
|
"""
|
|
30
|
-
POSTGRES_HOST: NonEmptyStr
|
|
31
|
+
POSTGRES_HOST: NonEmptyStr = "localhost"
|
|
31
32
|
"""
|
|
32
33
|
URL to the PostgreSQL server or path to a UNIX domain socket.
|
|
33
34
|
"""
|
|
34
|
-
POSTGRES_PORT:
|
|
35
|
+
POSTGRES_PORT: NonNegativeInt = 5432
|
|
35
36
|
"""
|
|
36
37
|
Port number to use when connecting to the PostgreSQL server.
|
|
37
38
|
"""
|
fractal_server/config/_main.py
CHANGED
|
@@ -0,0 +1,46 @@
|
|
|
1
|
+
"""Make resource_id FK non-nullable
|
|
2
|
+
|
|
3
|
+
Revision ID: 45fbb391d7af
|
|
4
|
+
Revises: caba9fb1ea5e
|
|
5
|
+
Create Date: 2025-11-11 16:39:12.813766
|
|
6
|
+
|
|
7
|
+
"""
|
|
8
|
+
import sqlalchemy as sa
|
|
9
|
+
from alembic import op
|
|
10
|
+
|
|
11
|
+
|
|
12
|
+
# revision identifiers, used by Alembic.
|
|
13
|
+
revision = "45fbb391d7af"
|
|
14
|
+
down_revision = "caba9fb1ea5e"
|
|
15
|
+
branch_labels = None
|
|
16
|
+
depends_on = None
|
|
17
|
+
|
|
18
|
+
|
|
19
|
+
def upgrade() -> None:
|
|
20
|
+
# ### commands auto generated by Alembic - please adjust! ###
|
|
21
|
+
with op.batch_alter_table("projectv2", schema=None) as batch_op:
|
|
22
|
+
batch_op.alter_column(
|
|
23
|
+
"resource_id", existing_type=sa.INTEGER(), nullable=False
|
|
24
|
+
)
|
|
25
|
+
|
|
26
|
+
with op.batch_alter_table("taskgroupv2", schema=None) as batch_op:
|
|
27
|
+
batch_op.alter_column(
|
|
28
|
+
"resource_id", existing_type=sa.INTEGER(), nullable=False
|
|
29
|
+
)
|
|
30
|
+
|
|
31
|
+
# ### end Alembic commands ###
|
|
32
|
+
|
|
33
|
+
|
|
34
|
+
def downgrade() -> None:
|
|
35
|
+
# ### commands auto generated by Alembic - please adjust! ###
|
|
36
|
+
with op.batch_alter_table("taskgroupv2", schema=None) as batch_op:
|
|
37
|
+
batch_op.alter_column(
|
|
38
|
+
"resource_id", existing_type=sa.INTEGER(), nullable=True
|
|
39
|
+
)
|
|
40
|
+
|
|
41
|
+
with op.batch_alter_table("projectv2", schema=None) as batch_op:
|
|
42
|
+
batch_op.alter_column(
|
|
43
|
+
"resource_id", existing_type=sa.INTEGER(), nullable=True
|
|
44
|
+
)
|
|
45
|
+
|
|
46
|
+
# ### end Alembic commands ###
|
|
@@ -0,0 +1,63 @@
|
|
|
1
|
+
"""Drop table
|
|
2
|
+
|
|
3
|
+
Revision ID: 49d0856e9569
|
|
4
|
+
Revises: 45fbb391d7af
|
|
5
|
+
Create Date: 2025-11-11 16:39:41.497832
|
|
6
|
+
|
|
7
|
+
"""
|
|
8
|
+
import sqlalchemy as sa
|
|
9
|
+
from alembic import op
|
|
10
|
+
from sqlalchemy.dialects import postgresql
|
|
11
|
+
|
|
12
|
+
# revision identifiers, used by Alembic.
|
|
13
|
+
revision = "49d0856e9569"
|
|
14
|
+
down_revision = "45fbb391d7af"
|
|
15
|
+
branch_labels = None
|
|
16
|
+
depends_on = None
|
|
17
|
+
|
|
18
|
+
|
|
19
|
+
def upgrade() -> None:
|
|
20
|
+
# ### commands auto generated by Alembic - please adjust! ###
|
|
21
|
+
op.drop_table("user_settings")
|
|
22
|
+
# ### end Alembic commands ###
|
|
23
|
+
|
|
24
|
+
|
|
25
|
+
def downgrade() -> None:
|
|
26
|
+
# ### commands auto generated by Alembic - please adjust! ###
|
|
27
|
+
op.create_table(
|
|
28
|
+
"user_settings",
|
|
29
|
+
sa.Column("id", sa.INTEGER(), autoincrement=True, nullable=False),
|
|
30
|
+
sa.Column(
|
|
31
|
+
"slurm_accounts",
|
|
32
|
+
postgresql.JSONB(astext_type=sa.Text()),
|
|
33
|
+
server_default=sa.text("'[]'::json"),
|
|
34
|
+
autoincrement=False,
|
|
35
|
+
nullable=False,
|
|
36
|
+
),
|
|
37
|
+
sa.Column(
|
|
38
|
+
"ssh_host", sa.VARCHAR(), autoincrement=False, nullable=True
|
|
39
|
+
),
|
|
40
|
+
sa.Column(
|
|
41
|
+
"ssh_username", sa.VARCHAR(), autoincrement=False, nullable=True
|
|
42
|
+
),
|
|
43
|
+
sa.Column(
|
|
44
|
+
"ssh_private_key_path",
|
|
45
|
+
sa.VARCHAR(),
|
|
46
|
+
autoincrement=False,
|
|
47
|
+
nullable=True,
|
|
48
|
+
),
|
|
49
|
+
sa.Column(
|
|
50
|
+
"ssh_tasks_dir", sa.VARCHAR(), autoincrement=False, nullable=True
|
|
51
|
+
),
|
|
52
|
+
sa.Column(
|
|
53
|
+
"ssh_jobs_dir", sa.VARCHAR(), autoincrement=False, nullable=True
|
|
54
|
+
),
|
|
55
|
+
sa.Column(
|
|
56
|
+
"slurm_user", sa.VARCHAR(), autoincrement=False, nullable=True
|
|
57
|
+
),
|
|
58
|
+
sa.Column(
|
|
59
|
+
"project_dir", sa.VARCHAR(), autoincrement=False, nullable=True
|
|
60
|
+
),
|
|
61
|
+
sa.PrimaryKeyConstraint("id", name=op.f("pk_user_settings")),
|
|
62
|
+
)
|
|
63
|
+
# ### end Alembic commands ###
|
|
@@ -0,0 +1,29 @@
|
|
|
1
|
+
"""Remove project_dir server_default
|
|
2
|
+
|
|
3
|
+
Revision ID: 7673fe18c05d
|
|
4
|
+
Revises: 49d0856e9569
|
|
5
|
+
Create Date: 2025-11-11 16:50:20.079193
|
|
6
|
+
|
|
7
|
+
"""
|
|
8
|
+
from alembic import op
|
|
9
|
+
|
|
10
|
+
|
|
11
|
+
# revision identifiers, used by Alembic.
|
|
12
|
+
revision = "7673fe18c05d"
|
|
13
|
+
down_revision = "49d0856e9569"
|
|
14
|
+
branch_labels = None
|
|
15
|
+
depends_on = None
|
|
16
|
+
|
|
17
|
+
|
|
18
|
+
def upgrade() -> None:
|
|
19
|
+
"""
|
|
20
|
+
Remove `server_default` for `project_dir` column - see
|
|
21
|
+
https://alembic.sqlalchemy.org/en/latest/ops.html#alembic.operations.Operations.alter_column.params.server_default
|
|
22
|
+
"""
|
|
23
|
+
with op.batch_alter_table("user_oauth") as batch_op:
|
|
24
|
+
batch_op.alter_column("project_dir", server_default=None)
|
|
25
|
+
|
|
26
|
+
|
|
27
|
+
def downgrade() -> None:
|
|
28
|
+
with op.batch_alter_table("user_oauth") as batch_op:
|
|
29
|
+
batch_op.alter_column("project_dir", server_default="/PLACEHOLDER")
|
|
@@ -0,0 +1,49 @@
|
|
|
1
|
+
"""Drop UserOAuth.user_settings_id
|
|
2
|
+
|
|
3
|
+
Revision ID: caba9fb1ea5e
|
|
4
|
+
Revises: 83bc2ad3ffcc
|
|
5
|
+
Create Date: 2025-11-11 16:38:27.243693
|
|
6
|
+
|
|
7
|
+
"""
|
|
8
|
+
import sqlalchemy as sa
|
|
9
|
+
from alembic import op
|
|
10
|
+
|
|
11
|
+
|
|
12
|
+
# revision identifiers, used by Alembic.
|
|
13
|
+
revision = "caba9fb1ea5e"
|
|
14
|
+
down_revision = "83bc2ad3ffcc"
|
|
15
|
+
branch_labels = None
|
|
16
|
+
depends_on = None
|
|
17
|
+
|
|
18
|
+
|
|
19
|
+
def upgrade() -> None:
|
|
20
|
+
# ### commands auto generated by Alembic - please adjust! ###
|
|
21
|
+
with op.batch_alter_table("user_oauth", schema=None) as batch_op:
|
|
22
|
+
batch_op.drop_constraint(
|
|
23
|
+
batch_op.f("fk_user_oauth_user_settings_id_user_settings"),
|
|
24
|
+
type_="foreignkey",
|
|
25
|
+
)
|
|
26
|
+
batch_op.drop_column("user_settings_id")
|
|
27
|
+
|
|
28
|
+
# ### end Alembic commands ###
|
|
29
|
+
|
|
30
|
+
|
|
31
|
+
def downgrade() -> None:
|
|
32
|
+
# ### commands auto generated by Alembic - please adjust! ###
|
|
33
|
+
with op.batch_alter_table("user_oauth", schema=None) as batch_op:
|
|
34
|
+
batch_op.add_column(
|
|
35
|
+
sa.Column(
|
|
36
|
+
"user_settings_id",
|
|
37
|
+
sa.INTEGER(),
|
|
38
|
+
autoincrement=False,
|
|
39
|
+
nullable=True,
|
|
40
|
+
)
|
|
41
|
+
)
|
|
42
|
+
batch_op.create_foreign_key(
|
|
43
|
+
batch_op.f("fk_user_oauth_user_settings_id_user_settings"),
|
|
44
|
+
"user_settings",
|
|
45
|
+
["user_settings_id"],
|
|
46
|
+
["id"],
|
|
47
|
+
)
|
|
48
|
+
|
|
49
|
+
# ### end Alembic commands ###
|
|
@@ -192,6 +192,30 @@ def execute_tasks_v2(
|
|
|
192
192
|
job_db.executor_error_log = None
|
|
193
193
|
db.merge(job_db)
|
|
194
194
|
db.commit()
|
|
195
|
+
db.expunge_all()
|
|
196
|
+
|
|
197
|
+
# Fail when running a non-converter task on an empty image list
|
|
198
|
+
if (
|
|
199
|
+
wftask.task_type
|
|
200
|
+
in [
|
|
201
|
+
TaskType.COMPOUND,
|
|
202
|
+
TaskType.PARALLEL,
|
|
203
|
+
TaskType.NON_PARALLEL,
|
|
204
|
+
]
|
|
205
|
+
and len(filtered_images) == 0
|
|
206
|
+
):
|
|
207
|
+
error_msg = (
|
|
208
|
+
f"Cannot run task '{task.name}' for an empty image list "
|
|
209
|
+
f"(obtained after applying {type_filters=} and "
|
|
210
|
+
f"attribute_filters={job_attribute_filters})."
|
|
211
|
+
)
|
|
212
|
+
logger.info(error_msg)
|
|
213
|
+
update_status_of_history_run(
|
|
214
|
+
history_run_id=history_run_id,
|
|
215
|
+
status=HistoryUnitStatus.FAILED,
|
|
216
|
+
db_sync=db,
|
|
217
|
+
)
|
|
218
|
+
raise JobExecutionError(error_msg)
|
|
195
219
|
|
|
196
220
|
# TASK EXECUTION (V2)
|
|
197
221
|
try:
|
|
@@ -45,7 +45,7 @@ class TasksPythonSettings(BaseModel):
|
|
|
45
45
|
if self.default_version not in self.versions.keys():
|
|
46
46
|
raise ValueError(
|
|
47
47
|
f"The default Python version ('{self.default_version}') is "
|
|
48
|
-
f"not available
|
|
48
|
+
f"not available in {list(self.versions.keys())}."
|
|
49
49
|
)
|
|
50
50
|
|
|
51
51
|
return self
|
|
@@ -51,8 +51,8 @@ def delete_local(
|
|
|
51
51
|
activity.log = get_current_log(log_file_path)
|
|
52
52
|
activity = add_commit_refresh(obj=activity, db=db)
|
|
53
53
|
|
|
54
|
-
db.commit()
|
|
55
54
|
db.delete(task_group)
|
|
55
|
+
db.commit()
|
|
56
56
|
logger.debug("Task group removed from database.")
|
|
57
57
|
|
|
58
58
|
if task_group.origin != TaskGroupV2OriginEnum.OTHER:
|
|
@@ -1,23 +1,22 @@
|
|
|
1
|
-
fractal_server/__init__.py,sha256=
|
|
2
|
-
fractal_server/__main__.py,sha256=
|
|
1
|
+
fractal_server/__init__.py,sha256=CUZtFSmkfStuE3L_hKV87CswJmhI3gnUuekCrCEQ1Yk,23
|
|
2
|
+
fractal_server/__main__.py,sha256=qLbUicU1Ulaob_Eo5pspi-IH2xAkLfifJTH9gYEhZss,11427
|
|
3
3
|
fractal_server/alembic.ini,sha256=MWwi7GzjzawI9cCAK1LW7NxIBQDUqD12-ptJoq5JpP0,3153
|
|
4
4
|
fractal_server/app/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
5
5
|
fractal_server/app/db/__init__.py,sha256=sttX0mHVV0ESI1SJ1kcxUKiuEwqeP-BWsst0o_9Yo44,2810
|
|
6
|
-
fractal_server/app/models/__init__.py,sha256=
|
|
6
|
+
fractal_server/app/models/__init__.py,sha256=93eFiAaiV6cU_Mf8WQDi_Ir7vqakAiS5H9Pl7qAb4JE,367
|
|
7
7
|
fractal_server/app/models/linkusergroup.py,sha256=3KkkE4QIUAlTrBAZs_tVy0pGvAxUAq6yOEjflct_z2M,678
|
|
8
8
|
fractal_server/app/models/linkuserproject.py,sha256=hvaxh3Lkiy2uUCwB8gvn8RorCpvxSSdzWdCS_U1GL7g,315
|
|
9
|
-
fractal_server/app/models/security.py,sha256=
|
|
10
|
-
fractal_server/app/models/user_settings.py,sha256=u0GOK1JdqDmXzA8hK2JV93rZxY_rF-0oKMkArRolnN8,1201
|
|
9
|
+
fractal_server/app/models/security.py,sha256=f44hOx4Tro0-KLj5N70_gv1UsFrr5ygHY7W_XHFaOJE,3546
|
|
11
10
|
fractal_server/app/models/v2/__init__.py,sha256=A668GF4z_UPar6kAOwC-o_qUo3CIRJ3SmBGYTs3Xc7k,923
|
|
12
11
|
fractal_server/app/models/v2/accounting.py,sha256=i-2TsjqyuclxFQ21C-TeDoss7ZBTRuXdzIJfVr2UxwE,1081
|
|
13
12
|
fractal_server/app/models/v2/dataset.py,sha256=P_zy4dPQAqrCALQ6737VkAFk1SvcgYjnslGUZhPI8sc,1226
|
|
14
13
|
fractal_server/app/models/v2/history.py,sha256=CBN2WVg9vW5pHU1RP8TkB_nnJrwnuifCcxgnd53UtEE,2163
|
|
15
14
|
fractal_server/app/models/v2/job.py,sha256=YYzt3ef2CU1WXFNjlltR3ft2kM9T0Hq8oskSipQSxuM,2042
|
|
16
15
|
fractal_server/app/models/v2/profile.py,sha256=QqOE7XGeq-ckQAbGhcgzDN5zFFaTNrtcuWgOXy9psR8,440
|
|
17
|
-
fractal_server/app/models/v2/project.py,sha256=
|
|
16
|
+
fractal_server/app/models/v2/project.py,sha256=DJgTZG1NTA_pbLlY0Jy3WFBE8X8fBMkVALfDWK5ZuHY,832
|
|
18
17
|
fractal_server/app/models/v2/resource.py,sha256=ReaBGtKb3e0_1PZOZncdGqrttkrC-bsgDCv3wPCGfOs,3512
|
|
19
18
|
fractal_server/app/models/v2/task.py,sha256=iBIQB8POQE5MyKvLZhw7jZWlBhbrThzCDzRTcgiAczQ,1493
|
|
20
|
-
fractal_server/app/models/v2/task_group.py,sha256=
|
|
19
|
+
fractal_server/app/models/v2/task_group.py,sha256=v9hI-R8mXhm0LvE_I_YG3M8VHz0V9DohhGNykdEgcB8,4667
|
|
21
20
|
fractal_server/app/models/v2/workflow.py,sha256=gBjDXO-RytVT81aAlesImBhmVHrwNUrmsF_UsGa1qLM,1057
|
|
22
21
|
fractal_server/app/models/v2/workflowtask.py,sha256=qkTc-hcFLpJUVsEUbnDq2BJL0qg9jagy2doZeusF1ek,1266
|
|
23
22
|
fractal_server/app/routes/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
@@ -26,11 +25,11 @@ fractal_server/app/routes/admin/v2/__init__.py,sha256=SzLurEzKBRWHcIFh2mpac2wrGA
|
|
|
26
25
|
fractal_server/app/routes/admin/v2/_aux_functions.py,sha256=fqA5sUCFuD2iVANQt2WUUfVOEVz5egQA7inzUKYGCw0,1684
|
|
27
26
|
fractal_server/app/routes/admin/v2/accounting.py,sha256=DjgMqzdrL8hYMn19cZj2pkwtLKl3QfBwxd473kH4KaI,3584
|
|
28
27
|
fractal_server/app/routes/admin/v2/impersonate.py,sha256=ictDjuvBr3iLv3YtwkVRMNQRq5qtPAeAXbbC7STSsEg,1125
|
|
29
|
-
fractal_server/app/routes/admin/v2/job.py,sha256=
|
|
28
|
+
fractal_server/app/routes/admin/v2/job.py,sha256=CRt_a6tULsuzN5ojcClzVH3vd7OXT06jK81Qro9fuBk,10171
|
|
30
29
|
fractal_server/app/routes/admin/v2/profile.py,sha256=0Y_1Qv-BA6cHVrxPTDDBOpttpfuJN8g1FqFlG6JiOD8,3164
|
|
31
30
|
fractal_server/app/routes/admin/v2/project.py,sha256=rRq7ZDngr_29skASnte1xfycZCjK-WPdeTf7siBXiCU,1182
|
|
32
31
|
fractal_server/app/routes/admin/v2/resource.py,sha256=UWimApUcL9HPu8NN0ccDbReuKly3aCqff2SA-Y1iEzs,6349
|
|
33
|
-
fractal_server/app/routes/admin/v2/task.py,sha256=
|
|
32
|
+
fractal_server/app/routes/admin/v2/task.py,sha256=tHWCPZRpfvu7k8sQA-634CUC7QbgdJa2RHiMXQL8a58,5599
|
|
34
33
|
fractal_server/app/routes/admin/v2/task_group.py,sha256=EDY9oliXq_xYVJ2HgRuE4-5MbL85j-y4LbWwupZxy38,6249
|
|
35
34
|
fractal_server/app/routes/admin/v2/task_group_lifecycle.py,sha256=W7LjIBAheyjrn0fEz0SsWINqcZK5HMB5GRGMjPrc6a4,9994
|
|
36
35
|
fractal_server/app/routes/api/__init__.py,sha256=ewprevw6hZ0FWM-GPHoQZU0w-yfItqLeQT-Jr_Nbjnw,1658
|
|
@@ -88,21 +87,20 @@ fractal_server/app/schemas/v2/project.py,sha256=7UC0aZLgtmkaAiPykeUj-9OZXhMkoyi3
|
|
|
88
87
|
fractal_server/app/schemas/v2/resource.py,sha256=4iXzZJeHVLcXYY08-okoJM_4gqpzhG4KglRPBm24Jwc,3718
|
|
89
88
|
fractal_server/app/schemas/v2/status_legacy.py,sha256=eQT1zGxbkzSwd0EqclsOdZ60n1x6J3DB1CZ3m4LYyxc,955
|
|
90
89
|
fractal_server/app/schemas/v2/task.py,sha256=IJv8loB4kx9FBkaIHoiMsswQyq02FxvyAnHK1u074fU,4364
|
|
91
|
-
fractal_server/app/schemas/v2/task_collection.py,sha256=
|
|
90
|
+
fractal_server/app/schemas/v2/task_collection.py,sha256=S4fBQ3cbL4YmyL6ZTZYDbIBm6o8D-v5CuYQy4ng3nWU,4581
|
|
92
91
|
fractal_server/app/schemas/v2/task_group.py,sha256=4hNZUXnWYSozpLXR3JqBvGzfZBG2TbjqydckHHu2Aq0,3506
|
|
93
92
|
fractal_server/app/schemas/v2/workflow.py,sha256=L-dW6SzCH_VNoH6ENip44lTgGGqVYHHBk_3PtM-Ooy8,1772
|
|
94
93
|
fractal_server/app/schemas/v2/workflowtask.py,sha256=6eweAMyziwaoMT-7R1fVJYunIeZKzT0-7fAVgPO_FEc,3639
|
|
95
94
|
fractal_server/app/security/__init__.py,sha256=k-La8Da89C1hSUGsiidrWo6Az4u6dbe5PzN1Ctt1t34,18394
|
|
96
95
|
fractal_server/app/security/signup_email.py,sha256=kphjq6TAygvPpYpg95QJWefyqmzdVrGz7fyRMctUJWE,1982
|
|
97
|
-
fractal_server/app/shutdown.py,sha256=
|
|
96
|
+
fractal_server/app/shutdown.py,sha256=QU4DfNvqwUXlHiLORtYJit4DxlFQo014SKTfs4dcE2U,2295
|
|
98
97
|
fractal_server/config/__init__.py,sha256=ZCmroNB50sUxJiFtkW0a4fFtmfyPnL4LWhtKY5FbQfg,737
|
|
99
98
|
fractal_server/config/_data.py,sha256=9Jyt83yrSsr_0_9ANWDAXz88_jjyFlcB5VWJGXq8aUY,2311
|
|
100
|
-
fractal_server/config/_database.py,sha256=
|
|
99
|
+
fractal_server/config/_database.py,sha256=k1z__MrslQjmel34yFvge0sroPUs1vBtT_OSlPY8pN8,1690
|
|
101
100
|
fractal_server/config/_email.py,sha256=j1QmZCyspNbD1xxkypc9Kv299tU3vTO1AqDFJ8-LZzQ,4201
|
|
102
|
-
fractal_server/config/_main.py,sha256=
|
|
101
|
+
fractal_server/config/_main.py,sha256=NmpNuNezVI7MpuAiZ9AOlo3Fc-4x73FIaFm-UMhppEw,1996
|
|
103
102
|
fractal_server/config/_oauth.py,sha256=7J4FphGVFfVmtQycCkas6scEJQJGZUGEzQ-t2PZiqSo,1934
|
|
104
103
|
fractal_server/config/_settings_config.py,sha256=tsyXQOnn9QKCFJD6hRo_dJXlQQyl70DbqgHMJoZ1xnY,144
|
|
105
|
-
fractal_server/data_migrations/2_17_0.py,sha256=ltMrezWWma-KPnZpOzQ12_yebBrjrdu0m7pQQh_RGWM,10894
|
|
106
104
|
fractal_server/data_migrations/README.md,sha256=_3AEFvDg9YkybDqCLlFPdDmGJvr6Tw7HRI14aZ3LOIw,398
|
|
107
105
|
fractal_server/data_migrations/tools.py,sha256=LeMeASwYGtEqd-3wOLle6WARdTGAimoyMmRbbJl-hAM,572
|
|
108
106
|
fractal_server/exceptions.py,sha256=7ftpWwNsTQmNonWCynhH5ErUh1haPPhIaVPrNHla7-o,53
|
|
@@ -122,7 +120,9 @@ fractal_server/migrations/versions/19eca0dd47a9_user_settings_project_dir.py,sha
|
|
|
122
120
|
fractal_server/migrations/versions/1a83a5260664_rename.py,sha256=EkzTAjbJm7CfsLraIUbH9hkTj4M6XvmziVb4K9ZjKmQ,790
|
|
123
121
|
fractal_server/migrations/versions/1eac13a26c83_drop_v1_tables.py,sha256=7OW3HmqAePHx53OWdEPzNxvtupxSR0lB_6tZF1b3JIM,1604
|
|
124
122
|
fractal_server/migrations/versions/316140ff7ee1_remove_usersettings_cache_dir.py,sha256=lANgTox0rz459_yo1Rw7fGCT1qw5sUCUXTLUMc_Bzf8,911
|
|
123
|
+
fractal_server/migrations/versions/45fbb391d7af_make_resource_id_fk_non_nullable.py,sha256=y9zr161YIWgnWbaMg1rahKN4b-vHjT3f5VSeoOAHaqI,1296
|
|
125
124
|
fractal_server/migrations/versions/47351f8c7ebc_drop_dataset_filters.py,sha256=vePkVm1iUHiPNKLQ3KR7BBLdHruqBdl87j_tUCbMbEA,1414
|
|
125
|
+
fractal_server/migrations/versions/49d0856e9569_drop_table.py,sha256=qoq7cGUQmrnUj_wpV2mRqVneyoKqglgbrgzW_8eS_5w,1835
|
|
126
126
|
fractal_server/migrations/versions/4c308bcaea2b_add_task_args_schema_and_task_args_.py,sha256=-wHe-fOffmYeAm0JXVl_lxZ7hhDkaEVqxgxpHkb_uL8,954
|
|
127
127
|
fractal_server/migrations/versions/4cedeb448a53_workflowtask_foreign_keys_not_nullables.py,sha256=Mob8McGYAcmgvrseyyYOa54E6Gsgr-4SiGdC-r9O4_A,1157
|
|
128
128
|
fractal_server/migrations/versions/501961cfcd85_remove_link_between_v1_and_v2_tasks_.py,sha256=JOrVa6mGzqZ6H61YCFVOed64vFRjTWGWyN3z7NE3T08,3270
|
|
@@ -130,6 +130,7 @@ fractal_server/migrations/versions/50a13d6138fd_initial_schema.py,sha256=zwXegXs
|
|
|
130
130
|
fractal_server/migrations/versions/5bf02391cfef_v2.py,sha256=axhNkr_H6R4rRbY7oGYazNbFvPXeSyBDWFVbKNmiqs8,8433
|
|
131
131
|
fractal_server/migrations/versions/70e77f1c38b0_add_applyworkflow_first_task_index_and_.py,sha256=Q-DsMzG3IcUV2Ol1dhJWosDvKERamBE6QvA2zzS5zpQ,1632
|
|
132
132
|
fractal_server/migrations/versions/71eefd1dd202_add_slurm_accounts.py,sha256=mbWuCkTpRAdGbRhW7lhXs_e5S6O37UAcCN6JfoY5H8A,1353
|
|
133
|
+
fractal_server/migrations/versions/7673fe18c05d_remove_project_dir_server_default.py,sha256=LAC1Uv4SeLkqjXPyqj5Mof8L0105gxqS1TYKzNVX4GE,795
|
|
133
134
|
fractal_server/migrations/versions/791ce783d3d8_add_indices.py,sha256=gNE6AgJgeJZY99Fbd336Z9see3gRMQvuNBC0xDk_5sw,1154
|
|
134
135
|
fractal_server/migrations/versions/83bc2ad3ffcc_2_17_0.py,sha256=U7t_8n58taRkd9sxCXOshrTr9M5AhlsQne8SGKa5Jt4,6377
|
|
135
136
|
fractal_server/migrations/versions/84bf0fffde30_add_dumps_to_applyworkflow.py,sha256=NSCuhANChsg76vBkShBl-9tQ4VEHubOjtAv1etHhlvY,2684
|
|
@@ -149,6 +150,7 @@ fractal_server/migrations/versions/af8673379a5c_drop_old_filter_columns.py,sha25
|
|
|
149
150
|
fractal_server/migrations/versions/b1e7f7a1ff71_task_group_for_pixi.py,sha256=loDrqBB-9U3vqLKePEeJy4gK4EuPs_1F345mdrnoCt0,1293
|
|
150
151
|
fractal_server/migrations/versions/b3ffb095f973_json_to_jsonb.py,sha256=Q01lPlBNQgi3hpoUquWj2QUEF7cTsyQ7uikUhWunzWY,10035
|
|
151
152
|
fractal_server/migrations/versions/c90a7c76e996_job_id_in_history_run.py,sha256=Y1cPwmFOZ4mx3v2XZM6adgu8u0L0VD_R4ADURyMb2ro,1102
|
|
153
|
+
fractal_server/migrations/versions/caba9fb1ea5e_drop_useroauth_user_settings_id.py,sha256=8tlWVmux-c-fB9hMO4JEsaPMXRwLN_X3PpC0rUuFrYw,1320
|
|
152
154
|
fractal_server/migrations/versions/d256a7379ab8_taskgroup_activity_and_venv_info_to_.py,sha256=HN3_Pk8G81SzdYjg4K1RZAyjKSlsZGvcYE2nWOUbwxQ,3861
|
|
153
155
|
fractal_server/migrations/versions/d4fe3708d309_make_applyworkflow_workflow_dump_non_.py,sha256=6cHEZFuTXiQg9yu32Y3RH1XAl71av141WQ6UMbiITIg,949
|
|
154
156
|
fractal_server/migrations/versions/da2cb2ac4255_user_group_viewer_paths.py,sha256=yGWSA2HIHUybcVy66xBITk08opV2DFYSCIIrulaUZhI,901
|
|
@@ -198,7 +200,7 @@ fractal_server/runner/v2/_slurm_sudo.py,sha256=f34B6KWtn1Wwm7Y-D95Ecub0vKpA03zYL
|
|
|
198
200
|
fractal_server/runner/v2/db_tools.py,sha256=kLB0sGYNuCMPPoP4XDSClPIf7x5lSseTKUjzGk6gAfc,3318
|
|
199
201
|
fractal_server/runner/v2/deduplicate_list.py,sha256=vidkd7K6u3w0A4zVgsGZkc9mwoP6ihTYJZQUhBNorfE,667
|
|
200
202
|
fractal_server/runner/v2/merge_outputs.py,sha256=0ahaSwdMFAoEhxVaEaO9nSJuKIcWg9pDZ356ktSHcC0,897
|
|
201
|
-
fractal_server/runner/v2/runner.py,sha256=
|
|
203
|
+
fractal_server/runner/v2/runner.py,sha256=PANeskwZpZbbaL8i5vk0voHfP-jli1sZaFKpiW1gq20,20245
|
|
202
204
|
fractal_server/runner/v2/runner_functions.py,sha256=w_i74LCt_9f07w511wslTFhoDUtoE1R-IKcglEGWPIc,19618
|
|
203
205
|
fractal_server/runner/v2/submit_workflow.py,sha256=s3mo5pLZH5x8V01IZjoJXcbpCl9geZwvISEr5StSBeI,11458
|
|
204
206
|
fractal_server/runner/v2/task_interface.py,sha256=ftPPpOU16rbJD8q-QV7o_3ey8W7MQTFuWJiYUr4OmF4,2532
|
|
@@ -210,7 +212,7 @@ fractal_server/syringe.py,sha256=3YJeIALH-wibuJ9R5VMNYUWh7x1-MkWT0SqGcWG5MY8,279
|
|
|
210
212
|
fractal_server/tasks/__init__.py,sha256=kadmVUoIghl8s190_Tt-8f-WBqMi8u8oU4Pvw39NHE8,23
|
|
211
213
|
fractal_server/tasks/config/__init__.py,sha256=f1kARvKJPdM3b0HmJhIyMRxi8RCgPajUfS7RBhNfLKY,152
|
|
212
214
|
fractal_server/tasks/config/_pixi.py,sha256=KVf0xYAf6hAnOKSrTmg_GYJ5VaA19kEMAydRJNHjeiA,3773
|
|
213
|
-
fractal_server/tasks/config/_python.py,sha256=
|
|
215
|
+
fractal_server/tasks/config/_python.py,sha256=uEBm4Z9Ah-7A7pXBcSTBhFgefSuJPwN8WrWQHoNdp9M,1353
|
|
214
216
|
fractal_server/tasks/utils.py,sha256=V7dj8o2AnoHhGSTYlqJHcRFhCIpmOrMOUhtiE_DvRVA,291
|
|
215
217
|
fractal_server/tasks/v2/__init__.py,sha256=47DEQpj8HBSa-_TImW-5JCeuQeRkm5NMpJWZG3hSuFU,0
|
|
216
218
|
fractal_server/tasks/v2/local/__init__.py,sha256=2dJqJIwn9UEANg8lPFprxVLk3ug-4xYIb5pvIsqPb5s,353
|
|
@@ -219,7 +221,7 @@ fractal_server/tasks/v2/local/collect.py,sha256=2dvR90b57gCz8F_bmNoMLtnvlrW1ZLBz
|
|
|
219
221
|
fractal_server/tasks/v2/local/collect_pixi.py,sha256=BTWwu07LZQB0zTvCpHJkyEHgUl8XlF1qSN5NsOC5MVo,11489
|
|
220
222
|
fractal_server/tasks/v2/local/deactivate.py,sha256=cPJbB-zoHjmxL5goaNAUItkW0o_fg-YAd2YuXu_n8Rg,9925
|
|
221
223
|
fractal_server/tasks/v2/local/deactivate_pixi.py,sha256=h1XdAwihACywC9qak5K2tlfz4y32zZJOA-OEkIVy67U,3627
|
|
222
|
-
fractal_server/tasks/v2/local/delete.py,sha256=
|
|
224
|
+
fractal_server/tasks/v2/local/delete.py,sha256=rmAjZ5i8x0jAIqOjG7eH9XFtUZkeo172S2WyZRFIVq0,2960
|
|
223
225
|
fractal_server/tasks/v2/local/reactivate.py,sha256=ZQTaqPquevpwdchUJIp59Lw01vErA3T2XN8g1jMBNAE,6013
|
|
224
226
|
fractal_server/tasks/v2/local/reactivate_pixi.py,sha256=R0D3bUwrAo3uyDNZRYsQ65NMl7-nhqd1GefBAh9Hk5c,8159
|
|
225
227
|
fractal_server/tasks/v2/ssh/__init__.py,sha256=dPK6BtEZVh1GiFP05j1RKTEnZvjJez8o2KkMC2hWXaw,339
|
|
@@ -255,8 +257,8 @@ fractal_server/types/validators/_workflow_task_arguments_validators.py,sha256=HL
|
|
|
255
257
|
fractal_server/urls.py,sha256=QjIKAC1a46bCdiPMu3AlpgFbcv6a4l3ABcd5xz190Og,471
|
|
256
258
|
fractal_server/utils.py,sha256=SYVVUuXe_nWyrJLsy7QA-KJscwc5PHEXjvsW4TK7XQI,2180
|
|
257
259
|
fractal_server/zip_tools.py,sha256=H0w7wS5yE4ebj7hw1_77YQ959dl2c-L0WX6J_ro1TY4,4884
|
|
258
|
-
fractal_server-2.17.
|
|
259
|
-
fractal_server-2.17.
|
|
260
|
-
fractal_server-2.17.
|
|
261
|
-
fractal_server-2.17.
|
|
262
|
-
fractal_server-2.17.
|
|
260
|
+
fractal_server-2.17.1.dist-info/METADATA,sha256=fxvj3AoIvayiofwYRfl1dwAp2cQcdCZ0cZPnQoEoOBg,4224
|
|
261
|
+
fractal_server-2.17.1.dist-info/WHEEL,sha256=zp0Cn7JsFoX2ATtOhtaFYIiE2rmFAD4OcMhtUki8W3U,88
|
|
262
|
+
fractal_server-2.17.1.dist-info/entry_points.txt,sha256=8tV2kynvFkjnhbtDnxAqImL6HMVKsopgGfew0DOp5UY,58
|
|
263
|
+
fractal_server-2.17.1.dist-info/licenses/LICENSE,sha256=QKAharUuhxL58kSoLizKJeZE3mTCBnX6ucmz8W0lxlk,1576
|
|
264
|
+
fractal_server-2.17.1.dist-info/RECORD,,
|
|
@@ -1,37 +0,0 @@
|
|
|
1
|
-
from sqlalchemy import Column
|
|
2
|
-
from sqlalchemy.dialects.postgresql import JSONB
|
|
3
|
-
from sqlmodel import Field
|
|
4
|
-
from sqlmodel import SQLModel
|
|
5
|
-
|
|
6
|
-
|
|
7
|
-
# TODO-2.17.1: Drop `UserSettings`
|
|
8
|
-
class UserSettings(SQLModel, table=True):
|
|
9
|
-
"""
|
|
10
|
-
Comprehensive list of user settings.
|
|
11
|
-
|
|
12
|
-
Attributes:
|
|
13
|
-
id: ID of database object
|
|
14
|
-
slurm_accounts:
|
|
15
|
-
List of SLURM accounts, to be used upon Fractal job submission.
|
|
16
|
-
ssh_host: SSH-reachable host where a SLURM client is available.
|
|
17
|
-
ssh_username: User on `ssh_host`.
|
|
18
|
-
ssh_private_key_path: Path of private SSH key for `ssh_username`.
|
|
19
|
-
slurm_user: Local user, to be impersonated via `sudo -u`
|
|
20
|
-
project_dir: Folder where `slurm_user` can write.
|
|
21
|
-
"""
|
|
22
|
-
|
|
23
|
-
__tablename__ = "user_settings"
|
|
24
|
-
|
|
25
|
-
id: int | None = Field(default=None, primary_key=True)
|
|
26
|
-
slurm_accounts: list[str] = Field(
|
|
27
|
-
sa_column=Column(JSONB, server_default="[]", nullable=False)
|
|
28
|
-
)
|
|
29
|
-
ssh_host: str | None = None
|
|
30
|
-
ssh_username: str | None = None
|
|
31
|
-
ssh_private_key_path: str | None = None
|
|
32
|
-
|
|
33
|
-
slurm_user: str | None = None
|
|
34
|
-
project_dir: str | None = None
|
|
35
|
-
|
|
36
|
-
ssh_tasks_dir: str | None = None
|
|
37
|
-
ssh_jobs_dir: str | None = None
|
|
@@ -1,321 +0,0 @@
|
|
|
1
|
-
"""
|
|
2
|
-
|
|
3
|
-
PRELIMINARY CHECKS (TO DO WITH 2.16)
|
|
4
|
-
* All users who are meant to actually use Fractal must be marked as active and verified.
|
|
5
|
-
* All users who are not active and verified will still be able to log in, but they won't have access to the rest of the API.
|
|
6
|
-
* All active users must have `project_dir` set, in their user settings.
|
|
7
|
-
* `FRACTAL_SLURM_WORKER_PYTHON` must be included explicitly in the old env file.
|
|
8
|
-
|
|
9
|
-
DATA-MIGRATION REQUIREMENTS
|
|
10
|
-
* Old `.fractal_server.env`, renamed into `.fractal_server.env.old`.
|
|
11
|
-
* New `.fractal_server.env` - see XXX for list of changes.
|
|
12
|
-
* Old JSON file with SLURM configuration.
|
|
13
|
-
* Old JSON file with pixi configuration - if applicable.
|
|
14
|
-
|
|
15
|
-
|
|
16
|
-
MANUAL FIXES POST DATA MIGRATION:
|
|
17
|
-
* Rename resource
|
|
18
|
-
* Rename profiles - if needed
|
|
19
|
-
"""
|
|
20
|
-
import json
|
|
21
|
-
import logging
|
|
22
|
-
import sys
|
|
23
|
-
from typing import Any
|
|
24
|
-
|
|
25
|
-
from devtools import debug
|
|
26
|
-
from dotenv.main import DotEnv
|
|
27
|
-
from pydantic import BaseModel
|
|
28
|
-
from sqlalchemy.orm import Session
|
|
29
|
-
from sqlalchemy.sql.operators import is_
|
|
30
|
-
from sqlalchemy.sql.operators import is_not
|
|
31
|
-
from sqlmodel import select
|
|
32
|
-
|
|
33
|
-
from fractal_server.app.db import get_sync_db
|
|
34
|
-
from fractal_server.app.models import Profile
|
|
35
|
-
from fractal_server.app.models import ProjectV2
|
|
36
|
-
from fractal_server.app.models import Resource
|
|
37
|
-
from fractal_server.app.models import TaskGroupV2
|
|
38
|
-
from fractal_server.app.models import UserOAuth
|
|
39
|
-
from fractal_server.app.models import UserSettings
|
|
40
|
-
from fractal_server.app.schemas.v2.profile import cast_serialize_profile
|
|
41
|
-
from fractal_server.app.schemas.v2.resource import cast_serialize_resource
|
|
42
|
-
from fractal_server.config import get_settings
|
|
43
|
-
from fractal_server.runner.config import JobRunnerConfigSLURM
|
|
44
|
-
from fractal_server.tasks.config import TasksPixiSettings
|
|
45
|
-
from fractal_server.tasks.config import TasksPythonSettings
|
|
46
|
-
from fractal_server.types import AbsolutePathStr
|
|
47
|
-
from fractal_server.types import ListUniqueNonEmptyString
|
|
48
|
-
from fractal_server.urls import normalize_url
|
|
49
|
-
|
|
50
|
-
logger = logging.getLogger("fix_db")
|
|
51
|
-
logger.setLevel(logging.INFO)
|
|
52
|
-
|
|
53
|
-
|
|
54
|
-
class UserUpdateInfo(BaseModel):
|
|
55
|
-
user_id: int
|
|
56
|
-
project_dir: AbsolutePathStr
|
|
57
|
-
slurm_accounts: ListUniqueNonEmptyString
|
|
58
|
-
|
|
59
|
-
|
|
60
|
-
class ProfileUsersUpdateInfo(BaseModel):
|
|
61
|
-
data: dict[str, Any]
|
|
62
|
-
user_updates: list[UserUpdateInfo]
|
|
63
|
-
|
|
64
|
-
|
|
65
|
-
def _get_user_settings(user: UserOAuth, db: Session) -> UserSettings:
|
|
66
|
-
if user.user_settings_id is None:
|
|
67
|
-
sys.exit(f"User {user.email} is active but {user.user_settings_id=}.")
|
|
68
|
-
user_settings = db.get(UserSettings, user.user_settings_id)
|
|
69
|
-
return user_settings
|
|
70
|
-
|
|
71
|
-
|
|
72
|
-
def assert_user_setting_key(
|
|
73
|
-
user: UserOAuth,
|
|
74
|
-
user_settings: UserSettings,
|
|
75
|
-
keys: list[str],
|
|
76
|
-
) -> None:
|
|
77
|
-
for key in keys:
|
|
78
|
-
if getattr(user_settings, key) is None:
|
|
79
|
-
sys.exit(
|
|
80
|
-
f"User {user.email} is active and verified but their "
|
|
81
|
-
f"user settings have {key}=None."
|
|
82
|
-
)
|
|
83
|
-
|
|
84
|
-
|
|
85
|
-
def prepare_profile_and_user_updates() -> dict[str, ProfileUsersUpdateInfo]:
|
|
86
|
-
settings = get_settings()
|
|
87
|
-
profiles_and_users: dict[str, ProfileUsersUpdateInfo] = {}
|
|
88
|
-
with next(get_sync_db()) as db:
|
|
89
|
-
# Get active&verified users
|
|
90
|
-
res = db.execute(
|
|
91
|
-
select(UserOAuth)
|
|
92
|
-
.where(is_(UserOAuth.is_active, True))
|
|
93
|
-
.where(is_(UserOAuth.is_verified, True))
|
|
94
|
-
.order_by(UserOAuth.id)
|
|
95
|
-
)
|
|
96
|
-
for user in res.unique().scalars().all():
|
|
97
|
-
# Get user settings
|
|
98
|
-
user_settings = _get_user_settings(user=user, db=db)
|
|
99
|
-
assert_user_setting_key(user, user_settings, ["project_dir"])
|
|
100
|
-
|
|
101
|
-
# Prepare profile data and user update
|
|
102
|
-
new_profile_data = dict()
|
|
103
|
-
if settings.FRACTAL_RUNNER_BACKEND == "slurm_sudo":
|
|
104
|
-
assert_user_setting_key(user, user_settings, ["slurm_user"])
|
|
105
|
-
username = user_settings.slurm_user
|
|
106
|
-
elif settings.FRACTAL_RUNNER_BACKEND == "slurm_ssh":
|
|
107
|
-
assert_user_setting_key(
|
|
108
|
-
user,
|
|
109
|
-
user_settings,
|
|
110
|
-
[
|
|
111
|
-
"ssh_username",
|
|
112
|
-
"ssh_private_key_path",
|
|
113
|
-
"ssh_tasks_dir",
|
|
114
|
-
"ssh_jobs_dir",
|
|
115
|
-
],
|
|
116
|
-
)
|
|
117
|
-
username = user_settings.ssh_username
|
|
118
|
-
new_profile_data.update(
|
|
119
|
-
ssh_key_path=user_settings.ssh_private_key_path,
|
|
120
|
-
tasks_remote_dir=normalize_url(
|
|
121
|
-
user_settings.ssh_tasks_dir
|
|
122
|
-
),
|
|
123
|
-
jobs_remote_dir=normalize_url(user_settings.ssh_jobs_dir),
|
|
124
|
-
)
|
|
125
|
-
|
|
126
|
-
new_profile_data.update(
|
|
127
|
-
name=f"Profile {username}",
|
|
128
|
-
username=username,
|
|
129
|
-
resource_type=settings.FRACTAL_RUNNER_BACKEND,
|
|
130
|
-
)
|
|
131
|
-
debug(new_profile_data)
|
|
132
|
-
cast_serialize_profile(new_profile_data)
|
|
133
|
-
|
|
134
|
-
user_update_info = UserUpdateInfo(
|
|
135
|
-
user_id=user.id,
|
|
136
|
-
project_dir=normalize_url(user_settings.project_dir),
|
|
137
|
-
slurm_accounts=user_settings.slurm_accounts or [],
|
|
138
|
-
)
|
|
139
|
-
|
|
140
|
-
if username in profiles_and_users.keys():
|
|
141
|
-
if profiles_and_users[username].data != new_profile_data:
|
|
142
|
-
# FIXME
|
|
143
|
-
debug(new_profile_data)
|
|
144
|
-
debug(profiles_and_users[username].data)
|
|
145
|
-
raise ValueError()
|
|
146
|
-
profiles_and_users[username].user_updates.append(
|
|
147
|
-
user_update_info
|
|
148
|
-
)
|
|
149
|
-
else:
|
|
150
|
-
profiles_and_users[username] = ProfileUsersUpdateInfo(
|
|
151
|
-
data=new_profile_data,
|
|
152
|
-
user_updates=[user_update_info],
|
|
153
|
-
)
|
|
154
|
-
|
|
155
|
-
return profiles_and_users
|
|
156
|
-
|
|
157
|
-
|
|
158
|
-
def get_old_dotenv_variables() -> dict[str, str | None]:
|
|
159
|
-
"""
|
|
160
|
-
See
|
|
161
|
-
https://github.com/fractal-analytics-platform/fractal-server/blob/2.16.x/fractal_server/config.py
|
|
162
|
-
"""
|
|
163
|
-
OLD_DOTENV_FILE = ".fractal_server.env.old"
|
|
164
|
-
return dict(
|
|
165
|
-
**DotEnv(
|
|
166
|
-
dotenv_path=OLD_DOTENV_FILE,
|
|
167
|
-
override=False,
|
|
168
|
-
).dict()
|
|
169
|
-
)
|
|
170
|
-
|
|
171
|
-
|
|
172
|
-
def get_TasksPythonSettings(
|
|
173
|
-
old_config: dict[str, str | None]
|
|
174
|
-
) -> dict[str, Any]:
|
|
175
|
-
versions = {}
|
|
176
|
-
for version_underscore in ["3_9", "3_10", "3_11", "3_12"]:
|
|
177
|
-
key = f"FRACTAL_TASKS_PYTHON_{version_underscore}"
|
|
178
|
-
version_dot = version_underscore.replace("_", ".")
|
|
179
|
-
value = old_config.get(key, None)
|
|
180
|
-
if value is not None:
|
|
181
|
-
versions[version_dot] = value
|
|
182
|
-
obj = TasksPythonSettings(
|
|
183
|
-
default_version=old_config["FRACTAL_TASKS_PYTHON_DEFAULT_VERSION"],
|
|
184
|
-
versions=versions,
|
|
185
|
-
pip_cache_dir=old_config.get("FRACTAL_PIP_CACHE_DIR", None),
|
|
186
|
-
)
|
|
187
|
-
return obj.model_dump()
|
|
188
|
-
|
|
189
|
-
|
|
190
|
-
def get_TasksPixiSettings(old_config: dict[str, str | None]) -> dict[str, Any]:
|
|
191
|
-
pixi_file = old_config.get("FRACTAL_PIXI_CONFIG_FILE", None)
|
|
192
|
-
if pixi_file is None:
|
|
193
|
-
return {}
|
|
194
|
-
with open(pixi_file) as f:
|
|
195
|
-
old_pixi_config = json.load(f)
|
|
196
|
-
TasksPixiSettings(**old_pixi_config)
|
|
197
|
-
return old_pixi_config
|
|
198
|
-
|
|
199
|
-
|
|
200
|
-
def get_JobRunnerConfigSLURM(
|
|
201
|
-
old_config: dict[str, str | None]
|
|
202
|
-
) -> dict[str, Any]:
|
|
203
|
-
slurm_file = old_config["FRACTAL_SLURM_CONFIG_FILE"]
|
|
204
|
-
with open(slurm_file) as f:
|
|
205
|
-
old_slurm_config = json.load(f)
|
|
206
|
-
JobRunnerConfigSLURM(**old_slurm_config)
|
|
207
|
-
return old_slurm_config
|
|
208
|
-
|
|
209
|
-
|
|
210
|
-
def get_ssh_host() -> str:
|
|
211
|
-
with next(get_sync_db()) as db:
|
|
212
|
-
res = db.execute(
|
|
213
|
-
select(UserSettings.ssh_host).where(
|
|
214
|
-
is_not(UserSettings.ssh_host, None)
|
|
215
|
-
)
|
|
216
|
-
)
|
|
217
|
-
hosts = res.scalars().all()
|
|
218
|
-
if len(set(hosts)) > 1:
|
|
219
|
-
host = max(set(hosts), key=hosts.count)
|
|
220
|
-
print(f"MOST FREQUENT HOST: {host}")
|
|
221
|
-
else:
|
|
222
|
-
host = hosts[0]
|
|
223
|
-
return host
|
|
224
|
-
|
|
225
|
-
|
|
226
|
-
def prepare_resource_data(old_config: dict[str, str | None]) -> dict[str, Any]:
|
|
227
|
-
settings = get_settings()
|
|
228
|
-
|
|
229
|
-
resource_data = dict(
|
|
230
|
-
type=settings.FRACTAL_RUNNER_BACKEND,
|
|
231
|
-
name="Resource Name",
|
|
232
|
-
tasks_python_config=get_TasksPythonSettings(old_config),
|
|
233
|
-
tasks_pixi_config=get_TasksPixiSettings(old_config),
|
|
234
|
-
jobs_runner_config=get_JobRunnerConfigSLURM(old_config),
|
|
235
|
-
tasks_local_dir=old_config["FRACTAL_TASKS_DIR"],
|
|
236
|
-
jobs_local_dir=old_config["FRACTAL_RUNNER_WORKING_BASE_DIR"],
|
|
237
|
-
jobs_slurm_python_worker=old_config["FRACTAL_SLURM_WORKER_PYTHON"],
|
|
238
|
-
jobs_poll_interval=int(
|
|
239
|
-
old_config.get("FRACTAL_SLURM_POLL_INTERVAL", 15)
|
|
240
|
-
),
|
|
241
|
-
)
|
|
242
|
-
if settings.FRACTAL_RUNNER_BACKEND == "slurm_ssh":
|
|
243
|
-
resource_data["host"] = get_ssh_host()
|
|
244
|
-
|
|
245
|
-
resource_data = cast_serialize_resource(resource_data)
|
|
246
|
-
|
|
247
|
-
return resource_data
|
|
248
|
-
|
|
249
|
-
|
|
250
|
-
def fix_db():
|
|
251
|
-
# READ-ONLY CHECK
|
|
252
|
-
|
|
253
|
-
settings = get_settings()
|
|
254
|
-
|
|
255
|
-
# Verify that we are in a SLURM instance
|
|
256
|
-
if settings.FRACTAL_RUNNER_BACKEND == "local":
|
|
257
|
-
sys.exit(
|
|
258
|
-
"ERROR: FRACTAL_RUNNER_BACKEND='local' is not "
|
|
259
|
-
"supported for this data migration."
|
|
260
|
-
)
|
|
261
|
-
|
|
262
|
-
# Read old env file
|
|
263
|
-
old_config = get_old_dotenv_variables()
|
|
264
|
-
|
|
265
|
-
# Prepare resource data
|
|
266
|
-
resource_data = prepare_resource_data(old_config)
|
|
267
|
-
|
|
268
|
-
# Prepare profile/users data
|
|
269
|
-
profile_and_user_updates = prepare_profile_and_user_updates()
|
|
270
|
-
|
|
271
|
-
# ---------------------------------------
|
|
272
|
-
|
|
273
|
-
# WRITES
|
|
274
|
-
|
|
275
|
-
with next(get_sync_db()) as db:
|
|
276
|
-
# Create new resource
|
|
277
|
-
resource = Resource(**resource_data)
|
|
278
|
-
db.add(resource)
|
|
279
|
-
db.commit()
|
|
280
|
-
db.refresh(resource)
|
|
281
|
-
db.expunge(resource)
|
|
282
|
-
resource_id = resource.id
|
|
283
|
-
debug(f"CREATED RESOURCE with {resource_id=}")
|
|
284
|
-
|
|
285
|
-
# Update task groups
|
|
286
|
-
res = db.execute(select(TaskGroupV2).order_by(TaskGroupV2.id))
|
|
287
|
-
for taskgroup in res.scalars().all():
|
|
288
|
-
taskgroup.resource_id = resource_id
|
|
289
|
-
db.add(taskgroup)
|
|
290
|
-
db.commit()
|
|
291
|
-
|
|
292
|
-
# Update projects
|
|
293
|
-
res = db.execute(select(ProjectV2).order_by(ProjectV2.id))
|
|
294
|
-
for project in res.scalars().all():
|
|
295
|
-
project.resource_id = resource_id
|
|
296
|
-
db.add(project)
|
|
297
|
-
db.commit()
|
|
298
|
-
|
|
299
|
-
db.expunge_all()
|
|
300
|
-
|
|
301
|
-
for _, info in profile_and_user_updates.items():
|
|
302
|
-
debug(info)
|
|
303
|
-
|
|
304
|
-
# Create profile
|
|
305
|
-
profile_data = info.data
|
|
306
|
-
profile_data["resource_id"] = resource_id
|
|
307
|
-
profile = Profile(**profile_data)
|
|
308
|
-
db.add(profile)
|
|
309
|
-
db.commit()
|
|
310
|
-
db.refresh(profile)
|
|
311
|
-
db.expunge(profile)
|
|
312
|
-
profile_id = profile.id
|
|
313
|
-
|
|
314
|
-
# Update users
|
|
315
|
-
for user_update in info.user_updates:
|
|
316
|
-
user = db.get(UserOAuth, user_update.user_id)
|
|
317
|
-
user.profile_id = profile_id
|
|
318
|
-
user.project_dir = user_update.project_dir
|
|
319
|
-
user.slurm_accounts = user_update.slurm_accounts
|
|
320
|
-
db.add(user)
|
|
321
|
-
db.commit()
|
|
File without changes
|
|
File without changes
|
|
File without changes
|