pybiolib 1.2.1056__py3-none-any.whl → 1.2.1727__py3-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of pybiolib might be problematic. Click here for more details.
- biolib/__init__.py +33 -10
- biolib/_data_record/data_record.py +103 -26
- biolib/_index/__init__.py +0 -0
- biolib/_index/index.py +51 -0
- biolib/_index/types.py +7 -0
- biolib/_internal/data_record/data_record.py +1 -1
- biolib/_internal/data_record/push_data.py +65 -16
- biolib/_internal/data_record/remote_storage_endpoint.py +3 -3
- biolib/_internal/file_utils.py +7 -4
- biolib/_internal/index/__init__.py +1 -0
- biolib/_internal/index/index.py +18 -0
- biolib/_internal/lfs/cache.py +4 -2
- biolib/_internal/push_application.py +89 -23
- biolib/_internal/runtime.py +2 -0
- biolib/_internal/templates/gui_template/App.tsx +38 -2
- biolib/_internal/templates/gui_template/Dockerfile +2 -0
- biolib/_internal/templates/gui_template/biolib-sdk.ts +37 -0
- biolib/_internal/templates/gui_template/dev-data/output.json +7 -0
- biolib/_internal/templates/gui_template/package.json +1 -0
- biolib/_internal/templates/gui_template/vite-plugin-dev-data.ts +49 -0
- biolib/_internal/templates/gui_template/vite.config.mts +2 -1
- biolib/_internal/templates/init_template/.github/workflows/biolib.yml +6 -1
- biolib/_internal/templates/init_template/Dockerfile +2 -0
- biolib/_internal/utils/__init__.py +40 -0
- biolib/_internal/utils/auth.py +46 -0
- biolib/_internal/utils/job_url.py +33 -0
- biolib/_runtime/runtime.py +9 -0
- biolib/_session/session.py +7 -5
- biolib/_shared/__init__.py +0 -0
- biolib/_shared/types/__init__.py +74 -0
- biolib/_shared/types/resource.py +37 -0
- biolib/_shared/types/resource_deploy_key.py +11 -0
- biolib/{_internal → _shared}/types/resource_version.py +8 -2
- biolib/_shared/utils/__init__.py +7 -0
- biolib/_shared/utils/resource_uri.py +75 -0
- biolib/api/client.py +3 -47
- biolib/app/app.py +57 -33
- biolib/biolib_api_client/api_client.py +3 -47
- biolib/biolib_api_client/app_types.py +1 -6
- biolib/biolib_api_client/biolib_app_api.py +17 -0
- biolib/biolib_binary_format/module_input.py +8 -0
- biolib/biolib_binary_format/remote_endpoints.py +3 -3
- biolib/biolib_binary_format/remote_stream_seeker.py +39 -25
- biolib/cli/__init__.py +2 -1
- biolib/cli/data_record.py +82 -0
- biolib/cli/index.py +32 -0
- biolib/cli/init.py +39 -1
- biolib/cli/lfs.py +1 -1
- biolib/cli/run.py +8 -5
- biolib/cli/start.py +14 -1
- biolib/compute_node/job_worker/executors/docker_executor.py +31 -9
- biolib/compute_node/job_worker/executors/docker_types.py +1 -1
- biolib/compute_node/job_worker/executors/types.py +6 -5
- biolib/compute_node/job_worker/job_worker.py +149 -93
- biolib/compute_node/job_worker/large_file_system.py +2 -6
- biolib/compute_node/job_worker/network_alloc.py +99 -0
- biolib/compute_node/job_worker/network_buffer.py +240 -0
- biolib/compute_node/job_worker/utilization_reporter_thread.py +2 -2
- biolib/compute_node/remote_host_proxy.py +139 -79
- biolib/compute_node/utils.py +2 -0
- biolib/compute_node/webserver/compute_node_results_proxy.py +188 -0
- biolib/compute_node/webserver/proxy_utils.py +28 -0
- biolib/compute_node/webserver/webserver.py +64 -19
- biolib/experiments/experiment.py +111 -16
- biolib/jobs/job.py +119 -29
- biolib/jobs/job_result.py +70 -33
- biolib/jobs/types.py +1 -0
- biolib/sdk/__init__.py +17 -2
- biolib/typing_utils.py +1 -1
- biolib/utils/cache_state.py +2 -2
- biolib/utils/multipart_uploader.py +24 -18
- biolib/utils/seq_util.py +1 -1
- pybiolib-1.2.1727.dist-info/METADATA +41 -0
- {pybiolib-1.2.1056.dist-info → pybiolib-1.2.1727.dist-info}/RECORD +103 -85
- {pybiolib-1.2.1056.dist-info → pybiolib-1.2.1727.dist-info}/WHEEL +1 -1
- pybiolib-1.2.1727.dist-info/entry_points.txt +2 -0
- biolib/_internal/types/__init__.py +0 -6
- biolib/_internal/types/resource.py +0 -18
- biolib/utils/app_uri.py +0 -57
- pybiolib-1.2.1056.dist-info/METADATA +0 -50
- pybiolib-1.2.1056.dist-info/entry_points.txt +0 -3
- /biolib/{_internal → _shared}/types/account.py +0 -0
- /biolib/{_internal → _shared}/types/account_member.py +0 -0
- /biolib/{_internal → _shared}/types/app.py +0 -0
- /biolib/{_internal → _shared}/types/data_record.py +0 -0
- /biolib/{_internal → _shared}/types/experiment.py +0 -0
- /biolib/{_internal → _shared}/types/file_node.py +0 -0
- /biolib/{_internal → _shared}/types/push.py +0 -0
- /biolib/{_internal → _shared}/types/resource_permission.py +0 -0
- /biolib/{_internal → _shared}/types/result.py +0 -0
- /biolib/{_internal → _shared}/types/typing.py +0 -0
- /biolib/{_internal → _shared}/types/user.py +0 -0
- {pybiolib-1.2.1056.dist-info → pybiolib-1.2.1727.dist-info/licenses}/LICENSE +0 -0
|
@@ -1,4 +1,5 @@
|
|
|
1
1
|
from biolib.biolib_binary_format.base_bbf_package import BioLibBinaryFormatBasePackage
|
|
2
|
+
from biolib.biolib_logging import logger
|
|
2
3
|
from biolib.typing_utils import TypedDict, Dict, List
|
|
3
4
|
|
|
4
5
|
|
|
@@ -14,6 +15,10 @@ class ModuleInput(BioLibBinaryFormatBasePackage):
|
|
|
14
15
|
self.package_type = 1
|
|
15
16
|
|
|
16
17
|
def serialize(self, stdin, arguments, files) -> bytes:
|
|
18
|
+
for path in files.keys():
|
|
19
|
+
if '//' in path:
|
|
20
|
+
raise ValueError(f"File path '{path}' contains double slashes which are not allowed")
|
|
21
|
+
|
|
17
22
|
bbf_data = bytearray()
|
|
18
23
|
bbf_data.extend(self.version.to_bytes(1, 'big'))
|
|
19
24
|
bbf_data.extend(self.package_type.to_bytes(1, 'big'))
|
|
@@ -67,6 +72,9 @@ class ModuleInput(BioLibBinaryFormatBasePackage):
|
|
|
67
72
|
data_len = self.get_data(8, output_type='int')
|
|
68
73
|
path = self.get_data(path_len, output_type='str')
|
|
69
74
|
data = self.get_data(data_len)
|
|
75
|
+
if '//' in path:
|
|
76
|
+
# TODO: Raise ValueError here once backwards compatibility period is over
|
|
77
|
+
logger.warning(f"File path '{path}' contains double slashes which are not allowed")
|
|
70
78
|
files[path] = bytes(data)
|
|
71
79
|
|
|
72
80
|
return ModuleInputDict(stdin=stdin, arguments=arguments, files=files)
|
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
from datetime import datetime, timedelta
|
|
1
|
+
from datetime import datetime, timedelta, timezone
|
|
2
2
|
|
|
3
3
|
from biolib.biolib_api_client.biolib_job_api import BiolibJobApi
|
|
4
4
|
from biolib.biolib_binary_format.utils import RemoteEndpoint
|
|
@@ -17,13 +17,13 @@ class RemoteJobStorageEndpoint(RemoteEndpoint):
|
|
|
17
17
|
self._storage_type: Literal['input', 'output'] = storage_type
|
|
18
18
|
|
|
19
19
|
def get_remote_url(self):
|
|
20
|
-
if not self._presigned_url or datetime.
|
|
20
|
+
if not self._presigned_url or not self._expires_at or datetime.now(timezone.utc) > self._expires_at:
|
|
21
21
|
self._presigned_url = BiolibJobApi.get_job_storage_download_url(
|
|
22
22
|
job_auth_token=self._job_auth_token,
|
|
23
23
|
job_uuid=self._job_uuid,
|
|
24
24
|
storage_type='results' if self._storage_type == 'output' else 'input',
|
|
25
25
|
)
|
|
26
|
-
self._expires_at = datetime.
|
|
26
|
+
self._expires_at = datetime.now(timezone.utc) + timedelta(minutes=8)
|
|
27
27
|
# TODO: Use expires at from url
|
|
28
28
|
# parsed_url = urlparse(self._presigned_url)
|
|
29
29
|
# query_params = parse_qs(parsed_url.query)
|
|
@@ -1,45 +1,59 @@
|
|
|
1
1
|
from biolib.biolib_binary_format.utils import IndexableBuffer
|
|
2
|
+
from biolib.biolib_logging import logger
|
|
2
3
|
from biolib.typing_utils import Iterable
|
|
3
4
|
|
|
4
5
|
|
|
5
6
|
class StreamSeeker:
|
|
6
7
|
def __init__(
|
|
7
|
-
|
|
8
|
-
|
|
9
|
-
|
|
10
|
-
|
|
11
|
-
|
|
8
|
+
self,
|
|
9
|
+
upstream_buffer: IndexableBuffer,
|
|
10
|
+
files_data_start: int,
|
|
11
|
+
files_data_end: int,
|
|
12
|
+
max_chunk_size: int,
|
|
12
13
|
):
|
|
13
14
|
self._upstream_buffer = upstream_buffer
|
|
14
15
|
self._files_data_end = files_data_end
|
|
15
|
-
self.
|
|
16
|
+
self._max_chunk_size = max_chunk_size
|
|
16
17
|
|
|
17
18
|
self._buffer_start = files_data_start
|
|
18
19
|
self._buffer = bytearray()
|
|
19
20
|
|
|
20
|
-
def seek_and_read(self, file_start: int, file_length: int) -> Iterable[bytes]:
|
|
21
|
+
def seek_and_read(self, file_start: int, file_length: int, read_ahead_bytes: int = 0) -> Iterable[bytes]:
|
|
21
22
|
assert file_start >= self._buffer_start
|
|
22
|
-
self._buffer = self._buffer[file_start - self._buffer_start:]
|
|
23
|
+
self._buffer = self._buffer[file_start - self._buffer_start :]
|
|
23
24
|
self._buffer_start = file_start
|
|
24
25
|
|
|
25
26
|
while True:
|
|
26
27
|
file_byte_count_remaining = file_length - (self._buffer_start - file_start)
|
|
27
|
-
if file_byte_count_remaining
|
|
28
|
+
if file_byte_count_remaining <= 0:
|
|
28
29
|
return
|
|
29
30
|
|
|
30
|
-
|
|
31
|
-
|
|
32
|
-
|
|
33
|
-
|
|
34
|
-
|
|
35
|
-
|
|
36
|
-
|
|
37
|
-
|
|
38
|
-
|
|
39
|
-
|
|
40
|
-
|
|
41
|
-
|
|
42
|
-
|
|
43
|
-
|
|
44
|
-
|
|
45
|
-
|
|
31
|
+
if len(self._buffer) > 0:
|
|
32
|
+
take = min(file_byte_count_remaining, len(self._buffer))
|
|
33
|
+
chunk = self._buffer[:take]
|
|
34
|
+
if chunk:
|
|
35
|
+
yield chunk
|
|
36
|
+
self._buffer = self._buffer[take:]
|
|
37
|
+
self._buffer_start += take
|
|
38
|
+
else:
|
|
39
|
+
start_of_fetch = self._buffer_start + len(self._buffer)
|
|
40
|
+
bytes_left_in_stream = self._files_data_end - start_of_fetch
|
|
41
|
+
if bytes_left_in_stream <= 0:
|
|
42
|
+
logger.error(
|
|
43
|
+
'StreamSeeker: no bytes left upstream (start_of_fetch=%d, files_data_end=%d)',
|
|
44
|
+
start_of_fetch,
|
|
45
|
+
self._files_data_end,
|
|
46
|
+
)
|
|
47
|
+
return
|
|
48
|
+
|
|
49
|
+
fetch_size = min(self._max_chunk_size, file_byte_count_remaining + read_ahead_bytes)
|
|
50
|
+
if fetch_size > bytes_left_in_stream:
|
|
51
|
+
logger.error(
|
|
52
|
+
'StreamSeeker: fetch_size (%d) > bytes_left_in_stream (%d); clamping',
|
|
53
|
+
fetch_size,
|
|
54
|
+
bytes_left_in_stream,
|
|
55
|
+
)
|
|
56
|
+
fetch_size = bytes_left_in_stream
|
|
57
|
+
|
|
58
|
+
fetched_data = self._upstream_buffer.get_data(start=start_of_fetch, length=fetch_size)
|
|
59
|
+
self._buffer.extend(fetched_data)
|
biolib/cli/__init__.py
CHANGED
|
@@ -5,7 +5,7 @@ import click
|
|
|
5
5
|
|
|
6
6
|
from biolib import utils
|
|
7
7
|
from biolib.biolib_logging import logger, logger_no_user_data
|
|
8
|
-
from biolib.cli import auth, data_record, download_container, init, lfs, push, run, runtime, sdk, start
|
|
8
|
+
from biolib.cli import auth, data_record, download_container, index, init, lfs, push, run, runtime, sdk, start
|
|
9
9
|
|
|
10
10
|
|
|
11
11
|
@click.version_option(version=utils.BIOLIB_PACKAGE_VERSION, prog_name='pybiolib')
|
|
@@ -31,6 +31,7 @@ cli.add_command(run.run)
|
|
|
31
31
|
cli.add_command(runtime.runtime)
|
|
32
32
|
cli.add_command(start.start)
|
|
33
33
|
cli.add_command(data_record.data_record)
|
|
34
|
+
cli.add_command(index.index)
|
|
34
35
|
cli.add_command(sdk.sdk)
|
|
35
36
|
|
|
36
37
|
# allow this script to be called without poetry in dev e.g. by an IDE debugger
|
biolib/cli/data_record.py
CHANGED
|
@@ -1,9 +1,11 @@
|
|
|
1
1
|
import json
|
|
2
2
|
import logging
|
|
3
3
|
import os
|
|
4
|
+
import sys
|
|
4
5
|
from typing import Dict, List
|
|
5
6
|
|
|
6
7
|
import click
|
|
8
|
+
import rich.progress
|
|
7
9
|
|
|
8
10
|
from biolib._data_record.data_record import DataRecord
|
|
9
11
|
from biolib.biolib_api_client import BiolibApiClient
|
|
@@ -80,3 +82,83 @@ def describe(uri: str, output_as_json: bool) -> None:
|
|
|
80
82
|
size_string = str(file_info['size_bytes'])
|
|
81
83
|
leading_space_string = ' ' * (10 - len(size_string))
|
|
82
84
|
print(f"{leading_space_string}{size_string} {file_info['path']}")
|
|
85
|
+
|
|
86
|
+
|
|
87
|
+
@data_record.command(help='Delete a Data Record')
|
|
88
|
+
@click.argument('uri', required=True)
|
|
89
|
+
def delete(uri: str) -> None:
|
|
90
|
+
record = DataRecord.get_by_uri(uri=uri)
|
|
91
|
+
|
|
92
|
+
print(f'You are about to delete the data record: {record.uri}')
|
|
93
|
+
print('This action cannot be undone.')
|
|
94
|
+
|
|
95
|
+
confirmation = input(f'To confirm deletion, please type the data record name "{record.name}": ')
|
|
96
|
+
if confirmation != record.name:
|
|
97
|
+
print('Data record name does not match. Deletion cancelled.')
|
|
98
|
+
return
|
|
99
|
+
|
|
100
|
+
record.delete()
|
|
101
|
+
print(f'Data record {record.uri} has been deleted.')
|
|
102
|
+
|
|
103
|
+
|
|
104
|
+
def _clone_data_record_with_progress(
|
|
105
|
+
source_record: DataRecord,
|
|
106
|
+
dest_record: DataRecord,
|
|
107
|
+
) -> None:
|
|
108
|
+
# pylint: disable=protected-access
|
|
109
|
+
total_size_in_bytes = source_record._get_zip_size_bytes()
|
|
110
|
+
# pylint: enable=protected-access
|
|
111
|
+
|
|
112
|
+
if total_size_in_bytes == 0:
|
|
113
|
+
logger.info('Source data record has no data to clone')
|
|
114
|
+
return
|
|
115
|
+
|
|
116
|
+
if sys.stdout.isatty():
|
|
117
|
+
with rich.progress.Progress(
|
|
118
|
+
rich.progress.TextColumn('[bold blue]{task.description}'),
|
|
119
|
+
rich.progress.BarColumn(),
|
|
120
|
+
rich.progress.TaskProgressColumn(),
|
|
121
|
+
rich.progress.TimeRemainingColumn(),
|
|
122
|
+
rich.progress.TransferSpeedColumn(),
|
|
123
|
+
) as progress:
|
|
124
|
+
task_id = progress.add_task('Cloning data record', total=total_size_in_bytes)
|
|
125
|
+
|
|
126
|
+
def on_progress(bytes_uploaded: int, _total_bytes: int) -> None:
|
|
127
|
+
progress.update(task_id, completed=bytes_uploaded)
|
|
128
|
+
|
|
129
|
+
DataRecord.clone(source=source_record, destination=dest_record, on_progress=on_progress)
|
|
130
|
+
else:
|
|
131
|
+
logger.info(f'Cloning ~{round(total_size_in_bytes / 10**6)}mb of data')
|
|
132
|
+
DataRecord.clone(source=source_record, destination=dest_record)
|
|
133
|
+
|
|
134
|
+
|
|
135
|
+
def _get_or_create_destination_record(destination_uri: str) -> Optional[DataRecord]:
|
|
136
|
+
try:
|
|
137
|
+
return DataRecord.get_by_uri(uri=destination_uri)
|
|
138
|
+
except Exception:
|
|
139
|
+
print(f'Destination data record "{destination_uri}" does not exist.')
|
|
140
|
+
confirmation = input('Would you like to create it? [y/N]: ')
|
|
141
|
+
if confirmation.lower() != 'y':
|
|
142
|
+
print('Clone cancelled.')
|
|
143
|
+
return None
|
|
144
|
+
|
|
145
|
+
return DataRecord.create(destination=destination_uri)
|
|
146
|
+
|
|
147
|
+
|
|
148
|
+
@data_record.command(help='Clone a Data Record to another location')
|
|
149
|
+
@click.argument('source_uri', required=True)
|
|
150
|
+
@click.argument('destination_uri', required=True)
|
|
151
|
+
def clone(source_uri: str, destination_uri: str) -> None:
|
|
152
|
+
BiolibApiClient.assert_is_signed_in(authenticated_action_description='clone a Data Record')
|
|
153
|
+
|
|
154
|
+
logger.info(f'Fetching source data record: {source_uri}')
|
|
155
|
+
source_record = DataRecord.get_by_uri(uri=source_uri)
|
|
156
|
+
|
|
157
|
+
logger.info(f'Checking destination data record: {destination_uri}')
|
|
158
|
+
dest_record = _get_or_create_destination_record(destination_uri)
|
|
159
|
+
if dest_record is None:
|
|
160
|
+
return
|
|
161
|
+
|
|
162
|
+
logger.info(f'Cloning from {source_record.uri} to {dest_record.uri}...')
|
|
163
|
+
_clone_data_record_with_progress(source_record=source_record, dest_record=dest_record)
|
|
164
|
+
logger.info('Clone completed successfully.')
|
biolib/cli/index.py
ADDED
|
@@ -0,0 +1,32 @@
|
|
|
1
|
+
import json
|
|
2
|
+
import logging
|
|
3
|
+
import sys
|
|
4
|
+
|
|
5
|
+
import click
|
|
6
|
+
|
|
7
|
+
from biolib._index.index import Index
|
|
8
|
+
from biolib.biolib_errors import BioLibError
|
|
9
|
+
from biolib.biolib_logging import logger, logger_no_user_data
|
|
10
|
+
|
|
11
|
+
|
|
12
|
+
@click.group(help='Manage Indexes')
|
|
13
|
+
def index() -> None:
|
|
14
|
+
logger.configure(default_log_level=logging.INFO)
|
|
15
|
+
logger_no_user_data.configure(default_log_level=logging.INFO)
|
|
16
|
+
|
|
17
|
+
|
|
18
|
+
@index.command(help='Create an Index')
|
|
19
|
+
@click.argument('uri', required=True)
|
|
20
|
+
@click.option('--config-path', required=True, type=click.Path(exists=True), help='Path to JSON config file')
|
|
21
|
+
def create(uri: str, config_path: str) -> None:
|
|
22
|
+
try:
|
|
23
|
+
Index.create_from_config_file(uri=uri, config_path=config_path)
|
|
24
|
+
except json.JSONDecodeError as error:
|
|
25
|
+
print(f'Error: Invalid JSON in config file: {error}', file=sys.stderr)
|
|
26
|
+
sys.exit(1)
|
|
27
|
+
except BioLibError as error:
|
|
28
|
+
print(f'Error creating index: {error.message}', file=sys.stderr)
|
|
29
|
+
sys.exit(1)
|
|
30
|
+
except Exception as error:
|
|
31
|
+
print(f'Error reading config file: {error}', file=sys.stderr)
|
|
32
|
+
sys.exit(1)
|
biolib/cli/init.py
CHANGED
|
@@ -10,18 +10,56 @@ from biolib import (
|
|
|
10
10
|
)
|
|
11
11
|
from biolib._internal.add_copilot_prompts import add_copilot_prompts
|
|
12
12
|
from biolib._internal.add_gui_files import add_gui_files
|
|
13
|
-
from biolib._internal.http_client import HttpError
|
|
13
|
+
from biolib._internal.http_client import HttpClient, HttpError
|
|
14
14
|
from biolib._internal.string_utils import normalize_for_docker_tag
|
|
15
15
|
from biolib._internal.templates import templates
|
|
16
16
|
from biolib.api import client as api_client
|
|
17
17
|
from biolib.biolib_api_client.api_client import BiolibApiClient
|
|
18
18
|
from biolib.biolib_api_client.biolib_app_api import BiolibAppApi
|
|
19
|
+
from biolib.biolib_logging import logger_no_user_data
|
|
20
|
+
from biolib.typing_utils import Optional
|
|
19
21
|
from biolib.user.sign_in import sign_in
|
|
20
22
|
from biolib.utils import BIOLIB_PACKAGE_VERSION
|
|
21
23
|
|
|
22
24
|
|
|
25
|
+
def _get_latest_pypi_version() -> Optional[str]:
|
|
26
|
+
try:
|
|
27
|
+
response = HttpClient.request(
|
|
28
|
+
url='https://pypi.org/pypi/pybiolib/json',
|
|
29
|
+
timeout_in_seconds=5,
|
|
30
|
+
retries=1,
|
|
31
|
+
)
|
|
32
|
+
data = response.json()
|
|
33
|
+
version = data.get('info', {}).get('version')
|
|
34
|
+
if isinstance(version, str):
|
|
35
|
+
return version
|
|
36
|
+
return None
|
|
37
|
+
except Exception as error:
|
|
38
|
+
logger_no_user_data.debug(f'Failed to fetch latest version from PyPI: {error}')
|
|
39
|
+
return None
|
|
40
|
+
|
|
41
|
+
|
|
42
|
+
def _is_current_version_outdated(current: str, latest: str) -> bool:
|
|
43
|
+
try:
|
|
44
|
+
current_parts = [int(x) for x in current.split('.')]
|
|
45
|
+
latest_parts = [int(x) for x in latest.split('.')]
|
|
46
|
+
return current_parts < latest_parts
|
|
47
|
+
except (ValueError, AttributeError):
|
|
48
|
+
return False
|
|
49
|
+
|
|
50
|
+
|
|
23
51
|
@click.command(help='Initialize a BioLib project', hidden=True)
|
|
24
52
|
def init() -> None:
|
|
53
|
+
latest_version = _get_latest_pypi_version()
|
|
54
|
+
if latest_version and _is_current_version_outdated(BIOLIB_PACKAGE_VERSION, latest_version):
|
|
55
|
+
print(f'A newer version of pybiolib is available: {latest_version} (current: {BIOLIB_PACKAGE_VERSION})')
|
|
56
|
+
print('To upgrade, run: pip install --upgrade pybiolib')
|
|
57
|
+
print()
|
|
58
|
+
continue_input = input('Do you want to continue with the current version? [y/N]: ')
|
|
59
|
+
if continue_input.lower() not in ['y', 'yes']:
|
|
60
|
+
print('Please upgrade pybiolib and run `biolib init` again.')
|
|
61
|
+
return
|
|
62
|
+
|
|
25
63
|
cwd = os.getcwd()
|
|
26
64
|
|
|
27
65
|
app_uri = input('What URI do you want to create the application under? (leave blank to skip): ')
|
biolib/cli/lfs.py
CHANGED
biolib/cli/run.py
CHANGED
|
@@ -13,19 +13,23 @@ from biolib.typing_utils import Optional, Tuple
|
|
|
13
13
|
help='Run an application on BioLib.',
|
|
14
14
|
)
|
|
15
15
|
@click.option('--experiment', type=str, required=False, help='Experiment name or URI to add the run to.')
|
|
16
|
-
@click.option('--local', is_flag=True, required=False,
|
|
16
|
+
@click.option('--local', is_flag=True, required=False, hidden=True)
|
|
17
17
|
@click.option('--non-blocking', is_flag=True, required=False, help='Run the application non blocking.')
|
|
18
18
|
@click.argument('uri', required=True)
|
|
19
19
|
@click.argument('args', nargs=-1, type=click.UNPROCESSED)
|
|
20
20
|
def run(experiment: Optional[str], local: bool, non_blocking: bool, uri: str, args: Tuple[str]) -> None:
|
|
21
|
+
if local:
|
|
22
|
+
print('Error: Running applications locally with --local is no longer supported.', file=sys.stderr)
|
|
23
|
+
sys.exit(1)
|
|
24
|
+
|
|
21
25
|
if experiment:
|
|
22
26
|
with Experiment(uri=experiment):
|
|
23
|
-
_run(
|
|
27
|
+
_run(non_blocking=non_blocking, uri=uri, args=args)
|
|
24
28
|
else:
|
|
25
|
-
_run(
|
|
29
|
+
_run(non_blocking=non_blocking, uri=uri, args=args)
|
|
26
30
|
|
|
27
31
|
|
|
28
|
-
def _run(
|
|
32
|
+
def _run(non_blocking: bool, uri: str, args: Tuple[str]) -> None:
|
|
29
33
|
try:
|
|
30
34
|
app = BioLibApp(uri=uri)
|
|
31
35
|
except biolib_errors.BioLibError as error:
|
|
@@ -43,7 +47,6 @@ def _run(local: bool, non_blocking: bool, uri: str, args: Tuple[str]) -> None:
|
|
|
43
47
|
args=list(args),
|
|
44
48
|
stdin=_get_stdin(),
|
|
45
49
|
files=None,
|
|
46
|
-
machine=('local' if local else ''),
|
|
47
50
|
blocking=blocking,
|
|
48
51
|
)
|
|
49
52
|
|
biolib/cli/start.py
CHANGED
|
@@ -13,7 +13,10 @@ from biolib.typing_utils import Optional
|
|
|
13
13
|
@click.option('--port', default=5000, type=click.IntRange(1, 65_535), required=False)
|
|
14
14
|
@click.option('--tls-certificate', type=click.Path(exists=True), required=False, hidden=True)
|
|
15
15
|
@click.option('--tls-key', type=click.Path(exists=True), required=False, hidden=True)
|
|
16
|
-
|
|
16
|
+
@click.option('--initialize-network-buffer', is_flag=True, help='Initialize the remote host network buffer and exit')
|
|
17
|
+
def start(
|
|
18
|
+
host: str, port: int, tls_certificate: Optional[str], tls_key: Optional[str], initialize_network_buffer: bool
|
|
19
|
+
) -> None:
|
|
17
20
|
logger.configure(default_log_level=logging.INFO)
|
|
18
21
|
logger_no_user_data.configure(default_log_level=logging.INFO)
|
|
19
22
|
if platform.system() == 'Windows':
|
|
@@ -22,6 +25,16 @@ def start(host: str, port: int, tls_certificate: Optional[str], tls_key: Optiona
|
|
|
22
25
|
if tls_certificate and not tls_key or tls_key and not tls_certificate:
|
|
23
26
|
raise Exception('Options --tls-certificate and --tls-key must be specified together')
|
|
24
27
|
|
|
28
|
+
if initialize_network_buffer:
|
|
29
|
+
from biolib.compute_node.job_worker.network_buffer import ( # pylint: disable=import-outside-toplevel
|
|
30
|
+
NetworkBuffer,
|
|
31
|
+
)
|
|
32
|
+
|
|
33
|
+
network_buffer = NetworkBuffer.get_instance()
|
|
34
|
+
created = network_buffer.fill_buffer()
|
|
35
|
+
logger_no_user_data.info(f'Initialized network buffer (created {created} networks)')
|
|
36
|
+
return
|
|
37
|
+
|
|
25
38
|
try:
|
|
26
39
|
from biolib.compute_node.webserver import webserver # pylint: disable=import-outside-toplevel
|
|
27
40
|
|
|
@@ -11,10 +11,10 @@ import zipfile
|
|
|
11
11
|
from copy import copy
|
|
12
12
|
from datetime import datetime
|
|
13
13
|
|
|
14
|
-
import docker
|
|
15
|
-
import docker.types
|
|
16
|
-
from docker.errors import APIError, ImageNotFound
|
|
17
|
-
from docker.models.containers import Container
|
|
14
|
+
import docker
|
|
15
|
+
import docker.types
|
|
16
|
+
from docker.errors import APIError, ImageNotFound
|
|
17
|
+
from docker.models.containers import Container
|
|
18
18
|
|
|
19
19
|
from biolib import utils
|
|
20
20
|
from biolib._internal.runtime import RuntimeJobDataDict
|
|
@@ -40,7 +40,7 @@ class DockerExecutor:
|
|
|
40
40
|
self._options: LocalExecutorOptions = options
|
|
41
41
|
self._is_cleaning_up = False
|
|
42
42
|
|
|
43
|
-
self._absolute_image_uri = f
|
|
43
|
+
self._absolute_image_uri = f'{utils.BIOLIB_SITE_HOSTNAME}/{self._options["module"]["image_uri"]}'
|
|
44
44
|
self._send_system_exception = options['send_system_exception']
|
|
45
45
|
self._send_stdout_and_stderr = options['send_stdout_and_stderr']
|
|
46
46
|
self._random_docker_id = compute_node_utils.random_string(15)
|
|
@@ -308,7 +308,8 @@ class DockerExecutor:
|
|
|
308
308
|
job_uuid = self._options['job']['public_id']
|
|
309
309
|
logger_no_user_data.debug(f'Job "{job_uuid}" initializing Docker container...')
|
|
310
310
|
module = self._options['module']
|
|
311
|
-
logger.debug(f
|
|
311
|
+
logger.debug(f'Initializing docker container with command: {module["command"]}')
|
|
312
|
+
docker_client = BiolibDockerClient.get_docker_client()
|
|
312
313
|
|
|
313
314
|
docker_volume_mounts = [lfs.docker_mount for lfs in self._options['large_file_systems'].values()]
|
|
314
315
|
|
|
@@ -318,10 +319,12 @@ class DockerExecutor:
|
|
|
318
319
|
biolib_system_secret = RuntimeJobDataDict(
|
|
319
320
|
version='1.0.0',
|
|
320
321
|
job_requested_machine=self._options['job']['requested_machine'],
|
|
322
|
+
job_requested_machine_spot=self._options['job'].get('requested_machine_spot', False),
|
|
321
323
|
job_uuid=self._options['job']['public_id'],
|
|
322
324
|
job_auth_token=self._options['job']['auth_token'],
|
|
323
325
|
app_uri=self._options['job']['app_uri'],
|
|
324
326
|
is_environment_biolib_cloud=bool(utils.IS_RUNNING_IN_CLOUD),
|
|
327
|
+
job_reserved_machines=self._options['job']['reserved_machines'],
|
|
325
328
|
)
|
|
326
329
|
docker_volume_mounts.append(
|
|
327
330
|
self._create_secrets_mount(
|
|
@@ -365,9 +368,11 @@ class DockerExecutor:
|
|
|
365
368
|
)
|
|
366
369
|
|
|
367
370
|
logger_no_user_data.debug(f'Job "{job_uuid}" initializing Docker container. Getting IPs for proxies...')
|
|
371
|
+
|
|
372
|
+
networks_to_connect = []
|
|
368
373
|
for proxy in self._options['remote_host_proxies']:
|
|
369
|
-
proxy_ip = proxy.get_ip_address_on_network(internal_network)
|
|
370
374
|
if proxy.is_app_caller_proxy:
|
|
375
|
+
proxy_ip = proxy.get_ip_address_on_network(internal_network)
|
|
371
376
|
logger_no_user_data.debug('Found app caller proxy, setting both base URLs in compute container')
|
|
372
377
|
environment_vars.update(
|
|
373
378
|
{
|
|
@@ -381,7 +386,11 @@ class DockerExecutor:
|
|
|
381
386
|
}
|
|
382
387
|
)
|
|
383
388
|
else:
|
|
384
|
-
extra_hosts
|
|
389
|
+
extra_hosts.update(proxy.get_hostname_to_ip_mapping())
|
|
390
|
+
|
|
391
|
+
for network in proxy.get_remote_host_networks():
|
|
392
|
+
if network != internal_network:
|
|
393
|
+
networks_to_connect.append(network)
|
|
385
394
|
|
|
386
395
|
logger_no_user_data.debug(f'Job "{job_uuid}" initializing Docker container. Constructing container args...')
|
|
387
396
|
create_container_args = {
|
|
@@ -391,6 +400,9 @@ class DockerExecutor:
|
|
|
391
400
|
'mounts': docker_volume_mounts,
|
|
392
401
|
'network': internal_network.name,
|
|
393
402
|
'working_dir': module['working_directory'],
|
|
403
|
+
'networking_config': {
|
|
404
|
+
internal_network.name: docker_client.api.create_endpoint_config(aliases=['main'])
|
|
405
|
+
},
|
|
394
406
|
}
|
|
395
407
|
|
|
396
408
|
if self._options['job'].get('arguments_override_command'):
|
|
@@ -429,9 +441,19 @@ class DockerExecutor:
|
|
|
429
441
|
if docker_runtime is not None:
|
|
430
442
|
create_container_args['runtime'] = docker_runtime
|
|
431
443
|
|
|
432
|
-
docker_client = BiolibDockerClient.get_docker_client()
|
|
433
444
|
logger_no_user_data.debug(f'Job "{job_uuid}" initializing Docker container. Creating container...')
|
|
434
445
|
self._docker_container = docker_client.containers.create(**create_container_args)
|
|
446
|
+
|
|
447
|
+
if networks_to_connect:
|
|
448
|
+
network_connection_start = time.time()
|
|
449
|
+
for network in networks_to_connect:
|
|
450
|
+
network.connect(self._docker_container.id)
|
|
451
|
+
logger_no_user_data.debug(f'Connected app container to network {network.name}')
|
|
452
|
+
network_connection_time = time.time() - network_connection_start
|
|
453
|
+
logger_no_user_data.debug(
|
|
454
|
+
f'Connected app container to {len(networks_to_connect)} networks in {network_connection_time:.2f}s'
|
|
455
|
+
)
|
|
456
|
+
|
|
435
457
|
logger_no_user_data.debug(f'Job "{job_uuid}" finished initializing Docker container.')
|
|
436
458
|
except Exception as exception:
|
|
437
459
|
raise ComputeProcessException(
|
|
@@ -1,11 +1,11 @@
|
|
|
1
|
-
from docker.models.networks import Network
|
|
1
|
+
from docker.models.networks import Network
|
|
2
2
|
|
|
3
|
+
from biolib.biolib_api_client.app_types import Module
|
|
4
|
+
from biolib.biolib_api_client.job_types import CloudJob, CreatedJobDict
|
|
3
5
|
from biolib.compute_node.job_worker.large_file_system import LargeFileSystem
|
|
4
|
-
from biolib.compute_node.webserver.webserver_types import ComputeNodeInfo
|
|
5
|
-
from biolib.typing_utils import TypedDict, Callable, Optional, List, Dict
|
|
6
6
|
from biolib.compute_node.remote_host_proxy import RemoteHostProxy
|
|
7
|
-
from biolib.
|
|
8
|
-
from biolib.
|
|
7
|
+
from biolib.compute_node.webserver.webserver_types import ComputeNodeInfo
|
|
8
|
+
from biolib.typing_utils import Callable, Dict, List, Optional, TypedDict
|
|
9
9
|
|
|
10
10
|
|
|
11
11
|
class StatusUpdate(TypedDict):
|
|
@@ -43,6 +43,7 @@ class LocalExecutorOptions(TypedDict):
|
|
|
43
43
|
send_system_exception: SendSystemExceptionType
|
|
44
44
|
send_stdout_and_stderr: SendStdoutAndStderrType
|
|
45
45
|
|
|
46
|
+
|
|
46
47
|
class MetadataToSaveOutput(TypedDict):
|
|
47
48
|
arguments: List[str]
|
|
48
49
|
startup_error_string: Optional[str]
|