rclone-api 1.4.15__tar.gz → 1.4.18__tar.gz
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- {rclone_api-1.4.15 → rclone_api-1.4.18}/PKG-INFO +1 -1
- {rclone_api-1.4.15 → rclone_api-1.4.18}/pyproject.toml +1 -1
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/cmd/copy_large_s3_finish.py +68 -13
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/process.py +65 -40
- rclone_api-1.4.18/src/rclone_api/s3/merge_state.py +115 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/s3/multipart/finished_piece.py +15 -1
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/s3/s3_multipart_uploader_by_copy.py +139 -81
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/util.py +84 -23
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api.egg-info/PKG-INFO +1 -1
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api.egg-info/SOURCES.txt +1 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/.aiderignore +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/.github/workflows/lint.yml +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/.github/workflows/push_macos.yml +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/.github/workflows/push_ubuntu.yml +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/.github/workflows/push_win.yml +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/.gitignore +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/.pylintrc +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/.vscode/launch.json +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/.vscode/settings.json +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/.vscode/tasks.json +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/LICENSE +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/MANIFEST.in +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/README.md +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/clean +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/install +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/lint +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/requirements.testing.txt +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/setup.cfg +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/setup.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/__init__.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/assets/example.txt +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/cli.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/cmd/analyze.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/cmd/copy_large_s3.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/cmd/list_files.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/cmd/save_to_db.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/completed_process.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/config.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/convert.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/db/__init__.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/db/db.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/db/models.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/deprecated.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/detail/copy_file_parts.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/detail/walk.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/diff.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/dir.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/dir_listing.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/exec.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/experimental/flags.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/experimental/flags_base.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/file.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/file_item.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/file_part.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/file_stream.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/filelist.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/group_files.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/http_server.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/log.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/mount.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/rclone_impl.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/remote.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/rpath.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/s3/api.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/s3/basic_ops.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/s3/chunk_task.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/s3/create.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/s3/multipart/file_info.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/s3/multipart/upload_info.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/s3/multipart/upload_state.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/s3/types.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/s3/upload_file_multipart.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/scan_missing_folders.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api/types.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api.egg-info/dependency_links.txt +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api.egg-info/entry_points.txt +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api.egg-info/requires.txt +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/src/rclone_api.egg-info/top_level.txt +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/test +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/tests/archive/test_paramiko.py.disabled +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/tests/test_cmd_list_files.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/tests/test_copy.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/tests/test_copy_bytes.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/tests/test_copy_file_resumable_s3.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/tests/test_copy_files.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/tests/test_db.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/tests/test_diff.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/tests/test_file_item.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/tests/test_group_files.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/tests/test_is_synced.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/tests/test_ls.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/tests/test_ls_stream_files.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/tests/test_mount.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/tests/test_mount_s3.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/tests/test_obscure.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/tests/test_rclone_config.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/tests/test_read_write_text.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/tests/test_remote_control.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/tests/test_remotes.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/tests/test_s3.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/tests/test_scan_missing_folders.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/tests/test_serve_http.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/tests/test_size_files.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/tests/test_size_suffix.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/tests/test_walk.py +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/tox.ini +0 -0
- {rclone_api-1.4.15 → rclone_api-1.4.18}/upload_package.sh +0 -0
@@ -6,12 +6,13 @@ from pathlib import Path
|
|
6
6
|
from rclone_api import Rclone
|
7
7
|
from rclone_api.detail.copy_file_parts import InfoJson
|
8
8
|
from rclone_api.s3.s3_multipart_uploader_by_copy import (
|
9
|
-
|
9
|
+
Part,
|
10
|
+
S3MultiPartUploader,
|
10
11
|
)
|
11
|
-
from rclone_api.types import SizeSuffix
|
12
12
|
|
13
13
|
_TIMEOUT_READ = 900
|
14
14
|
_TIMEOUT_CONNECTION = 900
|
15
|
+
_MAX_WORKERS = 10
|
15
16
|
|
16
17
|
|
17
18
|
@dataclass
|
@@ -53,7 +54,55 @@ def _parse_args() -> Args:
|
|
53
54
|
return out
|
54
55
|
|
55
56
|
|
56
|
-
def
|
57
|
+
# def finish_multipart_upload_from_keys(
|
58
|
+
# s3_client: BaseClient,
|
59
|
+
# source_bucket: str,
|
60
|
+
# parts: list[Part],
|
61
|
+
# destination_bucket: str,
|
62
|
+
# destination_key: str,
|
63
|
+
# chunk_size: int, # 5MB default
|
64
|
+
# max_workers: int = 100,
|
65
|
+
# retries: int = 3,
|
66
|
+
# ) -> str | Exception:
|
67
|
+
# """
|
68
|
+
# Finish a multipart upload by copying parts from existing S3 objects.
|
69
|
+
|
70
|
+
# Args:
|
71
|
+
# s3_client: Boto3 S3 client
|
72
|
+
# source_bucket: Source bucket name
|
73
|
+
# source_keys: List of source object keys to copy from
|
74
|
+
# destination_bucket: Destination bucket name
|
75
|
+
# destination_key: Destination object key
|
76
|
+
# chunk_size: Size of each part in bytes
|
77
|
+
# retries: Number of retry attempts
|
78
|
+
# byte_ranges: Optional list of byte ranges corresponding to source_keys
|
79
|
+
|
80
|
+
# Returns:
|
81
|
+
# The URL of the completed object
|
82
|
+
# """
|
83
|
+
|
84
|
+
# # Create upload info
|
85
|
+
# info = begin_upload(
|
86
|
+
# s3_client=s3_client,
|
87
|
+
# parts=parts,
|
88
|
+
# destination_bucket=destination_bucket,
|
89
|
+
# destination_key=destination_key,
|
90
|
+
# chunk_size=chunk_size,
|
91
|
+
# retries=retries,
|
92
|
+
# )
|
93
|
+
|
94
|
+
# out = do_body_work(
|
95
|
+
# info=info,
|
96
|
+
# source_bucket=source_bucket,
|
97
|
+
# parts=parts,
|
98
|
+
# max_workers=max_workers,
|
99
|
+
# retries=retries,
|
100
|
+
# )
|
101
|
+
|
102
|
+
# return out
|
103
|
+
|
104
|
+
|
105
|
+
def do_finish_part(rclone: Rclone, info: InfoJson, dst: str) -> Exception | None:
|
57
106
|
from rclone_api.s3.create import (
|
58
107
|
BaseClient,
|
59
108
|
S3Config,
|
@@ -71,6 +120,7 @@ def do_finish_part(rclone: Rclone, info: InfoJson, dst: str) -> None:
|
|
71
120
|
s3_client: BaseClient = create_s3_client(s3_creds=s3_creds, s3_config=s3_config)
|
72
121
|
s3_bucket = s3_creds.bucket_name
|
73
122
|
is_done = info.fetch_is_done()
|
123
|
+
size = info.size
|
74
124
|
assert is_done, f"Upload is not done: {info}"
|
75
125
|
|
76
126
|
parts_dir = info.parts_dir
|
@@ -86,11 +136,9 @@ def do_finish_part(rclone: Rclone, info: InfoJson, dst: str) -> None:
|
|
86
136
|
|
87
137
|
first_part: int | None = info.first_part
|
88
138
|
last_part: int | None = info.last_part
|
89
|
-
size: SizeSuffix | None = info.size
|
90
139
|
|
91
140
|
assert first_part is not None
|
92
141
|
assert last_part is not None
|
93
|
-
assert size is not None
|
94
142
|
|
95
143
|
def _to_s3_key(name: str | None) -> str:
|
96
144
|
if name:
|
@@ -99,11 +147,12 @@ def do_finish_part(rclone: Rclone, info: InfoJson, dst: str) -> None:
|
|
99
147
|
out = f"{parts_path}"
|
100
148
|
return out
|
101
149
|
|
102
|
-
parts: list[
|
150
|
+
parts: list[Part] = []
|
103
151
|
part_num = 1
|
104
152
|
for part_key in source_keys:
|
105
153
|
s3_key = _to_s3_key(name=part_key)
|
106
|
-
|
154
|
+
part = Part(part_number=part_num, s3_key=s3_key)
|
155
|
+
parts.append(part)
|
107
156
|
part_num += 1
|
108
157
|
|
109
158
|
chunksize = info.chunksize
|
@@ -114,30 +163,36 @@ def do_finish_part(rclone: Rclone, info: InfoJson, dst: str) -> None:
|
|
114
163
|
# dst_key =
|
115
164
|
dst_key = f"{dst_dir}/{dst_name}"
|
116
165
|
|
117
|
-
|
166
|
+
uploader: S3MultiPartUploader = S3MultiPartUploader(
|
118
167
|
s3_client=s3_client,
|
119
|
-
|
168
|
+
verbose=True,
|
169
|
+
)
|
170
|
+
|
171
|
+
from rclone_api.s3.s3_multipart_uploader_by_copy import MultipartUploadInfo
|
172
|
+
|
173
|
+
state: MultipartUploadInfo = uploader.begin_new_upload(
|
120
174
|
parts=parts,
|
121
175
|
destination_bucket=s3_creds.bucket_name,
|
122
176
|
destination_key=dst_key,
|
123
177
|
chunk_size=chunksize.as_int(),
|
124
|
-
final_size=size.as_int(),
|
125
|
-
max_workers=10,
|
126
178
|
)
|
127
179
|
|
180
|
+
uploader.start_upload(info=state, parts=parts, max_workers=_MAX_WORKERS)
|
181
|
+
|
128
182
|
# now check if the dst now exists, if so, delete the parts folder.
|
129
183
|
# if rclone.exists(dst):
|
130
184
|
# rclone.purge(parts_dir)
|
131
185
|
|
132
186
|
if not rclone.exists(dst):
|
133
|
-
|
187
|
+
return FileNotFoundError(f"Destination file not found: {dst}")
|
134
188
|
|
135
189
|
write_size = rclone.size_file(dst)
|
136
190
|
if write_size != size:
|
137
|
-
|
191
|
+
return ValueError(f"Size mismatch: {write_size} != {size}")
|
138
192
|
|
139
193
|
print(f"Upload complete: {dst}")
|
140
194
|
rclone.purge(parts_dir)
|
195
|
+
return None
|
141
196
|
|
142
197
|
|
143
198
|
def main() -> int:
|
@@ -1,12 +1,13 @@
|
|
1
1
|
import atexit
|
2
2
|
import subprocess
|
3
3
|
import threading
|
4
|
-
import time
|
5
4
|
import weakref
|
6
5
|
from dataclasses import dataclass
|
7
6
|
from pathlib import Path
|
8
7
|
from typing import Any
|
9
8
|
|
9
|
+
import psutil
|
10
|
+
|
10
11
|
from rclone_api.config import Config
|
11
12
|
from rclone_api.util import clear_temp_config_file, get_verbose, make_temp_config_file
|
12
13
|
|
@@ -24,20 +25,25 @@ class ProcessArgs:
|
|
24
25
|
|
25
26
|
class Process:
|
26
27
|
def __init__(self, args: ProcessArgs) -> None:
|
27
|
-
assert
|
28
|
+
assert (
|
29
|
+
args.rclone_exe.exists()
|
30
|
+
), f"rclone executable not found: {args.rclone_exe}"
|
28
31
|
self.args = args
|
29
32
|
self.log = args.log
|
30
33
|
self.tempfile: Path | None = None
|
34
|
+
|
31
35
|
verbose = get_verbose(args.verbose)
|
36
|
+
# Create a temporary config file if needed.
|
32
37
|
if isinstance(args.rclone_conf, Config):
|
33
|
-
self.
|
34
|
-
self.
|
35
|
-
rclone_conf = self.
|
38
|
+
self.tempfile = make_temp_config_file()
|
39
|
+
self.tempfile.write_text(args.rclone_conf.text, encoding="utf-8")
|
40
|
+
rclone_conf = self.tempfile
|
36
41
|
else:
|
37
42
|
rclone_conf = args.rclone_conf
|
38
43
|
|
39
|
-
assert rclone_conf.exists()
|
44
|
+
assert rclone_conf.exists(), f"rclone config not found: {rclone_conf}"
|
40
45
|
|
46
|
+
# Build the command.
|
41
47
|
self.cmd = (
|
42
48
|
[str(args.rclone_exe.resolve())]
|
43
49
|
+ ["--config", str(rclone_conf.resolve())]
|
@@ -49,16 +55,14 @@ class Process:
|
|
49
55
|
if verbose:
|
50
56
|
cmd_str = subprocess.list2cmdline(self.cmd)
|
51
57
|
print(f"Running: {cmd_str}")
|
52
|
-
kwargs: dict = {}
|
53
|
-
kwargs["shell"] = False
|
58
|
+
kwargs: dict = {"shell": False}
|
54
59
|
if args.capture_stdout:
|
55
60
|
kwargs["stdout"] = subprocess.PIPE
|
56
61
|
kwargs["stderr"] = subprocess.STDOUT
|
57
62
|
|
58
63
|
self.process = subprocess.Popen(self.cmd, **kwargs) # type: ignore
|
59
64
|
|
60
|
-
# Register an atexit callback using a weak reference to avoid
|
61
|
-
# keeping the Process instance alive solely due to the callback.
|
65
|
+
# Register an atexit callback using a weak reference to avoid keeping the Process instance alive.
|
62
66
|
self_ref = weakref.ref(self)
|
63
67
|
|
64
68
|
def exit_cleanup():
|
@@ -77,39 +81,60 @@ class Process:
|
|
77
81
|
self.cleanup()
|
78
82
|
|
79
83
|
def cleanup(self) -> None:
|
80
|
-
|
84
|
+
if self.tempfile:
|
85
|
+
clear_temp_config_file(self.tempfile)
|
81
86
|
|
82
|
-
def
|
87
|
+
def _kill_process_tree(self) -> None:
|
83
88
|
"""
|
84
|
-
|
85
|
-
If the process does not exit within a short timeout, it is aggressively killed.
|
89
|
+
Use psutil to recursively terminate the main process and all its child processes.
|
86
90
|
"""
|
87
|
-
|
88
|
-
|
89
|
-
|
91
|
+
try:
|
92
|
+
parent = psutil.Process(self.process.pid)
|
93
|
+
except psutil.NoSuchProcess:
|
94
|
+
return
|
95
|
+
|
96
|
+
# Terminate child processes.
|
97
|
+
children = parent.children(recursive=True)
|
98
|
+
if children:
|
99
|
+
print(f"Terminating {len(children)} child processes...")
|
100
|
+
for child in children:
|
90
101
|
try:
|
91
|
-
|
92
|
-
self.process.terminate()
|
102
|
+
child.terminate()
|
93
103
|
except Exception as e:
|
94
|
-
print(f"Error
|
95
|
-
|
96
|
-
|
97
|
-
|
98
|
-
|
99
|
-
time.sleep(0.1)
|
100
|
-
# If still running, kill aggressively.
|
101
|
-
if self.process.poll() is None:
|
104
|
+
print(f"Error terminating child process {child.pid}: {e}")
|
105
|
+
psutil.wait_procs(children, timeout=2)
|
106
|
+
# Kill any that remain.
|
107
|
+
for child in children:
|
108
|
+
if child.is_running():
|
102
109
|
try:
|
103
|
-
|
110
|
+
child.kill()
|
104
111
|
except Exception as e:
|
105
|
-
print(f"Error
|
106
|
-
|
112
|
+
print(f"Error killing child process {child.pid}: {e}")
|
113
|
+
|
114
|
+
# Terminate the parent process.
|
115
|
+
if parent.is_running():
|
116
|
+
try:
|
117
|
+
parent.terminate()
|
118
|
+
except Exception as e:
|
119
|
+
print(f"Error terminating process {parent.pid}: {e}")
|
120
|
+
try:
|
121
|
+
parent.wait(timeout=3)
|
122
|
+
except psutil.TimeoutExpired:
|
107
123
|
try:
|
108
|
-
|
109
|
-
except Exception:
|
110
|
-
|
124
|
+
parent.kill()
|
125
|
+
except Exception as e:
|
126
|
+
print(f"Error killing process {parent.pid}: {e}")
|
127
|
+
|
128
|
+
def _atexit_terminate(self) -> None:
|
129
|
+
"""
|
130
|
+
This method is registered via atexit and uses psutil to clean up the process tree.
|
131
|
+
It runs in a daemon thread so that termination happens without blocking interpreter shutdown.
|
132
|
+
"""
|
133
|
+
if self.process.poll() is None: # Process is still running.
|
134
|
+
|
135
|
+
def terminate_sequence():
|
136
|
+
self._kill_process_tree()
|
111
137
|
|
112
|
-
# Run the termination sequence in a separate daemon thread.
|
113
138
|
t = threading.Thread(target=terminate_sequence, daemon=True)
|
114
139
|
t.start()
|
115
140
|
t.join(timeout=3)
|
@@ -122,12 +147,12 @@ class Process:
|
|
122
147
|
self.cleanup()
|
123
148
|
|
124
149
|
def kill(self) -> None:
|
125
|
-
|
126
|
-
|
150
|
+
"""Forcefully kill the process tree."""
|
151
|
+
self._kill_process_tree()
|
127
152
|
|
128
153
|
def terminate(self) -> None:
|
129
|
-
|
130
|
-
|
154
|
+
"""Gracefully terminate the process tree."""
|
155
|
+
self._kill_process_tree()
|
131
156
|
|
132
157
|
@property
|
133
158
|
def returncode(self) -> int | None:
|
@@ -147,8 +172,8 @@ class Process:
|
|
147
172
|
def wait(self) -> int:
|
148
173
|
return self.process.wait()
|
149
174
|
|
150
|
-
def send_signal(self,
|
151
|
-
|
175
|
+
def send_signal(self, sig: int) -> None:
|
176
|
+
self.process.send_signal(sig)
|
152
177
|
|
153
178
|
def __str__(self) -> str:
|
154
179
|
state = ""
|
@@ -0,0 +1,115 @@
|
|
1
|
+
"""
|
2
|
+
https://boto3.amazonaws.com/v1/documentation/api/latest/reference/services/s3/client/upload_part_copy.html
|
3
|
+
* client.upload_part_copy
|
4
|
+
|
5
|
+
This module provides functionality for S3 multipart uploads, including copying parts
|
6
|
+
from existing S3 objects using upload_part_copy.
|
7
|
+
"""
|
8
|
+
|
9
|
+
import json
|
10
|
+
from dataclasses import dataclass
|
11
|
+
from typing import Any, Callable
|
12
|
+
|
13
|
+
from rclone_api.s3.multipart.finished_piece import FinishedPiece
|
14
|
+
|
15
|
+
|
16
|
+
@dataclass
|
17
|
+
class Part:
|
18
|
+
part_number: int
|
19
|
+
s3_key: str
|
20
|
+
|
21
|
+
def to_json(self) -> dict:
|
22
|
+
return {"part_number": self.part_number, "s3_key": self.s3_key}
|
23
|
+
|
24
|
+
@staticmethod
|
25
|
+
def from_json(json_dict: dict) -> "Part | Exception":
|
26
|
+
part_number = json_dict.get("part_number")
|
27
|
+
s3_key = json_dict.get("s3_key")
|
28
|
+
if part_number is None or s3_key is None:
|
29
|
+
return Exception(f"Invalid JSON: {json_dict}")
|
30
|
+
return Part(part_number=part_number, s3_key=s3_key)
|
31
|
+
|
32
|
+
@staticmethod
|
33
|
+
def from_json_array(json_array: list[dict]) -> list["Part"] | Exception:
|
34
|
+
try:
|
35
|
+
out: list[Part] = []
|
36
|
+
for j in json_array:
|
37
|
+
ok_or_err = Part.from_json(j)
|
38
|
+
if isinstance(ok_or_err, Exception):
|
39
|
+
return ok_or_err
|
40
|
+
else:
|
41
|
+
out.append(ok_or_err)
|
42
|
+
return out
|
43
|
+
except Exception as e:
|
44
|
+
return e
|
45
|
+
|
46
|
+
|
47
|
+
class MergeState:
|
48
|
+
|
49
|
+
def __init__(self, finished: list[FinishedPiece], all_parts: list[Part]) -> None:
|
50
|
+
self.finished: list[FinishedPiece] = list(finished)
|
51
|
+
self.all_parts: list[Part] = list(all_parts)
|
52
|
+
self.callbacks: list[Callable[[FinishedPiece], None]] = []
|
53
|
+
|
54
|
+
def add_callback(self, callback: Callable[[FinishedPiece], None]) -> None:
|
55
|
+
self.callbacks.append(callback)
|
56
|
+
|
57
|
+
def on_finished(self, finished_piece: FinishedPiece) -> None:
|
58
|
+
for callback in list(self.callbacks):
|
59
|
+
callback(finished_piece)
|
60
|
+
|
61
|
+
@staticmethod
|
62
|
+
def from_json_array(json_array: dict) -> "MergeState | Exception":
|
63
|
+
try:
|
64
|
+
finished: list[FinishedPiece] = FinishedPiece.from_json_array(
|
65
|
+
json_array["finished"]
|
66
|
+
)
|
67
|
+
all_parts: list[Part | Exception] = [
|
68
|
+
Part.from_json(j) for j in json_array["all"]
|
69
|
+
]
|
70
|
+
all_parts_no_err: list[Part] = [
|
71
|
+
p for p in all_parts if not isinstance(p, Exception)
|
72
|
+
]
|
73
|
+
errs: list[Exception] = [p for p in all_parts if isinstance(p, Exception)]
|
74
|
+
if len(errs):
|
75
|
+
return Exception(f"Errors in parts: {errs}")
|
76
|
+
return MergeState(finished=finished, all_parts=all_parts_no_err)
|
77
|
+
except Exception as e:
|
78
|
+
return e
|
79
|
+
|
80
|
+
def to_json(self) -> dict:
|
81
|
+
finished = self.finished.copy()
|
82
|
+
all_parts = self.all_parts.copy()
|
83
|
+
return {
|
84
|
+
"finished": FinishedPiece.to_json_array(finished),
|
85
|
+
"all": [part.to_json() for part in all_parts],
|
86
|
+
}
|
87
|
+
|
88
|
+
def to_json_str(self) -> str:
|
89
|
+
return json.dumps(self.to_json(), indent=1)
|
90
|
+
|
91
|
+
def __str__(self):
|
92
|
+
return self.to_json_str()
|
93
|
+
|
94
|
+
def __repr__(self):
|
95
|
+
return self.to_json_str()
|
96
|
+
|
97
|
+
def write(self, rclone_impl: Any, dst: str) -> None:
|
98
|
+
from rclone_api.rclone_impl import RcloneImpl
|
99
|
+
|
100
|
+
assert isinstance(rclone_impl, RcloneImpl)
|
101
|
+
json_str = self.to_json_str()
|
102
|
+
rclone_impl.write_text(dst, json_str)
|
103
|
+
|
104
|
+
def read(self, rclone_impl: Any, src: str) -> None:
|
105
|
+
from rclone_api.rclone_impl import RcloneImpl
|
106
|
+
|
107
|
+
assert isinstance(rclone_impl, RcloneImpl)
|
108
|
+
json_str = rclone_impl.read_text(src)
|
109
|
+
if isinstance(json_str, Exception):
|
110
|
+
raise json_str
|
111
|
+
json_dict = json.loads(json_str)
|
112
|
+
ok_or_err = FinishedPiece.from_json_array(json_dict["finished"])
|
113
|
+
if isinstance(ok_or_err, Exception):
|
114
|
+
raise ok_or_err
|
115
|
+
self.finished = ok_or_err
|
@@ -17,7 +17,9 @@ class FinishedPiece:
|
|
17
17
|
return json.dumps(self.to_json(), indent=0)
|
18
18
|
|
19
19
|
@staticmethod
|
20
|
-
def to_json_array(
|
20
|
+
def to_json_array(
|
21
|
+
parts: list["FinishedPiece | EndOfStream"] | list["FinishedPiece"],
|
22
|
+
) -> list[dict]:
|
21
23
|
non_none: list[FinishedPiece] = []
|
22
24
|
for p in parts:
|
23
25
|
if not isinstance(p, EndOfStream):
|
@@ -39,3 +41,15 @@ class FinishedPiece:
|
|
39
41
|
if json is None:
|
40
42
|
return EndOfStream()
|
41
43
|
return FinishedPiece(**json)
|
44
|
+
|
45
|
+
@staticmethod
|
46
|
+
def from_json_array(json: dict) -> list["FinishedPiece"]:
|
47
|
+
tmp = [FinishedPiece.from_json(j) for j in json]
|
48
|
+
out: list[FinishedPiece] = []
|
49
|
+
for t in tmp:
|
50
|
+
if isinstance(t, FinishedPiece):
|
51
|
+
out.append(t)
|
52
|
+
return out
|
53
|
+
|
54
|
+
def __hash__(self) -> int:
|
55
|
+
return hash(self.part_number)
|