rclone-api 1.4.14__tar.gz → 1.4.18__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (107) hide show
  1. {rclone_api-1.4.14 → rclone_api-1.4.18}/PKG-INFO +1 -1
  2. {rclone_api-1.4.14 → rclone_api-1.4.18}/pyproject.toml +1 -1
  3. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/cmd/copy_large_s3_finish.py +78 -17
  4. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/process.py +65 -40
  5. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/s3/api.py +3 -3
  6. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/s3/create.py +31 -27
  7. rclone_api-1.4.18/src/rclone_api/s3/merge_state.py +115 -0
  8. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/s3/multipart/finished_piece.py +15 -1
  9. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/s3/s3_multipart_uploader_by_copy.py +139 -81
  10. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/util.py +84 -23
  11. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api.egg-info/PKG-INFO +1 -1
  12. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api.egg-info/SOURCES.txt +1 -0
  13. {rclone_api-1.4.14 → rclone_api-1.4.18}/.aiderignore +0 -0
  14. {rclone_api-1.4.14 → rclone_api-1.4.18}/.github/workflows/lint.yml +0 -0
  15. {rclone_api-1.4.14 → rclone_api-1.4.18}/.github/workflows/push_macos.yml +0 -0
  16. {rclone_api-1.4.14 → rclone_api-1.4.18}/.github/workflows/push_ubuntu.yml +0 -0
  17. {rclone_api-1.4.14 → rclone_api-1.4.18}/.github/workflows/push_win.yml +0 -0
  18. {rclone_api-1.4.14 → rclone_api-1.4.18}/.gitignore +0 -0
  19. {rclone_api-1.4.14 → rclone_api-1.4.18}/.pylintrc +0 -0
  20. {rclone_api-1.4.14 → rclone_api-1.4.18}/.vscode/launch.json +0 -0
  21. {rclone_api-1.4.14 → rclone_api-1.4.18}/.vscode/settings.json +0 -0
  22. {rclone_api-1.4.14 → rclone_api-1.4.18}/.vscode/tasks.json +0 -0
  23. {rclone_api-1.4.14 → rclone_api-1.4.18}/LICENSE +0 -0
  24. {rclone_api-1.4.14 → rclone_api-1.4.18}/MANIFEST.in +0 -0
  25. {rclone_api-1.4.14 → rclone_api-1.4.18}/README.md +0 -0
  26. {rclone_api-1.4.14 → rclone_api-1.4.18}/clean +0 -0
  27. {rclone_api-1.4.14 → rclone_api-1.4.18}/install +0 -0
  28. {rclone_api-1.4.14 → rclone_api-1.4.18}/lint +0 -0
  29. {rclone_api-1.4.14 → rclone_api-1.4.18}/requirements.testing.txt +0 -0
  30. {rclone_api-1.4.14 → rclone_api-1.4.18}/setup.cfg +0 -0
  31. {rclone_api-1.4.14 → rclone_api-1.4.18}/setup.py +0 -0
  32. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/__init__.py +0 -0
  33. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/assets/example.txt +0 -0
  34. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/cli.py +0 -0
  35. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/cmd/analyze.py +0 -0
  36. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/cmd/copy_large_s3.py +0 -0
  37. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/cmd/list_files.py +0 -0
  38. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/cmd/save_to_db.py +0 -0
  39. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/completed_process.py +0 -0
  40. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/config.py +0 -0
  41. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/convert.py +0 -0
  42. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/db/__init__.py +0 -0
  43. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/db/db.py +0 -0
  44. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/db/models.py +0 -0
  45. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/deprecated.py +0 -0
  46. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/detail/copy_file_parts.py +0 -0
  47. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/detail/walk.py +0 -0
  48. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/diff.py +0 -0
  49. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/dir.py +0 -0
  50. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/dir_listing.py +0 -0
  51. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/exec.py +0 -0
  52. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/experimental/flags.py +0 -0
  53. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/experimental/flags_base.py +0 -0
  54. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/file.py +0 -0
  55. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/file_item.py +0 -0
  56. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/file_part.py +0 -0
  57. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/file_stream.py +0 -0
  58. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/filelist.py +0 -0
  59. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/group_files.py +0 -0
  60. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/http_server.py +0 -0
  61. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/log.py +0 -0
  62. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/mount.py +0 -0
  63. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/rclone_impl.py +0 -0
  64. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/remote.py +0 -0
  65. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/rpath.py +0 -0
  66. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/s3/basic_ops.py +0 -0
  67. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/s3/chunk_task.py +0 -0
  68. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/s3/multipart/file_info.py +0 -0
  69. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/s3/multipart/upload_info.py +0 -0
  70. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/s3/multipart/upload_state.py +0 -0
  71. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/s3/types.py +0 -0
  72. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/s3/upload_file_multipart.py +0 -0
  73. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/scan_missing_folders.py +0 -0
  74. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api/types.py +0 -0
  75. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api.egg-info/dependency_links.txt +0 -0
  76. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api.egg-info/entry_points.txt +0 -0
  77. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api.egg-info/requires.txt +0 -0
  78. {rclone_api-1.4.14 → rclone_api-1.4.18}/src/rclone_api.egg-info/top_level.txt +0 -0
  79. {rclone_api-1.4.14 → rclone_api-1.4.18}/test +0 -0
  80. {rclone_api-1.4.14 → rclone_api-1.4.18}/tests/archive/test_paramiko.py.disabled +0 -0
  81. {rclone_api-1.4.14 → rclone_api-1.4.18}/tests/test_cmd_list_files.py +0 -0
  82. {rclone_api-1.4.14 → rclone_api-1.4.18}/tests/test_copy.py +0 -0
  83. {rclone_api-1.4.14 → rclone_api-1.4.18}/tests/test_copy_bytes.py +0 -0
  84. {rclone_api-1.4.14 → rclone_api-1.4.18}/tests/test_copy_file_resumable_s3.py +0 -0
  85. {rclone_api-1.4.14 → rclone_api-1.4.18}/tests/test_copy_files.py +0 -0
  86. {rclone_api-1.4.14 → rclone_api-1.4.18}/tests/test_db.py +0 -0
  87. {rclone_api-1.4.14 → rclone_api-1.4.18}/tests/test_diff.py +0 -0
  88. {rclone_api-1.4.14 → rclone_api-1.4.18}/tests/test_file_item.py +0 -0
  89. {rclone_api-1.4.14 → rclone_api-1.4.18}/tests/test_group_files.py +0 -0
  90. {rclone_api-1.4.14 → rclone_api-1.4.18}/tests/test_is_synced.py +0 -0
  91. {rclone_api-1.4.14 → rclone_api-1.4.18}/tests/test_ls.py +0 -0
  92. {rclone_api-1.4.14 → rclone_api-1.4.18}/tests/test_ls_stream_files.py +0 -0
  93. {rclone_api-1.4.14 → rclone_api-1.4.18}/tests/test_mount.py +0 -0
  94. {rclone_api-1.4.14 → rclone_api-1.4.18}/tests/test_mount_s3.py +0 -0
  95. {rclone_api-1.4.14 → rclone_api-1.4.18}/tests/test_obscure.py +0 -0
  96. {rclone_api-1.4.14 → rclone_api-1.4.18}/tests/test_rclone_config.py +0 -0
  97. {rclone_api-1.4.14 → rclone_api-1.4.18}/tests/test_read_write_text.py +0 -0
  98. {rclone_api-1.4.14 → rclone_api-1.4.18}/tests/test_remote_control.py +0 -0
  99. {rclone_api-1.4.14 → rclone_api-1.4.18}/tests/test_remotes.py +0 -0
  100. {rclone_api-1.4.14 → rclone_api-1.4.18}/tests/test_s3.py +0 -0
  101. {rclone_api-1.4.14 → rclone_api-1.4.18}/tests/test_scan_missing_folders.py +0 -0
  102. {rclone_api-1.4.14 → rclone_api-1.4.18}/tests/test_serve_http.py +0 -0
  103. {rclone_api-1.4.14 → rclone_api-1.4.18}/tests/test_size_files.py +0 -0
  104. {rclone_api-1.4.14 → rclone_api-1.4.18}/tests/test_size_suffix.py +0 -0
  105. {rclone_api-1.4.14 → rclone_api-1.4.18}/tests/test_walk.py +0 -0
  106. {rclone_api-1.4.14 → rclone_api-1.4.18}/tox.ini +0 -0
  107. {rclone_api-1.4.14 → rclone_api-1.4.18}/upload_package.sh +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.2
2
2
  Name: rclone_api
3
- Version: 1.4.14
3
+ Version: 1.4.18
4
4
  Summary: rclone api in python
5
5
  Home-page: https://github.com/zackees/rclone-api
6
6
  License: BSD 3-Clause License
@@ -25,7 +25,7 @@ dependencies = [
25
25
  ]
26
26
 
27
27
  # Change this with the version number bump.
28
- version = "1.4.14"
28
+ version = "1.4.18"
29
29
 
30
30
  [tool.setuptools]
31
31
  package-dir = {"" = "src"}
@@ -6,9 +6,13 @@ from pathlib import Path
6
6
  from rclone_api import Rclone
7
7
  from rclone_api.detail.copy_file_parts import InfoJson
8
8
  from rclone_api.s3.s3_multipart_uploader_by_copy import (
9
- finish_multipart_upload_from_keys,
9
+ Part,
10
+ S3MultiPartUploader,
10
11
  )
11
- from rclone_api.types import SizeSuffix
12
+
13
+ _TIMEOUT_READ = 900
14
+ _TIMEOUT_CONNECTION = 900
15
+ _MAX_WORKERS = 10
12
16
 
13
17
 
14
18
  @dataclass
@@ -50,7 +54,55 @@ def _parse_args() -> Args:
50
54
  return out
51
55
 
52
56
 
53
- def do_finish_part(rclone: Rclone, info: InfoJson, dst: str) -> None:
57
+ # def finish_multipart_upload_from_keys(
58
+ # s3_client: BaseClient,
59
+ # source_bucket: str,
60
+ # parts: list[Part],
61
+ # destination_bucket: str,
62
+ # destination_key: str,
63
+ # chunk_size: int, # 5MB default
64
+ # max_workers: int = 100,
65
+ # retries: int = 3,
66
+ # ) -> str | Exception:
67
+ # """
68
+ # Finish a multipart upload by copying parts from existing S3 objects.
69
+
70
+ # Args:
71
+ # s3_client: Boto3 S3 client
72
+ # source_bucket: Source bucket name
73
+ # source_keys: List of source object keys to copy from
74
+ # destination_bucket: Destination bucket name
75
+ # destination_key: Destination object key
76
+ # chunk_size: Size of each part in bytes
77
+ # retries: Number of retry attempts
78
+ # byte_ranges: Optional list of byte ranges corresponding to source_keys
79
+
80
+ # Returns:
81
+ # The URL of the completed object
82
+ # """
83
+
84
+ # # Create upload info
85
+ # info = begin_upload(
86
+ # s3_client=s3_client,
87
+ # parts=parts,
88
+ # destination_bucket=destination_bucket,
89
+ # destination_key=destination_key,
90
+ # chunk_size=chunk_size,
91
+ # retries=retries,
92
+ # )
93
+
94
+ # out = do_body_work(
95
+ # info=info,
96
+ # source_bucket=source_bucket,
97
+ # parts=parts,
98
+ # max_workers=max_workers,
99
+ # retries=retries,
100
+ # )
101
+
102
+ # return out
103
+
104
+
105
+ def do_finish_part(rclone: Rclone, info: InfoJson, dst: str) -> Exception | None:
54
106
  from rclone_api.s3.create import (
55
107
  BaseClient,
56
108
  S3Config,
@@ -58,12 +110,17 @@ def do_finish_part(rclone: Rclone, info: InfoJson, dst: str) -> None:
58
110
  create_s3_client,
59
111
  )
60
112
 
61
- s3_creds: S3Credentials = rclone.impl.get_s3_credentials(remote=dst)
62
- s3_client: BaseClient = create_s3_client(
63
- s3_creds, S3Config(verbose=False, timeout_read=5 * 60)
113
+ s3_config = S3Config(
114
+ verbose=False,
115
+ timeout_read=_TIMEOUT_READ,
116
+ timeout_connection=_TIMEOUT_CONNECTION,
64
117
  )
118
+
119
+ s3_creds: S3Credentials = rclone.impl.get_s3_credentials(remote=dst)
120
+ s3_client: BaseClient = create_s3_client(s3_creds=s3_creds, s3_config=s3_config)
65
121
  s3_bucket = s3_creds.bucket_name
66
122
  is_done = info.fetch_is_done()
123
+ size = info.size
67
124
  assert is_done, f"Upload is not done: {info}"
68
125
 
69
126
  parts_dir = info.parts_dir
@@ -79,11 +136,9 @@ def do_finish_part(rclone: Rclone, info: InfoJson, dst: str) -> None:
79
136
 
80
137
  first_part: int | None = info.first_part
81
138
  last_part: int | None = info.last_part
82
- size: SizeSuffix | None = info.size
83
139
 
84
140
  assert first_part is not None
85
141
  assert last_part is not None
86
- assert size is not None
87
142
 
88
143
  def _to_s3_key(name: str | None) -> str:
89
144
  if name:
@@ -92,11 +147,12 @@ def do_finish_part(rclone: Rclone, info: InfoJson, dst: str) -> None:
92
147
  out = f"{parts_path}"
93
148
  return out
94
149
 
95
- parts: list[tuple[int, str]] = []
150
+ parts: list[Part] = []
96
151
  part_num = 1
97
152
  for part_key in source_keys:
98
153
  s3_key = _to_s3_key(name=part_key)
99
- parts.append((part_num, s3_key))
154
+ part = Part(part_number=part_num, s3_key=s3_key)
155
+ parts.append(part)
100
156
  part_num += 1
101
157
 
102
158
  chunksize = info.chunksize
@@ -107,31 +163,36 @@ def do_finish_part(rclone: Rclone, info: InfoJson, dst: str) -> None:
107
163
  # dst_key =
108
164
  dst_key = f"{dst_dir}/{dst_name}"
109
165
 
110
- finish_multipart_upload_from_keys(
166
+ uploader: S3MultiPartUploader = S3MultiPartUploader(
111
167
  s3_client=s3_client,
112
- source_bucket=s3_creds.bucket_name,
168
+ verbose=True,
169
+ )
170
+
171
+ from rclone_api.s3.s3_multipart_uploader_by_copy import MultipartUploadInfo
172
+
173
+ state: MultipartUploadInfo = uploader.begin_new_upload(
113
174
  parts=parts,
114
175
  destination_bucket=s3_creds.bucket_name,
115
176
  destination_key=dst_key,
116
177
  chunk_size=chunksize.as_int(),
117
- final_size=size.as_int(),
118
- max_workers=50,
119
- retries=3,
120
178
  )
121
179
 
180
+ uploader.start_upload(info=state, parts=parts, max_workers=_MAX_WORKERS)
181
+
122
182
  # now check if the dst now exists, if so, delete the parts folder.
123
183
  # if rclone.exists(dst):
124
184
  # rclone.purge(parts_dir)
125
185
 
126
186
  if not rclone.exists(dst):
127
- raise FileNotFoundError(f"Destination file not found: {dst}")
187
+ return FileNotFoundError(f"Destination file not found: {dst}")
128
188
 
129
189
  write_size = rclone.size_file(dst)
130
190
  if write_size != size:
131
- raise ValueError(f"Size mismatch: {write_size} != {size}")
191
+ return ValueError(f"Size mismatch: {write_size} != {size}")
132
192
 
133
193
  print(f"Upload complete: {dst}")
134
194
  rclone.purge(parts_dir)
195
+ return None
135
196
 
136
197
 
137
198
  def main() -> int:
@@ -1,12 +1,13 @@
1
1
  import atexit
2
2
  import subprocess
3
3
  import threading
4
- import time
5
4
  import weakref
6
5
  from dataclasses import dataclass
7
6
  from pathlib import Path
8
7
  from typing import Any
9
8
 
9
+ import psutil
10
+
10
11
  from rclone_api.config import Config
11
12
  from rclone_api.util import clear_temp_config_file, get_verbose, make_temp_config_file
12
13
 
@@ -24,20 +25,25 @@ class ProcessArgs:
24
25
 
25
26
  class Process:
26
27
  def __init__(self, args: ProcessArgs) -> None:
27
- assert args.rclone_exe.exists()
28
+ assert (
29
+ args.rclone_exe.exists()
30
+ ), f"rclone executable not found: {args.rclone_exe}"
28
31
  self.args = args
29
32
  self.log = args.log
30
33
  self.tempfile: Path | None = None
34
+
31
35
  verbose = get_verbose(args.verbose)
36
+ # Create a temporary config file if needed.
32
37
  if isinstance(args.rclone_conf, Config):
33
- self.tmpfile = make_temp_config_file()
34
- self.tmpfile.write_text(args.rclone_conf.text, encoding="utf-8")
35
- rclone_conf = self.tmpfile
38
+ self.tempfile = make_temp_config_file()
39
+ self.tempfile.write_text(args.rclone_conf.text, encoding="utf-8")
40
+ rclone_conf = self.tempfile
36
41
  else:
37
42
  rclone_conf = args.rclone_conf
38
43
 
39
- assert rclone_conf.exists()
44
+ assert rclone_conf.exists(), f"rclone config not found: {rclone_conf}"
40
45
 
46
+ # Build the command.
41
47
  self.cmd = (
42
48
  [str(args.rclone_exe.resolve())]
43
49
  + ["--config", str(rclone_conf.resolve())]
@@ -49,16 +55,14 @@ class Process:
49
55
  if verbose:
50
56
  cmd_str = subprocess.list2cmdline(self.cmd)
51
57
  print(f"Running: {cmd_str}")
52
- kwargs: dict = {}
53
- kwargs["shell"] = False
58
+ kwargs: dict = {"shell": False}
54
59
  if args.capture_stdout:
55
60
  kwargs["stdout"] = subprocess.PIPE
56
61
  kwargs["stderr"] = subprocess.STDOUT
57
62
 
58
63
  self.process = subprocess.Popen(self.cmd, **kwargs) # type: ignore
59
64
 
60
- # Register an atexit callback using a weak reference to avoid
61
- # keeping the Process instance alive solely due to the callback.
65
+ # Register an atexit callback using a weak reference to avoid keeping the Process instance alive.
62
66
  self_ref = weakref.ref(self)
63
67
 
64
68
  def exit_cleanup():
@@ -77,39 +81,60 @@ class Process:
77
81
  self.cleanup()
78
82
 
79
83
  def cleanup(self) -> None:
80
- clear_temp_config_file(self.tempfile)
84
+ if self.tempfile:
85
+ clear_temp_config_file(self.tempfile)
81
86
 
82
- def _atexit_terminate(self) -> None:
87
+ def _kill_process_tree(self) -> None:
83
88
  """
84
- Registered via atexit, this method attempts to gracefully terminate the process.
85
- If the process does not exit within a short timeout, it is aggressively killed.
89
+ Use psutil to recursively terminate the main process and all its child processes.
86
90
  """
87
- if self.process.poll() is None: # Process is still running
88
-
89
- def terminate_sequence():
91
+ try:
92
+ parent = psutil.Process(self.process.pid)
93
+ except psutil.NoSuchProcess:
94
+ return
95
+
96
+ # Terminate child processes.
97
+ children = parent.children(recursive=True)
98
+ if children:
99
+ print(f"Terminating {len(children)} child processes...")
100
+ for child in children:
90
101
  try:
91
- # Try to terminate gracefully.
92
- self.process.terminate()
102
+ child.terminate()
93
103
  except Exception as e:
94
- print(f"Error calling terminate on process {self.process.pid}: {e}")
95
- # Allow time for graceful shutdown.
96
- timeout = 2 # seconds
97
- start = time.time()
98
- while self.process.poll() is None and (time.time() - start) < timeout:
99
- time.sleep(0.1)
100
- # If still running, kill aggressively.
101
- if self.process.poll() is None:
104
+ print(f"Error terminating child process {child.pid}: {e}")
105
+ psutil.wait_procs(children, timeout=2)
106
+ # Kill any that remain.
107
+ for child in children:
108
+ if child.is_running():
102
109
  try:
103
- self.process.kill()
110
+ child.kill()
104
111
  except Exception as e:
105
- print(f"Error calling kill on process {self.process.pid}: {e}")
106
- # Optionally wait briefly for termination.
112
+ print(f"Error killing child process {child.pid}: {e}")
113
+
114
+ # Terminate the parent process.
115
+ if parent.is_running():
116
+ try:
117
+ parent.terminate()
118
+ except Exception as e:
119
+ print(f"Error terminating process {parent.pid}: {e}")
120
+ try:
121
+ parent.wait(timeout=3)
122
+ except psutil.TimeoutExpired:
107
123
  try:
108
- self.process.wait(timeout=1)
109
- except Exception:
110
- pass
124
+ parent.kill()
125
+ except Exception as e:
126
+ print(f"Error killing process {parent.pid}: {e}")
127
+
128
+ def _atexit_terminate(self) -> None:
129
+ """
130
+ This method is registered via atexit and uses psutil to clean up the process tree.
131
+ It runs in a daemon thread so that termination happens without blocking interpreter shutdown.
132
+ """
133
+ if self.process.poll() is None: # Process is still running.
134
+
135
+ def terminate_sequence():
136
+ self._kill_process_tree()
111
137
 
112
- # Run the termination sequence in a separate daemon thread.
113
138
  t = threading.Thread(target=terminate_sequence, daemon=True)
114
139
  t.start()
115
140
  t.join(timeout=3)
@@ -122,12 +147,12 @@ class Process:
122
147
  self.cleanup()
123
148
 
124
149
  def kill(self) -> None:
125
- self.cleanup()
126
- return self.process.kill()
150
+ """Forcefully kill the process tree."""
151
+ self._kill_process_tree()
127
152
 
128
153
  def terminate(self) -> None:
129
- self.cleanup()
130
- return self.process.terminate()
154
+ """Gracefully terminate the process tree."""
155
+ self._kill_process_tree()
131
156
 
132
157
  @property
133
158
  def returncode(self) -> int | None:
@@ -147,8 +172,8 @@ class Process:
147
172
  def wait(self) -> int:
148
173
  return self.process.wait()
149
174
 
150
- def send_signal(self, signal: int) -> None:
151
- return self.process.send_signal(signal)
175
+ def send_signal(self, sig: int) -> None:
176
+ self.process.send_signal(sig)
152
177
 
153
178
  def __str__(self) -> str:
154
179
  state = ""
@@ -20,11 +20,11 @@ _MIN_THRESHOLD_FOR_CHUNKING = 5 * 1024 * 1024
20
20
 
21
21
 
22
22
  class S3Client:
23
- def __init__(self, credentials: S3Credentials, verbose: bool = False) -> None:
23
+ def __init__(self, s3_creds: S3Credentials, verbose: bool = False) -> None:
24
24
  self.verbose = verbose
25
- self.credentials: S3Credentials = credentials
25
+ self.credentials: S3Credentials = s3_creds
26
26
  self.client: BaseClient = create_s3_client(
27
- credentials, config=S3Config(verbose=verbose)
27
+ s3_creds=s3_creds, s3_config=S3Config(verbose=verbose)
28
28
  )
29
29
 
30
30
  def list_bucket_contents(self, bucket_name: str) -> None:
@@ -8,7 +8,7 @@ from botocore.config import Config
8
8
  from rclone_api.s3.types import S3Credentials, S3Provider
9
9
 
10
10
  _DEFAULT_BACKBLAZE_ENDPOINT = "https://s3.us-west-002.backblazeb2.com"
11
- _MAX_CONNECTIONS = 50
11
+ _MAX_CONNECTIONS = 10
12
12
  _TIMEOUT_READ = 120
13
13
  _TIMEOUT_CONNECT = 60
14
14
 
@@ -30,14 +30,16 @@ class S3Config:
30
30
  # Create a Boto3 session and S3 client, this is back blaze specific.
31
31
  # Add a function if you want to use a different S3 provider.
32
32
  # If AWS support is added in a fork then please merge it back here.
33
- def _create_backblaze_s3_client(creds: S3Credentials, config: S3Config) -> BaseClient:
33
+ def _create_backblaze_s3_client(
34
+ s3_creds: S3Credentials, s3_config: S3Config
35
+ ) -> BaseClient:
34
36
  """Create and return an S3 client."""
35
- region_name = creds.region_name
36
- access_key = creds.access_key_id
37
- secret_key = creds.secret_access_key
38
- endpoint_url = creds.endpoint_url
37
+ region_name = s3_creds.region_name
38
+ access_key = s3_creds.access_key_id
39
+ secret_key = s3_creds.secret_access_key
40
+ endpoint_url = s3_creds.endpoint_url
39
41
  endpoint_url = endpoint_url or _DEFAULT_BACKBLAZE_ENDPOINT
40
- config.resolve_defaults()
42
+ s3_config.resolve_defaults()
41
43
  session = boto3.session.Session() # type: ignore
42
44
  return session.client(
43
45
  service_name="s3",
@@ -48,9 +50,9 @@ def _create_backblaze_s3_client(creds: S3Credentials, config: S3Config) -> BaseC
48
50
  config=Config(
49
51
  signature_version="s3v4",
50
52
  region_name=region_name,
51
- max_pool_connections=config.max_pool_connections,
52
- read_timeout=config.timeout_read,
53
- connect_timeout=config.timeout_connection,
53
+ max_pool_connections=s3_config.max_pool_connections,
54
+ read_timeout=s3_config.timeout_read,
55
+ connect_timeout=s3_config.timeout_connection,
54
56
  # Note that BackBlase has a boko3 bug where it doesn't support the new
55
57
  # checksum header, the following line was an attempt of fix it on the newest
56
58
  # version of boto3, but it didn't work.
@@ -59,18 +61,20 @@ def _create_backblaze_s3_client(creds: S3Credentials, config: S3Config) -> BaseC
59
61
  )
60
62
 
61
63
 
62
- def _create_unknown_s3_client(creds: S3Credentials, config: S3Config) -> BaseClient:
64
+ def _create_unknown_s3_client(
65
+ s3_creds: S3Credentials, s3_config: S3Config
66
+ ) -> BaseClient:
63
67
  """Create and return an S3 client."""
64
- access_key = creds.access_key_id
65
- secret_key = creds.secret_access_key
66
- endpoint_url = creds.endpoint_url
68
+ access_key = s3_creds.access_key_id
69
+ secret_key = s3_creds.secret_access_key
70
+ endpoint_url = s3_creds.endpoint_url
67
71
  if (endpoint_url is not None) and not (endpoint_url.startswith("http")):
68
- if config.verbose:
72
+ if s3_config.verbose:
69
73
  warnings.warn(
70
74
  f"Endpoint URL is schema naive: {endpoint_url}, assuming HTTPS"
71
75
  )
72
76
  endpoint_url = f"https://{endpoint_url}"
73
- config.resolve_defaults()
77
+ s3_config.resolve_defaults()
74
78
  session = boto3.session.Session() # type: ignore
75
79
  return session.client(
76
80
  service_name="s3",
@@ -79,25 +83,25 @@ def _create_unknown_s3_client(creds: S3Credentials, config: S3Config) -> BaseCli
79
83
  endpoint_url=endpoint_url,
80
84
  config=Config(
81
85
  signature_version="s3v4",
82
- region_name=creds.region_name,
83
- max_pool_connections=config.max_pool_connections,
84
- read_timeout=config.timeout_read,
85
- connect_timeout=config.timeout_connection,
86
+ region_name=s3_creds.region_name,
87
+ max_pool_connections=s3_config.max_pool_connections,
88
+ read_timeout=s3_config.timeout_read,
89
+ connect_timeout=s3_config.timeout_connection,
86
90
  ),
87
91
  )
88
92
 
89
93
 
90
94
  def create_s3_client(
91
- credentials: S3Credentials, config: S3Config | None = None
95
+ s3_creds: S3Credentials, s3_config: S3Config | None = None
92
96
  ) -> BaseClient:
93
97
  """Create and return an S3 client."""
94
- config = config or S3Config()
95
- provider = credentials.provider
98
+ s3_config = s3_config or S3Config()
99
+ provider = s3_creds.provider
96
100
  if provider == S3Provider.BACKBLAZE:
97
- if config.verbose:
101
+ if s3_config.verbose:
98
102
  print("Creating BackBlaze S3 client")
99
- return _create_backblaze_s3_client(creds=credentials, config=config)
103
+ return _create_backblaze_s3_client(s3_creds=s3_creds, s3_config=s3_config)
100
104
  else:
101
- if config.verbose:
105
+ if s3_config.verbose:
102
106
  print("Creating generic/unknown S3 client")
103
- return _create_unknown_s3_client(creds=credentials, config=config)
107
+ return _create_unknown_s3_client(s3_creds=s3_creds, s3_config=s3_config)
@@ -0,0 +1,115 @@
1
+ """
2
+ https://boto3.amazonaws.com/v1/documentation/api/latest/reference/services/s3/client/upload_part_copy.html
3
+ * client.upload_part_copy
4
+
5
+ This module provides functionality for S3 multipart uploads, including copying parts
6
+ from existing S3 objects using upload_part_copy.
7
+ """
8
+
9
+ import json
10
+ from dataclasses import dataclass
11
+ from typing import Any, Callable
12
+
13
+ from rclone_api.s3.multipart.finished_piece import FinishedPiece
14
+
15
+
16
+ @dataclass
17
+ class Part:
18
+ part_number: int
19
+ s3_key: str
20
+
21
+ def to_json(self) -> dict:
22
+ return {"part_number": self.part_number, "s3_key": self.s3_key}
23
+
24
+ @staticmethod
25
+ def from_json(json_dict: dict) -> "Part | Exception":
26
+ part_number = json_dict.get("part_number")
27
+ s3_key = json_dict.get("s3_key")
28
+ if part_number is None or s3_key is None:
29
+ return Exception(f"Invalid JSON: {json_dict}")
30
+ return Part(part_number=part_number, s3_key=s3_key)
31
+
32
+ @staticmethod
33
+ def from_json_array(json_array: list[dict]) -> list["Part"] | Exception:
34
+ try:
35
+ out: list[Part] = []
36
+ for j in json_array:
37
+ ok_or_err = Part.from_json(j)
38
+ if isinstance(ok_or_err, Exception):
39
+ return ok_or_err
40
+ else:
41
+ out.append(ok_or_err)
42
+ return out
43
+ except Exception as e:
44
+ return e
45
+
46
+
47
+ class MergeState:
48
+
49
+ def __init__(self, finished: list[FinishedPiece], all_parts: list[Part]) -> None:
50
+ self.finished: list[FinishedPiece] = list(finished)
51
+ self.all_parts: list[Part] = list(all_parts)
52
+ self.callbacks: list[Callable[[FinishedPiece], None]] = []
53
+
54
+ def add_callback(self, callback: Callable[[FinishedPiece], None]) -> None:
55
+ self.callbacks.append(callback)
56
+
57
+ def on_finished(self, finished_piece: FinishedPiece) -> None:
58
+ for callback in list(self.callbacks):
59
+ callback(finished_piece)
60
+
61
+ @staticmethod
62
+ def from_json_array(json_array: dict) -> "MergeState | Exception":
63
+ try:
64
+ finished: list[FinishedPiece] = FinishedPiece.from_json_array(
65
+ json_array["finished"]
66
+ )
67
+ all_parts: list[Part | Exception] = [
68
+ Part.from_json(j) for j in json_array["all"]
69
+ ]
70
+ all_parts_no_err: list[Part] = [
71
+ p for p in all_parts if not isinstance(p, Exception)
72
+ ]
73
+ errs: list[Exception] = [p for p in all_parts if isinstance(p, Exception)]
74
+ if len(errs):
75
+ return Exception(f"Errors in parts: {errs}")
76
+ return MergeState(finished=finished, all_parts=all_parts_no_err)
77
+ except Exception as e:
78
+ return e
79
+
80
+ def to_json(self) -> dict:
81
+ finished = self.finished.copy()
82
+ all_parts = self.all_parts.copy()
83
+ return {
84
+ "finished": FinishedPiece.to_json_array(finished),
85
+ "all": [part.to_json() for part in all_parts],
86
+ }
87
+
88
+ def to_json_str(self) -> str:
89
+ return json.dumps(self.to_json(), indent=1)
90
+
91
+ def __str__(self):
92
+ return self.to_json_str()
93
+
94
+ def __repr__(self):
95
+ return self.to_json_str()
96
+
97
+ def write(self, rclone_impl: Any, dst: str) -> None:
98
+ from rclone_api.rclone_impl import RcloneImpl
99
+
100
+ assert isinstance(rclone_impl, RcloneImpl)
101
+ json_str = self.to_json_str()
102
+ rclone_impl.write_text(dst, json_str)
103
+
104
+ def read(self, rclone_impl: Any, src: str) -> None:
105
+ from rclone_api.rclone_impl import RcloneImpl
106
+
107
+ assert isinstance(rclone_impl, RcloneImpl)
108
+ json_str = rclone_impl.read_text(src)
109
+ if isinstance(json_str, Exception):
110
+ raise json_str
111
+ json_dict = json.loads(json_str)
112
+ ok_or_err = FinishedPiece.from_json_array(json_dict["finished"])
113
+ if isinstance(ok_or_err, Exception):
114
+ raise ok_or_err
115
+ self.finished = ok_or_err
@@ -17,7 +17,9 @@ class FinishedPiece:
17
17
  return json.dumps(self.to_json(), indent=0)
18
18
 
19
19
  @staticmethod
20
- def to_json_array(parts: list["FinishedPiece | EndOfStream"]) -> list[dict]:
20
+ def to_json_array(
21
+ parts: list["FinishedPiece | EndOfStream"] | list["FinishedPiece"],
22
+ ) -> list[dict]:
21
23
  non_none: list[FinishedPiece] = []
22
24
  for p in parts:
23
25
  if not isinstance(p, EndOfStream):
@@ -39,3 +41,15 @@ class FinishedPiece:
39
41
  if json is None:
40
42
  return EndOfStream()
41
43
  return FinishedPiece(**json)
44
+
45
+ @staticmethod
46
+ def from_json_array(json: dict) -> list["FinishedPiece"]:
47
+ tmp = [FinishedPiece.from_json(j) for j in json]
48
+ out: list[FinishedPiece] = []
49
+ for t in tmp:
50
+ if isinstance(t, FinishedPiece):
51
+ out.append(t)
52
+ return out
53
+
54
+ def __hash__(self) -> int:
55
+ return hash(self.part_number)