mapillary-downloader 0.6.1__tar.gz → 0.7.0__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (19) hide show
  1. {mapillary_downloader-0.6.1 → mapillary_downloader-0.7.0}/PKG-INFO +24 -10
  2. {mapillary_downloader-0.6.1 → mapillary_downloader-0.7.0}/README.md +23 -9
  3. {mapillary_downloader-0.6.1 → mapillary_downloader-0.7.0}/pyproject.toml +1 -1
  4. {mapillary_downloader-0.6.1 → mapillary_downloader-0.7.0}/src/mapillary_downloader/tar_sequences.py +29 -28
  5. {mapillary_downloader-0.6.1 → mapillary_downloader-0.7.0}/src/mapillary_downloader/worker.py +15 -5
  6. {mapillary_downloader-0.6.1 → mapillary_downloader-0.7.0}/LICENSE.md +0 -0
  7. {mapillary_downloader-0.6.1 → mapillary_downloader-0.7.0}/src/mapillary_downloader/__init__.py +0 -0
  8. {mapillary_downloader-0.6.1 → mapillary_downloader-0.7.0}/src/mapillary_downloader/__main__.py +0 -0
  9. {mapillary_downloader-0.6.1 → mapillary_downloader-0.7.0}/src/mapillary_downloader/client.py +0 -0
  10. {mapillary_downloader-0.6.1 → mapillary_downloader-0.7.0}/src/mapillary_downloader/downloader.py +0 -0
  11. {mapillary_downloader-0.6.1 → mapillary_downloader-0.7.0}/src/mapillary_downloader/exif_writer.py +0 -0
  12. {mapillary_downloader-0.6.1 → mapillary_downloader-0.7.0}/src/mapillary_downloader/ia_check.py +0 -0
  13. {mapillary_downloader-0.6.1 → mapillary_downloader-0.7.0}/src/mapillary_downloader/ia_meta.py +0 -0
  14. {mapillary_downloader-0.6.1 → mapillary_downloader-0.7.0}/src/mapillary_downloader/ia_stats.py +0 -0
  15. {mapillary_downloader-0.6.1 → mapillary_downloader-0.7.0}/src/mapillary_downloader/logging_config.py +0 -0
  16. {mapillary_downloader-0.6.1 → mapillary_downloader-0.7.0}/src/mapillary_downloader/metadata_reader.py +0 -0
  17. {mapillary_downloader-0.6.1 → mapillary_downloader-0.7.0}/src/mapillary_downloader/utils.py +0 -0
  18. {mapillary_downloader-0.6.1 → mapillary_downloader-0.7.0}/src/mapillary_downloader/webp_converter.py +0 -0
  19. {mapillary_downloader-0.6.1 → mapillary_downloader-0.7.0}/src/mapillary_downloader/worker_pool.py +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: mapillary_downloader
3
- Version: 0.6.1
3
+ Version: 0.7.0
4
4
  Summary: Archive user data from Mapillary
5
5
  Author-email: Gareth Davidson <gaz@bitplane.net>
6
6
  Requires-Python: >=3.10
@@ -100,21 +100,28 @@ mapillary-downloader --no-webp USERNAME
100
100
 
101
101
  ## Tarballs
102
102
 
103
- Images are organized by sequence ID, bucketed by the first character of the
104
- sequence to reduce directory count:
103
+ Images are organized by capture date (YYYY-MM-DD) for incremental archiving:
105
104
 
106
105
  ```
107
106
  mapillary-username-quality/
108
- a/
107
+ 2024-01-15/
109
108
  abc123/
110
109
  image1.webp
111
110
  image2.webp
111
+ bcd456/
112
+ image3.webp
113
+ 2024-01-16/
114
+ def789/
115
+ image4.webp
112
116
  ```
113
117
 
114
- By default, these bucket directories are automatically tarred after download
115
- (resulting in `a.tar`, `b.tar`, etc. - about 62 tar files total). This is done
116
- because large collections with millions of images would otherwise create hundreds
117
- of thousands of tiny tars, and anger the archive gods.
118
+ By default, these date directories are automatically tarred after download
119
+ (resulting in `2024-01-15.tar`, `2024-01-16.tar`, etc.). This date-based
120
+ organization enables:
121
+
122
+ - **Incremental uploads** - Upload each day's tar as soon as it's ready
123
+ - **Manageable file counts** - ~365 days/year × 10 years = 3,650 tars max
124
+ - **Chronological organization** - Natural sorting and progress tracking
118
125
 
119
126
  To keep individual files instead of creating tars, use the `--no-tar` flag.
120
127
 
@@ -128,8 +135,15 @@ See inlay for details:
128
135
 
129
136
  * [📀 rip](https://bitplane.net/dev/sh/rip)
130
137
 
138
+ ## 📊 Stats
139
+
140
+ To see overall project progress, or an estimate, use `--stats`
141
+
142
+ ```bash
143
+ mapillary-downloader --stats
144
+ ```
131
145
 
132
- ## Development
146
+ ## 🚧 Development
133
147
 
134
148
  ```bash
135
149
  make dev # Setup dev environment
@@ -138,7 +152,7 @@ make dist # Build the distribution
138
152
  make help # See other make options
139
153
  ```
140
154
 
141
- ## Links
155
+ ## 🔗 Links
142
156
 
143
157
  * [🏠 home](https://bitplane.net/dev/python/mapillary_downloader)
144
158
  * [📖 pydoc](https://bitplane.net/dev/python/mapillary_downloader/pydoc)
@@ -70,21 +70,28 @@ mapillary-downloader --no-webp USERNAME
70
70
 
71
71
  ## Tarballs
72
72
 
73
- Images are organized by sequence ID, bucketed by the first character of the
74
- sequence to reduce directory count:
73
+ Images are organized by capture date (YYYY-MM-DD) for incremental archiving:
75
74
 
76
75
  ```
77
76
  mapillary-username-quality/
78
- a/
77
+ 2024-01-15/
79
78
  abc123/
80
79
  image1.webp
81
80
  image2.webp
81
+ bcd456/
82
+ image3.webp
83
+ 2024-01-16/
84
+ def789/
85
+ image4.webp
82
86
  ```
83
87
 
84
- By default, these bucket directories are automatically tarred after download
85
- (resulting in `a.tar`, `b.tar`, etc. - about 62 tar files total). This is done
86
- because large collections with millions of images would otherwise create hundreds
87
- of thousands of tiny tars, and anger the archive gods.
88
+ By default, these date directories are automatically tarred after download
89
+ (resulting in `2024-01-15.tar`, `2024-01-16.tar`, etc.). This date-based
90
+ organization enables:
91
+
92
+ - **Incremental uploads** - Upload each day's tar as soon as it's ready
93
+ - **Manageable file counts** - ~365 days/year × 10 years = 3,650 tars max
94
+ - **Chronological organization** - Natural sorting and progress tracking
88
95
 
89
96
  To keep individual files instead of creating tars, use the `--no-tar` flag.
90
97
 
@@ -98,8 +105,15 @@ See inlay for details:
98
105
 
99
106
  * [📀 rip](https://bitplane.net/dev/sh/rip)
100
107
 
108
+ ## 📊 Stats
109
+
110
+ To see overall project progress, or an estimate, use `--stats`
111
+
112
+ ```bash
113
+ mapillary-downloader --stats
114
+ ```
101
115
 
102
- ## Development
116
+ ## 🚧 Development
103
117
 
104
118
  ```bash
105
119
  make dev # Setup dev environment
@@ -108,7 +122,7 @@ make dist # Build the distribution
108
122
  make help # See other make options
109
123
  ```
110
124
 
111
- ## Links
125
+ ## 🔗 Links
112
126
 
113
127
  * [🏠 home](https://bitplane.net/dev/python/mapillary_downloader)
114
128
  * [📖 pydoc](https://bitplane.net/dev/python/mapillary_downloader/pydoc)
@@ -1,7 +1,7 @@
1
1
  [project]
2
2
  name = "mapillary_downloader"
3
3
  description = "Archive user data from Mapillary"
4
- version = "0.6.1"
4
+ version = "0.7.0"
5
5
  authors = [
6
6
  { name = "Gareth Davidson", email = "gaz@bitplane.net" }
7
7
  ]
@@ -1,6 +1,7 @@
1
1
  """Tar sequence directories for efficient Internet Archive uploads."""
2
2
 
3
3
  import logging
4
+ import re
4
5
  import tarfile
5
6
  from pathlib import Path
6
7
  from mapillary_downloader.utils import format_size
@@ -9,7 +10,9 @@ logger = logging.getLogger("mapillary_downloader")
9
10
 
10
11
 
11
12
  def tar_sequence_directories(collection_dir):
12
- """Tar all sequence directories in a collection for faster IA uploads.
13
+ """Tar all date directories in a collection for faster IA uploads.
14
+
15
+ Organizes by capture date (YYYY-MM-DD) for incremental archive.org uploads.
13
16
 
14
17
  Args:
15
18
  collection_dir: Path to collection directory (e.g., mapillary-user-quality/)
@@ -23,44 +26,44 @@ def tar_sequence_directories(collection_dir):
23
26
  logger.error(f"Collection directory not found: {collection_dir}")
24
27
  return 0, 0
25
28
 
26
- # Find all bucket directories (skip special dirs)
27
- # Now we tar entire bucket dirs (e.g., a/, b/, etc) to get ~62 tar files
29
+ # Find all date directories (skip special dirs)
30
+ # Date format: YYYY-MM-DD or unknown-date
28
31
  skip_dirs = {".meta", "__pycache__"}
29
- bucket_dirs = []
32
+ date_dirs = []
30
33
 
31
34
  for item in collection_dir.iterdir():
32
35
  if item.is_dir() and item.name not in skip_dirs:
33
- # Check if this is a bucket dir (single char)
34
- if len(item.name) == 1:
35
- bucket_dirs.append(item)
36
+ # Check if this is a date dir (YYYY-MM-DD) or unknown-date
37
+ if re.match(r"\d{4}-\d{2}-\d{2}$", item.name) or item.name == "unknown-date":
38
+ date_dirs.append(item)
36
39
 
37
- if not bucket_dirs:
38
- logger.info("No bucket directories to tar")
40
+ if not date_dirs:
41
+ logger.info("No date directories to tar")
39
42
  return 0, 0
40
43
 
41
- # Sort bucket directories alphabetically for consistent progress tracking
42
- bucket_dirs = sorted(bucket_dirs, key=lambda x: x.name)
44
+ # Sort date directories chronologically (YYYY-MM-DD sorts naturally)
45
+ date_dirs = sorted(date_dirs, key=lambda x: x.name)
43
46
 
44
- logger.info(f"Tarring {len(bucket_dirs)} bucket directories...")
47
+ logger.info(f"Tarring {len(date_dirs)} date directories...")
45
48
 
46
49
  tarred_count = 0
47
50
  total_files = 0
48
51
  total_tar_bytes = 0
49
52
 
50
- for bucket_dir in bucket_dirs:
51
- bucket_name = bucket_dir.name
52
- tar_path = collection_dir / f"{bucket_name}.tar"
53
+ for date_dir in date_dirs:
54
+ date_name = date_dir.name
55
+ tar_path = collection_dir / f"{date_name}.tar"
53
56
 
54
- # Count files in bucket
55
- files_to_tar = sorted([f for f in bucket_dir.rglob("*") if f.is_file()], key=lambda x: str(x))
57
+ # Count files in date directory
58
+ files_to_tar = sorted([f for f in date_dir.rglob("*") if f.is_file()], key=lambda x: str(x))
56
59
  file_count = len(files_to_tar)
57
60
 
58
61
  if file_count == 0:
59
- logger.warning(f"Skipping empty bucket directory: {bucket_name}")
62
+ logger.warning(f"Skipping empty date directory: {date_name}")
60
63
  continue
61
64
 
62
65
  try:
63
- logger.info(f"Tarring bucket '{bucket_name}' ({file_count} files)...")
66
+ logger.info(f"Tarring date '{date_name}' ({file_count} files)...")
64
67
 
65
68
  # Create reproducible uncompressed tar (WebP already compressed)
66
69
  with tarfile.open(tar_path, "w") as tar:
@@ -87,36 +90,34 @@ def tar_sequence_directories(collection_dir):
87
90
  tar_size = tar_path.stat().st_size
88
91
  total_tar_bytes += tar_size
89
92
 
90
- # Remove original bucket directory
91
- for file in bucket_dir.rglob("*"):
93
+ # Remove original date directory
94
+ for file in date_dir.rglob("*"):
92
95
  if file.is_file():
93
96
  file.unlink()
94
97
 
95
98
  # Remove empty subdirs and main dir
96
- for subdir in list(bucket_dir.rglob("*")):
99
+ for subdir in list(date_dir.rglob("*")):
97
100
  if subdir.is_dir():
98
101
  try:
99
102
  subdir.rmdir()
100
103
  except OSError:
101
104
  pass # Not empty yet
102
105
 
103
- bucket_dir.rmdir()
106
+ date_dir.rmdir()
104
107
 
105
108
  tarred_count += 1
106
109
  total_files += file_count
107
110
 
108
- logger.info(f"Tarred bucket '{bucket_name}': {file_count:,} files, {format_size(tar_size)}")
111
+ logger.info(f"Tarred date '{date_name}': {file_count:,} files, {format_size(tar_size)}")
109
112
  else:
110
113
  logger.error(f"Tar file empty or not created: {tar_path}")
111
114
  if tar_path.exists():
112
115
  tar_path.unlink()
113
116
 
114
117
  except Exception as e:
115
- logger.error(f"Error tarring bucket {bucket_name}: {e}")
118
+ logger.error(f"Error tarring date {date_name}: {e}")
116
119
  if tar_path.exists():
117
120
  tar_path.unlink()
118
121
 
119
- logger.info(
120
- f"Tarred {tarred_count} sequences ({total_files:,} files, {format_size(total_tar_bytes)} total tar size)"
121
- )
122
+ logger.info(f"Tarred {tarred_count} dates ({total_files:,} files, {format_size(total_tar_bytes)} total tar size)")
122
123
  return tarred_count, total_files
@@ -3,6 +3,7 @@
3
3
  import os
4
4
  import signal
5
5
  import tempfile
6
+ from datetime import datetime
6
7
  from pathlib import Path
7
8
  import requests
8
9
  from mapillary_downloader.exif_writer import write_exif_to_image
@@ -69,16 +70,25 @@ def download_and_convert_image(image_data, output_dir, quality, convert_webp, se
69
70
  if not image_url:
70
71
  return (image_id, 0, False, f"No {quality} URL")
71
72
 
72
- # Determine final output directory - organize by first char of sequence ID
73
+ # Determine final output directory - organize by capture date
73
74
  output_dir = Path(output_dir)
74
75
  sequence_id = image_data.get("sequence")
76
+
77
+ # Extract date from captured_at timestamp (milliseconds since epoch)
78
+ captured_at = image_data.get("captured_at")
79
+ if captured_at:
80
+ # Convert to UTC date string (YYYY-MM-DD)
81
+ date_str = datetime.utcfromtimestamp(captured_at / 1000).strftime("%Y-%m-%d")
82
+ else:
83
+ # Fallback for missing timestamp (should be rare per API docs)
84
+ date_str = "unknown-date"
85
+
75
86
  if sequence_id:
76
- # Use first character as bucket (gives us ~62 dirs instead of millions)
77
- first_char = sequence_id[0]
78
- img_dir = output_dir / first_char / sequence_id
87
+ img_dir = output_dir / date_str / sequence_id
79
88
  img_dir.mkdir(parents=True, exist_ok=True)
80
89
  else:
81
- img_dir = output_dir
90
+ img_dir = output_dir / date_str
91
+ img_dir.mkdir(parents=True, exist_ok=True)
82
92
 
83
93
  # If converting to WebP, use /tmp for intermediate JPEG
84
94
  # Otherwise write JPEG directly to final location