mapillary-downloader 0.7.0__py3-none-any.whl → 0.7.2__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -43,8 +43,8 @@ def main():
43
43
  parser.add_argument(
44
44
  "--max-workers",
45
45
  type=int,
46
- default=128,
47
- help="Maximum number of parallel workers (default: 128)",
46
+ default=os.cpu_count() or 8,
47
+ help=f"Maximum number of parallel workers (default: CPU count = {os.cpu_count() or 8})",
48
48
  )
49
49
  parser.add_argument(
50
50
  "--no-tar",
@@ -67,7 +67,6 @@ class MapillaryDownloader:
67
67
  self.username = username
68
68
  self.quality = quality
69
69
  self.max_workers = max_workers
70
- self.initial_workers = os.cpu_count() or 1 # Start with CPU count
71
70
  self.tar_sequences = tar_sequences
72
71
  self.convert_webp = convert_webp
73
72
  self.check_ia = check_ia
@@ -173,7 +172,7 @@ class MapillaryDownloader:
173
172
  logger.info(f"Downloading images for user: {self.username}")
174
173
  logger.info(f"Output directory: {self.output_dir}")
175
174
  logger.info(f"Quality: {self.quality}")
176
- logger.info(f"Worker pool: {self.initial_workers} initial, {self.max_workers} max")
175
+ logger.info(f"Worker pool: max {self.max_workers} workers")
177
176
 
178
177
  start_time = time.time()
179
178
 
@@ -188,9 +187,7 @@ class MapillaryDownloader:
188
187
  # Step 2: Start worker pool
189
188
  # Since workers do both I/O (download) and CPU (WebP), need many more workers
190
189
  # Start with CPU count and scale up based on throughput
191
- pool = AdaptiveWorkerPool(
192
- worker_process, min_workers=self.initial_workers, max_workers=self.max_workers, monitoring_interval=10
193
- )
190
+ pool = AdaptiveWorkerPool(worker_process, max_workers=self.max_workers, monitoring_interval=10)
194
191
  pool.start()
195
192
 
196
193
  # Step 3: Download images from metadata file while fetching new from API
@@ -0,0 +1,193 @@
1
+ """GraphQL web API utilities (unofficial, experimental).
2
+
3
+ This module provides access to Mapillary's GraphQL endpoint used by the web interface.
4
+ Unlike the official v4 REST API, this requires a public web token extracted from the
5
+ JavaScript bundle.
6
+
7
+ Use cases:
8
+ - Get user image counts without pagination
9
+ - Access leaderboard data
10
+ - Check for updates to existing downloads
11
+
12
+ WARNING: This is not officially documented and may break at any time.
13
+ """
14
+
15
+ import json
16
+ import logging
17
+ import re
18
+ from datetime import datetime
19
+ from urllib.parse import urlencode, quote
20
+ import requests
21
+
22
+ logger = logging.getLogger("mapillary_downloader")
23
+
24
+ # Fallback token (extracted from main JS bundle as of 2025-01-09)
25
+ FALLBACK_TOKEN = "MLY|4223665974375089|d62822dd792b6a823d0794ef26450398"
26
+
27
+
28
+ def extract_token_from_js():
29
+ """Extract public web token from Mapillary's JavaScript bundle.
30
+
31
+ This fetches the main page, finds the main JS bundle, and extracts
32
+ the hardcoded MLY token used for GraphQL queries.
33
+
34
+ Returns:
35
+ Token string (e.g., "MLY|123|abc...") or None if extraction failed
36
+ """
37
+ try:
38
+ # Fetch main page to find JS bundle URL
39
+ # Need consent cookie to get actual page (not GDPR banner)
40
+ logger.debug("Fetching Mapillary main page...")
41
+ # Generate today's date in the format YYYY_MM_DD for cookie
42
+ today = datetime.now().strftime("%Y_%m_%d")
43
+ cookies = {
44
+ "mly_cb": f'{{"version":"1","date":"{today}","third_party_consent":"withdrawn","categories":{{"content_and_media":"withdrawn"}},"integration_controls":{{"YOUTUBE":"withdrawn"}}}}'
45
+ }
46
+ headers = {
47
+ "User-Agent": "Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:144.0) Gecko/20100101 Firefox/144.0",
48
+ "Accept": "text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8",
49
+ "Accept-Language": "en-GB,en;q=0.5",
50
+ "Sec-GPC": "1",
51
+ "Upgrade-Insecure-Requests": "1",
52
+ "Sec-Fetch-Dest": "document",
53
+ "Sec-Fetch-Mode": "navigate",
54
+ "Sec-Fetch-Site": "none",
55
+ "Sec-Fetch-User": "?1",
56
+ }
57
+ response = requests.get("https://www.mapillary.com/app/", cookies=cookies, headers=headers, timeout=30)
58
+ response.raise_for_status()
59
+
60
+ # Find main JS file URL
61
+ # Pattern: <script src="main.{hash}.js" type="module"></script>
62
+ js_match = re.search(r'src="(main\.[a-f0-9]+\.js)"', response.text)
63
+ if not js_match:
64
+ logger.warning("Could not find main JS bundle URL in page")
65
+ return None
66
+
67
+ # URL is relative to /app/ base path
68
+ js_url = f"https://www.mapillary.com/app/{js_match.group(1)}"
69
+ logger.debug(f"Found JS bundle: {js_url}")
70
+
71
+ # Fetch JS bundle
72
+ logger.debug("Fetching JS bundle...")
73
+ js_response = requests.get(js_url, timeout=30)
74
+ js_response.raise_for_status()
75
+
76
+ # Extract token
77
+ # Pattern: "MLY|{client_id}|{secret}"
78
+ token_match = re.search(r'"(MLY\|[^"]+)"', js_response.text)
79
+ if not token_match:
80
+ logger.warning("Could not find MLY token in JS bundle")
81
+ return None
82
+
83
+ token = token_match.group(1)
84
+ logger.info(f"Extracted web token: {token[:20]}...")
85
+ return token
86
+
87
+ except requests.RequestException as e:
88
+ logger.error(f"Failed to extract web token: {e}")
89
+ return None
90
+ except Exception as e:
91
+ logger.error(f"Unexpected error extracting web token: {e}")
92
+ return None
93
+
94
+
95
+ def get_leaderboard(key="global", token=None):
96
+ """Get leaderboard data from Mapillary GraphQL API.
97
+
98
+ Args:
99
+ key: Leaderboard key (e.g., "global", country name, etc.)
100
+ token: MLY token (if None, will extract from JS bundle or use fallback)
101
+
102
+ Returns:
103
+ Dict with leaderboard data, or None on error
104
+ """
105
+ if token is None:
106
+ token = extract_token_from_js()
107
+ if token is None:
108
+ logger.warning("Failed to extract token, using fallback")
109
+ token = FALLBACK_TOKEN
110
+
111
+ # GraphQL query for leaderboard (lifetime stats only)
112
+ query = """query getUserLeaderboard($key: String!) {
113
+ user_leaderboards(key: $key) {
114
+ lifetime {
115
+ count
116
+ user {
117
+ id
118
+ username
119
+ profile_photo_url
120
+ __typename
121
+ }
122
+ __typename
123
+ }
124
+ __typename
125
+ }
126
+ }"""
127
+
128
+ try:
129
+ headers = {
130
+ "User-Agent": "Mozilla/5.0 (X11; Ubuntu; Linux x86_64; rv:144.0) Gecko/20100101 Firefox/144.0",
131
+ "Accept": "*/*",
132
+ "Accept-Language": "en-GB,en;q=0.5",
133
+ "Referer": "https://www.mapillary.com/",
134
+ "content-type": "application/json",
135
+ "authorization": f"OAuth {token}",
136
+ "Origin": "https://www.mapillary.com",
137
+ "Sec-Fetch-Dest": "empty",
138
+ "Sec-Fetch-Mode": "cors",
139
+ "Sec-Fetch-Site": "same-site",
140
+ }
141
+
142
+ # Build query params - use quote_via=quote to get %20 instead of +
143
+ # Note: both 'doc' and 'query' params seem to be required (from observed curl)
144
+ params = {
145
+ "doc": query,
146
+ "query": query,
147
+ "operationName": "getUserLeaderboard",
148
+ "variables": json.dumps({"key": key}, separators=(',', ':')),
149
+ }
150
+
151
+ # Build URL with proper percent encoding (not + for spaces)
152
+ # Don't encode parentheses to match curl behavior
153
+ query_string = urlencode(params, quote_via=lambda s, safe='', encoding=None, errors=None: quote(s, safe='()!'))
154
+ url = f"https://graph.mapillary.com/graphql?{query_string}"
155
+
156
+ logger.debug(f"Querying leaderboard for key: {key}")
157
+
158
+ response = requests.get(
159
+ url,
160
+ headers=headers,
161
+ timeout=30
162
+ )
163
+ response.raise_for_status()
164
+
165
+ return response.json()
166
+
167
+ except requests.RequestException as e:
168
+ logger.error(f"Failed to query leaderboard: {e}")
169
+ return None
170
+ except Exception as e:
171
+ logger.error(f"Unexpected error querying leaderboard: {e}")
172
+ return None
173
+
174
+
175
+ if __name__ == "__main__":
176
+ # Test the extraction and leaderboard query
177
+ logging.basicConfig(level=logging.DEBUG)
178
+
179
+ print("=== Extracting token ===")
180
+ token = extract_token_from_js()
181
+ if token:
182
+ print(f"Success! Token: {token}")
183
+ else:
184
+ print("Failed to extract token")
185
+ print(f"Fallback: {FALLBACK_TOKEN}")
186
+ token = FALLBACK_TOKEN
187
+
188
+ print("\n=== Querying global leaderboard ===")
189
+ data = get_leaderboard("global", token=token)
190
+ if data:
191
+ print(json.dumps(data, indent=2))
192
+ else:
193
+ print("Failed to get leaderboard data")
@@ -182,7 +182,7 @@ def generate_ia_metadata(collection_dir):
182
182
  write_meta_tag(meta_dir, "coverage", f"{first_date} - {last_date}")
183
183
  write_meta_tag(meta_dir, "licenseurl", "https://creativecommons.org/licenses/by-sa/4.0/")
184
184
  write_meta_tag(meta_dir, "mediatype", "data")
185
- write_meta_tag(meta_dir, "collection", "opensource_media")
185
+ write_meta_tag(meta_dir, "collection", "mapillary-images")
186
186
 
187
187
  # Source and scanner metadata
188
188
  write_meta_tag(meta_dir, "source", f"https://www.mapillary.com/app/user/{username}")
@@ -17,17 +17,15 @@ class AdaptiveWorkerPool:
17
17
  - If throughput plateauing/decreasing: reduce workers
18
18
  """
19
19
 
20
- def __init__(self, worker_func, min_workers=4, max_workers=16, monitoring_interval=10):
20
+ def __init__(self, worker_func, max_workers=16, monitoring_interval=10):
21
21
  """Initialize adaptive worker pool.
22
22
 
23
23
  Args:
24
24
  worker_func: Function to run in each worker (must accept work_queue, result_queue)
25
- min_workers: Minimum number of workers
26
25
  max_workers: Maximum number of workers
27
26
  monitoring_interval: Seconds between throughput checks
28
27
  """
29
28
  self.worker_func = worker_func
30
- self.min_workers = min_workers
31
29
  self.max_workers = max_workers
32
30
  self.monitoring_interval = monitoring_interval
33
31
 
@@ -37,7 +35,8 @@ class AdaptiveWorkerPool:
37
35
 
38
36
  # Worker management
39
37
  self.workers = []
40
- self.current_workers = min_workers # Start small and ramp up
38
+ # Start at 25% of max_workers (at least 1)
39
+ self.current_workers = max(1, int(max_workers * 0.25))
41
40
 
42
41
  # Throughput monitoring
43
42
  self.throughput_history = deque(maxlen=5) # Last 5 measurements
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.4
2
2
  Name: mapillary_downloader
3
- Version: 0.7.0
3
+ Version: 0.7.2
4
4
  Summary: Archive user data from Mapillary
5
5
  Author-email: Gareth Davidson <gaz@bitplane.net>
6
6
  Requires-Python: >=3.10
@@ -66,7 +66,7 @@ mapillary-downloader --output ./downloads USERNAME1
66
66
  | `--quality` | 256, 1024, 2048 or original | `original` |
67
67
  | `--bbox` | `west,south,east,north` | `None` |
68
68
  | `--no-webp` | Don't convert to WebP | `False` |
69
- | `--max-workers` | Maximum number of parallel download workers | `128` |
69
+ | `--max-workers` | Maximum number of parallel download workers | CPU count |
70
70
  | `--no-tar` | Don't tar bucket directories | `False` |
71
71
  | `--no-check-ia` | Don't check if exists on Internet Archive | `False` |
72
72
 
@@ -1,10 +1,11 @@
1
1
  mapillary_downloader/__init__.py,sha256=KEjiBRghXDeA7E15RJeLBfQm-yNJkowZarL59QOh_1w,120
2
- mapillary_downloader/__main__.py,sha256=jK0MU9Xn9KGb_V8x5giIRuwhUjNFQ-jjlprtbeW6b94,4817
2
+ mapillary_downloader/__main__.py,sha256=iuDGZoFVu8q_dTvJuExSpj4Jx1x9xASSjUITRGwd0RA,4864
3
3
  mapillary_downloader/client.py,sha256=a5n43FLHP45EHodEjl0ieziBK-b6Ey-rZJwYB6EFhNI,4743
4
- mapillary_downloader/downloader.py,sha256=TrFy9eTcZD_wyVh7L58HuVxgCIKbAYRnmr2gAWtXOuY,19738
4
+ mapillary_downloader/downloader.py,sha256=UFSKx-U5UYFJkmdmSN-X5vV81rp3GYb8jmvYMSqWhDM,19583
5
5
  mapillary_downloader/exif_writer.py,sha256=K_441EG1siWyNMmFGZSfnORUCjBThkeg4JFtbg9AOsA,5120
6
+ mapillary_downloader/graphql_web.py,sha256=69pmx8xDEuu-zn7xy5CkpUBjijGTS506ZJd5I_GAepk,6671
6
7
  mapillary_downloader/ia_check.py,sha256=L2MEbG_KmlAd5NLmo2HQkO8HWvRN0brE5wXXoyNMbq8,1100
7
- mapillary_downloader/ia_meta.py,sha256=78rcybHIPnQDsF02KGj6RYmDXzYzrU8sdVx4Q9Y0sfI,6266
8
+ mapillary_downloader/ia_meta.py,sha256=3aj1RunSLap2Kh3q9Mm_4WpcVT92elfR6k-um_GRTpk,6266
8
9
  mapillary_downloader/ia_stats.py,sha256=TSVCoaCcGFDPTYqxikGdvMo7uWtExRniYABjQQS26fw,7302
9
10
  mapillary_downloader/logging_config.py,sha256=Z-wNq34nt7aIhJWdeKc1feTY46P9-Or7HtiX7eUFjEI,2324
10
11
  mapillary_downloader/metadata_reader.py,sha256=Re-HN0Vfc7Hs1eOut7uOoW7jWJ2PIbKoNzC7Ak3ah5o,4933
@@ -12,9 +13,9 @@ mapillary_downloader/tar_sequences.py,sha256=UchKvvajBr5uaoE8xDHgyiFTkjh08EK7pPh
12
13
  mapillary_downloader/utils.py,sha256=VgcwbC8yb2XlTGerTNwHBU42K2IN14VU7P-I52Vb01c,2947
13
14
  mapillary_downloader/webp_converter.py,sha256=vYLLQxDmdnqRz0nm7wXwRUd4x9mQZNah-DrncpA8sNs,1901
14
15
  mapillary_downloader/worker.py,sha256=K2DkQgFzALKs20TsG1KibNUdFiWN_v8MtVnBX_0xVyc,5162
15
- mapillary_downloader/worker_pool.py,sha256=iGRq5uFwBNNVQnI4vEjbKHkbKTaEVCdmvMvXcRGuDMg,8203
16
- mapillary_downloader-0.7.0.dist-info/entry_points.txt,sha256=PdYtxOXHMJrUhmiPO4G-F98VuhUI4MN9D_T4KPrVZ5w,75
17
- mapillary_downloader-0.7.0.dist-info/licenses/LICENSE.md,sha256=7_BIuQ-veOrsF-WarH8kTkm0-xrCLvJ1PFE1C4Ebs64,146
18
- mapillary_downloader-0.7.0.dist-info/WHEEL,sha256=G2gURzTEtmeR8nrdXUJfNiB3VYVxigPQ-bEQujpNiNs,82
19
- mapillary_downloader-0.7.0.dist-info/METADATA,sha256=Ftc--29thU8dc-J_11_NlBUnf6SsOlvQP4r28nclsnk,5540
20
- mapillary_downloader-0.7.0.dist-info/RECORD,,
16
+ mapillary_downloader/worker_pool.py,sha256=ZcI2fG-nYZ2dxdcyYO6Ay-YaoPJILlDMOwQUaQ6j14k,8142
17
+ mapillary_downloader-0.7.2.dist-info/entry_points.txt,sha256=PdYtxOXHMJrUhmiPO4G-F98VuhUI4MN9D_T4KPrVZ5w,75
18
+ mapillary_downloader-0.7.2.dist-info/licenses/LICENSE.md,sha256=7_BIuQ-veOrsF-WarH8kTkm0-xrCLvJ1PFE1C4Ebs64,146
19
+ mapillary_downloader-0.7.2.dist-info/WHEEL,sha256=G2gURzTEtmeR8nrdXUJfNiB3VYVxigPQ-bEQujpNiNs,82
20
+ mapillary_downloader-0.7.2.dist-info/METADATA,sha256=d_CN0yVO5_l51jT3zRkI0OlrbBY3Sw4R3-2OoHZqQYM,5540
21
+ mapillary_downloader-0.7.2.dist-info/RECORD,,