openeo-gfmap 0.1.0__py3-none-any.whl → 0.2.0__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,8 +1,11 @@
1
1
  """This sub-module contains utilitary function and tools for OpenEO-GFMap"""
2
2
 
3
+ import logging
4
+
3
5
  from openeo_gfmap.utils.build_df import load_json
4
6
  from openeo_gfmap.utils.intervals import quintad_intervals
5
7
  from openeo_gfmap.utils.netcdf import update_nc_attributes
8
+ from openeo_gfmap.utils.split_stac import split_collection_by_epsg
6
9
  from openeo_gfmap.utils.tile_processing import (
7
10
  array_bounds,
8
11
  arrays_cosine_similarity,
@@ -11,6 +14,18 @@ from openeo_gfmap.utils.tile_processing import (
11
14
  select_sar_bands,
12
15
  )
13
16
 
17
+ _log = logging.getLogger(__name__)
18
+ _log.setLevel(logging.INFO)
19
+
20
+ ch = logging.StreamHandler()
21
+ ch.setLevel(logging.INFO)
22
+
23
+ formatter = logging.Formatter("%(asctime)s - %(name)s - %(levelname)s - %(message)s")
24
+ ch.setFormatter(formatter)
25
+
26
+ _log.addHandler(ch)
27
+
28
+
14
29
  __all__ = [
15
30
  "load_json",
16
31
  "normalize_array",
@@ -19,5 +34,6 @@ __all__ = [
19
34
  "select_sar_bands",
20
35
  "arrays_cosine_similarity",
21
36
  "quintad_intervals",
37
+ "split_collection_by_epsg",
22
38
  "update_nc_attributes",
23
39
  ]
@@ -1,11 +1,15 @@
1
1
  """Functionalities to interract with product catalogues."""
2
2
 
3
+ from typing import Optional
4
+
5
+ import geojson
6
+ import pandas as pd
3
7
  import requests
4
- from geojson import GeoJSON
5
8
  from pyproj.crs import CRS
6
9
  from rasterio.warp import transform_bounds
7
- from shapely import unary_union
10
+ from requests import adapters
8
11
  from shapely.geometry import box, shape
12
+ from shapely.ops import unary_union
9
13
 
10
14
  from openeo_gfmap import (
11
15
  Backend,
@@ -14,6 +18,21 @@ from openeo_gfmap import (
14
18
  SpatialContext,
15
19
  TemporalContext,
16
20
  )
21
+ from openeo_gfmap.utils import _log
22
+
23
+ request_sessions: Optional[requests.Session] = None
24
+
25
+
26
+ def _request_session() -> requests.Session:
27
+ global request_sessions
28
+
29
+ if request_sessions is None:
30
+ request_sessions = requests.Session()
31
+ retries = adapters.Retry(
32
+ total=5, backoff_factor=1, status_forcelist=[500, 502, 503, 504]
33
+ )
34
+ request_sessions.mount("https://", adapters.HTTPAdapter(max_retries=retries))
35
+ return request_sessions
17
36
 
18
37
 
19
38
  class UncoveredS1Exception(Exception):
@@ -24,13 +43,21 @@ class UncoveredS1Exception(Exception):
24
43
 
25
44
 
26
45
  def _parse_cdse_products(response: dict):
27
- """Parses the geometry of products from the CDSE catalogue."""
28
- geoemetries = []
46
+ """Parses the geometry and timestamps of products from the CDSE catalogue."""
47
+ geometries = []
48
+ timestamps = []
29
49
  products = response["features"]
30
50
 
31
51
  for product in products:
32
- geoemetries.append(shape(product["geometry"]))
33
- return geoemetries
52
+ if "geometry" in product and "startDate" in product["properties"]:
53
+ geometries.append(shape(product["geometry"]))
54
+ timestamps.append(pd.to_datetime(product["properties"]["startDate"]))
55
+ else:
56
+ _log.warning(
57
+ "Cannot parse product %s does not have a geometry or timestamp.",
58
+ product["properties"]["id"],
59
+ )
60
+ return geometries, timestamps
34
61
 
35
62
 
36
63
  def _query_cdse_catalogue(
@@ -39,6 +66,14 @@ def _query_cdse_catalogue(
39
66
  temporal_extent: TemporalContext,
40
67
  **additional_parameters: dict,
41
68
  ) -> dict:
69
+ """
70
+ Queries the CDSE catalogue for a given collection, spatio-temporal context and additional
71
+ parameters.
72
+
73
+ Params
74
+ ------
75
+
76
+ """
42
77
  minx, miny, maxx, maxy = bounds
43
78
 
44
79
  # The date format should be YYYY-MM-DD
@@ -48,13 +83,14 @@ def _query_cdse_catalogue(
48
83
  url = (
49
84
  f"https://catalogue.dataspace.copernicus.eu/resto/api/collections/"
50
85
  f"{collection}/search.json?box={minx},{miny},{maxx},{maxy}"
51
- f"&sortParam=startDate&maxRecords=100"
52
- f"&dataset=ESA-DATASET&startDate={start_date}&completionDate={end_date}"
86
+ f"&sortParam=startDate&maxRecords=1000&dataset=ESA-DATASET"
87
+ f"&startDate={start_date}&completionDate={end_date}"
53
88
  )
54
89
  for key, value in additional_parameters.items():
55
90
  url += f"&{key}={value}"
56
91
 
57
- response = requests.get(url)
92
+ session = _request_session()
93
+ response = session.get(url, timeout=60)
58
94
 
59
95
  if response.status_code != 200:
60
96
  raise Exception(
@@ -107,19 +143,51 @@ def _check_cdse_catalogue(
107
143
  return len(grd_tiles) > 0
108
144
 
109
145
 
110
- def s1_area_per_orbitstate(
146
+ def _compute_max_gap_days(
147
+ temporal_extent: TemporalContext, timestamps: list[pd.DatetimeIndex]
148
+ ) -> int:
149
+ """Computes the maximum temporal gap in days from the timestamps parsed from the catalogue.
150
+ Requires the start and end date to be included in the timestamps to compute the gap before
151
+ and after the first and last observation.
152
+
153
+ Parameters
154
+ ----------
155
+ temporal_extent : TemporalContext
156
+ The temporal extent to be checked. Same as used to query the catalogue.
157
+ timestamps : list[pd.DatetimeIndex]
158
+ The list of timestamps parsed from the catalogue and to compute the gap from.
159
+
160
+ Returns
161
+ -------
162
+ days : int
163
+ The maximum temporal gap in days.
164
+ """
165
+ # Computes max temporal gap. Include requested start and end date so we dont miss
166
+ # any start or end gap before first/last observation
167
+ timestamps = pd.DatetimeIndex(
168
+ sorted(
169
+ [pd.to_datetime(temporal_extent.start_date, utc=True)]
170
+ + timestamps
171
+ + [pd.to_datetime(temporal_extent.end_date, utc=True)]
172
+ )
173
+ )
174
+ return timestamps.to_series().diff().max().days
175
+
176
+
177
+ def s1_area_per_orbitstate_vvvh(
111
178
  backend: BackendContext,
112
179
  spatial_extent: SpatialContext,
113
180
  temporal_extent: TemporalContext,
114
181
  ) -> dict:
115
- """Evaluates for both the ascending and descending state orbits the area of interesection
116
- between the given spatio-temporal context and the products available in the backend's
117
- catalogue.
182
+ """
183
+ Evaluates for both the ascending and descending state orbits the area of interesection and
184
+ maximum temporal gap for the available products with a VV&VH polarisation.
118
185
 
119
186
  Parameters
120
187
  ----------
121
188
  backend : BackendContext
122
- The backend to be within, as each backend might use different catalogues.
189
+ The backend to be within, as each backend might use different catalogues. Only the CDSE,
190
+ CDSE_STAGING and FED backends are supported.
123
191
  spatial_extent : SpatialContext
124
192
  The spatial extent to be checked, it will check within its bounding box.
125
193
  temporal_extent : TemporalContext
@@ -128,13 +196,17 @@ def s1_area_per_orbitstate(
128
196
  Returns
129
197
  ------
130
198
  dict
131
- Keys containing the orbit state and values containing the total area of intersection in
132
- km^2
199
+ Keys containing the orbit state and values containing the total area of intersection and
200
+ in km^2 and maximum temporal gap in days.
133
201
  """
134
- if isinstance(spatial_extent, GeoJSON):
202
+ if isinstance(spatial_extent, geojson.FeatureCollection):
135
203
  # Transform geojson into shapely geometry and compute bounds
136
- bounds = shape(spatial_extent).bounds
137
- epsg = 4362
204
+ shapely_geometries = [
205
+ shape(feature["geometry"]) for feature in spatial_extent["features"]
206
+ ]
207
+ geometry = unary_union(shapely_geometries)
208
+ bounds = geometry.bounds
209
+ epsg = 4326
138
210
  elif isinstance(spatial_extent, BoundingBoxExtent):
139
211
  bounds = [
140
212
  spatial_extent.west,
@@ -153,17 +225,22 @@ def s1_area_per_orbitstate(
153
225
 
154
226
  # Queries the products in the catalogues
155
227
  if backend.backend in [Backend.CDSE, Backend.CDSE_STAGING, Backend.FED]:
156
- ascending_products = _parse_cdse_products(
228
+ ascending_products, ascending_timestamps = _parse_cdse_products(
157
229
  _query_cdse_catalogue(
158
- "Sentinel1", bounds, temporal_extent, orbitDirection="ASCENDING"
230
+ "Sentinel1",
231
+ bounds,
232
+ temporal_extent,
233
+ orbitDirection="ASCENDING",
234
+ polarisation="VV%26VH",
159
235
  )
160
236
  )
161
- descending_products = _parse_cdse_products(
237
+ descending_products, descending_timestamps = _parse_cdse_products(
162
238
  _query_cdse_catalogue(
163
239
  "Sentinel1",
164
240
  bounds,
165
241
  temporal_extent,
166
242
  orbitDirection="DESCENDING",
243
+ polarisation="VV%26VH",
167
244
  )
168
245
  )
169
246
  else:
@@ -185,6 +262,9 @@ def s1_area_per_orbitstate(
185
262
  return {
186
263
  "ASCENDING": {
187
264
  "full_overlap": ascending_covers,
265
+ "max_temporal_gap": _compute_max_gap_days(
266
+ temporal_extent, ascending_timestamps
267
+ ),
188
268
  "area": sum(
189
269
  product.intersection(spatial_extent).area
190
270
  for product in ascending_products
@@ -192,6 +272,9 @@ def s1_area_per_orbitstate(
192
272
  },
193
273
  "DESCENDING": {
194
274
  "full_overlap": descending_covers,
275
+ "max_temporal_gap": _compute_max_gap_days(
276
+ temporal_extent, descending_timestamps
277
+ ),
195
278
  "area": sum(
196
279
  product.intersection(spatial_extent).area
197
280
  for product in descending_products
@@ -200,22 +283,31 @@ def s1_area_per_orbitstate(
200
283
  }
201
284
 
202
285
 
203
- def select_S1_orbitstate(
286
+ def select_s1_orbitstate_vvvh(
204
287
  backend: BackendContext,
205
288
  spatial_extent: SpatialContext,
206
289
  temporal_extent: TemporalContext,
290
+ max_temporal_gap: int = 60,
207
291
  ) -> str:
208
- """Selects the orbit state that covers the most area of the given spatio-temporal context
209
- for the Sentinel-1 collection.
292
+ """Selects the orbit state based on some predefined rules that
293
+ are checked in sequential order:
294
+ 1. prefer an orbit with full coverage over the requested bounds
295
+ 2. prefer an orbit with a maximum temporal gap under a
296
+ predefined threshold
297
+ 3. prefer the orbit that covers the most area of intersection
298
+ for the available products
210
299
 
211
300
  Parameters
212
301
  ----------
213
302
  backend : BackendContext
214
- The backend to be within, as each backend might use different catalogues.
303
+ The backend to be within, as each backend might use different catalogues. Only the CDSE,
304
+ CDSE_STAGING and FED backends are supported.
215
305
  spatial_extent : SpatialContext
216
306
  The spatial extent to be checked, it will check within its bounding box.
217
307
  temporal_extent : TemporalContext
218
308
  The temporal period to be checked.
309
+ max_temporal_gap: int, optional, default: 30
310
+ The maximum temporal gap in days to be considered for the orbit state.
219
311
 
220
312
  Returns
221
313
  ------
@@ -224,25 +316,64 @@ def select_S1_orbitstate(
224
316
  """
225
317
 
226
318
  # Queries the products in the catalogues
227
- areas = s1_area_per_orbitstate(backend, spatial_extent, temporal_extent)
319
+ areas = s1_area_per_orbitstate_vvvh(backend, spatial_extent, temporal_extent)
228
320
 
229
321
  ascending_overlap = areas["ASCENDING"]["full_overlap"]
230
322
  descending_overlap = areas["DESCENDING"]["full_overlap"]
323
+ ascending_gap_too_large = areas["ASCENDING"]["max_temporal_gap"] > max_temporal_gap
324
+ descending_gap_too_large = (
325
+ areas["DESCENDING"]["max_temporal_gap"] > max_temporal_gap
326
+ )
327
+
328
+ orbit_choice = None
329
+
330
+ if not ascending_overlap and not descending_overlap:
331
+ raise UncoveredS1Exception(
332
+ "No product available to fully cover the requested area in both orbit states."
333
+ )
231
334
 
335
+ # Rule 1: Prefer an orbit with full coverage over the requested bounds
232
336
  if ascending_overlap and not descending_overlap:
233
- return "ASCENDING"
337
+ orbit_choice = "ASCENDING"
338
+ reason = "Only orbit fully covering the requested area."
234
339
  elif descending_overlap and not ascending_overlap:
235
- return "DESCENDING"
340
+ orbit_choice = "DESCENDING"
341
+ reason = "Only orbit fully covering the requested area."
342
+
343
+ # Rule 2: Prefer an orbit with a maximum temporal gap under a predefined threshold
344
+ elif ascending_gap_too_large and not descending_gap_too_large:
345
+ orbit_choice = "DESCENDING"
346
+ reason = (
347
+ "Only orbit with temporal gap under the threshold. "
348
+ f"{areas['DESCENDING']['max_temporal_gap']} days < {max_temporal_gap} days"
349
+ )
350
+ elif descending_gap_too_large and not ascending_gap_too_large:
351
+ orbit_choice = "ASCENDING"
352
+ reason = (
353
+ "Only orbit with temporal gap under the threshold. "
354
+ f"{areas['ASCENDING']['max_temporal_gap']} days < {max_temporal_gap} days"
355
+ )
356
+ # Rule 3: Prefer the orbit that covers the most area of intersection
357
+ # for the available products
236
358
  elif ascending_overlap and descending_overlap:
237
359
  ascending_cover_area = areas["ASCENDING"]["area"]
238
360
  descending_cover_area = areas["DESCENDING"]["area"]
239
361
 
240
362
  # Selects the orbit state that covers the most area
241
363
  if ascending_cover_area > descending_cover_area:
242
- return "ASCENDING"
364
+ orbit_choice = "ASCENDING"
365
+ reason = (
366
+ "Orbit has more cumulative intersected area. "
367
+ f"{ascending_cover_area} > {descending_cover_area}"
368
+ )
243
369
  else:
244
- return "DESCENDING"
370
+ reason = (
371
+ "Orbit has more cumulative intersected area. "
372
+ f"{descending_cover_area} > {ascending_cover_area}"
373
+ )
374
+ orbit_choice = "DESCENDING"
245
375
 
246
- raise UncoveredS1Exception(
247
- "No product available to fully cover the given spatio-temporal context."
248
- )
376
+ if orbit_choice is not None:
377
+ _log.info(f"Selected orbit state: {orbit_choice}. Reason: {reason}")
378
+ return orbit_choice
379
+ raise UncoveredS1Exception("Failed to select suitable Sentinel-1 orbit.")
@@ -0,0 +1,125 @@
1
+ """Utility function to split a STAC collection into multiple STAC collections based on CRS.
2
+ Requires the "proj:epsg" property to be present in all the STAC items.
3
+ """
4
+
5
+ import os
6
+ from pathlib import Path
7
+ from typing import Iterator, Union
8
+
9
+ import pystac
10
+
11
+
12
+ def _extract_epsg_from_stac_item(stac_item: pystac.Item) -> int:
13
+ """
14
+ Extract the EPSG code from a STAC item.
15
+
16
+ Parameters:
17
+ stac_item (pystac.Item): The STAC item.
18
+
19
+ Returns:
20
+ int: The EPSG code.
21
+
22
+ Raises:
23
+ KeyError: If the "proj:epsg" property is missing from the STAC item.
24
+ """
25
+
26
+ try:
27
+ epsg_code = stac_item.properties["proj:epsg"]
28
+ return epsg_code
29
+ except KeyError:
30
+ raise KeyError("The 'proj:epsg' property is missing from the STAC item.")
31
+
32
+
33
+ def _get_items_by_epsg(
34
+ collection: pystac.Collection,
35
+ ) -> Iterator[tuple[int, pystac.Item]]:
36
+ """
37
+ Generator function that yields items grouped by their EPSG code.
38
+
39
+ Parameters:
40
+ collection (pystac.Collection): The STAC collection.
41
+
42
+ Yields:
43
+ tuple[int, pystac.Item]: EPSG code and corresponding STAC item.
44
+ """
45
+ for item in collection.get_all_items():
46
+ epsg = _extract_epsg_from_stac_item(item)
47
+ yield epsg, item
48
+
49
+
50
+ def _create_collection_skeleton(
51
+ collection: pystac.Collection, epsg: int
52
+ ) -> pystac.Collection:
53
+ """
54
+ Create a skeleton for a new STAC collection with a given EPSG code.
55
+
56
+ Parameters:
57
+ collection (pystac.Collection): The original STAC collection.
58
+ epsg (int): The EPSG code.
59
+
60
+ Returns:
61
+ pystac.Collection: The skeleton of the new STAC collection.
62
+ """
63
+ new_collection = pystac.Collection(
64
+ id=f"{collection.id}_{epsg}",
65
+ description=f"{collection.description} Containing only items with EPSG code {epsg}",
66
+ extent=collection.extent.clone(),
67
+ summaries=collection.summaries,
68
+ license=collection.license,
69
+ stac_extensions=collection.stac_extensions,
70
+ )
71
+ if "item_assets" in collection.extra_fields:
72
+ item_assets_extension = pystac.extensions.item_assets.ItemAssetsExtension.ext(
73
+ collection
74
+ )
75
+
76
+ new_item_assets_extension = (
77
+ pystac.extensions.item_assets.ItemAssetsExtension.ext(
78
+ new_collection, add_if_missing=True
79
+ )
80
+ )
81
+
82
+ new_item_assets_extension.item_assets = item_assets_extension.item_assets
83
+ return new_collection
84
+
85
+
86
+ def split_collection_by_epsg(
87
+ collection: Union[str, Path, pystac.Collection], output_dir: Union[str, Path]
88
+ ):
89
+ """
90
+ Split a STAC collection into multiple STAC collections based on EPSG code.
91
+
92
+ Parameters
93
+ ----------
94
+ collection: Union[str, Path, pystac.Collection]
95
+ A collection of STAC items or a path to a STAC collection.
96
+ output_dir: Union[str, Path]
97
+ The directory where the split STAC collections will be saved.
98
+ """
99
+
100
+ if not isinstance(collection, pystac.Collection):
101
+ collection = Path(collection)
102
+ output_dir = Path(output_dir)
103
+ os.makedirs(output_dir, exist_ok=True)
104
+
105
+ try:
106
+ collection = pystac.read_file(collection)
107
+ except pystac.STACError:
108
+ print("Please provide a path to a valid STAC collection.")
109
+ return
110
+
111
+ collections_by_epsg = {}
112
+
113
+ for epsg, item in _get_items_by_epsg(collection):
114
+ if epsg not in collections_by_epsg:
115
+ collections_by_epsg[epsg] = _create_collection_skeleton(collection, epsg)
116
+
117
+ # Add item to the corresponding collection
118
+ collections_by_epsg[epsg].add_item(item)
119
+
120
+ # Write each collection to disk
121
+ for epsg, new_collection in collections_by_epsg.items():
122
+ new_collection.update_extent_from_items() # Update extent based on added items
123
+ collection_path = output_dir / f"collection-{epsg}"
124
+ new_collection.normalize_hrefs(str(collection_path))
125
+ new_collection.save()
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.3
2
2
  Name: openeo_gfmap
3
- Version: 0.1.0
3
+ Version: 0.2.0
4
4
  Summary: OpenEO General Framework for Mapping
5
5
  Project-URL: Homepage, https://github.com/Open-EO/openeo-gfmap
6
6
  Project-URL: Bug Tracker, https://github.com/Open-EO/openeo-gfmap/issues
@@ -4,19 +4,18 @@ openeo_gfmap/metadata.py,sha256=WlJ3zy78ff1E65HDbw7kOuYamqyvfV0BCNrv_ZvXX4Y,762
4
4
  openeo_gfmap/spatial.py,sha256=y1Gk9yM1j2eod127Pthn7SKxY37EFbzvPX0znqHgnMw,1353
5
5
  openeo_gfmap/temporal.py,sha256=qhKCgSoOc0CrscPTru-d7acaxsCVhftyYrb_8UVU1S4,583
6
6
  openeo_gfmap/features/__init__.py,sha256=UtQUPnglF9uURn9FqtegnazuE4_CgQP6a2Cdx1TOuZ0,419
7
- openeo_gfmap/features/feature_extractor.py,sha256=kMXIkb7OqEBDdZGhTZp0gFDtq9Ip_IVwszCoE6HOqWY,14491
8
- openeo_gfmap/fetching/__init__.py,sha256=vuG5qdBEY6TytLjH2NWjyg_2ivHBEX9zvXOwD-MrxzQ,521
9
- openeo_gfmap/fetching/commons.py,sha256=6LBQry1V18jSivhd7pO5XHLzeb0msC85frHThvVnH0E,7636
7
+ openeo_gfmap/features/feature_extractor.py,sha256=ApTCCbD1-S4VOOSEZh9i-Gqxso87xwe_z7CN35fP15A,14719
8
+ openeo_gfmap/fetching/__init__.py,sha256=KFDwqKTwXUDhFqPqeaIg5uCL2xp7lXmNzcbAph-EU1c,938
9
+ openeo_gfmap/fetching/commons.py,sha256=7kr34Lq3ZxKWQXJrtAHFJYkc93y6LHGfTe1jn7Adr64,7656
10
10
  openeo_gfmap/fetching/fetching.py,sha256=dHeOMzN6OPgD8YFfZtcCzEOwQqo47IeBgIdS2mrx3MY,3674
11
- openeo_gfmap/fetching/generic.py,sha256=QekJd2-BaEn_ObEaKEwpYTP7jkH5Enax7c7CgdpcRnY,5461
12
- openeo_gfmap/fetching/meteo.py,sha256=QVuyjPjzUzJyiYL3sxwSGA8bTgaSGI3ITkKkcW9WdzQ,3386
13
- openeo_gfmap/fetching/s1.py,sha256=bewQOzP5ClDhVzaKINAY3-IKPiBW7TCuVtUs_BRgSvA,6452
14
- openeo_gfmap/fetching/s2.py,sha256=Ulmjn57pCA2GyKV-OHnbTaLenMUK3_QpY06fDw3STaU,7699
11
+ openeo_gfmap/fetching/generic.py,sha256=ojSO52cnLsWpC6FAnLRoXxfQmTiC839DzFH8MAok2B8,5851
12
+ openeo_gfmap/fetching/s1.py,sha256=Ek9Ek-GExyKdb-9Ijja6I-izOmVvPY2C9w9gyyGGjII,6360
13
+ openeo_gfmap/fetching/s2.py,sha256=ytjrZiZIwXxrdiky2V0bAKLBU9Dpaa5b2XsHvI6jl1M,7718
15
14
  openeo_gfmap/inference/__init__.py,sha256=M6NnKGYCpHNYmRL9OkHi5GmfCtWoJ0wCNR6VXRuDgjE,165
16
- openeo_gfmap/inference/model_inference.py,sha256=-Pf0PHfcO8Y0ok9lsx4Vp2qee19RhrkzHVLVVKTRkVE,12478
15
+ openeo_gfmap/inference/model_inference.py,sha256=0qPUgrjI1hy5ZnyGwuuvvw5oxnMGdgvvu9Go6-e9LZQ,12550
17
16
  openeo_gfmap/manager/__init__.py,sha256=2bckkPiDQBgoBWD9spk1BKXy2UGkWKe50A3HmIwmqrA,795
18
- openeo_gfmap/manager/job_manager.py,sha256=PIsSBDNURwrunVTjqsxfKEFxi0xlBflQH8sC0j29Zoc,20056
19
- openeo_gfmap/manager/job_splitters.py,sha256=zV0Tr3aEa4WORJgY8Z-rRHqsosj-1pWpPsyBSUnuNwo,5123
17
+ openeo_gfmap/manager/job_manager.py,sha256=7NwpU6kvrtqtvhtDjymbR4AKySp8E2KyCfBJFqrd2Ns,27165
18
+ openeo_gfmap/manager/job_splitters.py,sha256=PD4DsZr34MEMMVMHMUiKyFvdgKYyQpL-Lohpak8g_vU,5947
20
19
  openeo_gfmap/preprocessing/__init__.py,sha256=-kJAy_WY4o8oqziRozcUuXtuGIM0IOvTCF6agTUgRWA,619
21
20
  openeo_gfmap/preprocessing/cloudmasking.py,sha256=d280H5fByjNbCVZHjPn_dUatNI-ejphu4A75sUVoRqo,10029
22
21
  openeo_gfmap/preprocessing/compositing.py,sha256=Jp9Ku5JpU7TJ4DYGc6YuqMeP1Ip7zns7NguC17BtFyA,2526
@@ -27,14 +26,15 @@ openeo_gfmap/preprocessing/udf_cldmask.py,sha256=WqqFLBK5rIQPkb_dlgUWWSzicsPtVSt
27
26
  openeo_gfmap/preprocessing/udf_rank.py,sha256=n2gSIY2ZHVVr9wJx1Bs2HtmvScAkz2NqhjxUM-iIKM0,1438
28
27
  openeo_gfmap/preprocessing/udf_score.py,sha256=L1d5do1lIRJFLWqbuSbXPdwR-hPoSZqVE8ffVtG5kI0,3330
29
28
  openeo_gfmap/stac/__init__.py,sha256=kVMJ9hrN4MjcRCOgRDCj5TfAWRXe0GHu2gJQjG-dS4Y,59
30
- openeo_gfmap/stac/constants.py,sha256=Oelktfzq3sE1leMI6_EC_Dt8yMvPA_8I2AeJSKqGxII,1485
31
- openeo_gfmap/utils/__init__.py,sha256=5UUoVq6PaNXar_1J82u-_6KInfuY7uNU4_u2E_Cj_HA,626
29
+ openeo_gfmap/stac/constants.py,sha256=O1bcijRBj6YRqR_aAcYO5JzJg7mdzhzUSm4vKnxMbtQ,1485
30
+ openeo_gfmap/utils/__init__.py,sha256=UDwkWUwsnV6ZLXeaJKOCos-MDG2ZaIFyg8s0IiRVtng,997
32
31
  openeo_gfmap/utils/build_df.py,sha256=OPmD_Onkl9ybYIiLxmU_GmanP8xD71F1ZybJc7xQmns,1515
33
- openeo_gfmap/utils/catalogue.py,sha256=-QjoxFZ9RIigPjXOEE-LQcLW88aoTxfI3oC0SYR3n68,8073
32
+ openeo_gfmap/utils/catalogue.py,sha256=-dG07sJ6mXOk2RguiOPcrWdCo5z8BhNmXN0ae7j3kr0,13348
34
33
  openeo_gfmap/utils/intervals.py,sha256=V6l3ofww50fN_pvWC4NuGQ2ZsyGdhAlTZTiRcC0foVE,2395
35
34
  openeo_gfmap/utils/netcdf.py,sha256=KkAAxnq-ZCMjDMd82638noYwxqNpMsnpiU04Q-qX26A,698
35
+ openeo_gfmap/utils/split_stac.py,sha256=asjT0jx6ic8GJFqqAisaWxOvQ_suSRv4sxyFOyHFvpI,3895
36
36
  openeo_gfmap/utils/tile_processing.py,sha256=QZ9bi5tPmyTVyyNvFZgd26s5dSnMl1grTKq2veK1C90,2068
37
- openeo_gfmap-0.1.0.dist-info/METADATA,sha256=BAThgJ7FwpCz2QLFWYkFiPkbg9UPrUDRGgAXjkq1iPc,4278
38
- openeo_gfmap-0.1.0.dist-info/WHEEL,sha256=zEMcRr9Kr03x1ozGwg5v9NQBKn3kndp6LSoSlVg-jhU,87
39
- openeo_gfmap-0.1.0.dist-info/licenses/LICENSE,sha256=aUuGpjieWiscTNtyLcSaeVsJ4pb6J9c4wUq1bR0e4t4,11349
40
- openeo_gfmap-0.1.0.dist-info/RECORD,,
37
+ openeo_gfmap-0.2.0.dist-info/METADATA,sha256=aSf8KibRnGJl6B8lRH1nT1RYWpR9TnMaKHy6Y3qIE-s,4278
38
+ openeo_gfmap-0.2.0.dist-info/WHEEL,sha256=1yFddiXMmvYK7QYTqtRNtX66WJ0Mz8PYEiEUoOUUxRY,87
39
+ openeo_gfmap-0.2.0.dist-info/licenses/LICENSE,sha256=aUuGpjieWiscTNtyLcSaeVsJ4pb6J9c4wUq1bR0e4t4,11349
40
+ openeo_gfmap-0.2.0.dist-info/RECORD,,
@@ -1,4 +1,4 @@
1
1
  Wheel-Version: 1.0
2
- Generator: hatchling 1.24.2
2
+ Generator: hatchling 1.25.0
3
3
  Root-Is-Purelib: true
4
4
  Tag: py3-none-any