ras-commander 0.48.0__py3-none-any.whl → 0.49.0__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -16,7 +16,6 @@ def log_call(func):
16
16
  return result
17
17
  return wrapper
18
18
 
19
-
20
19
  def standardize_input(file_type: str = 'plan_hdf'):
21
20
  """
22
21
  Decorator to standardize input for HDF file operations.
@@ -41,6 +40,14 @@ def standardize_input(file_type: str = 'plan_hdf'):
41
40
  def wrapper(*args, **kwargs):
42
41
  logger = logging.getLogger(func.__module__)
43
42
 
43
+ # Check if the function expects an hdf_path parameter
44
+ sig = inspect.signature(func)
45
+ param_names = list(sig.parameters.keys())
46
+
47
+ # If first parameter is 'hdf_file', skip path processing
48
+ if param_names and param_names[0] == 'hdf_file':
49
+ return func(*args, **kwargs)
50
+
44
51
  # Handle both static method calls and regular function calls
45
52
  if args and isinstance(args[0], type):
46
53
  # Static method call, remove the class argument
@@ -49,6 +56,16 @@ def standardize_input(file_type: str = 'plan_hdf'):
49
56
  hdf_input = kwargs.pop('hdf_path', None) or kwargs.pop('hdf_input', None) or (args[0] if args else None)
50
57
  ras_object = kwargs.pop('ras_object', None) or (args[1] if len(args) > 1 else None)
51
58
 
59
+ # If no hdf_input provided, return the function unmodified
60
+ if hdf_input is None:
61
+ return func(*args, **kwargs)
62
+
63
+ # NEW: If input is already a Path and exists, use it directly regardless of file_type
64
+ if isinstance(hdf_input, Path) and hdf_input.is_file():
65
+ logger.info(f"Using existing HDF file: {hdf_input}")
66
+ new_args = (hdf_input,) + args[1:]
67
+ return func(*new_args, **kwargs)
68
+
52
69
  hdf_path = None
53
70
 
54
71
  # If hdf_input is already an h5py.File object, use its filename
ras_commander/HdfBase.py CHANGED
@@ -1,197 +1,307 @@
1
- """
2
- Class: HdfBase
3
-
4
- Attribution: A substantial amount of code in this file is sourced or derived
5
- from the https://github.com/fema-ffrd/rashdf library,
6
- released under MIT license and Copyright (c) 2024 fema-ffrd
7
-
8
- The file has been forked and modified for use in RAS Commander.
9
- """
10
- import re
11
- from datetime import datetime, timedelta
12
- import h5py
13
- import numpy as np
14
- import pandas as pd
15
- import xarray as xr # Added import for xarray
16
- from typing import List, Tuple, Union, Optional, Dict
17
- from pathlib import Path
18
- import logging
19
-
20
- from .HdfUtils import HdfUtils
21
- from .Decorators import standardize_input, log_call
22
- from .LoggingConfig import setup_logging, get_logger
23
-
24
- logger = get_logger(__name__)
25
-
26
- class HdfBase:
27
- """
28
- Base class for HEC-RAS HDF file operations.
29
-
30
- This class provides fundamental methods for interacting with HEC-RAS HDF files,
31
- including time-related operations and mesh data retrieval. It serves as a foundation
32
- for more specialized HDF classes.
33
-
34
- The methods in this class are designed to work with both plan and geometry HDF files,
35
- providing low-level access to file structure and content.
36
-
37
- Note:
38
- - All methods in this class are static, allowing for use without instantiation.
39
- - This class is not meant to be used directly in most cases, but rather as a base
40
- for more specialized HDF classes.
41
- """
42
-
43
- @staticmethod
44
- def _get_simulation_start_time(hdf_file: h5py.File) -> datetime:
45
- """
46
- Get the simulation start time from the HDF file.
47
-
48
- Args:
49
- hdf_file (h5py.File): Open HDF file object.
50
-
51
- Returns:
52
- datetime: The simulation start time.
53
-
54
- Raises:
55
- ValueError: If Plan Information is not found in the HDF file.
56
- """
57
- plan_info = hdf_file.get("Plan Data/Plan Information")
58
- if plan_info is None:
59
- raise ValueError("Plan Information not found in HDF file")
60
- time_str = plan_info.attrs.get('Simulation Start Time')
61
- return datetime.strptime(time_str.decode('utf-8'), "%d%b%Y %H:%M:%S")
62
-
63
- @staticmethod
64
- def _get_unsteady_datetimes(hdf_file: h5py.File) -> List[datetime]:
65
- """
66
- Get the list of unsteady datetimes from the HDF file.
67
-
68
- Args:
69
- hdf_file (h5py.File): Open HDF file object.
70
-
71
- Returns:
72
- List[datetime]: A list of datetime objects representing the unsteady timestamps.
73
- """
74
- group_path = "Results/Unsteady/Output/Output Blocks/Base Output/Unsteady Time Series/Time Date Stamp (ms)"
75
- raw_datetimes = hdf_file[group_path][:]
76
- return [HdfBase._parse_ras_datetime_ms(x.decode("utf-8")) for x in raw_datetimes]
77
-
78
-
79
- @staticmethod
80
- def _get_2d_flow_area_names_and_counts(hdf_file: h5py.File) -> List[Tuple[str, int]]:
81
- """
82
- Get the names and cell counts of 2D flow areas from the HDF file.
83
-
84
- Args:
85
- hdf_file (h5py.File): Open HDF file object.
86
-
87
- Returns:
88
- List[Tuple[str, int]]: A list of tuples containing the name and cell count of each 2D flow area.
89
- """
90
- d2_flow_areas = hdf_file.get("Geometry/2D Flow Areas/Attributes")
91
- if d2_flow_areas is None:
92
- return []
93
- return [(HdfBase._convert_ras_hdf_string(d2_flow_area[0]), d2_flow_area[-1]) for d2_flow_area in d2_flow_areas[:]]
94
-
95
- @staticmethod
96
- def _parse_ras_datetime(datetime_str: str) -> datetime:
97
- """
98
- Parse a datetime string from a RAS file into a datetime object.
99
-
100
- Args:
101
- datetime_str (str): The datetime string to parse.
102
-
103
- Returns:
104
- datetime: The parsed datetime object.
105
- """
106
- return datetime.strptime(datetime_str, "%d%b%Y %H:%M:%S")
107
-
108
- @staticmethod
109
- def _parse_ras_simulation_window_datetime(datetime_str: str) -> datetime:
110
- """
111
- Parse a datetime string from a RAS simulation window into a datetime object.
112
-
113
- Args:
114
- datetime_str (str): The datetime string to parse.
115
-
116
- Returns:
117
- datetime: The parsed datetime object.
118
- """
119
- return datetime.strptime(datetime_str, "%d%b%Y %H%M")
120
-
121
- @staticmethod
122
- def _parse_duration(duration_str: str) -> timedelta:
123
- """
124
- Parse a duration string into a timedelta object.
125
-
126
- Args:
127
- duration_str (str): The duration string to parse.
128
-
129
- Returns:
130
- timedelta: The parsed duration as a timedelta object.
131
- """
132
- hours, minutes, seconds = map(int, duration_str.split(':'))
133
- return timedelta(hours=hours, minutes=minutes, seconds=seconds)
134
-
135
- @staticmethod
136
- def _parse_ras_datetime_ms(datetime_str: str) -> datetime:
137
- """
138
- Parse a datetime string with milliseconds from a RAS file.
139
-
140
- Args:
141
- datetime_str (str): The datetime string to parse.
142
-
143
- Returns:
144
- datetime: The parsed datetime object.
145
- """
146
- milliseconds = int(datetime_str[-3:])
147
- microseconds = milliseconds * 1000
148
- parsed_dt = HdfBase._parse_ras_datetime(datetime_str[:-4]).replace(microsecond=microseconds)
149
- return parsed_dt
150
-
151
- @staticmethod
152
- def _convert_ras_hdf_string(value: Union[str, bytes]) -> Union[bool, datetime, List[datetime], timedelta, str]:
153
- """
154
- Convert a string value from an HEC-RAS HDF file into a Python object.
155
-
156
- Args:
157
- value (Union[str, bytes]): The value to convert.
158
-
159
- Returns:
160
- Union[bool, datetime, List[datetime], timedelta, str]: The converted value.
161
- """
162
- if isinstance(value, bytes):
163
- s = value.decode("utf-8")
164
- else:
165
- s = value
166
-
167
- if s == "True":
168
- return True
169
- elif s == "False":
170
- return False
171
-
172
- ras_datetime_format1_re = r"\d{2}\w{3}\d{4} \d{2}:\d{2}:\d{2}"
173
- ras_datetime_format2_re = r"\d{2}\w{3}\d{4} \d{2}\d{2}"
174
- ras_duration_format_re = r"\d{2}:\d{2}:\d{2}"
175
-
176
- if re.match(rf"^{ras_datetime_format1_re}", s):
177
- if re.match(rf"^{ras_datetime_format1_re} to {ras_datetime_format1_re}$", s):
178
- split = s.split(" to ")
179
- return [
180
- HdfBase._parse_ras_datetime(split[0]),
181
- HdfBase._parse_ras_datetime(split[1]),
182
- ]
183
- return HdfBase._parse_ras_datetime(s)
184
- elif re.match(rf"^{ras_datetime_format2_re}", s):
185
- if re.match(rf"^{ras_datetime_format2_re} to {ras_datetime_format2_re}$", s):
186
- split = s.split(" to ")
187
- return [
188
- HdfBase._parse_ras_simulation_window_datetime(split[0]),
189
- HdfBase._parse_ras_simulation_window_datetime(split[1]),
190
- ]
191
- return HdfBase._parse_ras_simulation_window_datetime(s)
192
- elif re.match(rf"^{ras_duration_format_re}$", s):
193
- return HdfBase._parse_duration(s)
194
- return s
195
-
196
-
197
-
1
+ """
2
+ HdfBase: Core HDF File Operations for HEC-RAS
3
+
4
+ This module provides fundamental methods for interacting with HEC-RAS HDF files.
5
+ It serves as a foundation for more specialized HDF classes.
6
+
7
+ Attribution:
8
+ Derived from the rashdf library (https://github.com/fema-ffrd/rashdf)
9
+ Copyright (c) 2024 fema-ffrd - MIT License
10
+
11
+ Features:
12
+ - Time parsing and conversion utilities
13
+ - HDF attribute and dataset access
14
+ - Geometric data extraction
15
+ - 2D flow area information retrieval
16
+
17
+ Classes:
18
+ HdfBase: Base class containing static methods for HDF operations
19
+
20
+ Key Methods:
21
+ Time Operations:
22
+ - get_simulation_start_time(): Get simulation start datetime
23
+ - get_unsteady_timestamps(): Get unsteady output timestamps
24
+ - parse_ras_datetime(): Parse RAS datetime strings
25
+
26
+ Data Access:
27
+ - get_2d_flow_area_names_and_counts(): Get 2D flow area info
28
+ - get_projection(): Get spatial projection
29
+ - get_attrs(): Access HDF attributes
30
+ - get_dataset_info(): Explore HDF structure
31
+ - get_polylines_from_parts(): Extract geometric polylines
32
+
33
+ Example:
34
+ ```python
35
+ from ras_commander import HdfBase
36
+
37
+ with h5py.File('model.hdf', 'r') as hdf:
38
+ start_time = HdfBase.get_simulation_start_time(hdf)
39
+ timestamps = HdfBase.get_unsteady_timestamps(hdf)
40
+ ```
41
+ """
42
+ import re
43
+ from datetime import datetime, timedelta
44
+ import h5py
45
+ import numpy as np
46
+ import pandas as pd
47
+ import xarray as xr
48
+ from typing import List, Tuple, Union, Optional, Dict, Any
49
+ from pathlib import Path
50
+ import logging
51
+ from shapely.geometry import LineString, MultiLineString
52
+
53
+ from .HdfUtils import HdfUtils
54
+ from .Decorators import standardize_input, log_call
55
+ from .LoggingConfig import setup_logging, get_logger
56
+
57
+ logger = get_logger(__name__)
58
+
59
+ class HdfBase:
60
+ """
61
+ Base class for HEC-RAS HDF file operations.
62
+
63
+ This class provides static methods for fundamental HDF file operations,
64
+ including time parsing, attribute access, and geometric data extraction.
65
+ All methods are designed to work with h5py.File objects or pathlib.Path
66
+ inputs.
67
+
68
+ Note:
69
+ This class is not meant to be instantiated. All methods are static
70
+ and should be called directly from the class.
71
+ """
72
+
73
+ @staticmethod
74
+ def get_simulation_start_time(hdf_file: h5py.File) -> datetime:
75
+ """
76
+ Extract the simulation start time from the HDF file.
77
+
78
+ Args:
79
+ hdf_file: Open HDF file object containing RAS simulation data.
80
+
81
+ Returns:
82
+ datetime: Simulation start time as a datetime object.
83
+
84
+ Raises:
85
+ ValueError: If Plan Information is not found or start time cannot be parsed.
86
+
87
+ Note:
88
+ Expects 'Plan Data/Plan Information' group with 'Simulation Start Time' attribute.
89
+ """
90
+ plan_info = hdf_file.get("Plan Data/Plan Information")
91
+ if plan_info is None:
92
+ raise ValueError("Plan Information not found in HDF file")
93
+ time_str = plan_info.attrs.get('Simulation Start Time')
94
+ return HdfUtils.parse_ras_datetime(time_str.decode('utf-8'))
95
+
96
+ @staticmethod
97
+ def get_unsteady_timestamps(hdf_file: h5py.File) -> List[datetime]:
98
+ """
99
+ Extract the list of unsteady timestamps from the HDF file.
100
+
101
+ Args:
102
+ hdf_file (h5py.File): Open HDF file object.
103
+
104
+ Returns:
105
+ List[datetime]: A list of datetime objects representing the unsteady timestamps.
106
+ """
107
+ group_path = "Results/Unsteady/Output/Output Blocks/Base Output/Unsteady Time Series/Time Date Stamp (ms)"
108
+ raw_datetimes = hdf_file[group_path][:]
109
+ return [HdfUtils.parse_ras_datetime_ms(x.decode("utf-8")) for x in raw_datetimes]
110
+
111
+ @staticmethod
112
+ @standardize_input(file_type='plan_hdf')
113
+ def get_2d_flow_area_names_and_counts(hdf_path: Path) -> List[Tuple[str, int]]:
114
+ """
115
+ Get the names and cell counts of 2D flow areas from the HDF file.
116
+
117
+ Args:
118
+ hdf_path (Path): Path to the HDF file.
119
+
120
+ Returns:
121
+ List[Tuple[str, int]]: A list of tuples containing the name and cell count of each 2D flow area.
122
+
123
+ Raises:
124
+ ValueError: If there's an error reading the HDF file or accessing the required data.
125
+ """
126
+ try:
127
+ with h5py.File(hdf_path, 'r') as hdf_file:
128
+ flow_area_2d_path = "Geometry/2D Flow Areas"
129
+ if flow_area_2d_path not in hdf_file:
130
+ return []
131
+
132
+ attributes = hdf_file[f"{flow_area_2d_path}/Attributes"][()]
133
+ names = [HdfUtils.convert_ras_string(name) for name in attributes["Name"]]
134
+
135
+ cell_info = hdf_file[f"{flow_area_2d_path}/Cell Info"][()]
136
+ cell_counts = [info[1] for info in cell_info]
137
+
138
+ return list(zip(names, cell_counts))
139
+ except Exception as e:
140
+ logger.error(f"Error reading 2D flow area names and counts from {hdf_path}: {str(e)}")
141
+ raise ValueError(f"Failed to get 2D flow area names and counts: {str(e)}")
142
+
143
+ @staticmethod
144
+ @standardize_input(file_type='plan_hdf')
145
+ def get_projection(hdf_path: Path) -> Optional[str]:
146
+ """
147
+ Get the projection information from the HDF file.
148
+
149
+ Args:
150
+ hdf_path (Path): Path to the HDF file.
151
+
152
+ Returns:
153
+ Optional[str]: The projection information as a string, or None if not found.
154
+ """
155
+ try:
156
+ with h5py.File(hdf_path, 'r') as hdf_file:
157
+ proj_wkt = hdf_file.attrs.get("Projection")
158
+ if proj_wkt is None:
159
+ return None
160
+ if isinstance(proj_wkt, bytes) or isinstance(proj_wkt, np.bytes_):
161
+ proj_wkt = proj_wkt.decode("utf-8")
162
+ return proj_wkt
163
+ except Exception as e:
164
+ logger.error(f"Error reading projection from {hdf_path}: {str(e)}")
165
+ return None
166
+
167
+ @staticmethod
168
+ @standardize_input(file_type='plan_hdf')
169
+ def get_attrs(hdf_file: h5py.File, attr_path: str) -> Dict[str, Any]:
170
+ """
171
+ Get attributes from an HDF file at a specified path.
172
+
173
+ Args:
174
+ hdf_file (h5py.File): The opened HDF file.
175
+ attr_path (str): Path to the attributes in the HDF file.
176
+
177
+ Returns:
178
+ Dict[str, Any]: Dictionary of attributes.
179
+ """
180
+ try:
181
+ if attr_path not in hdf_file:
182
+ logger.warning(f"Path {attr_path} not found in HDF file")
183
+ return {}
184
+
185
+ return HdfUtils.hdf5_attrs_to_dict(hdf_file[attr_path].attrs)
186
+ except Exception as e:
187
+ logger.error(f"Error getting attributes from {attr_path}: {str(e)}")
188
+ return {}
189
+
190
+ @staticmethod
191
+ @standardize_input(file_type='plan_hdf')
192
+ def get_dataset_info(file_path: Path, group_path: str = '/') -> None:
193
+ """
194
+ Recursively explore and print the structure of an HDF5 file.
195
+
196
+ Displays detailed information about groups, datasets, and their attributes
197
+ in a hierarchical format.
198
+
199
+ Args:
200
+ file_path: Path to the HDF5 file.
201
+ group_path: Starting group path to explore (default: root '/').
202
+
203
+ Prints:
204
+ - Group and dataset names with hierarchical indentation
205
+ - Dataset shapes and data types
206
+ - All attributes for groups and datasets
207
+ """
208
+ def recurse(name, obj, indent=0):
209
+ spacer = " " * indent
210
+ if isinstance(obj, h5py.Group):
211
+ print(f"{spacer}Group: {name}")
212
+ HdfBase.print_attrs(name, obj)
213
+ for key in obj:
214
+ recurse(f"{name}/{key}", obj[key], indent+1)
215
+ elif isinstance(obj, h5py.Dataset):
216
+ print(f"{spacer}Dataset: {name}")
217
+ print(f"{spacer} Shape: {obj.shape}")
218
+ print(f"{spacer} Dtype: {obj.dtype}")
219
+ HdfBase.print_attrs(name, obj)
220
+ else:
221
+ print(f"{spacer}Unknown object: {name}")
222
+
223
+ try:
224
+ with h5py.File(file_path, 'r') as hdf_file:
225
+ if group_path in hdf_file:
226
+ print("")
227
+ print(f"Exploring group: {group_path}\n")
228
+ group = hdf_file[group_path]
229
+ for key in group:
230
+ print("")
231
+ recurse(f"{group_path}/{key}", group[key], indent=1)
232
+ else:
233
+ print(f"Group path '{group_path}' not found in the HDF5 file.")
234
+ except Exception as e:
235
+ print(f"Error exploring HDF5 file: {e}")
236
+
237
+ @staticmethod
238
+ @log_call
239
+ @standardize_input(file_type='plan_hdf')
240
+ def get_polylines_from_parts(hdf_path: Path, path: str, info_name: str = "Polyline Info",
241
+ parts_name: str = "Polyline Parts",
242
+ points_name: str = "Polyline Points") -> List[LineString]:
243
+ """
244
+ Extract polylines from HDF file parts data.
245
+
246
+ Args:
247
+ hdf_path: Path to the HDF file.
248
+ path: Internal HDF path to polyline data.
249
+ info_name: Name of polyline info dataset.
250
+ parts_name: Name of polyline parts dataset.
251
+ points_name: Name of polyline points dataset.
252
+
253
+ Returns:
254
+ List of Shapely LineString/MultiLineString geometries.
255
+
256
+ Note:
257
+ Expects HDF datasets containing:
258
+ - Polyline information (start points and counts)
259
+ - Parts information for multi-part lines
260
+ - Point coordinates
261
+ """
262
+ try:
263
+ with h5py.File(hdf_path, 'r') as hdf_file:
264
+ polyline_info_path = f"{path}/{info_name}"
265
+ polyline_parts_path = f"{path}/{parts_name}"
266
+ polyline_points_path = f"{path}/{points_name}"
267
+
268
+ polyline_info = hdf_file[polyline_info_path][()]
269
+ polyline_parts = hdf_file[polyline_parts_path][()]
270
+ polyline_points = hdf_file[polyline_points_path][()]
271
+
272
+ geoms = []
273
+ for pnt_start, pnt_cnt, part_start, part_cnt in polyline_info:
274
+ points = polyline_points[pnt_start : pnt_start + pnt_cnt]
275
+ if part_cnt == 1:
276
+ geoms.append(LineString(points))
277
+ else:
278
+ parts = polyline_parts[part_start : part_start + part_cnt]
279
+ geoms.append(
280
+ MultiLineString(
281
+ list(
282
+ points[part_pnt_start : part_pnt_start + part_pnt_cnt]
283
+ for part_pnt_start, part_pnt_cnt in parts
284
+ )
285
+ )
286
+ )
287
+ return geoms
288
+ except Exception as e:
289
+ logger.error(f"Error getting polylines: {str(e)}")
290
+ return []
291
+
292
+ @staticmethod
293
+ def print_attrs(name: str, obj: Union[h5py.Dataset, h5py.Group]) -> None:
294
+ """
295
+ Print the attributes of an HDF5 object (Dataset or Group).
296
+
297
+ Args:
298
+ name (str): Name of the object
299
+ obj (Union[h5py.Dataset, h5py.Group]): HDF5 object whose attributes are to be printed
300
+ """
301
+ if len(obj.attrs) > 0:
302
+ print(f" Attributes for {name}:")
303
+ for key, value in obj.attrs.items():
304
+ print(f" {key}: {value}")
305
+
306
+
307
+