tracdap-runtime 0.7.0rc1__py3-none-any.whl → 0.8.0b2__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (52) hide show
  1. tracdap/rt/_exec/actors.py +5 -4
  2. tracdap/rt/_exec/context.py +166 -74
  3. tracdap/rt/_exec/dev_mode.py +147 -71
  4. tracdap/rt/_exec/engine.py +224 -99
  5. tracdap/rt/_exec/functions.py +122 -80
  6. tracdap/rt/_exec/graph.py +23 -35
  7. tracdap/rt/_exec/graph_builder.py +250 -113
  8. tracdap/rt/_exec/runtime.py +24 -10
  9. tracdap/rt/_exec/server.py +4 -3
  10. tracdap/rt/_impl/config_parser.py +3 -2
  11. tracdap/rt/_impl/data.py +89 -16
  12. tracdap/rt/_impl/grpc/tracdap/metadata/file_pb2.py +3 -1
  13. tracdap/rt/_impl/grpc/tracdap/metadata/file_pb2.pyi +8 -0
  14. tracdap/rt/_impl/grpc/tracdap/metadata/job_pb2.py +64 -62
  15. tracdap/rt/_impl/grpc/tracdap/metadata/job_pb2.pyi +16 -2
  16. tracdap/rt/_impl/grpc/tracdap/metadata/model_pb2.py +27 -25
  17. tracdap/rt/_impl/grpc/tracdap/metadata/model_pb2.pyi +14 -4
  18. tracdap/rt/_impl/grpc/tracdap/metadata/object_id_pb2.py +3 -3
  19. tracdap/rt/_impl/grpc/tracdap/metadata/object_id_pb2.pyi +2 -0
  20. tracdap/rt/_impl/grpc/tracdap/metadata/object_pb2.py +4 -4
  21. tracdap/rt/_impl/grpc/tracdap/metadata/object_pb2.pyi +4 -2
  22. tracdap/rt/_impl/logging.py +195 -0
  23. tracdap/rt/_impl/models.py +11 -8
  24. tracdap/rt/_impl/repos.py +5 -3
  25. tracdap/rt/_impl/schemas.py +2 -2
  26. tracdap/rt/_impl/shim.py +3 -2
  27. tracdap/rt/_impl/static_api.py +53 -33
  28. tracdap/rt/_impl/storage.py +4 -3
  29. tracdap/rt/_impl/util.py +1 -111
  30. tracdap/rt/_impl/validation.py +57 -30
  31. tracdap/rt/_version.py +1 -1
  32. tracdap/rt/api/__init__.py +6 -3
  33. tracdap/rt/api/file_types.py +29 -0
  34. tracdap/rt/api/hook.py +15 -7
  35. tracdap/rt/api/model_api.py +16 -0
  36. tracdap/rt/api/static_api.py +211 -125
  37. tracdap/rt/config/__init__.py +6 -6
  38. tracdap/rt/config/common.py +11 -1
  39. tracdap/rt/config/platform.py +4 -6
  40. tracdap/rt/ext/plugins.py +2 -2
  41. tracdap/rt/launch/launch.py +9 -11
  42. tracdap/rt/metadata/__init__.py +11 -9
  43. tracdap/rt/metadata/file.py +8 -0
  44. tracdap/rt/metadata/job.py +16 -0
  45. tracdap/rt/metadata/model.py +12 -2
  46. tracdap/rt/metadata/object.py +2 -0
  47. tracdap/rt/metadata/object_id.py +2 -0
  48. {tracdap_runtime-0.7.0rc1.dist-info → tracdap_runtime-0.8.0b2.dist-info}/METADATA +15 -15
  49. {tracdap_runtime-0.7.0rc1.dist-info → tracdap_runtime-0.8.0b2.dist-info}/RECORD +52 -50
  50. {tracdap_runtime-0.7.0rc1.dist-info → tracdap_runtime-0.8.0b2.dist-info}/WHEEL +1 -1
  51. {tracdap_runtime-0.7.0rc1.dist-info → tracdap_runtime-0.8.0b2.dist-info}/LICENSE +0 -0
  52. {tracdap_runtime-0.7.0rc1.dist-info → tracdap_runtime-0.8.0b2.dist-info}/top_level.txt +0 -0
@@ -23,7 +23,7 @@ from tracdap.rt._impl.grpc.tracdap.metadata import custom_pb2 as tracdap_dot_rt_
23
23
  from tracdap.rt._impl.grpc.tracdap.metadata import stoarge_pb2 as tracdap_dot_rt_dot___impl_dot_grpc_dot_tracdap_dot_metadata_dot_stoarge__pb2
24
24
 
25
25
 
26
- DESCRIPTOR = _descriptor_pool.Default().AddSerializedFile(b'\n3tracdap/rt/_impl/grpc/tracdap/metadata/object.proto\x12\x10tracdap.metadata\x1a\x36tracdap/rt/_impl/grpc/tracdap/metadata/object_id.proto\x1a\x31tracdap/rt/_impl/grpc/tracdap/metadata/type.proto\x1a\x31tracdap/rt/_impl/grpc/tracdap/metadata/data.proto\x1a\x32tracdap/rt/_impl/grpc/tracdap/metadata/model.proto\x1a\x31tracdap/rt/_impl/grpc/tracdap/metadata/flow.proto\x1a\x30tracdap/rt/_impl/grpc/tracdap/metadata/job.proto\x1a\x31tracdap/rt/_impl/grpc/tracdap/metadata/file.proto\x1a\x33tracdap/rt/_impl/grpc/tracdap/metadata/custom.proto\x1a\x34tracdap/rt/_impl/grpc/tracdap/metadata/stoarge.proto\"\x87\x05\n\x10ObjectDefinition\x12\x30\n\nobjectType\x18\x01 \x01(\x0e\x32\x1c.tracdap.metadata.ObjectType\x12\x30\n\x04\x64\x61ta\x18\x02 \x01(\x0b\x32 .tracdap.metadata.DataDefinitionH\x00\x12\x32\n\x05model\x18\x03 \x01(\x0b\x32!.tracdap.metadata.ModelDefinitionH\x00\x12\x30\n\x04\x66low\x18\x04 \x01(\x0b\x32 .tracdap.metadata.FlowDefinitionH\x00\x12.\n\x03job\x18\x05 \x01(\x0b\x32\x1f.tracdap.metadata.JobDefinitionH\x00\x12\x30\n\x04\x66ile\x18\x06 \x01(\x0b\x32 .tracdap.metadata.FileDefinitionH\x00\x12\x34\n\x06\x63ustom\x18\x07 \x01(\x0b\x32\".tracdap.metadata.CustomDefinitionH\x00\x12\x36\n\x07storage\x18\x08 \x01(\x0b\x32#.tracdap.metadata.StorageDefinitionH\x00\x12\x34\n\x06schema\x18\t \x01(\x0b\x32\".tracdap.metadata.SchemaDefinitionH\x00\x12H\n\x0bobjectProps\x18\x64 \x03(\x0b\x32\x33.tracdap.metadata.ObjectDefinition.ObjectPropsEntry\x1aK\n\x10ObjectPropsEntry\x12\x0b\n\x03key\x18\x01 \x01(\t\x12&\n\x05value\x18\x02 \x01(\x0b\x32\x17.tracdap.metadata.Value:\x02\x38\x01\x42\x0c\n\ndefinitionB2\n\x1aorg.finos.tracdap.metadataB\x12ObjectProtoWrapperP\x01\x62\x06proto3')
26
+ DESCRIPTOR = _descriptor_pool.Default().AddSerializedFile(b'\n3tracdap/rt/_impl/grpc/tracdap/metadata/object.proto\x12\x10tracdap.metadata\x1a\x36tracdap/rt/_impl/grpc/tracdap/metadata/object_id.proto\x1a\x31tracdap/rt/_impl/grpc/tracdap/metadata/type.proto\x1a\x31tracdap/rt/_impl/grpc/tracdap/metadata/data.proto\x1a\x32tracdap/rt/_impl/grpc/tracdap/metadata/model.proto\x1a\x31tracdap/rt/_impl/grpc/tracdap/metadata/flow.proto\x1a\x30tracdap/rt/_impl/grpc/tracdap/metadata/job.proto\x1a\x31tracdap/rt/_impl/grpc/tracdap/metadata/file.proto\x1a\x33tracdap/rt/_impl/grpc/tracdap/metadata/custom.proto\x1a\x34tracdap/rt/_impl/grpc/tracdap/metadata/stoarge.proto\"\xbd\x05\n\x10ObjectDefinition\x12\x30\n\nobjectType\x18\x01 \x01(\x0e\x32\x1c.tracdap.metadata.ObjectType\x12\x30\n\x04\x64\x61ta\x18\x02 \x01(\x0b\x32 .tracdap.metadata.DataDefinitionH\x00\x12\x32\n\x05model\x18\x03 \x01(\x0b\x32!.tracdap.metadata.ModelDefinitionH\x00\x12\x30\n\x04\x66low\x18\x04 \x01(\x0b\x32 .tracdap.metadata.FlowDefinitionH\x00\x12.\n\x03job\x18\x05 \x01(\x0b\x32\x1f.tracdap.metadata.JobDefinitionH\x00\x12\x30\n\x04\x66ile\x18\x06 \x01(\x0b\x32 .tracdap.metadata.FileDefinitionH\x00\x12\x34\n\x06\x63ustom\x18\x07 \x01(\x0b\x32\".tracdap.metadata.CustomDefinitionH\x00\x12\x36\n\x07storage\x18\x08 \x01(\x0b\x32#.tracdap.metadata.StorageDefinitionH\x00\x12\x34\n\x06schema\x18\t \x01(\x0b\x32\".tracdap.metadata.SchemaDefinitionH\x00\x12\x34\n\x06result\x18\n \x01(\x0b\x32\".tracdap.metadata.ResultDefinitionH\x00\x12H\n\x0bobjectProps\x18\x64 \x03(\x0b\x32\x33.tracdap.metadata.ObjectDefinition.ObjectPropsEntry\x1aK\n\x10ObjectPropsEntry\x12\x0b\n\x03key\x18\x01 \x01(\t\x12&\n\x05value\x18\x02 \x01(\x0b\x32\x17.tracdap.metadata.Value:\x02\x38\x01\x42\x0c\n\ndefinitionB2\n\x1aorg.finos.tracdap.metadataB\x12ObjectProtoWrapperP\x01\x62\x06proto3')
27
27
 
28
28
  _globals = globals()
29
29
  _builder.BuildMessageAndEnumDescriptors(DESCRIPTOR, _globals)
@@ -34,7 +34,7 @@ if _descriptor._USE_C_DESCRIPTORS == False:
34
34
  _globals['_OBJECTDEFINITION_OBJECTPROPSENTRY']._options = None
35
35
  _globals['_OBJECTDEFINITION_OBJECTPROPSENTRY']._serialized_options = b'8\001'
36
36
  _globals['_OBJECTDEFINITION']._serialized_start=543
37
- _globals['_OBJECTDEFINITION']._serialized_end=1190
38
- _globals['_OBJECTDEFINITION_OBJECTPROPSENTRY']._serialized_start=1101
39
- _globals['_OBJECTDEFINITION_OBJECTPROPSENTRY']._serialized_end=1176
37
+ _globals['_OBJECTDEFINITION']._serialized_end=1244
38
+ _globals['_OBJECTDEFINITION_OBJECTPROPSENTRY']._serialized_start=1155
39
+ _globals['_OBJECTDEFINITION_OBJECTPROPSENTRY']._serialized_end=1230
40
40
  # @@protoc_insertion_point(module_scope)
@@ -15,7 +15,7 @@ from typing import ClassVar as _ClassVar, Mapping as _Mapping, Optional as _Opti
15
15
  DESCRIPTOR: _descriptor.FileDescriptor
16
16
 
17
17
  class ObjectDefinition(_message.Message):
18
- __slots__ = ("objectType", "data", "model", "flow", "job", "file", "custom", "storage", "schema", "objectProps")
18
+ __slots__ = ("objectType", "data", "model", "flow", "job", "file", "custom", "storage", "schema", "result", "objectProps")
19
19
  class ObjectPropsEntry(_message.Message):
20
20
  __slots__ = ("key", "value")
21
21
  KEY_FIELD_NUMBER: _ClassVar[int]
@@ -32,6 +32,7 @@ class ObjectDefinition(_message.Message):
32
32
  CUSTOM_FIELD_NUMBER: _ClassVar[int]
33
33
  STORAGE_FIELD_NUMBER: _ClassVar[int]
34
34
  SCHEMA_FIELD_NUMBER: _ClassVar[int]
35
+ RESULT_FIELD_NUMBER: _ClassVar[int]
35
36
  OBJECTPROPS_FIELD_NUMBER: _ClassVar[int]
36
37
  objectType: _object_id_pb2.ObjectType
37
38
  data: _data_pb2.DataDefinition
@@ -42,5 +43,6 @@ class ObjectDefinition(_message.Message):
42
43
  custom: _custom_pb2.CustomDefinition
43
44
  storage: _stoarge_pb2.StorageDefinition
44
45
  schema: _data_pb2.SchemaDefinition
46
+ result: _job_pb2.ResultDefinition
45
47
  objectProps: _containers.MessageMap[str, _type_pb2.Value]
46
- def __init__(self, objectType: _Optional[_Union[_object_id_pb2.ObjectType, str]] = ..., data: _Optional[_Union[_data_pb2.DataDefinition, _Mapping]] = ..., model: _Optional[_Union[_model_pb2.ModelDefinition, _Mapping]] = ..., flow: _Optional[_Union[_flow_pb2.FlowDefinition, _Mapping]] = ..., job: _Optional[_Union[_job_pb2.JobDefinition, _Mapping]] = ..., file: _Optional[_Union[_file_pb2.FileDefinition, _Mapping]] = ..., custom: _Optional[_Union[_custom_pb2.CustomDefinition, _Mapping]] = ..., storage: _Optional[_Union[_stoarge_pb2.StorageDefinition, _Mapping]] = ..., schema: _Optional[_Union[_data_pb2.SchemaDefinition, _Mapping]] = ..., objectProps: _Optional[_Mapping[str, _type_pb2.Value]] = ...) -> None: ...
48
+ def __init__(self, objectType: _Optional[_Union[_object_id_pb2.ObjectType, str]] = ..., data: _Optional[_Union[_data_pb2.DataDefinition, _Mapping]] = ..., model: _Optional[_Union[_model_pb2.ModelDefinition, _Mapping]] = ..., flow: _Optional[_Union[_flow_pb2.FlowDefinition, _Mapping]] = ..., job: _Optional[_Union[_job_pb2.JobDefinition, _Mapping]] = ..., file: _Optional[_Union[_file_pb2.FileDefinition, _Mapping]] = ..., custom: _Optional[_Union[_custom_pb2.CustomDefinition, _Mapping]] = ..., storage: _Optional[_Union[_stoarge_pb2.StorageDefinition, _Mapping]] = ..., schema: _Optional[_Union[_data_pb2.SchemaDefinition, _Mapping]] = ..., result: _Optional[_Union[_job_pb2.ResultDefinition, _Mapping]] = ..., objectProps: _Optional[_Mapping[str, _type_pb2.Value]] = ...) -> None: ...
@@ -0,0 +1,195 @@
1
+ # Licensed to the Fintech Open Source Foundation (FINOS) under one or
2
+ # more contributor license agreements. See the NOTICE file distributed
3
+ # with this work for additional information regarding copyright ownership.
4
+ # FINOS licenses this file to you under the Apache License, Version 2.0
5
+ # (the "License"); you may not use this file except in compliance with the
6
+ # License. You may obtain a copy of the License at
7
+ #
8
+ # http://www.apache.org/licenses/LICENSE-2.0
9
+ #
10
+ # Unless required by applicable law or agreed to in writing, software
11
+ # distributed under the License is distributed on an "AS IS" BASIS,
12
+ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13
+ # See the License for the specific language governing permissions and
14
+ # limitations under the License.
15
+
16
+ import io as _io
17
+ import sys as _sys
18
+ import typing as _tp
19
+
20
+ from logging import *
21
+
22
+
23
+ class PlainFormatter(Formatter):
24
+
25
+ FORMAT = f"%(asctime)s [%(threadName)s] %(levelname)s %(name)s" + \
26
+ f" - %(message)s"
27
+
28
+ def __init__(self):
29
+ super().__init__(self.FORMAT)
30
+
31
+
32
+ class ColorFormatter(Formatter):
33
+
34
+ _BLACK, _RED, _GREEN, _YELLOW, _BLUE, _MAGENTA, _CYAN, _WHITE, _DEFAULT_WHITE = range(9)
35
+ _DARK_BASE = 30
36
+ _LIGHT_BASE = 90
37
+
38
+ # DARK_BASE + WHITE = light grey
39
+ # DARK_BASE + DEFAULT_WHITE = regular console white
40
+ # LIGHT_BASE + WHITE = bright white (0xffffff), very bright!
41
+
42
+ def __init__(self, is_bright: bool):
43
+
44
+ super().__init__(self._base_fmt(is_bright))
45
+ self._level_colors = self._make_level_colors(is_bright)
46
+ self._default_color = self._make_default_color(is_bright)
47
+
48
+ def format(self, record):
49
+
50
+ level_name = record.levelname
51
+ level_color = self._level_colors.get(level_name)
52
+
53
+ if level_color:
54
+ record.levelname = level_color
55
+ else:
56
+ record.levelname = self._default_color + level_name
57
+
58
+ return Formatter.format(self, record)
59
+
60
+ def _base_fmt(self, is_bright: bool):
61
+
62
+ if is_bright:
63
+ base_color = self._make_ansi_code(self._DARK_BASE, self._DEFAULT_WHITE, is_bold=False)
64
+ message_color = self._make_ansi_code(self._LIGHT_BASE, self._CYAN, is_bold=False)
65
+ else:
66
+ base_color = self._make_ansi_code(self._DARK_BASE, self._WHITE, is_bold=False)
67
+ message_color = self._make_ansi_code(self._DARK_BASE, self._CYAN, is_bold=False)
68
+
69
+ return f"{base_color}%(asctime)s [%(threadName)s] %(levelname)s{base_color} %(name)s" + \
70
+ f" - {message_color}%(message)s"
71
+
72
+ def _make_level_colors(self, is_bright: bool):
73
+
74
+ base_code = self._LIGHT_BASE if is_bright else self._DARK_BASE
75
+
76
+ green = self._make_ansi_code(base_code, self._GREEN, is_bold=is_bright)
77
+ yellow = self._make_ansi_code(base_code, self._YELLOW, is_bold=is_bright)
78
+ red = self._make_ansi_code(base_code, self._RED, is_bold=is_bright)
79
+
80
+ level_colors = {
81
+ 'CRITICAL': f"{red}CRITICAL",
82
+ 'ERROR': f"{red}ERROR",
83
+ 'WARNING': f"{yellow}WARNING",
84
+ 'INFO': f"{green}INFO"
85
+ }
86
+
87
+ return level_colors
88
+
89
+ def _make_default_color(self, is_bright: bool):
90
+
91
+ base_code = self._LIGHT_BASE if is_bright else self._DARK_BASE
92
+ blue = self._make_ansi_code(base_code, self._BLUE, is_bold=is_bright)
93
+
94
+ return blue
95
+
96
+ @classmethod
97
+ def _make_ansi_code(cls, base_code: int, color_offset: int, is_bold: bool):
98
+ return f"\033[{1 if is_bold else 0};{base_code + color_offset}m"
99
+
100
+
101
+ def configure_logging(enable_debug=False):
102
+
103
+ root_logger = getLogger()
104
+ log_level = DEBUG if enable_debug else INFO
105
+
106
+ if not root_logger.hasHandlers():
107
+
108
+ console_formatter = ColorFormatter(is_bright=True)
109
+ console_handler = StreamHandler(_sys.stdout)
110
+ console_handler.setFormatter(console_formatter)
111
+ console_handler.setLevel(INFO)
112
+ root_logger.addHandler(console_handler)
113
+ root_logger.setLevel(log_level)
114
+
115
+ # Use is_bright=False for logs from the TRAC runtime, so model logs stand out
116
+
117
+ trac_logger = getLogger("tracdap.rt")
118
+
119
+ console_formatter = ColorFormatter(is_bright=False)
120
+ console_handler = StreamHandler(_sys.stdout)
121
+ console_handler.setFormatter(console_formatter)
122
+ console_handler.setLevel(log_level)
123
+ trac_logger.addHandler(console_handler)
124
+ trac_logger.propagate = False
125
+
126
+
127
+ def logger_for_object(obj: object) -> Logger:
128
+ return logger_for_class(obj.__class__)
129
+
130
+
131
+ def logger_for_class(clazz: type) -> Logger:
132
+ qualified_class_name = f"{clazz.__module__}.{clazz.__name__}"
133
+ return getLogger(qualified_class_name)
134
+
135
+
136
+ def logger_for_namespace(namespace: str) -> Logger:
137
+ return getLogger(namespace)
138
+
139
+
140
+ class JobLogger(Logger):
141
+
142
+ def __init__(self, sys_log: Logger, *handlers: Handler):
143
+
144
+ super().__init__(sys_log.name, sys_log.level)
145
+ self._sys_log = sys_log._log
146
+ self._job_log = super()._log
147
+
148
+ for handler in handlers:
149
+ self.addHandler(handler)
150
+
151
+ def _log(self, level, msg, args, exc_info=None, extra=None, stack_info=False, stacklevel=1):
152
+
153
+ self._sys_log(level, msg, args, exc_info, extra, stack_info, stacklevel)
154
+ self._job_log(level, msg, args, exc_info, extra, stack_info, stacklevel)
155
+
156
+
157
+ class LogProvider:
158
+
159
+ def logger_for_object(self, obj: object) -> Logger:
160
+ return logger_for_object(obj)
161
+
162
+ def logger_for_class(self, clazz: type) -> Logger:
163
+ return logger_for_class(clazz)
164
+
165
+ def logger_for_namespace(self, namespace: str) -> Logger:
166
+ return logger_for_namespace(namespace)
167
+
168
+
169
+ class JobLogProvider(LogProvider):
170
+
171
+ def __init__(self, *handlers: Handler):
172
+ self.__handlers = handlers
173
+
174
+ def logger_for_object(self, obj: object) -> Logger:
175
+ base_logger = logger_for_object(obj)
176
+ return JobLogger(base_logger, *self.__handlers)
177
+
178
+ def logger_for_class(self, clazz: type) -> Logger:
179
+ base_logger = logger_for_class(clazz)
180
+ return JobLogger(base_logger, *self.__handlers)
181
+
182
+ def logger_for_namespace(self, namespace: str) -> Logger:
183
+ base_logger = logger_for_namespace(namespace)
184
+ return JobLogger(base_logger, *self.__handlers)
185
+
186
+
187
+ def job_log_provider(target: _tp.BinaryIO) -> JobLogProvider:
188
+
189
+ stream = _io.TextIOWrapper(target, newline="\r\n")
190
+ formatter = PlainFormatter()
191
+
192
+ handler = StreamHandler(stream)
193
+ handler.setFormatter(formatter)
194
+
195
+ return JobLogProvider(handler)
@@ -25,9 +25,10 @@ import tracdap.rt.metadata as _meta
25
25
  import tracdap.rt.config as _cfg
26
26
  import tracdap.rt.exceptions as _ex
27
27
 
28
- import tracdap.rt._impl.type_system as _types
28
+ import tracdap.rt._impl.logging as _logging
29
29
  import tracdap.rt._impl.repos as _repos
30
30
  import tracdap.rt._impl.shim as _shim
31
+ import tracdap.rt._impl.type_system as _types
31
32
  import tracdap.rt._impl.util as _util
32
33
  import tracdap.rt._impl.validation as _val
33
34
 
@@ -43,7 +44,7 @@ class ModelLoader:
43
44
 
44
45
  def __init__(self, sys_config: _cfg.RuntimeConfig, scratch_dir: pathlib.Path):
45
46
 
46
- self.__log = _util.logger_for_object(self)
47
+ self.__log = _logging.logger_for_object(self)
47
48
 
48
49
  self.__scratch_dir = scratch_dir.joinpath("models")
49
50
  self.__repos = _repos.RepositoryManager(sys_config)
@@ -226,13 +227,15 @@ class ModelLoader:
226
227
  self.__log.info(f"Parameter [{name}] - {param.paramType.basicType.name}")
227
228
  param.paramProps = self._encoded_props(param.paramProps, "parameter", name)
228
229
 
229
- for name, schema in model_def.inputs.items():
230
- self.__log.info(f"Input [{name}] - {schema.schema.schemaType.name}")
231
- schema.inputProps = self._encoded_props(schema.inputProps, "input", name)
230
+ for name, input_def in model_def.inputs.items():
231
+ input_type = input_def.schema.schemaType.name if input_def.objectType == _meta.ObjectType.DATA else input_def.objectType.name
232
+ self.__log.info(f"Input [{name}] - {input_type}")
233
+ input_def.inputProps = self._encoded_props(input_def.inputProps, "input", name)
232
234
 
233
- for name, schema in model_def.outputs.items():
234
- self.__log.info(f"Output [{name}] - {schema.schema.schemaType.name}")
235
- schema.outputProps = self._encoded_props(schema.outputProps, "input", name)
235
+ for name, output_def in model_def.outputs.items():
236
+ output_type = output_def.schema.schemaType.name if output_def.objectType == _meta.ObjectType.DATA else output_def.objectType.name
237
+ self.__log.info(f"Output [{name}] - {output_type}")
238
+ output_def.outputProps = self._encoded_props(output_def.outputProps, "input", name)
236
239
 
237
240
  return model_def
238
241
 
tracdap/rt/_impl/repos.py CHANGED
@@ -13,10 +13,12 @@
13
13
  # See the License for the specific language governing permissions and
14
14
  # limitations under the License.
15
15
 
16
+ import typing as _tp
17
+
16
18
  import tracdap.rt.ext.plugins as plugins
17
19
  import tracdap.rt.config as cfg
18
20
  import tracdap.rt.exceptions as ex
19
- import tracdap.rt._impl.util as util
21
+ import tracdap.rt._impl.logging as _logging
20
22
 
21
23
  # Import repo interfaces
22
24
  from tracdap.rt.ext.repos import *
@@ -26,8 +28,8 @@ class RepositoryManager:
26
28
 
27
29
  def __init__(self, sys_config: cfg.RuntimeConfig):
28
30
 
29
- self._log = util.logger_for_object(self)
30
- self._repos: tp.Dict[str, IModelRepository] = dict()
31
+ self._log = _logging.logger_for_object(self)
32
+ self._repos: _tp.Dict[str, IModelRepository] = dict()
31
33
 
32
34
  # Initialize all repos in the system config
33
35
  # Any errors for missing repo types (plugins) will be raised during startup
@@ -22,9 +22,9 @@ import pyarrow as pa
22
22
  import tracdap.rt.metadata as _meta
23
23
  import tracdap.rt.exceptions as _ex
24
24
  import tracdap.rt._impl.data as _data
25
+ import tracdap.rt._impl.logging as _log
25
26
  import tracdap.rt._impl.storage as _storage
26
27
  import tracdap.rt._impl.shim as _shim
27
- import tracdap.rt._impl.util as _util
28
28
 
29
29
 
30
30
  class SchemaLoader:
@@ -169,4 +169,4 @@ class SchemaLoader:
169
169
  raise _ex.EDataConformance(err)
170
170
 
171
171
 
172
- SchemaLoader._log = _util.logger_for_class(SchemaLoader)
172
+ SchemaLoader._log = _log.logger_for_class(SchemaLoader)
tracdap/rt/_impl/shim.py CHANGED
@@ -31,6 +31,7 @@ import importlib.resources as _ilr
31
31
 
32
32
  import tracdap.rt.exceptions as _ex
33
33
  import tracdap.rt._impl.guard_rails as _guard
34
+ import tracdap.rt._impl.logging as _log
34
35
  import tracdap.rt._impl.util as _util
35
36
 
36
37
 
@@ -138,7 +139,7 @@ class _NamespaceShimFinder(_ila.MetaPathFinder):
138
139
  def __init__(self, shim_map: tp.Dict[str, _Shim], active_shim: _ActiveShim):
139
140
  self.__shim_map = shim_map
140
141
  self.__active_shim = active_shim
141
- self._log = _util.logger_for_class(ShimLoader)
142
+ self._log = _log.logger_for_class(ShimLoader)
142
143
 
143
144
  def find_spec(
144
145
  self, fullname: str,
@@ -561,5 +562,5 @@ class ShimLoader:
561
562
  raise _ex.ERuntimeValidation(err)
562
563
 
563
564
 
564
- ShimLoader._log = _util.logger_for_class(ShimLoader)
565
+ ShimLoader._log = _log.logger_for_class(ShimLoader)
565
566
  ShimLoader._init() # noqa
@@ -152,14 +152,18 @@ class StaticApiImpl(_StaticApiHook):
152
152
 
153
153
  def define_schema(
154
154
  self, *fields: _tp.Union[_meta.FieldSchema, _tp.List[_meta.FieldSchema]],
155
- schema_type: _meta.SchemaType = _meta.SchemaType.TABLE) \
155
+ schema_type: _meta.SchemaType = _meta.SchemaType.TABLE, dynamic: bool = False) \
156
156
  -> _meta.SchemaDefinition:
157
157
 
158
- _val.validate_signature(self.define_schema, *fields, schema_type=schema_type)
158
+ _val.validate_signature(self.define_schema, *fields, schema_type=schema_type, dynamic=dynamic)
159
159
 
160
160
  if schema_type == _meta.SchemaType.TABLE:
161
161
 
162
- table_schema = self._build_table_schema(*fields)
162
+ if dynamic and not fields:
163
+ table_schema = None
164
+ else:
165
+ table_schema = self._build_table_schema(*fields)
166
+
163
167
  return _meta.SchemaDefinition(_meta.SchemaType.TABLE, table=table_schema)
164
168
 
165
169
  raise _ex.ERuntimeValidation(f"Invalid schema type [{schema_type.name}]")
@@ -182,51 +186,67 @@ class StaticApiImpl(_StaticApiHook):
182
186
 
183
187
  return converter.infer_schema(dataset)
184
188
 
185
- def define_input_table(
186
- self, *fields: _tp.Union[_meta.FieldSchema, _tp.List[_meta.FieldSchema]],
187
- label: _tp.Optional[str] = None, optional: bool = False, dynamic: bool = False,
188
- input_props: _tp.Optional[_tp.Dict[str, _tp.Any]] = None) \
189
- -> _meta.ModelInputSchema:
189
+ def define_file_type(self, extension: str, mime_type: str) -> _meta.FileType:
190
+
191
+ _val.validate_signature(self.define_file_type, extension, mime_type)
192
+
193
+ return _meta.FileType(extension=extension, mimeType=mime_type)
194
+
195
+ def define_input(
196
+ self, requirement: _tp.Union[_meta.SchemaDefinition, _meta.FileType], *,
197
+ label: _tp.Optional[str] = None,
198
+ optional: bool = False, dynamic: bool = False,
199
+ input_props: _tp.Optional[_tp.Dict[str, _tp.Any]] = None):
190
200
 
191
201
  _val.validate_signature(
192
- self.define_input_table, *fields,
202
+ self.define_input, requirement,
193
203
  label=label, optional=optional, dynamic=dynamic,
194
204
  input_props=input_props)
195
205
 
196
- # Do not define details for dynamic schemas
206
+ if isinstance(requirement, _meta.SchemaDefinition):
197
207
 
198
- if dynamic:
199
- schema_def = _meta.SchemaDefinition(_meta.SchemaType.TABLE)
200
- else:
201
- schema_def = self.define_schema(*fields, schema_type=_meta.SchemaType.TABLE)
208
+ return _meta.ModelInputSchema(
209
+ objectType=_meta.ObjectType.DATA, schema=requirement,
210
+ label=label, optional=optional, dynamic=dynamic,
211
+ inputProps=input_props)
202
212
 
203
- return _meta.ModelInputSchema(
204
- schema=schema_def, label=label,
205
- optional=optional, dynamic=dynamic,
206
- inputProps=input_props)
213
+ elif isinstance(requirement, _meta.FileType):
207
214
 
208
- def define_output_table(
209
- self, *fields: _tp.Union[_meta.FieldSchema, _tp.List[_meta.FieldSchema]],
210
- label: _tp.Optional[str] = None, optional: bool = False, dynamic: bool = False,
211
- output_props: _tp.Optional[_tp.Dict[str, _tp.Any]] = None) \
212
- -> _meta.ModelOutputSchema:
215
+ return _meta.ModelInputSchema(
216
+ objectType=_meta.ObjectType.FILE, fileType=requirement,
217
+ label=label, optional=optional, dynamic=dynamic,
218
+ inputProps=input_props)
219
+
220
+ else:
221
+ raise _ex.EUnexpected()
222
+
223
+ def define_output(
224
+ self, requirement: _tp.Union[_meta.SchemaDefinition, _meta.FileType], *,
225
+ label: _tp.Optional[str] = None,
226
+ optional: bool = False, dynamic: bool = False,
227
+ output_props: _tp.Optional[_tp.Dict[str, _tp.Any]] = None):
213
228
 
214
229
  _val.validate_signature(
215
- self.define_output_table, *fields,
230
+ self.define_output, requirement,
216
231
  label=label, optional=optional, dynamic=dynamic,
217
232
  output_props=output_props)
218
233
 
219
- # Do not define details for dynamic schemas
234
+ if isinstance(requirement, _meta.SchemaDefinition):
220
235
 
221
- if dynamic:
222
- schema_def = _meta.SchemaDefinition(_meta.SchemaType.TABLE)
223
- else:
224
- schema_def = self.define_schema(*fields, schema_type=_meta.SchemaType.TABLE)
236
+ return _meta.ModelOutputSchema(
237
+ objectType=_meta.ObjectType.DATA, schema=requirement,
238
+ label=label, optional=optional, dynamic=dynamic,
239
+ outputProps=output_props)
225
240
 
226
- return _meta.ModelOutputSchema(
227
- schema=schema_def, label=label,
228
- optional=optional, dynamic=dynamic,
229
- outputProps=output_props)
241
+ elif isinstance(requirement, _meta.FileType):
242
+
243
+ return _meta.ModelOutputSchema(
244
+ objectType=_meta.ObjectType.FILE, fileType=requirement,
245
+ label=label, optional=optional, dynamic=dynamic,
246
+ outputProps=output_props)
247
+
248
+ else:
249
+ raise _ex.EUnexpected()
230
250
 
231
251
  @staticmethod
232
252
  def _build_named_dict(
@@ -30,6 +30,7 @@ import tracdap.rt.config as _cfg
30
30
  import tracdap.rt.exceptions as _ex
31
31
  import tracdap.rt.ext.plugins as plugins
32
32
  import tracdap.rt._impl.data as _data
33
+ import tracdap.rt._impl.logging as _logging
33
34
  import tracdap.rt._impl.util as _util
34
35
  import tracdap.rt._impl.validation as _val
35
36
 
@@ -76,7 +77,7 @@ class StorageManager:
76
77
 
77
78
  def __init__(self, sys_config: _cfg.RuntimeConfig):
78
79
 
79
- self.__log = _util.logger_for_object(self)
80
+ self.__log = _logging.logger_for_object(self)
80
81
  self.__file_storage: tp.Dict[str, IFileStorage] = dict()
81
82
  self.__data_storage: tp.Dict[str, IDataStorage] = dict()
82
83
  self.__external: tp.List[str] = list()
@@ -224,7 +225,7 @@ class CommonFileStorage(IFileStorage):
224
225
 
225
226
  def __init__(self, storage_key: str, storage_config: _cfg.PluginConfig, fs: pa_fs.SubTreeFileSystem):
226
227
 
227
- self._log = _util.logger_for_object(self)
228
+ self._log = _logging.logger_for_object(self)
228
229
  self._key = storage_key
229
230
  self._config = storage_config
230
231
  self._fs = fs
@@ -731,7 +732,7 @@ class CommonDataStorage(IDataStorage):
731
732
  self, config: _cfg.PluginConfig, file_storage: IFileStorage,
732
733
  pushdown_pandas: bool = False, pushdown_spark: bool = False):
733
734
 
734
- self.__log = _util.logger_for_object(self)
735
+ self.__log = _logging.logger_for_object(self)
735
736
 
736
737
  self.__config = config
737
738
  self.__file_storage = file_storage
tracdap/rt/_impl/util.py CHANGED
@@ -14,11 +14,9 @@
14
14
  # limitations under the License.
15
15
 
16
16
  import datetime as dt
17
- import logging
18
17
  import pathlib
19
18
  import platform
20
19
 
21
- import sys
22
20
  import typing as tp
23
21
  import uuid
24
22
 
@@ -38,114 +36,6 @@ def is_windows():
38
36
  return __IS_WINDOWS
39
37
 
40
38
 
41
- class ColorFormatter(logging.Formatter):
42
-
43
- _BLACK, _RED, _GREEN, _YELLOW, _BLUE, _MAGENTA, _CYAN, _WHITE, _DEFAULT_WHITE = range(9)
44
- _DARK_BASE = 30
45
- _LIGHT_BASE = 90
46
-
47
- # DARK_BASE + WHITE = light grey
48
- # DARK_BASE + DEFAULT_WHITE = regular console white
49
- # LIGHT_BASE + WHITE = bright white (0xffffff), very bright!
50
-
51
- def __init__(self, is_bright: bool):
52
-
53
- super().__init__(self._base_fmt(is_bright))
54
- self._level_colors = self._make_level_colors(is_bright)
55
- self._default_color = self._make_default_color(is_bright)
56
-
57
- def format(self, record):
58
-
59
- level_name = record.levelname
60
- level_color = self._level_colors.get(level_name)
61
-
62
- if level_color:
63
- record.levelname = level_color
64
- else:
65
- record.levelname = self._default_color + level_name
66
-
67
- return logging.Formatter.format(self, record)
68
-
69
- def _base_fmt(self, is_bright: bool):
70
-
71
- if is_bright:
72
- base_color = self._make_ansi_code(self._DARK_BASE, self._DEFAULT_WHITE, is_bold=False)
73
- message_color = self._make_ansi_code(self._LIGHT_BASE, self._CYAN, is_bold=False)
74
- else:
75
- base_color = self._make_ansi_code(self._DARK_BASE, self._WHITE, is_bold=False)
76
- message_color = self._make_ansi_code(self._DARK_BASE, self._CYAN, is_bold=False)
77
-
78
- return f"{base_color}%(asctime)s [%(threadName)s] %(levelname)s{base_color} %(name)s" + \
79
- f" - {message_color}%(message)s"
80
-
81
- def _make_level_colors(self, is_bright: bool):
82
-
83
- base_code = self._LIGHT_BASE if is_bright else self._DARK_BASE
84
-
85
- green = self._make_ansi_code(base_code, self._GREEN, is_bold=is_bright)
86
- yellow = self._make_ansi_code(base_code, self._YELLOW, is_bold=is_bright)
87
- red = self._make_ansi_code(base_code, self._RED, is_bold=is_bright)
88
-
89
- level_colors = {
90
- 'CRITICAL': f"{red}CRITICAL",
91
- 'ERROR': f"{red}ERROR",
92
- 'WARNING': f"{yellow}WARNING",
93
- 'INFO': f"{green}INFO"
94
- }
95
-
96
- return level_colors
97
-
98
- def _make_default_color(self, is_bright: bool):
99
-
100
- base_code = self._LIGHT_BASE if is_bright else self._DARK_BASE
101
- blue = self._make_ansi_code(base_code, self._BLUE, is_bold=is_bright)
102
-
103
- return blue
104
-
105
- @classmethod
106
- def _make_ansi_code(cls, base_code: int, color_offset: int, is_bold: bool):
107
- return f"\033[{1 if is_bold else 0};{base_code + color_offset}m"
108
-
109
-
110
- def configure_logging(enable_debug=False):
111
-
112
- root_logger = logging.getLogger()
113
- log_level = logging.DEBUG if enable_debug else logging.INFO
114
-
115
- if not root_logger.hasHandlers():
116
-
117
- console_formatter = ColorFormatter(is_bright=True)
118
- console_handler = logging.StreamHandler(sys.stdout)
119
- console_handler.setFormatter(console_formatter)
120
- console_handler.setLevel(logging.INFO)
121
- root_logger.addHandler(console_handler)
122
- root_logger.setLevel(log_level)
123
-
124
- # Use is_bright=False for logs from the TRAC runtime, so model logs stand out
125
-
126
- trac_logger = logging.getLogger("tracdap.rt")
127
-
128
- console_formatter = ColorFormatter(is_bright=False)
129
- console_handler = logging.StreamHandler(sys.stdout)
130
- console_handler.setFormatter(console_formatter)
131
- console_handler.setLevel(log_level)
132
- trac_logger.addHandler(console_handler)
133
- trac_logger.propagate = False
134
-
135
-
136
- def logger_for_object(obj: object) -> logging.Logger:
137
- return logger_for_class(obj.__class__)
138
-
139
-
140
- def logger_for_class(clazz: type) -> logging.Logger:
141
- qualified_class_name = f"{clazz.__module__}.{clazz.__name__}"
142
- return logging.getLogger(qualified_class_name)
143
-
144
-
145
- def logger_for_namespace(namespace: str) -> logging.Logger:
146
- return logging.getLogger(namespace)
147
-
148
-
149
39
  def format_file_size(size: int) -> str:
150
40
 
151
41
  if size < 1024:
@@ -235,7 +125,7 @@ def get_job_resource(
235
125
  if optional:
236
126
  return None
237
127
 
238
- err = f"Missing required {selector.objectType} resource [{object_key(selector)}]"
128
+ err = f"Missing required {selector.objectType.name} resource [{object_key(selector)}]"
239
129
  raise ex.ERuntimeValidation(err)
240
130
 
241
131