code-loader 1.0.56__tar.gz → 1.0.58__tar.gz

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (30) hide show
  1. {code_loader-1.0.56 → code_loader-1.0.58}/PKG-INFO +1 -1
  2. code_loader-1.0.58/code_loader/dualstream.py +32 -0
  3. {code_loader-1.0.56 → code_loader-1.0.58}/code_loader/inner_leap_binder/leapbinder_decorators.py +21 -11
  4. {code_loader-1.0.56 → code_loader-1.0.58}/code_loader/leaploader.py +5 -3
  5. {code_loader-1.0.56 → code_loader-1.0.58}/pyproject.toml +1 -1
  6. code_loader-1.0.56/code_loader/code_inegration_processes_manager.py +0 -83
  7. {code_loader-1.0.56 → code_loader-1.0.58}/LICENSE +0 -0
  8. {code_loader-1.0.56 → code_loader-1.0.58}/README.md +0 -0
  9. {code_loader-1.0.56 → code_loader-1.0.58}/code_loader/__init__.py +0 -0
  10. {code_loader-1.0.56 → code_loader-1.0.58}/code_loader/contract/__init__.py +0 -0
  11. {code_loader-1.0.56 → code_loader-1.0.58}/code_loader/contract/datasetclasses.py +0 -0
  12. {code_loader-1.0.56 → code_loader-1.0.58}/code_loader/contract/enums.py +0 -0
  13. {code_loader-1.0.56 → code_loader-1.0.58}/code_loader/contract/exceptions.py +0 -0
  14. {code_loader-1.0.56 → code_loader-1.0.58}/code_loader/contract/responsedataclasses.py +0 -0
  15. {code_loader-1.0.56 → code_loader-1.0.58}/code_loader/contract/visualizer_classes.py +0 -0
  16. {code_loader-1.0.56 → code_loader-1.0.58}/code_loader/experiment_api/__init__.py +0 -0
  17. {code_loader-1.0.56 → code_loader-1.0.58}/code_loader/experiment_api/api.py +0 -0
  18. {code_loader-1.0.56 → code_loader-1.0.58}/code_loader/experiment_api/cli_config_utils.py +0 -0
  19. {code_loader-1.0.56 → code_loader-1.0.58}/code_loader/experiment_api/client.py +0 -0
  20. {code_loader-1.0.56 → code_loader-1.0.58}/code_loader/experiment_api/epoch.py +0 -0
  21. {code_loader-1.0.56 → code_loader-1.0.58}/code_loader/experiment_api/experiment.py +0 -0
  22. {code_loader-1.0.56 → code_loader-1.0.58}/code_loader/experiment_api/experiment_context.py +0 -0
  23. {code_loader-1.0.56 → code_loader-1.0.58}/code_loader/experiment_api/types.py +0 -0
  24. {code_loader-1.0.56 → code_loader-1.0.58}/code_loader/experiment_api/utils.py +0 -0
  25. {code_loader-1.0.56 → code_loader-1.0.58}/code_loader/experiment_api/workingspace_config_utils.py +0 -0
  26. {code_loader-1.0.56 → code_loader-1.0.58}/code_loader/inner_leap_binder/__init__.py +0 -0
  27. {code_loader-1.0.56 → code_loader-1.0.58}/code_loader/inner_leap_binder/leapbinder.py +0 -0
  28. {code_loader-1.0.56 → code_loader-1.0.58}/code_loader/utils.py +0 -0
  29. {code_loader-1.0.56 → code_loader-1.0.58}/code_loader/visualizers/__init__.py +0 -0
  30. {code_loader-1.0.56 → code_loader-1.0.58}/code_loader/visualizers/default_visualizers.py +0 -0
@@ -1,6 +1,6 @@
1
1
  Metadata-Version: 2.1
2
2
  Name: code-loader
3
- Version: 1.0.56
3
+ Version: 1.0.58
4
4
  Summary:
5
5
  Home-page: https://github.com/tensorleap/code-loader
6
6
  License: MIT
@@ -0,0 +1,32 @@
1
+ from io import StringIO
2
+ from typing import IO
3
+
4
+
5
+ class DualStream(StringIO):
6
+ def __init__(self, stream1: IO[str], stream2: StringIO):
7
+ super().__init__()
8
+ self.stream1 = stream1 # Usually sys.stdout
9
+ self.stream2 = stream2 # The StringIO stream
10
+
11
+ def write(self, s: str) -> int:
12
+ # Write to both streams and return the length of the written string
13
+ self.stream1.write(s)
14
+ self.stream2.write(s)
15
+ return len(s)
16
+
17
+ def flush(self) -> None:
18
+ self.stream1.flush()
19
+ self.stream2.flush()
20
+
21
+ def close(self) -> None:
22
+ # Do not close sys.stdout
23
+ self.stream2.close()
24
+
25
+ def readable(self) -> bool:
26
+ return False
27
+
28
+ def writable(self) -> bool:
29
+ return True
30
+
31
+ def seekable(self) -> bool:
32
+ return False
@@ -52,20 +52,30 @@ def tensorleap_custom_metric(name: str, direction: Optional[MetricDirection] = M
52
52
  f'Metric has returned unsupported type. Supported types are List[float], '
53
53
  f'List[List[ConfusionMatrixElement]], NDArray[np.float32]. ')
54
54
 
55
- if isinstance(result, list):
56
- if isinstance(result[0], list):
57
- assert isinstance(result[0][0], ConfusionMatrixElement), \
58
- f'{supported_types_message}Got List[List[{type(result[0][0])}]].'
55
+ def _validate_single_metric(single_metric_result):
56
+ if isinstance(single_metric_result, list):
57
+ if isinstance(single_metric_result[0], list):
58
+ assert isinstance(single_metric_result[0][0], ConfusionMatrixElement), \
59
+ f'{supported_types_message}Got List[List[{type(single_metric_result[0][0])}]].'
60
+ else:
61
+ assert isinstance(single_metric_result[0], float), f'{supported_types_message}Got List[{type(single_metric_result[0])}].'
59
62
  else:
60
- assert isinstance(result[0], float), f'{supported_types_message}Got List[{type(result[0])}].'
63
+ assert isinstance(single_metric_result, np.ndarray), f'{supported_types_message}Got {type(single_metric_result)}.'
64
+ assert len(single_metric_result.shape) == 1, (f'tensorleap_custom_metric validation failed: '
65
+ f'The return shape should be 1D. Got {len(single_metric_result.shape)}D.')
61
66
 
67
+ if leap_binder.batch_size_to_validate:
68
+ assert len(single_metric_result) == leap_binder.batch_size_to_validate, \
69
+ f'tensorleap_custom_metrix validation failed: The return len should be as the batch size.'
70
+
71
+ if isinstance(result, dict):
72
+ for key, value in result.items():
73
+ assert isinstance(key, str), \
74
+ (f'tensorleap_custom_metric validation failed: '
75
+ f'Keys in the return dict should be of type str. Got {type(key)}.')
76
+ _validate_single_metric(value)
62
77
  else:
63
- assert isinstance(result, np.ndarray), f'{supported_types_message}Got {type(result)}.'
64
- assert len(result.shape) == 1, (f'tensorleap_custom_metric validation failed: '
65
- f'The return shape should be 1D. Got {len(result.shape)}D.')
66
- if leap_binder.batch_size_to_validate:
67
- assert len(result) == leap_binder.batch_size_to_validate, \
68
- f'tensorleap_custom_metrix validation failed: The return len should be as the batch size.'
78
+ _validate_single_metric(result)
69
79
 
70
80
  def inner(*args, **kwargs):
71
81
  _validate_input_args(*args, **kwargs)
@@ -19,6 +19,7 @@ from code_loader.contract.exceptions import DatasetScriptException
19
19
  from code_loader.contract.responsedataclasses import DatasetIntegParseResult, DatasetTestResultPayload, \
20
20
  DatasetPreprocess, DatasetSetup, DatasetInputInstance, DatasetOutputInstance, DatasetMetadataInstance, \
21
21
  VisualizerInstance, PredictionTypeInstance, ModelSetup, CustomLayerInstance, MetricInstance, CustomLossInstance
22
+ from code_loader.dualstream import DualStream
22
23
  from code_loader.inner_leap_binder import global_leap_binder
23
24
  from code_loader.utils import get_root_exception_file_and_line_number
24
25
 
@@ -123,8 +124,9 @@ class LeapLoader:
123
124
  test_payloads: List[DatasetTestResultPayload] = []
124
125
  setup_response = None
125
126
  general_error = None
126
- stdout_steam = io.StringIO()
127
- with redirect_stdout(stdout_steam):
127
+ stdout_stream = io.StringIO()
128
+ dual_stream = DualStream(sys.stdout, stdout_stream)
129
+ with redirect_stdout(dual_stream):
128
130
  try:
129
131
  self.exec_script()
130
132
  preprocess_test_payload = self._check_preprocess()
@@ -142,7 +144,7 @@ class LeapLoader:
142
144
  general_error = f"Something went wrong. {repr(e.__cause__)} in file {file_name}, line_number: {line_number}\nStacktrace:\n{stacktrace}"
143
145
  is_valid = False
144
146
 
145
- print_log = stdout_steam.getvalue()
147
+ print_log = dual_stream.stream2.getvalue()
146
148
  is_valid_for_model = bool(global_leap_binder.setup_container.custom_layers)
147
149
  model_setup = self.get_model_setup_response()
148
150
 
@@ -1,6 +1,6 @@
1
1
  [tool.poetry]
2
2
  name = "code-loader"
3
- version = "1.0.56"
3
+ version = "1.0.58"
4
4
  description = ""
5
5
  authors = ["dorhar <doron.harnoy@tensorleap.ai>"]
6
6
  license = "MIT"
@@ -1,83 +0,0 @@
1
- # mypy: ignore-errors
2
- import traceback
3
- from dataclasses import dataclass
4
-
5
- from typing import List, Tuple, Optional
6
-
7
- from multiprocessing import Process, Queue
8
-
9
- from code_loader.leap_loader_parallelized_base import LeapLoaderParallelizedBase
10
- from code_loader.leaploader import LeapLoader
11
- from code_loader.contract.enums import DataStateEnum
12
- from code_loader.metric_calculator_parallelized import MetricCalculatorParallelized
13
- from code_loader.samples_generator_parallelized import SamplesGeneratorParallelized
14
-
15
-
16
- @dataclass
17
- class SampleSerializableError:
18
- state: DataStateEnum
19
- index: int
20
- leap_script_trace: str
21
- exception_as_str: str
22
-
23
-
24
- class CodeIntegrationProcessesManager:
25
- def __init__(self, code_path: str, code_entry_name: str, n_workers: Optional[int] = 2,
26
- max_samples_in_queue: int = 128) -> None:
27
- self.metric_calculator_parallelized = MetricCalculatorParallelized(code_path, code_entry_name)
28
- self.samples_generator_parallelized = SamplesGeneratorParallelized(code_path, code_entry_name)
29
-
30
- def _create_and_start_process(self) -> Process:
31
- process = self.multiprocessing_context.Process(
32
- target=CodeIntegrationProcessesManager._process_func,
33
- args=(self.code_path, self.code_entry_name, self._inputs_waiting_to_be_process,
34
- self._ready_processed_results))
35
- process.daemon = True
36
- process.start()
37
- return process
38
-
39
- def _run_and_warm_first_process(self):
40
- process = self._create_and_start_process()
41
- self.processes = [process]
42
-
43
- # needed in order to make sure the preprocess func runs once in nonparallel
44
- self._start_process_inputs([(DataStateEnum.training, 0)])
45
- self._get_next_ready_processed_result()
46
-
47
- def _operation_decider(self):
48
- if self.metric_calculator_parallelized._ready_processed_results.empty() and not \
49
- self.metric_calculator_parallelized._inputs_waiting_to_be_process.empty():
50
- return 'metric'
51
-
52
- if self.samples_generator_parallelized._ready_processed_results.empty() and not \
53
- self.samples_generator_parallelized._inputs_waiting_to_be_process.empty():
54
- return 'dataset'
55
-
56
-
57
-
58
-
59
- @staticmethod
60
- def _process_func(code_path: str, code_entry_name: str,
61
- samples_to_process: Queue, ready_samples: Queue,
62
- metrics_to_process: Queue, ready_metrics: Queue) -> None:
63
- import os
64
- os.environ["CUDA_VISIBLE_DEVICES"] = "-1"
65
-
66
- leap_loader = LeapLoader(code_path, code_entry_name)
67
- while True:
68
-
69
- # decide on sample or metric to process
70
- state, idx = samples_to_process.get(block=True)
71
- leap_loader._preprocess_result()
72
- try:
73
- sample = leap_loader.get_sample(state, idx)
74
- except Exception as e:
75
- leap_script_trace = traceback.format_exc().split('File "<string>"')[-1]
76
- ready_samples.put(SampleSerializableError(state, idx, leap_script_trace, str(e)))
77
- continue
78
-
79
- ready_samples.put(sample)
80
-
81
- def generate_samples(self, sample_identities: List[Tuple[DataStateEnum, int]]):
82
- return self.start_process_inputs(sample_identities)
83
-
File without changes
File without changes