dtlpy 1.113.10__py3-none-any.whl → 1.114.13__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (243) hide show
  1. dtlpy/__init__.py +488 -488
  2. dtlpy/__version__.py +1 -1
  3. dtlpy/assets/__init__.py +26 -26
  4. dtlpy/assets/__pycache__/__init__.cpython-38.pyc +0 -0
  5. dtlpy/assets/code_server/config.yaml +2 -2
  6. dtlpy/assets/code_server/installation.sh +24 -24
  7. dtlpy/assets/code_server/launch.json +13 -13
  8. dtlpy/assets/code_server/settings.json +2 -2
  9. dtlpy/assets/main.py +53 -53
  10. dtlpy/assets/main_partial.py +18 -18
  11. dtlpy/assets/mock.json +11 -11
  12. dtlpy/assets/model_adapter.py +83 -83
  13. dtlpy/assets/package.json +61 -61
  14. dtlpy/assets/package_catalog.json +29 -29
  15. dtlpy/assets/package_gitignore +307 -307
  16. dtlpy/assets/service_runners/__init__.py +33 -33
  17. dtlpy/assets/service_runners/converter.py +96 -96
  18. dtlpy/assets/service_runners/multi_method.py +49 -49
  19. dtlpy/assets/service_runners/multi_method_annotation.py +54 -54
  20. dtlpy/assets/service_runners/multi_method_dataset.py +55 -55
  21. dtlpy/assets/service_runners/multi_method_item.py +52 -52
  22. dtlpy/assets/service_runners/multi_method_json.py +52 -52
  23. dtlpy/assets/service_runners/single_method.py +37 -37
  24. dtlpy/assets/service_runners/single_method_annotation.py +43 -43
  25. dtlpy/assets/service_runners/single_method_dataset.py +43 -43
  26. dtlpy/assets/service_runners/single_method_item.py +41 -41
  27. dtlpy/assets/service_runners/single_method_json.py +42 -42
  28. dtlpy/assets/service_runners/single_method_multi_input.py +45 -45
  29. dtlpy/assets/voc_annotation_template.xml +23 -23
  30. dtlpy/caches/base_cache.py +32 -32
  31. dtlpy/caches/cache.py +473 -473
  32. dtlpy/caches/dl_cache.py +201 -201
  33. dtlpy/caches/filesystem_cache.py +89 -89
  34. dtlpy/caches/redis_cache.py +84 -84
  35. dtlpy/dlp/__init__.py +20 -20
  36. dtlpy/dlp/cli_utilities.py +367 -367
  37. dtlpy/dlp/command_executor.py +764 -764
  38. dtlpy/dlp/dlp +1 -1
  39. dtlpy/dlp/dlp.bat +1 -1
  40. dtlpy/dlp/dlp.py +128 -128
  41. dtlpy/dlp/parser.py +651 -651
  42. dtlpy/entities/__init__.py +83 -83
  43. dtlpy/entities/analytic.py +311 -311
  44. dtlpy/entities/annotation.py +1879 -1879
  45. dtlpy/entities/annotation_collection.py +699 -699
  46. dtlpy/entities/annotation_definitions/__init__.py +20 -20
  47. dtlpy/entities/annotation_definitions/base_annotation_definition.py +100 -100
  48. dtlpy/entities/annotation_definitions/box.py +195 -195
  49. dtlpy/entities/annotation_definitions/classification.py +67 -67
  50. dtlpy/entities/annotation_definitions/comparison.py +72 -72
  51. dtlpy/entities/annotation_definitions/cube.py +204 -204
  52. dtlpy/entities/annotation_definitions/cube_3d.py +149 -149
  53. dtlpy/entities/annotation_definitions/description.py +32 -32
  54. dtlpy/entities/annotation_definitions/ellipse.py +124 -124
  55. dtlpy/entities/annotation_definitions/free_text.py +62 -62
  56. dtlpy/entities/annotation_definitions/gis.py +69 -69
  57. dtlpy/entities/annotation_definitions/note.py +139 -139
  58. dtlpy/entities/annotation_definitions/point.py +117 -117
  59. dtlpy/entities/annotation_definitions/polygon.py +182 -182
  60. dtlpy/entities/annotation_definitions/polyline.py +111 -111
  61. dtlpy/entities/annotation_definitions/pose.py +92 -92
  62. dtlpy/entities/annotation_definitions/ref_image.py +86 -86
  63. dtlpy/entities/annotation_definitions/segmentation.py +240 -240
  64. dtlpy/entities/annotation_definitions/subtitle.py +34 -34
  65. dtlpy/entities/annotation_definitions/text.py +85 -85
  66. dtlpy/entities/annotation_definitions/undefined_annotation.py +74 -74
  67. dtlpy/entities/app.py +220 -220
  68. dtlpy/entities/app_module.py +107 -107
  69. dtlpy/entities/artifact.py +174 -174
  70. dtlpy/entities/assignment.py +399 -399
  71. dtlpy/entities/base_entity.py +214 -214
  72. dtlpy/entities/bot.py +113 -113
  73. dtlpy/entities/codebase.py +296 -296
  74. dtlpy/entities/collection.py +38 -38
  75. dtlpy/entities/command.py +169 -169
  76. dtlpy/entities/compute.py +442 -442
  77. dtlpy/entities/dataset.py +1285 -1285
  78. dtlpy/entities/directory_tree.py +44 -44
  79. dtlpy/entities/dpk.py +470 -470
  80. dtlpy/entities/driver.py +222 -222
  81. dtlpy/entities/execution.py +397 -397
  82. dtlpy/entities/feature.py +124 -124
  83. dtlpy/entities/feature_set.py +145 -145
  84. dtlpy/entities/filters.py +641 -641
  85. dtlpy/entities/gis_item.py +107 -107
  86. dtlpy/entities/integration.py +184 -184
  87. dtlpy/entities/item.py +953 -953
  88. dtlpy/entities/label.py +123 -123
  89. dtlpy/entities/links.py +85 -85
  90. dtlpy/entities/message.py +175 -175
  91. dtlpy/entities/model.py +694 -691
  92. dtlpy/entities/node.py +1005 -1005
  93. dtlpy/entities/ontology.py +803 -803
  94. dtlpy/entities/organization.py +287 -287
  95. dtlpy/entities/package.py +657 -657
  96. dtlpy/entities/package_defaults.py +5 -5
  97. dtlpy/entities/package_function.py +185 -185
  98. dtlpy/entities/package_module.py +113 -113
  99. dtlpy/entities/package_slot.py +118 -118
  100. dtlpy/entities/paged_entities.py +290 -267
  101. dtlpy/entities/pipeline.py +593 -593
  102. dtlpy/entities/pipeline_execution.py +279 -279
  103. dtlpy/entities/project.py +394 -394
  104. dtlpy/entities/prompt_item.py +499 -499
  105. dtlpy/entities/recipe.py +301 -301
  106. dtlpy/entities/reflect_dict.py +102 -102
  107. dtlpy/entities/resource_execution.py +138 -138
  108. dtlpy/entities/service.py +958 -958
  109. dtlpy/entities/service_driver.py +117 -117
  110. dtlpy/entities/setting.py +294 -294
  111. dtlpy/entities/task.py +491 -491
  112. dtlpy/entities/time_series.py +143 -143
  113. dtlpy/entities/trigger.py +426 -426
  114. dtlpy/entities/user.py +118 -118
  115. dtlpy/entities/webhook.py +124 -124
  116. dtlpy/examples/__init__.py +19 -19
  117. dtlpy/examples/add_labels.py +135 -135
  118. dtlpy/examples/add_metadata_to_item.py +21 -21
  119. dtlpy/examples/annotate_items_using_model.py +65 -65
  120. dtlpy/examples/annotate_video_using_model_and_tracker.py +75 -75
  121. dtlpy/examples/annotations_convert_to_voc.py +9 -9
  122. dtlpy/examples/annotations_convert_to_yolo.py +9 -9
  123. dtlpy/examples/convert_annotation_types.py +51 -51
  124. dtlpy/examples/converter.py +143 -143
  125. dtlpy/examples/copy_annotations.py +22 -22
  126. dtlpy/examples/copy_folder.py +31 -31
  127. dtlpy/examples/create_annotations.py +51 -51
  128. dtlpy/examples/create_video_annotations.py +83 -83
  129. dtlpy/examples/delete_annotations.py +26 -26
  130. dtlpy/examples/filters.py +113 -113
  131. dtlpy/examples/move_item.py +23 -23
  132. dtlpy/examples/play_video_annotation.py +13 -13
  133. dtlpy/examples/show_item_and_mask.py +53 -53
  134. dtlpy/examples/triggers.py +49 -49
  135. dtlpy/examples/upload_batch_of_items.py +20 -20
  136. dtlpy/examples/upload_items_and_custom_format_annotations.py +55 -55
  137. dtlpy/examples/upload_items_with_modalities.py +43 -43
  138. dtlpy/examples/upload_segmentation_annotations_from_mask_image.py +44 -44
  139. dtlpy/examples/upload_yolo_format_annotations.py +70 -70
  140. dtlpy/exceptions.py +125 -125
  141. dtlpy/miscellaneous/__init__.py +20 -20
  142. dtlpy/miscellaneous/dict_differ.py +95 -95
  143. dtlpy/miscellaneous/git_utils.py +217 -217
  144. dtlpy/miscellaneous/json_utils.py +14 -14
  145. dtlpy/miscellaneous/list_print.py +105 -105
  146. dtlpy/miscellaneous/zipping.py +130 -130
  147. dtlpy/ml/__init__.py +20 -20
  148. dtlpy/ml/base_feature_extractor_adapter.py +27 -27
  149. dtlpy/ml/base_model_adapter.py +945 -940
  150. dtlpy/ml/metrics.py +461 -461
  151. dtlpy/ml/predictions_utils.py +274 -274
  152. dtlpy/ml/summary_writer.py +57 -57
  153. dtlpy/ml/train_utils.py +60 -60
  154. dtlpy/new_instance.py +252 -252
  155. dtlpy/repositories/__init__.py +56 -56
  156. dtlpy/repositories/analytics.py +85 -85
  157. dtlpy/repositories/annotations.py +916 -916
  158. dtlpy/repositories/apps.py +383 -383
  159. dtlpy/repositories/artifacts.py +452 -452
  160. dtlpy/repositories/assignments.py +599 -599
  161. dtlpy/repositories/bots.py +213 -213
  162. dtlpy/repositories/codebases.py +559 -559
  163. dtlpy/repositories/collections.py +332 -348
  164. dtlpy/repositories/commands.py +158 -158
  165. dtlpy/repositories/compositions.py +61 -61
  166. dtlpy/repositories/computes.py +434 -406
  167. dtlpy/repositories/datasets.py +1291 -1291
  168. dtlpy/repositories/downloader.py +895 -895
  169. dtlpy/repositories/dpks.py +433 -433
  170. dtlpy/repositories/drivers.py +266 -266
  171. dtlpy/repositories/executions.py +817 -817
  172. dtlpy/repositories/feature_sets.py +226 -226
  173. dtlpy/repositories/features.py +238 -238
  174. dtlpy/repositories/integrations.py +484 -484
  175. dtlpy/repositories/items.py +909 -915
  176. dtlpy/repositories/messages.py +94 -94
  177. dtlpy/repositories/models.py +877 -867
  178. dtlpy/repositories/nodes.py +80 -80
  179. dtlpy/repositories/ontologies.py +511 -511
  180. dtlpy/repositories/organizations.py +525 -525
  181. dtlpy/repositories/packages.py +1941 -1941
  182. dtlpy/repositories/pipeline_executions.py +448 -448
  183. dtlpy/repositories/pipelines.py +642 -642
  184. dtlpy/repositories/projects.py +539 -539
  185. dtlpy/repositories/recipes.py +399 -399
  186. dtlpy/repositories/resource_executions.py +137 -137
  187. dtlpy/repositories/schema.py +120 -120
  188. dtlpy/repositories/service_drivers.py +213 -213
  189. dtlpy/repositories/services.py +1704 -1704
  190. dtlpy/repositories/settings.py +339 -339
  191. dtlpy/repositories/tasks.py +1124 -1124
  192. dtlpy/repositories/times_series.py +278 -278
  193. dtlpy/repositories/triggers.py +536 -536
  194. dtlpy/repositories/upload_element.py +257 -257
  195. dtlpy/repositories/uploader.py +651 -651
  196. dtlpy/repositories/webhooks.py +249 -249
  197. dtlpy/services/__init__.py +22 -22
  198. dtlpy/services/aihttp_retry.py +131 -131
  199. dtlpy/services/api_client.py +1782 -1782
  200. dtlpy/services/api_reference.py +40 -40
  201. dtlpy/services/async_utils.py +133 -133
  202. dtlpy/services/calls_counter.py +44 -44
  203. dtlpy/services/check_sdk.py +68 -68
  204. dtlpy/services/cookie.py +115 -115
  205. dtlpy/services/create_logger.py +156 -156
  206. dtlpy/services/events.py +84 -84
  207. dtlpy/services/logins.py +235 -235
  208. dtlpy/services/reporter.py +256 -256
  209. dtlpy/services/service_defaults.py +91 -91
  210. dtlpy/utilities/__init__.py +20 -20
  211. dtlpy/utilities/annotations/__init__.py +16 -16
  212. dtlpy/utilities/annotations/annotation_converters.py +269 -269
  213. dtlpy/utilities/base_package_runner.py +264 -264
  214. dtlpy/utilities/converter.py +1650 -1650
  215. dtlpy/utilities/dataset_generators/__init__.py +1 -1
  216. dtlpy/utilities/dataset_generators/dataset_generator.py +670 -670
  217. dtlpy/utilities/dataset_generators/dataset_generator_tensorflow.py +23 -23
  218. dtlpy/utilities/dataset_generators/dataset_generator_torch.py +21 -21
  219. dtlpy/utilities/local_development/__init__.py +1 -1
  220. dtlpy/utilities/local_development/local_session.py +179 -179
  221. dtlpy/utilities/reports/__init__.py +2 -2
  222. dtlpy/utilities/reports/figures.py +343 -343
  223. dtlpy/utilities/reports/report.py +71 -71
  224. dtlpy/utilities/videos/__init__.py +17 -17
  225. dtlpy/utilities/videos/video_player.py +598 -598
  226. dtlpy/utilities/videos/videos.py +470 -470
  227. {dtlpy-1.113.10.data → dtlpy-1.114.13.data}/scripts/dlp +1 -1
  228. dtlpy-1.114.13.data/scripts/dlp.bat +2 -0
  229. {dtlpy-1.113.10.data → dtlpy-1.114.13.data}/scripts/dlp.py +128 -128
  230. {dtlpy-1.113.10.dist-info → dtlpy-1.114.13.dist-info}/LICENSE +200 -200
  231. {dtlpy-1.113.10.dist-info → dtlpy-1.114.13.dist-info}/METADATA +172 -172
  232. dtlpy-1.114.13.dist-info/RECORD +240 -0
  233. {dtlpy-1.113.10.dist-info → dtlpy-1.114.13.dist-info}/WHEEL +1 -1
  234. tests/features/environment.py +551 -550
  235. dtlpy-1.113.10.data/scripts/dlp.bat +0 -2
  236. dtlpy-1.113.10.dist-info/RECORD +0 -244
  237. tests/assets/__init__.py +0 -0
  238. tests/assets/models_flow/__init__.py +0 -0
  239. tests/assets/models_flow/failedmain.py +0 -52
  240. tests/assets/models_flow/main.py +0 -62
  241. tests/assets/models_flow/main_model.py +0 -54
  242. {dtlpy-1.113.10.dist-info → dtlpy-1.114.13.dist-info}/entry_points.txt +0 -0
  243. {dtlpy-1.113.10.dist-info → dtlpy-1.114.13.dist-info}/top_level.txt +0 -0
@@ -1,105 +1,105 @@
1
- import datetime
2
- import tabulate
3
- import typing
4
- import logging
5
- import pandas
6
-
7
- from .. import exceptions
8
-
9
- logger = logging.getLogger(name='dtlpy')
10
-
11
- T = typing.TypeVar('T')
12
-
13
-
14
- class List(list, typing.MutableSequence[T]):
15
- def to_df(self, show_all=False, columns=None):
16
- try:
17
- to_print = list()
18
- keys_list = list()
19
- for element in self.__iter__():
20
- if hasattr(element, 'to_json'):
21
- item_dict = element.to_json()
22
- else:
23
- item_dict = element
24
- to_print.append(item_dict)
25
- [keys_list.append(key) for key in list(item_dict.keys()) if key not in keys_list]
26
- try:
27
- # try sorting bt creation date
28
- to_print = sorted(to_print, key=lambda k: k['createdAt'] if k['createdAt'] is not None else "")
29
- except KeyError:
30
- pass
31
- except Exception:
32
- logger.exception('Error sorting printing:')
33
-
34
- remove_keys_list = ['contributors', 'url', 'annotations', 'items', 'export', 'directoryTree', 'org',
35
- '_contributors', 'role', 'account', 'featureConstraints',
36
- 'attributes', 'partitions', 'metadata', 'stream', 'updatedAt', 'arch',
37
- 'input', 'revisions', 'pipeline', # task fields
38
- 'feedbackQueue', # session fields
39
- '_ontology_ids', '_labels', # dataset
40
- 'esInstance', 'esIndex', # time series fields
41
- 'thumbnail', # item thumnail too long
42
- # services fields
43
- 'driverId', 'useUserJwt', 'versions', 'runtime', 'mq', 'global',
44
- # triggers
45
- 'scope',
46
- # Package
47
- 'modules'
48
- ]
49
- if not show_all:
50
- if columns is not None:
51
- # take columns from inputs
52
- if not isinstance(columns, list):
53
- if not isinstance(columns, str):
54
- raise exceptions.PlatformException(
55
- error='3002',
56
- message='"columns" input must be str or list. found: {}'.format(type(columns)))
57
- columns = [columns]
58
- keys_list = columns
59
- else:
60
- # take default columns
61
- for key in remove_keys_list:
62
- if key in keys_list:
63
- keys_list.remove(key)
64
-
65
- for element in to_print:
66
- # handle printing errors for not ascii string when in cli
67
- if 'name' in element:
68
- try:
69
- # check if ascii
70
- element['name'].encode('ascii')
71
- except UnicodeEncodeError:
72
- # if not - print bytes instead
73
- element['name'] = str(element['name']).encode('utf-8')
74
- if 'createdAt' in element:
75
- try:
76
- str_timestamp = str(element['createdAt'])
77
- if len(str_timestamp) > 10:
78
- str_timestamp = str_timestamp[:10]
79
- element['createdAt'] = datetime.datetime.fromtimestamp(int(str_timestamp), datetime.timezone.utc).isoformat()
80
- except Exception:
81
- pass
82
- df = pandas.DataFrame(to_print, columns=keys_list)
83
- return df
84
- except Exception:
85
- raise exceptions.PlatformException(error='3002',
86
- message='Failed converting to DataFrame')
87
-
88
- def print(self, show_all=False, level='print', to_return=False, columns=None):
89
- try:
90
- df = self.to_df(show_all=show_all, columns=columns)
91
- if 'name' in list(df.columns.values):
92
- df['name'] = df['name'].astype(str)
93
-
94
- if to_return:
95
- return tabulate.tabulate(df, headers='keys', tablefmt='psql')
96
- else:
97
- if level == 'print':
98
- print('\n{}'.format(tabulate.tabulate(df, headers='keys', tablefmt='psql')))
99
- elif level == 'debug':
100
- logger.debug('\n{}'.format(tabulate.tabulate(df, headers='keys', tablefmt='psql')))
101
- else:
102
- raise ValueError('unknown log level in printing: {}'.format(level))
103
-
104
- except Exception:
105
- raise exceptions.PlatformException(error='3002', message='Failed printing entity')
1
+ import datetime
2
+ import tabulate
3
+ import typing
4
+ import logging
5
+ import pandas
6
+
7
+ from .. import exceptions
8
+
9
+ logger = logging.getLogger(name='dtlpy')
10
+
11
+ T = typing.TypeVar('T')
12
+
13
+
14
+ class List(list, typing.MutableSequence[T]):
15
+ def to_df(self, show_all=False, columns=None):
16
+ try:
17
+ to_print = list()
18
+ keys_list = list()
19
+ for element in self.__iter__():
20
+ if hasattr(element, 'to_json'):
21
+ item_dict = element.to_json()
22
+ else:
23
+ item_dict = element
24
+ to_print.append(item_dict)
25
+ [keys_list.append(key) for key in list(item_dict.keys()) if key not in keys_list]
26
+ try:
27
+ # try sorting bt creation date
28
+ to_print = sorted(to_print, key=lambda k: k['createdAt'] if k['createdAt'] is not None else "")
29
+ except KeyError:
30
+ pass
31
+ except Exception:
32
+ logger.exception('Error sorting printing:')
33
+
34
+ remove_keys_list = ['contributors', 'url', 'annotations', 'items', 'export', 'directoryTree', 'org',
35
+ '_contributors', 'role', 'account', 'featureConstraints',
36
+ 'attributes', 'partitions', 'metadata', 'stream', 'updatedAt', 'arch',
37
+ 'input', 'revisions', 'pipeline', # task fields
38
+ 'feedbackQueue', # session fields
39
+ '_ontology_ids', '_labels', # dataset
40
+ 'esInstance', 'esIndex', # time series fields
41
+ 'thumbnail', # item thumnail too long
42
+ # services fields
43
+ 'driverId', 'useUserJwt', 'versions', 'runtime', 'mq', 'global',
44
+ # triggers
45
+ 'scope',
46
+ # Package
47
+ 'modules'
48
+ ]
49
+ if not show_all:
50
+ if columns is not None:
51
+ # take columns from inputs
52
+ if not isinstance(columns, list):
53
+ if not isinstance(columns, str):
54
+ raise exceptions.PlatformException(
55
+ error='3002',
56
+ message='"columns" input must be str or list. found: {}'.format(type(columns)))
57
+ columns = [columns]
58
+ keys_list = columns
59
+ else:
60
+ # take default columns
61
+ for key in remove_keys_list:
62
+ if key in keys_list:
63
+ keys_list.remove(key)
64
+
65
+ for element in to_print:
66
+ # handle printing errors for not ascii string when in cli
67
+ if 'name' in element:
68
+ try:
69
+ # check if ascii
70
+ element['name'].encode('ascii')
71
+ except UnicodeEncodeError:
72
+ # if not - print bytes instead
73
+ element['name'] = str(element['name']).encode('utf-8')
74
+ if 'createdAt' in element:
75
+ try:
76
+ str_timestamp = str(element['createdAt'])
77
+ if len(str_timestamp) > 10:
78
+ str_timestamp = str_timestamp[:10]
79
+ element['createdAt'] = datetime.datetime.fromtimestamp(int(str_timestamp), datetime.timezone.utc).isoformat()
80
+ except Exception:
81
+ pass
82
+ df = pandas.DataFrame(to_print, columns=keys_list)
83
+ return df
84
+ except Exception:
85
+ raise exceptions.PlatformException(error='3002',
86
+ message='Failed converting to DataFrame')
87
+
88
+ def print(self, show_all=False, level='print', to_return=False, columns=None):
89
+ try:
90
+ df = self.to_df(show_all=show_all, columns=columns)
91
+ if 'name' in list(df.columns.values):
92
+ df['name'] = df['name'].astype(str)
93
+
94
+ if to_return:
95
+ return tabulate.tabulate(df, headers='keys', tablefmt='psql')
96
+ else:
97
+ if level == 'print':
98
+ print('\n{}'.format(tabulate.tabulate(df, headers='keys', tablefmt='psql')))
99
+ elif level == 'debug':
100
+ logger.debug('\n{}'.format(tabulate.tabulate(df, headers='keys', tablefmt='psql')))
101
+ else:
102
+ raise ValueError('unknown log level in printing: {}'.format(level))
103
+
104
+ except Exception:
105
+ raise exceptions.PlatformException(error='3002', message='Failed printing entity')
@@ -1,130 +1,130 @@
1
- import logging
2
- import os
3
- import zipfile
4
- from typing import List
5
-
6
- import numpy as np
7
- import pathspec
8
-
9
- logger = logging.getLogger(name='dtlpy')
10
-
11
- MAX_ZIP_FILE = 100e6 # 100MB
12
-
13
-
14
- class Zipping:
15
- def __init__(self):
16
- pass
17
-
18
- @staticmethod
19
- def zip_directory(zip_filename, directory=None, ignore_max_file_size=False, ignore_directories: List[str] = None):
20
- """
21
- Zip Directory
22
- Will ignore .gitignore files
23
-
24
- :param directory: the directory to zip
25
- :param zip_filename: the name of the zipfile.
26
- :param ignore_max_file_size: ignore the limitation on the zip file size
27
- :param list[str] ignore_directories: directories to ignore.
28
- :return: None
29
- """
30
- # default path
31
- if directory is None:
32
- directory = os.getcwd()
33
- # check if directory
34
- assert os.path.isdir(directory), '[ERROR] Directory does not exists: {}'.format(directory)
35
-
36
- if '.gitignore' in os.listdir(directory):
37
- with open(os.path.join(directory, '.gitignore')) as f:
38
- spec_src = f.read()
39
- else:
40
- spec_src = ''
41
- ignore_lines = spec_src.splitlines() + ['.git', '.dataloop']
42
- if ignore_directories is not None:
43
- ignore_lines += ignore_directories
44
- spec = pathspec.PathSpec.from_lines(pathspec.patterns.GitWildMatchPattern, ignore_lines)
45
-
46
- # init zip file
47
- zip_file = zipfile.ZipFile(zip_filename, 'w', zipfile.ZIP_DEFLATED)
48
- try:
49
- for root, dirs, files in os.walk(directory):
50
- # remove dirs to avoid going file by file
51
- for d in dirs:
52
- if spec.match_file(os.path.relpath(os.path.join(root, d), directory)):
53
- dirs.remove(d)
54
- for file in files:
55
- filepath = os.path.join(root, file)
56
- if not spec.match_file(os.path.relpath(filepath, directory)):
57
- Zipping.__add_to_zip_file(directory, filepath, ignore_max_file_size, zip_file)
58
- finally:
59
- zip_file.close()
60
-
61
- @staticmethod
62
- def zip_directory_inclusive(zip_filename, directory=None, ignore_max_file_size=False,
63
- subpaths: List[str] = None):
64
- """
65
- Zip Directory
66
- Will ignore .gitignore files
67
-
68
- :param directory: the directory to zip.
69
- :param zip_filename: the name of the zipfile
70
- :param ignore_max_file_size: ignore the limitation on the zip file size
71
- :param list[str] subpaths: paths to include in the final zip (relative path).
72
- :return: None
73
- """
74
- # default path
75
- if directory is None:
76
- directory = os.getcwd()
77
- # check if directory
78
- assert os.path.isdir(directory), '[ERROR] Directory does not exists: %s' % directory
79
-
80
- if '.gitignore' in os.listdir(directory):
81
- with open(os.path.join(directory, '.gitignore')) as f:
82
- spec_src = f.read()
83
- else:
84
- spec_src = ''
85
- ignore_lines = spec_src.splitlines() + ['.git', '.dataloop']
86
- spec = pathspec.PathSpec.from_lines(pathspec.patterns.GitWildMatchPattern, ignore_lines)
87
-
88
- # init zip file
89
- zip_file = zipfile.ZipFile(zip_filename, 'w', zipfile.ZIP_DEFLATED)
90
- try:
91
- for root, dirs, files in os.walk(directory):
92
- for file in files:
93
- filepath = os.path.join(root, file)
94
- if not spec.match_file(os.path.relpath(filepath, directory)) \
95
- and Zipping.__check_filepath(os.path.relpath(filepath, directory), subpaths):
96
- Zipping.__add_to_zip_file(directory, filepath, ignore_max_file_size, zip_file)
97
- finally:
98
- zip_file.close()
99
-
100
- @staticmethod
101
- def __check_filepath(filepath: str, paths: List[str]):
102
- """
103
- Checks whether a specific file is inside one of the subdirectories
104
- """
105
- return any(filepath.startswith(directory) for directory in paths)
106
-
107
- @staticmethod
108
- def __add_to_zip_file(directory, filepath, ignore_max_file_size, zip_file):
109
- zip_file.write(filepath, arcname=os.path.relpath(filepath, directory))
110
- if not ignore_max_file_size:
111
- if np.sum([f.file_size for f in list(zip_file.NameToInfo.values())]) > MAX_ZIP_FILE:
112
- logger.error('Failed zipping in file: {}'.format(filepath))
113
- raise ValueError(
114
- 'Zip file cant be over 100MB. '
115
- 'Please verify that only code is being uploaded or '
116
- 'add files to .gitignore so they wont be zipped and uploaded as code.')
117
-
118
- @staticmethod
119
- def unzip_directory(zip_filename, to_directory=None):
120
- with zipfile.ZipFile(zip_filename) as zipdata:
121
- zipinfos = zipdata.infolist()
122
- # iterate through each file
123
- for zipinfo in zipinfos:
124
- # encode the file names
125
- # zip package make decode by cp437 for file that have name that not ascii
126
- # this happen when the flag_bits be different than 0
127
- # so we encode the name back
128
- if not zipinfo.flag_bits:
129
- zipinfo.filename = zipinfo.filename.encode('cp437').decode('utf-8')
130
- zipdata.extract(zipinfo, to_directory)
1
+ import logging
2
+ import os
3
+ import zipfile
4
+ from typing import List
5
+
6
+ import numpy as np
7
+ import pathspec
8
+
9
+ logger = logging.getLogger(name='dtlpy')
10
+
11
+ MAX_ZIP_FILE = 100e6 # 100MB
12
+
13
+
14
+ class Zipping:
15
+ def __init__(self):
16
+ pass
17
+
18
+ @staticmethod
19
+ def zip_directory(zip_filename, directory=None, ignore_max_file_size=False, ignore_directories: List[str] = None):
20
+ """
21
+ Zip Directory
22
+ Will ignore .gitignore files
23
+
24
+ :param directory: the directory to zip
25
+ :param zip_filename: the name of the zipfile.
26
+ :param ignore_max_file_size: ignore the limitation on the zip file size
27
+ :param list[str] ignore_directories: directories to ignore.
28
+ :return: None
29
+ """
30
+ # default path
31
+ if directory is None:
32
+ directory = os.getcwd()
33
+ # check if directory
34
+ assert os.path.isdir(directory), '[ERROR] Directory does not exists: {}'.format(directory)
35
+
36
+ if '.gitignore' in os.listdir(directory):
37
+ with open(os.path.join(directory, '.gitignore')) as f:
38
+ spec_src = f.read()
39
+ else:
40
+ spec_src = ''
41
+ ignore_lines = spec_src.splitlines() + ['.git', '.dataloop']
42
+ if ignore_directories is not None:
43
+ ignore_lines += ignore_directories
44
+ spec = pathspec.PathSpec.from_lines(pathspec.patterns.GitWildMatchPattern, ignore_lines)
45
+
46
+ # init zip file
47
+ zip_file = zipfile.ZipFile(zip_filename, 'w', zipfile.ZIP_DEFLATED)
48
+ try:
49
+ for root, dirs, files in os.walk(directory):
50
+ # remove dirs to avoid going file by file
51
+ for d in dirs:
52
+ if spec.match_file(os.path.relpath(os.path.join(root, d), directory)):
53
+ dirs.remove(d)
54
+ for file in files:
55
+ filepath = os.path.join(root, file)
56
+ if not spec.match_file(os.path.relpath(filepath, directory)):
57
+ Zipping.__add_to_zip_file(directory, filepath, ignore_max_file_size, zip_file)
58
+ finally:
59
+ zip_file.close()
60
+
61
+ @staticmethod
62
+ def zip_directory_inclusive(zip_filename, directory=None, ignore_max_file_size=False,
63
+ subpaths: List[str] = None):
64
+ """
65
+ Zip Directory
66
+ Will ignore .gitignore files
67
+
68
+ :param directory: the directory to zip.
69
+ :param zip_filename: the name of the zipfile
70
+ :param ignore_max_file_size: ignore the limitation on the zip file size
71
+ :param list[str] subpaths: paths to include in the final zip (relative path).
72
+ :return: None
73
+ """
74
+ # default path
75
+ if directory is None:
76
+ directory = os.getcwd()
77
+ # check if directory
78
+ assert os.path.isdir(directory), '[ERROR] Directory does not exists: %s' % directory
79
+
80
+ if '.gitignore' in os.listdir(directory):
81
+ with open(os.path.join(directory, '.gitignore')) as f:
82
+ spec_src = f.read()
83
+ else:
84
+ spec_src = ''
85
+ ignore_lines = spec_src.splitlines() + ['.git', '.dataloop']
86
+ spec = pathspec.PathSpec.from_lines(pathspec.patterns.GitWildMatchPattern, ignore_lines)
87
+
88
+ # init zip file
89
+ zip_file = zipfile.ZipFile(zip_filename, 'w', zipfile.ZIP_DEFLATED)
90
+ try:
91
+ for root, dirs, files in os.walk(directory):
92
+ for file in files:
93
+ filepath = os.path.join(root, file)
94
+ if not spec.match_file(os.path.relpath(filepath, directory)) \
95
+ and Zipping.__check_filepath(os.path.relpath(filepath, directory), subpaths):
96
+ Zipping.__add_to_zip_file(directory, filepath, ignore_max_file_size, zip_file)
97
+ finally:
98
+ zip_file.close()
99
+
100
+ @staticmethod
101
+ def __check_filepath(filepath: str, paths: List[str]):
102
+ """
103
+ Checks whether a specific file is inside one of the subdirectories
104
+ """
105
+ return any(filepath.startswith(directory) for directory in paths)
106
+
107
+ @staticmethod
108
+ def __add_to_zip_file(directory, filepath, ignore_max_file_size, zip_file):
109
+ zip_file.write(filepath, arcname=os.path.relpath(filepath, directory))
110
+ if not ignore_max_file_size:
111
+ if np.sum([f.file_size for f in list(zip_file.NameToInfo.values())]) > MAX_ZIP_FILE:
112
+ logger.error('Failed zipping in file: {}'.format(filepath))
113
+ raise ValueError(
114
+ 'Zip file cant be over 100MB. '
115
+ 'Please verify that only code is being uploaded or '
116
+ 'add files to .gitignore so they wont be zipped and uploaded as code.')
117
+
118
+ @staticmethod
119
+ def unzip_directory(zip_filename, to_directory=None):
120
+ with zipfile.ZipFile(zip_filename) as zipdata:
121
+ zipinfos = zipdata.infolist()
122
+ # iterate through each file
123
+ for zipinfo in zipinfos:
124
+ # encode the file names
125
+ # zip package make decode by cp437 for file that have name that not ascii
126
+ # this happen when the flag_bits be different than 0
127
+ # so we encode the name back
128
+ if not zipinfo.flag_bits:
129
+ zipinfo.filename = zipinfo.filename.encode('cp437').decode('utf-8')
130
+ zipdata.extract(zipinfo, to_directory)
dtlpy/ml/__init__.py CHANGED
@@ -1,20 +1,20 @@
1
- #! /usr/bin/env python3
2
- # This file is part of DTLPY.
3
- #
4
- # DTLPY is free software: you can redistribute it and/or modify
5
- # it under the terms of the GNU General Public License as published by
6
- # the Free Software Foundation, either version 3 of the License, or
7
- # (at your option) any later version.
8
- #
9
- # DTLPY is distributed in the hope that it will be useful,
10
- # but WITHOUT ANY WARRANTY; without even the implied warranty of
11
- # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
12
- # GNU General Public License for more details.
13
- #
14
- # You should have received a copy of the GNU General Public License
15
- # along with DTLPY. If not, see <http://www.gnu.org/licenses/>.
16
- from .base_model_adapter import BaseModelAdapter
17
-
18
- from . import metrics
19
- from . import predictions_utils
20
- from . import train_utils
1
+ #! /usr/bin/env python3
2
+ # This file is part of DTLPY.
3
+ #
4
+ # DTLPY is free software: you can redistribute it and/or modify
5
+ # it under the terms of the GNU General Public License as published by
6
+ # the Free Software Foundation, either version 3 of the License, or
7
+ # (at your option) any later version.
8
+ #
9
+ # DTLPY is distributed in the hope that it will be useful,
10
+ # but WITHOUT ANY WARRANTY; without even the implied warranty of
11
+ # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
12
+ # GNU General Public License for more details.
13
+ #
14
+ # You should have received a copy of the GNU General Public License
15
+ # along with DTLPY. If not, see <http://www.gnu.org/licenses/>.
16
+ from .base_model_adapter import BaseModelAdapter
17
+
18
+ from . import metrics
19
+ from . import predictions_utils
20
+ from . import train_utils
@@ -1,28 +1,28 @@
1
- from abc import ABC
2
-
3
- from base_model_adapter import BaseModelAdapter
4
- from .. import entities
5
-
6
-
7
- class BaseFeatureExtractorAdapter(BaseModelAdapter, ABC):
8
- def __int__(self, model_entity: entities.Model = None):
9
- super().__init__(model_entity)
10
-
11
- def extract_features(self, batch: list, **kwargs):
12
- """ Runs inference with the model, but does not predict. Instead, extracts features for the input batch.
13
-
14
- Virtual method - need to implement
15
-
16
- :param batch: `list` a list containing a batch of items whose features will be extracted
17
- """
18
- raise NotImplementedError("Please implement 'extract_features' method in {}".format(self.__class__.__name__))
19
-
20
- def extract_dataset_features(self, dataset: entities.Dataset, **kwargs):
21
- """ Runs inference to extract features for all items in a dataset.
22
-
23
- Virtual method - need to implement
24
-
25
- :param dataset: `entities.Dataset` dataset entity whose items will have their features extracted
26
- """
27
- raise NotImplementedError("Please implement 'extract_dataset_features' method in "
1
+ from abc import ABC
2
+
3
+ from base_model_adapter import BaseModelAdapter
4
+ from .. import entities
5
+
6
+
7
+ class BaseFeatureExtractorAdapter(BaseModelAdapter, ABC):
8
+ def __int__(self, model_entity: entities.Model = None):
9
+ super().__init__(model_entity)
10
+
11
+ def extract_features(self, batch: list, **kwargs):
12
+ """ Runs inference with the model, but does not predict. Instead, extracts features for the input batch.
13
+
14
+ Virtual method - need to implement
15
+
16
+ :param batch: `list` a list containing a batch of items whose features will be extracted
17
+ """
18
+ raise NotImplementedError("Please implement 'extract_features' method in {}".format(self.__class__.__name__))
19
+
20
+ def extract_dataset_features(self, dataset: entities.Dataset, **kwargs):
21
+ """ Runs inference to extract features for all items in a dataset.
22
+
23
+ Virtual method - need to implement
24
+
25
+ :param dataset: `entities.Dataset` dataset entity whose items will have their features extracted
26
+ """
27
+ raise NotImplementedError("Please implement 'extract_dataset_features' method in "
28
28
  "{}".format(self.__class__.__name__))