mindstudio-probe 1.0.3__py3-none-any.whl → 1.1.0__py3-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (278) hide show
  1. {mindstudio_probe-1.0.3.dist-info → mindstudio_probe-1.1.0.dist-info}/LICENSE +201 -201
  2. {mindstudio_probe-1.0.3.dist-info → mindstudio_probe-1.1.0.dist-info}/METADATA +36 -34
  3. mindstudio_probe-1.1.0.dist-info/RECORD +287 -0
  4. {mindstudio_probe-1.0.3.dist-info → mindstudio_probe-1.1.0.dist-info}/WHEEL +1 -1
  5. {mindstudio_probe-1.0.3.dist-info → mindstudio_probe-1.1.0.dist-info}/entry_points.txt +1 -0
  6. msprobe/README.md +131 -237
  7. msprobe/__init__.py +16 -1
  8. msprobe/{config/config.json → config.json} +47 -49
  9. msprobe/core/advisor/advisor.py +124 -124
  10. msprobe/core/advisor/advisor_const.py +58 -59
  11. msprobe/core/advisor/advisor_result.py +58 -58
  12. msprobe/core/common/const.py +402 -318
  13. msprobe/core/common/exceptions.py +99 -99
  14. msprobe/core/common/{file_check.py → file_utils.py} +523 -283
  15. msprobe/core/common/inplace_op_checker.py +38 -0
  16. msprobe/core/common/inplace_ops.yaml +251 -0
  17. msprobe/core/common/log.py +86 -69
  18. msprobe/core/common/utils.py +371 -616
  19. msprobe/core/common_config.py +78 -71
  20. msprobe/core/compare/acc_compare.py +472 -298
  21. msprobe/core/compare/check.py +180 -95
  22. msprobe/core/compare/compare_cli.py +69 -49
  23. msprobe/core/compare/highlight.py +259 -222
  24. msprobe/core/compare/multiprocessing_compute.py +174 -149
  25. msprobe/core/compare/npy_compare.py +310 -295
  26. msprobe/core/compare/utils.py +464 -429
  27. msprobe/core/data_dump/data_collector.py +153 -144
  28. msprobe/core/data_dump/data_processor/base.py +337 -293
  29. msprobe/core/data_dump/data_processor/factory.py +76 -59
  30. msprobe/core/data_dump/data_processor/mindspore_processor.py +192 -198
  31. msprobe/core/data_dump/data_processor/pytorch_processor.py +383 -389
  32. msprobe/core/data_dump/json_writer.py +117 -116
  33. msprobe/core/data_dump/scope.py +194 -178
  34. msprobe/core/grad_probe/constant.py +74 -70
  35. msprobe/core/grad_probe/grad_compare.py +170 -175
  36. msprobe/core/grad_probe/utils.py +77 -52
  37. msprobe/docs/01.installation.md +99 -0
  38. msprobe/docs/02.config_introduction.md +137 -0
  39. msprobe/docs/03.config_examples.md +237 -0
  40. msprobe/docs/04.acl_config_examples.md +78 -0
  41. msprobe/docs/05.data_dump_PyTorch.md +326 -0
  42. msprobe/docs/06.data_dump_MindSpore.md +285 -0
  43. msprobe/docs/07.accuracy_checker_PyTorch.md +297 -0
  44. msprobe/docs/08.accuracy_checker_online_PyTorch.md +238 -0
  45. msprobe/docs/09.accuracy_checker_MindSpore.md +68 -0
  46. msprobe/docs/10.accuracy_compare_PyTorch.md +327 -0
  47. msprobe/docs/11.accuracy_compare_MindSpore.md +333 -0
  48. msprobe/docs/12.overflow_check_PyTorch.md +79 -0
  49. msprobe/docs/13.overflow_check_MindSpore.md +31 -0
  50. msprobe/{pytorch/doc/parse_tool.md → docs/14.data_parse_PyTorch.md} +283 -286
  51. msprobe/docs/15.free_benchmarking_PyTorch.md +170 -0
  52. msprobe/docs/16.free_benchmarking_MindSpore.md +140 -0
  53. msprobe/{doc/grad_probe/grad_probe.md → docs/17.grad_probe.md} +205 -207
  54. msprobe/{pytorch/doc//321/205/320/254/320/270/321/207/342/225/221/342/224/220/321/207/342/226/223/342/225/233/321/205/342/225/221/320/266/321/206/320/277/320/244/321/205/320/277/342/225/243.md → docs/18.online_dispatch.md} +89 -90
  55. msprobe/docs/FAQ.md +189 -0
  56. msprobe/docs/S02.report_free_benchmarking_validation_performance_baseline.md +146 -0
  57. msprobe/docs/img/free_benchmark_framework.png +0 -0
  58. msprobe/docs/img/ms_dump.png +0 -0
  59. msprobe/docs/img/ms_layer.png +0 -0
  60. msprobe/docs/img/pt_dump.png +0 -0
  61. msprobe/mindspore/__init__.py +2 -1
  62. msprobe/mindspore/api_accuracy_checker/api_accuracy_checker.py +278 -245
  63. msprobe/mindspore/api_accuracy_checker/api_info.py +76 -69
  64. msprobe/mindspore/api_accuracy_checker/api_runner.py +155 -151
  65. msprobe/mindspore/api_accuracy_checker/base_compare_algorithm.py +196 -196
  66. msprobe/mindspore/api_accuracy_checker/cmd_parser.py +6 -0
  67. msprobe/mindspore/api_accuracy_checker/compute_element.py +238 -223
  68. msprobe/mindspore/api_accuracy_checker/main.py +8 -15
  69. msprobe/mindspore/api_accuracy_checker/type_mapping.py +113 -113
  70. msprobe/mindspore/api_accuracy_checker/utils.py +79 -62
  71. msprobe/mindspore/cell_processor.py +58 -34
  72. msprobe/mindspore/common/const.py +108 -87
  73. msprobe/mindspore/common/log.py +37 -37
  74. msprobe/mindspore/common/utils.py +97 -57
  75. msprobe/mindspore/compare/distributed_compare.py +62 -75
  76. msprobe/mindspore/compare/layer_mapping.py +146 -0
  77. msprobe/mindspore/compare/modify_mapping.py +107 -0
  78. msprobe/mindspore/compare/ms_compare.py +357 -117
  79. msprobe/mindspore/compare/ms_graph_compare.py +364 -317
  80. msprobe/mindspore/compare/ms_to_pt_api.yaml +399 -399
  81. msprobe/mindspore/debugger/debugger_config.py +69 -74
  82. msprobe/mindspore/debugger/precision_debugger.py +150 -107
  83. msprobe/mindspore/dump/dump_tool_factory.py +50 -35
  84. msprobe/mindspore/dump/hook_cell/api_registry.py +128 -104
  85. msprobe/mindspore/dump/hook_cell/hook_cell.py +55 -53
  86. msprobe/mindspore/dump/hook_cell/primitive_hooks.py +206 -0
  87. msprobe/mindspore/dump/hook_cell/support_wrap_ops.yaml +994 -925
  88. msprobe/mindspore/dump/hook_cell/wrap_api.py +121 -0
  89. msprobe/mindspore/dump/jit_dump.py +96 -56
  90. msprobe/mindspore/dump/kernel_graph_dump.py +75 -60
  91. msprobe/mindspore/dump/kernel_kbyk_dump.py +79 -65
  92. msprobe/mindspore/free_benchmark/api_pynative_self_check.py +131 -116
  93. msprobe/mindspore/free_benchmark/common/config.py +27 -12
  94. msprobe/mindspore/free_benchmark/common/handler_params.py +32 -17
  95. msprobe/mindspore/free_benchmark/common/utils.py +85 -71
  96. msprobe/mindspore/free_benchmark/data/support_wrap_ops.yaml +842 -842
  97. msprobe/mindspore/free_benchmark/decorator/dec_forward.py +57 -42
  98. msprobe/mindspore/free_benchmark/decorator/decorator_factory.py +122 -107
  99. msprobe/mindspore/free_benchmark/handler/base_handler.py +105 -90
  100. msprobe/mindspore/free_benchmark/handler/check_handler.py +56 -41
  101. msprobe/mindspore/free_benchmark/handler/fix_handler.py +51 -36
  102. msprobe/mindspore/free_benchmark/handler/handler_factory.py +36 -21
  103. msprobe/mindspore/free_benchmark/perturbation/add_noise.py +82 -67
  104. msprobe/mindspore/free_benchmark/perturbation/base_perturbation.py +36 -21
  105. msprobe/mindspore/free_benchmark/perturbation/bit_noise.py +78 -63
  106. msprobe/mindspore/free_benchmark/perturbation/exchange_value.py +77 -0
  107. msprobe/mindspore/free_benchmark/perturbation/improve_precision.py +49 -34
  108. msprobe/mindspore/free_benchmark/perturbation/no_change.py +27 -12
  109. msprobe/mindspore/free_benchmark/perturbation/perturbation_factory.py +44 -27
  110. msprobe/mindspore/free_benchmark/self_check_tool_factory.py +48 -33
  111. msprobe/mindspore/grad_probe/global_context.py +100 -91
  112. msprobe/mindspore/grad_probe/grad_analyzer.py +231 -231
  113. msprobe/mindspore/grad_probe/grad_monitor.py +27 -27
  114. msprobe/mindspore/grad_probe/grad_stat_csv.py +131 -131
  115. msprobe/mindspore/grad_probe/hook.py +94 -92
  116. msprobe/mindspore/grad_probe/utils.py +29 -28
  117. msprobe/mindspore/ms_config.py +128 -126
  118. msprobe/mindspore/overflow_check/kernel_graph_overflow_check.py +60 -45
  119. msprobe/mindspore/overflow_check/overflow_check_tool_factory.py +49 -34
  120. msprobe/mindspore/runtime.py +4 -4
  121. msprobe/mindspore/service.py +297 -354
  122. msprobe/mindspore/task_handler_factory.py +24 -24
  123. msprobe/msprobe.py +105 -107
  124. msprobe/pytorch/__init__.py +23 -4
  125. msprobe/pytorch/api_accuracy_checker/common/config.py +70 -55
  126. msprobe/pytorch/api_accuracy_checker/common/utils.py +246 -165
  127. msprobe/pytorch/api_accuracy_checker/compare/algorithm.py +230 -213
  128. msprobe/pytorch/api_accuracy_checker/compare/api_precision_compare.py +632 -581
  129. msprobe/pytorch/api_accuracy_checker/compare/api_precision_standard.yaml +132 -132
  130. msprobe/pytorch/api_accuracy_checker/compare/api_precision_threshold.yaml +390 -390
  131. msprobe/pytorch/api_accuracy_checker/compare/compare.py +416 -381
  132. msprobe/pytorch/api_accuracy_checker/compare/compare_column.py +90 -73
  133. msprobe/pytorch/api_accuracy_checker/compare/compare_utils.py +265 -244
  134. msprobe/pytorch/api_accuracy_checker/config.yaml +10 -10
  135. msprobe/pytorch/api_accuracy_checker/run_ut/data_generate.py +370 -332
  136. msprobe/pytorch/api_accuracy_checker/run_ut/multi_run_ut.py +221 -199
  137. msprobe/pytorch/api_accuracy_checker/run_ut/run_overflow_check.py +150 -134
  138. msprobe/pytorch/api_accuracy_checker/run_ut/run_ut.py +518 -581
  139. msprobe/pytorch/api_accuracy_checker/run_ut/run_ut_utils.py +213 -74
  140. msprobe/pytorch/api_accuracy_checker/run_ut/torch_ut_setting.json +7 -4
  141. msprobe/pytorch/api_accuracy_checker/tensor_transport_layer/attl.py +218 -202
  142. msprobe/pytorch/api_accuracy_checker/tensor_transport_layer/client.py +370 -324
  143. msprobe/pytorch/api_accuracy_checker/tensor_transport_layer/device_dispatch.py +227 -204
  144. msprobe/pytorch/api_accuracy_checker/tensor_transport_layer/dump_dispatch.py +110 -0
  145. msprobe/pytorch/api_accuracy_checker/tensor_transport_layer/server.py +244 -218
  146. msprobe/pytorch/api_accuracy_checker/tensor_transport_layer/torch_ops_config.yaml +63 -0
  147. msprobe/pytorch/api_accuracy_checker/tensor_transport_layer/utils.py +44 -0
  148. msprobe/pytorch/bench_functions/__init__.py +30 -15
  149. msprobe/pytorch/bench_functions/apply_adam_w.py +43 -28
  150. msprobe/pytorch/bench_functions/confusion_transpose.py +34 -19
  151. msprobe/pytorch/bench_functions/fast_gelu.py +70 -55
  152. msprobe/pytorch/bench_functions/layer_norm_eval.py +21 -6
  153. msprobe/pytorch/bench_functions/linear.py +27 -12
  154. msprobe/pytorch/bench_functions/matmul_backward.py +63 -48
  155. msprobe/pytorch/bench_functions/npu_fusion_attention.py +538 -421
  156. msprobe/pytorch/bench_functions/rms_norm.py +30 -15
  157. msprobe/pytorch/bench_functions/rotary_mul.py +71 -52
  158. msprobe/pytorch/bench_functions/scaled_mask_softmax.py +41 -26
  159. msprobe/pytorch/bench_functions/swiglu.py +70 -55
  160. msprobe/pytorch/common/__init__.py +17 -2
  161. msprobe/pytorch/common/compare_script.template +14 -14
  162. msprobe/pytorch/common/log.py +33 -32
  163. msprobe/pytorch/common/parse_json.py +54 -39
  164. msprobe/pytorch/common/utils.py +310 -300
  165. msprobe/pytorch/compare/distributed_compare.py +66 -66
  166. msprobe/pytorch/compare/mapping.yaml +607 -607
  167. msprobe/pytorch/compare/match.py +49 -33
  168. msprobe/pytorch/compare/pt_compare.py +82 -40
  169. msprobe/pytorch/debugger/debugger_config.py +108 -95
  170. msprobe/pytorch/debugger/precision_debugger.py +173 -125
  171. msprobe/pytorch/free_benchmark/__init__.py +23 -8
  172. msprobe/pytorch/free_benchmark/common/constant.py +70 -70
  173. msprobe/pytorch/free_benchmark/common/counter.py +71 -71
  174. msprobe/pytorch/free_benchmark/common/enums.py +65 -37
  175. msprobe/pytorch/free_benchmark/common/params.py +144 -129
  176. msprobe/pytorch/free_benchmark/common/utils.py +118 -102
  177. msprobe/pytorch/free_benchmark/compare/grad_saver.py +200 -179
  178. msprobe/pytorch/free_benchmark/compare/single_benchmark.py +119 -104
  179. msprobe/pytorch/free_benchmark/main.py +120 -105
  180. msprobe/pytorch/free_benchmark/perturbed_layers/base_layer.py +28 -13
  181. msprobe/pytorch/free_benchmark/perturbed_layers/layer_factory.py +56 -41
  182. msprobe/pytorch/free_benchmark/perturbed_layers/npu/add_noise.py +105 -90
  183. msprobe/pytorch/free_benchmark/perturbed_layers/npu/bit_noise.py +119 -104
  184. msprobe/pytorch/free_benchmark/perturbed_layers/npu/change_value.py +87 -63
  185. msprobe/pytorch/free_benchmark/perturbed_layers/npu/improve_precision.py +83 -68
  186. msprobe/pytorch/free_benchmark/perturbed_layers/npu/no_change.py +43 -28
  187. msprobe/pytorch/free_benchmark/perturbed_layers/npu/npu_base_layser.py +60 -45
  188. msprobe/pytorch/free_benchmark/perturbed_layers/run_cpu.py +34 -19
  189. msprobe/pytorch/free_benchmark/result_handlers/base_handler.py +256 -217
  190. msprobe/pytorch/free_benchmark/result_handlers/check_handler.py +54 -39
  191. msprobe/pytorch/free_benchmark/result_handlers/fix_handler.py +38 -23
  192. msprobe/pytorch/free_benchmark/result_handlers/handler_factory.py +45 -30
  193. msprobe/pytorch/free_benchmark/result_handlers/preheat_handler.py +185 -170
  194. msprobe/pytorch/function_factory.py +91 -75
  195. msprobe/pytorch/functional/module_dump.py +84 -0
  196. msprobe/pytorch/grad_probe/grad_monitor.py +91 -90
  197. msprobe/pytorch/grad_probe/grad_stat_csv.py +128 -128
  198. msprobe/pytorch/hook_module/__init__.py +16 -1
  199. msprobe/pytorch/hook_module/api_registry.py +166 -161
  200. msprobe/pytorch/hook_module/hook_module.py +118 -120
  201. msprobe/pytorch/hook_module/support_wrap_ops.yaml +1879 -1877
  202. msprobe/pytorch/hook_module/utils.py +28 -29
  203. msprobe/pytorch/hook_module/wrap_aten.py +111 -110
  204. msprobe/pytorch/hook_module/wrap_distributed.py +77 -78
  205. msprobe/pytorch/hook_module/wrap_functional.py +104 -105
  206. msprobe/pytorch/hook_module/wrap_npu_custom.py +85 -84
  207. msprobe/pytorch/hook_module/wrap_tensor.py +69 -71
  208. msprobe/pytorch/hook_module/wrap_torch.py +84 -86
  209. msprobe/pytorch/hook_module/wrap_vf.py +60 -62
  210. msprobe/pytorch/module_processer.py +153 -138
  211. msprobe/pytorch/online_dispatch/__init__.py +20 -20
  212. msprobe/pytorch/online_dispatch/compare.py +235 -236
  213. msprobe/pytorch/online_dispatch/dispatch.py +271 -271
  214. msprobe/pytorch/online_dispatch/dump_compare.py +155 -156
  215. msprobe/pytorch/online_dispatch/single_compare.py +391 -391
  216. msprobe/pytorch/online_dispatch/torch_ops_config.yaml +57 -49
  217. msprobe/pytorch/online_dispatch/utils.py +127 -146
  218. msprobe/pytorch/parse.py +19 -4
  219. msprobe/pytorch/parse_tool/cli.py +31 -32
  220. msprobe/pytorch/parse_tool/lib/compare.py +259 -271
  221. msprobe/pytorch/parse_tool/lib/config.py +52 -52
  222. msprobe/pytorch/parse_tool/lib/file_desc.py +31 -31
  223. msprobe/pytorch/parse_tool/lib/interactive_cli.py +102 -102
  224. msprobe/pytorch/parse_tool/lib/parse_exception.py +54 -54
  225. msprobe/pytorch/parse_tool/lib/parse_tool.py +161 -158
  226. msprobe/pytorch/parse_tool/lib/utils.py +320 -321
  227. msprobe/pytorch/parse_tool/lib/visualization.py +85 -91
  228. msprobe/pytorch/pt_config.py +317 -187
  229. msprobe/pytorch/service.py +311 -252
  230. mindstudio_probe-1.0.3.dist-info/RECORD +0 -272
  231. msprobe/config/README.md +0 -539
  232. msprobe/mindspore/doc/compare.md +0 -58
  233. msprobe/mindspore/doc/dump.md +0 -217
  234. msprobe/mindspore/dump/hook_cell/wrap_functional.py +0 -91
  235. msprobe/mindspore/dump/hook_cell/wrap_tensor.py +0 -63
  236. msprobe/pytorch/api_accuracy_checker/tensor_transport_layer/ssl_config.py +0 -10
  237. msprobe/pytorch/doc/FAQ.md +0 -193
  238. msprobe/pytorch/doc/api_accuracy_checker.md +0 -313
  239. msprobe/pytorch/doc/api_accuracy_checker_online.md +0 -187
  240. msprobe/pytorch/doc/dump.md +0 -260
  241. msprobe/pytorch/doc/msprobe/321/207/342/226/223/342/225/233/321/205/342/225/221/320/266/321/205/342/225/226/320/265/321/205/320/225/342/225/226/321/206/320/245/342/226/221/321/206/320/235/320/276dump/321/206/320/260/320/227/321/205/320/227/320/226/321/206/320/220/320/267/321/210/320/223/342/225/234/321/205/320/257/342/225/221/321/207/342/225/221/342/224/220/321/206/320/232/320/265/321/205/320/241/320/232.md +0 -182
  242. msprobe/pytorch/doc/ptdbg_ascend_compare.md +0 -240
  243. msprobe/pytorch/doc/ptdbg_ascend_overview.md +0 -68
  244. msprobe/pytorch/doc/ptdbg_ascend_quickstart.md +0 -381
  245. msprobe/pytorch/doc/run_overflow_check.md +0 -25
  246. msprobe/pytorch/doc//321/206/320/247/320/260/321/206/320/260/320/227/321/206/320/255/320/226/321/205/342/225/226/320/265/321/205/320/225/342/225/226/321/205/320/254/342/225/221/321/206/320/251/320/277/321/211/320/272/320/234/321/210/320/277/320/221/321/205/320/242/320/234/321/206/320/220/320/267/321/210/320/223/342/225/234/321/205/320/257/342/225/221/321/207/342/225/221/342/224/220/321/206/320/232/320/265/321/205/320/241/320/232.md +0 -151
  247. msprobe/pytorch/functional/data_processor.py +0 -0
  248. msprobe/pytorch/functional/dump_module.py +0 -39
  249. {mindstudio_probe-1.0.3.dist-info → mindstudio_probe-1.1.0.dist-info}/top_level.txt +0 -0
  250. /msprobe/{pytorch/doc → docs}/img/BLOOM-7B_1.png +0 -0
  251. /msprobe/{pytorch/doc → docs}/img/BLOOM-7B_2.png +0 -0
  252. /msprobe/{pytorch/doc → docs}/img/BLOOM-7B_3.png +0 -0
  253. /msprobe/{pytorch/doc → docs}/img/BLOOM-7B_4.png +0 -0
  254. /msprobe/{pytorch/doc → docs}/img/GPT-3_1.png +0 -0
  255. /msprobe/{pytorch/doc → docs}/img/GPT-3_2.png +0 -0
  256. /msprobe/{pytorch/doc → docs}/img/GPT-3_3.png +0 -0
  257. /msprobe/{pytorch/doc → docs}/img/GPT-3_4.png +0 -0
  258. /msprobe/{pytorch/doc → docs}/img/GPT-3_5.png +0 -0
  259. /msprobe/{pytorch/doc → docs}/img/GPT-3_6.png +0 -0
  260. /msprobe/{pytorch/doc → docs}/img/GPT-3_7.png +0 -0
  261. /msprobe/{pytorch/doc → docs}/img/GPT-3_8.png +0 -0
  262. /msprobe/{pytorch/doc → docs}/img/YOLOV5S_1.png +0 -0
  263. /msprobe/{pytorch/doc → docs}/img/YOLOV5S_2.png +0 -0
  264. /msprobe/{pytorch/doc → docs}/img/accuracy_checking_details.png +0 -0
  265. /msprobe/{pytorch/doc → docs}/img/accuracy_checking_result.png +0 -0
  266. /msprobe/{pytorch/doc → docs}/img/api_precision_compare_details.png +0 -0
  267. /msprobe/{pytorch/doc → docs}/img/api_precision_compare_result.png +0 -0
  268. /msprobe/{pytorch/doc → docs}/img/auto_analyze_log.png +0 -0
  269. /msprobe/{pytorch/doc → docs}/img/compare_result_pkl.png +0 -0
  270. /msprobe/{pytorch/doc → docs}/img/compare_result_pkl_md5.png.png +0 -0
  271. /msprobe/{pytorch/doc → docs}/img/cpu_info.png +0 -0
  272. /msprobe/{config → docs}/img/free_benchmark.png +0 -0
  273. /msprobe/{doc/grad_probe/img/image-1.png → docs/img/grad_probe_image-1.png} +0 -0
  274. /msprobe/{doc/grad_probe/img/image-2.png → docs/img/grad_probe_image-2.png} +0 -0
  275. /msprobe/{doc/grad_probe/img/image-3.png → docs/img/grad_probe_image-3.png} +0 -0
  276. /msprobe/{doc/grad_probe/img/image-4.png → docs/img/grad_probe_image-4.png} +0 -0
  277. /msprobe/{doc/grad_probe/img/image.png → docs/img/grad_probe_image.png} +0 -0
  278. /msprobe/{pytorch/doc → docs}/img/module_compare.png +0 -0
@@ -1,86 +1,84 @@
1
- #!/usr/bin/env python3
2
- # -*- coding: utf-8 -*-
3
- """
4
- # Copyright (C) 2019-2020. Huawei Technologies Co., Ltd. All rights reserved.
5
- # Licensed under the Apache License, Version 2.0 (the "License");
6
- # you may not use this file except in compliance with the License.
7
- # You may obtain a copy of the License at
8
- #
9
- # http://www.apache.org/licenses/LICENSE-2.0
10
- #
11
- # Unless required by applicable law or agreed to in writing, software
12
- # distributed under the License is distributed on an "AS IS" BASIS,
13
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
14
- # See the License for the specific language governing permissions and
15
- # limitations under the License.
16
- """
17
-
18
- import os
19
- import torch
20
-
21
- from msprobe.pytorch.hook_module.hook_module import HOOKModule
22
- from msprobe.pytorch.common.utils import torch_device_guard
23
- from msprobe.core.common.const import Const
24
- from msprobe.core.common.utils import load_yaml
25
-
26
-
27
- cur_path = os.path.dirname(os.path.realpath(__file__))
28
- yaml_path = os.path.join(cur_path, "support_wrap_ops.yaml")
29
-
30
-
31
- def get_torch_ops():
32
- _torch_ops = []
33
- yaml_data = load_yaml(yaml_path)
34
- wrap_torch_ops = yaml_data.get('torch')
35
- for operation in wrap_torch_ops:
36
- if '.' in operation:
37
- operation_sub_module_name, operation_sub_op = operation.rsplit('.', 1)
38
- operation_sub_module = getattr(torch, operation_sub_module_name)
39
- if operation_sub_op in dir(operation_sub_module):
40
- _torch_ops.append(operation)
41
- else:
42
- if hasattr(torch, operation):
43
- _torch_ops.append(operation)
44
- return set(_torch_ops)
45
-
46
-
47
- TorchOps = {}
48
- for op in get_torch_ops():
49
- if '.' in op:
50
- sub_module_name, sub_op = op.rsplit('.', 1)
51
- sub_module = getattr(torch, sub_module_name)
52
- TorchOps[op] = getattr(sub_module, sub_op)
53
- else:
54
- TorchOps[op] = getattr(torch, op)
55
-
56
-
57
-
58
- class HOOKTorchOP(object):
59
- pass
60
-
61
-
62
- class TorchOPTemplate(HOOKModule):
63
-
64
- def __init__(self, op_name, hook, need_hook=True):
65
- self.op_name_ = op_name
66
- self.prefix_op_name_ = "Torch" + Const.SEP + str(op_name) + Const.SEP
67
- if need_hook:
68
- super().__init__(hook)
69
-
70
- @torch_device_guard
71
- def forward(self, *args, **kwargs):
72
- return TorchOps[str(self.op_name_)](*args, **kwargs)
73
-
74
-
75
- def wrap_torch_op(op_name, hook):
76
-
77
- def torch_op_template(*args, **kwargs):
78
- return TorchOPTemplate(op_name, hook)(*args, **kwargs)
79
-
80
- return torch_op_template
81
-
82
-
83
- def wrap_torch_ops_and_bind(hook):
84
- _torch_ops = get_torch_ops()
85
- for op_name in _torch_ops:
86
- setattr(HOOKTorchOP, "wrap_" + op_name, wrap_torch_op(op_name, hook))
1
+ # Copyright (c) 2024-2024, Huawei Technologies Co., Ltd.
2
+ # All rights reserved.
3
+ #
4
+ # Licensed under the Apache License, Version 2.0 (the "License");
5
+ # you may not use this file except in compliance with the License.
6
+ # You may obtain a copy of the License at
7
+ #
8
+ # http://www.apache.org/licenses/LICENSE-2.0
9
+ #
10
+ # Unless required by applicable law or agreed to in writing, software
11
+ # distributed under the License is distributed on an "AS IS" BASIS,
12
+ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13
+ # See the License for the specific language governing permissions and
14
+ # limitations under the License.
15
+
16
+ import os
17
+ import torch
18
+
19
+ from msprobe.pytorch.hook_module.hook_module import HOOKModule
20
+ from msprobe.pytorch.common.utils import torch_device_guard
21
+ from msprobe.core.common.const import Const
22
+ from msprobe.core.common.file_utils import load_yaml
23
+
24
+
25
+ cur_path = os.path.dirname(os.path.realpath(__file__))
26
+ yaml_path = os.path.join(cur_path, "support_wrap_ops.yaml")
27
+
28
+
29
+ def get_torch_ops():
30
+ _torch_ops = []
31
+ yaml_data = load_yaml(yaml_path)
32
+ wrap_torch_ops = yaml_data.get('torch')
33
+ for operation in wrap_torch_ops:
34
+ if '.' in operation:
35
+ operation_sub_module_name, operation_sub_op = operation.rsplit('.', 1)
36
+ operation_sub_module = getattr(torch, operation_sub_module_name)
37
+ if operation_sub_op in dir(operation_sub_module):
38
+ _torch_ops.append(operation)
39
+ else:
40
+ if hasattr(torch, operation):
41
+ _torch_ops.append(operation)
42
+ return set(_torch_ops)
43
+
44
+
45
+ TorchOps = {}
46
+ for op in get_torch_ops():
47
+ if '.' in op:
48
+ sub_module_name, sub_op = op.rsplit('.', 1)
49
+ sub_module = getattr(torch, sub_module_name)
50
+ TorchOps[op] = getattr(sub_module, sub_op)
51
+ else:
52
+ TorchOps[op] = getattr(torch, op)
53
+
54
+
55
+
56
+ class HOOKTorchOP(object):
57
+ pass
58
+
59
+
60
+ class TorchOPTemplate(HOOKModule):
61
+
62
+ def __init__(self, op_name, hook, need_hook=True):
63
+ self.op_name_ = op_name
64
+ self.prefix_op_name_ = "Torch" + Const.SEP + str(op_name) + Const.SEP
65
+ if need_hook:
66
+ super().__init__(hook)
67
+
68
+ @torch_device_guard
69
+ def forward(self, *args, **kwargs):
70
+ return TorchOps[str(self.op_name_)](*args, **kwargs)
71
+
72
+
73
+ def wrap_torch_op(op_name, hook):
74
+
75
+ def torch_op_template(*args, **kwargs):
76
+ return TorchOPTemplate(op_name, hook)(*args, **kwargs)
77
+
78
+ return torch_op_template
79
+
80
+
81
+ def wrap_torch_ops_and_bind(hook):
82
+ _torch_ops = get_torch_ops()
83
+ for op_name in _torch_ops:
84
+ setattr(HOOKTorchOP, "wrap_" + op_name, wrap_torch_op(op_name, hook))
@@ -1,62 +1,60 @@
1
- #!/usr/bin/env python3
2
- # -*- coding: utf-8 -*-
3
- """
4
- # Copyright (C) 2019-2020. Huawei Technologies Co., Ltd. All rights reserved.
5
- # Licensed under the Apache License, Version 2.0 (the "License");
6
- # you may not use this file except in compliance with the License.
7
- # You may obtain a copy of the License at
8
- #
9
- # http://www.apache.org/licenses/LICENSE-2.0
10
- #
11
- # Unless required by applicable law or agreed to in writing, software
12
- # distributed under the License is distributed on an "AS IS" BASIS,
13
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
14
- # See the License for the specific language governing permissions and
15
- # limitations under the License.
16
- """
17
-
18
- import os
19
- import torch
20
-
21
- from msprobe.core.common.const import Const
22
- from msprobe.core.common.utils import load_yaml
23
- from msprobe.pytorch.hook_module.hook_module import HOOKModule
24
- from msprobe.pytorch.common.utils import torch_device_guard
25
-
26
-
27
- cur_path = os.path.dirname(os.path.realpath(__file__))
28
- yaml_path = os.path.join(cur_path, "support_wrap_ops.yaml")
29
-
30
-
31
- def get_vf_ops():
32
- yaml_data = load_yaml(yaml_path)
33
- wrap_vf_ops = yaml_data.get('_VF')
34
- return wrap_vf_ops
35
-
36
-
37
- class HOOKVfOP(object):
38
- pass
39
-
40
-
41
- class VfOPTemplate(HOOKModule):
42
- def __init__(self, op_name, hook):
43
- self.op_name_ = op_name
44
- self.prefix_op_name_ = "VF" + Const.SEP + str(op_name) + Const.SEP
45
- super().__init__(hook)
46
-
47
- @torch_device_guard
48
- def forward(self, *args, **kwargs):
49
- return getattr(torch._C._VariableFunctionsClass, str(self.op_name_))(*args, **kwargs)
50
-
51
-
52
- def wrap_vf_op(op_name, hook):
53
- def vf_op_template(*args, **kwargs):
54
- return VfOPTemplate(op_name, hook)(*args, **kwargs)
55
-
56
- return vf_op_template
57
-
58
-
59
- def wrap_vf_ops_and_bind(hook):
60
- _vf_ops = get_vf_ops()
61
- for op_name in _vf_ops:
62
- setattr(HOOKVfOP, "wrap_" + op_name, wrap_vf_op(op_name, hook))
1
+ # Copyright (c) 2024-2024, Huawei Technologies Co., Ltd.
2
+ # All rights reserved.
3
+ #
4
+ # Licensed under the Apache License, Version 2.0 (the "License");
5
+ # you may not use this file except in compliance with the License.
6
+ # You may obtain a copy of the License at
7
+ #
8
+ # http://www.apache.org/licenses/LICENSE-2.0
9
+ #
10
+ # Unless required by applicable law or agreed to in writing, software
11
+ # distributed under the License is distributed on an "AS IS" BASIS,
12
+ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13
+ # See the License for the specific language governing permissions and
14
+ # limitations under the License.
15
+
16
+ import os
17
+ import torch
18
+
19
+ from msprobe.core.common.const import Const
20
+ from msprobe.core.common.file_utils import load_yaml
21
+ from msprobe.pytorch.hook_module.hook_module import HOOKModule
22
+ from msprobe.pytorch.common.utils import torch_device_guard
23
+
24
+
25
+ cur_path = os.path.dirname(os.path.realpath(__file__))
26
+ yaml_path = os.path.join(cur_path, "support_wrap_ops.yaml")
27
+
28
+
29
+ def get_vf_ops():
30
+ yaml_data = load_yaml(yaml_path)
31
+ wrap_vf_ops = yaml_data.get('_VF')
32
+ return wrap_vf_ops
33
+
34
+
35
+ class HOOKVfOP(object):
36
+ pass
37
+
38
+
39
+ class VfOPTemplate(HOOKModule):
40
+ def __init__(self, op_name, hook):
41
+ self.op_name_ = op_name
42
+ self.prefix_op_name_ = "VF" + Const.SEP + str(op_name) + Const.SEP
43
+ super().__init__(hook)
44
+
45
+ @torch_device_guard
46
+ def forward(self, *args, **kwargs):
47
+ return getattr(torch._C._VariableFunctionsClass, str(self.op_name_))(*args, **kwargs)
48
+
49
+
50
+ def wrap_vf_op(op_name, hook):
51
+ def vf_op_template(*args, **kwargs):
52
+ return VfOPTemplate(op_name, hook)(*args, **kwargs)
53
+
54
+ return vf_op_template
55
+
56
+
57
+ def wrap_vf_ops_and_bind(hook):
58
+ _vf_ops = get_vf_ops()
59
+ for op_name in _vf_ops:
60
+ setattr(HOOKVfOP, "wrap_" + op_name, wrap_vf_op(op_name, hook))
@@ -1,138 +1,153 @@
1
- from functools import wraps
2
-
3
- import torch
4
- from torch.utils.hooks import BackwardHook
5
-
6
- from msprobe.core.common.const import Const
7
- from msprobe.core.data_dump.scope import ModuleRangeScope
8
- torch_version_above_or_equal_2 = torch.__version__.split('+')[0] >= '2.0'
9
-
10
-
11
- class ModuleProcesser:
12
- module_count = {}
13
- module_stack = []
14
- api_parent_node = ""
15
- module_node = {}
16
-
17
- def __init__(self, scope):
18
- if isinstance(scope, ModuleRangeScope):
19
- self.scope = scope
20
- else:
21
- self.scope = None
22
- BackwardHook.setup_input_hook = ModuleProcesser.clone_return_value(BackwardHook.setup_input_hook)
23
- BackwardHook.setup_output_hook = ModuleProcesser.clone_return_value(BackwardHook.setup_output_hook)
24
- BackwardHook.setup_output_hook = ModuleProcesser.filter_tensor_and_tuple(BackwardHook.setup_output_hook)
25
-
26
- @staticmethod
27
- def filter_tensor_and_tuple(func):
28
- @wraps(func)
29
- def wrap_by_filter_tensor_and_tuple(*args, **kwargs):
30
- # setup_output_hook传入非tensor数据,工具后续dump会报错,处理方式是解析非tensor数据的属性,对tensor属性挂hook
31
- # setup_output_hook定义为setup_output_hook(self, args),因此处理第二个位置参数,即*args[1]
32
- if not isinstance(args[1], (torch.Tensor, tuple)):
33
- for item_str in dir(args[1]):
34
- item = getattr(args[1], item_str)
35
- # 处理tensor或者只包含tensor的元组
36
- if isinstance(item, torch.Tensor) or \
37
- (isinstance(item, tuple) and all(isinstance(x, torch.Tensor) for x in item)):
38
- args_new = (args[0], item)
39
- result = func(*args_new, **kwargs)
40
- setattr(args[1], item_str, result)
41
- return args[1]
42
- return func(*args, **kwargs)
43
-
44
- return wrap_by_filter_tensor_and_tuple
45
-
46
- @staticmethod
47
- def clone_return_value(func):
48
- @wraps(func)
49
- def clone_return_value_func(*args, **kwargs):
50
- result = func(*args, **kwargs)
51
- return ModuleProcesser.clone_if_tensor(result)
52
-
53
- return clone_return_value_func
54
-
55
- @staticmethod
56
- def clone_if_tensor(result):
57
- if isinstance(result, torch.Tensor):
58
- return result.clone()
59
- elif isinstance(result, tuple):
60
- return tuple(ModuleProcesser.clone_if_tensor(x) for x in result)
61
- elif isinstance(result, list):
62
- return list(ModuleProcesser.clone_if_tensor(x) for x in result)
63
- elif isinstance(result, dict):
64
- return {k: ModuleProcesser.clone_if_tensor(v) for k, v in result.items()}
65
- else:
66
- return result
67
-
68
- @staticmethod
69
- def module_count_func(module_name):
70
- if module_name not in ModuleProcesser.module_count:
71
- ModuleProcesser.module_count[module_name] = 0
72
- else:
73
- ModuleProcesser.module_count[module_name] += 1
74
- return ModuleProcesser.module_count[module_name]
75
-
76
- @classmethod
77
- def reset_module_stats(cls):
78
- cls.module_count = {}
79
- cls.module_stack = []
80
- cls.api_parent_node = ""
81
- cls.module_node = {}
82
-
83
- def node_hook(self, name_prefix, start_or_stop, **kwargs):
84
-
85
- def pre_hook(module, input, output=None):
86
- try:
87
- index = ModuleProcesser.module_count_func(name_prefix)
88
- except IndexError as e:
89
- index = None
90
- pass
91
- module.mindstudio_reserved_name = full_name = name_prefix + Const.SEP + str(index)
92
- if self.module_stack:
93
- ModuleProcesser.module_node[full_name] = self.module_stack[-1]
94
- else:
95
- ModuleProcesser.module_node[full_name] = None
96
-
97
- ModuleProcesser.module_stack.append(full_name)
98
- if self.module_stack:
99
- ModuleProcesser.api_parent_node = self.module_stack[-1]
100
- if self.scope:
101
- self.scope.begin_module(full_name)
102
-
103
- def end_hook(module, input, output=None):
104
- if self.module_stack:
105
- ModuleProcesser.module_stack.pop()
106
- if self.module_stack:
107
- ModuleProcesser.api_parent_node = self.module_stack[-1]
108
- else:
109
- ModuleProcesser.api_parent_node = None
110
- if self.scope:
111
- self.scope.end_module(module.mindstudio_reserved_name)
112
-
113
- def backward_hook(module, input, output=None):
114
- try:
115
- index = ModuleProcesser.module_count_func(name_prefix)
116
- except IndexError as e:
117
- index = None
118
- pass
119
- module.mindstudio_reserved_name = full_name = name_prefix + Const.SEP + str(index)
120
- forward_full_name = full_name.replace(Const.BACKWARD, Const.FORWARD)
121
- ModuleProcesser.module_node[full_name] = ModuleProcesser.module_node[forward_full_name].replace(
122
- Const.FORWARD, Const.BACKWARD) if ModuleProcesser.module_node[forward_full_name] else None
123
- ModuleProcesser.api_parent_node = None
124
- if self.scope:
125
- self.scope.begin_module(full_name)
126
-
127
- if torch_version_above_or_equal_2:
128
- if Const.START in start_or_stop:
129
- return pre_hook
130
- else:
131
- return end_hook
132
- else:
133
- if Const.FORWARD in name_prefix and Const.START in start_or_stop:
134
- return pre_hook
135
- elif Const.BACKWARD in name_prefix:
136
- return backward_hook
137
- else:
138
- return end_hook
1
+ # Copyright (c) 2024-2024, Huawei Technologies Co., Ltd.
2
+ # All rights reserved.
3
+ #
4
+ # Licensed under the Apache License, Version 2.0 (the "License");
5
+ # you may not use this file except in compliance with the License.
6
+ # You may obtain a copy of the License at
7
+ #
8
+ # http://www.apache.org/licenses/LICENSE-2.0
9
+ #
10
+ # Unless required by applicable law or agreed to in writing, software
11
+ # distributed under the License is distributed on an "AS IS" BASIS,
12
+ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
13
+ # See the License for the specific language governing permissions and
14
+ # limitations under the License.
15
+
16
+ from functools import wraps
17
+
18
+ import torch
19
+ from msprobe.core.common.const import Const
20
+ from msprobe.core.data_dump.scope import ModuleRangeScope
21
+ from torch.utils.hooks import BackwardHook
22
+
23
+ torch_version_above_or_equal_2 = torch.__version__.split('+')[0] >= '2.0'
24
+
25
+
26
+ class ModuleProcesser:
27
+ module_count = {}
28
+ module_stack = []
29
+ api_parent_node = ""
30
+ module_node = {}
31
+
32
+ def __init__(self, scope):
33
+ if isinstance(scope, ModuleRangeScope):
34
+ self.scope = scope
35
+ else:
36
+ self.scope = None
37
+ BackwardHook.setup_input_hook = ModuleProcesser.clone_return_value(BackwardHook.setup_input_hook)
38
+ BackwardHook.setup_output_hook = ModuleProcesser.clone_return_value(BackwardHook.setup_output_hook)
39
+ BackwardHook.setup_output_hook = ModuleProcesser.filter_tensor_and_tuple(BackwardHook.setup_output_hook)
40
+
41
+ @staticmethod
42
+ def filter_tensor_and_tuple(func):
43
+ @wraps(func)
44
+ def wrap_by_filter_tensor_and_tuple(*args, **kwargs):
45
+ # setup_output_hook传入非tensor数据,工具后续dump会报错,处理方式是解析非tensor数据的属性,对tensor属性挂hook
46
+ # setup_output_hook定义为setup_output_hook(self, args),因此处理第二个位置参数,即*args[1]
47
+ if not isinstance(args[1], (torch.Tensor, tuple)):
48
+ for item_str in dir(args[1]):
49
+ item = getattr(args[1], item_str)
50
+ # 处理tensor或者只包含tensor的元组
51
+ if isinstance(item, torch.Tensor) or \
52
+ (isinstance(item, tuple) and all(isinstance(x, torch.Tensor) for x in item)):
53
+ args_new = (args[0], item)
54
+ result = func(*args_new, **kwargs)
55
+ setattr(args[1], item_str, result)
56
+ return args[1]
57
+ return func(*args, **kwargs)
58
+
59
+ return wrap_by_filter_tensor_and_tuple
60
+
61
+ @staticmethod
62
+ def clone_return_value(func):
63
+ @wraps(func)
64
+ def clone_return_value_func(*args, **kwargs):
65
+ result = func(*args, **kwargs)
66
+ return ModuleProcesser.clone_if_tensor(result)
67
+
68
+ return clone_return_value_func
69
+
70
+ @staticmethod
71
+ def clone_if_tensor(result):
72
+ if isinstance(result, torch.Tensor):
73
+ return result.clone()
74
+ elif isinstance(result, tuple):
75
+ return tuple(ModuleProcesser.clone_if_tensor(x) for x in result)
76
+ elif isinstance(result, list):
77
+ return list(ModuleProcesser.clone_if_tensor(x) for x in result)
78
+ elif isinstance(result, dict):
79
+ return {k: ModuleProcesser.clone_if_tensor(v) for k, v in result.items()}
80
+ else:
81
+ return result
82
+
83
+ @staticmethod
84
+ def module_count_func(module_name):
85
+ if module_name not in ModuleProcesser.module_count:
86
+ ModuleProcesser.module_count[module_name] = 0
87
+ else:
88
+ ModuleProcesser.module_count[module_name] += 1
89
+ return ModuleProcesser.module_count[module_name]
90
+
91
+ @classmethod
92
+ def reset_module_stats(cls):
93
+ cls.module_count = {}
94
+ cls.module_stack = []
95
+ cls.api_parent_node = ""
96
+ cls.module_node = {}
97
+
98
+ def node_hook(self, name_prefix, start_or_stop, **kwargs):
99
+
100
+ def pre_hook(module, input, output=None):
101
+ try:
102
+ index = ModuleProcesser.module_count_func(name_prefix)
103
+ except IndexError as e:
104
+ index = None
105
+ pass
106
+ module.mindstudio_reserved_name = full_name = name_prefix + Const.SEP + str(index)
107
+ if self.module_stack:
108
+ ModuleProcesser.module_node[full_name] = self.module_stack[-1]
109
+ else:
110
+ ModuleProcesser.module_node[full_name] = None
111
+
112
+ ModuleProcesser.module_stack.append(full_name)
113
+ if self.module_stack:
114
+ ModuleProcesser.api_parent_node = self.module_stack[-1]
115
+ if self.scope:
116
+ self.scope.begin_module(full_name)
117
+
118
+ def end_hook(module, input, output=None):
119
+ if self.module_stack:
120
+ ModuleProcesser.module_stack.pop()
121
+ if self.module_stack:
122
+ ModuleProcesser.api_parent_node = self.module_stack[-1]
123
+ else:
124
+ ModuleProcesser.api_parent_node = None
125
+ if self.scope:
126
+ self.scope.end_module(module.mindstudio_reserved_name)
127
+
128
+ def backward_hook(module, input, output=None):
129
+ try:
130
+ index = ModuleProcesser.module_count_func(name_prefix)
131
+ except IndexError as e:
132
+ index = None
133
+ pass
134
+ module.mindstudio_reserved_name = full_name = name_prefix + Const.SEP + str(index)
135
+ forward_full_name = full_name.replace(Const.BACKWARD, Const.FORWARD)
136
+ ModuleProcesser.module_node[full_name] = ModuleProcesser.module_node[forward_full_name].replace(
137
+ Const.FORWARD, Const.BACKWARD) if ModuleProcesser.module_node[forward_full_name] else None
138
+ ModuleProcesser.api_parent_node = None
139
+ if self.scope:
140
+ self.scope.begin_module(full_name)
141
+
142
+ if torch_version_above_or_equal_2:
143
+ if Const.START in start_or_stop:
144
+ return pre_hook
145
+ else:
146
+ return end_hook
147
+ else:
148
+ if Const.FORWARD in name_prefix and Const.START in start_or_stop:
149
+ return pre_hook
150
+ elif Const.BACKWARD in name_prefix:
151
+ return backward_hook
152
+ else:
153
+ return end_hook
@@ -1,20 +1,20 @@
1
- # Copyright (c) 2024-2024 Huawei Technologies Co., Ltd.
2
- #
3
- # Licensed under the Apache License, Version 2.0 (the "License");
4
- # you may not use this file except in compliance with the License.
5
- # You may obtain a copy of the License at
6
- #
7
- # http://www.apache.org/licenses/LICENSE-2.0
8
- #
9
- # Unless required by applicable law or agreed to in writing, software
10
- # distributed under the License is distributed on an "AS IS" BASIS,
11
- # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12
- # See the License for the specific language governing permissions and
13
- # limitations under the License.
14
-
15
- from signal import signal, SIGPIPE, SIG_DFL
16
- from .dispatch import PtdbgDispatch
17
- signal(SIGPIPE, SIG_DFL)
18
-
19
-
20
- __all__ = ["PtdbgDispatch"]
1
+ # Copyright (c) 2024-2024 Huawei Technologies Co., Ltd.
2
+ #
3
+ # Licensed under the Apache License, Version 2.0 (the "License");
4
+ # you may not use this file except in compliance with the License.
5
+ # You may obtain a copy of the License at
6
+ #
7
+ # http://www.apache.org/licenses/LICENSE-2.0
8
+ #
9
+ # Unless required by applicable law or agreed to in writing, software
10
+ # distributed under the License is distributed on an "AS IS" BASIS,
11
+ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12
+ # See the License for the specific language governing permissions and
13
+ # limitations under the License.
14
+
15
+ from signal import signal, SIGPIPE, SIG_DFL
16
+ from .dispatch import PtdbgDispatch
17
+ signal(SIGPIPE, SIG_DFL)
18
+
19
+
20
+ __all__ = ["PtdbgDispatch"]