mindspore 2.5.0__cp310-cp310-win_amd64.whl → 2.6.0__cp310-cp310-win_amd64.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of mindspore might be problematic. Click here for more details.

Files changed (493) hide show
  1. mindspore/.commit_id +1 -1
  2. mindspore/Microsoft.VisualStudio.Telemetry.dll +0 -0
  3. mindspore/Newtonsoft.Json.dll +0 -0
  4. mindspore/__init__.py +6 -4
  5. mindspore/_c_dataengine.cp310-win_amd64.pyd +0 -0
  6. mindspore/_c_expression.cp310-win_amd64.pyd +0 -0
  7. mindspore/_c_mindrecord.cp310-win_amd64.pyd +0 -0
  8. mindspore/_check_jit_forbidden_api.py +3 -0
  9. mindspore/_checkparam.py +3 -33
  10. mindspore/_deprecated/__init__.py +17 -0
  11. mindspore/_deprecated/jit.py +198 -0
  12. mindspore/_extends/builtin_operations.py +1 -1
  13. mindspore/_extends/parse/__init__.py +6 -7
  14. mindspore/_extends/parse/compile_config.py +19 -0
  15. mindspore/_extends/parse/deprecated/deprecated_tensor_method.py +22 -3
  16. mindspore/_extends/parse/jit_fallback_modules/__init__.py +0 -0
  17. mindspore/_extends/parse/jit_fallback_modules/check_utils.py +123 -0
  18. mindspore/_extends/parse/jit_fallback_modules/third_party_modules.py +50 -0
  19. mindspore/_extends/parse/parser.py +25 -194
  20. mindspore/_extends/parse/resources.py +1 -5
  21. mindspore/_extends/parse/standard_method.py +109 -75
  22. mindspore/_extends/pijit/__init__.py +2 -2
  23. mindspore/_extends/pijit/pijit_func_white_list.py +16 -11
  24. mindspore/_extends/pijit/tensor_func_list.py +27 -0
  25. mindspore/_extends/utils.py +1 -1
  26. mindspore/amp.py +4 -4
  27. mindspore/atlprov.dll +0 -0
  28. mindspore/avcodec-59.dll +0 -0
  29. mindspore/avdevice-59.dll +0 -0
  30. mindspore/avfilter-8.dll +0 -0
  31. mindspore/avformat-59.dll +0 -0
  32. mindspore/avutil-57.dll +0 -0
  33. mindspore/boost/__init__.py +2 -2
  34. mindspore/boost/base.py +3 -7
  35. mindspore/boost/boost_cell_wrapper.py +2 -2
  36. mindspore/c1.dll +0 -0
  37. mindspore/c1xx.dll +0 -0
  38. mindspore/c2.dll +0 -0
  39. mindspore/common/__init__.py +4 -3
  40. mindspore/common/_grad_function.py +56 -0
  41. mindspore/common/_pijit_context.py +14 -5
  42. mindspore/common/_register_for_tensor.py +1 -1
  43. mindspore/common/_stub_tensor.py +5 -10
  44. mindspore/common/_tensor_cpp_method.py +1 -1
  45. mindspore/common/_tensor_docs.py +2014 -3386
  46. mindspore/common/api.py +386 -355
  47. mindspore/common/auto_dynamic_shape.py +41 -44
  48. mindspore/common/dtype.py +5 -2
  49. mindspore/common/dump.py +7 -5
  50. mindspore/common/file_system.py +3 -0
  51. mindspore/common/generator.py +3 -0
  52. mindspore/common/hook_handle.py +5 -3
  53. mindspore/common/initializer.py +10 -6
  54. mindspore/common/jit_begin_end.py +94 -0
  55. mindspore/common/jit_config.py +6 -1
  56. mindspore/common/jit_context.py +76 -0
  57. mindspore/common/jit_trace.py +378 -0
  58. mindspore/common/lazy_inline.py +2 -2
  59. mindspore/common/mutable.py +5 -4
  60. mindspore/common/parameter.py +106 -39
  61. mindspore/common/seed.py +2 -2
  62. mindspore/common/sparse_tensor.py +23 -17
  63. mindspore/common/tensor.py +332 -714
  64. mindspore/communication/__init__.py +7 -5
  65. mindspore/communication/_comm_helper.py +47 -2
  66. mindspore/communication/comm_func.py +70 -53
  67. mindspore/communication/management.py +83 -17
  68. mindspore/context.py +228 -571
  69. mindspore/dataset/__init__.py +44 -20
  70. mindspore/dataset/audio/__init__.py +2 -8
  71. mindspore/dataset/audio/transforms.py +3 -17
  72. mindspore/dataset/core/config.py +3 -3
  73. mindspore/dataset/engine/cache_client.py +1 -1
  74. mindspore/dataset/engine/datasets.py +102 -120
  75. mindspore/dataset/engine/datasets_audio.py +22 -22
  76. mindspore/dataset/engine/datasets_standard_format.py +43 -24
  77. mindspore/dataset/engine/datasets_text.py +78 -85
  78. mindspore/dataset/engine/datasets_user_defined.py +109 -77
  79. mindspore/dataset/engine/datasets_vision.py +111 -108
  80. mindspore/dataset/engine/iterators.py +5 -3
  81. mindspore/dataset/engine/obs/obs_mindrecord_dataset.py +1 -1
  82. mindspore/dataset/engine/samplers.py +279 -57
  83. mindspore/dataset/engine/serializer_deserializer.py +2 -1
  84. mindspore/dataset/engine/validators.py +10 -0
  85. mindspore/dataset/text/__init__.py +7 -6
  86. mindspore/dataset/text/transforms.py +6 -5
  87. mindspore/dataset/text/utils.py +3 -3
  88. mindspore/dataset/transforms/__init__.py +0 -9
  89. mindspore/dataset/transforms/transforms.py +3 -3
  90. mindspore/dataset/utils/browse_dataset.py +1 -1
  91. mindspore/dataset/vision/__init__.py +2 -9
  92. mindspore/dataset/vision/transforms.py +202 -158
  93. mindspore/dataset/vision/utils.py +7 -5
  94. mindspore/device_context/ascend/op_debug.py +60 -1
  95. mindspore/device_context/ascend/op_tuning.py +0 -4
  96. mindspore/device_manager.py +39 -3
  97. mindspore/dnnl.dll +0 -0
  98. mindspore/dpcmi.dll +0 -0
  99. mindspore/experimental/es/embedding_service.py +35 -27
  100. mindspore/experimental/llm_boost/ascend_native/llama_boost_ascend_native.py +0 -2
  101. mindspore/experimental/map_parameter.py +4 -4
  102. mindspore/experimental/optim/adadelta.py +22 -26
  103. mindspore/experimental/optim/adagrad.py +4 -4
  104. mindspore/experimental/optim/adam.py +4 -0
  105. mindspore/experimental/optim/adamax.py +4 -4
  106. mindspore/experimental/optim/adamw.py +4 -0
  107. mindspore/experimental/optim/asgd.py +1 -1
  108. mindspore/experimental/optim/lr_scheduler.py +40 -22
  109. mindspore/experimental/optim/radam.py +5 -5
  110. mindspore/experimental/optim/rprop.py +1 -1
  111. mindspore/experimental/optim/sgd.py +1 -1
  112. mindspore/hal/contiguous_tensors_handle.py +6 -10
  113. mindspore/hal/device.py +55 -81
  114. mindspore/hal/event.py +38 -55
  115. mindspore/hal/memory.py +115 -147
  116. mindspore/hal/stream.py +81 -125
  117. mindspore/include/dataset/constants.h +7 -4
  118. mindspore/include/dataset/execute.h +2 -2
  119. mindspore/jpeg62.dll +0 -0
  120. mindspore/log.py +40 -2
  121. mindspore/mindrecord/__init__.py +20 -7
  122. mindspore/mindspore_backend_common.dll +0 -0
  123. mindspore/mindspore_backend_manager.dll +0 -0
  124. mindspore/mindspore_common.dll +0 -0
  125. mindspore/mindspore_core.dll +0 -0
  126. mindspore/mindspore_dump.dll +0 -0
  127. mindspore/mindspore_frontend.dll +0 -0
  128. mindspore/mindspore_glog.dll +0 -0
  129. mindspore/mindspore_memory_pool.dll +0 -0
  130. mindspore/mindspore_ms_backend.dll +0 -0
  131. mindspore/mindspore_ops.dll +0 -0
  132. mindspore/{mindspore_backend.dll → mindspore_ops_host.dll} +0 -0
  133. mindspore/mindspore_ops_kernel_common.dll +0 -0
  134. mindspore/mindspore_profiler.dll +0 -0
  135. mindspore/mindspore_pyboost.dll +0 -0
  136. mindspore/mindspore_pynative.dll +0 -0
  137. mindspore/mindspore_res_manager.dll +0 -0
  138. mindspore/mindspore_runtime_pipeline.dll +0 -0
  139. mindspore/mint/__init__.py +133 -702
  140. mindspore/mint/distributed/__init__.py +5 -1
  141. mindspore/mint/distributed/distributed.py +198 -113
  142. mindspore/mint/linalg/__init__.py +2 -0
  143. mindspore/mint/nn/__init__.py +280 -18
  144. mindspore/mint/nn/functional.py +282 -64
  145. mindspore/mint/nn/layer/__init__.py +4 -0
  146. mindspore/mint/nn/layer/_functions.py +7 -3
  147. mindspore/mint/nn/layer/activation.py +120 -13
  148. mindspore/mint/nn/layer/conv.py +234 -28
  149. mindspore/mint/nn/layer/normalization.py +15 -16
  150. mindspore/mint/nn/layer/padding.py +1 -1
  151. mindspore/mint/nn/layer/pooling.py +66 -1
  152. mindspore/mint/optim/__init__.py +2 -1
  153. mindspore/mint/optim/sgd.py +171 -0
  154. mindspore/msobj140.dll +0 -0
  155. mindspore/mspdb140.dll +0 -0
  156. mindspore/mspdbcore.dll +0 -0
  157. mindspore/mspdbst.dll +0 -0
  158. mindspore/mspft140.dll +0 -0
  159. mindspore/msvcdis140.dll +0 -0
  160. mindspore/msvcp140_1.dll +0 -0
  161. mindspore/msvcp140_2.dll +0 -0
  162. mindspore/msvcp140_atomic_wait.dll +0 -0
  163. mindspore/msvcp140_codecvt_ids.dll +0 -0
  164. mindspore/nn/__init__.py +4 -1
  165. mindspore/nn/cell.py +1253 -179
  166. mindspore/nn/layer/activation.py +23 -21
  167. mindspore/nn/layer/basic.py +22 -16
  168. mindspore/nn/layer/container.py +1 -1
  169. mindspore/nn/layer/conv.py +53 -42
  170. mindspore/nn/layer/embedding.py +9 -8
  171. mindspore/nn/layer/normalization.py +48 -42
  172. mindspore/nn/layer/pooling.py +75 -31
  173. mindspore/nn/layer/transformer.py +11 -10
  174. mindspore/nn/learning_rate_schedule.py +4 -2
  175. mindspore/nn/loss/loss.py +27 -19
  176. mindspore/nn/optim/ada_grad.py +6 -5
  177. mindspore/nn/optim/adadelta.py +9 -7
  178. mindspore/nn/optim/adafactor.py +1 -1
  179. mindspore/nn/optim/adam.py +18 -14
  180. mindspore/nn/optim/adamax.py +8 -7
  181. mindspore/nn/optim/adasum.py +5 -5
  182. mindspore/nn/optim/asgd.py +3 -1
  183. mindspore/nn/optim/ftrl.py +11 -9
  184. mindspore/nn/optim/lamb.py +1 -1
  185. mindspore/nn/optim/lazyadam.py +12 -10
  186. mindspore/nn/optim/momentum.py +7 -6
  187. mindspore/nn/optim/optimizer.py +2 -2
  188. mindspore/nn/optim/proximal_ada_grad.py +12 -10
  189. mindspore/nn/optim/rmsprop.py +13 -12
  190. mindspore/nn/optim/rprop.py +9 -7
  191. mindspore/nn/optim/sgd.py +9 -6
  192. mindspore/nn/optim/tft_wrapper.py +5 -2
  193. mindspore/nn/probability/bijector/bijector.py +17 -11
  194. mindspore/nn/probability/bijector/gumbel_cdf.py +5 -5
  195. mindspore/nn/probability/bijector/invert.py +2 -2
  196. mindspore/nn/probability/bijector/scalar_affine.py +3 -3
  197. mindspore/nn/probability/bijector/softplus.py +3 -2
  198. mindspore/nn/probability/distribution/beta.py +3 -3
  199. mindspore/nn/probability/distribution/categorical.py +1 -1
  200. mindspore/nn/probability/distribution/cauchy.py +4 -2
  201. mindspore/nn/probability/distribution/exponential.py +6 -7
  202. mindspore/nn/probability/distribution/gamma.py +2 -2
  203. mindspore/nn/probability/distribution/gumbel.py +2 -2
  204. mindspore/nn/probability/distribution/half_normal.py +5 -3
  205. mindspore/nn/probability/distribution/logistic.py +5 -3
  206. mindspore/nn/probability/distribution/poisson.py +1 -1
  207. mindspore/nn/probability/distribution/uniform.py +5 -3
  208. mindspore/nn/reinforcement/_tensors_queue.py +1 -1
  209. mindspore/nn/reinforcement/tensor_array.py +1 -1
  210. mindspore/nn/wrap/__init__.py +6 -6
  211. mindspore/nn/wrap/cell_wrapper.py +178 -117
  212. mindspore/nn/wrap/grad_reducer.py +45 -36
  213. mindspore/nn/wrap/loss_scale.py +3 -3
  214. mindspore/numpy/array_creations.py +3 -3
  215. mindspore/numpy/array_ops.py +1 -1
  216. mindspore/numpy/utils.py +1 -2
  217. mindspore/numpy/utils_const.py +1 -2
  218. mindspore/opencv_core452.dll +0 -0
  219. mindspore/opencv_imgcodecs452.dll +0 -0
  220. mindspore/opencv_imgproc452.dll +0 -0
  221. mindspore/ops/__init__.py +3 -2
  222. mindspore/ops/_grad_experimental/grad_comm_ops.py +18 -3
  223. mindspore/ops/_grad_experimental/grad_debug_ops.py +8 -1
  224. mindspore/ops/_grad_experimental/taylor_rule.py +29 -0
  225. mindspore/ops/_register_for_op.py +0 -11
  226. mindspore/{ops_generate → ops/_utils}/arg_dtype_cast.py +123 -4
  227. mindspore/{ops_generate → ops/_utils}/arg_handler.py +3 -4
  228. mindspore/ops/_vmap/vmap_array_ops.py +32 -6
  229. mindspore/ops/_vmap/vmap_grad_nn_ops.py +2 -1
  230. mindspore/ops/_vmap/vmap_math_ops.py +4 -7
  231. mindspore/ops/_vmap/vmap_nn_ops.py +9 -8
  232. mindspore/ops/auto_generate/__init__.py +4 -3
  233. mindspore/ops/auto_generate/cpp_create_prim_instance_helper.py +127 -52
  234. mindspore/ops/auto_generate/gen_extend_func.py +286 -208
  235. mindspore/ops/auto_generate/gen_ops_def.py +2783 -2335
  236. mindspore/ops/auto_generate/gen_ops_prim.py +8992 -2686
  237. mindspore/ops/auto_generate/pyboost_inner_prim.py +106 -76
  238. mindspore/ops/composite/__init__.py +2 -1
  239. mindspore/ops/composite/base.py +19 -24
  240. mindspore/ops/composite/math_ops.py +6 -16
  241. mindspore/ops/composite/multitype_ops/__init__.py +5 -2
  242. mindspore/ops/composite/multitype_ops/_compile_utils.py +4 -5
  243. mindspore/ops/composite/multitype_ops/_constexpr_utils.py +1 -2
  244. mindspore/ops/composite/multitype_ops/add_impl.py +2 -1
  245. mindspore/ops/composite/multitype_ops/bitwise_and_impl.py +2 -1
  246. mindspore/ops/composite/multitype_ops/bitwise_or_impl.py +2 -1
  247. mindspore/ops/composite/multitype_ops/bitwise_xor_impl.py +2 -1
  248. mindspore/ops/composite/multitype_ops/div_impl.py +6 -4
  249. mindspore/ops/composite/multitype_ops/equal_impl.py +4 -3
  250. mindspore/ops/composite/multitype_ops/floordiv_impl.py +2 -1
  251. mindspore/ops/composite/multitype_ops/getitem_impl.py +3 -2
  252. mindspore/ops/composite/multitype_ops/greater_equal_impl.py +4 -3
  253. mindspore/ops/composite/multitype_ops/greater_impl.py +4 -3
  254. mindspore/ops/composite/multitype_ops/in_impl.py +2 -1
  255. mindspore/ops/composite/multitype_ops/invert_impl.py +50 -0
  256. mindspore/ops/composite/multitype_ops/left_shift_impl.py +2 -1
  257. mindspore/ops/composite/multitype_ops/less_equal_impl.py +4 -3
  258. mindspore/ops/composite/multitype_ops/less_impl.py +4 -3
  259. mindspore/ops/composite/multitype_ops/logic_not_impl.py +3 -2
  260. mindspore/ops/composite/multitype_ops/logical_and_impl.py +2 -1
  261. mindspore/ops/composite/multitype_ops/logical_or_impl.py +2 -1
  262. mindspore/ops/composite/multitype_ops/mod_impl.py +2 -1
  263. mindspore/ops/composite/multitype_ops/mul_impl.py +3 -2
  264. mindspore/ops/composite/multitype_ops/negative_impl.py +2 -1
  265. mindspore/ops/composite/multitype_ops/not_equal_impl.py +2 -1
  266. mindspore/ops/composite/multitype_ops/not_in_impl.py +2 -1
  267. mindspore/ops/composite/multitype_ops/ones_like_impl.py +18 -0
  268. mindspore/ops/composite/multitype_ops/pow_impl.py +2 -1
  269. mindspore/ops/composite/multitype_ops/right_shift_impl.py +2 -1
  270. mindspore/ops/composite/multitype_ops/setitem_impl.py +2 -1
  271. mindspore/ops/composite/multitype_ops/sub_impl.py +2 -1
  272. mindspore/ops/function/__init__.py +28 -2
  273. mindspore/ops/function/_add_attr_func.py +58 -0
  274. mindspore/ops/function/array_func.py +1631 -2347
  275. mindspore/ops/function/clip_func.py +38 -45
  276. mindspore/ops/function/debug_func.py +36 -44
  277. mindspore/ops/function/grad/__init__.py +1 -0
  278. mindspore/ops/function/grad/grad_func.py +104 -71
  279. mindspore/ops/function/image_func.py +1 -1
  280. mindspore/ops/function/linalg_func.py +46 -78
  281. mindspore/ops/function/math_func.py +3024 -3855
  282. mindspore/ops/function/nn_func.py +678 -274
  283. mindspore/ops/function/other_func.py +159 -1
  284. mindspore/ops/function/parameter_func.py +17 -30
  285. mindspore/ops/function/random_func.py +216 -361
  286. mindspore/ops/function/reshard_func.py +4 -70
  287. mindspore/ops/function/sparse_func.py +3 -3
  288. mindspore/ops/function/sparse_unary_func.py +5 -5
  289. mindspore/ops/function/spectral_func.py +25 -58
  290. mindspore/ops/function/vmap_func.py +26 -18
  291. mindspore/ops/functional.py +8 -5
  292. mindspore/ops/functional_overload.py +655 -4
  293. mindspore/ops/op_info_register.py +32 -244
  294. mindspore/ops/operations/__init__.py +21 -14
  295. mindspore/ops/operations/_custom_ops_utils.py +235 -0
  296. mindspore/ops/operations/_grad_ops.py +1 -10
  297. mindspore/ops/operations/_inner_ops.py +5 -76
  298. mindspore/ops/operations/_ms_kernel.py +4 -10
  299. mindspore/ops/operations/_rl_inner_ops.py +1 -1
  300. mindspore/ops/operations/_scalar_ops.py +3 -2
  301. mindspore/ops/operations/_sequence_ops.py +1 -1
  302. mindspore/ops/operations/_tensor_array.py +1 -1
  303. mindspore/ops/operations/array_ops.py +39 -24
  304. mindspore/ops/operations/comm_ops.py +150 -107
  305. mindspore/ops/operations/custom_ops.py +287 -32
  306. mindspore/ops/operations/debug_ops.py +119 -16
  307. mindspore/ops/operations/inner_ops.py +1 -1
  308. mindspore/ops/operations/linalg_ops.py +1 -58
  309. mindspore/ops/operations/manually_defined/_inner.py +1 -1
  310. mindspore/ops/operations/manually_defined/ops_def.py +746 -79
  311. mindspore/ops/operations/math_ops.py +21 -18
  312. mindspore/ops/operations/nn_ops.py +67 -224
  313. mindspore/ops/operations/other_ops.py +62 -9
  314. mindspore/ops/operations/random_ops.py +13 -7
  315. mindspore/ops/operations/reshard_ops.py +1 -1
  316. mindspore/ops/operations/sparse_ops.py +2 -2
  317. mindspore/ops/primitive.py +43 -32
  318. mindspore/ops/tensor_method.py +243 -17
  319. mindspore/ops_generate/__init__.py +0 -5
  320. mindspore/ops_generate/aclnn/__init__.py +0 -0
  321. mindspore/ops_generate/{aclnn_kernel_register_auto_cc_generator.py → aclnn/aclnn_kernel_register_auto_cc_generator.py} +43 -18
  322. mindspore/ops_generate/{gen_aclnn_implement.py → aclnn/gen_aclnn_implement.py} +49 -51
  323. mindspore/ops_generate/api/__init__.py +0 -0
  324. mindspore/ops_generate/{add_tensor_docs_generator.py → api/add_tensor_docs_generator.py} +9 -7
  325. mindspore/ops_generate/{cpp_create_prim_instance_helper_generator.py → api/cpp_create_prim_instance_helper_generator.py} +6 -9
  326. mindspore/ops_generate/{functional_map_cpp_generator.py → api/functional_map_cpp_generator.py} +25 -12
  327. mindspore/ops_generate/{functional_overload_py_generator.py → api/functional_overload_py_generator.py} +8 -6
  328. mindspore/ops_generate/{functions_cc_generator.py → api/functions_cc_generator.py} +14 -10
  329. mindspore/ops_generate/api/gen_api.py +103 -0
  330. mindspore/ops_generate/{op_api_proto.py → api/op_api_proto.py} +98 -69
  331. mindspore/ops_generate/{tensor_func_reg_cpp_generator.py → api/tensor_func_reg_cpp_generator.py} +82 -43
  332. mindspore/ops_generate/common/__init__.py +0 -0
  333. mindspore/ops_generate/common/gen_constants.py +91 -0
  334. mindspore/ops_generate/{gen_utils.py → common/gen_utils.py} +72 -19
  335. mindspore/ops_generate/{op_proto.py → common/op_proto.py} +64 -1
  336. mindspore/ops_generate/{template.py → common/template.py} +96 -84
  337. mindspore/ops_generate/gen_ops.py +23 -325
  338. mindspore/ops_generate/op_def/__init__.py +0 -0
  339. mindspore/ops_generate/op_def/gen_op_def.py +90 -0
  340. mindspore/ops_generate/{lite_ops_cpp_generator.py → op_def/lite_ops_cpp_generator.py} +47 -11
  341. mindspore/ops_generate/{ops_def_cc_generator.py → op_def/ops_def_cc_generator.py} +18 -10
  342. mindspore/ops_generate/{ops_def_h_generator.py → op_def/ops_def_h_generator.py} +5 -5
  343. mindspore/ops_generate/{ops_name_h_generator.py → op_def/ops_name_h_generator.py} +30 -15
  344. mindspore/ops_generate/op_def/ops_primitive_h_generator.py +125 -0
  345. mindspore/ops_generate/op_def_py/__init__.py +0 -0
  346. mindspore/ops_generate/op_def_py/gen_op_def_py.py +47 -0
  347. mindspore/ops_generate/{op_def_py_generator.py → op_def_py/op_def_py_generator.py} +6 -5
  348. mindspore/ops_generate/{op_prim_py_generator.py → op_def_py/op_prim_py_generator.py} +24 -15
  349. mindspore/ops_generate/pyboost/__init__.py +0 -0
  350. mindspore/ops_generate/{auto_grad_impl_cc_generator.py → pyboost/auto_grad_impl_cc_generator.py} +11 -7
  351. mindspore/ops_generate/{auto_grad_reg_cc_generator.py → pyboost/auto_grad_reg_cc_generator.py} +7 -7
  352. mindspore/ops_generate/{gen_pyboost_func.py → pyboost/gen_pyboost_func.py} +40 -16
  353. mindspore/ops_generate/{op_template_parser.py → pyboost/op_template_parser.py} +105 -24
  354. mindspore/ops_generate/{pyboost_functions_cpp_generator.py → pyboost/pyboost_functions_cpp_generator.py} +55 -18
  355. mindspore/ops_generate/{pyboost_functions_h_generator.py → pyboost/pyboost_functions_h_generator.py} +42 -10
  356. mindspore/ops_generate/{pyboost_functions_py_generator.py → pyboost/pyboost_functions_py_generator.py} +6 -6
  357. mindspore/ops_generate/{pyboost_grad_function_cpp_generator.py → pyboost/pyboost_grad_function_cpp_generator.py} +11 -10
  358. mindspore/ops_generate/{pyboost_inner_prim_generator.py → pyboost/pyboost_inner_prim_generator.py} +8 -7
  359. mindspore/ops_generate/{pyboost_native_grad_functions_generator.py → pyboost/pyboost_native_grad_functions_generator.py} +14 -10
  360. mindspore/ops_generate/{pyboost_op_cpp_code_generator.py → pyboost/pyboost_op_cpp_code_generator.py} +140 -53
  361. mindspore/ops_generate/{pyboost_overload_functions_cpp_generator.py → pyboost/pyboost_overload_functions_cpp_generator.py} +28 -15
  362. mindspore/ops_generate/{pyboost_utils.py → pyboost/pyboost_utils.py} +88 -4
  363. mindspore/ops_generate/resources/__init__.py +0 -0
  364. mindspore/ops_generate/resources/resource_list.py +30 -0
  365. mindspore/ops_generate/resources/resource_loader.py +36 -0
  366. mindspore/ops_generate/resources/resource_manager.py +64 -0
  367. mindspore/ops_generate/resources/yaml_loader.py +88 -0
  368. mindspore/ops_generate/tensor_py_cc_generator.py +122 -0
  369. mindspore/parallel/__init__.py +6 -2
  370. mindspore/parallel/_auto_parallel_context.py +140 -12
  371. mindspore/parallel/_cell_wrapper.py +132 -15
  372. mindspore/parallel/_parallel_serialization.py +95 -4
  373. mindspore/parallel/_ps_context.py +1 -1
  374. mindspore/parallel/_recovery_context.py +7 -2
  375. mindspore/parallel/_tensor.py +142 -18
  376. mindspore/parallel/_utils.py +198 -25
  377. mindspore/parallel/algo_parameter_config.py +3 -3
  378. mindspore/parallel/auto_parallel.py +732 -0
  379. mindspore/parallel/checkpoint_convert.py +159 -0
  380. mindspore/parallel/checkpoint_transform.py +658 -37
  381. mindspore/parallel/cluster/process_entity/_api.py +151 -19
  382. mindspore/parallel/cluster/run.py +1 -1
  383. mindspore/parallel/function/__init__.py +24 -0
  384. mindspore/parallel/function/reshard_func.py +258 -0
  385. mindspore/parallel/nn/__init__.py +25 -0
  386. mindspore/parallel/nn/parallel_cell_wrapper.py +263 -0
  387. mindspore/parallel/nn/parallel_grad_reducer.py +169 -0
  388. mindspore/parallel/parameter_broadcast.py +24 -13
  389. mindspore/parallel/shard.py +137 -62
  390. mindspore/parallel/transform_safetensors.py +288 -95
  391. mindspore/pgodb140.dll +0 -0
  392. mindspore/pgort140.dll +0 -0
  393. mindspore/profiler/__init__.py +9 -5
  394. mindspore/profiler/analysis/parser/ascend_cann_parser.py +6 -2
  395. mindspore/profiler/analysis/parser/ms_framework_parser.py +4 -4
  396. mindspore/profiler/analysis/parser/timeline_assembly_factory/ascend_timeline_assembler.py +7 -4
  397. mindspore/profiler/analysis/parser/timeline_assembly_factory/trace_view_container.py +25 -0
  398. mindspore/profiler/analysis/parser/timeline_creator/fwk_timeline_creator.py +3 -3
  399. mindspore/profiler/analysis/parser/timeline_event/fwk_event.py +241 -86
  400. mindspore/profiler/analysis/viewer/ascend_communication_viewer.py +41 -2
  401. mindspore/profiler/analysis/viewer/ascend_kernel_details_viewer.py +33 -35
  402. mindspore/profiler/analysis/viewer/ascend_memory_viewer.py +7 -0
  403. mindspore/profiler/analysis/viewer/ascend_op_memory_viewer.py +8 -3
  404. mindspore/profiler/analysis/viewer/ascend_step_trace_time_viewer.py +141 -30
  405. mindspore/profiler/analysis/viewer/ms_dataset_viewer.py +5 -6
  406. mindspore/profiler/common/ascend_msprof_exporter.py +5 -4
  407. mindspore/profiler/common/constant.py +12 -0
  408. mindspore/profiler/common/msprof_cmd_tool.py +42 -23
  409. mindspore/profiler/common/path_manager.py +24 -0
  410. mindspore/profiler/common/profiler_context.py +26 -2
  411. mindspore/profiler/common/profiler_meta_data.py +74 -0
  412. mindspore/profiler/common/profiler_parameters.py +59 -18
  413. mindspore/profiler/common/profiler_path_manager.py +66 -7
  414. mindspore/profiler/dynamic_profiler.py +112 -79
  415. mindspore/profiler/envprofiler.py +26 -1
  416. mindspore/profiler/experimental_config.py +197 -0
  417. mindspore/profiler/mstx.py +57 -14
  418. mindspore/profiler/platform/npu_profiler.py +33 -7
  419. mindspore/profiler/profiler.py +541 -45
  420. mindspore/profiler/profiler_action_controller.py +1 -1
  421. mindspore/profiler/profiler_interface.py +4 -0
  422. mindspore/profiler/schedule.py +57 -22
  423. mindspore/rewrite/api/node.py +15 -13
  424. mindspore/rewrite/api/symbol_tree.py +1 -1
  425. mindspore/run_check/_check_version.py +25 -14
  426. mindspore/run_check/run_check.py +1 -1
  427. mindspore/runtime/__init__.py +2 -2
  428. mindspore/runtime/executor.py +40 -11
  429. mindspore/runtime/memory.py +37 -13
  430. mindspore/safeguard/rewrite_obfuscation.py +12 -9
  431. mindspore/swresample-4.dll +0 -0
  432. mindspore/swscale-6.dll +0 -0
  433. mindspore/tbbmalloc.dll +0 -0
  434. mindspore/tinyxml2.dll +0 -0
  435. mindspore/train/__init__.py +8 -8
  436. mindspore/train/_utils.py +43 -9
  437. mindspore/train/amp.py +1 -1
  438. mindspore/train/callback/__init__.py +2 -2
  439. mindspore/train/callback/_callback.py +2 -16
  440. mindspore/train/callback/_checkpoint.py +24 -40
  441. mindspore/train/callback/_cluster_monitor.py +14 -18
  442. mindspore/train/callback/_flops_collector.py +2 -3
  443. mindspore/train/callback/_history.py +7 -4
  444. mindspore/train/callback/_lambda_callback.py +2 -2
  445. mindspore/train/callback/_landscape.py +0 -3
  446. mindspore/train/callback/_loss_monitor.py +2 -1
  447. mindspore/train/callback/_on_request_exit.py +6 -5
  448. mindspore/train/callback/_reduce_lr_on_plateau.py +11 -6
  449. mindspore/train/callback/_summary_collector.py +8 -13
  450. mindspore/train/callback/_time_monitor.py +2 -1
  451. mindspore/train/callback/{_tft_register.py → _train_fault_tolerance.py} +204 -105
  452. mindspore/train/data_sink.py +25 -2
  453. mindspore/train/dataset_helper.py +4 -5
  454. mindspore/train/loss_scale_manager.py +8 -7
  455. mindspore/train/metrics/accuracy.py +3 -3
  456. mindspore/train/metrics/confusion_matrix.py +9 -9
  457. mindspore/train/metrics/error.py +3 -3
  458. mindspore/train/metrics/hausdorff_distance.py +4 -4
  459. mindspore/train/metrics/mean_surface_distance.py +3 -3
  460. mindspore/train/metrics/metric.py +0 -12
  461. mindspore/train/metrics/occlusion_sensitivity.py +4 -2
  462. mindspore/train/metrics/precision.py +8 -6
  463. mindspore/train/metrics/recall.py +9 -9
  464. mindspore/train/metrics/root_mean_square_surface_distance.py +2 -2
  465. mindspore/train/mind_ir_pb2.py +19 -12
  466. mindspore/train/model.py +262 -127
  467. mindspore/train/serialization.py +246 -988
  468. mindspore/train/summary/_summary_adapter.py +2 -2
  469. mindspore/train/summary/summary_record.py +1 -1
  470. mindspore/turbojpeg.dll +0 -0
  471. mindspore/utils/__init__.py +3 -2
  472. mindspore/utils/dryrun.py +4 -2
  473. mindspore/utils/hooks.py +81 -0
  474. mindspore/utils/runtime_execution_order_check.py +2 -0
  475. mindspore/utils/utils.py +138 -4
  476. mindspore/vcmeta.dll +0 -0
  477. mindspore/vcruntime140.dll +0 -0
  478. mindspore/vcruntime140_1.dll +0 -0
  479. mindspore/version.py +1 -1
  480. {mindspore-2.5.0.dist-info → mindspore-2.6.0.dist-info}/METADATA +2 -1
  481. {mindspore-2.5.0.dist-info → mindspore-2.6.0.dist-info}/RECORD +485 -440
  482. mindspore/_install_custom.py +0 -43
  483. mindspore/common/_register_for_adapter.py +0 -74
  484. mindspore/ops/auto_generate/gen_arg_dtype_cast.py +0 -252
  485. mindspore/ops/auto_generate/gen_arg_handler.py +0 -136
  486. mindspore/ops/operations/_opaque_predicate_registry.py +0 -41
  487. mindspore/ops_generate/gen_constants.py +0 -190
  488. mindspore/ops_generate/gen_ops_inner_prim.py +0 -131
  489. mindspore/ops_generate/ops_primitive_h_generator.py +0 -81
  490. /mindspore/ops_generate/{base_generator.py → common/base_generator.py} +0 -0
  491. {mindspore-2.5.0.dist-info → mindspore-2.6.0.dist-info}/WHEEL +0 -0
  492. {mindspore-2.5.0.dist-info → mindspore-2.6.0.dist-info}/entry_points.txt +0 -0
  493. {mindspore-2.5.0.dist-info → mindspore-2.6.0.dist-info}/top_level.txt +0 -0
mindspore/hal/event.py CHANGED
@@ -24,41 +24,37 @@ function_event_status = {'Event': False, 'wait': False}
24
24
 
25
25
  class Event(Event_):
26
26
  r"""
27
- Wrapper around a device event.
27
+ Wrapper around a device event, this api will be deprecated and removed in future versions, please use
28
+ the api :func:`mindspore.runtime.Event` instead.
28
29
 
29
30
  Device events are synchronization markers that can be used to monitor the device’s progress,
30
31
  to accurately measure timing, and to synchronize device streams.
31
32
 
32
33
  The underlying device events are lazily initialized when the event is first recorded.
33
34
 
34
- Note:
35
- - The api will be deprecated, please use the api :func:`mindspore.runtime.Event` instead.
36
-
37
35
  Args:
38
- enable_timing (bool, optional): indicates if the event should measure time (default: ``False``)
39
- blocking (bool, optional): if ``True``, `wait` will be blocking (default: ``False``)
36
+ enable_timing (bool, optional): indicates if the event should measure time. Default ``False``.
37
+ blocking (bool, optional): if ``True``, `wait` will be blocking. Default ``False``.
40
38
 
41
39
  Examples:
42
- >>> import mindspore as ms
43
- >>> import numpy as np
44
- >>> from mindspore import Tensor, ops
45
- >>> start = ms.hal.Event(enable_timing=True)
46
- >>> end = ms.hal.Event(enable_timing=True)
47
- >>> s1 = ms.hal.Stream()
48
- >>> s2 = ms.hal.Stream()
49
- >>> a = Tensor(np.ones([2, 2]), ms.float32)
50
- >>> b = Tensor(np.ones([2, 2]), ms.float32)
51
- >>> c = Tensor(np.ones([2, 2]), ms.float32)
52
- >>> with ms.hal.StreamCtx(s1):
53
- ... d = ops.matmul(a, b)
40
+ >>> import mindspore
41
+ >>> start = mindspore.hal.Event(enable_timing=True)
42
+ >>> end = mindspore.hal.Event(enable_timing=True)
43
+ >>> s1 = mindspore.hal.Stream()
44
+ >>> s2 = mindspore.hal.Stream()
45
+ >>> a = mindspore.tensor(mindspore.ops.ones([2, 2]), mindspore.float32)
46
+ >>> b = mindspore.tensor(mindspore.ops.ones([2, 2]), mindspore.float32)
47
+ >>> c = mindspore.tensor(mindspore.ops.ones([2, 2]), mindspore.float32)
48
+ >>> with mindspore.hal.StreamCtx(s1):
49
+ ... d = mindspore.ops.matmul(a, b)
54
50
  ... start.record()
55
51
  >>> c += 2
56
52
  >>> end.record()
57
- >>> with ms.hal.StreamCtx(s2):
53
+ >>> with mindspore.hal.StreamCtx(s2):
58
54
  ... start.synchronize()
59
55
  ... end.synchronize()
60
56
  ... e = c + d
61
- >>> ms.hal.synchronize()
57
+ >>> mindspore.hal.synchronize()
62
58
  >>> print(e)
63
59
  [[5. 5.]
64
60
  [5. 5.]]
@@ -78,17 +74,14 @@ class Event(Event_):
78
74
 
79
75
  def record(self, stream=None):
80
76
  r"""
81
- Records the event in a given stream.
77
+ Record the event in a given stream.
82
78
 
83
79
  Uses :func:`mindspore.hal.current_stream()` if no `stream` is specified. The
84
80
  stream's device must match the event's device.
85
81
 
86
82
  Args:
87
83
  stream (Stream, optional): a stream to record. If this argument is ``None``,
88
- current stream will be used. Default value: ``None``.
89
-
90
- Raises:
91
- TypeError: If 'stream' is neither a :class:`mindspore.hal.Stream` nor a ``None``.
84
+ current stream will be used. Default ``None``.
92
85
  """
93
86
  if stream is None:
94
87
  stream = current_stream_()
@@ -99,33 +92,28 @@ class Event(Event_):
99
92
 
100
93
  def wait(self, stream=None):
101
94
  r"""
102
- Makes all future work submitted to the given stream wait for this
95
+ Make all future work submitted to the given stream wait for this
103
96
  event.
104
97
 
105
98
  Use :func:`mindspore.hal.current_stream()` if no `stream` is specified.
106
99
 
107
100
  Args:
108
101
  stream (Stream, optional): a stream to record. If this argument is ``None``,
109
- current stream will be used. Default value: ``None``.
110
-
111
- Raises:
112
- TypeError: If 'stream' is neither a :class:`mindspore.hal.Stream` nor a ``None``.
102
+ current stream will be used. Default ``None``.
113
103
 
114
104
  Examples:
115
- >>> import mindspore as ms
116
- >>> import numpy as np
117
- >>> from mindspore import Tensor, ops
118
- >>> event = ms.hal.Event()
119
- >>> s1 = ms.hal.Stream()
120
- >>> s2 = ms.hal.Stream()
121
- >>> a = Tensor(np.ones([2, 2]), ms.float32)
122
- >>> b = Tensor(np.ones([2, 2]), ms.float32)
123
- >>> with ms.hal.StreamCtx(s1):
124
- ... c = ops.matmul(a, b)
105
+ >>> import mindspore
106
+ >>> event = mindspore.hal.Event()
107
+ >>> s1 = mindspore.hal.Stream()
108
+ >>> s2 = mindspore.hal.Stream()
109
+ >>> a = mindspore.tensor(mindspore.ops.ones([2, 2]), mindspore.float32)
110
+ >>> b = mindspore.tensor(mindspore.ops.ones([2, 2]), mindspore.float32)
111
+ >>> with mindspore.hal.StreamCtx(s1):
112
+ ... c = mindspore.ops.matmul(a, b)
125
113
  ... event.record()
126
114
  >>> event.wait()
127
115
  >>> d = c + 2
128
- >>> ms.hal.synchronize()
116
+ >>> mindspore.hal.synchronize()
129
117
  >>> print(d)
130
118
  [[4. 4.]
131
119
  [4. 4.]]
@@ -144,7 +132,7 @@ class Event(Event_):
144
132
 
145
133
  def synchronize(self):
146
134
  r"""
147
- Waits for the event to complete.
135
+ Wait for the event to complete.
148
136
 
149
137
  Waits until the completion of all work currently captured in this event.
150
138
  This prevents the CPU thread from proceeding until the event completes.
@@ -154,20 +142,18 @@ class Event(Event_):
154
142
 
155
143
  def query(self):
156
144
  r"""
157
- Checks if all work currently captured by event has completed.
145
+ Check if all work currently captured by event has completed.
158
146
 
159
147
  Returns:
160
148
  A boolean indicating if all work currently captured by event has completed.
161
149
 
162
150
  Examples:
163
- >>> import mindspore as ms
164
- >>> import numpy as np
165
- >>> from mindspore import Tensor, ops
166
- >>> a = Tensor(np.ones([1024, 2048]), ms.float32)
167
- >>> b = Tensor(np.ones([2048, 4096]), ms.float32)
168
- >>> s1 = ms.hal.Stream()
169
- >>> with ms.hal.StreamCtx(s1):
170
- ... c = ops.matmul(a, b)
151
+ >>> import mindspore
152
+ >>> a = mindspore.tensor(mindspore.ops.ones([1024, 2048]), mindspore.float32)
153
+ >>> b = mindspore.tensor(mindspore.ops.ones([2048, 4096]), mindspore.float32)
154
+ >>> s1 = mindspore.hal.Stream()
155
+ >>> with mindspore.hal.StreamCtx(s1):
156
+ ... c = mindspore.ops.matmul(a, b)
171
157
  ... ev = s1.record_event()
172
158
  >>> s1.synchronize()
173
159
  >>> assert ev.query()
@@ -177,7 +163,7 @@ class Event(Event_):
177
163
 
178
164
  def elapsed_time(self, end_event):
179
165
  r"""
180
- Returns the time elapsed in milliseconds after the event was
166
+ Return the time elapsed in milliseconds after the event was
181
167
  recorded and before the end_event was recorded.
182
168
 
183
169
  Args:
@@ -185,9 +171,6 @@ class Event(Event_):
185
171
 
186
172
  Returns:
187
173
  float, the time elapsed in milliseconds.
188
-
189
- Raises:
190
- TypeError: If 'end_event' is not a :class:`mindspore.hal.Event`.
191
174
  """
192
175
  # pylint: disable=useless-super-delegation
193
176
  if not isinstance(end_event, Event):
mindspore/hal/memory.py CHANGED
@@ -14,43 +14,42 @@
14
14
  # ============================================================================
15
15
 
16
16
  """Hardware memory interfaces."""
17
- from mindspore._c_expression import _memory_stats, _reset_max_mem_reserved, _reset_max_mem_allocated
17
+ from mindspore._c_expression import _memory_stats, _reset_max_mem_reserved, _reset_max_mem_allocated, _empty_cache, \
18
+ DeviceContextManager
18
19
  from mindspore import log as logger
20
+ import mindspore as ms
19
21
  from .device import _check_inputs_validation, is_initialized
20
22
 
23
+
21
24
  function_memory_status = {'memory_stats': False, 'memory_reserved': False, 'max_memory_reserved': False,
22
25
  'empty_cache': False, 'reset_peak_memory_stats': False, 'memory_summary': False,
23
26
  'memory_allocated': False, 'max_memory_allocated': False,
24
27
  'reset_max_memory_reserved': False, 'reset_max_memory_allocated': False}
28
+ _device_context_mgr = DeviceContextManager.get_instance()
25
29
 
26
30
 
27
31
  @_check_inputs_validation
28
32
  def memory_stats(device_target=None):
29
33
  """
30
- Returns status information queried from the memory pool.
31
-
32
- Note:
33
- - The api will be deprecated, please use the api :func:`mindspore.runtime.memory_stats` instead.
34
+ Returns status information queried from the memory pool, this api will be deprecated and removed in future
35
+ versions, please use the api :func:`mindspore.runtime.memory_stats` instead.
34
36
 
35
37
  Note:
36
- - If `device_target` is not specified, get the device capability of the current backend set by context.
37
- - For the `CPU` backend, a dictionary with empty data is always returned.
38
+ - For the `CPU` device, a dictionary with empty data is always returned.
38
39
 
39
40
  Args:
40
- device_target (str, optional): The device name of backend, should be one of "CPU", "GPU" and "Ascend".
41
- Default value: ``None``.
41
+ device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
42
+ ``"Ascend"`` . Default ``None`` , represents the current device set by context.
42
43
 
43
44
  Returns:
44
45
  dict, the queried memory information.
45
46
 
46
47
  Examples:
47
- >>> import mindspore as ms
48
- >>> import numpy as np
49
- >>> from mindspore import Tensor, ops
50
- >>> a = Tensor(np.ones([1, 2]), ms.float32)
51
- >>> b = Tensor(np.ones([1, 2]), ms.float32)
52
- >>> c = ops.add(a, b).asnumpy()
53
- >>> print(ms.hal.memory_stats())
48
+ >>> import mindspore
49
+ >>> a = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
50
+ >>> b = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
51
+ >>> c = mindspore.ops.add(a, b).asnumpy()
52
+ >>> print(mindspore.hal.memory_stats())
54
53
  {'total_reserved_memory': 1073741824, 'total_allocated_memory': 1024, 'total_idle_memory': 1073740800,
55
54
  'total_eager_free_memory': 0, 'max_reserved_memory': 1073741824, 'max_allocated_memory': 1536,
56
55
  'common_mem_pool_stats': {'block_unit_size': 1073741824, 'block_counts': 1, 'blocks_info':
@@ -72,30 +71,25 @@ def memory_stats(device_target=None):
72
71
  @_check_inputs_validation
73
72
  def memory_reserved(device_target=None):
74
73
  """
75
- Returns the total amount of memory currently managed by the memory pool.
76
-
77
- Note:
78
- - The api will be deprecated, please use the api :func:`mindspore.runtime.memory_reserved` instead.
74
+ Returns the total amount of memory currently managed by the memory pool, this api will be deprecated and removed in
75
+ future versions, please use the api :func:`mindspore.runtime.memory_reserved` instead.
79
76
 
80
77
  Note:
81
- - If `device_target` is not specified, get the device capability of the current backend set by context.
82
- - For the `CPU` backend, 0 is always returned.
78
+ - For the `CPU` device, 0 is always returned.
83
79
 
84
80
  Args:
85
- device_target (str, optional): The device name of backend, should be one of "CPU", "GPU" and "Ascend".
86
- Default value: ``None``.
81
+ device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
82
+ ``"Ascend"`` . Default ``None`` , represents the current device set by context.
87
83
 
88
84
  Returns:
89
85
  int, in Byte.
90
86
 
91
87
  Examples:
92
- >>> import mindspore as ms
93
- >>> import numpy as np
94
- >>> from mindspore import Tensor, ops
95
- >>> a = Tensor(np.ones([1, 2]), ms.float32)
96
- >>> b = Tensor(np.ones([1, 2]), ms.float32)
97
- >>> c = ops.add(a, b).asnumpy()
98
- >>> print(ms.hal.memory_reserved())
88
+ >>> import mindspore
89
+ >>> a = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
90
+ >>> b = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
91
+ >>> c = mindspore.ops.add(a, b).asnumpy()
92
+ >>> print(mindspore.hal.memory_reserved())
99
93
  1073741824
100
94
  """
101
95
  if not function_memory_status['memory_reserved']:
@@ -111,29 +105,25 @@ def memory_reserved(device_target=None):
111
105
  def max_memory_reserved(device_target=None):
112
106
  """
113
107
  Returns the peak value of the total memory managed by the memory pool since the process was started.
108
+ This api will be deprecated and removed in future versions, please use
109
+ the api :func:`mindspore.runtime.max_memory_reserved` instead.
114
110
 
115
111
  Note:
116
- - The api will be deprecated, please use the api :func:`mindspore.runtime.max_memory_reserved` instead.
117
-
118
- Note:
119
- - If `device_target` is not specified, get the device capability of the current backend set by context.
120
- - For the `CPU` backend, 0 is always returned.
112
+ - For the `CPU` device, 0 is always returned.
121
113
 
122
114
  Args:
123
- device_target (str, optional): The device name of backend, should be one of "CPU", "GPU" and "Ascend".
124
- Default value: ``None``.
115
+ device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
116
+ ``"Ascend"`` . Default ``None`` , represents the current device set by context.
125
117
 
126
118
  Returns:
127
119
  int, in Byte.
128
120
 
129
121
  Examples:
130
- >>> import mindspore as ms
131
- >>> import numpy as np
132
- >>> from mindspore import Tensor, ops
133
- >>> a = Tensor(np.ones([1, 2]), ms.float32)
134
- >>> b = Tensor(np.ones([1, 2]), ms.float32)
135
- >>> c = ops.add(a, b).asnumpy()
136
- >>> print(ms.hal.max_memory_reserved())
122
+ >>> import mindspore
123
+ >>> a = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
124
+ >>> b = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
125
+ >>> c = mindspore.ops.add(a, b).asnumpy()
126
+ >>> print(mindspore.hal.max_memory_reserved())
137
127
  1073741824
138
128
  """
139
129
  if not function_memory_status['max_memory_reserved']:
@@ -145,54 +135,60 @@ def max_memory_reserved(device_target=None):
145
135
  return _memory_stats(device_target).get("max_reserved_memory", 0)
146
136
 
147
137
 
138
+ def _is_initialized(device_target):
139
+ """
140
+ Returns whether specified backend is initialized.
141
+ """
142
+ _device_context = _device_context_mgr.get_device_context(device_target)
143
+ if _device_context is None:
144
+ return False
145
+ return _device_context.initialized()
146
+
148
147
  @_check_inputs_validation
149
148
  def empty_cache():
150
149
  """
151
- Release all memory fragments in the memory pool, so that memory arrangement
152
- will be optimized.
150
+ Empty cache in the memory pool, this api will be deprecated and removed in future versions.
151
+ Please use the api :func:`mindspore.runtime.empty_cache` instead.
153
152
 
154
153
  Note:
155
- - The api will be deprecated, please use the api :func:`mindspore.runtime.empty_cache` instead.
154
+ - Empty cache help reduce the fragmentation of device memory.
155
+ - Support Atlas A2 series products.
156
156
 
157
- Note:
158
- Currently, the MindSpore memory pool does not have the function of releasing memory fragments.
159
- This interface is reserved but implemented as an empty method and prompted in log mode.
157
+ Supported Platforms:
158
+ ``Ascend``
160
159
  """
161
160
  if not function_memory_status['empty_cache']:
162
161
  function_memory_status['empty_cache'] = True
163
- logger.warning(f"The empty_cache operation is currently not supported.")
162
+ device_target = ms.context.get_context("device_target")
163
+ if not _is_initialized(device_target):
164
+ logger.warning(f"Backend {device_target} is not initialized yet.")
165
+ return
166
+ _empty_cache(device_target)
164
167
 
165
168
 
166
169
  @_check_inputs_validation
167
170
  def reset_peak_memory_stats(device_target=None):
168
171
  """
169
- Reset the "peak" stats tracked by memory manager.
170
-
171
- Note:
172
- - The api will be deprecated, please use the api :func:`mindspore.runtime.reset_peak_memory_stats` instead.
173
-
174
- Note:
175
- If `device_target` is not specified, get the device capability of the current backend set by context.
172
+ Reset the "peak" stats tracked by memory manager, this api will be deprecated and removed in future versions.
173
+ Please use the api :func:`mindspore.runtime.reset_peak_memory_stats` instead.
176
174
 
177
175
  Args:
178
- device_target (str, optional): The device name of backend, should be one of "CPU", "GPU" and "Ascend".
179
- Default value: ``None``.
176
+ device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
177
+ ``"Ascend"`` . Default ``None`` , represents the current device set by context.
180
178
 
181
179
  Examples:
182
- >>> import mindspore as ms
183
- >>> import numpy as np
184
- >>> from mindspore import Tensor, ops
185
- >>> a = Tensor(np.ones([1, 2]), ms.float32)
186
- >>> b = Tensor(np.ones([1, 2]), ms.float32)
187
- >>> c = ops.add(a, b).asnumpy()
188
- >>> print(ms.hal.max_memory_reserved())
180
+ >>> import mindspore
181
+ >>> a = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
182
+ >>> b = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
183
+ >>> c = mindspore.ops.add(a, b).asnumpy()
184
+ >>> print(mindspore.hal.max_memory_reserved())
189
185
  1073741824
190
- >>> print(ms.hal.max_memory_allocated())
186
+ >>> print(mindspore.hal.max_memory_allocated())
191
187
  1536
192
- >>> ms.hal.reset_peak_memory_stats()
193
- >>> print(ms.hal.max_memory_reserved())
188
+ >>> mindspore.hal.reset_peak_memory_stats()
189
+ >>> print(mindspore.hal.max_memory_reserved())
194
190
  0
195
- >>> print(ms.hal.max_memory_allocated())
191
+ >>> print(mindspore.hal.max_memory_allocated())
196
192
  0
197
193
  """
198
194
  if not function_memory_status['reset_peak_memory_stats']:
@@ -208,17 +204,12 @@ def reset_peak_memory_stats(device_target=None):
208
204
  @_check_inputs_validation
209
205
  def memory_summary(device_target=None):
210
206
  """
211
- Returns readable memory pool status information.
212
-
213
- Note:
214
- - The api will be deprecated, please use the api :func:`mindspore.runtime.memory_summary` instead.
215
-
216
- Note:
217
- If `device_target` is not specified, get the device capability of the current backend set by context.
207
+ Returns readable memory pool status information, this api will be deprecated and removed in future versions.
208
+ Please use the api :func:`mindspore.runtime.memory_summary` instead.
218
209
 
219
210
  Args:
220
- device_target (str, optional): The device name of backend, should be one of "CPU", "GPU" and "Ascend".
221
- Default value: ``None``.
211
+ device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
212
+ ``"Ascend"`` . Default ``None`` , represents the current device set by context.
222
213
 
223
214
  Returns:
224
215
  str, readable memory pool status information in tabular form.
@@ -270,30 +261,25 @@ def memory_summary(device_target=None):
270
261
  @_check_inputs_validation
271
262
  def memory_allocated(device_target=None):
272
263
  """
273
- Returns the actual memory size currently occupied by Tensor.
274
-
275
- Note:
276
- - The api will be deprecated, please use the api :func:`mindspore.runtime.memory_allocated` instead.
264
+ Returns the actual memory size currently occupied by Tensor, this api will be deprecated and removed in future
265
+ versions, please use the api :func:`mindspore.runtime.memory_allocated` instead.
277
266
 
278
267
  Note:
279
- - If `device_target` is not specified, get the device capability of the current backend set by context.
280
- - For the `CPU` backend, 0 is always returned.
268
+ - For the `CPU` device, 0 is always returned.
281
269
 
282
270
  Args:
283
- device_target (str, optional): The device name of backend, should be one of "CPU", "GPU" and "Ascend".
284
- Default value: ``None``.
271
+ device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
272
+ ``"Ascend"`` . Default ``None`` , represents the current device set by context.
285
273
 
286
274
  Returns:
287
275
  int, in Byte.
288
276
 
289
277
  Examples:
290
- >>> import mindspore as ms
291
- >>> import numpy as np
292
- >>> from mindspore import Tensor, ops
293
- >>> a = Tensor(np.ones([1, 2]), ms.float32)
294
- >>> b = Tensor(np.ones([1, 2]), ms.float32)
295
- >>> c = ops.add(a, b).asnumpy()
296
- >>> print(ms.hal.memory_allocated())
278
+ >>> import mindspore
279
+ >>> a = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
280
+ >>> b = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
281
+ >>> c = mindspore.ops.add(a, b).asnumpy()
282
+ >>> print(mindspore.hal.memory_allocated())
297
283
  1024
298
284
  """
299
285
  if not function_memory_status['memory_allocated']:
@@ -308,30 +294,26 @@ def memory_allocated(device_target=None):
308
294
  @_check_inputs_validation
309
295
  def max_memory_allocated(device_target=None):
310
296
  """
311
- Returns the peak memory size of the memory pool actually occupied by Tensor since the process was started.
312
-
313
- Note:
314
- - The api will be deprecated, please use the api :func:`mindspore.runtime.max_memory_allocated` instead.
297
+ Return the peak memory size of the memory pool actually occupied by Tensor since the process was started.
298
+ This api will be deprecated and removed in future versions, please use
299
+ the api :func:`mindspore.runtime.max_memory_allocated` instead.
315
300
 
316
301
  Note:
317
- - If `device_target` is not specified, get the device capability of the current backend set by context.
318
- - For the `CPU` backend, 0 is always returned.
302
+ - For the `CPU` device, 0 is always returned.
319
303
 
320
304
  Args:
321
- device_target (str, optional): The device name of backend, should be one of "CPU", "GPU" and "Ascend".
322
- Default value: ``None``.
305
+ device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
306
+ ``"Ascend"`` . Default ``None`` , represents the current device set by context.
323
307
 
324
308
  Returns:
325
309
  int, in Byte.
326
310
 
327
311
  Examples:
328
- >>> import mindspore as ms
329
- >>> import numpy as np
330
- >>> from mindspore import Tensor, ops
331
- >>> a = Tensor(np.ones([1, 2]), ms.float32)
332
- >>> b = Tensor(np.ones([1, 2]), ms.float32)
333
- >>> c = ops.add(a, b).asnumpy()
334
- >>> print(ms.hal.max_memory_allocated())
312
+ >>> import mindspore
313
+ >>> a = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
314
+ >>> b = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
315
+ >>> c = mindspore.ops.add(a, b).asnumpy()
316
+ >>> print(mindspore.hal.max_memory_allocated())
335
317
  1536
336
318
  """
337
319
  if not function_memory_status['max_memory_allocated']:
@@ -346,29 +328,22 @@ def max_memory_allocated(device_target=None):
346
328
  @_check_inputs_validation
347
329
  def reset_max_memory_reserved(device_target=None):
348
330
  """
349
- Reset the peak memory size managed by the memory pool.
350
-
351
- Note:
352
- - The api will be deprecated, please use the api :func:`mindspore.runtime.reset_max_memory_reserved` instead.
353
-
354
- Note:
355
- If `device_target` is not specified, get the device capability of the current backend set by context.
331
+ Reset the peak memory size managed by the memory pool, this api will be deprecated and removed in future versions.
332
+ Please use the api :func:`mindspore.runtime.reset_max_memory_reserved` instead.
356
333
 
357
334
  Args:
358
- device_target (str, optional): The device name of backend, should be one of "CPU", "GPU" and "Ascend".
359
- Default value: ``None``.
335
+ device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
336
+ ``"Ascend"`` . Default ``None`` , represents the current device set by context.
360
337
 
361
338
  Examples:
362
- >>> import mindspore as ms
363
- >>> import numpy as np
364
- >>> from mindspore import Tensor, ops
365
- >>> a = Tensor(np.ones([1, 2]), ms.float32)
366
- >>> b = Tensor(np.ones([1, 2]), ms.float32)
367
- >>> c = ops.add(a, b).asnumpy()
368
- >>> print(ms.hal.max_memory_reserved())
339
+ >>> import mindspore
340
+ >>> a = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
341
+ >>> b = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
342
+ >>> c = mindspore.ops.add(a, b).asnumpy()
343
+ >>> print(mindspore.hal.max_memory_reserved())
369
344
  1073741824
370
- >>> ms.hal.reset_max_memory_reserved()
371
- >>> print(ms.hal.max_memory_reserved())
345
+ >>> mindspore.hal.reset_max_memory_reserved()
346
+ >>> print(mindspore.hal.max_memory_reserved())
372
347
  0
373
348
  """
374
349
  if not function_memory_status['reset_max_memory_reserved']:
@@ -383,29 +358,22 @@ def reset_max_memory_reserved(device_target=None):
383
358
  @_check_inputs_validation
384
359
  def reset_max_memory_allocated(device_target=None):
385
360
  """
386
- Reset the peak memory size of the memory pool actually occupied by Tensor.
387
-
388
- Note:
389
- - The api will be deprecated, please use the api :func:`mindspore.runtime.reset_max_memory_allocated` instead.
390
-
391
- Note:
392
- If `device_target` is not specified, get the device capability of the current backend set by context.
361
+ Reset the peak memory size of the memory pool actually occupied by Tensor, this api will be deprecated and removed
362
+ in future versions, please use the api :func:`mindspore.runtime.reset_max_memory_allocated` instead.
393
363
 
394
364
  Args:
395
- device_target (str, optional): The device name of backend, should be one of "CPU", "GPU" and "Ascend".
396
- Default value: ``None``.
365
+ device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
366
+ ``"Ascend"`` . Default ``None`` , represents the current device set by context.
397
367
 
398
368
  Examples:
399
- >>> import mindspore as ms
400
- >>> import numpy as np
401
- >>> from mindspore import Tensor, ops
402
- >>> a = Tensor(np.ones([1, 2]), ms.float32)
403
- >>> b = Tensor(np.ones([1, 2]), ms.float32)
404
- >>> c = ops.add(a, b).asnumpy()
405
- >>> print(ms.hal.max_memory_allocated())
369
+ >>> import mindspore
370
+ >>> a = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
371
+ >>> b = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
372
+ >>> c = mindspore.ops.add(a, b).asnumpy()
373
+ >>> print(mindspore.hal.max_memory_allocated())
406
374
  1536
407
- >>> ms.hal.reset_max_memory_allocated()
408
- >>> print(ms.hal.max_memory_allocated())
375
+ >>> mindspore.hal.reset_max_memory_allocated()
376
+ >>> print(mindspore.hal.max_memory_allocated())
409
377
  0
410
378
  """
411
379
  if not function_memory_status['reset_max_memory_allocated']: