mindspore 2.5.0__cp311-cp311-win_amd64.whl → 2.6.0rc1__cp311-cp311-win_amd64.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of mindspore might be problematic. Click here for more details.

Files changed (491) hide show
  1. mindspore/.commit_id +1 -1
  2. mindspore/Microsoft.VisualStudio.Telemetry.dll +0 -0
  3. mindspore/Newtonsoft.Json.dll +0 -0
  4. mindspore/__init__.py +6 -4
  5. mindspore/_c_dataengine.cp311-win_amd64.pyd +0 -0
  6. mindspore/_c_expression.cp311-win_amd64.pyd +0 -0
  7. mindspore/_c_mindrecord.cp311-win_amd64.pyd +0 -0
  8. mindspore/_check_jit_forbidden_api.py +3 -0
  9. mindspore/_checkparam.py +3 -33
  10. mindspore/_deprecated/__init__.py +17 -0
  11. mindspore/_deprecated/jit.py +198 -0
  12. mindspore/_extends/builtin_operations.py +1 -1
  13. mindspore/_extends/parse/__init__.py +6 -7
  14. mindspore/_extends/parse/compile_config.py +19 -0
  15. mindspore/_extends/parse/deprecated/deprecated_tensor_method.py +22 -3
  16. mindspore/_extends/parse/jit_fallback_modules/__init__.py +0 -0
  17. mindspore/_extends/parse/jit_fallback_modules/check_utils.py +123 -0
  18. mindspore/_extends/parse/jit_fallback_modules/third_party_modules.py +50 -0
  19. mindspore/_extends/parse/parser.py +24 -193
  20. mindspore/_extends/parse/resources.py +1 -5
  21. mindspore/_extends/parse/standard_method.py +97 -74
  22. mindspore/_extends/pijit/__init__.py +2 -2
  23. mindspore/_extends/pijit/pijit_func_white_list.py +16 -11
  24. mindspore/_extends/pijit/tensor_func_list.py +27 -0
  25. mindspore/_extends/utils.py +1 -1
  26. mindspore/amp.py +4 -4
  27. mindspore/atlprov.dll +0 -0
  28. mindspore/avcodec-59.dll +0 -0
  29. mindspore/avdevice-59.dll +0 -0
  30. mindspore/avfilter-8.dll +0 -0
  31. mindspore/avformat-59.dll +0 -0
  32. mindspore/avutil-57.dll +0 -0
  33. mindspore/boost/__init__.py +2 -2
  34. mindspore/boost/base.py +3 -7
  35. mindspore/boost/boost_cell_wrapper.py +2 -2
  36. mindspore/c1.dll +0 -0
  37. mindspore/c1xx.dll +0 -0
  38. mindspore/c2.dll +0 -0
  39. mindspore/common/__init__.py +4 -3
  40. mindspore/common/_grad_function.py +56 -0
  41. mindspore/common/_pijit_context.py +14 -5
  42. mindspore/common/_register_for_tensor.py +1 -1
  43. mindspore/common/_stub_tensor.py +5 -10
  44. mindspore/common/_tensor_cpp_method.py +1 -1
  45. mindspore/common/_tensor_docs.py +1915 -3287
  46. mindspore/common/api.py +341 -354
  47. mindspore/common/auto_dynamic_shape.py +41 -44
  48. mindspore/common/dtype.py +5 -2
  49. mindspore/common/dump.py +7 -5
  50. mindspore/common/file_system.py +3 -0
  51. mindspore/common/hook_handle.py +5 -3
  52. mindspore/common/initializer.py +10 -6
  53. mindspore/common/jit_begin_end.py +94 -0
  54. mindspore/common/jit_config.py +6 -1
  55. mindspore/common/jit_context.py +76 -0
  56. mindspore/common/jit_trace.py +378 -0
  57. mindspore/common/lazy_inline.py +2 -2
  58. mindspore/common/mutable.py +5 -4
  59. mindspore/common/parameter.py +106 -39
  60. mindspore/common/seed.py +2 -2
  61. mindspore/common/sparse_tensor.py +23 -17
  62. mindspore/common/tensor.py +297 -714
  63. mindspore/communication/__init__.py +7 -5
  64. mindspore/communication/_comm_helper.py +47 -2
  65. mindspore/communication/comm_func.py +70 -53
  66. mindspore/communication/management.py +83 -17
  67. mindspore/context.py +214 -560
  68. mindspore/dataset/__init__.py +44 -20
  69. mindspore/dataset/audio/__init__.py +2 -8
  70. mindspore/dataset/audio/transforms.py +3 -17
  71. mindspore/dataset/core/config.py +3 -3
  72. mindspore/dataset/engine/cache_client.py +1 -1
  73. mindspore/dataset/engine/datasets.py +102 -120
  74. mindspore/dataset/engine/datasets_audio.py +22 -22
  75. mindspore/dataset/engine/datasets_standard_format.py +43 -24
  76. mindspore/dataset/engine/datasets_text.py +78 -85
  77. mindspore/dataset/engine/datasets_user_defined.py +108 -76
  78. mindspore/dataset/engine/datasets_vision.py +111 -108
  79. mindspore/dataset/engine/iterators.py +5 -3
  80. mindspore/dataset/engine/obs/obs_mindrecord_dataset.py +1 -1
  81. mindspore/dataset/engine/samplers.py +279 -57
  82. mindspore/dataset/engine/serializer_deserializer.py +2 -1
  83. mindspore/dataset/engine/validators.py +10 -0
  84. mindspore/dataset/text/__init__.py +7 -6
  85. mindspore/dataset/text/transforms.py +6 -5
  86. mindspore/dataset/text/utils.py +3 -3
  87. mindspore/dataset/transforms/__init__.py +0 -9
  88. mindspore/dataset/transforms/transforms.py +3 -3
  89. mindspore/dataset/utils/browse_dataset.py +1 -1
  90. mindspore/dataset/vision/__init__.py +2 -9
  91. mindspore/dataset/vision/transforms.py +202 -158
  92. mindspore/dataset/vision/utils.py +7 -5
  93. mindspore/device_context/ascend/op_debug.py +60 -1
  94. mindspore/device_context/ascend/op_tuning.py +0 -4
  95. mindspore/device_manager.py +39 -3
  96. mindspore/dnnl.dll +0 -0
  97. mindspore/dpcmi.dll +0 -0
  98. mindspore/experimental/es/embedding_service.py +35 -27
  99. mindspore/experimental/map_parameter.py +4 -4
  100. mindspore/experimental/optim/adadelta.py +22 -26
  101. mindspore/experimental/optim/adagrad.py +4 -4
  102. mindspore/experimental/optim/adam.py +4 -0
  103. mindspore/experimental/optim/adamax.py +4 -4
  104. mindspore/experimental/optim/adamw.py +4 -0
  105. mindspore/experimental/optim/asgd.py +1 -1
  106. mindspore/experimental/optim/lr_scheduler.py +40 -22
  107. mindspore/experimental/optim/radam.py +5 -5
  108. mindspore/experimental/optim/rprop.py +1 -1
  109. mindspore/experimental/optim/sgd.py +1 -1
  110. mindspore/hal/contiguous_tensors_handle.py +6 -10
  111. mindspore/hal/device.py +55 -81
  112. mindspore/hal/event.py +38 -55
  113. mindspore/hal/memory.py +93 -144
  114. mindspore/hal/stream.py +81 -125
  115. mindspore/include/dataset/constants.h +7 -4
  116. mindspore/include/dataset/execute.h +2 -2
  117. mindspore/jpeg62.dll +0 -0
  118. mindspore/log.py +40 -2
  119. mindspore/mindrecord/__init__.py +20 -7
  120. mindspore/mindspore_backend_common.dll +0 -0
  121. mindspore/mindspore_backend_manager.dll +0 -0
  122. mindspore/mindspore_common.dll +0 -0
  123. mindspore/mindspore_core.dll +0 -0
  124. mindspore/mindspore_dump.dll +0 -0
  125. mindspore/mindspore_frontend.dll +0 -0
  126. mindspore/mindspore_glog.dll +0 -0
  127. mindspore/mindspore_memory_pool.dll +0 -0
  128. mindspore/mindspore_ms_backend.dll +0 -0
  129. mindspore/mindspore_ops.dll +0 -0
  130. mindspore/{mindspore_backend.dll → mindspore_ops_host.dll} +0 -0
  131. mindspore/mindspore_ops_kernel_common.dll +0 -0
  132. mindspore/mindspore_profiler.dll +0 -0
  133. mindspore/mindspore_pyboost.dll +0 -0
  134. mindspore/mindspore_pynative.dll +0 -0
  135. mindspore/mindspore_res_manager.dll +0 -0
  136. mindspore/mindspore_runtime_pipeline.dll +0 -0
  137. mindspore/mint/__init__.py +131 -700
  138. mindspore/mint/distributed/__init__.py +5 -1
  139. mindspore/mint/distributed/distributed.py +194 -109
  140. mindspore/mint/linalg/__init__.py +2 -0
  141. mindspore/mint/nn/__init__.py +280 -18
  142. mindspore/mint/nn/functional.py +282 -64
  143. mindspore/mint/nn/layer/__init__.py +4 -0
  144. mindspore/mint/nn/layer/_functions.py +7 -3
  145. mindspore/mint/nn/layer/activation.py +120 -13
  146. mindspore/mint/nn/layer/conv.py +218 -24
  147. mindspore/mint/nn/layer/normalization.py +15 -16
  148. mindspore/mint/nn/layer/padding.py +1 -1
  149. mindspore/mint/nn/layer/pooling.py +66 -1
  150. mindspore/mint/optim/__init__.py +2 -1
  151. mindspore/mint/optim/sgd.py +171 -0
  152. mindspore/msobj140.dll +0 -0
  153. mindspore/mspdb140.dll +0 -0
  154. mindspore/mspdbcore.dll +0 -0
  155. mindspore/mspdbst.dll +0 -0
  156. mindspore/mspft140.dll +0 -0
  157. mindspore/msvcdis140.dll +0 -0
  158. mindspore/msvcp140_1.dll +0 -0
  159. mindspore/msvcp140_2.dll +0 -0
  160. mindspore/msvcp140_atomic_wait.dll +0 -0
  161. mindspore/msvcp140_codecvt_ids.dll +0 -0
  162. mindspore/nn/__init__.py +4 -1
  163. mindspore/nn/cell.py +1250 -176
  164. mindspore/nn/layer/activation.py +23 -21
  165. mindspore/nn/layer/basic.py +22 -16
  166. mindspore/nn/layer/container.py +1 -1
  167. mindspore/nn/layer/conv.py +22 -17
  168. mindspore/nn/layer/embedding.py +9 -8
  169. mindspore/nn/layer/normalization.py +48 -42
  170. mindspore/nn/layer/pooling.py +75 -31
  171. mindspore/nn/layer/transformer.py +11 -10
  172. mindspore/nn/learning_rate_schedule.py +4 -2
  173. mindspore/nn/loss/loss.py +27 -19
  174. mindspore/nn/optim/ada_grad.py +6 -5
  175. mindspore/nn/optim/adadelta.py +9 -7
  176. mindspore/nn/optim/adafactor.py +1 -1
  177. mindspore/nn/optim/adam.py +16 -12
  178. mindspore/nn/optim/adamax.py +8 -7
  179. mindspore/nn/optim/adasum.py +5 -5
  180. mindspore/nn/optim/asgd.py +1 -1
  181. mindspore/nn/optim/ftrl.py +11 -9
  182. mindspore/nn/optim/lamb.py +1 -1
  183. mindspore/nn/optim/lazyadam.py +12 -10
  184. mindspore/nn/optim/momentum.py +7 -6
  185. mindspore/nn/optim/optimizer.py +2 -2
  186. mindspore/nn/optim/proximal_ada_grad.py +12 -10
  187. mindspore/nn/optim/rmsprop.py +13 -12
  188. mindspore/nn/optim/rprop.py +9 -7
  189. mindspore/nn/optim/sgd.py +9 -6
  190. mindspore/nn/optim/tft_wrapper.py +5 -2
  191. mindspore/nn/probability/bijector/bijector.py +17 -11
  192. mindspore/nn/probability/bijector/gumbel_cdf.py +5 -5
  193. mindspore/nn/probability/bijector/invert.py +2 -2
  194. mindspore/nn/probability/bijector/scalar_affine.py +3 -3
  195. mindspore/nn/probability/bijector/softplus.py +3 -2
  196. mindspore/nn/probability/distribution/beta.py +3 -3
  197. mindspore/nn/probability/distribution/categorical.py +1 -1
  198. mindspore/nn/probability/distribution/cauchy.py +4 -2
  199. mindspore/nn/probability/distribution/exponential.py +6 -7
  200. mindspore/nn/probability/distribution/gamma.py +2 -2
  201. mindspore/nn/probability/distribution/gumbel.py +2 -2
  202. mindspore/nn/probability/distribution/half_normal.py +5 -3
  203. mindspore/nn/probability/distribution/logistic.py +5 -3
  204. mindspore/nn/probability/distribution/poisson.py +1 -1
  205. mindspore/nn/probability/distribution/uniform.py +5 -3
  206. mindspore/nn/reinforcement/_tensors_queue.py +1 -1
  207. mindspore/nn/reinforcement/tensor_array.py +1 -1
  208. mindspore/nn/wrap/__init__.py +6 -6
  209. mindspore/nn/wrap/cell_wrapper.py +178 -117
  210. mindspore/nn/wrap/grad_reducer.py +45 -36
  211. mindspore/nn/wrap/loss_scale.py +3 -3
  212. mindspore/numpy/array_creations.py +3 -3
  213. mindspore/numpy/array_ops.py +1 -1
  214. mindspore/numpy/math_ops.py +4 -4
  215. mindspore/numpy/utils.py +1 -2
  216. mindspore/numpy/utils_const.py +1 -2
  217. mindspore/opencv_core452.dll +0 -0
  218. mindspore/opencv_imgcodecs452.dll +0 -0
  219. mindspore/opencv_imgproc452.dll +0 -0
  220. mindspore/ops/__init__.py +3 -2
  221. mindspore/ops/_grad_experimental/grad_comm_ops.py +18 -3
  222. mindspore/ops/_grad_experimental/grad_debug_ops.py +8 -1
  223. mindspore/ops/_grad_experimental/taylor_rule.py +29 -0
  224. mindspore/ops/_register_for_op.py +0 -11
  225. mindspore/{ops_generate → ops/_utils}/arg_dtype_cast.py +123 -4
  226. mindspore/{ops_generate → ops/_utils}/arg_handler.py +3 -4
  227. mindspore/ops/_vmap/vmap_array_ops.py +7 -6
  228. mindspore/ops/_vmap/vmap_grad_nn_ops.py +2 -1
  229. mindspore/ops/_vmap/vmap_math_ops.py +4 -7
  230. mindspore/ops/_vmap/vmap_nn_ops.py +9 -8
  231. mindspore/ops/auto_generate/__init__.py +4 -3
  232. mindspore/ops/auto_generate/cpp_create_prim_instance_helper.py +102 -49
  233. mindspore/ops/auto_generate/gen_extend_func.py +281 -135
  234. mindspore/ops/auto_generate/gen_ops_def.py +2574 -2326
  235. mindspore/ops/auto_generate/gen_ops_prim.py +8566 -2755
  236. mindspore/ops/auto_generate/pyboost_inner_prim.py +106 -76
  237. mindspore/ops/composite/__init__.py +2 -1
  238. mindspore/ops/composite/base.py +19 -24
  239. mindspore/ops/composite/math_ops.py +6 -16
  240. mindspore/ops/composite/multitype_ops/__init__.py +5 -2
  241. mindspore/ops/composite/multitype_ops/_compile_utils.py +2 -3
  242. mindspore/ops/composite/multitype_ops/_constexpr_utils.py +1 -2
  243. mindspore/ops/composite/multitype_ops/add_impl.py +2 -1
  244. mindspore/ops/composite/multitype_ops/bitwise_and_impl.py +2 -1
  245. mindspore/ops/composite/multitype_ops/bitwise_or_impl.py +2 -1
  246. mindspore/ops/composite/multitype_ops/bitwise_xor_impl.py +2 -1
  247. mindspore/ops/composite/multitype_ops/div_impl.py +6 -4
  248. mindspore/ops/composite/multitype_ops/equal_impl.py +4 -3
  249. mindspore/ops/composite/multitype_ops/floordiv_impl.py +2 -1
  250. mindspore/ops/composite/multitype_ops/getitem_impl.py +3 -2
  251. mindspore/ops/composite/multitype_ops/greater_equal_impl.py +4 -3
  252. mindspore/ops/composite/multitype_ops/greater_impl.py +4 -3
  253. mindspore/ops/composite/multitype_ops/in_impl.py +2 -1
  254. mindspore/ops/composite/multitype_ops/invert_impl.py +50 -0
  255. mindspore/ops/composite/multitype_ops/left_shift_impl.py +2 -1
  256. mindspore/ops/composite/multitype_ops/less_equal_impl.py +4 -3
  257. mindspore/ops/composite/multitype_ops/less_impl.py +4 -3
  258. mindspore/ops/composite/multitype_ops/logic_not_impl.py +3 -2
  259. mindspore/ops/composite/multitype_ops/logical_and_impl.py +2 -1
  260. mindspore/ops/composite/multitype_ops/logical_or_impl.py +2 -1
  261. mindspore/ops/composite/multitype_ops/mod_impl.py +2 -1
  262. mindspore/ops/composite/multitype_ops/mul_impl.py +3 -2
  263. mindspore/ops/composite/multitype_ops/negative_impl.py +2 -1
  264. mindspore/ops/composite/multitype_ops/not_equal_impl.py +2 -1
  265. mindspore/ops/composite/multitype_ops/not_in_impl.py +2 -1
  266. mindspore/ops/composite/multitype_ops/ones_like_impl.py +18 -0
  267. mindspore/ops/composite/multitype_ops/pow_impl.py +2 -1
  268. mindspore/ops/composite/multitype_ops/right_shift_impl.py +2 -1
  269. mindspore/ops/composite/multitype_ops/setitem_impl.py +2 -1
  270. mindspore/ops/composite/multitype_ops/sub_impl.py +2 -1
  271. mindspore/ops/function/__init__.py +28 -2
  272. mindspore/ops/function/_add_attr_func.py +58 -0
  273. mindspore/ops/function/array_func.py +1629 -2345
  274. mindspore/ops/function/clip_func.py +38 -45
  275. mindspore/ops/function/debug_func.py +36 -44
  276. mindspore/ops/function/grad/__init__.py +1 -0
  277. mindspore/ops/function/grad/grad_func.py +104 -71
  278. mindspore/ops/function/image_func.py +1 -1
  279. mindspore/ops/function/linalg_func.py +46 -78
  280. mindspore/ops/function/math_func.py +3035 -3705
  281. mindspore/ops/function/nn_func.py +676 -241
  282. mindspore/ops/function/other_func.py +159 -1
  283. mindspore/ops/function/parameter_func.py +17 -30
  284. mindspore/ops/function/random_func.py +204 -361
  285. mindspore/ops/function/reshard_func.py +4 -70
  286. mindspore/ops/function/sparse_func.py +3 -3
  287. mindspore/ops/function/sparse_unary_func.py +5 -5
  288. mindspore/ops/function/spectral_func.py +25 -58
  289. mindspore/ops/function/vmap_func.py +24 -17
  290. mindspore/ops/functional.py +6 -4
  291. mindspore/ops/functional_overload.py +547 -4
  292. mindspore/ops/op_info_register.py +32 -244
  293. mindspore/ops/operations/__init__.py +10 -5
  294. mindspore/ops/operations/_custom_ops_utils.py +247 -0
  295. mindspore/ops/operations/_grad_ops.py +1 -10
  296. mindspore/ops/operations/_inner_ops.py +5 -76
  297. mindspore/ops/operations/_ms_kernel.py +4 -10
  298. mindspore/ops/operations/_rl_inner_ops.py +1 -1
  299. mindspore/ops/operations/_scalar_ops.py +3 -2
  300. mindspore/ops/operations/_sequence_ops.py +1 -1
  301. mindspore/ops/operations/_tensor_array.py +1 -1
  302. mindspore/ops/operations/array_ops.py +37 -22
  303. mindspore/ops/operations/comm_ops.py +150 -107
  304. mindspore/ops/operations/custom_ops.py +221 -23
  305. mindspore/ops/operations/debug_ops.py +115 -16
  306. mindspore/ops/operations/inner_ops.py +1 -1
  307. mindspore/ops/operations/linalg_ops.py +1 -58
  308. mindspore/ops/operations/manually_defined/_inner.py +1 -1
  309. mindspore/ops/operations/manually_defined/ops_def.py +746 -79
  310. mindspore/ops/operations/math_ops.py +21 -18
  311. mindspore/ops/operations/nn_ops.py +65 -191
  312. mindspore/ops/operations/other_ops.py +62 -9
  313. mindspore/ops/operations/random_ops.py +13 -7
  314. mindspore/ops/operations/reshard_ops.py +1 -1
  315. mindspore/ops/operations/sparse_ops.py +2 -2
  316. mindspore/ops/primitive.py +43 -32
  317. mindspore/ops/tensor_method.py +232 -13
  318. mindspore/ops_generate/__init__.py +0 -5
  319. mindspore/ops_generate/aclnn/__init__.py +0 -0
  320. mindspore/ops_generate/{aclnn_kernel_register_auto_cc_generator.py → aclnn/aclnn_kernel_register_auto_cc_generator.py} +43 -18
  321. mindspore/ops_generate/{gen_aclnn_implement.py → aclnn/gen_aclnn_implement.py} +49 -51
  322. mindspore/ops_generate/api/__init__.py +0 -0
  323. mindspore/ops_generate/{add_tensor_docs_generator.py → api/add_tensor_docs_generator.py} +9 -7
  324. mindspore/ops_generate/{cpp_create_prim_instance_helper_generator.py → api/cpp_create_prim_instance_helper_generator.py} +6 -9
  325. mindspore/ops_generate/{functional_map_cpp_generator.py → api/functional_map_cpp_generator.py} +25 -12
  326. mindspore/ops_generate/{functional_overload_py_generator.py → api/functional_overload_py_generator.py} +8 -6
  327. mindspore/ops_generate/{functions_cc_generator.py → api/functions_cc_generator.py} +14 -10
  328. mindspore/ops_generate/api/gen_api.py +103 -0
  329. mindspore/ops_generate/{op_api_proto.py → api/op_api_proto.py} +98 -69
  330. mindspore/ops_generate/{tensor_func_reg_cpp_generator.py → api/tensor_func_reg_cpp_generator.py} +82 -43
  331. mindspore/ops_generate/common/__init__.py +0 -0
  332. mindspore/ops_generate/common/gen_constants.py +91 -0
  333. mindspore/ops_generate/{gen_utils.py → common/gen_utils.py} +72 -19
  334. mindspore/ops_generate/{op_proto.py → common/op_proto.py} +64 -1
  335. mindspore/ops_generate/{template.py → common/template.py} +96 -84
  336. mindspore/ops_generate/gen_ops.py +23 -325
  337. mindspore/ops_generate/op_def/__init__.py +0 -0
  338. mindspore/ops_generate/op_def/gen_op_def.py +90 -0
  339. mindspore/ops_generate/{lite_ops_cpp_generator.py → op_def/lite_ops_cpp_generator.py} +47 -11
  340. mindspore/ops_generate/{ops_def_cc_generator.py → op_def/ops_def_cc_generator.py} +18 -7
  341. mindspore/ops_generate/{ops_def_h_generator.py → op_def/ops_def_h_generator.py} +5 -5
  342. mindspore/ops_generate/{ops_name_h_generator.py → op_def/ops_name_h_generator.py} +30 -15
  343. mindspore/ops_generate/op_def/ops_primitive_h_generator.py +125 -0
  344. mindspore/ops_generate/op_def_py/__init__.py +0 -0
  345. mindspore/ops_generate/op_def_py/gen_op_def_py.py +47 -0
  346. mindspore/ops_generate/{op_def_py_generator.py → op_def_py/op_def_py_generator.py} +6 -5
  347. mindspore/ops_generate/{op_prim_py_generator.py → op_def_py/op_prim_py_generator.py} +24 -15
  348. mindspore/ops_generate/pyboost/__init__.py +0 -0
  349. mindspore/ops_generate/{auto_grad_impl_cc_generator.py → pyboost/auto_grad_impl_cc_generator.py} +11 -7
  350. mindspore/ops_generate/{auto_grad_reg_cc_generator.py → pyboost/auto_grad_reg_cc_generator.py} +7 -7
  351. mindspore/ops_generate/{gen_pyboost_func.py → pyboost/gen_pyboost_func.py} +40 -16
  352. mindspore/ops_generate/{op_template_parser.py → pyboost/op_template_parser.py} +105 -24
  353. mindspore/ops_generate/{pyboost_functions_cpp_generator.py → pyboost/pyboost_functions_cpp_generator.py} +55 -18
  354. mindspore/ops_generate/{pyboost_functions_h_generator.py → pyboost/pyboost_functions_h_generator.py} +42 -10
  355. mindspore/ops_generate/{pyboost_functions_py_generator.py → pyboost/pyboost_functions_py_generator.py} +6 -6
  356. mindspore/ops_generate/{pyboost_grad_function_cpp_generator.py → pyboost/pyboost_grad_function_cpp_generator.py} +11 -10
  357. mindspore/ops_generate/{pyboost_inner_prim_generator.py → pyboost/pyboost_inner_prim_generator.py} +8 -7
  358. mindspore/ops_generate/{pyboost_native_grad_functions_generator.py → pyboost/pyboost_native_grad_functions_generator.py} +14 -10
  359. mindspore/ops_generate/{pyboost_op_cpp_code_generator.py → pyboost/pyboost_op_cpp_code_generator.py} +140 -53
  360. mindspore/ops_generate/{pyboost_overload_functions_cpp_generator.py → pyboost/pyboost_overload_functions_cpp_generator.py} +28 -15
  361. mindspore/ops_generate/{pyboost_utils.py → pyboost/pyboost_utils.py} +88 -4
  362. mindspore/ops_generate/resources/__init__.py +0 -0
  363. mindspore/ops_generate/resources/resource_list.py +30 -0
  364. mindspore/ops_generate/resources/resource_loader.py +36 -0
  365. mindspore/ops_generate/resources/resource_manager.py +64 -0
  366. mindspore/ops_generate/resources/yaml_loader.py +88 -0
  367. mindspore/ops_generate/tensor_py_cc_generator.py +122 -0
  368. mindspore/parallel/__init__.py +6 -2
  369. mindspore/parallel/_auto_parallel_context.py +133 -6
  370. mindspore/parallel/_cell_wrapper.py +130 -15
  371. mindspore/parallel/_parallel_serialization.py +95 -4
  372. mindspore/parallel/_ps_context.py +1 -1
  373. mindspore/parallel/_recovery_context.py +7 -2
  374. mindspore/parallel/_tensor.py +142 -18
  375. mindspore/parallel/_utils.py +198 -25
  376. mindspore/parallel/algo_parameter_config.py +3 -3
  377. mindspore/parallel/auto_parallel.py +732 -0
  378. mindspore/parallel/checkpoint_convert.py +159 -0
  379. mindspore/parallel/checkpoint_transform.py +656 -37
  380. mindspore/parallel/cluster/process_entity/_api.py +151 -19
  381. mindspore/parallel/cluster/run.py +1 -1
  382. mindspore/parallel/function/__init__.py +24 -0
  383. mindspore/parallel/function/reshard_func.py +259 -0
  384. mindspore/parallel/nn/__init__.py +25 -0
  385. mindspore/parallel/nn/parallel_cell_wrapper.py +263 -0
  386. mindspore/parallel/nn/parallel_grad_reducer.py +169 -0
  387. mindspore/parallel/parameter_broadcast.py +24 -13
  388. mindspore/parallel/shard.py +137 -61
  389. mindspore/parallel/transform_safetensors.py +287 -95
  390. mindspore/pgodb140.dll +0 -0
  391. mindspore/pgort140.dll +0 -0
  392. mindspore/profiler/__init__.py +9 -5
  393. mindspore/profiler/analysis/parser/ascend_cann_parser.py +6 -2
  394. mindspore/profiler/analysis/parser/ms_framework_parser.py +4 -4
  395. mindspore/profiler/analysis/parser/timeline_assembly_factory/ascend_timeline_assembler.py +7 -4
  396. mindspore/profiler/analysis/parser/timeline_assembly_factory/trace_view_container.py +22 -0
  397. mindspore/profiler/analysis/parser/timeline_creator/fwk_timeline_creator.py +3 -3
  398. mindspore/profiler/analysis/parser/timeline_event/fwk_event.py +241 -86
  399. mindspore/profiler/analysis/viewer/ascend_communication_viewer.py +41 -2
  400. mindspore/profiler/analysis/viewer/ascend_kernel_details_viewer.py +33 -35
  401. mindspore/profiler/analysis/viewer/ascend_memory_viewer.py +7 -0
  402. mindspore/profiler/analysis/viewer/ascend_op_memory_viewer.py +8 -3
  403. mindspore/profiler/analysis/viewer/ascend_step_trace_time_viewer.py +141 -30
  404. mindspore/profiler/analysis/viewer/ms_dataset_viewer.py +5 -6
  405. mindspore/profiler/common/ascend_msprof_exporter.py +5 -4
  406. mindspore/profiler/common/constant.py +12 -0
  407. mindspore/profiler/common/msprof_cmd_tool.py +42 -23
  408. mindspore/profiler/common/path_manager.py +24 -0
  409. mindspore/profiler/common/profiler_context.py +26 -2
  410. mindspore/profiler/common/profiler_meta_data.py +74 -0
  411. mindspore/profiler/common/profiler_parameters.py +59 -18
  412. mindspore/profiler/common/profiler_path_manager.py +66 -7
  413. mindspore/profiler/dynamic_profiler.py +112 -79
  414. mindspore/profiler/envprofiler.py +26 -1
  415. mindspore/profiler/experimental_config.py +197 -0
  416. mindspore/profiler/mstx.py +57 -14
  417. mindspore/profiler/platform/npu_profiler.py +33 -7
  418. mindspore/profiler/profiler.py +541 -45
  419. mindspore/profiler/profiler_action_controller.py +1 -1
  420. mindspore/profiler/profiler_interface.py +4 -0
  421. mindspore/profiler/schedule.py +57 -22
  422. mindspore/rewrite/api/node.py +15 -13
  423. mindspore/rewrite/api/symbol_tree.py +1 -1
  424. mindspore/run_check/_check_version.py +25 -14
  425. mindspore/run_check/run_check.py +1 -1
  426. mindspore/runtime/__init__.py +2 -2
  427. mindspore/runtime/executor.py +40 -11
  428. mindspore/runtime/memory.py +25 -8
  429. mindspore/safeguard/rewrite_obfuscation.py +12 -9
  430. mindspore/swresample-4.dll +0 -0
  431. mindspore/swscale-6.dll +0 -0
  432. mindspore/tbbmalloc.dll +0 -0
  433. mindspore/tinyxml2.dll +0 -0
  434. mindspore/train/__init__.py +8 -8
  435. mindspore/train/_utils.py +35 -7
  436. mindspore/train/amp.py +1 -1
  437. mindspore/train/callback/__init__.py +2 -2
  438. mindspore/train/callback/_callback.py +2 -16
  439. mindspore/train/callback/_checkpoint.py +24 -40
  440. mindspore/train/callback/_cluster_monitor.py +14 -18
  441. mindspore/train/callback/_flops_collector.py +2 -3
  442. mindspore/train/callback/_history.py +7 -4
  443. mindspore/train/callback/_lambda_callback.py +2 -2
  444. mindspore/train/callback/_landscape.py +0 -3
  445. mindspore/train/callback/_loss_monitor.py +2 -1
  446. mindspore/train/callback/_on_request_exit.py +6 -5
  447. mindspore/train/callback/_reduce_lr_on_plateau.py +11 -6
  448. mindspore/train/callback/_summary_collector.py +8 -13
  449. mindspore/train/callback/_time_monitor.py +2 -1
  450. mindspore/train/callback/{_tft_register.py → _train_fault_tolerance.py} +179 -103
  451. mindspore/train/data_sink.py +25 -2
  452. mindspore/train/dataset_helper.py +4 -5
  453. mindspore/train/loss_scale_manager.py +8 -7
  454. mindspore/train/metrics/accuracy.py +3 -3
  455. mindspore/train/metrics/confusion_matrix.py +9 -9
  456. mindspore/train/metrics/error.py +3 -3
  457. mindspore/train/metrics/hausdorff_distance.py +4 -4
  458. mindspore/train/metrics/mean_surface_distance.py +3 -3
  459. mindspore/train/metrics/metric.py +0 -12
  460. mindspore/train/metrics/occlusion_sensitivity.py +4 -2
  461. mindspore/train/metrics/precision.py +8 -6
  462. mindspore/train/metrics/recall.py +9 -9
  463. mindspore/train/metrics/root_mean_square_surface_distance.py +2 -2
  464. mindspore/train/mind_ir_pb2.py +19 -12
  465. mindspore/train/model.py +176 -103
  466. mindspore/train/serialization.py +246 -988
  467. mindspore/train/summary/_summary_adapter.py +2 -2
  468. mindspore/train/summary/summary_record.py +1 -1
  469. mindspore/turbojpeg.dll +0 -0
  470. mindspore/utils/__init__.py +3 -2
  471. mindspore/utils/dryrun.py +4 -2
  472. mindspore/utils/hooks.py +81 -0
  473. mindspore/utils/utils.py +138 -4
  474. mindspore/vcmeta.dll +0 -0
  475. mindspore/vcruntime140.dll +0 -0
  476. mindspore/vcruntime140_1.dll +0 -0
  477. mindspore/version.py +1 -1
  478. {mindspore-2.5.0.dist-info → mindspore-2.6.0rc1.dist-info}/METADATA +2 -1
  479. {mindspore-2.5.0.dist-info → mindspore-2.6.0rc1.dist-info}/RECORD +483 -438
  480. mindspore/_install_custom.py +0 -43
  481. mindspore/common/_register_for_adapter.py +0 -74
  482. mindspore/ops/auto_generate/gen_arg_dtype_cast.py +0 -252
  483. mindspore/ops/auto_generate/gen_arg_handler.py +0 -136
  484. mindspore/ops/operations/_opaque_predicate_registry.py +0 -41
  485. mindspore/ops_generate/gen_constants.py +0 -190
  486. mindspore/ops_generate/gen_ops_inner_prim.py +0 -131
  487. mindspore/ops_generate/ops_primitive_h_generator.py +0 -81
  488. /mindspore/ops_generate/{base_generator.py → common/base_generator.py} +0 -0
  489. {mindspore-2.5.0.dist-info → mindspore-2.6.0rc1.dist-info}/WHEEL +0 -0
  490. {mindspore-2.5.0.dist-info → mindspore-2.6.0rc1.dist-info}/entry_points.txt +0 -0
  491. {mindspore-2.5.0.dist-info → mindspore-2.6.0rc1.dist-info}/top_level.txt +0 -0
@@ -41,10 +41,11 @@ class LRScheduler:
41
41
 
42
42
  Args:
43
43
  optimizer (:class:`mindspore.experimental.optim.Optimizer`): The optimizer instance.
44
- last_epoch (int, optional): The index of the last epoch. Default: ``-1``.
44
+ last_epoch (int, optional): The number of times the `step()` method of
45
+ the current learning rate adjustment strategy has been executed. Default: ``-1``.
45
46
 
46
47
  Raises:
47
- TypeError: If `optimizer` is not an Optimizer.
48
+ TypeError: If `optimizer` does not satisfy the type requirement.
48
49
  KeyError: If `last_epoch` != -1 and ``'initial_lr'`` not in param groups.
49
50
  ValueError: if `last_epoch` is not int.
50
51
  ValueError: If `last_epoch` is not greater than -1.
@@ -700,9 +701,8 @@ class ConstantLR(LRScheduler):
700
701
  @jit_class
701
702
  class SequentialLR:
702
703
  r"""
703
- Receives the list of schedulers that is expected to be called sequentially during
704
- optimization process and milestone points that provides exact intervals to reflect
705
- which scheduler is supposed to be called at a given epoch.
704
+ Concatenate multiple learning rate adjustment strategies in `schedulers` in sequence,
705
+ switching to the next learning rate adjustment strategy at `milestone`.
706
706
 
707
707
  .. warning::
708
708
  This is an experimental lr scheduler module that is subject to change.
@@ -713,8 +713,10 @@ class SequentialLR:
713
713
  optimizer (:class:`mindspore.experimental.optim.Optimizer`): Wrapped optimizer.
714
714
  schedulers (list[:class:`mindspore.experimental.optim.lr_scheduler.LRScheduler`]):
715
715
  List of learning rate schedulers.
716
- milestones (list): List of integers that reflects milestone points.
717
- last_epoch (int, optional): The index of the last epoch. Default: ``-1``.
716
+ milestones (list): List of integers of milestone points,
717
+ sets which learning rate adjustment strategy is invoked for each epoch.
718
+ last_epoch (int, optional): The number of times the `step()` method
719
+ of the current learning rate adjustment strategy has been executed. Default: ``-1``.
718
720
 
719
721
  Raises:
720
722
  ValueError: The optimizer in `schedulers` is different from the `optimizer` passed in.
@@ -805,9 +807,8 @@ class ReduceLROnPlateau:
805
807
  """
806
808
  Reduce learning rate when a metric has stopped improving.
807
809
  Models often benefit from reducing the learning rate by a factor
808
- of 2-10 once learning stagnates. This scheduler reads a metrics
809
- quantity and if no improvement is seen for a 'patience' number
810
- of epochs, the learning rate is reduced.
810
+ of 2-10 once learning stagnates. The scheduler reads the metrics `metrics` during execution
811
+ and adjusts the learning rate via the `step` method if the metrics do not improve within `patience` cycles.
811
812
 
812
813
  .. warning::
813
814
  This is an experimental lr scheduler module that is subject to change.
@@ -816,7 +817,8 @@ class ReduceLROnPlateau:
816
817
 
817
818
  Args:
818
819
  optimizer (:class:`mindspore.experimental.optim.Optimizer`): Wrapped optimizer.
819
- mode (str, optional): One of `min`, `max`. In `min` mode, lr will
820
+ mode (str, optional): Trigger mode that triggers a reduction in learning rate
821
+ when the monitoring metrics are at their `min` / `max` point. In `min` mode, lr will
820
822
  be reduced when the quantity monitored has stopped
821
823
  decreasing; in `max` mode it will be reduced when the
822
824
  quantity monitored has stopped increasing. Default: ``'min'``.
@@ -830,12 +832,25 @@ class ReduceLROnPlateau:
830
832
  Default: ``10``.
831
833
  threshold (float, optional): Threshold for measuring the new optimum,
832
834
  to only focus on significant changes. Default: ``1e-4``.
833
- threshold_mode (str, optional): One of `rel`, `abs`. Given dynamic_threshold is the benchmark to
834
- define whether the current metric is improvement,
835
- in ``'rel'`` mode, dynamic_threshold = best * ( 1 + threshold ) in ``'max'`` mode
836
- or best * ( 1 - threshold ) in ``'min'`` mode.
837
- In ``'abs'`` mode, dynamic_threshold = best + threshold in ``'max'`` mode or
838
- best - threshold in ``'min'`` mode. Default: ``'rel'``.
835
+ threshold_mode (str, optional): A mode for measuring indicators of change for the better.
836
+ One of `rel`, `abs`. Default: ``'rel'``.
837
+
838
+ Assume that `best` represents the best value of the current performance metric.
839
+
840
+ - In ``'rel'`` mode, the indicator is compared to a `threshold` in proportional form:
841
+
842
+ - When `mode` is ``'max'``, the indicator is considered better if it exceeds best * ( 1 + threshold ).
843
+
844
+ - When `mode` is ``'min'``, the indicator is considered better
845
+ if it is lower than best * ( 1 - threshold ).
846
+
847
+ - In ``'abs'`` mode, the indicator is compared to `threshold` in absolute value form:
848
+
849
+ - When `mode` is ``'max'``, the indicator is considered better if it exceeds best + threshold.
850
+
851
+ - When `mode` is ``'min'``, the indicator is considered better
852
+ if it is lower than best - threshold.
853
+
839
854
  cooldown (int, optional): Number of epochs to wait before resuming
840
855
  normal operation after lr has been reduced. Default: ``0``.
841
856
  min_lr (Union(float, list), optional): A scalar or a list of scalars. A
@@ -1164,15 +1179,17 @@ class CyclicLR(LRScheduler):
1164
1179
  class CosineAnnealingWarmRestarts(LRScheduler):
1165
1180
  r"""
1166
1181
  Set the learning rate of each parameter group using a cosine annealing warm restarts
1167
- schedule. Where :math:`\eta_{max}` is set to the initial lr, :math:`\eta_{min}` is the minimum value
1168
- for learning rate, :math:`\eta_{t}` is the current learning rate, :math:`T_{0}` is the number of iterations for the
1169
- first restar, :math:`T_{i}` is the current number of iterations between two warm restarts in SGDR,
1170
- :math:`T_{cur}` is the number of epochs since the last restart in SGDR.
1182
+ schedule.
1171
1183
 
1172
1184
  .. math::
1173
1185
  \eta_t = \eta_{min} + \frac{1}{2}(\eta_{max} - \eta_{min})\left(1 +
1174
1186
  \cos\left(\frac{T_{cur}}{T_{i}}\pi\right)\right)
1175
1187
 
1188
+ Where :math:`\eta_{max}` is set to the initial lr, :math:`\eta_{min}` is the minimum value
1189
+ for learning rate, :math:`\eta_{t}` is the current learning rate, :math:`T_{0}` is the number of iterations for the
1190
+ first restar, :math:`T_{i}` is the current number of iterations between two warm restarts in SGDR,
1191
+ :math:`T_{cur}` is the number of epochs since the last restart in SGDR.
1192
+
1176
1193
  When :math:`T_{cur}=T_{i}`, set :math:`\eta_t = \eta_{min}`.
1177
1194
  When :math:`T_{cur}=0` after restart, set :math:`\eta_t=\eta_{max}`.
1178
1195
 
@@ -1189,7 +1206,8 @@ class CosineAnnealingWarmRestarts(LRScheduler):
1189
1206
  T_0 (int): Number of iterations for the first restart.
1190
1207
  T_mult (int, optional): A factor increases :math:`T_{i}` after a restart. Default: ``1``.
1191
1208
  eta_min (Union(float, int), optional): Minimum learning rate. Default: ``0``.
1192
- last_epoch (int, optional): The index of the last epoch. Default: ``-1``.
1209
+ last_epoch (int, optional): The number of times the `step()` method of
1210
+ the current learning rate adjustment strategy has been executed. Default: ``-1``.
1193
1211
 
1194
1212
  Raises:
1195
1213
  ValueError: `T_0` is less than or equal than 0 or not an int.
@@ -56,7 +56,7 @@ class RAdam(Optimizer):
56
56
 
57
57
  .. math::
58
58
  \begin{align*}
59
- &\rule{110mm}{0.4pt} \\
59
+ &\rule{180mm}{0.4pt} \\
60
60
  &\textbf{Input}:
61
61
  \gamma \text{ (lr)}, \: \beta_1, \beta_2 \text{ (betas)}, \: \theta_0 \text{ (params)}, \:f(\theta)
62
62
  \text{ (objective)}, \:
@@ -67,7 +67,7 @@ class RAdam(Optimizer):
67
67
  v_0 \leftarrow 0 \text{ (second moment)} \\
68
68
  \rho_{\infty} \xleftarrow{\text{def}} \dfrac{2}{1 - \beta_2} - 1
69
69
  \end{cases} \\
70
- &\rule{110mm}{0.4pt} \\
70
+ &\rule{180mm}{0.4pt} \\
71
71
  &\textbf{For } t = 1 \text{ to } \ldots \text{ do}: \\
72
72
  &\quad g_t \leftarrow \nabla_{\theta} f_t(\theta_{t - 1}) \\
73
73
  &\quad \text{If } \lambda \neq 0: \\
@@ -84,9 +84,9 @@ class RAdam(Optimizer):
84
84
  &\quad\quad \theta_t \leftarrow \theta_{t - 1} - \gamma \widehat{m_t} r_t l_t \\
85
85
  &\quad \text{Else}: \\
86
86
  &\quad\quad \theta_t \leftarrow \theta_{t - 1} - \gamma \widehat{m_t} \\
87
- &\rule{110mm}{0.4pt} \\
87
+ &\rule{180mm}{0.4pt} \\
88
88
  &\bf{Return}: \theta_t \\
89
- &\rule{110mm}{0.4pt}
89
+ &\rule{180mm}{0.4pt}
90
90
  \end{align*}
91
91
 
92
92
  .. warning::
@@ -157,7 +157,7 @@ class RAdam(Optimizer):
157
157
  self.increase_tensor = Tensor(1, mstype.int32)
158
158
  self.assignadd = P.AssignAdd()
159
159
 
160
- @jit
160
+ @jit(backend="ms_backend")
161
161
  def implementation(self, lr, beta1, beta2, weight_decay, eps, start_id, end_id, gradients):
162
162
  """Extract the common computing part for acceleration"""
163
163
  params = self.parameters[start_id: end_id]
@@ -133,7 +133,7 @@ class Rprop(Optimizer):
133
133
  self.increase_tensor = Tensor(1, mstype.int32)
134
134
  self.op_cast = P.Cast()
135
135
 
136
- @jit
136
+ @jit(backend="ms_backend")
137
137
  def implementation(self, etaminus, etaplus, group_id, lr, gradients, maximize, step_size_min, step_size_max):
138
138
  """Extract the common computing part for acceleration"""
139
139
  etaminus, etaplus = op_cast(etaminus, mstype.float32), op_cast(etaplus, mstype.float32)
@@ -51,7 +51,7 @@ class SGD(Optimizer):
51
51
 
52
52
  To be noticed, for the first step, :math:`v_{t+1} = gradient`.
53
53
 
54
- Here : where p, v and u denote the parameters, accum, and momentum respectively.
54
+ Here: where :math:`p`, :math:`v` and :math:`u` denote the parameters, accum, and momentum respectively.
55
55
 
56
56
  .. warning::
57
57
  This is an experimental optimizer API that is subject to change.
@@ -19,7 +19,7 @@ from mindspore.common.tensor import Tensor
19
19
  from mindspore.common.api import _convert_python_data
20
20
  from mindspore.common.dtype import type_size_in_bytes
21
21
  from mindspore._c_expression import slice_by_tensor_index, slice_by_padding_shape, \
22
- combine_tensor_list_contiguous as combine_tensor_list, Tensor as Tensor_
22
+ combine_tensor_list_contiguous as combine_tensor_list, TensorPy as Tensor_
23
23
 
24
24
 
25
25
  def combine_tensor_list_contiguous(tensor_list, enable_mem_align=True):
@@ -29,7 +29,7 @@ def combine_tensor_list_contiguous(tensor_list, enable_mem_align=True):
29
29
  Args:
30
30
  tensor_list (list[Tensor], Tuple[Tensor]): The tensor list to be stored.
31
31
  enable_mem_align (bool, optional): Whether to enable the memory alignment function.
32
- False is not supported. Default: ``True`` .
32
+ False is not supported. Default ``True`` .
33
33
 
34
34
  Returns:
35
35
  ContiguousTensorsHandle, a manager with contiguous memory.
@@ -59,7 +59,7 @@ class ContiguousTensorsHandle:
59
59
  Args:
60
60
  tensor_list (list[Tensor], Tuple[Tensor]): The tensor list to be stored.
61
61
  enable_mem_align (bool, optional): Whether to enable the memory alignment function.
62
- False is not supported. Default: ``True`` .
62
+ False is not supported. Default ``True`` .
63
63
 
64
64
  Returns:
65
65
  ContiguousTensorsHandle, a manager with contiguous memory.
@@ -137,15 +137,11 @@ class ContiguousTensorsHandle:
137
137
  Return the tensor which is sliced by tensor index.
138
138
 
139
139
  Args:
140
- start(int, None): Starting position. Default:``None``.
141
- end(int, None): Deadline position. Default:``None``.
140
+ start(int, None): Starting position. Default ``None``.
141
+ end(int, None): Deadline position. Default ``None``.
142
142
 
143
143
  Returns:
144
- Tensor,is sliced by tensor index.
145
-
146
- Raises:
147
- TypeError: If `start` or `end`, is neither an 'int' nor a 'none'.
148
- ValueError: If values of `start` or `end` are negative, or out of the list range, or start >= end.
144
+ Tensor
149
145
 
150
146
  Examples:
151
147
  >>> import numpy as np
mindspore/hal/device.py CHANGED
@@ -120,23 +120,21 @@ def _check_device_id(fn):
120
120
 
121
121
  def is_initialized(device_target):
122
122
  """
123
- Returns whether specified backend is initialized.
123
+ Return whether specified device is initialized, this api will be deprecated and removed in future versions.
124
124
 
125
125
  Note:
126
- - The api will be deprecated.
126
+ MindSpore's devices "CPU", "GPU" and "Ascend" will be initialized in the following scenarios:
127
127
 
128
- Note:
129
- MindSpore's backends "CPU", "GPU" and "Ascend" will be initialized in the following scenarios:
130
-
131
- - For distributed job, backend will be initialized after `mindspore.communication.init` method is called.
132
- - For standalone job, backend will be initialized after running
128
+ - For distributed job, device will be initialized after `mindspore.communication.init` method is called.
129
+ - For standalone job, device will be initialized after running
133
130
  the first operator or calling creating stream/event interfaces.
134
131
 
135
132
  Args:
136
- device_target (str): The device name of backend, should be one of "CPU", "GPU" and "Ascend".
133
+ device_target (str): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
134
+ ``"Ascend"`` .
137
135
 
138
136
  Returns:
139
- Bool, whether the specified backend is initialized.
137
+ bool
140
138
 
141
139
  Examples:
142
140
  >>> import mindspore as ms
@@ -162,25 +160,21 @@ def is_initialized(device_target):
162
160
 
163
161
  def is_available(device_target):
164
162
  """
165
- Returns whether specified backend is available.
166
- All dependent libraries should be successfully loaded if this backend is available.
167
-
168
- Note:
169
- - The api will be deprecated.
170
- - CPU hardware, please use the interface :func:`mindspore.device_context.cpu.is_available`.
171
- - GPU hardware, please use the interface :func:`mindspore.device_context.gpu.is_available`.
172
- - Ascend hardware, please use the interface :func:`mindspore.device_context.ascend.is_available`.
163
+ Return whether specified device is available, this api will be deprecated and removed in future versions, please
164
+ use the api :func:`mindspore.device_context.cpu.is_available` , :func:`mindspore.device_context.gpu.is_available`,
165
+ :func:`mindspore.device_context.ascend.is_available` instead.
173
166
 
174
167
  Args:
175
- device_target (str): The device name of backend, should be one of "CPU", "GPU" and "Ascend".
168
+ device_target (str): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
169
+ ``"Ascend"`` . Default ``None`` , represents the current device set by context.
176
170
 
177
171
  Returns:
178
- Bool, whether the specified backend is available for this MindSpore package.
172
+ bool
179
173
 
180
174
  Examples:
181
- >>> import mindspore as ms
182
- >>> device_target = ms.context.get_context("device_target")
183
- >>> print(ms.hal.is_available(device_target))
175
+ >>> import mindspore
176
+ >>> device_target = mindspore.context.get_context("device_target")
177
+ >>> print(mindspore.hal.is_available(device_target))
184
178
  True
185
179
  """
186
180
  if device_target not in valid_targets:
@@ -201,29 +195,24 @@ def is_available(device_target):
201
195
  @_check_inputs_validation
202
196
  def device_count(device_target=None):
203
197
  """
204
- Returns device count of specified backend.
198
+ Return device count of specified device, this api will be deprecated and removed in future versions, please use
199
+ the api :func:`mindspore.device_context.cpu.device_count` , :func:`mindspore.device_context.gpu.device_count`,
200
+ :func:`mindspore.device_context.ascend.device_count` instead.
205
201
 
206
202
  Note:
207
- - The api will be deprecated.
208
- - CPU hardware, please use the interface :func:`mindspore.device_context.cpu.device_count`.
209
- - GPU hardware, please use the interface :func:`mindspore.device_context.gpu.device_count`.
210
- - Ascend hardware, please use the interface :func:`mindspore.device_context.ascend.device_count`.
211
-
212
-
213
- Note:
214
- If `device_target` is not specified, get the device count of the current backend set by context.
215
- For CPU backend, this method always returns 1.
203
+ For CPU device, this method always returns 1.
216
204
 
217
205
  Args:
218
- device_target (str, optional): The device name of backend, should be one of "CPU", "GPU" and "Ascend".
206
+ device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
207
+ ``"Ascend"`` . Default ``None`` , represents the current device set by context.
219
208
 
220
209
  Returns:
221
- int.
210
+ int
222
211
 
223
212
  Examples:
224
- >>> import mindspore as ms
225
- >>> device_target = ms.context.get_context("device_target")
226
- >>> print(ms.hal.device_count(device_target))
213
+ >>> import mindspore
214
+ >>> device_target = mindspore.context.get_context("device_target")
215
+ >>> print(mindspore.hal.device_count(device_target))
227
216
  """
228
217
  hal_instance = hal_instances.get(device_target)
229
218
  if hal_instance is None:
@@ -235,20 +224,15 @@ def device_count(device_target=None):
235
224
  @_check_inputs_validation
236
225
  def get_device_capability(device_id, device_target=None):
237
226
  """
238
- Get specified device's capability.
239
-
240
- Note:
241
- - The api will be deprecated.
242
-
243
- Note:
244
- If `device_target` is not specified, get the device capability of the current backend set by context.
227
+ Get specified device's capability, this api will be deprecated and removed in future versions.
245
228
 
246
229
  Args:
247
230
  device_id (int): The device id of which the capability will be returned.
248
- device_target (str, optional): The device name of backend, should be one of "CPU", "GPU" and "Ascend".
231
+ device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
232
+ ``"Ascend"`` . Default ``None`` , represents the current device set by context.
249
233
 
250
234
  Returns:
251
- tuple(int, int) for GPU.
235
+ tuple(param1, param2) for GPU.
252
236
 
253
237
  - param1 - int, cuda major revision number.
254
238
  - param2 - int, cuda minor revision number.
@@ -256,9 +240,9 @@ def get_device_capability(device_id, device_target=None):
256
240
  None for Ascend and CPU.
257
241
 
258
242
  Examples:
259
- >>> import mindspore as ms
260
- >>> device_target = ms.context.get_context("device_target")
261
- >>> print(ms.hal.get_device_capability(0, device_target))
243
+ >>> import mindspore
244
+ >>> device_target = mindspore.context.get_context("device_target")
245
+ >>> print(mindspore.hal.get_device_capability(0, device_target))
262
246
  """
263
247
  hal_instance = hal_instances.get(device_target)
264
248
  if hal_instance is None:
@@ -270,20 +254,17 @@ def get_device_capability(device_id, device_target=None):
270
254
  @_check_inputs_validation
271
255
  def get_device_properties(device_id, device_target=None):
272
256
  """
273
- Get specified device's properties.
274
-
275
- Note:
276
- - The api will be deprecated.
257
+ Get specified device's properties, this api will be deprecated and removed in future versions.
277
258
 
278
259
  Note:
279
- If `device_target` is not specified, get the device properties of the current backend set by context.
280
- For Ascend, backend must be initialized before calling this method,
260
+ For Ascend, device must be initialized before calling this method,
281
261
  or `total_memory` and `free_memory` will be 0,
282
262
  and `device_id` will be ignored since this method only returns current device's properties.
283
263
 
284
264
  Args:
285
265
  device_id (int): The device id of which the properties will be returned.
286
- device_target (str, optional): The device name of backend, should be one of "CPU", "GPU" and "Ascend".
266
+ device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
267
+ ``"Ascend"`` . Default ``None`` , represents the current device set by context.
287
268
 
288
269
  Returns:
289
270
  - `cudaDeviceProp` for GPU.
@@ -314,9 +295,9 @@ def get_device_properties(device_id, device_target=None):
314
295
  - None for CPU.
315
296
 
316
297
  Examples:
317
- >>> import mindspore as ms
318
- >>> device_target = ms.context.get_context("device_target")
319
- >>> print(ms.hal.get_device_properties(0, device_target))
298
+ >>> import mindspore
299
+ >>> device_target = mindspore.context.get_context("device_target")
300
+ >>> print(mindspore.hal.get_device_properties(0, device_target))
320
301
  """
321
302
  hal_instance = hal_instances.get(device_target)
322
303
  if hal_instance is None:
@@ -328,26 +309,23 @@ def get_device_properties(device_id, device_target=None):
328
309
  @_check_inputs_validation
329
310
  def get_device_name(device_id, device_target=None):
330
311
  """
331
- Get specified device's name.
332
-
333
- Note:
334
- - The api will be deprecated.
312
+ Get specified device's name, this api will be deprecated and removed in future versions.
335
313
 
336
314
  Note:
337
- If `device_target` is not specified, get the device name of the current backend set by context.
338
- This method always returns "CPU" for CPU backend.
315
+ This method always returns "CPU" for CPU device.
339
316
 
340
317
  Args:
341
318
  device_id (int): The device id of which the name will be returned.
342
- device_target (str, optional): The device name of backend, should be one of "CPU", "GPU" and "Ascend".
319
+ device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
320
+ ``"Ascend"`` . Default ``None`` , represents the current device set by context.
343
321
 
344
322
  Returns:
345
- str.
323
+ str
346
324
 
347
325
  Examples:
348
- >>> import mindspore as ms
349
- >>> device_target = ms.context.get_context("device_target")
350
- >>> print(ms.hal.get_device_name(0, device_target))
326
+ >>> import mindspore
327
+ >>> device_target = mindspore.context.get_context("device_target")
328
+ >>> print(mindspore.hal.get_device_name(0, device_target))
351
329
  """
352
330
  hal_instance = hal_instances.get(device_target)
353
331
  if hal_instance is None:
@@ -358,25 +336,21 @@ def get_device_name(device_id, device_target=None):
358
336
  @_check_inputs_validation
359
337
  def get_arch_list(device_target=None):
360
338
  """
361
- Get the architecture list this MindSpore was compiled for.
362
-
363
- Note:
364
- - The api will be deprecated.
365
-
366
- Note:
367
- If `device_target` is not specified, get the device name of the current backend set by context.
339
+ Get the architecture list this MindSpore was compiled for, this api will be deprecated and removed in future
340
+ versions.
368
341
 
369
342
  Args:
370
- device_target (str, optional): The device name of backend, should be one of "CPU", "GPU" and "Ascend".
343
+ device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
344
+ ``"Ascend"`` . Default ``None`` , represents the current device set by context.
371
345
 
372
346
  Returns:
373
347
  str for GPU.
374
348
  None for Ascend and CPU.
375
349
 
376
350
  Examples:
377
- >>> import mindspore as ms
378
- >>> device_target = ms.context.get_context("device_target")
379
- >>> print(ms.hal.get_arch_list(device_target))
351
+ >>> import mindspore
352
+ >>> device_target = mindspore.context.get_context("device_target")
353
+ >>> print(mindspore.hal.get_arch_list(device_target))
380
354
  """
381
355
  hal_instance = hal_instances.get(device_target)
382
356
  if hal_instance is None: