mindspore 2.3.0rc1__cp37-none-any.whl → 2.3.0rc2__cp37-none-any.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of mindspore might be problematic. Click here for more details.

Files changed (316) hide show
  1. mindspore/.commit_id +1 -1
  2. mindspore/__init__.py +1 -1
  3. mindspore/_akg/akg/utils/tbe_codegen_utils.py +13 -3
  4. mindspore/_c_dataengine.cpython-37m-aarch64-linux-gnu.so +0 -0
  5. mindspore/_c_expression.cpython-37m-aarch64-linux-gnu.so +0 -0
  6. mindspore/_checkparam.py +20 -0
  7. mindspore/_extends/parse/parser.py +1 -1
  8. mindspore/_extends/parse/standard_method.py +6 -5
  9. mindspore/_mindspore_offline_debug.cpython-37m-aarch64-linux-gnu.so +0 -0
  10. mindspore/amp.py +5 -5
  11. mindspore/bin/cache_admin +0 -0
  12. mindspore/bin/cache_server +0 -0
  13. mindspore/boost/boost_cell_wrapper.py +1 -1
  14. mindspore/boost/group_loss_scale_manager.py +1 -1
  15. mindspore/common/__init__.py +4 -2
  16. mindspore/common/_register_for_recompute.py +48 -0
  17. mindspore/common/_stub_tensor.py +1 -0
  18. mindspore/common/api.py +56 -4
  19. mindspore/common/dtype.py +5 -3
  20. mindspore/common/dump.py +2 -2
  21. mindspore/common/hook_handle.py +51 -4
  22. mindspore/common/initializer.py +1 -1
  23. mindspore/common/jit_config.py +17 -6
  24. mindspore/common/parameter.py +7 -2
  25. mindspore/common/recompute.py +247 -0
  26. mindspore/common/sparse_tensor.py +2 -2
  27. mindspore/common/symbol.py +1 -1
  28. mindspore/common/tensor.py +74 -36
  29. mindspore/communication/__init__.py +3 -3
  30. mindspore/communication/management.py +30 -30
  31. mindspore/context.py +28 -15
  32. mindspore/dataset/__init__.py +5 -5
  33. mindspore/dataset/audio/__init__.py +2 -2
  34. mindspore/dataset/audio/transforms.py +51 -51
  35. mindspore/dataset/callback/ds_callback.py +2 -2
  36. mindspore/dataset/engine/cache_client.py +1 -1
  37. mindspore/dataset/engine/datasets.py +3 -3
  38. mindspore/dataset/engine/datasets_audio.py +14 -14
  39. mindspore/dataset/engine/datasets_standard_format.py +3 -3
  40. mindspore/dataset/engine/datasets_text.py +38 -38
  41. mindspore/dataset/engine/datasets_user_defined.py +3 -3
  42. mindspore/dataset/engine/datasets_vision.py +68 -68
  43. mindspore/dataset/text/__init__.py +3 -3
  44. mindspore/dataset/text/transforms.py +26 -26
  45. mindspore/dataset/transforms/__init__.py +1 -1
  46. mindspore/dataset/vision/__init__.py +3 -3
  47. mindspore/dataset/vision/transforms.py +92 -92
  48. mindspore/dataset/vision/utils.py +1 -1
  49. mindspore/experimental/optim/adadelta.py +2 -2
  50. mindspore/experimental/optim/adagrad.py +2 -2
  51. mindspore/experimental/optim/adam.py +2 -2
  52. mindspore/experimental/optim/adamax.py +2 -2
  53. mindspore/experimental/optim/adamw.py +2 -2
  54. mindspore/experimental/optim/asgd.py +2 -2
  55. mindspore/experimental/optim/lr_scheduler.py +24 -20
  56. mindspore/experimental/optim/nadam.py +2 -2
  57. mindspore/experimental/optim/optimizer.py +1 -1
  58. mindspore/experimental/optim/radam.py +2 -2
  59. mindspore/experimental/optim/rmsprop.py +2 -2
  60. mindspore/experimental/optim/rprop.py +2 -2
  61. mindspore/experimental/optim/sgd.py +2 -2
  62. mindspore/hal/stream.py +2 -0
  63. mindspore/include/mindapi/base/types.h +5 -0
  64. mindspore/lib/libdnnl.so.2 +0 -0
  65. mindspore/lib/libmindspore.so +0 -0
  66. mindspore/lib/libmindspore_backend.so +0 -0
  67. mindspore/lib/libmindspore_common.so +0 -0
  68. mindspore/lib/libmindspore_core.so +0 -0
  69. mindspore/lib/libmindspore_gpr.so.15 +0 -0
  70. mindspore/lib/libmindspore_grpc++.so.1 +0 -0
  71. mindspore/lib/libmindspore_grpc.so.15 +0 -0
  72. mindspore/lib/libmindspore_shared_lib.so +0 -0
  73. mindspore/lib/libopencv_core.so.4.5 +0 -0
  74. mindspore/lib/plugin/ascend/custom_aicpu_ops/op_impl/cpu/aicpu_kernel/impl/libcust_cpu_kernels.so +0 -0
  75. mindspore/lib/plugin/ascend/custom_aicpu_ops/op_impl/cpu/config/cust_aicpu_kernel.json +6 -6
  76. mindspore/lib/plugin/ascend/custom_aicpu_ops/op_proto/libcust_op_proto.so +0 -0
  77. mindspore/lib/plugin/ascend/libdvpp_utils.so +0 -0
  78. mindspore/lib/plugin/ascend/liblowlatency_collective.so +0 -0
  79. mindspore/lib/plugin/ascend/libmindspore_cpu_kernels.so +0 -0
  80. mindspore/lib/plugin/ascend/ms_kernels_internal/asdops/bin/DeviceBin +0 -0
  81. mindspore/lib/plugin/ascend/ms_kernels_internal/asdops/bin/PkgInspect +0 -0
  82. mindspore/lib/plugin/ascend/ms_kernels_internal/asdops/bin/op_man +0 -0
  83. mindspore/lib/plugin/ascend/ms_kernels_internal/asdops/device/ascend910b/bin/ascend910b.bin +101787 -98559
  84. mindspore/lib/plugin/ascend/ms_kernels_internal/asdops/host/libasdops_cann_host.so +0 -0
  85. mindspore/lib/plugin/ascend/ms_kernels_internal/asdops/host/libasdops_host.so +0 -0
  86. mindspore/lib/plugin/ascend/ms_kernels_internal/asdops/include/asdops/base/op_register.h +2 -2
  87. mindspore/lib/plugin/ascend/ms_kernels_internal/asdops/include/asdops/params/mix.h +8 -1
  88. mindspore/lib/plugin/ascend/ms_kernels_internal/asdops/include/asdops/params/norm.h +5 -3
  89. mindspore/lib/plugin/ascend/ms_kernels_internal/asdops/include/asdops/params/reduce.h +2 -2
  90. mindspore/lib/plugin/ascend/ms_kernels_internal/asdops/include/asdops/utils/rt/backend/backend.h +3 -3
  91. mindspore/lib/plugin/ascend/ms_kernels_internal/asdops/include/asdops/utils/rt/backend/rtbackend.h +3 -3
  92. mindspore/lib/plugin/ascend/ms_kernels_internal/asdops/include/asdops/utils/rt/base/types.h +0 -1
  93. mindspore/lib/plugin/ascend/ms_kernels_internal/asdops/include/asdops/utils/rt/module/module.h +3 -3
  94. mindspore/lib/plugin/ascend/ms_kernels_internal/asdops/include/asdops/utils/svector/svector.h +3 -2
  95. mindspore/lib/plugin/ascend/ms_kernels_internal/asdops/lib/libasdops.so +0 -0
  96. mindspore/lib/plugin/ascend/ms_kernels_internal/asdops/lib/libasdops_static.a +0 -0
  97. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/add/tiling/add_tiling.h +9 -9
  98. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/apply_rotary_pos_emb/apply_rotary_pos_emb_impl.h +2 -6
  99. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/apply_rotary_pos_emb/kernel/apply_rotary_pos_emb.h +2 -2
  100. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/apply_rotary_pos_emb/kernel/apply_rotary_pos_emb_base.h +460 -0
  101. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/apply_rotary_pos_emb/kernel/apply_rotary_pos_emb_bf16.h +217 -0
  102. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/apply_rotary_pos_emb/kernel/apply_rotary_pos_emb_fp16.h +116 -0
  103. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/apply_rotary_pos_emb/kernel/apply_rotary_pos_emb_tiling.h +16 -24
  104. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/apply_rotary_pos_emb/kernel/apply_rotary_pos_emb_value.h +27 -0
  105. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/asdop/asd_op_impl.h +0 -4
  106. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/{attention/FlashAttentionScore_impl.h → flash_attention_score/flash_attention_score_impl.h} +2 -1
  107. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/{attention/bs_attention_tiling.h → flash_attention_score/flash_attention_score_tiling.h} +15 -19
  108. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/gelu/tiling/gelu_tiling.h +7 -9
  109. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/lccl/lccl_wrapper.h +58 -0
  110. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/matmul/matmul_impl.h +19 -8
  111. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/{matmul → matmul_common}/pp_matmul_common_tiling.h +18 -8
  112. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/{matmul → matmul_common}/pp_matmul_info.h +7 -4
  113. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/{matmul → matmul_common}/tiling_data.h +44 -6
  114. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/matmul_common/tiling_utils.h +65 -0
  115. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/matmul_stridedslice/matmul_stridedslice_fusion_impl.h +10 -6
  116. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/op_param.h +4 -1
  117. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/paged_attention/kernel/paged_attention_mix_hwsync.h +41 -0
  118. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/{attention/PagedAttention_impl.h → paged_attention/paged_attention_impl.h} +1 -1
  119. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/paged_attention/paged_attention_tiling.h +63 -0
  120. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/param/add_param.h +2 -2
  121. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/{attention_param.h → param/attention_param.h} +11 -2
  122. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/param/matmul_ext_param.h +37 -0
  123. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/param/sub_param.h +45 -0
  124. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/reshape_and_cache/reshape_and_cache_tiling.h +1 -2
  125. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/rms_norm/kernel/rms_norm.h +23 -0
  126. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/rms_norm/kernel/rms_norm_base.h +175 -0
  127. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/rms_norm/kernel/rms_norm_normal.h +276 -0
  128. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/rms_norm/kernel/rms_norm_split_d.h +280 -0
  129. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/rms_norm/kernel/tiling_data.h +35 -0
  130. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/rms_norm/rms_norm_impl.h +45 -0
  131. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/sub/kernel/sub_kernel.h +20 -0
  132. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/sub/sub_impl.h +47 -0
  133. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/sub/sub_tiling.h +25 -0
  134. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/tune_repo/matmul_table.h +323 -23
  135. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/types.h +15 -4
  136. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/utils/log/log_tiling.h +8 -0
  137. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libAdd_impl.so +0 -0
  138. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libSub_impl.so +0 -0
  139. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libadd_layernorm_impl.so +0 -0
  140. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libadd_rms_norm_impl.so +0 -0
  141. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libapply_rotary_pos_emb_impl.so +0 -0
  142. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libcast_impl.so +0 -0
  143. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libgelu_impl.so +0 -0
  144. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libmatmul_impl.so +0 -0
  145. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libmatmul_stridedslice_fusion_impl.so +0 -0
  146. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libms_kernels_internal.so +0 -0
  147. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libnot_equal_impl.so +0 -0
  148. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libreshape_and_cache_impl.so +0 -0
  149. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/librms_norm_impl.so +0 -0
  150. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/BSAttention/flash_attention_score_bf16_bnsd_full_mix.o +0 -0
  151. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/BSAttention/flash_attention_score_bf16_bnsd_tri_mix.o +0 -0
  152. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/BSAttention/flash_attention_score_bf16_bsh_full_mix.o +0 -0
  153. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/BSAttention/flash_attention_score_bf16_bsh_tri_mix.o +0 -0
  154. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/BSAttention/flash_attention_score_fp16_bnsd_full_mix.o +0 -0
  155. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/BSAttention/flash_attention_score_fp16_bnsd_tri_mix.o +0 -0
  156. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/BSAttention/flash_attention_score_fp16_bsh_full_mix.o +0 -0
  157. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/BSAttention/flash_attention_score_fp16_bsh_tri_mix.o +0 -0
  158. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/BSAttention/paged_attention_bf16_bnsd_full_mix.o +0 -0
  159. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/BSAttention/paged_attention_bf16_bsh_full_mix.o +0 -0
  160. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/BSAttention/paged_attention_fp16_bnsd_full_mix.o +0 -0
  161. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/BSAttention/paged_attention_fp16_bsh_full_mix.o +0 -0
  162. mindspore/lib/plugin/ascend/ms_kernels_internal/lccl/include/lcal.h +22 -0
  163. mindspore/lib/plugin/ascend/ms_kernels_internal/lccl/include/lcal_comm.h +70 -0
  164. mindspore/lib/plugin/ascend/ms_kernels_internal/lccl/include/lcal_types.h +103 -0
  165. mindspore/lib/plugin/ascend/ms_kernels_internal/lccl/include/lccl.h +47 -0
  166. mindspore/lib/plugin/ascend/ms_kernels_internal/lccl/include/lccl_wrapper.h +58 -0
  167. mindspore/lib/plugin/ascend/ms_kernels_internal/lccl/include/lcoc.h +154 -0
  168. mindspore/lib/plugin/ascend/ms_kernels_internal/lccl/lib/liblcal.so +0 -0
  169. mindspore/lib/plugin/ascend/ms_kernels_internal/lccl/lib/liblccl_wrapper.so +0 -0
  170. mindspore/lib/plugin/libmindspore_ascend.so.2 +0 -0
  171. mindspore/log.py +2 -2
  172. mindspore/mint/__init__.py +457 -0
  173. mindspore/mint/nn/__init__.py +430 -0
  174. mindspore/mint/nn/functional.py +424 -0
  175. mindspore/mint/optim/__init__.py +24 -0
  176. mindspore/mint/optim/adamw.py +186 -0
  177. mindspore/multiprocessing/__init__.py +4 -0
  178. mindspore/nn/__init__.py +3 -0
  179. mindspore/nn/cell.py +51 -47
  180. mindspore/nn/extend/__init__.py +29 -0
  181. mindspore/nn/extend/basic.py +140 -0
  182. mindspore/nn/extend/embedding.py +143 -0
  183. mindspore/nn/extend/layer/__init__.py +27 -0
  184. mindspore/nn/extend/layer/normalization.py +107 -0
  185. mindspore/nn/extend/pooling.py +117 -0
  186. mindspore/nn/generator.py +297 -0
  187. mindspore/nn/layer/basic.py +109 -1
  188. mindspore/nn/layer/container.py +2 -2
  189. mindspore/nn/layer/conv.py +6 -6
  190. mindspore/nn/layer/embedding.py +1 -1
  191. mindspore/nn/layer/normalization.py +21 -43
  192. mindspore/nn/layer/padding.py +4 -0
  193. mindspore/nn/optim/ada_grad.py +2 -2
  194. mindspore/nn/optim/adadelta.py +1 -1
  195. mindspore/nn/optim/adafactor.py +1 -1
  196. mindspore/nn/optim/adam.py +7 -7
  197. mindspore/nn/optim/adamax.py +2 -2
  198. mindspore/nn/optim/adasum.py +2 -2
  199. mindspore/nn/optim/asgd.py +2 -2
  200. mindspore/nn/optim/ftrl.py +1 -1
  201. mindspore/nn/optim/lamb.py +3 -3
  202. mindspore/nn/optim/lars.py +1 -1
  203. mindspore/nn/optim/lazyadam.py +2 -2
  204. mindspore/nn/optim/momentum.py +2 -2
  205. mindspore/nn/optim/optimizer.py +2 -2
  206. mindspore/nn/optim/proximal_ada_grad.py +2 -2
  207. mindspore/nn/optim/rmsprop.py +2 -2
  208. mindspore/nn/optim/rprop.py +2 -2
  209. mindspore/nn/optim/sgd.py +2 -2
  210. mindspore/nn/optim/thor.py +2 -2
  211. mindspore/nn/wrap/cell_wrapper.py +9 -9
  212. mindspore/nn/wrap/grad_reducer.py +5 -5
  213. mindspore/ops/_grad_experimental/grad_comm_ops.py +4 -2
  214. mindspore/ops/_vmap/vmap_grad_nn_ops.py +41 -2
  215. mindspore/ops/_vmap/vmap_math_ops.py +27 -8
  216. mindspore/ops/_vmap/vmap_nn_ops.py +66 -8
  217. mindspore/ops/auto_generate/cpp_create_prim_instance_helper.py +73 -1
  218. mindspore/ops/auto_generate/gen_arg_dtype_cast.py +12 -3
  219. mindspore/ops/auto_generate/gen_arg_handler.py +24 -0
  220. mindspore/ops/auto_generate/gen_extend_func.py +274 -0
  221. mindspore/ops/auto_generate/gen_ops_def.py +889 -22
  222. mindspore/ops/auto_generate/gen_ops_prim.py +3541 -253
  223. mindspore/ops/auto_generate/pyboost_inner_prim.py +282 -0
  224. mindspore/ops/composite/multitype_ops/_compile_utils.py +2 -1
  225. mindspore/ops/composite/multitype_ops/_constexpr_utils.py +9 -0
  226. mindspore/ops/extend/__init__.py +9 -1
  227. mindspore/ops/extend/array_func.py +134 -27
  228. mindspore/ops/extend/math_func.py +3 -3
  229. mindspore/ops/extend/nn_func.py +363 -2
  230. mindspore/ops/function/__init__.py +19 -2
  231. mindspore/ops/function/array_func.py +463 -439
  232. mindspore/ops/function/clip_func.py +7 -18
  233. mindspore/ops/function/grad/grad_func.py +5 -5
  234. mindspore/ops/function/linalg_func.py +4 -4
  235. mindspore/ops/function/math_func.py +260 -243
  236. mindspore/ops/function/nn_func.py +825 -62
  237. mindspore/ops/function/random_func.py +73 -4
  238. mindspore/ops/function/sparse_unary_func.py +1 -1
  239. mindspore/ops/function/vmap_func.py +1 -1
  240. mindspore/ops/functional.py +2 -2
  241. mindspore/ops/op_info_register.py +1 -31
  242. mindspore/ops/operations/__init__.py +2 -3
  243. mindspore/ops/operations/_grad_ops.py +2 -107
  244. mindspore/ops/operations/_inner_ops.py +5 -5
  245. mindspore/ops/operations/_sequence_ops.py +2 -2
  246. mindspore/ops/operations/array_ops.py +11 -233
  247. mindspore/ops/operations/comm_ops.py +32 -32
  248. mindspore/ops/operations/custom_ops.py +7 -89
  249. mindspore/ops/operations/manually_defined/ops_def.py +329 -4
  250. mindspore/ops/operations/math_ops.py +13 -163
  251. mindspore/ops/operations/nn_ops.py +9 -316
  252. mindspore/ops/operations/random_ops.py +1 -1
  253. mindspore/ops/operations/sparse_ops.py +3 -3
  254. mindspore/ops/primitive.py +2 -2
  255. mindspore/ops_generate/arg_dtype_cast.py +12 -3
  256. mindspore/ops_generate/arg_handler.py +24 -0
  257. mindspore/ops_generate/gen_ops_inner_prim.py +2 -0
  258. mindspore/ops_generate/gen_pyboost_func.py +13 -6
  259. mindspore/ops_generate/pyboost_utils.py +2 -17
  260. mindspore/parallel/__init__.py +3 -2
  261. mindspore/parallel/_auto_parallel_context.py +106 -1
  262. mindspore/parallel/_parallel_serialization.py +34 -2
  263. mindspore/parallel/_utils.py +16 -0
  264. mindspore/parallel/algo_parameter_config.py +4 -4
  265. mindspore/parallel/checkpoint_transform.py +249 -77
  266. mindspore/parallel/cluster/process_entity/_api.py +1 -1
  267. mindspore/parallel/parameter_broadcast.py +1 -1
  268. mindspore/parallel/shard.py +1 -1
  269. mindspore/profiler/parser/ascend_analysis/fwk_cann_parser.py +1 -0
  270. mindspore/profiler/parser/ascend_analysis/profiler_info_parser.py +17 -5
  271. mindspore/profiler/parser/ascend_msprof_exporter.py +3 -3
  272. mindspore/profiler/parser/ascend_msprof_generator.py +10 -3
  273. mindspore/profiler/parser/ascend_op_generator.py +26 -9
  274. mindspore/profiler/parser/ascend_timeline_generator.py +7 -4
  275. mindspore/profiler/parser/profiler_info.py +11 -1
  276. mindspore/profiler/profiling.py +13 -5
  277. mindspore/rewrite/api/node.py +12 -12
  278. mindspore/rewrite/api/symbol_tree.py +11 -11
  279. mindspore/run_check/_check_version.py +1 -1
  280. mindspore/safeguard/rewrite_obfuscation.py +2 -2
  281. mindspore/train/amp.py +4 -4
  282. mindspore/train/anf_ir_pb2.py +8 -2
  283. mindspore/train/callback/_backup_and_restore.py +2 -2
  284. mindspore/train/callback/_callback.py +4 -4
  285. mindspore/train/callback/_checkpoint.py +2 -2
  286. mindspore/train/callback/_early_stop.py +2 -2
  287. mindspore/train/callback/_landscape.py +4 -4
  288. mindspore/train/callback/_loss_monitor.py +2 -2
  289. mindspore/train/callback/_on_request_exit.py +2 -2
  290. mindspore/train/callback/_reduce_lr_on_plateau.py +2 -2
  291. mindspore/train/callback/_summary_collector.py +2 -2
  292. mindspore/train/callback/_time_monitor.py +2 -2
  293. mindspore/train/dataset_helper.py +8 -3
  294. mindspore/train/loss_scale_manager.py +2 -2
  295. mindspore/train/metrics/metric.py +3 -3
  296. mindspore/train/mind_ir_pb2.py +22 -17
  297. mindspore/train/model.py +15 -15
  298. mindspore/train/serialization.py +18 -18
  299. mindspore/train/summary/summary_record.py +7 -7
  300. mindspore/train/train_thor/convert_utils.py +3 -3
  301. mindspore/version.py +1 -1
  302. {mindspore-2.3.0rc1.dist-info → mindspore-2.3.0rc2.dist-info}/METADATA +1 -1
  303. {mindspore-2.3.0rc1.dist-info → mindspore-2.3.0rc2.dist-info}/RECORD +307 -260
  304. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/matmul_stridedslice/tiling_data.h +0 -59
  305. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/BSAttention/FlashAttentionScore_bf16_BNSD_mix.o +0 -0
  306. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/BSAttention/FlashAttentionScore_bf16_BSH_mix.o +0 -0
  307. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/BSAttention/FlashAttentionScore_fp16_BNSD_mix.o +0 -0
  308. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/BSAttention/FlashAttentionScore_fp16_BSH_mix.o +0 -0
  309. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/BSAttention/PagedAttention_bf16_BNSD_mix.o +0 -0
  310. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/BSAttention/PagedAttention_bf16_BSH_mix.o +0 -0
  311. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/BSAttention/PagedAttention_fp16_BNSD_mix.o +0 -0
  312. mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/BSAttention/PagedAttention_fp16_BSH_mix.o +0 -0
  313. /mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/{attention/bs_attention_mix_hwsync.h → flash_attention_score/kernel/flash_attention_score_mix_hwsync.h} +0 -0
  314. {mindspore-2.3.0rc1.dist-info → mindspore-2.3.0rc2.dist-info}/WHEEL +0 -0
  315. {mindspore-2.3.0rc1.dist-info → mindspore-2.3.0rc2.dist-info}/entry_points.txt +0 -0
  316. {mindspore-2.3.0rc1.dist-info → mindspore-2.3.0rc2.dist-info}/top_level.txt +0 -0
@@ -65,6 +65,19 @@ class _ParallelOptimizerConfig:
65
65
  OPTIMIZER_WEIGHT_SHARD_SIZE = "optimizer_weight_shard_size"
66
66
 
67
67
 
68
+ class _PipelineConfig:
69
+ """
70
+ The key of the Pipeline parallelism.
71
+ """
72
+ PIPELINE_INTERLEAVE = "pipeline_interleave"
73
+ PIPELINE_SCHEDULER = "pipeline_scheduler"
74
+
75
+
76
+ class _PipelineScheduler:
77
+ PIPELINE_1F1B = "1f1b"
78
+ PIPELINE_GPIPE = "gpipe"
79
+
80
+
68
81
  class _AutoParallelContext:
69
82
  """
70
83
  _AutoParallelContext is the environment in which operations are executed
@@ -248,6 +261,16 @@ class _AutoParallelContext:
248
261
  self.check_context_handle()
249
262
  return self._context_handle.get_pipeline_result_broadcast()
250
263
 
264
+ def get_pipeline_interleave(self):
265
+ """Get pipeline interleave flag"""
266
+ self.check_context_handle()
267
+ return self._context_handle.get_pipeline_interleave()
268
+
269
+ def get_pipeline_scheduler(self):
270
+ """Get pipeline scheduler"""
271
+ self.check_context_handle()
272
+ return self._context_handle.get_pipeline_scheduler()
273
+
251
274
  def set_pipeline_segments(self, segments):
252
275
  """Set the segments of the pipeline"""
253
276
  if isinstance(segments, bool) or not isinstance(segments, int):
@@ -796,16 +819,87 @@ class _AutoParallelContext:
796
819
  .format(type(enable_parallel_optimizer)))
797
820
  self._context_handle.set_enable_parallel_optimizer(enable_parallel_optimizer)
798
821
 
822
+ def set_force_fp32_communication(self, force_fp32_communication):
823
+ """
824
+ Set enable/disable force fp32 communication.
825
+
826
+ Args:
827
+ set_force_fp32_communication (bool): Enable/disable force fp32 communication.
828
+ """
829
+ self.check_context_handle()
830
+ if not isinstance(force_fp32_communication, bool):
831
+ raise TypeError("For 'set_auto_parallel_context', "
832
+ "the argument 'force_fp32_communication' must be bool, but got the type : {}."
833
+ .format(type(force_fp32_communication)))
834
+ self._context_handle.set_force_fp32_communication(force_fp32_communication)
835
+
799
836
  def get_enable_fold_pipeline(self):
800
837
  """Get parallel optimizer flag."""
801
838
  self.check_context_handle()
802
839
  return self._context_handle.get_enable_fold_pipeline()
803
840
 
841
+ def set_pipeline_config(self, pipeline_config):
842
+ r"""
843
+ Set the configuration for pipeline parallelism. The configuration provides more detailed behavior control about
844
+ parallel training when pipeline parallelism is enabled.
845
+
846
+ Args:
847
+ pipeline_config (dict): The configuration for pipeline parallelism. It supports following keys:
848
+
849
+ - pipeline_interleave(bool): Setting true enable interleave scheduler for pipeline parallelism. This
850
+ scheduler requires more memory but less bubble.
851
+ - pipeline_scheduler(string): There are two choices, "1f1b" and "gpipe". default is "1f1b"
852
+
853
+ - 1f1b: It requires less memory and bubble ratio, for it run backward pass when corresponding forward pass
854
+ finished.
855
+ - gpipe: It requires more memory and bubble ratio, for it run backward pass after all forward pass
856
+ finished.
857
+
858
+ Raises:
859
+ TypeError: If the type of `pipeline_config` is not `dict`.
860
+ ValueError: If the key in `pipeline_config` not in ["pipeline_interleave", "pipeline_scheduler"].
861
+ ValueError: If pipeline interleave is False, pipeline scheduler is not `1f1b`.
862
+ """
863
+ self.check_context_handle()
864
+
865
+ if not isinstance(pipeline_config, dict):
866
+ raise TypeError("For 'set_pipeline_config', the argument 'pipeine_config' "
867
+ "must be dict, but got the type : {}.".format(type(pipeline_config)))
868
+
869
+ pp_interleave = _PipelineConfig.PIPELINE_INTERLEAVE
870
+ pp_scheduler = _PipelineConfig.PIPELINE_SCHEDULER
871
+
872
+ for config_name in pipeline_config:
873
+ unknown_config = []
874
+ if config_name not in [pp_interleave, pp_scheduler]:
875
+ unknown_config.append(config_name)
876
+
877
+ if unknown_config:
878
+ raise ValueError("Unknown config: {}".format(unknown_config))
879
+
880
+ Validator.check_bool(
881
+ pipeline_config[pp_interleave], pp_interleave, pp_interleave)
882
+ self._context_handle.set_pipeline_interleave(
883
+ pipeline_config[pp_interleave])
884
+
885
+ Validator.check_string(pipeline_config[pp_scheduler], [_PipelineScheduler.PIPELINE_1F1B,
886
+ _PipelineScheduler.PIPELINE_GPIPE])
887
+ if not pipeline_config[pp_interleave] and pipeline_config[pp_scheduler] != _PipelineScheduler.PIPELINE_1F1B:
888
+ raise ValueError(f"When pipeline_interleave is False, {pp_scheduler} is not supported")
889
+
890
+ self._context_handle.set_pipeline_scheduler(pipeline_config[pp_scheduler])
891
+
804
892
  def get_enable_parallel_optimizer(self):
805
893
  """Get parallel optimizer flag."""
806
894
  self.check_context_handle()
807
895
  return self._context_handle.get_enable_parallel_optimizer()
808
896
 
897
+ def get_force_fp32_communication(self):
898
+ """Get force fp32 communication flag."""
899
+ self.check_context_handle()
900
+ return self._context_handle.get_force_fp32_communication()
901
+
902
+
809
903
  def set_parallel_optimizer_config(self, parallel_optimizer_config):
810
904
  r"""
811
905
  Set the configure for parallel optimizer. The configure provides more detailed behavior control about parallel
@@ -1087,6 +1181,7 @@ class _AutoParallelContext:
1087
1181
  self.set_enable_all_gather_fusion(openstate)
1088
1182
  self.set_enable_reduce_scatter_fusion(openstate)
1089
1183
 
1184
+
1090
1185
  def _set_ops_strategy_json_config(type="SAVE", path="", mode="all"):
1091
1186
  """
1092
1187
  Set strategy json configuration.
@@ -1110,6 +1205,7 @@ def _set_ops_strategy_json_config(type="SAVE", path="", mode="all"):
1110
1205
  else:
1111
1206
  raise KeyError("Type must be 'SAVE' or 'LOAD' and mode must be 'all' or 'principal'")
1112
1207
 
1208
+
1113
1209
  _AUTO_PARALLEL_CONTEXT = None
1114
1210
 
1115
1211
 
@@ -1145,7 +1241,9 @@ _set_auto_parallel_context_func_map = {
1145
1241
  "full_batch": auto_parallel_context().set_full_batch,
1146
1242
  "dataset_strategy": auto_parallel_context().set_dataset_strategy,
1147
1243
  "enable_parallel_optimizer": auto_parallel_context().set_enable_parallel_optimizer,
1244
+ "force_fp32_communication": auto_parallel_context().set_force_fp32_communication,
1148
1245
  "parallel_optimizer_config": auto_parallel_context().set_parallel_optimizer_config,
1246
+ "pipeline_config": auto_parallel_context().set_pipeline_config,
1149
1247
  "grad_accumulation_step": auto_parallel_context().set_grad_accumulation_step,
1150
1248
  "all_reduce_fusion_config": auto_parallel_context().set_all_reduce_fusion_split_indices,
1151
1249
  "communi_parallel_mode": auto_parallel_context().set_communi_parallel_mode,
@@ -1164,6 +1262,8 @@ _get_auto_parallel_context_func_map = {
1164
1262
  "loss_repeated_mean": auto_parallel_context().get_loss_repeated_mean,
1165
1263
  "pipeline_stages": auto_parallel_context().get_pipeline_stages,
1166
1264
  "pipeline_result_broadcast": auto_parallel_context().get_pipeline_result_broadcast,
1265
+ "pipeline_interleave": auto_parallel_context().get_pipeline_interleave,
1266
+ "pipeline_scheduler": auto_parallel_context().get_pipeline_scheduler,
1167
1267
  "parallel_mode": auto_parallel_context().get_parallel_mode,
1168
1268
  "search_mode": auto_parallel_context().get_strategy_search_mode,
1169
1269
  "auto_parallel_search_mode": auto_parallel_context().get_auto_parallel_search_mode,
@@ -1173,6 +1273,7 @@ _get_auto_parallel_context_func_map = {
1173
1273
  "full_batch": auto_parallel_context().get_full_batch,
1174
1274
  "dataset_strategy": auto_parallel_context().get_dataset_strategy,
1175
1275
  "enable_parallel_optimizer": auto_parallel_context().get_enable_parallel_optimizer,
1276
+ "force_fp32_communication": auto_parallel_context().get_force_fp32_communication,
1176
1277
  "grad_accumulation_step": auto_parallel_context().get_grad_accumulation_step,
1177
1278
  "all_reduce_fusion_config": auto_parallel_context().get_all_reduce_fusion_split_indices,
1178
1279
  "communi_parallel_mode": auto_parallel_context().get_communi_parallel_mode,
@@ -1192,7 +1293,7 @@ _get_auto_parallel_context_func_map = {
1192
1293
  grad_accumulation_step=int, all_reduce_fusion_config=list, group_ckpt_save_file=str,
1193
1294
  communi_parallel_mode=str, optimizer_weight_shard_size=int, sharding_propagation=bool,
1194
1295
  optimizer_weight_shard_aggregated_save=bool, enable_alltoall=bool, comm_fusion=dict,
1195
- strategy_ckpt_config=dict)
1296
+ strategy_ckpt_config=dict, force_fp32_communication=bool)
1196
1297
  def _set_auto_parallel_context(**kwargs):
1197
1298
  """
1198
1299
  Set auto parallel context.
@@ -1240,6 +1341,9 @@ def _set_auto_parallel_context(**kwargs):
1240
1341
  full_batch (bool): Whether to load the whole batch on each device. Default: ``False``.
1241
1342
  dataset_strategy Union[str, tuple]: Dataset sharding strategy. Default: "data_parallel".
1242
1343
  enable_parallel_optimizer (bool): Enable using optimizer segmentation or not. Default: ``False``.
1344
+ force_fp32_communication (bool): A switch that determines whether reduce operators (AllReduce, ReduceScatter)
1345
+ are forced to use the fp32 data type for communication during communication. True is the enable
1346
+ switch. Default: ``False`` .
1243
1347
  all_reduce_fusion_config (list): Set allreduce fusion strategy by parameters indices.
1244
1348
  pipeline_stages (int): Set the stage information for pipeline parallel. This indicates how
1245
1349
  the devices are distributed alone the pipeline. The total devices will be divided into
@@ -1330,6 +1434,7 @@ def _reset_auto_parallel_context():
1330
1434
  - strategy_ckpt_load_file: ""
1331
1435
  - strategy_ckpt_save_file: ""
1332
1436
  - enable_parallel_optimizer: False
1437
+ - force_fp32_communication: False
1333
1438
  - search_mode: 'recursive_programming
1334
1439
  - auto_parallel_search_mode: 'recursive_programming
1335
1440
  - sharding_propagation: False
@@ -259,6 +259,33 @@ def _extract_pipeline_stage_num(strategy_file):
259
259
  return pipeline_stage_num
260
260
 
261
261
 
262
+ def _extract_src_dst_layout_map_by_src(src_strategy_file=None, dst_strategy_file=None):
263
+ """Extract strategy list by src strategy"""
264
+ src_layout_map = _extract_layout_map(src_strategy_file)
265
+ dst_layout_map = _extract_layout_map(dst_strategy_file)
266
+ if dst_layout_map is None:
267
+ return src_layout_map, dst_layout_map
268
+ for param_name in list(dst_layout_map.keys()):
269
+ if param_name in src_layout_map.keys():
270
+ continue
271
+ dst_layout_map.pop(param_name)
272
+ stage_id = 0
273
+ if src_strategy_file[-5:] == ".json":
274
+ with open(src_strategy_file, 'r') as f:
275
+ json_content = json.load(f)
276
+ strategy_items = json_content.get("parallel_strategy_item")
277
+ if not strategy_items:
278
+ raise ValueError("The strategy file {} if empty.".format(src_strategy_file))
279
+ stage_id = strategy_items.get(list(strategy_items.keys())[0]).get('stage')
280
+ else:
281
+ src_parallel_strategy_map = _load_protobuf_strategy(src_strategy_file)
282
+ strategy_items = src_parallel_strategy_map.parallel_strategy_item
283
+ if not strategy_items:
284
+ raise ValueError("The strategy file {} if empty.".format(src_strategy_file))
285
+ stage_id = strategy_items[0].parallel_strategys.stage
286
+ return src_layout_map, dst_layout_map, stage_id
287
+
288
+
262
289
  def _extract_src_dst_layout_map(rank_id, src_strategy_file=None, dst_strategy_file=None):
263
290
  """Extract strategy list"""
264
291
  src_layout_map = _extract_layout_map(src_strategy_file, None)
@@ -357,6 +384,7 @@ def _transform_parallel_checkpoint(rank_id, param_total_dict, param_attr_dict, s
357
384
  Transform model parallel dimension for distributed checkpoint files.
358
385
  """
359
386
  transform_param_dict = {}
387
+ device_num = -1
360
388
  for param_name, _ in param_total_dict.items():
361
389
  tensor_shape = list(param_total_dict[param_name].values())[0].shape
362
390
  from_dev_matrix = [1]
@@ -410,14 +438,18 @@ def _transform_parallel_checkpoint(rank_id, param_total_dict, param_attr_dict, s
410
438
  to_info_tuple = (to_opt_shard_size, to_dev_matrix_origin, to_tensor_map_origin, origin_tensor_shape)
411
439
  _insert_opt_shard_reshape(param_rank_map, from_info_tuple, to_info_tuple)
412
440
  transform_operator_stack = _generate_transform_operator_stack(param_rank_map, rank_id)
413
- _apply_tensor_transform_operators(transform_operator_stack, param_total_dict[param_name], device_num)
414
- transform_tensor = ms.Tensor(param_total_dict[param_name][rank_id % device_num])
441
+ param_total_dict_copy = param_total_dict[param_name].copy()
442
+ _apply_tensor_transform_operators(transform_operator_stack, param_total_dict_copy, device_num)
443
+ transform_tensor = ms.Tensor(param_total_dict_copy[rank_id % device_num])
415
444
  requires_grad = param_attr_dict[param_name][rank_id % device_num][0]
416
445
  layerwise_parallel = param_attr_dict[param_name][rank_id % device_num][1]
417
446
  transform_para = ms.Parameter(transform_tensor, param_name, requires_grad, layerwise_parallel)
418
447
  if param_type_dict[param_name][rank_id % device_num] == "BFloat16":
419
448
  transform_para.set_dtype(ms.bfloat16)
420
449
  transform_param_dict[param_name] = transform_para
450
+ if device_num < 1:
451
+ raise ValueError("None of the parameters in checkpoint file are in either src strategy or "
452
+ "dst strategy. Please check correctness of strategy files.")
421
453
 
422
454
  # Handle those parameter like learning_rate, global_step which not in strategy_file.
423
455
  for param_name, _ in param_total_dict.items():
@@ -191,6 +191,22 @@ def _origin_shapes(shapes):
191
191
  return new_shapes
192
192
 
193
193
 
194
+ def _dynamic_shape_for_dataset(dataset_shapes, dynamic_shapes):
195
+ """convert static dataset shapes to dynamic shape"""
196
+ if len(dataset_shapes) != len(dynamic_shapes):
197
+ raise ValueError("The dataset shapes size of {} is not equal to "
198
+ "dynamic shapes size of {}".format(dataset_shapes, dynamic_shapes))
199
+ ret = dataset_shapes
200
+ for i in range(len(dynamic_shapes)):
201
+ if len(dataset_shapes[i]) != len(dynamic_shapes[i]):
202
+ raise ValueError("The dataset shapes size of {} is not equal to "
203
+ "dynamic shapes size of {}".format(dataset_shapes, dynamic_shapes))
204
+ for j in range(len(dynamic_shapes[i])):
205
+ if dynamic_shapes[i][j] == -1:
206
+ ret[i][j] = -1
207
+ return ret
208
+
209
+
194
210
  def _to_full_tensor(elem, global_device_num, global_rank, scaling_sens=None):
195
211
  """Convert numpy to tensor, expanding batch dimension according to device_num, adapt to feed the data
196
212
  from host solution.
@@ -229,7 +229,7 @@ def set_algo_parameters(**kwargs):
229
229
  """
230
230
  Set parameters in the algorithm for parallel strategy searching. See a typical use in
231
231
  `test_auto_parallel_resnet.py
232
- <https://gitee.com/mindspore/mindspore/blob/r2.3.q1/tests/ut/python/parallel/test_auto_parallel_resnet.py>`_.
232
+ <https://gitee.com/mindspore/mindspore/blob/master/tests/ut/python/parallel/test_auto_parallel_resnet.py>`_.
233
233
 
234
234
  Note:
235
235
  The attribute name is required. This interface works ONLY in AUTO_PARALLEL mode.
@@ -266,14 +266,14 @@ def set_algo_parameters(**kwargs):
266
266
 
267
267
  For the Ascend devices, users need to prepare the rank table, set rank_id and device_id.
268
268
  Please see the `rank table startup
269
- <https://www.mindspore.cn/tutorials/experts/en/r2.3.q1/parallel/rank_table.html>`_
269
+ <https://www.mindspore.cn/tutorials/experts/en/master/parallel/rank_table.html>`_
270
270
  for more details.
271
271
 
272
272
  For the GPU devices, users need to prepare the host file and mpi, please see the `mpirun startup
273
- <https://www.mindspore.cn/tutorials/experts/en/r2.3.q1/parallel/mpirun.html>`_ .
273
+ <https://www.mindspore.cn/tutorials/experts/en/master/parallel/mpirun.html>`_ .
274
274
 
275
275
  For the CPU device, users need to write a dynamic cluster startup script, please see the `Dynamic Cluster
276
- Startup <https://www.mindspore.cn/tutorials/experts/en/r2.3.q1/parallel/dynamic_cluster.html>`_ .
276
+ Startup <https://www.mindspore.cn/tutorials/experts/en/master/parallel/dynamic_cluster.html>`_ .
277
277
 
278
278
  >>> import numpy as np
279
279
  >>> import mindspore as ms