mindspore 2.5.0__cp310-cp310-win_amd64.whl → 2.6.0__cp310-cp310-win_amd64.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of mindspore might be problematic. Click here for more details.

Files changed (493) hide show
  1. mindspore/.commit_id +1 -1
  2. mindspore/Microsoft.VisualStudio.Telemetry.dll +0 -0
  3. mindspore/Newtonsoft.Json.dll +0 -0
  4. mindspore/__init__.py +6 -4
  5. mindspore/_c_dataengine.cp310-win_amd64.pyd +0 -0
  6. mindspore/_c_expression.cp310-win_amd64.pyd +0 -0
  7. mindspore/_c_mindrecord.cp310-win_amd64.pyd +0 -0
  8. mindspore/_check_jit_forbidden_api.py +3 -0
  9. mindspore/_checkparam.py +3 -33
  10. mindspore/_deprecated/__init__.py +17 -0
  11. mindspore/_deprecated/jit.py +198 -0
  12. mindspore/_extends/builtin_operations.py +1 -1
  13. mindspore/_extends/parse/__init__.py +6 -7
  14. mindspore/_extends/parse/compile_config.py +19 -0
  15. mindspore/_extends/parse/deprecated/deprecated_tensor_method.py +22 -3
  16. mindspore/_extends/parse/jit_fallback_modules/__init__.py +0 -0
  17. mindspore/_extends/parse/jit_fallback_modules/check_utils.py +123 -0
  18. mindspore/_extends/parse/jit_fallback_modules/third_party_modules.py +50 -0
  19. mindspore/_extends/parse/parser.py +25 -194
  20. mindspore/_extends/parse/resources.py +1 -5
  21. mindspore/_extends/parse/standard_method.py +109 -75
  22. mindspore/_extends/pijit/__init__.py +2 -2
  23. mindspore/_extends/pijit/pijit_func_white_list.py +16 -11
  24. mindspore/_extends/pijit/tensor_func_list.py +27 -0
  25. mindspore/_extends/utils.py +1 -1
  26. mindspore/amp.py +4 -4
  27. mindspore/atlprov.dll +0 -0
  28. mindspore/avcodec-59.dll +0 -0
  29. mindspore/avdevice-59.dll +0 -0
  30. mindspore/avfilter-8.dll +0 -0
  31. mindspore/avformat-59.dll +0 -0
  32. mindspore/avutil-57.dll +0 -0
  33. mindspore/boost/__init__.py +2 -2
  34. mindspore/boost/base.py +3 -7
  35. mindspore/boost/boost_cell_wrapper.py +2 -2
  36. mindspore/c1.dll +0 -0
  37. mindspore/c1xx.dll +0 -0
  38. mindspore/c2.dll +0 -0
  39. mindspore/common/__init__.py +4 -3
  40. mindspore/common/_grad_function.py +56 -0
  41. mindspore/common/_pijit_context.py +14 -5
  42. mindspore/common/_register_for_tensor.py +1 -1
  43. mindspore/common/_stub_tensor.py +5 -10
  44. mindspore/common/_tensor_cpp_method.py +1 -1
  45. mindspore/common/_tensor_docs.py +2014 -3386
  46. mindspore/common/api.py +386 -355
  47. mindspore/common/auto_dynamic_shape.py +41 -44
  48. mindspore/common/dtype.py +5 -2
  49. mindspore/common/dump.py +7 -5
  50. mindspore/common/file_system.py +3 -0
  51. mindspore/common/generator.py +3 -0
  52. mindspore/common/hook_handle.py +5 -3
  53. mindspore/common/initializer.py +10 -6
  54. mindspore/common/jit_begin_end.py +94 -0
  55. mindspore/common/jit_config.py +6 -1
  56. mindspore/common/jit_context.py +76 -0
  57. mindspore/common/jit_trace.py +378 -0
  58. mindspore/common/lazy_inline.py +2 -2
  59. mindspore/common/mutable.py +5 -4
  60. mindspore/common/parameter.py +106 -39
  61. mindspore/common/seed.py +2 -2
  62. mindspore/common/sparse_tensor.py +23 -17
  63. mindspore/common/tensor.py +332 -714
  64. mindspore/communication/__init__.py +7 -5
  65. mindspore/communication/_comm_helper.py +47 -2
  66. mindspore/communication/comm_func.py +70 -53
  67. mindspore/communication/management.py +83 -17
  68. mindspore/context.py +228 -571
  69. mindspore/dataset/__init__.py +44 -20
  70. mindspore/dataset/audio/__init__.py +2 -8
  71. mindspore/dataset/audio/transforms.py +3 -17
  72. mindspore/dataset/core/config.py +3 -3
  73. mindspore/dataset/engine/cache_client.py +1 -1
  74. mindspore/dataset/engine/datasets.py +102 -120
  75. mindspore/dataset/engine/datasets_audio.py +22 -22
  76. mindspore/dataset/engine/datasets_standard_format.py +43 -24
  77. mindspore/dataset/engine/datasets_text.py +78 -85
  78. mindspore/dataset/engine/datasets_user_defined.py +109 -77
  79. mindspore/dataset/engine/datasets_vision.py +111 -108
  80. mindspore/dataset/engine/iterators.py +5 -3
  81. mindspore/dataset/engine/obs/obs_mindrecord_dataset.py +1 -1
  82. mindspore/dataset/engine/samplers.py +279 -57
  83. mindspore/dataset/engine/serializer_deserializer.py +2 -1
  84. mindspore/dataset/engine/validators.py +10 -0
  85. mindspore/dataset/text/__init__.py +7 -6
  86. mindspore/dataset/text/transforms.py +6 -5
  87. mindspore/dataset/text/utils.py +3 -3
  88. mindspore/dataset/transforms/__init__.py +0 -9
  89. mindspore/dataset/transforms/transforms.py +3 -3
  90. mindspore/dataset/utils/browse_dataset.py +1 -1
  91. mindspore/dataset/vision/__init__.py +2 -9
  92. mindspore/dataset/vision/transforms.py +202 -158
  93. mindspore/dataset/vision/utils.py +7 -5
  94. mindspore/device_context/ascend/op_debug.py +60 -1
  95. mindspore/device_context/ascend/op_tuning.py +0 -4
  96. mindspore/device_manager.py +39 -3
  97. mindspore/dnnl.dll +0 -0
  98. mindspore/dpcmi.dll +0 -0
  99. mindspore/experimental/es/embedding_service.py +35 -27
  100. mindspore/experimental/llm_boost/ascend_native/llama_boost_ascend_native.py +0 -2
  101. mindspore/experimental/map_parameter.py +4 -4
  102. mindspore/experimental/optim/adadelta.py +22 -26
  103. mindspore/experimental/optim/adagrad.py +4 -4
  104. mindspore/experimental/optim/adam.py +4 -0
  105. mindspore/experimental/optim/adamax.py +4 -4
  106. mindspore/experimental/optim/adamw.py +4 -0
  107. mindspore/experimental/optim/asgd.py +1 -1
  108. mindspore/experimental/optim/lr_scheduler.py +40 -22
  109. mindspore/experimental/optim/radam.py +5 -5
  110. mindspore/experimental/optim/rprop.py +1 -1
  111. mindspore/experimental/optim/sgd.py +1 -1
  112. mindspore/hal/contiguous_tensors_handle.py +6 -10
  113. mindspore/hal/device.py +55 -81
  114. mindspore/hal/event.py +38 -55
  115. mindspore/hal/memory.py +115 -147
  116. mindspore/hal/stream.py +81 -125
  117. mindspore/include/dataset/constants.h +7 -4
  118. mindspore/include/dataset/execute.h +2 -2
  119. mindspore/jpeg62.dll +0 -0
  120. mindspore/log.py +40 -2
  121. mindspore/mindrecord/__init__.py +20 -7
  122. mindspore/mindspore_backend_common.dll +0 -0
  123. mindspore/mindspore_backend_manager.dll +0 -0
  124. mindspore/mindspore_common.dll +0 -0
  125. mindspore/mindspore_core.dll +0 -0
  126. mindspore/mindspore_dump.dll +0 -0
  127. mindspore/mindspore_frontend.dll +0 -0
  128. mindspore/mindspore_glog.dll +0 -0
  129. mindspore/mindspore_memory_pool.dll +0 -0
  130. mindspore/mindspore_ms_backend.dll +0 -0
  131. mindspore/mindspore_ops.dll +0 -0
  132. mindspore/{mindspore_backend.dll → mindspore_ops_host.dll} +0 -0
  133. mindspore/mindspore_ops_kernel_common.dll +0 -0
  134. mindspore/mindspore_profiler.dll +0 -0
  135. mindspore/mindspore_pyboost.dll +0 -0
  136. mindspore/mindspore_pynative.dll +0 -0
  137. mindspore/mindspore_res_manager.dll +0 -0
  138. mindspore/mindspore_runtime_pipeline.dll +0 -0
  139. mindspore/mint/__init__.py +133 -702
  140. mindspore/mint/distributed/__init__.py +5 -1
  141. mindspore/mint/distributed/distributed.py +198 -113
  142. mindspore/mint/linalg/__init__.py +2 -0
  143. mindspore/mint/nn/__init__.py +280 -18
  144. mindspore/mint/nn/functional.py +282 -64
  145. mindspore/mint/nn/layer/__init__.py +4 -0
  146. mindspore/mint/nn/layer/_functions.py +7 -3
  147. mindspore/mint/nn/layer/activation.py +120 -13
  148. mindspore/mint/nn/layer/conv.py +234 -28
  149. mindspore/mint/nn/layer/normalization.py +15 -16
  150. mindspore/mint/nn/layer/padding.py +1 -1
  151. mindspore/mint/nn/layer/pooling.py +66 -1
  152. mindspore/mint/optim/__init__.py +2 -1
  153. mindspore/mint/optim/sgd.py +171 -0
  154. mindspore/msobj140.dll +0 -0
  155. mindspore/mspdb140.dll +0 -0
  156. mindspore/mspdbcore.dll +0 -0
  157. mindspore/mspdbst.dll +0 -0
  158. mindspore/mspft140.dll +0 -0
  159. mindspore/msvcdis140.dll +0 -0
  160. mindspore/msvcp140_1.dll +0 -0
  161. mindspore/msvcp140_2.dll +0 -0
  162. mindspore/msvcp140_atomic_wait.dll +0 -0
  163. mindspore/msvcp140_codecvt_ids.dll +0 -0
  164. mindspore/nn/__init__.py +4 -1
  165. mindspore/nn/cell.py +1253 -179
  166. mindspore/nn/layer/activation.py +23 -21
  167. mindspore/nn/layer/basic.py +22 -16
  168. mindspore/nn/layer/container.py +1 -1
  169. mindspore/nn/layer/conv.py +53 -42
  170. mindspore/nn/layer/embedding.py +9 -8
  171. mindspore/nn/layer/normalization.py +48 -42
  172. mindspore/nn/layer/pooling.py +75 -31
  173. mindspore/nn/layer/transformer.py +11 -10
  174. mindspore/nn/learning_rate_schedule.py +4 -2
  175. mindspore/nn/loss/loss.py +27 -19
  176. mindspore/nn/optim/ada_grad.py +6 -5
  177. mindspore/nn/optim/adadelta.py +9 -7
  178. mindspore/nn/optim/adafactor.py +1 -1
  179. mindspore/nn/optim/adam.py +18 -14
  180. mindspore/nn/optim/adamax.py +8 -7
  181. mindspore/nn/optim/adasum.py +5 -5
  182. mindspore/nn/optim/asgd.py +3 -1
  183. mindspore/nn/optim/ftrl.py +11 -9
  184. mindspore/nn/optim/lamb.py +1 -1
  185. mindspore/nn/optim/lazyadam.py +12 -10
  186. mindspore/nn/optim/momentum.py +7 -6
  187. mindspore/nn/optim/optimizer.py +2 -2
  188. mindspore/nn/optim/proximal_ada_grad.py +12 -10
  189. mindspore/nn/optim/rmsprop.py +13 -12
  190. mindspore/nn/optim/rprop.py +9 -7
  191. mindspore/nn/optim/sgd.py +9 -6
  192. mindspore/nn/optim/tft_wrapper.py +5 -2
  193. mindspore/nn/probability/bijector/bijector.py +17 -11
  194. mindspore/nn/probability/bijector/gumbel_cdf.py +5 -5
  195. mindspore/nn/probability/bijector/invert.py +2 -2
  196. mindspore/nn/probability/bijector/scalar_affine.py +3 -3
  197. mindspore/nn/probability/bijector/softplus.py +3 -2
  198. mindspore/nn/probability/distribution/beta.py +3 -3
  199. mindspore/nn/probability/distribution/categorical.py +1 -1
  200. mindspore/nn/probability/distribution/cauchy.py +4 -2
  201. mindspore/nn/probability/distribution/exponential.py +6 -7
  202. mindspore/nn/probability/distribution/gamma.py +2 -2
  203. mindspore/nn/probability/distribution/gumbel.py +2 -2
  204. mindspore/nn/probability/distribution/half_normal.py +5 -3
  205. mindspore/nn/probability/distribution/logistic.py +5 -3
  206. mindspore/nn/probability/distribution/poisson.py +1 -1
  207. mindspore/nn/probability/distribution/uniform.py +5 -3
  208. mindspore/nn/reinforcement/_tensors_queue.py +1 -1
  209. mindspore/nn/reinforcement/tensor_array.py +1 -1
  210. mindspore/nn/wrap/__init__.py +6 -6
  211. mindspore/nn/wrap/cell_wrapper.py +178 -117
  212. mindspore/nn/wrap/grad_reducer.py +45 -36
  213. mindspore/nn/wrap/loss_scale.py +3 -3
  214. mindspore/numpy/array_creations.py +3 -3
  215. mindspore/numpy/array_ops.py +1 -1
  216. mindspore/numpy/utils.py +1 -2
  217. mindspore/numpy/utils_const.py +1 -2
  218. mindspore/opencv_core452.dll +0 -0
  219. mindspore/opencv_imgcodecs452.dll +0 -0
  220. mindspore/opencv_imgproc452.dll +0 -0
  221. mindspore/ops/__init__.py +3 -2
  222. mindspore/ops/_grad_experimental/grad_comm_ops.py +18 -3
  223. mindspore/ops/_grad_experimental/grad_debug_ops.py +8 -1
  224. mindspore/ops/_grad_experimental/taylor_rule.py +29 -0
  225. mindspore/ops/_register_for_op.py +0 -11
  226. mindspore/{ops_generate → ops/_utils}/arg_dtype_cast.py +123 -4
  227. mindspore/{ops_generate → ops/_utils}/arg_handler.py +3 -4
  228. mindspore/ops/_vmap/vmap_array_ops.py +32 -6
  229. mindspore/ops/_vmap/vmap_grad_nn_ops.py +2 -1
  230. mindspore/ops/_vmap/vmap_math_ops.py +4 -7
  231. mindspore/ops/_vmap/vmap_nn_ops.py +9 -8
  232. mindspore/ops/auto_generate/__init__.py +4 -3
  233. mindspore/ops/auto_generate/cpp_create_prim_instance_helper.py +127 -52
  234. mindspore/ops/auto_generate/gen_extend_func.py +286 -208
  235. mindspore/ops/auto_generate/gen_ops_def.py +2783 -2335
  236. mindspore/ops/auto_generate/gen_ops_prim.py +8992 -2686
  237. mindspore/ops/auto_generate/pyboost_inner_prim.py +106 -76
  238. mindspore/ops/composite/__init__.py +2 -1
  239. mindspore/ops/composite/base.py +19 -24
  240. mindspore/ops/composite/math_ops.py +6 -16
  241. mindspore/ops/composite/multitype_ops/__init__.py +5 -2
  242. mindspore/ops/composite/multitype_ops/_compile_utils.py +4 -5
  243. mindspore/ops/composite/multitype_ops/_constexpr_utils.py +1 -2
  244. mindspore/ops/composite/multitype_ops/add_impl.py +2 -1
  245. mindspore/ops/composite/multitype_ops/bitwise_and_impl.py +2 -1
  246. mindspore/ops/composite/multitype_ops/bitwise_or_impl.py +2 -1
  247. mindspore/ops/composite/multitype_ops/bitwise_xor_impl.py +2 -1
  248. mindspore/ops/composite/multitype_ops/div_impl.py +6 -4
  249. mindspore/ops/composite/multitype_ops/equal_impl.py +4 -3
  250. mindspore/ops/composite/multitype_ops/floordiv_impl.py +2 -1
  251. mindspore/ops/composite/multitype_ops/getitem_impl.py +3 -2
  252. mindspore/ops/composite/multitype_ops/greater_equal_impl.py +4 -3
  253. mindspore/ops/composite/multitype_ops/greater_impl.py +4 -3
  254. mindspore/ops/composite/multitype_ops/in_impl.py +2 -1
  255. mindspore/ops/composite/multitype_ops/invert_impl.py +50 -0
  256. mindspore/ops/composite/multitype_ops/left_shift_impl.py +2 -1
  257. mindspore/ops/composite/multitype_ops/less_equal_impl.py +4 -3
  258. mindspore/ops/composite/multitype_ops/less_impl.py +4 -3
  259. mindspore/ops/composite/multitype_ops/logic_not_impl.py +3 -2
  260. mindspore/ops/composite/multitype_ops/logical_and_impl.py +2 -1
  261. mindspore/ops/composite/multitype_ops/logical_or_impl.py +2 -1
  262. mindspore/ops/composite/multitype_ops/mod_impl.py +2 -1
  263. mindspore/ops/composite/multitype_ops/mul_impl.py +3 -2
  264. mindspore/ops/composite/multitype_ops/negative_impl.py +2 -1
  265. mindspore/ops/composite/multitype_ops/not_equal_impl.py +2 -1
  266. mindspore/ops/composite/multitype_ops/not_in_impl.py +2 -1
  267. mindspore/ops/composite/multitype_ops/ones_like_impl.py +18 -0
  268. mindspore/ops/composite/multitype_ops/pow_impl.py +2 -1
  269. mindspore/ops/composite/multitype_ops/right_shift_impl.py +2 -1
  270. mindspore/ops/composite/multitype_ops/setitem_impl.py +2 -1
  271. mindspore/ops/composite/multitype_ops/sub_impl.py +2 -1
  272. mindspore/ops/function/__init__.py +28 -2
  273. mindspore/ops/function/_add_attr_func.py +58 -0
  274. mindspore/ops/function/array_func.py +1631 -2347
  275. mindspore/ops/function/clip_func.py +38 -45
  276. mindspore/ops/function/debug_func.py +36 -44
  277. mindspore/ops/function/grad/__init__.py +1 -0
  278. mindspore/ops/function/grad/grad_func.py +104 -71
  279. mindspore/ops/function/image_func.py +1 -1
  280. mindspore/ops/function/linalg_func.py +46 -78
  281. mindspore/ops/function/math_func.py +3024 -3855
  282. mindspore/ops/function/nn_func.py +678 -274
  283. mindspore/ops/function/other_func.py +159 -1
  284. mindspore/ops/function/parameter_func.py +17 -30
  285. mindspore/ops/function/random_func.py +216 -361
  286. mindspore/ops/function/reshard_func.py +4 -70
  287. mindspore/ops/function/sparse_func.py +3 -3
  288. mindspore/ops/function/sparse_unary_func.py +5 -5
  289. mindspore/ops/function/spectral_func.py +25 -58
  290. mindspore/ops/function/vmap_func.py +26 -18
  291. mindspore/ops/functional.py +8 -5
  292. mindspore/ops/functional_overload.py +655 -4
  293. mindspore/ops/op_info_register.py +32 -244
  294. mindspore/ops/operations/__init__.py +21 -14
  295. mindspore/ops/operations/_custom_ops_utils.py +235 -0
  296. mindspore/ops/operations/_grad_ops.py +1 -10
  297. mindspore/ops/operations/_inner_ops.py +5 -76
  298. mindspore/ops/operations/_ms_kernel.py +4 -10
  299. mindspore/ops/operations/_rl_inner_ops.py +1 -1
  300. mindspore/ops/operations/_scalar_ops.py +3 -2
  301. mindspore/ops/operations/_sequence_ops.py +1 -1
  302. mindspore/ops/operations/_tensor_array.py +1 -1
  303. mindspore/ops/operations/array_ops.py +39 -24
  304. mindspore/ops/operations/comm_ops.py +150 -107
  305. mindspore/ops/operations/custom_ops.py +287 -32
  306. mindspore/ops/operations/debug_ops.py +119 -16
  307. mindspore/ops/operations/inner_ops.py +1 -1
  308. mindspore/ops/operations/linalg_ops.py +1 -58
  309. mindspore/ops/operations/manually_defined/_inner.py +1 -1
  310. mindspore/ops/operations/manually_defined/ops_def.py +746 -79
  311. mindspore/ops/operations/math_ops.py +21 -18
  312. mindspore/ops/operations/nn_ops.py +67 -224
  313. mindspore/ops/operations/other_ops.py +62 -9
  314. mindspore/ops/operations/random_ops.py +13 -7
  315. mindspore/ops/operations/reshard_ops.py +1 -1
  316. mindspore/ops/operations/sparse_ops.py +2 -2
  317. mindspore/ops/primitive.py +43 -32
  318. mindspore/ops/tensor_method.py +243 -17
  319. mindspore/ops_generate/__init__.py +0 -5
  320. mindspore/ops_generate/aclnn/__init__.py +0 -0
  321. mindspore/ops_generate/{aclnn_kernel_register_auto_cc_generator.py → aclnn/aclnn_kernel_register_auto_cc_generator.py} +43 -18
  322. mindspore/ops_generate/{gen_aclnn_implement.py → aclnn/gen_aclnn_implement.py} +49 -51
  323. mindspore/ops_generate/api/__init__.py +0 -0
  324. mindspore/ops_generate/{add_tensor_docs_generator.py → api/add_tensor_docs_generator.py} +9 -7
  325. mindspore/ops_generate/{cpp_create_prim_instance_helper_generator.py → api/cpp_create_prim_instance_helper_generator.py} +6 -9
  326. mindspore/ops_generate/{functional_map_cpp_generator.py → api/functional_map_cpp_generator.py} +25 -12
  327. mindspore/ops_generate/{functional_overload_py_generator.py → api/functional_overload_py_generator.py} +8 -6
  328. mindspore/ops_generate/{functions_cc_generator.py → api/functions_cc_generator.py} +14 -10
  329. mindspore/ops_generate/api/gen_api.py +103 -0
  330. mindspore/ops_generate/{op_api_proto.py → api/op_api_proto.py} +98 -69
  331. mindspore/ops_generate/{tensor_func_reg_cpp_generator.py → api/tensor_func_reg_cpp_generator.py} +82 -43
  332. mindspore/ops_generate/common/__init__.py +0 -0
  333. mindspore/ops_generate/common/gen_constants.py +91 -0
  334. mindspore/ops_generate/{gen_utils.py → common/gen_utils.py} +72 -19
  335. mindspore/ops_generate/{op_proto.py → common/op_proto.py} +64 -1
  336. mindspore/ops_generate/{template.py → common/template.py} +96 -84
  337. mindspore/ops_generate/gen_ops.py +23 -325
  338. mindspore/ops_generate/op_def/__init__.py +0 -0
  339. mindspore/ops_generate/op_def/gen_op_def.py +90 -0
  340. mindspore/ops_generate/{lite_ops_cpp_generator.py → op_def/lite_ops_cpp_generator.py} +47 -11
  341. mindspore/ops_generate/{ops_def_cc_generator.py → op_def/ops_def_cc_generator.py} +18 -10
  342. mindspore/ops_generate/{ops_def_h_generator.py → op_def/ops_def_h_generator.py} +5 -5
  343. mindspore/ops_generate/{ops_name_h_generator.py → op_def/ops_name_h_generator.py} +30 -15
  344. mindspore/ops_generate/op_def/ops_primitive_h_generator.py +125 -0
  345. mindspore/ops_generate/op_def_py/__init__.py +0 -0
  346. mindspore/ops_generate/op_def_py/gen_op_def_py.py +47 -0
  347. mindspore/ops_generate/{op_def_py_generator.py → op_def_py/op_def_py_generator.py} +6 -5
  348. mindspore/ops_generate/{op_prim_py_generator.py → op_def_py/op_prim_py_generator.py} +24 -15
  349. mindspore/ops_generate/pyboost/__init__.py +0 -0
  350. mindspore/ops_generate/{auto_grad_impl_cc_generator.py → pyboost/auto_grad_impl_cc_generator.py} +11 -7
  351. mindspore/ops_generate/{auto_grad_reg_cc_generator.py → pyboost/auto_grad_reg_cc_generator.py} +7 -7
  352. mindspore/ops_generate/{gen_pyboost_func.py → pyboost/gen_pyboost_func.py} +40 -16
  353. mindspore/ops_generate/{op_template_parser.py → pyboost/op_template_parser.py} +105 -24
  354. mindspore/ops_generate/{pyboost_functions_cpp_generator.py → pyboost/pyboost_functions_cpp_generator.py} +55 -18
  355. mindspore/ops_generate/{pyboost_functions_h_generator.py → pyboost/pyboost_functions_h_generator.py} +42 -10
  356. mindspore/ops_generate/{pyboost_functions_py_generator.py → pyboost/pyboost_functions_py_generator.py} +6 -6
  357. mindspore/ops_generate/{pyboost_grad_function_cpp_generator.py → pyboost/pyboost_grad_function_cpp_generator.py} +11 -10
  358. mindspore/ops_generate/{pyboost_inner_prim_generator.py → pyboost/pyboost_inner_prim_generator.py} +8 -7
  359. mindspore/ops_generate/{pyboost_native_grad_functions_generator.py → pyboost/pyboost_native_grad_functions_generator.py} +14 -10
  360. mindspore/ops_generate/{pyboost_op_cpp_code_generator.py → pyboost/pyboost_op_cpp_code_generator.py} +140 -53
  361. mindspore/ops_generate/{pyboost_overload_functions_cpp_generator.py → pyboost/pyboost_overload_functions_cpp_generator.py} +28 -15
  362. mindspore/ops_generate/{pyboost_utils.py → pyboost/pyboost_utils.py} +88 -4
  363. mindspore/ops_generate/resources/__init__.py +0 -0
  364. mindspore/ops_generate/resources/resource_list.py +30 -0
  365. mindspore/ops_generate/resources/resource_loader.py +36 -0
  366. mindspore/ops_generate/resources/resource_manager.py +64 -0
  367. mindspore/ops_generate/resources/yaml_loader.py +88 -0
  368. mindspore/ops_generate/tensor_py_cc_generator.py +122 -0
  369. mindspore/parallel/__init__.py +6 -2
  370. mindspore/parallel/_auto_parallel_context.py +140 -12
  371. mindspore/parallel/_cell_wrapper.py +132 -15
  372. mindspore/parallel/_parallel_serialization.py +95 -4
  373. mindspore/parallel/_ps_context.py +1 -1
  374. mindspore/parallel/_recovery_context.py +7 -2
  375. mindspore/parallel/_tensor.py +142 -18
  376. mindspore/parallel/_utils.py +198 -25
  377. mindspore/parallel/algo_parameter_config.py +3 -3
  378. mindspore/parallel/auto_parallel.py +732 -0
  379. mindspore/parallel/checkpoint_convert.py +159 -0
  380. mindspore/parallel/checkpoint_transform.py +658 -37
  381. mindspore/parallel/cluster/process_entity/_api.py +151 -19
  382. mindspore/parallel/cluster/run.py +1 -1
  383. mindspore/parallel/function/__init__.py +24 -0
  384. mindspore/parallel/function/reshard_func.py +258 -0
  385. mindspore/parallel/nn/__init__.py +25 -0
  386. mindspore/parallel/nn/parallel_cell_wrapper.py +263 -0
  387. mindspore/parallel/nn/parallel_grad_reducer.py +169 -0
  388. mindspore/parallel/parameter_broadcast.py +24 -13
  389. mindspore/parallel/shard.py +137 -62
  390. mindspore/parallel/transform_safetensors.py +288 -95
  391. mindspore/pgodb140.dll +0 -0
  392. mindspore/pgort140.dll +0 -0
  393. mindspore/profiler/__init__.py +9 -5
  394. mindspore/profiler/analysis/parser/ascend_cann_parser.py +6 -2
  395. mindspore/profiler/analysis/parser/ms_framework_parser.py +4 -4
  396. mindspore/profiler/analysis/parser/timeline_assembly_factory/ascend_timeline_assembler.py +7 -4
  397. mindspore/profiler/analysis/parser/timeline_assembly_factory/trace_view_container.py +25 -0
  398. mindspore/profiler/analysis/parser/timeline_creator/fwk_timeline_creator.py +3 -3
  399. mindspore/profiler/analysis/parser/timeline_event/fwk_event.py +241 -86
  400. mindspore/profiler/analysis/viewer/ascend_communication_viewer.py +41 -2
  401. mindspore/profiler/analysis/viewer/ascend_kernel_details_viewer.py +33 -35
  402. mindspore/profiler/analysis/viewer/ascend_memory_viewer.py +7 -0
  403. mindspore/profiler/analysis/viewer/ascend_op_memory_viewer.py +8 -3
  404. mindspore/profiler/analysis/viewer/ascend_step_trace_time_viewer.py +141 -30
  405. mindspore/profiler/analysis/viewer/ms_dataset_viewer.py +5 -6
  406. mindspore/profiler/common/ascend_msprof_exporter.py +5 -4
  407. mindspore/profiler/common/constant.py +12 -0
  408. mindspore/profiler/common/msprof_cmd_tool.py +42 -23
  409. mindspore/profiler/common/path_manager.py +24 -0
  410. mindspore/profiler/common/profiler_context.py +26 -2
  411. mindspore/profiler/common/profiler_meta_data.py +74 -0
  412. mindspore/profiler/common/profiler_parameters.py +59 -18
  413. mindspore/profiler/common/profiler_path_manager.py +66 -7
  414. mindspore/profiler/dynamic_profiler.py +112 -79
  415. mindspore/profiler/envprofiler.py +26 -1
  416. mindspore/profiler/experimental_config.py +197 -0
  417. mindspore/profiler/mstx.py +57 -14
  418. mindspore/profiler/platform/npu_profiler.py +33 -7
  419. mindspore/profiler/profiler.py +541 -45
  420. mindspore/profiler/profiler_action_controller.py +1 -1
  421. mindspore/profiler/profiler_interface.py +4 -0
  422. mindspore/profiler/schedule.py +57 -22
  423. mindspore/rewrite/api/node.py +15 -13
  424. mindspore/rewrite/api/symbol_tree.py +1 -1
  425. mindspore/run_check/_check_version.py +25 -14
  426. mindspore/run_check/run_check.py +1 -1
  427. mindspore/runtime/__init__.py +2 -2
  428. mindspore/runtime/executor.py +40 -11
  429. mindspore/runtime/memory.py +37 -13
  430. mindspore/safeguard/rewrite_obfuscation.py +12 -9
  431. mindspore/swresample-4.dll +0 -0
  432. mindspore/swscale-6.dll +0 -0
  433. mindspore/tbbmalloc.dll +0 -0
  434. mindspore/tinyxml2.dll +0 -0
  435. mindspore/train/__init__.py +8 -8
  436. mindspore/train/_utils.py +43 -9
  437. mindspore/train/amp.py +1 -1
  438. mindspore/train/callback/__init__.py +2 -2
  439. mindspore/train/callback/_callback.py +2 -16
  440. mindspore/train/callback/_checkpoint.py +24 -40
  441. mindspore/train/callback/_cluster_monitor.py +14 -18
  442. mindspore/train/callback/_flops_collector.py +2 -3
  443. mindspore/train/callback/_history.py +7 -4
  444. mindspore/train/callback/_lambda_callback.py +2 -2
  445. mindspore/train/callback/_landscape.py +0 -3
  446. mindspore/train/callback/_loss_monitor.py +2 -1
  447. mindspore/train/callback/_on_request_exit.py +6 -5
  448. mindspore/train/callback/_reduce_lr_on_plateau.py +11 -6
  449. mindspore/train/callback/_summary_collector.py +8 -13
  450. mindspore/train/callback/_time_monitor.py +2 -1
  451. mindspore/train/callback/{_tft_register.py → _train_fault_tolerance.py} +204 -105
  452. mindspore/train/data_sink.py +25 -2
  453. mindspore/train/dataset_helper.py +4 -5
  454. mindspore/train/loss_scale_manager.py +8 -7
  455. mindspore/train/metrics/accuracy.py +3 -3
  456. mindspore/train/metrics/confusion_matrix.py +9 -9
  457. mindspore/train/metrics/error.py +3 -3
  458. mindspore/train/metrics/hausdorff_distance.py +4 -4
  459. mindspore/train/metrics/mean_surface_distance.py +3 -3
  460. mindspore/train/metrics/metric.py +0 -12
  461. mindspore/train/metrics/occlusion_sensitivity.py +4 -2
  462. mindspore/train/metrics/precision.py +8 -6
  463. mindspore/train/metrics/recall.py +9 -9
  464. mindspore/train/metrics/root_mean_square_surface_distance.py +2 -2
  465. mindspore/train/mind_ir_pb2.py +19 -12
  466. mindspore/train/model.py +262 -127
  467. mindspore/train/serialization.py +246 -988
  468. mindspore/train/summary/_summary_adapter.py +2 -2
  469. mindspore/train/summary/summary_record.py +1 -1
  470. mindspore/turbojpeg.dll +0 -0
  471. mindspore/utils/__init__.py +3 -2
  472. mindspore/utils/dryrun.py +4 -2
  473. mindspore/utils/hooks.py +81 -0
  474. mindspore/utils/runtime_execution_order_check.py +2 -0
  475. mindspore/utils/utils.py +138 -4
  476. mindspore/vcmeta.dll +0 -0
  477. mindspore/vcruntime140.dll +0 -0
  478. mindspore/vcruntime140_1.dll +0 -0
  479. mindspore/version.py +1 -1
  480. {mindspore-2.5.0.dist-info → mindspore-2.6.0.dist-info}/METADATA +2 -1
  481. {mindspore-2.5.0.dist-info → mindspore-2.6.0.dist-info}/RECORD +485 -440
  482. mindspore/_install_custom.py +0 -43
  483. mindspore/common/_register_for_adapter.py +0 -74
  484. mindspore/ops/auto_generate/gen_arg_dtype_cast.py +0 -252
  485. mindspore/ops/auto_generate/gen_arg_handler.py +0 -136
  486. mindspore/ops/operations/_opaque_predicate_registry.py +0 -41
  487. mindspore/ops_generate/gen_constants.py +0 -190
  488. mindspore/ops_generate/gen_ops_inner_prim.py +0 -131
  489. mindspore/ops_generate/ops_primitive_h_generator.py +0 -81
  490. /mindspore/ops_generate/{base_generator.py → common/base_generator.py} +0 -0
  491. {mindspore-2.5.0.dist-info → mindspore-2.6.0.dist-info}/WHEEL +0 -0
  492. {mindspore-2.5.0.dist-info → mindspore-2.6.0.dist-info}/entry_points.txt +0 -0
  493. {mindspore-2.5.0.dist-info → mindspore-2.6.0.dist-info}/top_level.txt +0 -0
@@ -286,37 +286,40 @@ class BatchNorm2d(_BatchNorm):
286
286
  Note that the formula for updating the :math:`moving\_mean` and :math:`moving\_var` is
287
287
 
288
288
  .. math::
289
- \text{moving_mean}=\text{moving_mean*momentum}+μ_β\text{*(1momentum)}\\
290
- \text{moving_var}=\text{moving_var*momentum}+σ^2_β\text{*(1momentum)}
289
+ \text{moving_mean}=\text{moving_mean*momentum}+μ_β\text{*(1-momentum)}\\
290
+ \text{moving_var}=\text{moving_var*momentum}+σ^2_β\text{*(1-momentum)}
291
291
 
292
292
  where :math:`moving\_mean` is the updated mean, :math:`moving\_var` is the updated variance,
293
- :math:`μ_β, σ^2_β` are the observed value (mean and variance) of each batch of data.
293
+ :math:`μ_β, σ^2_β` are the observed value (mean and variance respectively) of each batch of data.
294
294
 
295
295
  Args:
296
296
  num_features (int): The number of channels of the input tensor. Expected input size is :math:`(N, C, H, W)`,
297
297
  `C` represents the number of channels.
298
- eps (float): :math:`\epsilon` added to the denominator for numerical stability. Default: ``1e-5`` .
299
- momentum (float): A floating hyperparameter of the momentum for the
298
+ eps (float, optional): :math:`\epsilon` added to the denominator for numerical stability. Default: ``1e-5`` .
299
+ momentum (float, optional): A floating hyperparameter of the momentum for the
300
300
  running_mean and running_var computation. Default: ``0.9`` .
301
- affine (bool): A bool value. When set to ``True`` , :math:`\gamma` and :math:`\beta` can be learned.
301
+ affine (bool, optional): A bool value. When set to ``True`` , :math:`\gamma` and :math:`\beta` can be learned.
302
302
  Default: ``True`` .
303
- gamma_init (Union[Tensor, str, Initializer, numbers.Number]): Initializer for the :math:`\gamma` weight.
303
+ gamma_init (Union[Tensor, str, Initializer, numbers.Number], optional): Initializer for the
304
+ :math:`\gamma` weight.
304
305
  The values of str refer to the function `mindspore.common.initializer
305
306
  <https://www.mindspore.cn/docs/en/master/api_python/mindspore.common.initializer.html>`_
306
307
  including ``'zeros'`` , ``'ones'`` , etc. Default: ``'ones'`` .
307
- beta_init (Union[Tensor, str, Initializer, numbers.Number]): Initializer for the :math:`\beta` weight.
308
+ beta_init (Union[Tensor, str, Initializer, numbers.Number], optional): Initializer for the
309
+ :math:`\beta` weight.
308
310
  The values of str refer to the function `mindspore.common.initializer
309
311
  <https://www.mindspore.cn/docs/en/master/api_python/mindspore.common.initializer.html>`_
310
312
  including ``'zeros'`` , ``'ones'`` , etc. Default: ``'zeros'`` .
311
- moving_mean_init (Union[Tensor, str, Initializer, numbers.Number]): Initializer for the moving mean.
313
+ moving_mean_init (Union[Tensor, str, Initializer, numbers.Number], optional): Initializer for the moving mean.
312
314
  The values of str refer to the function `mindspore.common.initializer
313
315
  <https://www.mindspore.cn/docs/en/master/api_python/mindspore.common.initializer.html>`_
314
316
  including ``'zeros'`` , ``'ones'`` , etc. Default: ``'zeros'`` .
315
- moving_var_init (Union[Tensor, str, Initializer, numbers.Number]): Initializer for the moving variance.
317
+ moving_var_init (Union[Tensor, str, Initializer, numbers.Number], optional): Initializer for
318
+ the moving variance.
316
319
  The values of str refer to the function `mindspore.common.initializer
317
320
  <https://www.mindspore.cn/docs/en/master/api_python/mindspore.common.initializer.html>`_
318
321
  including ``'zeros'`` , ``'ones'`` , etc. Default: ``'ones'`` .
319
- use_batch_statistics (bool): Default: ``None`` .
322
+ use_batch_statistics (bool, optional): Default: ``None`` .
320
323
 
321
324
  - If ``true`` , use the mean value and variance value of current batch data and track running mean
322
325
  and running variance.
@@ -325,9 +328,9 @@ class BatchNorm2d(_BatchNorm):
325
328
  training and evaluation mode. During training, the parameter is set to true, and during evaluation, the
326
329
  parameter is set to false.
327
330
 
328
- data_format (str): The optional value for data format, is ``'NHWC'`` or ``'NCHW'`` .
331
+ data_format (str, optional): The optional value for data format, is ``'NHWC'`` or ``'NCHW'`` .
329
332
  Default: ``'NCHW'`` .
330
- dtype (:class:`mindspore.dtype`): Dtype of Parameters. Default: ``mstype.float32`` .
333
+ dtype (:class:`mindspore.dtype`, optional): Dtype of Parameters. Default: ``mstype.float32`` .
331
334
 
332
335
  Inputs:
333
336
  - **x** (Tensor) - Tensor of shape :math:`(N, C, H, W)`. Supported types: float16, float32.
@@ -340,7 +343,7 @@ class BatchNorm2d(_BatchNorm):
340
343
  TypeError: If `eps` is not a float.
341
344
  ValueError: If `num_features` is less than 1.
342
345
  ValueError: If `momentum` is not in range [0, 1].
343
- ValueError: If `data_format` is neither 'NHWC' not 'NCHW'.
346
+ ValueError: If `data_format` is neither ``'NHWC'`` not ``'NCHW'``.
344
347
 
345
348
  Supported Platforms:
346
349
  ``Ascend`` ``GPU`` ``CPU``
@@ -508,32 +511,34 @@ class SyncBatchNorm(_BatchNorm):
508
511
 
509
512
  Args:
510
513
  num_features (int): `C` from an expected input of size :math:`(N, C, H, W)`.
511
- eps (float): :math:`\epsilon`, a value added to the denominator for numerical stability. Default: ``1e-5`` .
512
- momentum (float): A floating hyperparameter of the momentum for the
514
+ eps (float, optional): :math:`\epsilon`, a value added to the denominator for numerical stability.
515
+ Default: ``1e-5`` .
516
+ momentum (float, optional): A floating hyperparameter of the momentum for the
513
517
  running_mean and running_var computation. Default: ``0.9`` .
514
- affine (bool): A bool value. When set to ``True`` , :math:`\gamma` and :math:`\beta` can be learned.
518
+ affine (bool, optional): A bool value. When set to ``True`` , :math:`\gamma` and :math:`\beta` are learnable
519
+ parameters. When set to ``False`` , :math:`\gamma` and :math:`\beta` are unlearnable parameters.
515
520
  Default: ``True`` .
516
- gamma_init (Union[Tensor, str, Initializer, numbers.Number]): Initializer for the :math:`\gamma` weight.
517
- The values of str refer to the function `initializer` including ``'zeros'`` , ``'ones'`` ,
521
+ gamma_init (Union[Tensor, str, Initializer, numbers.Number], optional): Initializer for the :math:`\gamma`
522
+ weight. The values of str refer to the function `initializer` including ``'zeros'`` , ``'ones'`` ,
518
523
  ``'xavier_uniform'`` , ``'he_uniform'`` , etc. Default: ``'ones'`` .
519
- beta_init (Union[Tensor, str, Initializer, numbers.Number]): Initializer for the :math:`\beta` weight.
524
+ beta_init (Union[Tensor, str, Initializer, numbers.Number], optional): Initializer for the :math:`\beta` weight.
520
525
  The values of str refer to the function `initializer` including ``'zeros'`` , ``'ones'`` ,
521
526
  ``'xavier_uniform'`` , ``'he_uniform'`` , etc. Default: ``'zeros'`` .
522
- moving_mean_init (Union[Tensor, str, Initializer, numbers.Number]): Initializer for the moving mean.
527
+ moving_mean_init (Union[Tensor, str, Initializer, numbers.Number], optional): Initializer for the moving mean.
523
528
  The values of str refer to the function `initializer` including ``'zeros'`` , ``'ones'`` ,
524
529
  ``'xavier_uniform'`` , ``'he_uniform'`` , etc. Default: ``'zeros'`` .
525
- moving_var_init (Union[Tensor, str, Initializer, numbers.Number]): Initializer for the moving variance.
526
- The values of str refer to the function `initializer` including ``'zeros'`` , ``'ones'`` ,
530
+ moving_var_init (Union[Tensor, str, Initializer, numbers.Number], optional): Initializer for the moving
531
+ variance. The values of str refer to the function `initializer` including ``'zeros'`` , ``'ones'`` ,
527
532
  ``'xavier_uniform'`` , ``'he_uniform'`` , etc. Default: ``'ones'`` .
528
- use_batch_statistics (bool): If ``true`` , use the mean value and variance value of current batch data. If
529
- ``false`` , use the mean value and variance value of specified value. If ``None`` , training process will
530
- use the mean and variance of current batch data and track the running mean and variance, eval process will
531
- use the running mean and variance. Default: ``None`` .
532
- process_groups (list): A list to divide devices into different sync groups, containing N subtraction lists.
533
- Each subtraction list contains int numbers identifying rank ids which need to be synchronized in the same
534
- group. All int values must be in [0, rank_size) and different from each other. Default: ``None`` ,
533
+ use_batch_statistics (bool, optional): If ``true`` , use the mean value and variance value of current batch
534
+ data. If ``false`` , use the mean value and variance value of specified value. If ``None`` , training
535
+ process will use the mean and variance of current batch data and track the running mean and variance, eval
536
+ process will use the running mean and variance. Default: ``None`` .
537
+ process_groups (list, optional): A list to divide devices into different sync groups, containing N subtraction
538
+ lists. Each subtraction list contains int numbers identifying rank ids which need to be synchronized in the
539
+ same group. All int values must be in [0, rank_size) and different from each other. Default: ``None`` ,
535
540
  indicating synchronization across all devices.
536
- dtype (:class:`mindspore.dtype`): Dtype of Parameters. Default: ``mstype.float32`` .
541
+ dtype (:class:`mindspore.dtype`, optional): Dtype of Parameters. Default: ``mstype.float32`` .
537
542
 
538
543
  Inputs:
539
544
  - **x** (Tensor) - Tensor of shape :math:`(N, C_{in}, H_{in}, W_{in})`.
@@ -558,14 +563,14 @@ class SyncBatchNorm(_BatchNorm):
558
563
 
559
564
  For the Ascend devices, users need to prepare the rank table, set rank_id and device_id.
560
565
  Please see the `Ascend tutorial
561
- <https://www.mindspore.cn/docs/en/master/model_train/parallel/rank_table.html>`_
566
+ <https://www.mindspore.cn/tutorials/en/master/parallel/rank_table.html>`_
562
567
  for more details.
563
568
 
564
569
  For the GPU devices, users need to prepare the host file and mpi, please see the `mpirun Startup
565
- <https://www.mindspore.cn/docs/en/master/model_train/parallel/mpirun.html>`_ .
570
+ <https://www.mindspore.cn/tutorials/en/master/parallel/mpirun.html>`_ .
566
571
 
567
572
  For the CPU device, users need to write a dynamic cluster startup script, please see the `Dynamic Cluster
568
- Startup <https://www.mindspore.cn/docs/en/master/model_train/parallel/dynamic_cluster.html>`_ .
573
+ Startup <https://www.mindspore.cn/tutorials/en/master/parallel/dynamic_cluster.html>`_ .
569
574
 
570
575
  This example should be run with multiple devices.
571
576
 
@@ -987,8 +992,8 @@ class InstanceNorm1d(_InstanceNorm):
987
992
  ValueError: If `num_features` is less than 1.
988
993
  ValueError: If `momentum` is not in range [0, 1].
989
994
  ValueError: If the shape of `gamma_init` / `beta_init` is not :math:`(C)`.
990
- KeyError: If any of `gamma_init`/`beta_init` is str and the homonymous class inheriting from `Initializer` not
991
- exists.
995
+ KeyError: If any of `gamma_init`/`beta_init` is str and
996
+ there is no homonymous class inheriting from `Initializer`.
992
997
 
993
998
  Supported Platforms:
994
999
  ``GPU``
@@ -1065,8 +1070,8 @@ class InstanceNorm2d(_InstanceNorm):
1065
1070
  ValueError: If `num_features` is less than 1.
1066
1071
  ValueError: If `momentum` is not in range [0, 1].
1067
1072
  ValueError: If the shape of `gamma_init` / `beta_init` is not :math:`(C)`.
1068
- KeyError: If any of `gamma_init`/`beta_init` is str and the homonymous class inheriting from `Initializer` not
1069
- exists.
1073
+ KeyError: If any of `gamma_init`/`beta_init` is str and
1074
+ there is no homonymous class inheriting from `Initializer`.
1070
1075
 
1071
1076
  Supported Platforms:
1072
1077
  ``GPU``
@@ -1171,10 +1176,11 @@ class GroupNorm(Cell):
1171
1176
 
1172
1177
  Group Normalization is widely used in recurrent neural networks. It applies
1173
1178
  normalization on a mini-batch of inputs for each single training case as described
1174
- in the paper `Group Normalization <https://arxiv.org/pdf/1803.08494.pdf>`_. Group Normalization
1175
- divides the channels into groups and computes within each group the mean and variance for normalization,
1176
- and it performs very stable over a wide range of batch size. :math:`\gamma` and :math:`\beta` are trainable scale
1177
- and shift.
1179
+ in the paper `Group Normalization <https://arxiv.org/pdf/1803.08494.pdf>`_.
1180
+ Group Normalization
1181
+ divides the channels into groups and computes within each group the mean and variance for normalization.
1182
+ :math:`\gamma` and :math:`\beta` are scale
1183
+ and shift values obtained by training learning.
1178
1184
  It can be described using the following formula:
1179
1185
 
1180
1186
  .. math::
@@ -32,7 +32,7 @@ from mindspore.ops.auto_generate import avg_pool1d_ext
32
32
  __all__ = ['AvgPool3d', 'MaxPool3d', 'AvgPool2d', 'MaxPool2d', 'AvgPool1d', 'MaxPool1d', 'FractionalMaxPool2d',
33
33
  'FractionalMaxPool3d', 'AdaptiveAvgPool1d', 'AdaptiveMaxPool1d', 'AdaptiveMaxPool2d', 'AdaptiveMaxPool3d',
34
34
  'AdaptiveAvgPool2d', 'AdaptiveAvgPool3d', 'MaxUnpool1d', 'MaxUnpool2d', 'MaxUnpool3d', 'LPPool1d',
35
- 'LPPool2d', 'AvgPool2dExt', 'MaxPool2dExt', 'AvgPool1dExt']
35
+ 'LPPool2d', 'AvgPool2dExt', 'AvgPool3dExt', 'MaxPool2dExt', 'AvgPool1dExt']
36
36
 
37
37
 
38
38
  class _PoolNd(Cell):
@@ -299,11 +299,12 @@ class MaxPool3d(_PoolNd):
299
299
  For Atlas training series products, this interface is not supported.
300
300
 
301
301
  Args:
302
- kernel_size (Union[int, tuple[int]]): The size of kernel used to take the maximum value,
302
+ kernel_size (Union[int, tuple[int]], optional): The size of kernel used to take the maximum value,
303
303
  is an int number or a single element tuple that represents depth, height and width of the kernel, or a tuple
304
304
  of three int numbers that represent depth, height and width respectively.
305
305
  The value must be a positive integer. Default: ``1`` .
306
- stride (Union[int, tuple[int]]): The moving stride of pooling operation, an int number or a single element tuple
306
+ stride (Union[int, tuple[int]], optional): The moving stride of pooling operation,
307
+ an int number or a single element tuple
307
308
  that represents the moving stride of pooling kernel in the directions of depth, height and the width,
308
309
  or a tuple of three int numbers that represent depth, height and width of movement respectively.
309
310
  The value must be a positive integer. If the value is None, the default value `kernel_size` is used.
@@ -324,18 +325,19 @@ class MaxPool3d(_PoolNd):
324
325
  in the depth, height and width dimension is determined by the `padding` parameter.
325
326
  If this mode is set, `padding` must be greater than or equal to 0.
326
327
 
327
- padding (Union(int, tuple[int], list[int])): Pooling padding value. Default: ``0`` .
328
+ padding (Union(int, tuple[int], list[int]), optional): Pooling padding value. Default: ``0`` .
328
329
  `padding` can only be an integer or a tuple/list containing one or three integers.
329
330
  If `padding` is an integer or a tuple/list containing one integer, it will be padded in six directions of
330
331
  front, back, top, bottom, left and right of the input. If `padding` is a tuple/list containing three
331
332
  integers, it will be padded in front and back of the input `padding[0]` times, up and down `padding[1]`
332
333
  times, and left and right of the input `padding[2]` times.
333
- dilation (Union(int, tuple[int])): The spacing between the elements of the kernel in convolution,
334
+ dilation (Union(int, tuple[int]), optional): The spacing between the elements of the kernel in convolution,
334
335
  used to increase the receptive field of the pooling operation. If it is a tuple, it must contain one or
335
336
  three integers. Default: ``1`` .
336
- return_indices (bool): If ``True`` , output is a Tuple of 2 Tensors, representing the maxpool result and where
337
+ return_indices (bool, optional): If ``True`` , output is a Tuple of 2 Tensors,
338
+ representing the maxpool result and where
337
339
  the max values are generated. Otherwise, only the maxpool result is returned. Default: ``False`` .
338
- ceil_mode (bool): If ``True``, use ceil to calculate output shape.
340
+ ceil_mode (bool, optional): If ``True``, use ceil to calculate output shape.
339
341
  If ``False``, use ceil to calculate output shape. Default: ``False`` .
340
342
 
341
343
  Inputs:
@@ -713,9 +715,9 @@ class MaxPool1d(_PoolNd):
713
715
  \text{input}(N_i, C_j, s_0 \times l + n)
714
716
 
715
717
  Args:
716
- kernel_size (int): The size of kernel used to take the max value, Default: ``1`` .
717
- stride (int): The distance of kernel moving, an int number that represents
718
- the width of movement is stride, Default: ``1`` .
718
+ kernel_size (int, optional): The size of kernel used to take the max value. Default: ``1`` .
719
+ stride (int, optional): The distance of kernel moving, an int number that represents
720
+ the width of movement is stride. Default: ``1`` .
719
721
  pad_mode (str, optional): Specifies the padding mode with a padding value of 0. It can be set to:
720
722
  ``"same"`` , ``"valid"`` or ``"pad"`` . Default: ``"valid"`` .
721
723
 
@@ -731,24 +733,25 @@ class MaxPool1d(_PoolNd):
731
733
  at the begin and end is determined by the `padding` parameter.
732
734
  If this mode is set, `padding` must be greater than or equal to 0.
733
735
 
734
- padding (Union(int, tuple[int], list[int])): Padding value for the pooling. Default value is ``0``.
736
+ padding (Union(int, tuple[int], list[int]), optional): Padding value for the pooling. Default value is ``0``.
735
737
  padding can only be an integer or a tuple/list containing a single integer, in which case padding times or
736
738
  padding[0] times are padded on both sides of the input.
737
- dilation (Union(int, tuple[int])): The spacing between the elements of the kernel in convolution,
739
+ dilation (Union(int, tuple[int]), optional): The spacing between the elements of the kernel in convolution,
738
740
  used to increase the receptive field of the pooling operation. If it is a tuple, its length can only be 1.
739
741
  Default: ``1`` .
740
- return_indices (bool): If ``True`` , the function will return both the result of max pooling and the indices of
742
+ return_indices (bool, optional): If ``True`` , the function will return
743
+ both the result of max pooling and the indices of
741
744
  the max elements. Default: ``False`` .
742
- ceil_mode (bool): If True, use ceil to compute the output shape instead of floor. Default: ``False`` .
745
+ ceil_mode (bool, optional): If True, use ceil to compute the output shape instead of floor. Default: ``False`` .
743
746
 
744
747
  Inputs:
745
748
  - **x** (Tensor) - Tensor of shape :math:`(N, C_{in}, L_{in})` or :math:`(C_{in}, L_{in})`.
746
749
 
747
750
  Outputs:
748
- If `return_indices` is False, output is a Tensor, with shape :math:`(N, C_{out}, L_{out})` or
751
+ If `return_indices` is ``False``, output is a Tensor, with shape :math:`(N, C_{out}, L_{out})` or
749
752
  :math:`(C_{out}, L_{out})`. It has the same data type as `x`.
750
753
 
751
- If `return_indices` is True, output is a Tuple of 2 Tensors, representing the maxpool result and where
754
+ If `return_indices` is ``True``, output is a Tuple of 2 Tensors, representing the maxpool result and where
752
755
  the max values are generated.
753
756
 
754
757
  - **output** (Tensor) - Maxpooling result, with shape :math:`(N, C_{out}, L_{out})` or
@@ -1021,6 +1024,47 @@ class AvgPool3d(_PoolNd):
1021
1024
  return out
1022
1025
 
1023
1026
 
1027
+ class AvgPool3dExt(Cell):
1028
+ r"""
1029
+ Applies a 3D average pooling over an input Tensor which can be regarded as
1030
+ a composition of 3D input planes.
1031
+
1032
+ .. warning::
1033
+ This is an experimental API that is subject to change or deletion.
1034
+
1035
+ For details, please refer to :func:`mindspore.mint.nn.functional.avg_pool3d`.
1036
+
1037
+ Supported Platforms:
1038
+ ``Ascend``
1039
+
1040
+ Examples:
1041
+ >>> import mindspore as ms
1042
+ >>> pool = ms.nn.AvgPool3dExt(kernel_size=3, stride=1)
1043
+ >>> x = ms.ops.randn(1, 2, 4, 4, 5).astype(ms.float32)
1044
+ >>> output = pool(x)
1045
+ >>> print(output.shape)
1046
+ (1, 2, 2, 2, 3)
1047
+ >>> x1 = ms.ops.randn(6, 5, 7, 7, 5).astype(ms.float32)
1048
+ >>> pool2 = ms.nn.AvgPool3dExt(4, stride=2, padding=(2, 2, 1), divisor_override=10)
1049
+ >>> output2 = pool2(x1)
1050
+ >>> print(output2.shape)
1051
+ (6, 5, 4, 4, 2)
1052
+ """
1053
+ def __init__(self, kernel_size, stride=None, padding=0, ceil_mode=False,
1054
+ count_include_pad=True, divisor_override=None):
1055
+ super(AvgPool3dExt, self).__init__()
1056
+ self.kernel_size = kernel_size
1057
+ self.stride = stride
1058
+ self.padding = padding
1059
+ self.ceil_mode = ceil_mode
1060
+ self.count_include_pad = count_include_pad
1061
+ self.divisor_override = divisor_override
1062
+
1063
+ def construct(self, input):
1064
+ return ops.function.nn_func.avg_pool3d_ext(input, self.kernel_size, self.stride, self.padding,
1065
+ self.ceil_mode, self.count_include_pad, self.divisor_override)
1066
+
1067
+
1024
1068
  class AvgPool1dExt(Cell):
1025
1069
  r"""
1026
1070
  Applies a 1D average pooling over an input Tensor which can be regarded as
@@ -1270,8 +1314,8 @@ class AvgPool1d(_PoolNd):
1270
1314
  This interface currently does not support Atlas A2 training series products.
1271
1315
 
1272
1316
  Args:
1273
- kernel_size (int): The size of kernel window used to take the average value, Default: ``1`` .
1274
- stride (int): The distance of kernel moving, an int number that represents
1317
+ kernel_size (int, optional): The size of kernel window used to take the average value, Default: ``1`` .
1318
+ stride (int, optional): The distance of kernel moving, an int number that represents
1275
1319
  the width of movement is strides, Default: ``1`` .
1276
1320
  pad_mode (str, optional): Specifies the padding mode with a padding value of 0. It can be set to:
1277
1321
  ``"same"`` , ``"valid"`` or ``"pad"`` . Default: ``"valid"`` .
@@ -1282,17 +1326,20 @@ class AvgPool1d(_PoolNd):
1282
1326
  uniformly distributed around the input, if it is odd, the excess padding is goes to the right side.
1283
1327
  If this mode is set, `padding` must be 0.
1284
1328
  - ``"valid"``: No padding is applied to the input, and the output returns the maximum
1285
- possible length. Extra pixels that could not complete a full stride will
1286
- be discarded. If this mode is set, `padding` must be 0.
1329
+ possible length. If a full stride cannot be formed, the extra pixels will be discarded.
1330
+ If this mode is set, `padding` must be 0.
1287
1331
  - ``"pad"``: Pad the input with a specified amount. In this mode, the amount of padding
1288
1332
  at the begin and end is determined by the `padding` parameter.
1289
1333
  If this mode is set, `padding` must be greater than or equal to 0.
1290
1334
 
1291
- padding (Union(int, tuple[int], list[int])): Pooling padding value, only ``"pad"`` mode can be set to non-zero.
1335
+ padding (Union(int, tuple[int], list[int]), optional): Pooling padding value,
1336
+ only ``"pad"`` mode can be set to non-zero.
1292
1337
  Default: ``0`` . padding can only be an integer or a tuple/list containing a single integer, in which case
1293
1338
  padding times or padding[0] times are padded on both sides of the input.
1294
- ceil_mode (bool): If ``True`` , use ceil to compute the output shape instead of floor. Default: ``False`` .
1295
- count_include_pad (bool): If ``True`` , averaging calculation will include the zero-padding. Default: ``True`` .
1339
+ ceil_mode (bool, optional): If ``True`` , use ceil to compute the output shape instead of floor.
1340
+ Default: ``False`` .
1341
+ count_include_pad (bool, optional): If ``True`` , averaging calculation will include the zero-padding.
1342
+ Default: ``True`` .
1296
1343
 
1297
1344
  Inputs:
1298
1345
  - **x** (Tensor) - Tensor of shape :math:`(N, C_{in}, L_{in})` or :math:`(C_{in}, L_{in})`.
@@ -1728,13 +1775,14 @@ class AdaptiveMaxPool2d(Cell):
1728
1775
  \end{align}
1729
1776
 
1730
1777
  Note:
1731
- Ascend platform only supports float16 type for input.
1778
+ In KBK mode, `output_size` does not support mutable.
1732
1779
 
1733
1780
  Args:
1734
1781
  output_size (Union[int, tuple]): The target output size. `output_size` can be a tuple :math:`(H, W)`,
1735
1782
  or an int H for :math:`(H, H)`. :math:`H` and :math:`W` can be int or None.
1736
1783
  If it is None, it means the output size is the same as the input size.
1737
- return_indices (bool): If `return_indices` is ``True`` , the indices of max value would be output.
1784
+ return_indices (bool, optional): Whether to output the index of the maximum value.
1785
+ If `return_indices` is ``True`` , the indices of max value would be output.
1738
1786
  Default: ``False`` .
1739
1787
 
1740
1788
  Inputs:
@@ -1797,15 +1845,11 @@ class AdaptiveMaxPool2d(Cell):
1797
1845
  def __init__(self, output_size, return_indices=False):
1798
1846
  """Initialize AdaptiveMaxPool2d."""
1799
1847
  super(AdaptiveMaxPool2d, self).__init__()
1800
- validator.check_value_type('return_indices', return_indices, [bool], self.cls_name)
1801
- self.adaptive_max_pool2d = ops.AdaptiveMaxPool2D(output_size)
1848
+ self.output_size = output_size
1802
1849
  self.return_indices = return_indices
1803
1850
 
1804
1851
  def construct(self, input):
1805
- output = self.adaptive_max_pool2d(input)
1806
- if self.return_indices:
1807
- return output
1808
- return output[0]
1852
+ return ops.adaptive_max_pool2d(input, self.output_size, self.return_indices)
1809
1853
 
1810
1854
 
1811
1855
  class AdaptiveMaxPool3d(Cell):
@@ -54,16 +54,16 @@ class MultiheadAttention(Cell):
54
54
  embed_dim (int): Total dimension of MultiheadAttention.
55
55
  num_heads (int): Number of attention heads. Note that `embed_dim` will be split
56
56
  across `num_heads` (i.e. each head will have dimension `embed_dim // num_heads`).
57
- dropout (float): Dropout probability of `attn_output_weights`. Default: ``0.0``.
58
- has_bias (bool): Whether adds bias to input / output projection layers. Default: ``True``.
59
- add_bias_kv (bool): Whether adds bias to the key and value sequences at axis=0. Default: ``False``.
60
- add_zero_attn (bool): Whether adds a new batch of zeros to the key and value sequences at axis=1.
57
+ dropout (float, optional): Dropout probability of `attn_output_weights`. Default: ``0.0``.
58
+ has_bias (bool, optional): Whether adds bias to input / output projection layers. Default: ``True``.
59
+ add_bias_kv (bool, optional): Whether adds bias to the key and value sequences at axis=0. Default: ``False``.
60
+ add_zero_attn (bool, optional): Whether adds a new batch of zeros to the key and value sequences at axis=1.
61
61
  Default: ``False``.
62
- kdim (int): Total number of features for keys. Default: ``None`` (`kdim=embed_dim`).
63
- vdim (int): Total number of features for values. Default: ``None`` (`vdim=embed_dim`).
64
- batch_first (bool): If ``True``, then the input and output shape are :math:`(batch, seq, feature)` ,
62
+ kdim (int, optional): Total number of features for keys. Default: ``None`` (`kdim=embed_dim`).
63
+ vdim (int, optional): Total number of features for values. Default: ``None`` (`vdim=embed_dim`).
64
+ batch_first (bool, optional): If ``True``, then the input and output shape are :math:`(batch, seq, feature)` ,
65
65
  else :math:`(seq, batch, feature)` . Default: ``False``.
66
- dtype (:class:`mindspore.dtype`): Data type of Parameter. Default: ``mstype.float32`` .
66
+ dtype (:class:`mindspore.dtype`, optional): Data type of Parameter. Default: ``mstype.float32`` .
67
67
 
68
68
  Inputs:
69
69
  - **query** (Tensor) - The query embeddings. If `query` is unbatched, the shape is :math:`(L, E_q)`,
@@ -85,7 +85,7 @@ class MultiheadAttention(Cell):
85
85
  For a binary mask, a ``True`` value indicates that the corresponding `key` value will be ignored for
86
86
  the purpose of attention. For a float mask, it will be directly added to the corresponding `key` value.
87
87
  Supported float types: float16, float32, float64. Default: ``None``.
88
- - **need_weights** (bool) - Whether returns `attn_output_weights` in addition to `attn_outputs`.
88
+ - **need_weights** (bool, optional) - Whether returns `attn_output_weights` in addition to `attn_outputs`.
89
89
  Default: ``True``.
90
90
  - **attn_mask** (Tensor, optional) - If specified, a 2D or 3D mask preventing attention to certain positions.
91
91
  Must be of shape :math:`(L, S)` or :math:`(N\cdot\text{num_heads}, L, S)`, where :math:`N` is the
@@ -94,7 +94,8 @@ class MultiheadAttention(Cell):
94
94
  in the batch. For a binary mask, a ``True`` value indicates that the corresponding position is not allowed
95
95
  to attend. For a float mask, the mask values will be added to the attention weight.
96
96
  Supported float types: float16, float32, float64. Default: ``None``.
97
- - **average_attn_weights** (bool) - If true, indicates that the returned `attn_weights` should be averaged
97
+ - **average_attn_weights** (bool, optional) - If true, indicates that
98
+ the returned `attn_weights` should be averaged
98
99
  across heads. Otherwise, `attn_weights` are provided separately per head. Note that this flag only
99
100
  has an effect when `need_weights=True`. Default: ``True`` (i.e. average weights across heads)
100
101
 
@@ -80,7 +80,8 @@ class ExponentialDecayLR(LearningRateSchedule):
80
80
  learning_rate (float): The initial value of learning rate.
81
81
  decay_rate (float): The decay rate.
82
82
  decay_steps (int): Number of steps to decay over.
83
- is_stair (bool): If true, learning rate is decayed once every `decay_steps` time. Default: ``False`` .
83
+ is_stair (bool, optional): If ``True``, learning rate is decayed once every `decay_steps` time.
84
+ Default: ``False`` .
84
85
 
85
86
  Inputs:
86
87
  - **global_step** (Tensor) - The current step number. :math:`current\_step` in the above formula.
@@ -223,7 +224,8 @@ class InverseDecayLR(LearningRateSchedule):
223
224
  learning_rate (float): The initial value of learning rate.
224
225
  decay_rate (float): The decay rate.
225
226
  decay_steps (int): Number of steps to decay over.
226
- is_stair (bool): If true, learning rate decay once every `decay_steps` times. If False, the learning rate
227
+ is_stair (bool, optional): If true, learning rate decay once every `decay_steps` times.
228
+ If False, the learning rate
227
229
  decays for every step. Default: ``False`` .
228
230
 
229
231
  Inputs:
mindspore/nn/loss/loss.py CHANGED
@@ -127,7 +127,8 @@ class LossBase(Cell):
127
127
  Args:
128
128
  x (Tensor): Tensor of shape :math:`(N, *)` where :math:`*` means, any number of
129
129
  additional dimensions.
130
- weights (Union[float, Tensor]): Optional `Tensor` whose rank is either 0, or the same rank as inputs,
130
+ weights (Union[float, Tensor], optional): Weights. When `weights` is a Tensor,
131
+ the rank is either 0, or the same rank as inputs,
131
132
  and must be broadcastable to inputs (i.e., all dimensions must be either `1`,
132
133
  or the same as the corresponding inputs dimension). Default: ``1.0`` .
133
134
 
@@ -617,7 +618,8 @@ class MarginRankingLoss(LossBase):
617
618
 
618
619
  class SmoothL1Loss(LossBase):
619
620
  r"""
620
- SmoothL1 loss function, if the absolute error element-wise between the predicted value and the target value
621
+ SmoothL1 loss function. Compare the error value element-wise and
622
+ if the absolute error between the predicted value and the target value
621
623
  is less than the set threshold `beta`, the square term is used, otherwise the absolute error term is used.
622
624
 
623
625
  Given two input :math:`x,\ y`, the SmoothL1Loss can be described as follows:
@@ -667,11 +669,13 @@ class SmoothL1Loss(LossBase):
667
669
 
668
670
  - Ascend: float16, float32, bfloat16.
669
671
  - CPU/GPU: float16, float32, float64.
672
+
670
673
  - **labels** (Tensor) - Ground truth data.
671
674
 
672
675
  - CPU/Ascend: has the same shape as the `logits`,
673
676
  `logits` and `labels` comply with the implicit type conversion rules to make the data types consistent.
674
677
  - GPU: has the same shape and dtype as the `logits`.
678
+
675
679
  Outputs:
676
680
  Tensor, if `reduction` is ``'none'``, then output is a tensor with the same shape as `logits`.
677
681
  Otherwise the shape of output tensor is :math:`()`.
@@ -732,16 +736,19 @@ class SoftMarginLoss(LossBase):
732
736
  - ``'sum'``: the output elements will be summed.
733
737
 
734
738
  Inputs:
735
- - **logits** (Tensor) - Predict data. Data type must be float16 or float32.
736
- - **labels** (Tensor) - Ground truth data, with the same type and shape as `logits`.
739
+ - **logits** (Tensor) - Predict data. Data type must be float16, float32,
740
+ bfloat16 (Among them, the Atlas training series products do not support bfloat16).
741
+ - **labels** (Tensor) - Ground truth data, with the same shape as `logits`.
742
+ In GE mode, the data type should be the same as `logits`.
737
743
 
738
744
  Outputs:
739
- Tensor or Scalar, if `reduction` is ``"none"``, its shape is the same as `logits`.
745
+ Tensor or Scalar, if `reduction` is ``'none'``, its shape is the same as `logits`.
740
746
  Otherwise, a scalar value will be returned.
741
747
 
742
748
  Raises:
743
749
  TypeError: If `logits` or `labels` is not a Tensor.
744
- TypeError: If dtype of `logits` or `labels` is neither float16 nor float32.
750
+ TypeError: If dtype of `logits` or `labels` is not float16, float32,
751
+ bfloat16 (Among them, the Atlas training series products do not support bfloat16).
745
752
  ValueError: If shape of `logits` is not the same as `labels`.
746
753
  ValueError: If `reduction` is not one of ``'none'``, ``'mean'``, ``'sum'``.
747
754
 
@@ -762,10 +769,10 @@ class SoftMarginLoss(LossBase):
762
769
 
763
770
  def __init__(self, reduction='mean'):
764
771
  super(SoftMarginLoss, self).__init__()
765
- self.soft_margin_loss = P.SoftMarginLoss(reduction)
772
+ self.reduction = reduction
766
773
 
767
774
  def construct(self, logits, labels):
768
- return self.soft_margin_loss(logits, labels)
775
+ return F.soft_margin_loss(logits, labels, self.reduction)
769
776
 
770
777
 
771
778
  class SoftmaxCrossEntropyWithLogits(LossBase):
@@ -813,8 +820,8 @@ class SoftmaxCrossEntropyWithLogits(LossBase):
813
820
 
814
821
  Raises:
815
822
  TypeError: If `sparse` is not a bool.
816
- TypeError: If `sparse` is True and dtype of `labels` is neither int32 nor int64.
817
- TypeError: If `sparse` is False and dtype of `labels` is neither float16 not float32.
823
+ TypeError: If `sparse` is ``True`` and dtype of `labels` is neither int32 nor int64.
824
+ TypeError: If `sparse` is ``False`` and dtype of `labels` is neither float16 not float32.
818
825
  ValueError: If `reduction` is not one of ``'none'``, ``'mean'``, ``'sum'``.
819
826
 
820
827
  Supported Platforms:
@@ -893,8 +900,8 @@ class DiceLoss(LossBase):
893
900
  :math:`pred` represent `logits`, :math:`true` represent `labels` .
894
901
 
895
902
  Args:
896
- smooth (float): A term added to the denominator to improve numerical stability. Should be greater than 0.
897
- Default: ``1e-5`` .
903
+ smooth (float, optional): A term added to the denominator to improve numerical stability.
904
+ Should be greater than 0. Default: ``1e-5`` .
898
905
 
899
906
  Inputs:
900
907
  - **logits** (Tensor) - Input predicted value. The data type must be float16 or float32.
@@ -938,11 +945,12 @@ class DiceLoss(LossBase):
938
945
  if label.dtype == mstype.uint8:
939
946
  raise TypeError(f"For '{self.cls_name}', the dtype of 'labels' can not be uint8.")
940
947
  intersection = self.reduce_sum(self.mul(logits.view(-1), label.view(-1)))
941
- unionset = self.reduce_sum(self.mul(logits.view(-1), logits.view(-1))) + \
942
- self.reduce_sum(self.mul(label.view(-1), label.view(-1)))
948
+ unionset_part1 = self.reduce_sum(self.mul(logits.view(-1), logits.view(-1)))
949
+ unionset_part2 = self.reduce_sum(self.mul(label.view(-1), label.view(-1)))
950
+ unionset = ops.add(unionset_part1, unionset_part2)
943
951
 
944
- single_dice_coeff = (2 * intersection) / (unionset + self.smooth)
945
- dice_loss = 1 - single_dice_coeff
952
+ single_dice_coeff = (2 * intersection) / ops.add(unionset, self.smooth)
953
+ dice_loss = ops.sub(1, single_dice_coeff)
946
954
 
947
955
  return dice_loss
948
956
 
@@ -1058,7 +1066,7 @@ class MultiClassDiceLoss(LossBase):
1058
1066
  dice_loss = self.binarydiceloss(logits[:, i], label[:, i])
1059
1067
  if self.weights is not None:
1060
1068
  _check_weights(self.weights.shape[0], label.shape[1], self.cls_name)
1061
- dice_loss *= self.weights[i]
1069
+ dice_loss = dice_loss * self.weights[i]
1062
1070
  total_loss += dice_loss
1063
1071
 
1064
1072
  return total_loss / label.shape[1]
@@ -2571,7 +2579,7 @@ class KLDivLoss(LossBase):
2571
2579
  the updating formulas of KLDivLoss algorithm are as follows,
2572
2580
 
2573
2581
  .. math::
2574
- L(x, target) = target \cdot (\log target - x)
2582
+ L(x, target) = target \cdot (\log target - \log x)
2575
2583
 
2576
2584
  Then,
2577
2585
 
@@ -2865,7 +2873,7 @@ class HingeEmbeddingLoss(LossBase):
2865
2873
  where :math:`L = \{l_1,\dots,l_N\}^\top`.
2866
2874
 
2867
2875
  Args:
2868
- margin (float, int): Threshold defined by Hinge Embedding Loss :math:`margin`.
2876
+ margin (float, int, optional): Threshold defined by Hinge Embedding Loss :math:`margin`.
2869
2877
  Represented as :math:`\Delta` in the formula. Default: ``1.0`` .
2870
2878
  reduction (str, optional): Apply specific reduction method to the output: ``'none'`` , ``'mean'`` ,
2871
2879
  ``'sum'`` . Default: ``'mean'`` .
@@ -113,8 +113,8 @@ class Adagrad(Optimizer):
113
113
  If `order_params` in the keys, other keys will be ignored and the element of 'order_params' must be in
114
114
  one group of `params`.
115
115
 
116
- accum (float): The starting value for :math:`h`, must be zero or positive values. Default: ``0.1`` .
117
- learning_rate (Union[float, int, Tensor, Iterable, LearningRateSchedule]): Default: ``0.001`` .
116
+ accum (float, optional): The starting value for :math:`h`, must be zero or positive values. Default: ``0.1`` .
117
+ learning_rate (Union[float, int, Tensor, Iterable, LearningRateSchedule], optional): Default: ``0.001`` .
118
118
 
119
119
  - float: The fixed learning rate value. Must be equal to or greater than 0.
120
120
 
@@ -130,13 +130,14 @@ class Adagrad(Optimizer):
130
130
  <https://www.mindspore.cn/docs/en/master/api_python/mindspore.nn.html#learningrateschedule-class>`_
131
131
  with step as the input to get the learning rate of current step.
132
132
 
133
- update_slots (bool): Whether the :math:`h` will be updated. Default: ``True`` .
134
- loss_scale (float): Value for the loss scale. It must be greater than 0.0. In general, use the default value.
133
+ update_slots (bool, optional): Whether the :math:`h` will be updated. Default: ``True`` .
134
+ loss_scale (float, optional): Value for the loss scale. It must be greater than 0.0. In general,
135
+ use the default value.
135
136
  Only when `FixedLossScaleManager` is used for training and the `drop_overflow_update` in
136
137
  `FixedLossScaleManager` is set to False, then this value needs to be the same as the `loss_scale` in
137
138
  `FixedLossScaleManager`. Refer to class :class:`mindspore.amp.FixedLossScaleManager` for more details.
138
139
  Default: ``1.0`` .
139
- weight_decay (Union[float, int, Cell]): Weight decay (L2 penalty). Default: ``0.0`` .
140
+ weight_decay (Union[float, int, Cell], optional): Weight decay (L2 penalty). Default: ``0.0`` .
140
141
 
141
142
  - float: The fixed weight decay value. Must be equal to or greater than 0.
142
143