mindspore 2.7.0rc1__cp311-cp311-win_amd64.whl → 2.7.1__cp311-cp311-win_amd64.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of mindspore might be problematic. Click here for more details.

Files changed (370) hide show
  1. mindspore/.commit_id +1 -1
  2. mindspore/__init__.py +5 -2
  3. mindspore/_c_dataengine.cp311-win_amd64.pyd +0 -0
  4. mindspore/_c_expression.cp311-win_amd64.pyd +0 -0
  5. mindspore/_c_mindrecord.cp311-win_amd64.pyd +0 -0
  6. mindspore/_checkparam.py +2 -2
  7. mindspore/_extends/builtin_operations.py +3 -3
  8. mindspore/_extends/parallel_compile/akg_compiler/custom.py +1109 -0
  9. mindspore/_extends/parallel_compile/akg_compiler/gen_custom_op_files.py +1 -1
  10. mindspore/_extends/parse/__init__.py +3 -3
  11. mindspore/_extends/parse/compile_config.py +24 -1
  12. mindspore/_extends/parse/deprecated/deprecated_tensor_method.py +6 -3
  13. mindspore/_extends/parse/parser.py +28 -22
  14. mindspore/_extends/parse/resources.py +1 -1
  15. mindspore/_extends/parse/standard_method.py +23 -2
  16. mindspore/_extends/parse/trope.py +2 -1
  17. mindspore/_extends/pijit/pijit_func_white_list.py +9 -27
  18. mindspore/amp.py +0 -18
  19. mindspore/avcodec-59.dll +0 -0
  20. mindspore/avdevice-59.dll +0 -0
  21. mindspore/avfilter-8.dll +0 -0
  22. mindspore/avformat-59.dll +0 -0
  23. mindspore/avutil-57.dll +0 -0
  24. mindspore/boost/base.py +29 -2
  25. mindspore/common/__init__.py +18 -12
  26. mindspore/common/_decorator.py +3 -2
  27. mindspore/common/_grad_function.py +3 -1
  28. mindspore/common/_tensor_cpp_method.py +1 -1
  29. mindspore/common/_tensor_docs.py +371 -96
  30. mindspore/common/_utils.py +7 -43
  31. mindspore/common/api.py +434 -135
  32. mindspore/common/dtype.py +98 -57
  33. mindspore/common/dump.py +7 -108
  34. mindspore/common/dynamic_shape/__init__.py +0 -0
  35. mindspore/common/{auto_dynamic_shape.py → dynamic_shape/auto_dynamic_shape.py} +15 -23
  36. mindspore/common/dynamic_shape/enable_dynamic.py +197 -0
  37. mindspore/common/file_system.py +59 -9
  38. mindspore/common/hook_handle.py +82 -3
  39. mindspore/common/jit_config.py +5 -1
  40. mindspore/common/jit_trace.py +27 -12
  41. mindspore/common/lazy_inline.py +5 -3
  42. mindspore/common/np_dtype.py +3 -3
  43. mindspore/common/parameter.py +17 -127
  44. mindspore/common/recompute.py +4 -13
  45. mindspore/common/tensor.py +50 -217
  46. mindspore/communication/_comm_helper.py +11 -1
  47. mindspore/communication/comm_func.py +138 -4
  48. mindspore/communication/management.py +85 -1
  49. mindspore/config/op_info.config +0 -15
  50. mindspore/context.py +20 -106
  51. mindspore/dataset/__init__.py +1 -1
  52. mindspore/dataset/audio/transforms.py +1 -1
  53. mindspore/dataset/core/config.py +35 -1
  54. mindspore/dataset/engine/datasets.py +338 -319
  55. mindspore/dataset/engine/datasets_user_defined.py +38 -22
  56. mindspore/dataset/engine/datasets_vision.py +1 -1
  57. mindspore/dataset/engine/validators.py +1 -15
  58. mindspore/dataset/transforms/c_transforms.py +2 -2
  59. mindspore/dataset/transforms/transforms.py +3 -3
  60. mindspore/dataset/vision/__init__.py +1 -1
  61. mindspore/dataset/vision/py_transforms.py +8 -8
  62. mindspore/dataset/vision/transforms.py +17 -5
  63. mindspore/dataset/vision/utils.py +632 -21
  64. mindspore/device_context/ascend/op_tuning.py +35 -1
  65. mindspore/dnnl.dll +0 -0
  66. mindspore/{profiler/common/validator → graph}/__init__.py +9 -1
  67. mindspore/graph/custom_pass.py +55 -0
  68. mindspore/include/api/cell.h +28 -4
  69. mindspore/include/api/cfg.h +24 -7
  70. mindspore/include/api/context.h +1 -0
  71. mindspore/include/api/delegate.h +0 -2
  72. mindspore/include/api/dual_abi_helper.h +100 -19
  73. mindspore/include/api/graph.h +14 -1
  74. mindspore/include/api/kernel.h +16 -3
  75. mindspore/include/api/kernel_api.h +9 -1
  76. mindspore/include/api/metrics/accuracy.h +9 -0
  77. mindspore/include/api/model.h +5 -1
  78. mindspore/include/api/model_group.h +4 -0
  79. mindspore/include/api/model_parallel_runner.h +2 -0
  80. mindspore/include/api/status.h +48 -10
  81. mindspore/include/api/types.h +6 -1
  82. mindspore/include/dataset/constants.h +9 -0
  83. mindspore/include/dataset/execute.h +2 -2
  84. mindspore/jpeg62.dll +0 -0
  85. mindspore/mindrecord/__init__.py +3 -3
  86. mindspore/mindrecord/common/exceptions.py +1 -0
  87. mindspore/mindrecord/config.py +1 -1
  88. mindspore/{parallel/mpi → mindrecord/core}/__init__.py +4 -1
  89. mindspore/mindrecord/{shardheader.py → core/shardheader.py} +2 -1
  90. mindspore/mindrecord/{shardindexgenerator.py → core/shardindexgenerator.py} +1 -1
  91. mindspore/mindrecord/{shardreader.py → core/shardreader.py} +2 -1
  92. mindspore/mindrecord/{shardsegment.py → core/shardsegment.py} +2 -2
  93. mindspore/mindrecord/{shardutils.py → core/shardutils.py} +1 -1
  94. mindspore/mindrecord/{shardwriter.py → core/shardwriter.py} +1 -1
  95. mindspore/mindrecord/filereader.py +4 -4
  96. mindspore/mindrecord/filewriter.py +5 -5
  97. mindspore/mindrecord/mindpage.py +2 -2
  98. mindspore/mindrecord/tools/cifar10.py +4 -3
  99. mindspore/mindrecord/tools/cifar100.py +1 -1
  100. mindspore/mindrecord/tools/cifar100_to_mr.py +1 -1
  101. mindspore/mindrecord/tools/cifar10_to_mr.py +6 -6
  102. mindspore/mindrecord/tools/csv_to_mr.py +1 -1
  103. mindspore/mindrecord/tools/imagenet_to_mr.py +1 -1
  104. mindspore/mindrecord/tools/mnist_to_mr.py +1 -1
  105. mindspore/mindrecord/tools/tfrecord_to_mr.py +1 -1
  106. mindspore/mindspore_backend_common.dll +0 -0
  107. mindspore/mindspore_backend_manager.dll +0 -0
  108. mindspore/mindspore_cluster.dll +0 -0
  109. mindspore/mindspore_common.dll +0 -0
  110. mindspore/mindspore_core.dll +0 -0
  111. mindspore/mindspore_cpu.dll +0 -0
  112. mindspore/mindspore_dump.dll +0 -0
  113. mindspore/mindspore_frontend.dll +0 -0
  114. mindspore/mindspore_glog.dll +0 -0
  115. mindspore/mindspore_hardware_abstract.dll +0 -0
  116. mindspore/mindspore_memory_pool.dll +0 -0
  117. mindspore/mindspore_ms_backend.dll +0 -0
  118. mindspore/mindspore_ops.dll +0 -0
  119. mindspore/{mindspore_ops_host.dll → mindspore_ops_cpu.dll} +0 -0
  120. mindspore/mindspore_profiler.dll +0 -0
  121. mindspore/mindspore_pyboost.dll +0 -0
  122. mindspore/mindspore_pynative.dll +0 -0
  123. mindspore/mindspore_runtime_pipeline.dll +0 -0
  124. mindspore/mindspore_runtime_utils.dll +0 -0
  125. mindspore/mindspore_tools.dll +0 -0
  126. mindspore/mint/__init__.py +15 -10
  127. mindspore/mint/distributed/__init__.py +4 -0
  128. mindspore/mint/distributed/distributed.py +392 -69
  129. mindspore/mint/nn/__init__.py +2 -16
  130. mindspore/mint/nn/functional.py +4 -110
  131. mindspore/mint/nn/layer/__init__.py +0 -2
  132. mindspore/mint/nn/layer/_functions.py +1 -2
  133. mindspore/mint/nn/layer/activation.py +0 -6
  134. mindspore/mint/nn/layer/basic.py +0 -47
  135. mindspore/mint/nn/layer/conv.py +10 -10
  136. mindspore/mint/nn/layer/normalization.py +11 -16
  137. mindspore/mint/nn/layer/pooling.py +0 -4
  138. mindspore/nn/__init__.py +1 -3
  139. mindspore/nn/cell.py +231 -239
  140. mindspore/nn/layer/activation.py +4 -2
  141. mindspore/nn/layer/basic.py +56 -14
  142. mindspore/nn/layer/container.py +16 -0
  143. mindspore/nn/layer/embedding.py +4 -169
  144. mindspore/nn/layer/image.py +1 -1
  145. mindspore/nn/layer/normalization.py +2 -1
  146. mindspore/nn/layer/thor_layer.py +4 -85
  147. mindspore/nn/optim/ada_grad.py +0 -1
  148. mindspore/nn/optim/adafactor.py +0 -1
  149. mindspore/nn/optim/adam.py +32 -127
  150. mindspore/nn/optim/adamax.py +0 -1
  151. mindspore/nn/optim/asgd.py +0 -1
  152. mindspore/nn/optim/ftrl.py +8 -102
  153. mindspore/nn/optim/lamb.py +1 -4
  154. mindspore/nn/optim/lars.py +0 -3
  155. mindspore/nn/optim/lazyadam.py +25 -218
  156. mindspore/nn/optim/momentum.py +5 -43
  157. mindspore/nn/optim/optimizer.py +6 -55
  158. mindspore/nn/optim/proximal_ada_grad.py +0 -1
  159. mindspore/nn/optim/rmsprop.py +0 -1
  160. mindspore/nn/optim/rprop.py +0 -1
  161. mindspore/nn/optim/sgd.py +0 -1
  162. mindspore/nn/optim/tft_wrapper.py +2 -4
  163. mindspore/nn/optim/thor.py +0 -2
  164. mindspore/nn/probability/bijector/bijector.py +7 -8
  165. mindspore/nn/probability/bijector/gumbel_cdf.py +2 -2
  166. mindspore/nn/probability/bijector/power_transform.py +20 -21
  167. mindspore/nn/probability/bijector/scalar_affine.py +5 -5
  168. mindspore/nn/probability/bijector/softplus.py +13 -14
  169. mindspore/nn/probability/distribution/_utils/utils.py +2 -2
  170. mindspore/nn/wrap/cell_wrapper.py +39 -5
  171. mindspore/nn/wrap/grad_reducer.py +4 -89
  172. mindspore/numpy/array_creations.py +4 -4
  173. mindspore/numpy/fft.py +9 -9
  174. mindspore/numpy/utils_const.py +1 -1
  175. mindspore/{nn/reinforcement → onnx}/__init__.py +5 -8
  176. mindspore/onnx/onnx_export.py +137 -0
  177. mindspore/opencv_core4110.dll +0 -0
  178. mindspore/opencv_imgcodecs4110.dll +0 -0
  179. mindspore/{opencv_imgproc452.dll → opencv_imgproc4110.dll} +0 -0
  180. mindspore/ops/__init__.py +2 -0
  181. mindspore/ops/_grad_experimental/grad_comm_ops.py +38 -2
  182. mindspore/ops/_grad_experimental/grad_inner_ops.py +0 -9
  183. mindspore/ops/_op_impl/aicpu/__init__.py +0 -10
  184. mindspore/ops/_op_impl/cpu/__init__.py +1 -5
  185. mindspore/ops/_op_impl/cpu/{buffer_append.py → joinedstr_op.py} +8 -8
  186. mindspore/ops/auto_generate/cpp_create_prim_instance_helper.py +28 -24
  187. mindspore/ops/auto_generate/gen_extend_func.py +6 -11
  188. mindspore/ops/auto_generate/gen_ops_def.py +385 -154
  189. mindspore/ops/auto_generate/gen_ops_prim.py +5676 -5167
  190. mindspore/ops/communication.py +97 -0
  191. mindspore/ops/composite/__init__.py +5 -2
  192. mindspore/ops/composite/base.py +16 -2
  193. mindspore/ops/composite/multitype_ops/__init__.py +3 -1
  194. mindspore/ops/composite/multitype_ops/_compile_utils.py +150 -8
  195. mindspore/ops/composite/multitype_ops/_constexpr_utils.py +1 -1
  196. mindspore/ops/composite/multitype_ops/add_impl.py +7 -0
  197. mindspore/ops/composite/multitype_ops/mod_impl.py +27 -0
  198. mindspore/ops/function/__init__.py +2 -0
  199. mindspore/ops/function/array_func.py +24 -18
  200. mindspore/ops/function/comm_func.py +3883 -0
  201. mindspore/ops/function/debug_func.py +7 -6
  202. mindspore/ops/function/grad/grad_func.py +4 -12
  203. mindspore/ops/function/math_func.py +89 -86
  204. mindspore/ops/function/nn_func.py +92 -313
  205. mindspore/ops/function/random_func.py +9 -18
  206. mindspore/ops/functional.py +4 -1
  207. mindspore/ops/functional_overload.py +377 -30
  208. mindspore/ops/operations/__init__.py +2 -5
  209. mindspore/ops/operations/_custom_ops_utils.py +7 -9
  210. mindspore/ops/operations/_inner_ops.py +12 -50
  211. mindspore/ops/operations/_rl_inner_ops.py +0 -933
  212. mindspore/ops/operations/array_ops.py +5 -50
  213. mindspore/ops/operations/comm_ops.py +95 -17
  214. mindspore/ops/operations/custom_ops.py +237 -22
  215. mindspore/ops/operations/debug_ops.py +33 -35
  216. mindspore/ops/operations/manually_defined/ops_def.py +39 -318
  217. mindspore/ops/operations/math_ops.py +5 -5
  218. mindspore/ops/operations/nn_ops.py +3 -3
  219. mindspore/ops/operations/sparse_ops.py +0 -83
  220. mindspore/ops/primitive.py +4 -27
  221. mindspore/ops/tensor_method.py +88 -10
  222. mindspore/ops_generate/aclnn/aclnn_kernel_register_auto_cc_generator.py +5 -5
  223. mindspore/ops_generate/aclnn/gen_aclnn_implement.py +8 -8
  224. mindspore/ops_generate/api/functions_cc_generator.py +53 -4
  225. mindspore/ops_generate/api/tensor_func_reg_cpp_generator.py +25 -11
  226. mindspore/ops_generate/common/gen_constants.py +11 -10
  227. mindspore/ops_generate/common/op_proto.py +18 -1
  228. mindspore/ops_generate/common/template.py +102 -245
  229. mindspore/ops_generate/common/template_utils.py +212 -0
  230. mindspore/ops_generate/gen_custom_ops.py +69 -0
  231. mindspore/ops_generate/op_def/ops_def_cc_generator.py +78 -7
  232. mindspore/ops_generate/op_def_py/base_op_prim_py_generator.py +360 -0
  233. mindspore/ops_generate/op_def_py/custom_op_prim_py_generator.py +140 -0
  234. mindspore/ops_generate/op_def_py/op_def_py_generator.py +54 -7
  235. mindspore/ops_generate/op_def_py/op_prim_py_generator.py +5 -312
  236. mindspore/ops_generate/pyboost/auto_grad_impl_cc_generator.py +74 -17
  237. mindspore/ops_generate/pyboost/auto_grad_reg_cc_generator.py +22 -5
  238. mindspore/ops_generate/pyboost/gen_pyboost_func.py +0 -16
  239. mindspore/ops_generate/pyboost/op_template_parser.py +3 -2
  240. mindspore/ops_generate/pyboost/pyboost_functions_cpp_generator.py +21 -5
  241. mindspore/ops_generate/pyboost/pyboost_functions_h_generator.py +2 -2
  242. mindspore/ops_generate/pyboost/pyboost_functions_impl_cpp_generator.py +30 -10
  243. mindspore/ops_generate/pyboost/pyboost_grad_function_cpp_generator.py +10 -3
  244. mindspore/ops_generate/pyboost/pyboost_internal_kernel_info_adapter_generator.py +1 -1
  245. mindspore/ops_generate/pyboost/pyboost_native_grad_functions_generator.py +19 -9
  246. mindspore/ops_generate/pyboost/pyboost_op_cpp_code_generator.py +71 -28
  247. mindspore/ops_generate/pyboost/pyboost_overload_functions_cpp_generator.py +10 -9
  248. mindspore/ops_generate/pyboost/pyboost_utils.py +27 -16
  249. mindspore/ops_generate/resources/yaml_loader.py +13 -0
  250. mindspore/ops_generate/tensor_py_cc_generator.py +2 -2
  251. mindspore/parallel/_auto_parallel_context.py +5 -15
  252. mindspore/parallel/_cell_wrapper.py +1 -1
  253. mindspore/parallel/_parallel_serialization.py +4 -6
  254. mindspore/parallel/_ps_context.py +2 -2
  255. mindspore/parallel/_utils.py +34 -17
  256. mindspore/parallel/auto_parallel.py +23 -9
  257. mindspore/parallel/checkpoint_transform.py +20 -2
  258. mindspore/parallel/cluster/process_entity/_api.py +28 -33
  259. mindspore/parallel/cluster/process_entity/_utils.py +9 -5
  260. mindspore/parallel/cluster/run.py +5 -3
  261. mindspore/{experimental/llm_boost/ascend_native → parallel/distributed}/__init__.py +21 -22
  262. mindspore/parallel/distributed/distributed_data_parallel.py +393 -0
  263. mindspore/parallel/distributed/flatten_grad_buffer.py +295 -0
  264. mindspore/parallel/function/reshard_func.py +6 -5
  265. mindspore/parallel/nn/parallel_cell_wrapper.py +40 -3
  266. mindspore/parallel/nn/parallel_grad_reducer.py +0 -8
  267. mindspore/parallel/shard.py +7 -21
  268. mindspore/parallel/strategy.py +336 -0
  269. mindspore/parallel/transform_safetensors.py +127 -20
  270. mindspore/profiler/analysis/viewer/ascend_kernel_details_viewer.py +13 -9
  271. mindspore/profiler/analysis/viewer/ascend_op_memory_viewer.py +1 -1
  272. mindspore/profiler/analysis/viewer/ms_minddata_viewer.py +1 -1
  273. mindspore/profiler/common/constant.py +5 -0
  274. mindspore/profiler/common/file_manager.py +9 -0
  275. mindspore/profiler/common/msprof_cmd_tool.py +40 -4
  276. mindspore/profiler/common/path_manager.py +65 -24
  277. mindspore/profiler/common/profiler_context.py +27 -14
  278. mindspore/profiler/common/profiler_info.py +3 -3
  279. mindspore/profiler/common/profiler_meta_data.py +1 -0
  280. mindspore/profiler/common/profiler_op_analyse.py +10 -6
  281. mindspore/profiler/common/profiler_path_manager.py +13 -0
  282. mindspore/profiler/common/util.py +30 -3
  283. mindspore/profiler/dynamic_profiler.py +91 -46
  284. mindspore/profiler/envprofiler.py +30 -5
  285. mindspore/profiler/experimental_config.py +18 -2
  286. mindspore/profiler/platform/cpu_profiler.py +10 -4
  287. mindspore/profiler/platform/npu_profiler.py +34 -7
  288. mindspore/profiler/profiler.py +193 -145
  289. mindspore/profiler/profiler_action_controller.py +1 -1
  290. mindspore/profiler/profiler_interface.py +2 -2
  291. mindspore/rewrite/symbol_tree/symbol_tree.py +1 -1
  292. mindspore/run_check/_check_version.py +108 -24
  293. mindspore/runtime/__init__.py +9 -6
  294. mindspore/runtime/executor.py +35 -0
  295. mindspore/runtime/memory.py +113 -0
  296. mindspore/runtime/thread_bind_core.py +1 -1
  297. mindspore/swresample-4.dll +0 -0
  298. mindspore/swscale-6.dll +0 -0
  299. mindspore/tinyxml2.dll +0 -0
  300. mindspore/{experimental/llm_boost → tools}/__init__.py +5 -5
  301. mindspore/tools/data_dump.py +130 -0
  302. mindspore/tools/sdc_detect.py +91 -0
  303. mindspore/tools/stress_detect.py +63 -0
  304. mindspore/train/__init__.py +6 -6
  305. mindspore/train/_utils.py +8 -21
  306. mindspore/train/amp.py +6 -7
  307. mindspore/train/callback/_callback.py +2 -1
  308. mindspore/train/callback/_checkpoint.py +1 -17
  309. mindspore/train/callback/_flops_collector.py +10 -6
  310. mindspore/train/callback/_train_fault_tolerance.py +72 -25
  311. mindspore/train/data_sink.py +5 -9
  312. mindspore/train/dataset_helper.py +5 -5
  313. mindspore/train/model.py +41 -230
  314. mindspore/train/serialization.py +160 -401
  315. mindspore/train/train_thor/model_thor.py +2 -2
  316. mindspore/turbojpeg.dll +0 -0
  317. mindspore/utils/__init__.py +6 -3
  318. mindspore/utils/dlpack.py +92 -0
  319. mindspore/utils/dryrun.py +1 -1
  320. mindspore/utils/runtime_execution_order_check.py +10 -0
  321. mindspore/utils/sdc_detect.py +14 -12
  322. mindspore/utils/stress_detect.py +43 -0
  323. mindspore/utils/utils.py +152 -16
  324. mindspore/version.py +1 -1
  325. {mindspore-2.7.0rc1.dist-info → mindspore-2.7.1.dist-info}/METADATA +3 -2
  326. {mindspore-2.7.0rc1.dist-info → mindspore-2.7.1.dist-info}/RECORD +330 -344
  327. mindspore/_extends/remote/kernel_build_server_ascend.py +0 -75
  328. mindspore/communication/_hccl_management.py +0 -297
  329. mindspore/experimental/llm_boost/ascend_native/llama_boost_ascend_native.py +0 -207
  330. mindspore/experimental/llm_boost/ascend_native/llm_boost.py +0 -52
  331. mindspore/experimental/llm_boost/atb/__init__.py +0 -23
  332. mindspore/experimental/llm_boost/atb/boost_base.py +0 -385
  333. mindspore/experimental/llm_boost/atb/llama_boost.py +0 -137
  334. mindspore/experimental/llm_boost/atb/qwen_boost.py +0 -124
  335. mindspore/experimental/llm_boost/register.py +0 -130
  336. mindspore/experimental/llm_boost/utils.py +0 -31
  337. mindspore/include/OWNERS +0 -7
  338. mindspore/mindspore_cpu_res_manager.dll +0 -0
  339. mindspore/mindspore_ops_kernel_common.dll +0 -0
  340. mindspore/mindspore_res_manager.dll +0 -0
  341. mindspore/nn/optim/_dist_optimizer_registry.py +0 -111
  342. mindspore/nn/reinforcement/_batch_read_write.py +0 -142
  343. mindspore/nn/reinforcement/_tensors_queue.py +0 -152
  344. mindspore/nn/reinforcement/tensor_array.py +0 -145
  345. mindspore/opencv_core452.dll +0 -0
  346. mindspore/opencv_imgcodecs452.dll +0 -0
  347. mindspore/ops/_op_impl/aicpu/priority_replay_buffer.py +0 -113
  348. mindspore/ops/_op_impl/aicpu/reservoir_replay_buffer.py +0 -96
  349. mindspore/ops/_op_impl/aicpu/sparse_cross.py +0 -42
  350. mindspore/ops/_op_impl/cpu/buffer_get.py +0 -28
  351. mindspore/ops/_op_impl/cpu/buffer_sample.py +0 -28
  352. mindspore/ops/_op_impl/cpu/priority_replay_buffer.py +0 -42
  353. mindspore/ops/operations/_tensor_array.py +0 -359
  354. mindspore/ops/operations/rl_ops.py +0 -288
  355. mindspore/parallel/_offload_context.py +0 -275
  356. mindspore/parallel/_recovery_context.py +0 -115
  357. mindspore/parallel/_transformer/__init__.py +0 -35
  358. mindspore/parallel/_transformer/layers.py +0 -765
  359. mindspore/parallel/_transformer/loss.py +0 -251
  360. mindspore/parallel/_transformer/moe.py +0 -693
  361. mindspore/parallel/_transformer/op_parallel_config.py +0 -222
  362. mindspore/parallel/_transformer/transformer.py +0 -3124
  363. mindspore/parallel/mpi/_mpi_config.py +0 -116
  364. mindspore/profiler/common/validator/validate_path.py +0 -84
  365. mindspore/train/memory_profiling_pb2.py +0 -298
  366. mindspore/utils/hooks.py +0 -81
  367. /mindspore/common/{_auto_dynamic.py → dynamic_shape/_auto_dynamic.py} +0 -0
  368. {mindspore-2.7.0rc1.dist-info → mindspore-2.7.1.dist-info}/WHEEL +0 -0
  369. {mindspore-2.7.0rc1.dist-info → mindspore-2.7.1.dist-info}/entry_points.txt +0 -0
  370. {mindspore-2.7.0rc1.dist-info → mindspore-2.7.1.dist-info}/top_level.txt +0 -0
@@ -1269,8 +1269,10 @@ class PReLUExt(Cell):
1269
1269
  :align: center
1270
1270
 
1271
1271
  .. note::
1272
- Channel dim is the 2nd dim of input. When input has dims < 2, then there is
1273
- no channel dim and the number of channels = 1.
1272
+ - Channel dim is the 2nd dim of input. When input has dims < 2, then there is
1273
+ no channel dim and the number of channels = 1.
1274
+ - In GE mode, the rank of the input tensor must be greater than 1;
1275
+ otherwise, an error will be triggered.
1274
1276
 
1275
1277
  Args:
1276
1278
  num_parameters (int, optional): number of `w` to learn. Although it takes an int as input,
@@ -38,7 +38,7 @@ from mindspore.common.generator import default_generator
38
38
 
39
39
  __all__ = ['Dropout', 'Flatten', 'Dense', 'Linear', 'ClipByNorm', 'Norm', 'OneHot', 'Pad', 'Unfold', 'Tril', 'Triu',
40
40
  'MatrixDiag', 'MatrixDiagPart', 'MatrixSetDiag', 'L1Regularizer', 'Dropout1d',
41
- 'Dropout2d', 'Dropout3d', 'Upsample', 'Roll', 'Identity', 'Unflatten', 'DropoutExt']
41
+ 'Dropout2d', 'Dropout3d', 'Upsample', 'Roll', 'Identity', 'Unflatten', 'DropoutExt', 'Dropout2dExt']
42
42
 
43
43
 
44
44
  class L1Regularizer(Cell):
@@ -209,7 +209,7 @@ class DropoutExt(Cell):
209
209
  Dropout is a means of regularization that reduces overfitting by preventing correlations between neuronal nodes.
210
210
  The operator randomly sets some neurons output to 0 according to `p`, which means the probability of discarding
211
211
  during training. And the return will be multiplied by :math:`\frac{1}{1-p}` during training.
212
- During the reasoning, this layer returns the same Tensor as the `x`.
212
+ During the reasoning, this layer returns the same Tensor as the `input`.
213
213
 
214
214
  This technique is proposed in paper `Dropout: A Simple Way to Prevent Neural Networks from Overfitting
215
215
  <http://www.cs.toronto.edu/~rsalakhu/papers/srivastava14a.pdf>`_ and proved to be effective to reduce
@@ -228,14 +228,13 @@ class DropoutExt(Cell):
228
228
  If set to ``True`` , will do this operation in-place. Default: ``False`` .
229
229
 
230
230
  Inputs:
231
- - **x** (Tensor) - The input of Dropout.
231
+ - **input** (Tensor) - The input of Dropout.
232
232
 
233
233
  Outputs:
234
- Tensor, output tensor with the same shape as the `x`.
234
+ Tensor, output tensor with the same shape as the `input`.
235
235
 
236
236
  Raises:
237
- TypeError: If the dtype of `p` is not float.
238
- ValueError: If length of shape of `x` is less than 1.
237
+ TypeError: If the dtype of `inplace` is not bool.
239
238
 
240
239
  Supported Platforms:
241
240
  ``Ascend``
@@ -259,17 +258,12 @@ class DropoutExt(Cell):
259
258
  self.inplace = inplace
260
259
  self.generator_step = Tensor(12, mstype.int64)
261
260
 
262
- def construct(self, x):
261
+ def construct(self, input):
263
262
  if not self.training or self.p == 0:
264
- return x
263
+ return input
265
264
 
266
265
  seed, offset = default_generator._step(self.generator_step) # pylint: disable=protected-access
267
- out, _ = ops.auto_generate.dropout_ext_op(x, self.p, seed, offset)
268
-
269
- if self.inplace:
270
- x.copy_(out)
271
- return x
272
- return out
266
+ return ops.auto_generate.func_dropout_ext_op(input, self.p, self.training, self.inplace, seed, offset)
273
267
 
274
268
 
275
269
  class Dropout1d(Cell):
@@ -388,6 +382,54 @@ class Dropout2d(Cell):
388
382
  return f"p={self.keep_prob}"
389
383
 
390
384
 
385
+
386
+ class Dropout2dExt(Cell):
387
+ r"""
388
+ During training, randomly zeroes some channels of the input tensor with probability `p`
389
+ from a Bernoulli distribution (For a 4-dimensional tensor with a shape of :math:`NCHW`,
390
+ the channel feature map refers to a 2-dimensional feature map with the shape of :math:`HW`).
391
+
392
+ For example, the :math:`j\_th` channel of the :math:`i\_th` sample in the batched input is a to-be-processed
393
+ `2D` tensor input[i,j].
394
+ Each channel will be zeroed out independently on every forward call with probability `p` using samples
395
+ from a Bernoulli distribution.
396
+
397
+ `Dropout2d` can improve the independence between channel feature maps.
398
+
399
+ .. warning::
400
+ This is an experimental API that is subject to change or deletion.
401
+
402
+ Refer to :func:`mindspore.mint.nn.functional.dropout2d` for more details.
403
+
404
+ Supported Platforms:
405
+ ``Ascend``
406
+
407
+ Examples:
408
+ >>> import mindspore
409
+ >>> from mindspore import Tensor, mint
410
+ >>> import numpy as np
411
+ >>> dropout = mint.nn.Dropout2d(p=0.5)
412
+ >>> x = Tensor(np.ones([2, 1, 2, 3]), mindspore.float32)
413
+ >>> output = dropout(x)
414
+ >>> print(output.shape)
415
+ (2, 1, 2, 3)
416
+ """
417
+
418
+ def __init__(self, p=0.5, inplace=False):
419
+ """Initialize Dropout2d."""
420
+ super(Dropout2dExt, self).__init__()
421
+ self.p = p
422
+ self.inplace = inplace
423
+ self.generator_step = Tensor(12, mstype.int64)
424
+
425
+ def construct(self, input):
426
+ if not self.training or self.p == 0:
427
+ return input
428
+
429
+ seed, offset = default_generator._step(self.generator_step) # pylint: disable=protected-access
430
+ return ops.auto_generate.dropout2d_ext_op(input, self.p, self.training, self.inplace, seed, offset)
431
+
432
+
391
433
  class Dropout3d(Cell):
392
434
  r"""
393
435
  During training, randomly zeroes some channels of the input tensor
@@ -247,6 +247,14 @@ class SequentialCell(Cell):
247
247
  self._cells = temp_dict
248
248
  self.cell_list = list(self._cells.values())
249
249
 
250
+ def __setattr__(self, name, value):
251
+ from mindspore.common.recompute import _RecomputeCell
252
+ if isinstance(value, _RecomputeCell):
253
+ # RecomputeCell should not be executed
254
+ object.__setattr__(self, name, value)
255
+ else:
256
+ super().__setattr__(name, value)
257
+
250
258
  def __bool__(self):
251
259
  return len(self._cells) != 0
252
260
 
@@ -594,6 +602,14 @@ class CellDict(_CellDictBase, Cell):
594
602
  def __delitem__(self, key):
595
603
  del self._cells[key]
596
604
 
605
+ def __setattr__(self, name, value):
606
+ from mindspore.common.recompute import _RecomputeCell
607
+ if isinstance(value, _RecomputeCell):
608
+ # RecomputeCell should not be executed
609
+ object.__setattr__(self, name, value)
610
+ else:
611
+ super().__setattr__(name, value)
612
+
597
613
  def __len__(self):
598
614
  return len(self._cells)
599
615
 
@@ -17,17 +17,12 @@ from __future__ import absolute_import
17
17
 
18
18
  import mindspore.common.dtype as mstype
19
19
  import mindspore.ops as ops
20
- from mindspore import log as logger
21
20
  from mindspore.common.tensor import Tensor
22
21
  from mindspore.common.parameter import Parameter
23
- from mindspore.common.parameter import _get_unique_parameter_key
24
22
  from mindspore.common.initializer import initializer, Normal
25
- from mindspore.communication.management import get_group_size, get_rank
23
+ from mindspore.communication.management import get_group_size
26
24
  from mindspore.context import ParallelMode
27
25
  from mindspore.parallel._utils import _get_parallel_mode, _get_full_batch
28
- from mindspore.parallel._ps_context import _get_ps_context, _enable_distributed_mindrt
29
- from mindspore.parallel._ps_context import _is_role_worker, _is_role_pserver
30
- from mindspore.parallel._ps_context import _insert_hash_table_size, _set_cache_enable, _set_rank_id
31
26
  from mindspore import _checkparam as Validator
32
27
  from mindspore.ops.primitive import constexpr, _primexpr
33
28
  from mindspore.nn.layer.basic import ClipByNorm
@@ -341,10 +336,6 @@ class EmbeddingLookup(Cell):
341
336
  max_norm (Union[float, None]): A maximum clipping value. The data type must be float16, float32
342
337
  or None. Default: ``None`` .
343
338
  sparse (bool): Using sparse mode. When 'target' is set to 'CPU', 'sparse' has to be true. Default: ``True`` .
344
- vocab_cache_size (int): Cache size of the dictionary of embeddings. Default: ``0`` . It is valid only in
345
- parameter server trainning mode and 'DEVICE' target. And the moment parameter of corresponding
346
- optimizer will also be set to the cache size. In addition, it should be noted that it will cost the 'DEVICE'
347
- memory, so suggests setting a reasonable value to avoid insufficient memory.
348
339
  dtype (:class:`mindspore.dtype`): Dtype of Parameters. Default: ``mstype.float32`` .
349
340
 
350
341
  Inputs:
@@ -358,10 +349,9 @@ class EmbeddingLookup(Cell):
358
349
  Tensor, the shape of tensor is :math:`(z_1, z_2, ..., z_N)`.
359
350
 
360
351
  Raises:
361
- TypeError: If `vocab_size` or `embedding_size` or `vocab_cache_size` is not an int.
352
+ TypeError: If `vocab_size` or `embedding_size` is not an int.
362
353
  TypeError: If `sparse` is not a bool or `manual_shapes` is not a tuple.
363
354
  ValueError: If `vocab_size` or `embedding_size` is less than 1.
364
- ValueError: If `vocab_cache_size` is less than 0.
365
355
  ValueError: If `target` is neither 'CPU' nor 'DEVICE'.
366
356
  ValueError: If `slice_mode` is not one of 'batch_slice' or 'field_slice' or
367
357
  'table_row_slice' or 'table_column_slice'.
@@ -387,17 +377,14 @@ class EmbeddingLookup(Cell):
387
377
 
388
378
  def __init__(self, vocab_size, embedding_size, param_init='normal',
389
379
  target='CPU', slice_mode='batch_slice', manual_shapes=None,
390
- max_norm=None, sparse=True, vocab_cache_size=0, dtype=mstype.float32):
380
+ max_norm=None, sparse=True, dtype=mstype.float32):
391
381
  """Initialize EmbeddingLookup."""
392
382
  super(EmbeddingLookup, self).__init__()
393
383
  Validator.check_value_type('sparse', sparse, [bool], self.cls_name)
394
384
  self.vocab_size = Validator.check_positive_int(
395
385
  vocab_size, 'vocab_size')
396
- self.vocab_cache_size = Validator.check_non_negative_int(
397
- vocab_cache_size, 'vocab_cache_size')
398
386
  self.target = target
399
387
  self.sparse = sparse
400
- self.cache_enable = self.vocab_cache_size > 0
401
388
  self.forward_unique = False
402
389
  Validator.check_string(
403
390
  target, ['CPU', 'DEVICE'], 'target', self.cls_name)
@@ -409,10 +396,6 @@ class EmbeddingLookup(Cell):
409
396
  else:
410
397
  self.gatherv2 = ops.Gather()
411
398
  self.embeddinglookup = ops.EmbeddingLookup().set_device('CPU')
412
- self.is_ps_server = False
413
- enable_ps = _get_ps_context("enable_ps")
414
- if enable_ps:
415
- self._process_vocab_cache(slice_mode)
416
399
  self.embedding_size = Validator.check_positive_int(
417
400
  embedding_size, 'embedding_size', self.cls_name)
418
401
  self.embedding_table = Parameter(initializer(param_init, [self.vocab_size, self.embedding_size],
@@ -427,11 +410,6 @@ class EmbeddingLookup(Cell):
427
410
  self.shape = ops.Shape()
428
411
  if is_auto_parallel:
429
412
  self.unique = ops.Unique().shard(((1,),))
430
- if self.cache_enable and enable_ps:
431
- self._set_voacb_cache_enable_for_ps(
432
- vocab_cache_size, embedding_size, vocab_size, param_init, dtype=dtype)
433
- if is_auto_parallel:
434
- self.unique.add_prim_attr('cache_enable', True)
435
413
  indices_shape_size = 2
436
414
  if slice_mode == "field_slice" and is_auto_parallel:
437
415
  if not manual_shapes:
@@ -450,7 +428,7 @@ class EmbeddingLookup(Cell):
450
428
  ((get_group_size(), 1), (1, get_group_size())))
451
429
  elif slice_mode == "table_row_slice" and is_auto_parallel:
452
430
  full_batch = _get_full_batch()
453
- if (target == 'DEVICE' and not full_batch) or (self.cache_enable and enable_ps and sparse):
431
+ if (target == 'DEVICE' and not full_batch):
454
432
  indices_shape_size = 1
455
433
  self.gather_revert.shard(((1, 1), (get_group_size(),)))
456
434
  self.forward_unique = True
@@ -479,9 +457,6 @@ class EmbeddingLookup(Cell):
479
457
  "table_column_slice", "batch_slice"]
480
458
  raise ValueError(f"For '{self.cls_name}', the 'slice_mode' must be in {support_mode}, "
481
459
  f"but got \"{slice_mode}\".")
482
- if self.cache_enable and not enable_ps:
483
- raise ValueError(
484
- f"For '{self.cls_name}', haven't supported cache enable for not ps mode.")
485
460
  self.embedding_table.unique = self.forward_unique
486
461
  self.max_norm = max_norm
487
462
  if self.max_norm is not None:
@@ -489,149 +464,9 @@ class EmbeddingLookup(Cell):
489
464
  self.max_norm, 'max_norm', self.cls_name)
490
465
  self.max_norm = Tensor(self.max_norm, dtype=mstype.float32)
491
466
 
492
- def _process_vocab_cache(self, slice_mode):
493
- """PS embeddingLookup cache check and process."""
494
- self.cache_enable = False
495
- if self.vocab_cache_size > 0:
496
- if self.target == 'CPU':
497
- logger.warning("The configuration of 'vocab_cache_size' is valid only in 'DEVICE' target, "
498
- "current target is CPU, so it will be ignored.")
499
- return
500
- enable_ps = _get_ps_context("enable_ps")
501
- if not enable_ps:
502
- logger.warning("The configuration of 'vocab_cache_size' is valid only in parameter server training "
503
- "mode, current mode is not parameter server trainning mode, so it will be ignored.")
504
- return
505
- self.is_ps_server = _is_role_pserver() and _enable_distributed_mindrt()
506
- parallel_mode = _get_parallel_mode()
507
- is_auto_parallel = parallel_mode in (
508
- ParallelMode.SEMI_AUTO_PARALLEL, ParallelMode.AUTO_PARALLEL)
509
- if is_auto_parallel:
510
- rank_size = get_group_size()
511
- rank_id = get_rank()
512
- full_batch = _get_full_batch()
513
- if rank_size > 1 and not (full_batch and slice_mode == "table_row_slice"):
514
- raise ValueError(f"For '{self.cls_name}', the cache of parameter server parallel should only be "
515
- f"used in \"full_batch\" and the value of \"full_batch\" must be True. "
516
- f"Meanwhile, the value of 'slice_mode' must be \"table_row_slice\"."
517
- f"But got full_batch: {full_batch} and 'slice_mode': \"{slice_mode}\".")
518
- self.vocab_cache_size = self.vocab_cache_size * rank_size
519
- _set_rank_id(rank_id)
520
-
521
- self.cache_enable = True
522
- _set_cache_enable(True)
523
-
524
- if _is_role_worker():
525
- self.vocab_size = self.vocab_cache_size
526
-
527
- def _set_voacb_cache_enable_for_ps(self, vocab_cache_size, embedding_size, vocab_size, param_init,
528
- dtype=mstype.float32):
529
- """PS embeddingLookup cache enable set."""
530
- if self.sparse:
531
- self.forward_unique = True
532
- param_key = _get_unique_parameter_key()
533
- if _is_role_worker():
534
- self.embedding_table.is_param_ps = True
535
- self.embedding_table.cache_enable = True
536
- self.embedding_table.key = param_key
537
- _insert_hash_table_size(
538
- self.embedding_table.name, vocab_cache_size, embedding_size, vocab_size, param_key)
539
-
540
- if _enable_distributed_mindrt():
541
- self.rank_id = get_rank()
542
- if self.is_ps_server:
543
- self._slice_pserver_embeddings("zeros", dtype=dtype)
544
- self._set_cache_enable_and_key_for_pserver(param_key)
545
-
546
- def _slice_pserver_embeddings(self, param_init, dtype=mstype.float32):
547
- '''
548
- Method to slice embedding tables on Parameter Servers.
549
- It helps to train with a large scale embedding table and is used only in Parameter Server training mode.
550
- So EmbeddingLookup op is on CPU device.
551
- '''
552
- self.embedding_lookup_list = []
553
- # The dimension of each embedding table on servers could be different according to the slicing algorithm.
554
- self.embedding_table_vocab_dim_list = []
555
- self.embedding_table_list = []
556
- # For different servers, the offset of their embedding table should be different.
557
- self.embedding_offset = []
558
-
559
- server_num = _get_ps_context("server_num")
560
- if server_num == 0:
561
- raise ValueError("The Parameter Server number is zero.")
562
- # Assign the embedding table dimensions.
563
- for _ in range(server_num):
564
- self.embedding_table_vocab_dim_list.append(
565
- self.vocab_size // server_num)
566
- rest_vocab_size = self.vocab_size % server_num
567
- if rest_vocab_size != 0:
568
- for i in range(rest_vocab_size):
569
- self.embedding_table_vocab_dim_list[i] += 1
570
-
571
- offset = 0
572
- for i in range(server_num):
573
- self.embedding_table_list.append(Parameter(initializer(param_init,
574
- [self.embedding_table_vocab_dim_list[i],
575
- self.embedding_size], dtype=dtype),
576
- name="embedding_table_server_" + str(i)))
577
-
578
- self.embedding_offset.append(offset)
579
- offset += self.embedding_table_vocab_dim_list[i]
580
-
581
- # Add EmbeddingLookup ops on different servers.
582
- if self.target == 'CPU':
583
- embedding_lookup = ops.EmbeddingLookup().set_device('CPU')
584
- else:
585
- if self.sparse:
586
- embedding_lookup = ops.SparseGatherV2()
587
- else:
588
- embedding_lookup = ops.Gather()
589
- embedding_lookup.add_prim_attr(
590
- 'offset', self.embedding_offset[i])
591
- embedding_lookup.add_prim_attr('rank_id', i)
592
- embedding_lookup.add_prim_attr('ms_role', 'MS_PSERVER')
593
- self.embedding_lookup_list.append(embedding_lookup)
594
-
595
- # For now unique operation is not applied,
596
- # so we need to reduce the lookup results from different servers with AddN.
597
- self.reduce_lookup_result = ops.AddN()
598
-
599
- def _do_server_embedding_lookup(self, indices):
600
- '''
601
- Construct backbone for EmbeddingLookup operators on servers.
602
- '''
603
- result_from_servers = []
604
- for i in range(_get_ps_context("server_num")):
605
- result = self.embedding_lookup_list[i](self.embedding_table_list[i],
606
- indices, self.embedding_offset[i])
607
- result_from_servers.append(result)
608
- final_result = self.reduce_lookup_result(result_from_servers)
609
- return final_result
610
-
611
- def _set_cache_enable_and_key_for_pserver(self, param_key):
612
- '''
613
- Set cache enable and parameter key for embedding table on parameter servers.
614
- '''
615
- # Parameter The Embedding Table on the Server side will be divided according to the number of servers.
616
- # The divided Embedding Table will be used instead of the complete Embedding Table.
617
- self.embedding_table = self.embedding_table_list[self.rank_id]
618
- self.embedding_table.cache_enable = True
619
- self.embedding_table.key = param_key
620
-
621
- def _pserver_embedding_lookup(self, indices):
622
- '''
623
- Construct backbone for EmbeddingLookup operators on servers for embedding cache lookup.
624
- '''
625
- if self.target == 'CPU':
626
- return self.embedding_lookup_list[self.rank_id](self.embedding_table, indices,
627
- self.embedding_offset[self.rank_id])
628
- return self.embedding_lookup_list[self.rank_id](self.embedding_table, indices, 0)
629
-
630
467
  def construct(self, indices):
631
468
  if self.target == "CPU":
632
469
  out = self.embeddinglookup(self.embedding_table, indices, 0)
633
- elif self.is_ps_server:
634
- out = self._pserver_embedding_lookup(indices)
635
470
  else:
636
471
  if self.forward_unique:
637
472
  shp = self.shape(indices) + (self.embedding_size,)
@@ -111,7 +111,7 @@ def _convert_img_dtype_to_float32(img, max_val):
111
111
  @constexpr
112
112
  def _get_dtype_max(dtype):
113
113
  """get max of the dtype"""
114
- np_type = mstype.dtype_to_nptype(dtype)
114
+ np_type = mstype._dtype_to_nptype(dtype) # pylint:disable=protected-access
115
115
  if issubclass(np_type, numbers.Integral):
116
116
  dtype_max = np.float64(np.iinfo(np_type).max).item()
117
117
  else:
@@ -21,6 +21,7 @@ import numbers
21
21
  import hashlib
22
22
  import numpy as np
23
23
  import mindspore.ops as ops
24
+ from mindspore.ops import operations as P
24
25
  from mindspore.ops.operations import _inner_ops as inner
25
26
  from mindspore.common.parameter import Parameter
26
27
  from mindspore.common.initializer import initializer, Initializer
@@ -917,7 +918,7 @@ class _InstanceNorm(Cell):
917
918
 
918
919
  self.shape = ops.Shape()
919
920
  self.momentum = momentum
920
- self.instance_bn = ops.InstanceNorm(epsilon=self.eps, momentum=self.momentum)
921
+ self.instance_bn = P.InstanceNorm(epsilon=self.eps, momentum=self.momentum)
921
922
 
922
923
  def construct(self, x):
923
924
  self._check_input_dim(self.shape(x), self.cls_name)
@@ -22,7 +22,7 @@ import mindspore.common.dtype as mstype
22
22
  import mindspore.log as logger
23
23
  from mindspore.common.tensor import Tensor
24
24
  from mindspore.common.initializer import initializer, Initializer
25
- from mindspore.communication.management import get_group_size, get_rank
25
+ from mindspore.communication.management import get_group_size
26
26
  from mindspore.ops.operations._thor_ops import ThorIm2Col
27
27
  from mindspore.common.parameter import Parameter
28
28
  from mindspore import _checkparam as Validator
@@ -30,8 +30,6 @@ from mindspore._checkparam import twice
30
30
  from mindspore import context
31
31
  from mindspore.nn.cell import Cell
32
32
  from mindspore.nn.layer.activation import get_activation
33
- from mindspore.parallel._ps_context import _is_role_worker, _get_ps_context, \
34
- _set_rank_id, _insert_hash_table_size, _set_cache_enable
35
33
  from mindspore.parallel._utils import _get_parallel_mode, _get_full_batch
36
34
  from mindspore.context import ParallelMode
37
35
  from mindspore.nn.layer.basic import ClipByNorm
@@ -695,10 +693,6 @@ class EmbeddingLookupThor(Cell):
695
693
  Default: ``None`` .
696
694
  sparse (bool): Using sparse mode. When 'target' is set to 'CPU', 'sparse' has to be ``true`` .
697
695
  Default: ``True`` .
698
- vocab_cache_size (int): Cache size of the dictionary of embeddings. Default: ``0`` . It is valid only in
699
- 'DEVICE' target. And the moment parameter of corresponding optimizer will also be set to the cache size.
700
- In addition, it should be noted that it will cost the 'DEVICE' memory, so suggests setting a reasonable
701
- value to avoid insufficient memory.
702
696
 
703
697
  Inputs:
704
698
  - **input_indices** (Tensor) - The shape of tensor is :math:`(y_1, y_2, ..., y_S)`.
@@ -712,10 +706,9 @@ class EmbeddingLookupThor(Cell):
712
706
  'table_row_slice' or 'table_column_slice'.
713
707
  ValueError: If `sparse` is False and `target` is 'CPU'.
714
708
  ValueError: If `slice_mode` is 'field_slice' and `manual_shapes` is None.
715
- TypeError: If `vocab_size` or `embedding_size` or `vocab_cache_size` is not an int.
709
+ TypeError: If `vocab_size` or `embedding_size` is not an int.
716
710
  TypeError: If `sparse` is not a bool or `manual_shapes` is not a tuple.
717
711
  ValueError: If `vocab_size` or `embedding_size` is less than 1.
718
- ValueError: If `vocab_cache_size` is less than 0.
719
712
 
720
713
 
721
714
  Supported Platforms:
@@ -736,14 +729,12 @@ class EmbeddingLookupThor(Cell):
736
729
 
737
730
  def __init__(self, vocab_size, embedding_size, param_init='normal',
738
731
  target='CPU', slice_mode='batch_slice', manual_shapes=None,
739
- max_norm=None, sparse=True, vocab_cache_size=0):
732
+ max_norm=None, sparse=True):
740
733
  super(EmbeddingLookupThor, self).__init__()
741
734
  Validator.check_value_type('sparse', sparse, [bool], self.cls_name)
742
735
  self.vocab_size = Validator.check_positive_int(vocab_size, 'vocab_size', self.cls_name)
743
- self.vocab_cache_size = Validator.check_non_negative_int(vocab_cache_size, 'vocab_cache_size', self.cls_name)
744
736
  self.target = target
745
737
  self.sparse = sparse
746
- self.cache_enable = self.vocab_cache_size > 0
747
738
  self.forward_unique = False
748
739
  self.dtype = mstype.float16
749
740
  if target not in ('CPU', 'DEVICE'):
@@ -757,9 +748,6 @@ class EmbeddingLookupThor(Cell):
757
748
  else:
758
749
  self.gatherv2 = ops.Gather()
759
750
  self.embeddinglookup = ops.EmbeddingLookup().set_device('CPU')
760
- enable_ps = _get_ps_context("enable_ps")
761
- if enable_ps:
762
- self._process_vocab_cache(slice_mode)
763
751
  self.embedding_size = Validator.check_positive_int(embedding_size, 'embedding_size', self.cls_name)
764
752
  self.embedding_table = Parameter(initializer(param_init, [self.vocab_size, self.embedding_size],
765
753
  mstype.float16), name='embedding_table')
@@ -772,10 +760,6 @@ class EmbeddingLookupThor(Cell):
772
760
  self.shape = ops.Shape()
773
761
  if is_auto_parallel:
774
762
  self.unique = ops.Unique().shard(((1,),))
775
- if self.cache_enable and enable_ps:
776
- self._set_voacb_cache_enable_for_ps(vocab_cache_size, embedding_size, vocab_size)
777
- if is_auto_parallel:
778
- self.unique.add_prim_attr('cache_enable', True)
779
763
  indices_shape_size = 2
780
764
  if slice_mode == "field_slice" and is_auto_parallel:
781
765
  if not manual_shapes:
@@ -792,7 +776,7 @@ class EmbeddingLookupThor(Cell):
792
776
  self.embeddinglookup.shard(((get_group_size(), 1), (1, get_group_size())))
793
777
  elif slice_mode == "table_row_slice" and is_auto_parallel:
794
778
  full_batch = _get_full_batch()
795
- if (target == 'DEVICE' and not full_batch) or (self.cache_enable and enable_ps and sparse):
779
+ if (target == 'DEVICE' and not full_batch):
796
780
  indices_shape_size = 1
797
781
  self.gather_revert.shard(((1, 1), (get_group_size(),)))
798
782
  self.forward_unique = True
@@ -818,11 +802,6 @@ class EmbeddingLookupThor(Cell):
818
802
  raise ValueError(f"For '{self.cls_name}', the 'slice_mode' must be one of values in "
819
803
  f"['field_slice', 'table_row_slice', 'table_column_slice', 'batch_slice'], "
820
804
  f"but got 'slice_mode': {slice_mode}")
821
- if self.cache_enable and not enable_ps:
822
- if parallel_mode != ParallelMode.STAND_ALONE:
823
- raise ValueError(f"For '{self.cls_name}', the 'parallel_mode' must be equal to "
824
- f"'ParallelMode.STAND_ALONE', but got {parallel_mode}.")
825
- self._set_cache_enable()
826
805
  self.embedding_table.unique = self.forward_unique
827
806
  self.max_norm = max_norm
828
807
  if self.max_norm is not None:
@@ -859,66 +838,6 @@ class EmbeddingLookupThor(Cell):
859
838
  self.matrix_g = matrix_g
860
839
  return out
861
840
 
862
- def _set_cache_enable(self):
863
- """EmbeddingLookup cache check for not ps env, which is only support 'ascend'."""
864
- if self.target != 'DEVICE':
865
- raise ValueError(f"For '{self.cls_name}', the configuration of 'vocab_cache_size' is valid "
866
- f"only when 'target' is 'DEVICE', but got 'target': {self.target}.")
867
- if not self.sparse:
868
- raise ValueError(f"For '{self.cls_name}', the configuration of 'vocab_cache_size' is valid "
869
- f"only when 'sparse' is true, but got 'sparse': {self.sparse}.")
870
- if context.get_context("device_target") != 'Ascend':
871
- raise ValueError(f"For '{self.cls_name}', the configuration of 'vocab_cache_size' is valid "
872
- f"only when 'device_target' is 'Ascend', but got {context.get_context('device_target')}.")
873
-
874
- logger.info("EmbeddingLookup cache enable takes effect.")
875
- self.forward_unique = True
876
- self.unique = ops.Unique().set_device('CPU')
877
- self.unique.add_prim_attr('cache_enable', True)
878
- self.embedding_table.cache_enable = self.cache_enable
879
- self.embedding_table.cache_shape = (self.vocab_cache_size, self.embedding_size)
880
- self.reshape_first = ops.Reshape().set_device('CPU')
881
-
882
- def _process_vocab_cache(self, slice_mode):
883
- """PS embeddingLookup cache check and process."""
884
- self.cache_enable = False
885
- if self.vocab_cache_size > 0:
886
- if self.target == 'CPU':
887
- logger.warning("The configuration of 'vocab_cache_size' is valid only in 'DEVICE' target, "
888
- "current target is CPU, so it will be ignored.")
889
- return
890
- enable_ps = _get_ps_context("enable_ps")
891
- if not enable_ps:
892
- logger.warning(
893
- "The configuration of 'vocab_cache_size' is valid only in parameter server trainning "
894
- "mode, current mode is not parameter server trainning mode, so it will be ignored.")
895
- return
896
- parallel_mode = _get_parallel_mode()
897
- is_auto_parallel = parallel_mode in (ParallelMode.SEMI_AUTO_PARALLEL, ParallelMode.AUTO_PARALLEL)
898
- if is_auto_parallel:
899
- rank_size = get_group_size()
900
- rank_id = get_rank()
901
- full_batch = _get_full_batch()
902
- if rank_size > 1 and not (full_batch and slice_mode == "table_row_slice"):
903
- raise ValueError(f"For '{self.cls_name}', the embeddingLookup cache of parameter server parallel "
904
- f"only be used in 'full_batch' and 'table_row_slice' parallel strategy, but got "
905
- f"'full_batch': {full_batch}, 'slice_mode': {slice_mode}.")
906
- self.vocab_cache_size = self.vocab_cache_size * rank_size
907
- _set_rank_id(rank_id)
908
- self.cache_enable = True
909
- if _is_role_worker():
910
- self.vocab_size = self.vocab_cache_size
911
-
912
- def _set_voacb_cache_enable_for_ps(self, vocab_cache_size, embedding_size, vocab_size):
913
- """PS embeddingLookup cache enable set."""
914
- self.embedding_table.cache_enable = True
915
- self.embedding_table.is_param_ps = True
916
- _set_cache_enable(True)
917
- if self.sparse:
918
- self.forward_unique = True
919
- if _is_role_worker():
920
- _insert_hash_table_size(self.embedding_table.name, vocab_cache_size, embedding_size, vocab_size)
921
-
922
841
  def construct(self, indices):
923
842
  if self.target == "CPU":
924
843
  out = self.embeddinglookup(self.embedding_table, indices, 0)
@@ -204,7 +204,6 @@ class Adagrad(Optimizer):
204
204
  def construct(self, grads):
205
205
  params = self._parameters
206
206
  accum = self.accum
207
- grads = self.flatten_gradients(grads)
208
207
  grads = self.decay_weight(grads)
209
208
  grads = self.gradients_centralization(grads)
210
209
  grads = self.scale_grad(grads)
@@ -408,7 +408,6 @@ class AdaFactor(Optimizer):
408
408
 
409
409
  @jit(backend="ms_backend")
410
410
  def construct(self, gradients):
411
- gradients = self.flatten_gradients(gradients)
412
411
  lr = self.get_lr()
413
412
  self.assignadd(self.global_step, self.global_step_increase_tensor)
414
413
  step = F.assign_add(self.step, 1)