mindspore 2.5.0__cp39-cp39-win_amd64.whl → 2.6.0rc1__cp39-cp39-win_amd64.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of mindspore might be problematic. Click here for more details.

Files changed (491) hide show
  1. mindspore/.commit_id +1 -1
  2. mindspore/Microsoft.VisualStudio.Telemetry.dll +0 -0
  3. mindspore/Newtonsoft.Json.dll +0 -0
  4. mindspore/__init__.py +6 -4
  5. mindspore/_c_dataengine.cp39-win_amd64.pyd +0 -0
  6. mindspore/_c_expression.cp39-win_amd64.pyd +0 -0
  7. mindspore/_c_mindrecord.cp39-win_amd64.pyd +0 -0
  8. mindspore/_check_jit_forbidden_api.py +3 -0
  9. mindspore/_checkparam.py +3 -33
  10. mindspore/_deprecated/__init__.py +17 -0
  11. mindspore/_deprecated/jit.py +198 -0
  12. mindspore/_extends/builtin_operations.py +1 -1
  13. mindspore/_extends/parse/__init__.py +6 -7
  14. mindspore/_extends/parse/compile_config.py +19 -0
  15. mindspore/_extends/parse/deprecated/deprecated_tensor_method.py +22 -3
  16. mindspore/_extends/parse/jit_fallback_modules/__init__.py +0 -0
  17. mindspore/_extends/parse/jit_fallback_modules/check_utils.py +123 -0
  18. mindspore/_extends/parse/jit_fallback_modules/third_party_modules.py +50 -0
  19. mindspore/_extends/parse/parser.py +24 -193
  20. mindspore/_extends/parse/resources.py +1 -5
  21. mindspore/_extends/parse/standard_method.py +97 -74
  22. mindspore/_extends/pijit/__init__.py +2 -2
  23. mindspore/_extends/pijit/pijit_func_white_list.py +16 -11
  24. mindspore/_extends/pijit/tensor_func_list.py +27 -0
  25. mindspore/_extends/utils.py +1 -1
  26. mindspore/amp.py +4 -4
  27. mindspore/atlprov.dll +0 -0
  28. mindspore/avcodec-59.dll +0 -0
  29. mindspore/avdevice-59.dll +0 -0
  30. mindspore/avfilter-8.dll +0 -0
  31. mindspore/avformat-59.dll +0 -0
  32. mindspore/avutil-57.dll +0 -0
  33. mindspore/boost/__init__.py +2 -2
  34. mindspore/boost/base.py +3 -7
  35. mindspore/boost/boost_cell_wrapper.py +2 -2
  36. mindspore/c1.dll +0 -0
  37. mindspore/c1xx.dll +0 -0
  38. mindspore/c2.dll +0 -0
  39. mindspore/common/__init__.py +4 -3
  40. mindspore/common/_grad_function.py +56 -0
  41. mindspore/common/_pijit_context.py +14 -5
  42. mindspore/common/_register_for_tensor.py +1 -1
  43. mindspore/common/_stub_tensor.py +5 -10
  44. mindspore/common/_tensor_cpp_method.py +1 -1
  45. mindspore/common/_tensor_docs.py +1915 -3287
  46. mindspore/common/api.py +341 -354
  47. mindspore/common/auto_dynamic_shape.py +41 -44
  48. mindspore/common/dtype.py +5 -2
  49. mindspore/common/dump.py +7 -5
  50. mindspore/common/file_system.py +3 -0
  51. mindspore/common/hook_handle.py +5 -3
  52. mindspore/common/initializer.py +10 -6
  53. mindspore/common/jit_begin_end.py +94 -0
  54. mindspore/common/jit_config.py +6 -1
  55. mindspore/common/jit_context.py +76 -0
  56. mindspore/common/jit_trace.py +378 -0
  57. mindspore/common/lazy_inline.py +2 -2
  58. mindspore/common/mutable.py +5 -4
  59. mindspore/common/parameter.py +106 -39
  60. mindspore/common/seed.py +2 -2
  61. mindspore/common/sparse_tensor.py +23 -17
  62. mindspore/common/tensor.py +297 -714
  63. mindspore/communication/__init__.py +7 -5
  64. mindspore/communication/_comm_helper.py +47 -2
  65. mindspore/communication/comm_func.py +70 -53
  66. mindspore/communication/management.py +83 -17
  67. mindspore/context.py +214 -560
  68. mindspore/dataset/__init__.py +44 -20
  69. mindspore/dataset/audio/__init__.py +2 -8
  70. mindspore/dataset/audio/transforms.py +3 -17
  71. mindspore/dataset/core/config.py +3 -3
  72. mindspore/dataset/engine/cache_client.py +1 -1
  73. mindspore/dataset/engine/datasets.py +102 -120
  74. mindspore/dataset/engine/datasets_audio.py +22 -22
  75. mindspore/dataset/engine/datasets_standard_format.py +43 -24
  76. mindspore/dataset/engine/datasets_text.py +78 -85
  77. mindspore/dataset/engine/datasets_user_defined.py +108 -76
  78. mindspore/dataset/engine/datasets_vision.py +111 -108
  79. mindspore/dataset/engine/iterators.py +5 -3
  80. mindspore/dataset/engine/obs/obs_mindrecord_dataset.py +1 -1
  81. mindspore/dataset/engine/samplers.py +279 -57
  82. mindspore/dataset/engine/serializer_deserializer.py +2 -1
  83. mindspore/dataset/engine/validators.py +10 -0
  84. mindspore/dataset/text/__init__.py +7 -6
  85. mindspore/dataset/text/transforms.py +6 -5
  86. mindspore/dataset/text/utils.py +3 -3
  87. mindspore/dataset/transforms/__init__.py +0 -9
  88. mindspore/dataset/transforms/transforms.py +3 -3
  89. mindspore/dataset/utils/browse_dataset.py +1 -1
  90. mindspore/dataset/vision/__init__.py +2 -9
  91. mindspore/dataset/vision/transforms.py +202 -158
  92. mindspore/dataset/vision/utils.py +7 -5
  93. mindspore/device_context/ascend/op_debug.py +60 -1
  94. mindspore/device_context/ascend/op_tuning.py +0 -4
  95. mindspore/device_manager.py +39 -3
  96. mindspore/dnnl.dll +0 -0
  97. mindspore/dpcmi.dll +0 -0
  98. mindspore/experimental/es/embedding_service.py +35 -27
  99. mindspore/experimental/map_parameter.py +4 -4
  100. mindspore/experimental/optim/adadelta.py +22 -26
  101. mindspore/experimental/optim/adagrad.py +4 -4
  102. mindspore/experimental/optim/adam.py +4 -0
  103. mindspore/experimental/optim/adamax.py +4 -4
  104. mindspore/experimental/optim/adamw.py +4 -0
  105. mindspore/experimental/optim/asgd.py +1 -1
  106. mindspore/experimental/optim/lr_scheduler.py +40 -22
  107. mindspore/experimental/optim/radam.py +5 -5
  108. mindspore/experimental/optim/rprop.py +1 -1
  109. mindspore/experimental/optim/sgd.py +1 -1
  110. mindspore/hal/contiguous_tensors_handle.py +6 -10
  111. mindspore/hal/device.py +55 -81
  112. mindspore/hal/event.py +38 -55
  113. mindspore/hal/memory.py +93 -144
  114. mindspore/hal/stream.py +81 -125
  115. mindspore/include/dataset/constants.h +7 -4
  116. mindspore/include/dataset/execute.h +2 -2
  117. mindspore/jpeg62.dll +0 -0
  118. mindspore/log.py +40 -2
  119. mindspore/mindrecord/__init__.py +20 -7
  120. mindspore/mindspore_backend_common.dll +0 -0
  121. mindspore/mindspore_backend_manager.dll +0 -0
  122. mindspore/mindspore_common.dll +0 -0
  123. mindspore/mindspore_core.dll +0 -0
  124. mindspore/mindspore_dump.dll +0 -0
  125. mindspore/mindspore_frontend.dll +0 -0
  126. mindspore/mindspore_glog.dll +0 -0
  127. mindspore/mindspore_memory_pool.dll +0 -0
  128. mindspore/mindspore_ms_backend.dll +0 -0
  129. mindspore/mindspore_ops.dll +0 -0
  130. mindspore/{mindspore_backend.dll → mindspore_ops_host.dll} +0 -0
  131. mindspore/mindspore_ops_kernel_common.dll +0 -0
  132. mindspore/mindspore_profiler.dll +0 -0
  133. mindspore/mindspore_pyboost.dll +0 -0
  134. mindspore/mindspore_pynative.dll +0 -0
  135. mindspore/mindspore_res_manager.dll +0 -0
  136. mindspore/mindspore_runtime_pipeline.dll +0 -0
  137. mindspore/mint/__init__.py +131 -700
  138. mindspore/mint/distributed/__init__.py +5 -1
  139. mindspore/mint/distributed/distributed.py +194 -109
  140. mindspore/mint/linalg/__init__.py +2 -0
  141. mindspore/mint/nn/__init__.py +280 -18
  142. mindspore/mint/nn/functional.py +282 -64
  143. mindspore/mint/nn/layer/__init__.py +4 -0
  144. mindspore/mint/nn/layer/_functions.py +7 -3
  145. mindspore/mint/nn/layer/activation.py +120 -13
  146. mindspore/mint/nn/layer/conv.py +218 -24
  147. mindspore/mint/nn/layer/normalization.py +15 -16
  148. mindspore/mint/nn/layer/padding.py +1 -1
  149. mindspore/mint/nn/layer/pooling.py +66 -1
  150. mindspore/mint/optim/__init__.py +2 -1
  151. mindspore/mint/optim/sgd.py +171 -0
  152. mindspore/msobj140.dll +0 -0
  153. mindspore/mspdb140.dll +0 -0
  154. mindspore/mspdbcore.dll +0 -0
  155. mindspore/mspdbst.dll +0 -0
  156. mindspore/mspft140.dll +0 -0
  157. mindspore/msvcdis140.dll +0 -0
  158. mindspore/msvcp140_1.dll +0 -0
  159. mindspore/msvcp140_2.dll +0 -0
  160. mindspore/msvcp140_atomic_wait.dll +0 -0
  161. mindspore/msvcp140_codecvt_ids.dll +0 -0
  162. mindspore/nn/__init__.py +4 -1
  163. mindspore/nn/cell.py +1250 -176
  164. mindspore/nn/layer/activation.py +23 -21
  165. mindspore/nn/layer/basic.py +22 -16
  166. mindspore/nn/layer/container.py +1 -1
  167. mindspore/nn/layer/conv.py +22 -17
  168. mindspore/nn/layer/embedding.py +9 -8
  169. mindspore/nn/layer/normalization.py +48 -42
  170. mindspore/nn/layer/pooling.py +75 -31
  171. mindspore/nn/layer/transformer.py +11 -10
  172. mindspore/nn/learning_rate_schedule.py +4 -2
  173. mindspore/nn/loss/loss.py +27 -19
  174. mindspore/nn/optim/ada_grad.py +6 -5
  175. mindspore/nn/optim/adadelta.py +9 -7
  176. mindspore/nn/optim/adafactor.py +1 -1
  177. mindspore/nn/optim/adam.py +16 -12
  178. mindspore/nn/optim/adamax.py +8 -7
  179. mindspore/nn/optim/adasum.py +5 -5
  180. mindspore/nn/optim/asgd.py +1 -1
  181. mindspore/nn/optim/ftrl.py +11 -9
  182. mindspore/nn/optim/lamb.py +1 -1
  183. mindspore/nn/optim/lazyadam.py +12 -10
  184. mindspore/nn/optim/momentum.py +7 -6
  185. mindspore/nn/optim/optimizer.py +2 -2
  186. mindspore/nn/optim/proximal_ada_grad.py +12 -10
  187. mindspore/nn/optim/rmsprop.py +13 -12
  188. mindspore/nn/optim/rprop.py +9 -7
  189. mindspore/nn/optim/sgd.py +9 -6
  190. mindspore/nn/optim/tft_wrapper.py +5 -2
  191. mindspore/nn/probability/bijector/bijector.py +17 -11
  192. mindspore/nn/probability/bijector/gumbel_cdf.py +5 -5
  193. mindspore/nn/probability/bijector/invert.py +2 -2
  194. mindspore/nn/probability/bijector/scalar_affine.py +3 -3
  195. mindspore/nn/probability/bijector/softplus.py +3 -2
  196. mindspore/nn/probability/distribution/beta.py +3 -3
  197. mindspore/nn/probability/distribution/categorical.py +1 -1
  198. mindspore/nn/probability/distribution/cauchy.py +4 -2
  199. mindspore/nn/probability/distribution/exponential.py +6 -7
  200. mindspore/nn/probability/distribution/gamma.py +2 -2
  201. mindspore/nn/probability/distribution/gumbel.py +2 -2
  202. mindspore/nn/probability/distribution/half_normal.py +5 -3
  203. mindspore/nn/probability/distribution/logistic.py +5 -3
  204. mindspore/nn/probability/distribution/poisson.py +1 -1
  205. mindspore/nn/probability/distribution/uniform.py +5 -3
  206. mindspore/nn/reinforcement/_tensors_queue.py +1 -1
  207. mindspore/nn/reinforcement/tensor_array.py +1 -1
  208. mindspore/nn/wrap/__init__.py +6 -6
  209. mindspore/nn/wrap/cell_wrapper.py +178 -117
  210. mindspore/nn/wrap/grad_reducer.py +45 -36
  211. mindspore/nn/wrap/loss_scale.py +3 -3
  212. mindspore/numpy/array_creations.py +3 -3
  213. mindspore/numpy/array_ops.py +1 -1
  214. mindspore/numpy/math_ops.py +4 -4
  215. mindspore/numpy/utils.py +1 -2
  216. mindspore/numpy/utils_const.py +1 -2
  217. mindspore/opencv_core452.dll +0 -0
  218. mindspore/opencv_imgcodecs452.dll +0 -0
  219. mindspore/opencv_imgproc452.dll +0 -0
  220. mindspore/ops/__init__.py +3 -2
  221. mindspore/ops/_grad_experimental/grad_comm_ops.py +18 -3
  222. mindspore/ops/_grad_experimental/grad_debug_ops.py +8 -1
  223. mindspore/ops/_grad_experimental/taylor_rule.py +29 -0
  224. mindspore/ops/_register_for_op.py +0 -11
  225. mindspore/{ops_generate → ops/_utils}/arg_dtype_cast.py +123 -4
  226. mindspore/{ops_generate → ops/_utils}/arg_handler.py +3 -4
  227. mindspore/ops/_vmap/vmap_array_ops.py +7 -6
  228. mindspore/ops/_vmap/vmap_grad_nn_ops.py +2 -1
  229. mindspore/ops/_vmap/vmap_math_ops.py +4 -7
  230. mindspore/ops/_vmap/vmap_nn_ops.py +9 -8
  231. mindspore/ops/auto_generate/__init__.py +4 -3
  232. mindspore/ops/auto_generate/cpp_create_prim_instance_helper.py +102 -49
  233. mindspore/ops/auto_generate/gen_extend_func.py +281 -135
  234. mindspore/ops/auto_generate/gen_ops_def.py +2574 -2326
  235. mindspore/ops/auto_generate/gen_ops_prim.py +8566 -2755
  236. mindspore/ops/auto_generate/pyboost_inner_prim.py +106 -76
  237. mindspore/ops/composite/__init__.py +2 -1
  238. mindspore/ops/composite/base.py +19 -24
  239. mindspore/ops/composite/math_ops.py +6 -16
  240. mindspore/ops/composite/multitype_ops/__init__.py +5 -2
  241. mindspore/ops/composite/multitype_ops/_compile_utils.py +2 -3
  242. mindspore/ops/composite/multitype_ops/_constexpr_utils.py +1 -2
  243. mindspore/ops/composite/multitype_ops/add_impl.py +2 -1
  244. mindspore/ops/composite/multitype_ops/bitwise_and_impl.py +2 -1
  245. mindspore/ops/composite/multitype_ops/bitwise_or_impl.py +2 -1
  246. mindspore/ops/composite/multitype_ops/bitwise_xor_impl.py +2 -1
  247. mindspore/ops/composite/multitype_ops/div_impl.py +6 -4
  248. mindspore/ops/composite/multitype_ops/equal_impl.py +4 -3
  249. mindspore/ops/composite/multitype_ops/floordiv_impl.py +2 -1
  250. mindspore/ops/composite/multitype_ops/getitem_impl.py +3 -2
  251. mindspore/ops/composite/multitype_ops/greater_equal_impl.py +4 -3
  252. mindspore/ops/composite/multitype_ops/greater_impl.py +4 -3
  253. mindspore/ops/composite/multitype_ops/in_impl.py +2 -1
  254. mindspore/ops/composite/multitype_ops/invert_impl.py +50 -0
  255. mindspore/ops/composite/multitype_ops/left_shift_impl.py +2 -1
  256. mindspore/ops/composite/multitype_ops/less_equal_impl.py +4 -3
  257. mindspore/ops/composite/multitype_ops/less_impl.py +4 -3
  258. mindspore/ops/composite/multitype_ops/logic_not_impl.py +3 -2
  259. mindspore/ops/composite/multitype_ops/logical_and_impl.py +2 -1
  260. mindspore/ops/composite/multitype_ops/logical_or_impl.py +2 -1
  261. mindspore/ops/composite/multitype_ops/mod_impl.py +2 -1
  262. mindspore/ops/composite/multitype_ops/mul_impl.py +3 -2
  263. mindspore/ops/composite/multitype_ops/negative_impl.py +2 -1
  264. mindspore/ops/composite/multitype_ops/not_equal_impl.py +2 -1
  265. mindspore/ops/composite/multitype_ops/not_in_impl.py +2 -1
  266. mindspore/ops/composite/multitype_ops/ones_like_impl.py +18 -0
  267. mindspore/ops/composite/multitype_ops/pow_impl.py +2 -1
  268. mindspore/ops/composite/multitype_ops/right_shift_impl.py +2 -1
  269. mindspore/ops/composite/multitype_ops/setitem_impl.py +2 -1
  270. mindspore/ops/composite/multitype_ops/sub_impl.py +2 -1
  271. mindspore/ops/function/__init__.py +28 -2
  272. mindspore/ops/function/_add_attr_func.py +58 -0
  273. mindspore/ops/function/array_func.py +1629 -2345
  274. mindspore/ops/function/clip_func.py +38 -45
  275. mindspore/ops/function/debug_func.py +36 -44
  276. mindspore/ops/function/grad/__init__.py +1 -0
  277. mindspore/ops/function/grad/grad_func.py +104 -71
  278. mindspore/ops/function/image_func.py +1 -1
  279. mindspore/ops/function/linalg_func.py +46 -78
  280. mindspore/ops/function/math_func.py +3035 -3705
  281. mindspore/ops/function/nn_func.py +676 -241
  282. mindspore/ops/function/other_func.py +159 -1
  283. mindspore/ops/function/parameter_func.py +17 -30
  284. mindspore/ops/function/random_func.py +204 -361
  285. mindspore/ops/function/reshard_func.py +4 -70
  286. mindspore/ops/function/sparse_func.py +3 -3
  287. mindspore/ops/function/sparse_unary_func.py +5 -5
  288. mindspore/ops/function/spectral_func.py +25 -58
  289. mindspore/ops/function/vmap_func.py +24 -17
  290. mindspore/ops/functional.py +6 -4
  291. mindspore/ops/functional_overload.py +547 -4
  292. mindspore/ops/op_info_register.py +32 -244
  293. mindspore/ops/operations/__init__.py +10 -5
  294. mindspore/ops/operations/_custom_ops_utils.py +247 -0
  295. mindspore/ops/operations/_grad_ops.py +1 -10
  296. mindspore/ops/operations/_inner_ops.py +5 -76
  297. mindspore/ops/operations/_ms_kernel.py +4 -10
  298. mindspore/ops/operations/_rl_inner_ops.py +1 -1
  299. mindspore/ops/operations/_scalar_ops.py +3 -2
  300. mindspore/ops/operations/_sequence_ops.py +1 -1
  301. mindspore/ops/operations/_tensor_array.py +1 -1
  302. mindspore/ops/operations/array_ops.py +37 -22
  303. mindspore/ops/operations/comm_ops.py +150 -107
  304. mindspore/ops/operations/custom_ops.py +221 -23
  305. mindspore/ops/operations/debug_ops.py +115 -16
  306. mindspore/ops/operations/inner_ops.py +1 -1
  307. mindspore/ops/operations/linalg_ops.py +1 -58
  308. mindspore/ops/operations/manually_defined/_inner.py +1 -1
  309. mindspore/ops/operations/manually_defined/ops_def.py +746 -79
  310. mindspore/ops/operations/math_ops.py +21 -18
  311. mindspore/ops/operations/nn_ops.py +65 -191
  312. mindspore/ops/operations/other_ops.py +62 -9
  313. mindspore/ops/operations/random_ops.py +13 -7
  314. mindspore/ops/operations/reshard_ops.py +1 -1
  315. mindspore/ops/operations/sparse_ops.py +2 -2
  316. mindspore/ops/primitive.py +43 -32
  317. mindspore/ops/tensor_method.py +232 -13
  318. mindspore/ops_generate/__init__.py +0 -5
  319. mindspore/ops_generate/aclnn/__init__.py +0 -0
  320. mindspore/ops_generate/{aclnn_kernel_register_auto_cc_generator.py → aclnn/aclnn_kernel_register_auto_cc_generator.py} +43 -18
  321. mindspore/ops_generate/{gen_aclnn_implement.py → aclnn/gen_aclnn_implement.py} +49 -51
  322. mindspore/ops_generate/api/__init__.py +0 -0
  323. mindspore/ops_generate/{add_tensor_docs_generator.py → api/add_tensor_docs_generator.py} +9 -7
  324. mindspore/ops_generate/{cpp_create_prim_instance_helper_generator.py → api/cpp_create_prim_instance_helper_generator.py} +6 -9
  325. mindspore/ops_generate/{functional_map_cpp_generator.py → api/functional_map_cpp_generator.py} +25 -12
  326. mindspore/ops_generate/{functional_overload_py_generator.py → api/functional_overload_py_generator.py} +8 -6
  327. mindspore/ops_generate/{functions_cc_generator.py → api/functions_cc_generator.py} +14 -10
  328. mindspore/ops_generate/api/gen_api.py +103 -0
  329. mindspore/ops_generate/{op_api_proto.py → api/op_api_proto.py} +98 -69
  330. mindspore/ops_generate/{tensor_func_reg_cpp_generator.py → api/tensor_func_reg_cpp_generator.py} +82 -43
  331. mindspore/ops_generate/common/__init__.py +0 -0
  332. mindspore/ops_generate/common/gen_constants.py +91 -0
  333. mindspore/ops_generate/{gen_utils.py → common/gen_utils.py} +72 -19
  334. mindspore/ops_generate/{op_proto.py → common/op_proto.py} +64 -1
  335. mindspore/ops_generate/{template.py → common/template.py} +96 -84
  336. mindspore/ops_generate/gen_ops.py +23 -325
  337. mindspore/ops_generate/op_def/__init__.py +0 -0
  338. mindspore/ops_generate/op_def/gen_op_def.py +90 -0
  339. mindspore/ops_generate/{lite_ops_cpp_generator.py → op_def/lite_ops_cpp_generator.py} +47 -11
  340. mindspore/ops_generate/{ops_def_cc_generator.py → op_def/ops_def_cc_generator.py} +18 -7
  341. mindspore/ops_generate/{ops_def_h_generator.py → op_def/ops_def_h_generator.py} +5 -5
  342. mindspore/ops_generate/{ops_name_h_generator.py → op_def/ops_name_h_generator.py} +30 -15
  343. mindspore/ops_generate/op_def/ops_primitive_h_generator.py +125 -0
  344. mindspore/ops_generate/op_def_py/__init__.py +0 -0
  345. mindspore/ops_generate/op_def_py/gen_op_def_py.py +47 -0
  346. mindspore/ops_generate/{op_def_py_generator.py → op_def_py/op_def_py_generator.py} +6 -5
  347. mindspore/ops_generate/{op_prim_py_generator.py → op_def_py/op_prim_py_generator.py} +24 -15
  348. mindspore/ops_generate/pyboost/__init__.py +0 -0
  349. mindspore/ops_generate/{auto_grad_impl_cc_generator.py → pyboost/auto_grad_impl_cc_generator.py} +11 -7
  350. mindspore/ops_generate/{auto_grad_reg_cc_generator.py → pyboost/auto_grad_reg_cc_generator.py} +7 -7
  351. mindspore/ops_generate/{gen_pyboost_func.py → pyboost/gen_pyboost_func.py} +40 -16
  352. mindspore/ops_generate/{op_template_parser.py → pyboost/op_template_parser.py} +105 -24
  353. mindspore/ops_generate/{pyboost_functions_cpp_generator.py → pyboost/pyboost_functions_cpp_generator.py} +55 -18
  354. mindspore/ops_generate/{pyboost_functions_h_generator.py → pyboost/pyboost_functions_h_generator.py} +42 -10
  355. mindspore/ops_generate/{pyboost_functions_py_generator.py → pyboost/pyboost_functions_py_generator.py} +6 -6
  356. mindspore/ops_generate/{pyboost_grad_function_cpp_generator.py → pyboost/pyboost_grad_function_cpp_generator.py} +11 -10
  357. mindspore/ops_generate/{pyboost_inner_prim_generator.py → pyboost/pyboost_inner_prim_generator.py} +8 -7
  358. mindspore/ops_generate/{pyboost_native_grad_functions_generator.py → pyboost/pyboost_native_grad_functions_generator.py} +14 -10
  359. mindspore/ops_generate/{pyboost_op_cpp_code_generator.py → pyboost/pyboost_op_cpp_code_generator.py} +140 -53
  360. mindspore/ops_generate/{pyboost_overload_functions_cpp_generator.py → pyboost/pyboost_overload_functions_cpp_generator.py} +28 -15
  361. mindspore/ops_generate/{pyboost_utils.py → pyboost/pyboost_utils.py} +88 -4
  362. mindspore/ops_generate/resources/__init__.py +0 -0
  363. mindspore/ops_generate/resources/resource_list.py +30 -0
  364. mindspore/ops_generate/resources/resource_loader.py +36 -0
  365. mindspore/ops_generate/resources/resource_manager.py +64 -0
  366. mindspore/ops_generate/resources/yaml_loader.py +88 -0
  367. mindspore/ops_generate/tensor_py_cc_generator.py +122 -0
  368. mindspore/parallel/__init__.py +6 -2
  369. mindspore/parallel/_auto_parallel_context.py +133 -6
  370. mindspore/parallel/_cell_wrapper.py +130 -15
  371. mindspore/parallel/_parallel_serialization.py +95 -4
  372. mindspore/parallel/_ps_context.py +1 -1
  373. mindspore/parallel/_recovery_context.py +7 -2
  374. mindspore/parallel/_tensor.py +142 -18
  375. mindspore/parallel/_utils.py +198 -25
  376. mindspore/parallel/algo_parameter_config.py +3 -3
  377. mindspore/parallel/auto_parallel.py +732 -0
  378. mindspore/parallel/checkpoint_convert.py +159 -0
  379. mindspore/parallel/checkpoint_transform.py +656 -37
  380. mindspore/parallel/cluster/process_entity/_api.py +151 -19
  381. mindspore/parallel/cluster/run.py +1 -1
  382. mindspore/parallel/function/__init__.py +24 -0
  383. mindspore/parallel/function/reshard_func.py +259 -0
  384. mindspore/parallel/nn/__init__.py +25 -0
  385. mindspore/parallel/nn/parallel_cell_wrapper.py +263 -0
  386. mindspore/parallel/nn/parallel_grad_reducer.py +169 -0
  387. mindspore/parallel/parameter_broadcast.py +24 -13
  388. mindspore/parallel/shard.py +137 -61
  389. mindspore/parallel/transform_safetensors.py +287 -95
  390. mindspore/pgodb140.dll +0 -0
  391. mindspore/pgort140.dll +0 -0
  392. mindspore/profiler/__init__.py +9 -5
  393. mindspore/profiler/analysis/parser/ascend_cann_parser.py +6 -2
  394. mindspore/profiler/analysis/parser/ms_framework_parser.py +4 -4
  395. mindspore/profiler/analysis/parser/timeline_assembly_factory/ascend_timeline_assembler.py +7 -4
  396. mindspore/profiler/analysis/parser/timeline_assembly_factory/trace_view_container.py +22 -0
  397. mindspore/profiler/analysis/parser/timeline_creator/fwk_timeline_creator.py +3 -3
  398. mindspore/profiler/analysis/parser/timeline_event/fwk_event.py +241 -86
  399. mindspore/profiler/analysis/viewer/ascend_communication_viewer.py +41 -2
  400. mindspore/profiler/analysis/viewer/ascend_kernel_details_viewer.py +33 -35
  401. mindspore/profiler/analysis/viewer/ascend_memory_viewer.py +7 -0
  402. mindspore/profiler/analysis/viewer/ascend_op_memory_viewer.py +8 -3
  403. mindspore/profiler/analysis/viewer/ascend_step_trace_time_viewer.py +141 -30
  404. mindspore/profiler/analysis/viewer/ms_dataset_viewer.py +5 -6
  405. mindspore/profiler/common/ascend_msprof_exporter.py +5 -4
  406. mindspore/profiler/common/constant.py +12 -0
  407. mindspore/profiler/common/msprof_cmd_tool.py +42 -23
  408. mindspore/profiler/common/path_manager.py +24 -0
  409. mindspore/profiler/common/profiler_context.py +26 -2
  410. mindspore/profiler/common/profiler_meta_data.py +74 -0
  411. mindspore/profiler/common/profiler_parameters.py +59 -18
  412. mindspore/profiler/common/profiler_path_manager.py +66 -7
  413. mindspore/profiler/dynamic_profiler.py +112 -79
  414. mindspore/profiler/envprofiler.py +26 -1
  415. mindspore/profiler/experimental_config.py +197 -0
  416. mindspore/profiler/mstx.py +57 -14
  417. mindspore/profiler/platform/npu_profiler.py +33 -7
  418. mindspore/profiler/profiler.py +541 -45
  419. mindspore/profiler/profiler_action_controller.py +1 -1
  420. mindspore/profiler/profiler_interface.py +4 -0
  421. mindspore/profiler/schedule.py +57 -22
  422. mindspore/rewrite/api/node.py +15 -13
  423. mindspore/rewrite/api/symbol_tree.py +1 -1
  424. mindspore/run_check/_check_version.py +25 -14
  425. mindspore/run_check/run_check.py +1 -1
  426. mindspore/runtime/__init__.py +2 -2
  427. mindspore/runtime/executor.py +40 -11
  428. mindspore/runtime/memory.py +25 -8
  429. mindspore/safeguard/rewrite_obfuscation.py +12 -9
  430. mindspore/swresample-4.dll +0 -0
  431. mindspore/swscale-6.dll +0 -0
  432. mindspore/tbbmalloc.dll +0 -0
  433. mindspore/tinyxml2.dll +0 -0
  434. mindspore/train/__init__.py +8 -8
  435. mindspore/train/_utils.py +35 -7
  436. mindspore/train/amp.py +1 -1
  437. mindspore/train/callback/__init__.py +2 -2
  438. mindspore/train/callback/_callback.py +2 -16
  439. mindspore/train/callback/_checkpoint.py +24 -40
  440. mindspore/train/callback/_cluster_monitor.py +14 -18
  441. mindspore/train/callback/_flops_collector.py +2 -3
  442. mindspore/train/callback/_history.py +7 -4
  443. mindspore/train/callback/_lambda_callback.py +2 -2
  444. mindspore/train/callback/_landscape.py +0 -3
  445. mindspore/train/callback/_loss_monitor.py +2 -1
  446. mindspore/train/callback/_on_request_exit.py +6 -5
  447. mindspore/train/callback/_reduce_lr_on_plateau.py +11 -6
  448. mindspore/train/callback/_summary_collector.py +8 -13
  449. mindspore/train/callback/_time_monitor.py +2 -1
  450. mindspore/train/callback/{_tft_register.py → _train_fault_tolerance.py} +179 -103
  451. mindspore/train/data_sink.py +25 -2
  452. mindspore/train/dataset_helper.py +4 -5
  453. mindspore/train/loss_scale_manager.py +8 -7
  454. mindspore/train/metrics/accuracy.py +3 -3
  455. mindspore/train/metrics/confusion_matrix.py +9 -9
  456. mindspore/train/metrics/error.py +3 -3
  457. mindspore/train/metrics/hausdorff_distance.py +4 -4
  458. mindspore/train/metrics/mean_surface_distance.py +3 -3
  459. mindspore/train/metrics/metric.py +0 -12
  460. mindspore/train/metrics/occlusion_sensitivity.py +4 -2
  461. mindspore/train/metrics/precision.py +8 -6
  462. mindspore/train/metrics/recall.py +9 -9
  463. mindspore/train/metrics/root_mean_square_surface_distance.py +2 -2
  464. mindspore/train/mind_ir_pb2.py +19 -12
  465. mindspore/train/model.py +176 -103
  466. mindspore/train/serialization.py +246 -988
  467. mindspore/train/summary/_summary_adapter.py +2 -2
  468. mindspore/train/summary/summary_record.py +1 -1
  469. mindspore/turbojpeg.dll +0 -0
  470. mindspore/utils/__init__.py +3 -2
  471. mindspore/utils/dryrun.py +4 -2
  472. mindspore/utils/hooks.py +81 -0
  473. mindspore/utils/utils.py +138 -4
  474. mindspore/vcmeta.dll +0 -0
  475. mindspore/vcruntime140.dll +0 -0
  476. mindspore/vcruntime140_1.dll +0 -0
  477. mindspore/version.py +1 -1
  478. {mindspore-2.5.0.dist-info → mindspore-2.6.0rc1.dist-info}/METADATA +2 -1
  479. {mindspore-2.5.0.dist-info → mindspore-2.6.0rc1.dist-info}/RECORD +483 -438
  480. mindspore/_install_custom.py +0 -43
  481. mindspore/common/_register_for_adapter.py +0 -74
  482. mindspore/ops/auto_generate/gen_arg_dtype_cast.py +0 -252
  483. mindspore/ops/auto_generate/gen_arg_handler.py +0 -136
  484. mindspore/ops/operations/_opaque_predicate_registry.py +0 -41
  485. mindspore/ops_generate/gen_constants.py +0 -190
  486. mindspore/ops_generate/gen_ops_inner_prim.py +0 -131
  487. mindspore/ops_generate/ops_primitive_h_generator.py +0 -81
  488. /mindspore/ops_generate/{base_generator.py → common/base_generator.py} +0 -0
  489. {mindspore-2.5.0.dist-info → mindspore-2.6.0rc1.dist-info}/WHEEL +0 -0
  490. {mindspore-2.5.0.dist-info → mindspore-2.6.0rc1.dist-info}/entry_points.txt +0 -0
  491. {mindspore-2.5.0.dist-info → mindspore-2.6.0rc1.dist-info}/top_level.txt +0 -0
@@ -25,7 +25,8 @@ After declaring the dataset object, you can further apply dataset operations
25
25
  """
26
26
  import mindspore._c_dataengine as cde
27
27
 
28
- from .datasets import TextBaseDataset, SourceDataset, MappableDataset, Shuffle
28
+ from .samplers import Shuffle
29
+ from .datasets import TextBaseDataset, SourceDataset, MappableDataset
29
30
  from .validators import check_imdb_dataset, check_iwslt2016_dataset, check_iwslt2017_dataset, \
30
31
  check_penn_treebank_dataset, check_ag_news_dataset, check_amazon_review_dataset, check_udpos_dataset, \
31
32
  check_wiki_text_dataset, check_conll2000_dataset, check_cluedataset, \
@@ -61,18 +62,17 @@ class AGNewsDataset(SourceDataset, TextBaseDataset):
61
62
  Set the mode of data shuffling by passing in enumeration variables:
62
63
 
63
64
  - ``Shuffle.GLOBAL``: Shuffle both the files and samples.
64
-
65
65
  - ``Shuffle.FILES``: Shuffle files only.
66
66
 
67
67
  num_shards (int, optional): Number of shards that the dataset will be divided into.
68
68
  Default: ``None``. When this argument is specified, `num_samples` reflects the
69
69
  max sample number of per shard.
70
- Used in `data parallel training <https://www.mindspore.cn/docs/en/master/model_train/
71
- parallel/data_parallel.html#data-parallel-mode-loads-datasets>`_ .
70
+ Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
71
+ parallel/data_parallel.html#loading-datasets>`_ .
72
72
  shard_id (int, optional): The shard ID within `num_shards` . This
73
73
  argument can only be specified when `num_shards` is also specified. Default: ``None``.
74
74
  cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
75
- `Single-Node Data Cache <https://www.mindspore.cn/docs/en/master/model_train/dataset/cache.html>`_ .
75
+ `Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
76
76
  Default: ``None``, which means no cache is used.
77
77
 
78
78
  Raises:
@@ -90,9 +90,9 @@ class AGNewsDataset(SourceDataset, TextBaseDataset):
90
90
  >>> ag_news_dataset_dir = "/path/to/ag_news_dataset_file"
91
91
  >>> dataset = ds.AGNewsDataset(dataset_dir=ag_news_dataset_dir, usage='all')
92
92
 
93
- About AGNews dataset:
93
+ About AG News dataset:
94
94
 
95
- AG is a collection of over 1 million news articles. The news articles were collected
95
+ AG News is a collection of over 1 million news articles. The news articles were collected
96
96
  by ComeToMyHead from over 2,000 news sources in over 1 year of activity. ComeToMyHead
97
97
  is an academic news search engine that has been in operation since July 2004.
98
98
  The dataset is provided by academics for research purposes such as data mining
@@ -172,17 +172,16 @@ class AmazonReviewDataset(SourceDataset, TextBaseDataset):
172
172
  Set the mode of data shuffling by passing in enumeration variables:
173
173
 
174
174
  - ``Shuffle.GLOBAL`` : Shuffle both the files and samples.
175
-
176
175
  - ``Shuffle.FILES`` : Shuffle files only.
177
176
 
178
177
  num_shards (int, optional): Number of shards that the dataset will be divided into. Default: ``None`` .
179
178
  When this argument is specified, `num_samples` reflects the max sample number of per shard.
180
- Used in `data parallel training <https://www.mindspore.cn/docs/en/master/model_train/
181
- parallel/data_parallel.html#data-parallel-mode-loads-datasets>`_ .
179
+ Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
180
+ parallel/data_parallel.html#loading-datasets>`_ .
182
181
  shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
183
182
  argument can only be specified when `num_shards` is also specified.
184
183
  cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
185
- `Single-Node Data Cache <https://www.mindspore.cn/docs/en/master/model_train/dataset/cache.html>`_ .
184
+ `Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
186
185
  Default: ``None`` , which means no cache is used.
187
186
 
188
187
  Raises:
@@ -272,17 +271,16 @@ class CLUEDataset(SourceDataset, TextBaseDataset):
272
271
  There are three levels of shuffling, desired shuffle enum defined by :class:`mindspore.dataset.Shuffle` .
273
272
 
274
273
  - ``Shuffle.GLOBAL`` : Shuffle both the files and samples, same as setting `shuffle` to ``True``.
275
-
276
274
  - ``Shuffle.FILES`` : Shuffle files only.
277
275
 
278
276
  num_shards (int, optional): Number of shards that the dataset will be divided into. Default: ``None`` .
279
277
  When this argument is specified, `num_samples` reflects the maximum sample number of per shard.
280
- Used in `data parallel training <https://www.mindspore.cn/docs/en/master/model_train/
281
- parallel/data_parallel.html#data-parallel-mode-loads-datasets>`_ .
278
+ Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
279
+ parallel/data_parallel.html#loading-datasets>`_ .
282
280
  shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
283
281
  argument can only be specified when `num_shards` is also specified.
284
282
  cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
285
- `Single-Node Data Cache <https://www.mindspore.cn/docs/en/master/model_train/dataset/cache.html>`_ .
283
+ `Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
286
284
  Default: ``None`` , which means no cache is used.
287
285
 
288
286
  The generated dataset with different task setting has different output columns:
@@ -521,15 +519,15 @@ class CoNLL2000Dataset(SourceDataset, TextBaseDataset):
521
519
 
522
520
  num_shards (int, optional): Number of shards that the dataset will be divided into.
523
521
  When this argument is specified, `num_samples` reflects the max sample number of per shard.
524
- Default: ``None`` . Used in `data parallel training <https://www.mindspore.cn/docs/en/master/model_train/
525
- parallel/data_parallel.html#data-parallel-mode-loads-datasets>`_ .
522
+ Default: ``None`` . Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
523
+ parallel/data_parallel.html#loading-datasets>`_ .
526
524
  shard_id (int, optional): The shard ID within `num_shards` . This
527
525
  argument can only be specified when `num_shards` is also specified. Default: ``None`` .
528
526
  num_parallel_workers (int, optional): Number of worker threads to read the data.
529
527
  Default: ``None`` , will use global default workers(8), it can be set
530
528
  by :func:`mindspore.dataset.config.set_num_parallel_workers` .
531
529
  cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
532
- `Single-Node Data Cache <https://www.mindspore.cn/docs/en/master/model_train/dataset/cache.html>`_ .
530
+ `Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
533
531
  Default: ``None`` , which means no cache is used.
534
532
 
535
533
  Raises:
@@ -620,17 +618,16 @@ class DBpediaDataset(SourceDataset, TextBaseDataset):
620
618
  Set the mode of data shuffling by passing in enumeration variables:
621
619
 
622
620
  - ``Shuffle.GLOBAL`` : Shuffle both the files and samples.
623
-
624
621
  - ``Shuffle.FILES`` : Shuffle files only.
625
622
 
626
623
  num_shards (int, optional): Number of shards that the dataset will be divided into. Default: ``None`` .
627
624
  When this argument is specified, `num_samples` reflects the maximum sample number of per shard.
628
- Used in `data parallel training <https://www.mindspore.cn/docs/en/master/model_train/
629
- parallel/data_parallel.html#data-parallel-mode-loads-datasets>`_ .
625
+ Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
626
+ parallel/data_parallel.html#loading-datasets>`_ .
630
627
  shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
631
628
  argument can only be specified when `num_shards` is also specified.
632
629
  cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
633
- `Single-Node Data Cache <https://www.mindspore.cn/docs/en/master/model_train/dataset/cache.html>`_ .
630
+ `Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
634
631
  Default: ``None`` , which means no cache is used.
635
632
 
636
633
  Raises:
@@ -721,17 +718,16 @@ class EnWik9Dataset(SourceDataset, TextBaseDataset):
721
718
  Set the mode of data shuffling by passing in enumeration variables:
722
719
 
723
720
  - ``Shuffle.GLOBAL`` : Shuffle both the files and samples.
724
-
725
721
  - ``Shuffle.FILES`` : Shuffle files only.
726
722
 
727
723
  num_shards (int, optional): Number of shards that the dataset will be divided into. Default: ``None`` .
728
724
  When this argument is specified, `num_samples` reflects the maximum sample number of per shard.
729
- Used in `data parallel training <https://www.mindspore.cn/docs/en/master/model_train/
730
- parallel/data_parallel.html#data-parallel-mode-loads-datasets>`_ .
725
+ Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
726
+ parallel/data_parallel.html#loading-datasets>`_ .
731
727
  shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
732
728
  argument can only be specified when `num_shards` is also specified.
733
729
  cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
734
- `Single-Node Data Cache <https://www.mindspore.cn/docs/en/master/model_train/dataset/cache.html>`_ .
730
+ `Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
735
731
  Default: ``None`` , which means no cache is used.
736
732
 
737
733
  Raises:
@@ -816,12 +812,12 @@ class IMDBDataset(MappableDataset, TextBaseDataset):
816
812
  num_shards (int, optional): Number of shards that the dataset will be divided
817
813
  into. Default: ``None`` . When this argument is specified, `num_samples` reflects
818
814
  the maximum sample number of per shard.
819
- Used in `data parallel training <https://www.mindspore.cn/docs/en/master/model_train/
820
- parallel/data_parallel.html#data-parallel-mode-loads-datasets>`_ .
815
+ Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
816
+ parallel/data_parallel.html#loading-datasets>`_ .
821
817
  shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
822
818
  argument can only be specified when `num_shards` is also specified.
823
819
  cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
824
- `Single-Node Data Cache <https://www.mindspore.cn/docs/en/master/model_train/dataset/cache.html>`_ .
820
+ `Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
825
821
  Default: ``None`` , which means no cache is used.
826
822
 
827
823
  Raises:
@@ -949,20 +945,19 @@ class IWSLT2016Dataset(SourceDataset, TextBaseDataset):
949
945
  Set the mode of data shuffling by passing in enumeration variables:
950
946
 
951
947
  - ``Shuffle.GLOBAL`` : Shuffle both the files and samples.
952
-
953
948
  - ``Shuffle.FILES`` : Shuffle files only.
954
949
 
955
950
  num_shards (int, optional): Number of shards that the dataset will be divided into. Default: ``None`` .
956
951
  When this argument is specified, `num_samples` reflects the max sample number of per shard.
957
- Used in `data parallel training <https://www.mindspore.cn/docs/en/master/model_train/
958
- parallel/data_parallel.html#data-parallel-mode-loads-datasets>`_ .
952
+ Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
953
+ parallel/data_parallel.html#loading-datasets>`_ .
959
954
  shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
960
955
  argument can only be specified when `num_shards` is also specified.
961
956
  num_parallel_workers (int, optional): Number of worker threads to read the data.
962
957
  Default: ``None`` , will use global default workers(8), it can be set
963
958
  by :func:`mindspore.dataset.config.set_num_parallel_workers` .
964
959
  cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
965
- `Single-Node Data Cache <https://www.mindspore.cn/docs/en/master/model_train/dataset/cache.html>`_ .
960
+ `Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
966
961
  Default: ``None`` , which means no cache is used.
967
962
 
968
963
  Raises:
@@ -1083,20 +1078,19 @@ class IWSLT2017Dataset(SourceDataset, TextBaseDataset):
1083
1078
  Set the mode of data shuffling by passing in enumeration variables:
1084
1079
 
1085
1080
  - ``Shuffle.GLOBAL`` : Shuffle both the files and samples.
1086
-
1087
1081
  - ``Shuffle.FILES`` : Shuffle files only.
1088
1082
 
1089
1083
  num_shards (int, optional): Number of shards that the dataset will be divided into. Default: ``None`` .
1090
1084
  When this argument is specified, `num_samples` reflects the max sample number of per shard.
1091
- Used in `data parallel training <https://www.mindspore.cn/docs/en/master/model_train/
1092
- parallel/data_parallel.html#data-parallel-mode-loads-datasets>`_ .
1085
+ Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
1086
+ parallel/data_parallel.html#loading-datasets>`_ .
1093
1087
  shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
1094
1088
  argument can only be specified when `num_shards` is also specified.
1095
1089
  num_parallel_workers (int, optional): Number of worker threads to read the data.
1096
1090
  Default: ``None`` , will use global default workers(8), it can be set
1097
1091
  by :func:`mindspore.dataset.config.set_num_parallel_workers` .
1098
1092
  cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
1099
- `Single-Node Data Cache <https://www.mindspore.cn/docs/en/master/model_train/dataset/cache.html>`_ .
1093
+ `Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
1100
1094
  Default: ``None`` , which means no cache is used.
1101
1095
 
1102
1096
  Raises:
@@ -1198,12 +1192,12 @@ class Multi30kDataset(SourceDataset, TextBaseDataset):
1198
1192
  num_shards (int, optional): Number of shards that the dataset will be divided
1199
1193
  into. Default: ``None`` . When this argument is specified, `num_samples` reflects
1200
1194
  the max sample number of per shard.
1201
- Used in `data parallel training <https://www.mindspore.cn/docs/en/master/model_train/
1202
- parallel/data_parallel.html#data-parallel-mode-loads-datasets>`_ .
1195
+ Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
1196
+ parallel/data_parallel.html#loading-datasets>`_ .
1203
1197
  shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
1204
1198
  argument can only be specified when `num_shards` is also specified.
1205
1199
  cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
1206
- `Single-Node Data Cache <https://www.mindspore.cn/docs/en/master/model_train/dataset/cache.html>`_ .
1200
+ `Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
1207
1201
  Default: ``None`` , which means no cache is used.
1208
1202
 
1209
1203
  Raises:
@@ -1229,7 +1223,7 @@ class Multi30kDataset(SourceDataset, TextBaseDataset):
1229
1223
 
1230
1224
  Multi30K is a multilingual dataset that features approximately 31,000 standardized images
1231
1225
  described in multiple languages. The images are sourced from Flickr and each image comes
1232
- with sentence descripitions in both English and German, as well as descriptions in other
1226
+ with sentence descriptions in both English and German, as well as descriptions in other
1233
1227
  languages. Multi30k is used primarily for training and testing in tasks such as image
1234
1228
  captioning, machine translation, and visual question answering.
1235
1229
 
@@ -1304,17 +1298,16 @@ class PennTreebankDataset(SourceDataset, TextBaseDataset):
1304
1298
  Set the mode of data shuffling by passing in enumeration variables:
1305
1299
 
1306
1300
  - ``Shuffle.GLOBAL`` : Shuffle both the files and samples.
1307
-
1308
1301
  - ``Shuffle.FILES`` : Shuffle files only.
1309
1302
 
1310
1303
  num_shards (int, optional): Number of shards that the dataset will be divided into. Default: ``None`` .
1311
1304
  When this argument is specified, `num_samples` reflects the max sample number of per shard.
1312
- Used in `data parallel training <https://www.mindspore.cn/docs/en/master/model_train/
1313
- parallel/data_parallel.html#data-parallel-mode-loads-datasets>`_ .
1305
+ Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
1306
+ parallel/data_parallel.html#loading-datasets>`_ .
1314
1307
  shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
1315
1308
  argument can only be specified when `num_shards` is also specified.
1316
1309
  cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
1317
- `Single-Node Data Cache <https://www.mindspore.cn/docs/en/master/model_train/dataset/cache.html>`_ .
1310
+ `Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
1318
1311
  Default: ``None`` , which means no cache is used.
1319
1312
 
1320
1313
  Raises:
@@ -1335,7 +1328,7 @@ class PennTreebankDataset(SourceDataset, TextBaseDataset):
1335
1328
  About PennTreebank dataset:
1336
1329
 
1337
1330
  Penn Treebank (PTB) dataset, is widely used in machine learning for NLP (Natural Language Processing)
1338
- research. Word-level PTB does not contain capital letters, numbers, and punctuations, and the vocabulary
1331
+ research. Word-level PTB does not contain capital letters, numbers, and punctuation, and the vocabulary
1339
1332
  is capped at 10k unique words, which is relatively small in comparison to most modern datasets which
1340
1333
  can result in a larger number of out of vocabulary tokens.
1341
1334
 
@@ -1406,19 +1399,19 @@ class SogouNewsDataset(SourceDataset, TextBaseDataset):
1406
1399
  Set the mode of data shuffling by passing in enumeration variables:
1407
1400
 
1408
1401
  - ``Shuffle.GLOBAL`` : Shuffle both the files and samples, same as setting shuffle to True.
1409
-
1410
1402
  - ``Shuffle.FILES`` : Shuffle files only.
1403
+
1411
1404
  num_shards (int, optional): Number of shards that the dataset will be divided into. Default: ``None`` .
1412
1405
  When this argument is specified, `num_samples` reflects the max sample number of per shard.
1413
- Used in `data parallel training <https://www.mindspore.cn/docs/en/master/model_train/
1414
- parallel/data_parallel.html#data-parallel-mode-loads-datasets>`_ .
1406
+ Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
1407
+ parallel/data_parallel.html#loading-datasets>`_ .
1415
1408
  shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
1416
1409
  argument can only be specified when `num_shards` is also specified.
1417
1410
  num_parallel_workers (int, optional): Number of worker threads to read the data.
1418
1411
  Default: ``None`` , will use global default workers(8), it can be set
1419
1412
  by :func:`mindspore.dataset.config.set_num_parallel_workers` .
1420
1413
  cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
1421
- `Single-Node Data Cache <https://www.mindspore.cn/docs/en/master/model_train/dataset/cache.html>`_ .
1414
+ `Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
1422
1415
  Default: ``None`` , which means no cache is used.
1423
1416
 
1424
1417
  Raises:
@@ -1436,11 +1429,11 @@ class SogouNewsDataset(SourceDataset, TextBaseDataset):
1436
1429
  >>> sogou_news_dataset_dir = "/path/to/sogou_news_dataset_dir"
1437
1430
  >>> dataset = ds.SogouNewsDataset(dataset_dir=sogou_news_dataset_dir, usage='all')
1438
1431
 
1439
- About SogouNews Dataset:
1432
+ About Sogou News Dataset:
1440
1433
 
1441
- SogouNews dataset includes 3 columns, corresponding to class index (1 to 5), title and content. The title and
1434
+ Sogou News dataset includes 3 columns, corresponding to class index (1 to 5), title and content. The title and
1442
1435
  content are escaped using double quotes ("), and any internal double quote is escaped by 2 double quotes ("").
1443
- New lines are escaped by a backslash followed with an "n" character, that is "\n".
1436
+ New lines are escaped by a backslash followed with an "n" character, that is "\\n".
1444
1437
 
1445
1438
  You can unzip the dataset files into the following structure and read by MindSpore's API:
1446
1439
 
@@ -1513,12 +1506,12 @@ class SQuADDataset(SourceDataset, TextBaseDataset):
1513
1506
 
1514
1507
  num_shards (int, optional): Number of shards that the dataset will be divided into. Default: ``None`` .
1515
1508
  When this argument is specified, `num_samples` reflects the maximum sample number of per shard.
1516
- Used in `data parallel training <https://www.mindspore.cn/docs/en/master/model_train/
1517
- parallel/data_parallel.html#data-parallel-mode-loads-datasets>`_ .
1509
+ Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
1510
+ parallel/data_parallel.html#loading-datasets>`_ .
1518
1511
  shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
1519
1512
  argument can only be specified when `num_shards` is also specified.
1520
1513
  cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
1521
- `Single-Node Data Cache <https://www.mindspore.cn/docs/en/master/model_train/dataset/cache.html>`_ .
1514
+ `Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
1522
1515
  Default: ``None`` , which means no cache is used.
1523
1516
 
1524
1517
  Raises:
@@ -1629,16 +1622,17 @@ class SST2Dataset(SourceDataset, TextBaseDataset):
1629
1622
  If `shuffle` is ``True`` , the behavior is the same as setting shuffle to be Shuffle.GLOBAL
1630
1623
  Set the mode of data shuffling by passing in enumeration variables:
1631
1624
 
1632
- - ``Shuffle.GLOBAL`` : Shuffle the samples.
1625
+ - ``Shuffle.GLOBAL`` : Shuffle both the files and samples.
1626
+ - ``Shuffle.FILES`` : Shuffle files only.
1633
1627
 
1634
1628
  num_shards (int, optional): Number of shards that the dataset will be divided into. Default: ``None`` .
1635
1629
  When this argument is specified, `num_samples` reflects the maximum sample number of per shard.
1636
- Used in `data parallel training <https://www.mindspore.cn/docs/en/master/model_train/
1637
- parallel/data_parallel.html#data-parallel-mode-loads-datasets>`_ .
1630
+ Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
1631
+ parallel/data_parallel.html#loading-datasets>`_ .
1638
1632
  shard_id (int, optional): The shard ID within `num_shards`. This argument can only be specified when
1639
1633
  `num_shards` is also specified. Default: ``None`` .
1640
1634
  cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
1641
- `Single-Node Data Cache <https://www.mindspore.cn/docs/en/master/model_train/dataset/cache.html>`_ .
1635
+ `Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
1642
1636
  Default: ``None`` , which means no cache is used.
1643
1637
 
1644
1638
  Raises:
@@ -1733,17 +1727,16 @@ class TextFileDataset(SourceDataset, TextBaseDataset):
1733
1727
  There are three levels of shuffling, desired shuffle enum defined by :class:`mindspore.dataset.Shuffle` .
1734
1728
 
1735
1729
  - ``Shuffle.GLOBAL`` : Shuffle both the files and samples, same as setting shuffle to True.
1736
-
1737
1730
  - ``Shuffle.FILES`` : Shuffle files only.
1738
1731
 
1739
1732
  num_shards (int, optional): Number of shards that the dataset will be divided into. Default: ``None`` .
1740
1733
  When this argument is specified, `num_samples` reflects the maximum sample number of per shard.
1741
- Used in `data parallel training <https://www.mindspore.cn/docs/en/master/model_train/
1742
- parallel/data_parallel.html#data-parallel-mode-loads-datasets>`_ .
1734
+ Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
1735
+ parallel/data_parallel.html#loading-datasets>`_ .
1743
1736
  shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
1744
1737
  argument can only be specified when `num_shards` is also specified.
1745
1738
  cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
1746
- `Single-Node Data Cache <https://www.mindspore.cn/docs/en/master/model_train/dataset/cache.html>`_ .
1739
+ `Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
1747
1740
  Default: ``None`` , which means no cache is used.
1748
1741
 
1749
1742
  Raises:
@@ -1799,20 +1792,19 @@ class UDPOSDataset(SourceDataset, TextBaseDataset):
1799
1792
  Set the mode of data shuffling by passing in enumeration variables:
1800
1793
 
1801
1794
  - ``Shuffle.GLOBAL`` : Shuffle both the files and samples.
1802
-
1803
1795
  - ``Shuffle.FILES`` : Shuffle files only.
1804
1796
 
1805
1797
  num_shards (int, optional): Number of shards that the dataset will be divided into. Default: ``None`` .
1806
1798
  When this argument is specified, `num_samples` reflects the max sample number of per shard.
1807
- Used in `data parallel training <https://www.mindspore.cn/docs/en/master/model_train/
1808
- parallel/data_parallel.html#data-parallel-mode-loads-datasets>`_ .
1799
+ Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
1800
+ parallel/data_parallel.html#loading-datasets>`_ .
1809
1801
  shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
1810
1802
  argument can only be specified when `num_shards` is also specified.
1811
1803
  num_parallel_workers (int, optional): Number of worker threads to read the data.
1812
1804
  Default: ``None`` , will use global default workers(8), it can be set
1813
1805
  by :func:`mindspore.dataset.config.set_num_parallel_workers` .
1814
1806
  cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
1815
- `Single-Node Data Cache <https://www.mindspore.cn/docs/en/master/model_train/dataset/cache.html>`_ .
1807
+ `Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
1816
1808
  Default: ``None`` , which means no cache is used.
1817
1809
 
1818
1810
  Raises:
@@ -1887,17 +1879,16 @@ class WikiTextDataset(SourceDataset, TextBaseDataset):
1887
1879
  Set the mode of data shuffling by passing in enumeration variables:
1888
1880
 
1889
1881
  - ``Shuffle.GLOBAL`` : Shuffle both the files and samples.
1890
-
1891
1882
  - ``Shuffle.FILES`` : Shuffle files only.
1892
1883
 
1893
1884
  num_shards (int, optional): Number of shards that the dataset will be divided into. Default: ``None`` .
1894
1885
  When this argument is specified, `num_samples` reflects the max sample number of per shard.
1895
- Used in `data parallel training <https://www.mindspore.cn/docs/en/master/model_train/
1896
- parallel/data_parallel.html#data-parallel-mode-loads-datasets>`_ .
1886
+ Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
1887
+ parallel/data_parallel.html#loading-datasets>`_ .
1897
1888
  shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
1898
1889
  argument can only be specified when `num_shards` is also specified.
1899
1890
  cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
1900
- `Single-Node Data Cache <https://www.mindspore.cn/docs/en/master/model_train/dataset/cache.html>`_ .
1891
+ `Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
1901
1892
  Default: ``None`` , which means no cache is used.
1902
1893
 
1903
1894
  Raises:
@@ -1986,17 +1977,16 @@ class YahooAnswersDataset(SourceDataset, TextBaseDataset):
1986
1977
  Set the mode of data shuffling by passing in enumeration variables:
1987
1978
 
1988
1979
  - ``Shuffle.GLOBAL`` : Shuffle both the files and samples.
1989
-
1990
1980
  - ``Shuffle.FILES`` : Shuffle files only.
1991
1981
 
1992
1982
  num_shards (int, optional): Number of shards that the dataset will be divided into. Default: ``None`` .
1993
1983
  When this argument is specified, `num_samples` reflects the maximum sample number of per shard.
1994
- Used in `data parallel training <https://www.mindspore.cn/docs/en/master/model_train/
1995
- parallel/data_parallel.html#data-parallel-mode-loads-datasets>`_ .
1984
+ Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
1985
+ parallel/data_parallel.html#loading-datasets>`_ .
1996
1986
  shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
1997
1987
  argument can only be specified when `num_shards` is also specified.
1998
1988
  cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
1999
- `Single-Node Data Cache <https://www.mindspore.cn/docs/en/master/model_train/dataset/cache.html>`_ .
1989
+ `Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
2000
1990
  Default: ``None`` , which means no cache is used.
2001
1991
 
2002
1992
  Raises:
@@ -2074,11 +2064,14 @@ class YelpReviewDataset(SourceDataset, TextBaseDataset):
2074
2064
  Args:
2075
2065
  dataset_dir (str): Path to the root directory that contains the dataset.
2076
2066
  usage (str, optional): Usage of this dataset, can be ``'train'`` , ``'test'`` or ``'all'`` .
2077
- For Polarity, ``'train'`` will read from 560,000 train samples,
2078
- ``'test'`` will read from 38,000 test samples,
2079
- ``'all'`` will read from all 598,000 samples.
2080
- For Full, ``'train'`` will read from 650,000 train samples, ``'test'`` will read from 50,000 test samples,
2081
- ``'all'`` will read from all 700,000 samples. Default: ``None`` , all samples.
2067
+ Default: ``None`` , all samples.
2068
+
2069
+ - For Polarity, ``'train'`` will read from 560,000 train samples,
2070
+ ``'test'`` will read from 38,000 test samples,
2071
+ ``'all'`` will read from all 598,000 samples.
2072
+ - For Full, ``'train'`` will read from 650,000 train samples, ``'test'`` will read from 50,000 test samples,
2073
+ ``'all'`` will read from all 700,000 samples.
2074
+
2082
2075
  num_samples (int, optional): Number of samples (rows) to read. Default: ``None`` , reads all samples.
2083
2076
  shuffle (Union[bool, Shuffle], optional): Perform reshuffling of the data every epoch.
2084
2077
  Bool type and Shuffle enum are both supported to pass in.
@@ -2089,19 +2082,19 @@ class YelpReviewDataset(SourceDataset, TextBaseDataset):
2089
2082
  Set the mode of data shuffling by passing in enumeration variables:
2090
2083
 
2091
2084
  - ``Shuffle.GLOBAL`` : Shuffle both the files and samples.
2092
-
2093
2085
  - ``Shuffle.FILES`` : Shuffle files only.
2086
+
2094
2087
  num_shards (int, optional): Number of shards that the dataset will be divided into. Default: ``None`` .
2095
2088
  When this argument is specified, `num_samples` reflects the max sample number of per shard.
2096
- Used in `data parallel training <https://www.mindspore.cn/docs/en/master/model_train/
2097
- parallel/data_parallel.html#data-parallel-mode-loads-datasets>`_ .
2089
+ Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
2090
+ parallel/data_parallel.html#loading-datasets>`_ .
2098
2091
  shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
2099
2092
  argument can only be specified when `num_shards` is also specified.
2100
2093
  num_parallel_workers (int, optional): Number of worker threads to read the data.
2101
2094
  Default: ``None`` , will use global default workers(8), it can be set
2102
2095
  by :func:`mindspore.dataset.config.set_num_parallel_workers` .
2103
2096
  cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
2104
- `Single-Node Data Cache <https://www.mindspore.cn/docs/en/master/model_train/dataset/cache.html>`_ .
2097
+ `Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
2105
2098
  Default: ``None`` , which means no cache is used.
2106
2099
 
2107
2100
  Raises: