mindspore 2.7.0__cp311-cp311-win_amd64.whl → 2.7.1__cp311-cp311-win_amd64.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of mindspore might be problematic. Click here for more details.

Files changed (290) hide show
  1. mindspore/.commit_id +1 -1
  2. mindspore/__init__.py +4 -1
  3. mindspore/_c_dataengine.cp311-win_amd64.pyd +0 -0
  4. mindspore/_c_expression.cp311-win_amd64.pyd +0 -0
  5. mindspore/_c_mindrecord.cp311-win_amd64.pyd +0 -0
  6. mindspore/_extends/parse/compile_config.py +24 -1
  7. mindspore/_extends/parse/deprecated/deprecated_tensor_method.py +6 -2
  8. mindspore/_extends/parse/resources.py +1 -1
  9. mindspore/_extends/parse/standard_method.py +8 -1
  10. mindspore/_extends/parse/trope.py +2 -1
  11. mindspore/_extends/pijit/pijit_func_white_list.py +7 -22
  12. mindspore/avcodec-59.dll +0 -0
  13. mindspore/avdevice-59.dll +0 -0
  14. mindspore/avfilter-8.dll +0 -0
  15. mindspore/avformat-59.dll +0 -0
  16. mindspore/avutil-57.dll +0 -0
  17. mindspore/boost/base.py +29 -2
  18. mindspore/common/_decorator.py +3 -2
  19. mindspore/common/_grad_function.py +3 -1
  20. mindspore/common/_tensor_cpp_method.py +1 -1
  21. mindspore/common/_tensor_docs.py +275 -64
  22. mindspore/common/_utils.py +0 -44
  23. mindspore/common/api.py +285 -35
  24. mindspore/common/dump.py +7 -108
  25. mindspore/common/dynamic_shape/auto_dynamic_shape.py +1 -3
  26. mindspore/common/hook_handle.py +60 -0
  27. mindspore/common/jit_config.py +5 -1
  28. mindspore/common/jit_trace.py +27 -12
  29. mindspore/common/lazy_inline.py +5 -3
  30. mindspore/common/parameter.py +13 -107
  31. mindspore/common/recompute.py +4 -11
  32. mindspore/common/tensor.py +16 -169
  33. mindspore/communication/_comm_helper.py +11 -1
  34. mindspore/communication/comm_func.py +138 -4
  35. mindspore/communication/management.py +85 -1
  36. mindspore/config/op_info.config +0 -15
  37. mindspore/context.py +5 -85
  38. mindspore/dataset/engine/datasets.py +8 -4
  39. mindspore/dataset/engine/datasets_vision.py +1 -1
  40. mindspore/dataset/engine/validators.py +1 -15
  41. mindspore/dnnl.dll +0 -0
  42. mindspore/{experimental/llm_boost/ascend_native → graph}/__init__.py +7 -7
  43. mindspore/graph/custom_pass.py +55 -0
  44. mindspore/include/dataset/execute.h +2 -2
  45. mindspore/jpeg62.dll +0 -0
  46. mindspore/mindrecord/__init__.py +3 -3
  47. mindspore/mindrecord/common/exceptions.py +1 -0
  48. mindspore/mindrecord/config.py +1 -1
  49. mindspore/{parallel/mpi → mindrecord/core}/__init__.py +4 -1
  50. mindspore/mindrecord/{shardheader.py → core/shardheader.py} +2 -1
  51. mindspore/mindrecord/{shardindexgenerator.py → core/shardindexgenerator.py} +1 -1
  52. mindspore/mindrecord/{shardreader.py → core/shardreader.py} +2 -1
  53. mindspore/mindrecord/{shardsegment.py → core/shardsegment.py} +2 -2
  54. mindspore/mindrecord/{shardutils.py → core/shardutils.py} +1 -1
  55. mindspore/mindrecord/{shardwriter.py → core/shardwriter.py} +1 -1
  56. mindspore/mindrecord/filereader.py +4 -4
  57. mindspore/mindrecord/filewriter.py +5 -5
  58. mindspore/mindrecord/mindpage.py +2 -2
  59. mindspore/mindrecord/tools/cifar10.py +1 -1
  60. mindspore/mindrecord/tools/cifar100.py +1 -1
  61. mindspore/mindrecord/tools/cifar100_to_mr.py +1 -1
  62. mindspore/mindrecord/tools/cifar10_to_mr.py +1 -1
  63. mindspore/mindrecord/tools/csv_to_mr.py +1 -1
  64. mindspore/mindrecord/tools/imagenet_to_mr.py +1 -1
  65. mindspore/mindrecord/tools/mnist_to_mr.py +1 -1
  66. mindspore/mindrecord/tools/tfrecord_to_mr.py +1 -1
  67. mindspore/mindspore_backend_common.dll +0 -0
  68. mindspore/mindspore_backend_manager.dll +0 -0
  69. mindspore/mindspore_cluster.dll +0 -0
  70. mindspore/mindspore_common.dll +0 -0
  71. mindspore/mindspore_core.dll +0 -0
  72. mindspore/mindspore_cpu.dll +0 -0
  73. mindspore/mindspore_dump.dll +0 -0
  74. mindspore/mindspore_frontend.dll +0 -0
  75. mindspore/mindspore_glog.dll +0 -0
  76. mindspore/mindspore_hardware_abstract.dll +0 -0
  77. mindspore/mindspore_memory_pool.dll +0 -0
  78. mindspore/mindspore_ms_backend.dll +0 -0
  79. mindspore/mindspore_ops.dll +0 -0
  80. mindspore/{mindspore_ops_host.dll → mindspore_ops_cpu.dll} +0 -0
  81. mindspore/mindspore_profiler.dll +0 -0
  82. mindspore/mindspore_pyboost.dll +0 -0
  83. mindspore/mindspore_pynative.dll +0 -0
  84. mindspore/mindspore_runtime_pipeline.dll +0 -0
  85. mindspore/mindspore_runtime_utils.dll +0 -0
  86. mindspore/mindspore_tools.dll +0 -0
  87. mindspore/mint/__init__.py +15 -10
  88. mindspore/mint/distributed/distributed.py +182 -62
  89. mindspore/mint/nn/__init__.py +2 -16
  90. mindspore/mint/nn/functional.py +4 -110
  91. mindspore/mint/nn/layer/__init__.py +0 -2
  92. mindspore/mint/nn/layer/activation.py +0 -6
  93. mindspore/mint/nn/layer/basic.py +0 -47
  94. mindspore/mint/nn/layer/conv.py +4 -4
  95. mindspore/mint/nn/layer/normalization.py +8 -13
  96. mindspore/mint/nn/layer/pooling.py +0 -4
  97. mindspore/nn/__init__.py +1 -3
  98. mindspore/nn/cell.py +16 -66
  99. mindspore/nn/layer/basic.py +49 -1
  100. mindspore/nn/layer/container.py +16 -0
  101. mindspore/nn/layer/embedding.py +4 -169
  102. mindspore/nn/layer/normalization.py +2 -1
  103. mindspore/nn/layer/thor_layer.py +4 -85
  104. mindspore/nn/optim/ada_grad.py +0 -1
  105. mindspore/nn/optim/adafactor.py +0 -1
  106. mindspore/nn/optim/adam.py +31 -124
  107. mindspore/nn/optim/adamax.py +0 -1
  108. mindspore/nn/optim/asgd.py +0 -1
  109. mindspore/nn/optim/ftrl.py +8 -102
  110. mindspore/nn/optim/lamb.py +0 -1
  111. mindspore/nn/optim/lars.py +0 -3
  112. mindspore/nn/optim/lazyadam.py +25 -218
  113. mindspore/nn/optim/momentum.py +5 -43
  114. mindspore/nn/optim/optimizer.py +6 -55
  115. mindspore/nn/optim/proximal_ada_grad.py +0 -1
  116. mindspore/nn/optim/rmsprop.py +0 -1
  117. mindspore/nn/optim/rprop.py +0 -1
  118. mindspore/nn/optim/sgd.py +0 -1
  119. mindspore/nn/optim/tft_wrapper.py +0 -1
  120. mindspore/nn/optim/thor.py +0 -2
  121. mindspore/nn/probability/bijector/bijector.py +7 -8
  122. mindspore/nn/probability/bijector/gumbel_cdf.py +2 -2
  123. mindspore/nn/probability/bijector/power_transform.py +20 -21
  124. mindspore/nn/probability/bijector/scalar_affine.py +5 -5
  125. mindspore/nn/probability/bijector/softplus.py +13 -14
  126. mindspore/nn/wrap/grad_reducer.py +4 -74
  127. mindspore/numpy/array_creations.py +2 -2
  128. mindspore/numpy/fft.py +9 -9
  129. mindspore/{nn/reinforcement → onnx}/__init__.py +5 -8
  130. mindspore/onnx/onnx_export.py +137 -0
  131. mindspore/opencv_core4110.dll +0 -0
  132. mindspore/opencv_imgcodecs4110.dll +0 -0
  133. mindspore/{opencv_imgproc452.dll → opencv_imgproc4110.dll} +0 -0
  134. mindspore/ops/__init__.py +2 -0
  135. mindspore/ops/_grad_experimental/grad_comm_ops.py +38 -2
  136. mindspore/ops/_op_impl/aicpu/__init__.py +0 -10
  137. mindspore/ops/_op_impl/cpu/__init__.py +0 -5
  138. mindspore/ops/auto_generate/cpp_create_prim_instance_helper.py +16 -22
  139. mindspore/ops/auto_generate/gen_extend_func.py +2 -7
  140. mindspore/ops/auto_generate/gen_ops_def.py +98 -141
  141. mindspore/ops/auto_generate/gen_ops_prim.py +12708 -12686
  142. mindspore/ops/communication.py +97 -0
  143. mindspore/ops/composite/__init__.py +5 -2
  144. mindspore/ops/composite/base.py +15 -1
  145. mindspore/ops/composite/multitype_ops/__init__.py +3 -1
  146. mindspore/ops/composite/multitype_ops/_compile_utils.py +150 -8
  147. mindspore/ops/composite/multitype_ops/add_impl.py +7 -0
  148. mindspore/ops/composite/multitype_ops/mod_impl.py +27 -0
  149. mindspore/ops/function/__init__.py +1 -0
  150. mindspore/ops/function/array_func.py +14 -12
  151. mindspore/ops/function/comm_func.py +3883 -0
  152. mindspore/ops/function/debug_func.py +3 -4
  153. mindspore/ops/function/math_func.py +45 -54
  154. mindspore/ops/function/nn_func.py +75 -294
  155. mindspore/ops/function/random_func.py +9 -18
  156. mindspore/ops/functional.py +2 -0
  157. mindspore/ops/functional_overload.py +354 -18
  158. mindspore/ops/operations/__init__.py +2 -5
  159. mindspore/ops/operations/_custom_ops_utils.py +7 -9
  160. mindspore/ops/operations/_inner_ops.py +1 -38
  161. mindspore/ops/operations/_rl_inner_ops.py +0 -933
  162. mindspore/ops/operations/array_ops.py +1 -0
  163. mindspore/ops/operations/comm_ops.py +94 -2
  164. mindspore/ops/operations/custom_ops.py +228 -19
  165. mindspore/ops/operations/debug_ops.py +27 -29
  166. mindspore/ops/operations/manually_defined/ops_def.py +27 -306
  167. mindspore/ops/operations/nn_ops.py +2 -2
  168. mindspore/ops/operations/sparse_ops.py +0 -83
  169. mindspore/ops/primitive.py +1 -17
  170. mindspore/ops/tensor_method.py +72 -3
  171. mindspore/ops_generate/aclnn/aclnn_kernel_register_auto_cc_generator.py +5 -5
  172. mindspore/ops_generate/aclnn/gen_aclnn_implement.py +8 -8
  173. mindspore/ops_generate/api/functions_cc_generator.py +53 -4
  174. mindspore/ops_generate/api/tensor_func_reg_cpp_generator.py +25 -11
  175. mindspore/ops_generate/common/gen_constants.py +11 -10
  176. mindspore/ops_generate/common/op_proto.py +18 -1
  177. mindspore/ops_generate/common/template.py +102 -245
  178. mindspore/ops_generate/common/template_utils.py +212 -0
  179. mindspore/ops_generate/gen_custom_ops.py +69 -0
  180. mindspore/ops_generate/op_def/ops_def_cc_generator.py +78 -7
  181. mindspore/ops_generate/op_def_py/base_op_prim_py_generator.py +360 -0
  182. mindspore/ops_generate/op_def_py/custom_op_prim_py_generator.py +140 -0
  183. mindspore/ops_generate/op_def_py/op_def_py_generator.py +54 -7
  184. mindspore/ops_generate/op_def_py/op_prim_py_generator.py +5 -312
  185. mindspore/ops_generate/pyboost/auto_grad_impl_cc_generator.py +74 -17
  186. mindspore/ops_generate/pyboost/auto_grad_reg_cc_generator.py +22 -5
  187. mindspore/ops_generate/pyboost/op_template_parser.py +3 -2
  188. mindspore/ops_generate/pyboost/pyboost_functions_cpp_generator.py +21 -5
  189. mindspore/ops_generate/pyboost/pyboost_functions_h_generator.py +2 -2
  190. mindspore/ops_generate/pyboost/pyboost_functions_impl_cpp_generator.py +30 -10
  191. mindspore/ops_generate/pyboost/pyboost_grad_function_cpp_generator.py +10 -3
  192. mindspore/ops_generate/pyboost/pyboost_internal_kernel_info_adapter_generator.py +1 -1
  193. mindspore/ops_generate/pyboost/pyboost_native_grad_functions_generator.py +19 -9
  194. mindspore/ops_generate/pyboost/pyboost_op_cpp_code_generator.py +71 -28
  195. mindspore/ops_generate/pyboost/pyboost_overload_functions_cpp_generator.py +10 -9
  196. mindspore/ops_generate/pyboost/pyboost_utils.py +27 -16
  197. mindspore/ops_generate/resources/yaml_loader.py +13 -0
  198. mindspore/ops_generate/tensor_py_cc_generator.py +2 -2
  199. mindspore/parallel/_cell_wrapper.py +1 -1
  200. mindspore/parallel/_parallel_serialization.py +1 -4
  201. mindspore/parallel/_utils.py +29 -6
  202. mindspore/parallel/checkpoint_transform.py +18 -2
  203. mindspore/parallel/cluster/process_entity/_api.py +24 -32
  204. mindspore/parallel/cluster/process_entity/_utils.py +9 -5
  205. mindspore/{experimental/llm_boost/atb → parallel/distributed}/__init__.py +21 -23
  206. mindspore/parallel/distributed/distributed_data_parallel.py +393 -0
  207. mindspore/parallel/distributed/flatten_grad_buffer.py +295 -0
  208. mindspore/parallel/strategy.py +336 -0
  209. mindspore/parallel/transform_safetensors.py +117 -16
  210. mindspore/profiler/analysis/viewer/ascend_kernel_details_viewer.py +3 -0
  211. mindspore/profiler/analysis/viewer/ms_minddata_viewer.py +1 -1
  212. mindspore/profiler/common/constant.py +5 -0
  213. mindspore/profiler/common/file_manager.py +9 -0
  214. mindspore/profiler/common/msprof_cmd_tool.py +38 -2
  215. mindspore/profiler/common/path_manager.py +56 -24
  216. mindspore/profiler/common/profiler_context.py +2 -12
  217. mindspore/profiler/common/profiler_info.py +3 -3
  218. mindspore/profiler/common/profiler_path_manager.py +13 -0
  219. mindspore/profiler/common/util.py +30 -3
  220. mindspore/profiler/experimental_config.py +2 -1
  221. mindspore/profiler/platform/npu_profiler.py +33 -6
  222. mindspore/run_check/_check_version.py +108 -24
  223. mindspore/runtime/__init__.py +3 -2
  224. mindspore/runtime/executor.py +11 -3
  225. mindspore/runtime/memory.py +112 -0
  226. mindspore/swresample-4.dll +0 -0
  227. mindspore/swscale-6.dll +0 -0
  228. mindspore/tinyxml2.dll +0 -0
  229. mindspore/{experimental/llm_boost → tools}/__init__.py +5 -5
  230. mindspore/tools/data_dump.py +130 -0
  231. mindspore/tools/sdc_detect.py +91 -0
  232. mindspore/tools/stress_detect.py +63 -0
  233. mindspore/train/__init__.py +6 -6
  234. mindspore/train/_utils.py +5 -18
  235. mindspore/train/amp.py +6 -4
  236. mindspore/train/callback/_checkpoint.py +0 -9
  237. mindspore/train/callback/_train_fault_tolerance.py +69 -18
  238. mindspore/train/data_sink.py +1 -5
  239. mindspore/train/model.py +38 -211
  240. mindspore/train/serialization.py +126 -387
  241. mindspore/turbojpeg.dll +0 -0
  242. mindspore/utils/__init__.py +6 -3
  243. mindspore/utils/dlpack.py +92 -0
  244. mindspore/utils/dryrun.py +1 -1
  245. mindspore/utils/runtime_execution_order_check.py +10 -0
  246. mindspore/utils/sdc_detect.py +14 -12
  247. mindspore/utils/stress_detect.py +43 -0
  248. mindspore/utils/utils.py +144 -8
  249. mindspore/version.py +1 -1
  250. {mindspore-2.7.0.dist-info → mindspore-2.7.1.dist-info}/METADATA +3 -2
  251. {mindspore-2.7.0.dist-info → mindspore-2.7.1.dist-info}/RECORD +254 -267
  252. mindspore/experimental/llm_boost/ascend_native/llama_boost_ascend_native.py +0 -210
  253. mindspore/experimental/llm_boost/ascend_native/llm_boost.py +0 -52
  254. mindspore/experimental/llm_boost/atb/boost_base.py +0 -385
  255. mindspore/experimental/llm_boost/atb/llama_boost.py +0 -137
  256. mindspore/experimental/llm_boost/atb/qwen_boost.py +0 -124
  257. mindspore/experimental/llm_boost/register.py +0 -130
  258. mindspore/experimental/llm_boost/utils.py +0 -31
  259. mindspore/include/OWNERS +0 -7
  260. mindspore/mindspore_cpu_res_manager.dll +0 -0
  261. mindspore/mindspore_ops_kernel_common.dll +0 -0
  262. mindspore/mindspore_res_manager.dll +0 -0
  263. mindspore/nn/optim/_dist_optimizer_registry.py +0 -111
  264. mindspore/nn/reinforcement/_batch_read_write.py +0 -142
  265. mindspore/nn/reinforcement/_tensors_queue.py +0 -152
  266. mindspore/nn/reinforcement/tensor_array.py +0 -145
  267. mindspore/opencv_core452.dll +0 -0
  268. mindspore/opencv_imgcodecs452.dll +0 -0
  269. mindspore/ops/_op_impl/aicpu/priority_replay_buffer.py +0 -113
  270. mindspore/ops/_op_impl/aicpu/reservoir_replay_buffer.py +0 -96
  271. mindspore/ops/_op_impl/aicpu/sparse_cross.py +0 -42
  272. mindspore/ops/_op_impl/cpu/buffer_append.py +0 -28
  273. mindspore/ops/_op_impl/cpu/buffer_get.py +0 -28
  274. mindspore/ops/_op_impl/cpu/buffer_sample.py +0 -28
  275. mindspore/ops/_op_impl/cpu/priority_replay_buffer.py +0 -42
  276. mindspore/ops/operations/_tensor_array.py +0 -359
  277. mindspore/ops/operations/rl_ops.py +0 -288
  278. mindspore/parallel/_offload_context.py +0 -275
  279. mindspore/parallel/_recovery_context.py +0 -115
  280. mindspore/parallel/_transformer/__init__.py +0 -35
  281. mindspore/parallel/_transformer/layers.py +0 -765
  282. mindspore/parallel/_transformer/loss.py +0 -251
  283. mindspore/parallel/_transformer/moe.py +0 -693
  284. mindspore/parallel/_transformer/op_parallel_config.py +0 -222
  285. mindspore/parallel/_transformer/transformer.py +0 -3124
  286. mindspore/parallel/mpi/_mpi_config.py +0 -116
  287. mindspore/train/memory_profiling_pb2.py +0 -298
  288. {mindspore-2.7.0.dist-info → mindspore-2.7.1.dist-info}/WHEEL +0 -0
  289. {mindspore-2.7.0.dist-info → mindspore-2.7.1.dist-info}/entry_points.txt +0 -0
  290. {mindspore-2.7.0.dist-info → mindspore-2.7.1.dist-info}/top_level.txt +0 -0
mindspore/context.py CHANGED
@@ -33,13 +33,12 @@ from mindspore.parallel._auto_parallel_context import _set_auto_parallel_context
33
33
  _reset_auto_parallel_context
34
34
  from mindspore.parallel._ps_context import _set_ps_context, _get_ps_context, _reset_ps_context, \
35
35
  _need_reset_device_target_for_ps
36
- from mindspore.parallel._offload_context import _set_offload_context, _get_offload_context
37
36
  from mindspore.hal.device import is_initialized
38
37
  from mindspore.common import api
39
38
 
40
39
  __all__ = ['GRAPH_MODE', 'PYNATIVE_MODE', 'STRICT', 'COMPATIBLE', 'LAX', 'set_context', 'get_context',
41
40
  'set_auto_parallel_context', 'get_auto_parallel_context', 'reset_auto_parallel_context', 'ParallelMode',
42
- 'set_ps_context', 'get_ps_context', 'reset_ps_context', 'set_offload_context', 'get_offload_context']
41
+ 'set_ps_context', 'get_ps_context']
43
42
 
44
43
  GRAPH_MODE = 0
45
44
  PYNATIVE_MODE = 1
@@ -255,22 +254,6 @@ class _Context:
255
254
  f"{exec_order_modes}, but got {exec_order}.")
256
255
  self.set_param(ms_ctx_param.exec_order, exec_order)
257
256
 
258
- def set_memory_offload(self, memory_offload):
259
- """
260
- Enable memory offload or not, support "ON", "OFF".
261
-
262
- Args:
263
- memory_offload (str): "ON", "OFF"
264
- """
265
- memory_offload_options = ["ON", "OFF"]
266
- if memory_offload not in memory_offload_options:
267
- raise ValueError(f"For 'context.set_context', the argument 'memory_offload' must be one of "
268
- f"{memory_offload_options}, but got {memory_offload}.")
269
- if memory_offload == "ON":
270
- self.set_param(ms_ctx_param.memory_offload, True)
271
- else:
272
- self.set_param(ms_ctx_param.memory_offload, False)
273
-
274
257
  def set_deterministic(self, deterministic):
275
258
  """
276
259
  Enable model run in deterministic, and support the values "ON" and "OFF".
@@ -686,7 +669,6 @@ class _Context:
686
669
  'memory_optimize_level': set_memory_optimize_level,
687
670
  'exec_order': set_exec_order,
688
671
  'op_timeout': set_op_timeout,
689
- 'memory_offload': set_memory_offload,
690
672
  'deterministic': set_deterministic,
691
673
  'ascend_config': set_ascend_config,
692
674
  'jit_syntax_level': set_jit_syntax_level,
@@ -1266,65 +1248,6 @@ def reset_auto_parallel_context():
1266
1248
  api.ms_compile_cache.clear()
1267
1249
 
1268
1250
 
1269
- @args_type_check(offload_config=dict)
1270
- def set_offload_context(offload_config):
1271
- r"""
1272
- Configure heterogeneous training detailed parameters to adjust the offload strategy, this api will be deprecated
1273
- and removed in future versions.
1274
-
1275
- Note:
1276
- The offload configuration is only used if the memory offload feature is enabled
1277
- via mindspore.set_context(memory_offload="ON"), and the memory_optimize_level must be set to O0. On the Ascend
1278
- hardware platform, the graph compilation level must be O0.
1279
-
1280
- Args:
1281
- offload_config (dict): A dict contains the keys and values for setting the offload context
1282
- configure.It supports the following keys.
1283
-
1284
- - offload_path (str): The path of offload, relative paths are supported. Default: ``"./offload"``.
1285
- - offload_cpu_size (str): The cpu memory size for offload. The format is "xxGB".
1286
- - offload_disk_size (str): The disk size for offload. The format is "xxGB"
1287
- - hbm_ratio (float): The ratio that can be used based on the maximum device memory.
1288
- The range is (0,1], Default: ``1.0``.
1289
- - cpu_ratio (float): The ratio that can be used based on the maximum host memory.
1290
- The range is (0,1], Default: ``1.0``.
1291
- - enable_pinned_mem (bool): The flag of whether enabling Pinned Memory. Default: ``True``.
1292
- - enable_aio (bool): The flag of whether enabling aio. Default: ``True``.
1293
- - aio_block_size (str): The size of aio block. The format is "xxGB".
1294
- - aio_queue_depth (int): The depth of aio queue.
1295
- - offload_param (str): The param for offload destination, cpu or disk, Default: ``""``.
1296
- - offload_checkpoint (str): The checkpoint for offload destination, only valid if recompute is turned on,
1297
- cpu or disk, Default: ``""``.
1298
- - auto_offload (bool): The flag of whether auto offload. Default: ``True``.
1299
- - host_mem_block_size (str): The memory block size of host memory pool. The format is "xxGB"
1300
-
1301
- Raises:
1302
- ValueError: If input key is not attribute in auto parallel context.
1303
-
1304
- Examples:
1305
- >>> from mindspore import context
1306
- >>> context.set_offload_context(offload_config={"offload_param":"cpu"})
1307
- """
1308
- _set_offload_context(offload_config)
1309
-
1310
-
1311
- def get_offload_context():
1312
- """
1313
- Gets the offload configuration parameters, this api will be deprecated and removed in future versions.
1314
-
1315
- Configure through interface mindspore.set_offload_context(). If the user is not set, the default configuration is
1316
- obtained.
1317
-
1318
- Returns:
1319
- Dict, heterogeneous training offload detailed configuration parameters.
1320
-
1321
- Examples:
1322
- >>> from mindspore import context
1323
- >>> offload_config = context.get_offload_context()
1324
- """
1325
- return _get_offload_context()
1326
-
1327
-
1328
1251
  def _check_target_specific_cfgs(device, arg_key):
1329
1252
  """Checking whether a config is suitable for a specified device"""
1330
1253
  device_cfgs = {
@@ -1400,8 +1323,7 @@ def _check_context_deprecated(key):
1400
1323
  mindspore.device_context.gpu.op_precision.conv_fprop_algo(),
1401
1324
  mindspore.device_context.gpu.op_precision.conv_wgrad_algo(),
1402
1325
  mindspore.device_context.gpu.op_precision.conv_dgrad_algo()''',
1403
- 'runtime_num_threads': 'api mindspore.device_context.cpu.op_tuning.threads_num()',
1404
- 'memory_offload': "`device` parameter of `mindspore.Parameter`"}
1326
+ 'runtime_num_threads': 'api mindspore.device_context.cpu.op_tuning.threads_num()'}
1405
1327
  invalid_context_dict = {
1406
1328
  'exception_dump': {'version': '2.6', 'interface': 'device_context.ascend.op_debug.aclinit_config()'}
1407
1329
  }
@@ -1468,9 +1390,6 @@ def set_context(**kwargs):
1468
1390
  inter_op_parallel_num(int): The thread number of op parallel at the same time.
1469
1391
  Default ``0`` . This parameter will be deprecated and removed in future versions.
1470
1392
  Please use the api :func:`mindspore.runtime.dispatch_threads_num` instead.
1471
- memory_offload (str): Whether to enable the memory offload function. Default ``"OFF"`` .
1472
- This parameter will be deprecated and removed in future versions. Please use the api
1473
- :func:`mindspore.nn.Cell.offload` instead.
1474
1393
  disable_format_transform (bool): Whether to disable the automatic format transform function from NCHW
1475
1394
  to NHWC. Default ``False`` . This parameter will be deprecated and removed in future versions. Please
1476
1395
  use the related parameter of :func:`mindspore.jit` instead.
@@ -1540,7 +1459,9 @@ def set_context(**kwargs):
1540
1459
  This parameter will be deprecated and removed in future versions. Please use the
1541
1460
  api :func:`mindspore.parallel.auto_parallel.AutoParallel.transformer_opt` instead.
1542
1461
  - hccl_watchdog (bool): Enable a thread to monitor the failure of collective communication.
1543
- Default ``True`` .
1462
+ Default ``True`` . This parameter will be deprecated and removed in future versions. Please use the
1463
+ environment variable `MS_ENABLE_THM="{HCCL_WATCHDOG:1}"` instead.
1464
+
1544
1465
  gpu_config (dict): Set the parameters specific to gpu hardware platform. It is not set by default.
1545
1466
 
1546
1467
  - conv_fprop_algo (str): Specifies convolution forward algorithm. Default ``"normal"`` .
@@ -1595,7 +1516,6 @@ def set_context(**kwargs):
1595
1516
  >>> ms.set_context(inter_op_parallel_num=4)
1596
1517
  >>> ms.set_context(disable_format_transform=True)
1597
1518
  >>> ms.set_context(memory_optimize_level='O0')
1598
- >>> ms.set_context(memory_offload='ON')
1599
1519
  >>> ms.set_context(deterministic='ON')
1600
1520
  >>> ms.set_context(ascend_config={"precision_mode": "force_fp16", "jit_compile": True,
1601
1521
  ... "atomic_clean_policy": 1, "op_precision_mode": "./op_precision_config_file",
@@ -457,8 +457,10 @@ class Dataset:
457
457
  each bucket. Must contain len(bucket_boundaries)+1 elements.
458
458
  element_length_function (Callable, optional): A function that takes in
459
459
  M arguments where M = len(column_names) and returns an integer. If no value
460
- provided, parameter M the len(column_names) must be 1, and the size of the first
461
- dimension of that column will be taken as the length. Default: ``None``.
460
+ provided, parameter M the len(column_names) must be 1. At this time, the length of the data in this
461
+ column is determined based on its ndim. If ndim=0, the data length is 0, indicating a str, bool, int,
462
+ or float scalar; if it is an array with ndim > 0, the length of the data is array.shape[0].
463
+ Default: ``None`` , indicating this parameter is not specified.
462
464
  pad_info (dict, optional): The information about how to batch each column. The key
463
465
  corresponds to the column name, and the value must be a tuple of 2 elements.
464
466
  The first element corresponds to the shape to pad to, and the second
@@ -819,8 +821,7 @@ class Dataset:
819
821
  return dataset
820
822
 
821
823
  @check_map
822
- def map(self, operations, input_columns=None, output_columns=None, column_order=None,
823
- num_parallel_workers=None, **kwargs):
824
+ def map(self, operations, input_columns=None, output_columns=None, num_parallel_workers=None, **kwargs):
824
825
  """
825
826
  Apply each operation in operations to this dataset.
826
827
 
@@ -3073,6 +3074,9 @@ def _worker_loop(quit_signal, operations, worker_id, op_type, key, video_backend
3073
3074
  and shared memory. This logic no longer uses the Python multi-process pool, in_queue, and out_queue for
3074
3075
  data transferring.
3075
3076
  """
3077
+ # Release the lock which had been holded in map_op.cc::Launch()/batch_op.cc::Launch()
3078
+ cde.unlock_shm_id_and_msg_id_mutex()
3079
+
3076
3080
  # Initialize C++ side signal handlers
3077
3081
  cde.register_worker_handlers()
3078
3082
 
@@ -2179,7 +2179,7 @@ class ImageFolderDataset(MappableDataset, VisionBaseDataset):
2179
2179
  <https://www.mindspore.cn/docs/en/master/api_python/samples/dataset/dataset_gallery.html>`_
2180
2180
 
2181
2181
  Note:
2182
- - The shape of the image column is [image_size] if `decode` flag is ``False``, or [H,W,C] otherwise.
2182
+ - The shape of the image column is [undecoded_image_size] if `decode` flag is ``False``, or [H,W,C] otherwise.
2183
2183
  - The parameters `num_samples` , `shuffle` , `num_shards` , `shard_id` can be used to control the sampler
2184
2184
  used in the dataset, and their effects when combined with parameter `sampler` are as follows.
2185
2185
 
@@ -1481,23 +1481,9 @@ def check_map(method):
1481
1481
  @wraps(method)
1482
1482
  def new_method(self, *args, **kwargs):
1483
1483
  from mindspore.dataset.callback import DSCallback
1484
- [operations, input_columns, output_columns, column_order, num_parallel_workers, param_dict], _ = \
1484
+ [operations, input_columns, output_columns, num_parallel_workers, param_dict], _ = \
1485
1485
  parse_user_args(method, *args, **kwargs)
1486
1486
 
1487
- if column_order is not None:
1488
- raise ValueError("The parameter 'column_order' had been deleted in map operation. "
1489
- "Please use '.project' operation instead.\n"
1490
- ">> # Usage of old api:\n"
1491
- ">> dataset = dataset.map(operations=PyFunc,\n"
1492
- ">> input_columns=[\"column_a\"],\n"
1493
- ">> output_columns=[\"column_b\", \"column_c\"],\n"
1494
- ">> column_order=[\"column_b\", \"column_c\"])\n"
1495
- ">> # Usage of new api:\n"
1496
- ">> dataset = dataset.map(operations=PyFunc,\n"
1497
- ">> input_columns=[\"column_a\"],\n"
1498
- ">> output_columns=[\"column_b\", \"column_c\"])\n"
1499
- ">> dataset = dataset.project([\"column_b\", \"column_c\"])")
1500
-
1501
1487
  (python_multiprocessing, max_rowsize, cache, callbacks, offload) = get_map_kwargs_from_dict(param_dict)
1502
1488
 
1503
1489
  # check whether network computing operator exist in input operations(python function)
mindspore/dnnl.dll CHANGED
Binary file
@@ -1,4 +1,4 @@
1
- # Copyright 2024 Huawei Technologies Co., Ltd
1
+ # Copyright 2025 Huawei Technologies Co., Ltd
2
2
  #
3
3
  # Licensed under the Apache License, Version 2.0 (the "License");
4
4
  # you may not use this file except in compliance with the License.
@@ -12,11 +12,11 @@
12
12
  # See the License for the specific language governing permissions and
13
13
  # limitations under the License.
14
14
  # ============================================================================
15
- """
16
- Provide llm boost for inference, such as LlamaBoost.
17
- """
18
- from __future__ import absolute_import
19
15
 
20
- from mindspore.experimental.llm_boost.ascend_native.llama_boost_ascend_native import LlamaBoostAscendNative
16
+ """Graph optimization and transformation functionality."""
21
17
 
22
- __all__ = ['LlamaBoostAscendNative']
18
+ from mindspore.graph.custom_pass import register_custom_pass
19
+
20
+ __all__ = [
21
+ "register_custom_pass"
22
+ ]
@@ -0,0 +1,55 @@
1
+ # Copyright 2025 Huawei Technologies Co., Ltd
2
+ #
3
+ # Licensed under the Apache License, Version 2.0 (the "License");
4
+ # you may not use this file except in compliance with the License.
5
+ # You may obtain a copy of the License at
6
+ #
7
+ # http://www.apache.org/licenses/LICENSE-2.0
8
+ #
9
+ # Unless required by applicable law or agreed to in writing, software
10
+ # distributed under the License is distributed on an "AS IS" BASIS,
11
+ # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
12
+ # See the License for the specific language governing permissions and
13
+ # limitations under the License.
14
+ # ============================================================================
15
+
16
+ """
17
+ Experimental custom pass registration functionality.
18
+
19
+ This module provides experimental APIs for registering custom optimization passes.
20
+ These APIs are subject to change and should be used with caution in production code.
21
+ """
22
+
23
+ import mindspore._c_expression as _c_expression
24
+
25
+
26
+ def register_custom_pass(pass_name: str, plugin_so_path: str, device: str = "all", stage: str = "") -> bool:
27
+ """Register a custom pass plugin.
28
+
29
+ .. warning::
30
+ This is an experimental API that is subject to change or deletion.
31
+
32
+ Args:
33
+ pass_name (str): Name of the pass expected to be provided by the plugin.
34
+ plugin_so_path (str): Absolute path to the plugin shared library (.so file).
35
+ device (str): Target device for the pass. Supported values: "cpu", "gpu", "ascend", or "all".
36
+ Default: "all".
37
+ stage (str): Pass stage. Reserved field for future use. Default: "".
38
+
39
+ Returns:
40
+ bool: True if plugin loaded successfully, False otherwise.
41
+
42
+ Examples:
43
+ >>> import mindspore.graph as graph
44
+ >>> # Register a custom optimization pass
45
+ >>> success = graph.register_custom_pass(
46
+ ... pass_name="my_fusion_pass",
47
+ ... plugin_so_path="/path/to/my_plugin.so",
48
+ ... device="ascend"
49
+ ... )
50
+ >>> print(f"Registration successful: {success}")
51
+ """
52
+ return _c_expression.register_custom_pass(pass_name, plugin_so_path, device, stage)
53
+
54
+
55
+ __all__ = ["register_custom_pass"]
@@ -27,8 +27,8 @@
27
27
  #include "include/dataset/constants.h"
28
28
  #include "include/dataset/transforms.h"
29
29
  #if defined(ENABLE_D)
30
- #include "runtime/hardware/device_context.h"
31
- #include "runtime/hardware/device_context_manager.h"
30
+ #include "runtime/hardware_abstract/device_context/device_context.h"
31
+ #include "runtime/hardware_abstract/device_context/device_context_manager.h"
32
32
  #endif
33
33
 
34
34
  namespace mindspore {
mindspore/jpeg62.dll CHANGED
Binary file
@@ -31,8 +31,8 @@ datasets using `mindspore.dataset.MindDataset <https://www.mindspore.cn/docs/en/
31
31
  mindspore.dataset.MindDataset.html>`_ .
32
32
 
33
33
  Users can also convert datasets from other formats to the MindRecord format.
34
- For more details, please refer to `Converting Dataset to MindRecord <https://www.mindspore.cn/docs/en/master/
35
- model_train/dataset/record.html>`_ .
34
+ For more details, please refer to
35
+ `Converting Dataset to MindRecord <https://www.mindspore.cn/tutorials/en/master/dataset/record.html>`_ .
36
36
  Additionally, MindRecord supports file encryption, decryption,
37
37
  and integrity checks to ensure the security of MindRecord format datasets.
38
38
  """
@@ -41,7 +41,7 @@ from .filewriter import FileWriter
41
41
  from .filereader import FileReader
42
42
  from .mindpage import MindPage
43
43
  from .common.exceptions import *
44
- from .shardutils import SUCCESS, FAILED
44
+ from .core.shardutils import SUCCESS, FAILED
45
45
  from .tools.cifar10_to_mr import Cifar10ToMR
46
46
  from .tools.cifar100_to_mr import Cifar100ToMR
47
47
  from .tools.csv_to_mr import CsvToMR
@@ -26,6 +26,7 @@ Error rule:
26
26
  """
27
27
  from .enums import LogRuntime, ErrorCodeType, ErrorLevel
28
28
  from .constant import SYS_ID
29
+
29
30
  EXCEPTIONS = dict(
30
31
  # the format of list is [exception No, base error message]
31
32
  UnknownError=[0, 'Unknown Error.'],
@@ -30,7 +30,7 @@ import time
30
30
 
31
31
  from mindspore import log as logger
32
32
  from mindspore._c_expression import _encrypt, _decrypt_data
33
- from .shardutils import MIN_FILE_SIZE
33
+ from .core.shardutils import MIN_FILE_SIZE
34
34
 
35
35
 
36
36
  __all__ = ['set_enc_key',
@@ -1,4 +1,4 @@
1
- # Copyright 2023 Huawei Technologies Co., Ltd
1
+ # Copyright 2025 Huawei Technologies Co., Ltd
2
2
  #
3
3
  # Licensed under the Apache License, Version 2.0 (the "License");
4
4
  # you may not use this file except in compliance with the License.
@@ -12,3 +12,6 @@
12
12
  # See the License for the specific language governing permissions and
13
13
  # limitations under the License.
14
14
  # ============================================================================
15
+ """
16
+ init core
17
+ """
@@ -17,7 +17,8 @@ This module is to write data into mindrecord.
17
17
  """
18
18
  import mindspore._c_mindrecord as ms
19
19
  from mindspore import log as logger
20
- from .common.exceptions import MRMAddSchemaError, MRMAddIndexError, MRMBuildSchemaError, MRMGetMetaError
20
+ from mindspore.mindrecord.common.exceptions import MRMAddSchemaError, MRMAddIndexError, MRMBuildSchemaError, \
21
+ MRMGetMetaError
21
22
 
22
23
  __all__ = ['ShardHeader']
23
24
 
@@ -17,7 +17,7 @@ This module is to write data into mindrecord.
17
17
  """
18
18
  import mindspore._c_mindrecord as ms
19
19
  from mindspore import log as logger
20
- from .common.exceptions import MRMIndexGeneratorError, MRMGenerateIndexError
20
+ from mindspore.mindrecord.common.exceptions import MRMIndexGeneratorError, MRMGenerateIndexError
21
21
 
22
22
  __all__ = ['ShardIndexGenerator']
23
23
 
@@ -17,7 +17,8 @@ This module is to read data from mindrecord.
17
17
  """
18
18
  import mindspore._c_mindrecord as ms
19
19
  from mindspore import log as logger
20
- from .common.exceptions import MRMOpenError, MRMLaunchError
20
+ from mindspore.mindrecord.common.exceptions import MRMOpenError, MRMLaunchError
21
+
21
22
  __all__ = ['ShardReader']
22
23
 
23
24
 
@@ -17,8 +17,8 @@ This module is to read page from mindrecord.
17
17
  """
18
18
  import mindspore._c_mindrecord as ms
19
19
  from mindspore import log as logger
20
- from .shardutils import populate_data, SUCCESS
21
- from .shardheader import ShardHeader
20
+ from mindspore.mindrecord.core.shardutils import populate_data, SUCCESS
21
+ from mindspore.mindrecord.core.shardheader import ShardHeader
22
22
 
23
23
  __all__ = ['ShardSegment']
24
24
 
@@ -25,7 +25,7 @@ from functools import wraps
25
25
 
26
26
  import numpy as np
27
27
  import mindspore._c_mindrecord as ms
28
- from .common.exceptions import ParamValueError, MRMUnsupportedSchemaError
28
+ from mindspore.mindrecord.common.exceptions import ParamValueError, MRMUnsupportedSchemaError
29
29
 
30
30
  SUCCESS = ms.MSRStatus.SUCCESS
31
31
  FAILED = ms.MSRStatus.FAILED
@@ -18,7 +18,7 @@ This module is to write data into mindrecord.
18
18
  import numpy as np
19
19
  import mindspore._c_mindrecord as ms
20
20
  from mindspore import log as logger
21
- from .common.exceptions import MRMOpenError, MRMOpenForAppendError, MRMInvalidHeaderSizeError, \
21
+ from mindspore.mindrecord.common.exceptions import MRMOpenError, MRMOpenForAppendError, MRMInvalidHeaderSizeError, \
22
22
  MRMInvalidPageSizeError, MRMSetHeaderError, MRMWriteDatasetError, MRMCommitError
23
23
 
24
24
  __all__ = ['ShardWriter']
@@ -18,10 +18,10 @@ This module is to read data from MindRecord.
18
18
  import platform
19
19
  from mindspore import log as logger
20
20
 
21
- from .shardreader import ShardReader
22
- from .shardheader import ShardHeader
23
- from .shardutils import populate_data
24
- from .shardutils import check_parameter
21
+ from .core.shardreader import ShardReader
22
+ from .core.shardheader import ShardHeader
23
+ from .core.shardutils import populate_data
24
+ from .core.shardutils import check_parameter
25
25
  from .common.exceptions import ParamTypeError
26
26
  from .config import _get_enc_key, _get_dec_mode, decrypt
27
27
 
@@ -25,11 +25,11 @@ import time
25
25
  import multiprocessing as mp
26
26
  import numpy as np
27
27
  from mindspore import log as logger
28
- from .shardwriter import ShardWriter
29
- from .shardreader import ShardReader
30
- from .shardheader import ShardHeader
31
- from .shardindexgenerator import ShardIndexGenerator
32
- from .shardutils import MIN_SHARD_COUNT, MAX_SHARD_COUNT, VALID_ATTRIBUTES, VALID_ARRAY_ATTRIBUTES, \
28
+ from .core.shardwriter import ShardWriter
29
+ from .core.shardreader import ShardReader
30
+ from .core.shardheader import ShardHeader
31
+ from .core.shardindexgenerator import ShardIndexGenerator
32
+ from .core.shardutils import MIN_SHARD_COUNT, MAX_SHARD_COUNT, VALID_ATTRIBUTES, VALID_ARRAY_ATTRIBUTES, \
33
33
  check_filename, VALUE_TYPE_MAP, SUCCESS
34
34
  from .common.exceptions import ParamValueError, ParamTypeError, MRMInvalidSchemaError, MRMDefineIndexError
35
35
  from .config import _get_enc_key, _get_enc_mode, _get_dec_mode, encrypt, decrypt
@@ -16,8 +16,8 @@
16
16
  This module is to support reading page from MindRecord.
17
17
  """
18
18
 
19
- from .shardsegment import ShardSegment
20
- from .shardutils import check_parameter
19
+ from .core.shardsegment import ShardSegment
20
+ from .core.shardutils import check_parameter
21
21
  from .common.exceptions import ParamValueError, ParamTypeError, MRMDefineCategoryError
22
22
  from .config import _get_enc_key, _get_dec_mode, decrypt
23
23
 
@@ -23,7 +23,7 @@ import re
23
23
  import numpy as np
24
24
 
25
25
  from mindspore import log as logger
26
- from ..shardutils import check_filename
26
+ from ..core.shardutils import check_filename
27
27
 
28
28
  __all__ = ['Cifar10']
29
29
 
@@ -21,7 +21,7 @@ import pickle
21
21
  import os
22
22
  import numpy as np
23
23
 
24
- from ..shardutils import check_filename
24
+ from ..core.shardutils import check_filename
25
25
 
26
26
  __all__ = ['Cifar100']
27
27
 
@@ -23,7 +23,7 @@ from mindspore import log as logger
23
23
  from .cifar100 import Cifar100
24
24
  from ..common.exceptions import PathNotExistsError
25
25
  from ..filewriter import FileWriter
26
- from ..shardutils import check_filename, ExceptionThread, SUCCESS
26
+ from ..core.shardutils import check_filename, ExceptionThread, SUCCESS
27
27
 
28
28
 
29
29
  __all__ = ['Cifar100ToMR']
@@ -23,7 +23,7 @@ from mindspore import log as logger
23
23
  from .cifar10 import Cifar10
24
24
  from ..common.exceptions import PathNotExistsError
25
25
  from ..filewriter import FileWriter
26
- from ..shardutils import check_filename, ExceptionThread, SUCCESS
26
+ from ..core.shardutils import check_filename, ExceptionThread, SUCCESS
27
27
 
28
28
 
29
29
  __all__ = ['Cifar10ToMR']
@@ -20,7 +20,7 @@ import os
20
20
 
21
21
  from mindspore import log as logger
22
22
  from ..filewriter import FileWriter
23
- from ..shardutils import check_filename, ExceptionThread
23
+ from ..core.shardutils import check_filename, ExceptionThread
24
24
 
25
25
 
26
26
  __all__ = ['CsvToMR']
@@ -21,7 +21,7 @@ import time
21
21
  from mindspore import log as logger
22
22
  from ..common.exceptions import PathNotExistsError
23
23
  from ..filewriter import FileWriter
24
- from ..shardutils import check_filename, ExceptionThread
24
+ from ..core.shardutils import check_filename, ExceptionThread
25
25
 
26
26
  __all__ = ['ImageNetToMR']
27
27
 
@@ -23,7 +23,7 @@ import numpy as np
23
23
 
24
24
  from mindspore import log as logger
25
25
  from ..filewriter import FileWriter
26
- from ..shardutils import check_filename, ExceptionThread, SUCCESS, FAILED
26
+ from ..core.shardutils import check_filename, ExceptionThread, SUCCESS, FAILED
27
27
 
28
28
 
29
29
  __all__ = ['MnistToMR']
@@ -21,7 +21,7 @@ import numpy as np
21
21
 
22
22
  from mindspore import log as logger
23
23
  from ..filewriter import FileWriter
24
- from ..shardutils import check_filename, ExceptionThread
24
+ from ..core.shardutils import check_filename, ExceptionThread
25
25
 
26
26
  __all__ = ['TFRecordToMR']
27
27
 
Binary file
Binary file
Binary file
Binary file
Binary file
Binary file
Binary file
Binary file
Binary file
Binary file
Binary file
Binary file
Binary file
Binary file
Binary file
Binary file
Binary file
Binary file