mindspore 2.7.0__cp310-cp310-win_amd64.whl → 2.7.1__cp310-cp310-win_amd64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of mindspore might be problematic. Click here for more details.
- mindspore/.commit_id +1 -1
- mindspore/__init__.py +4 -1
- mindspore/_c_dataengine.cp310-win_amd64.pyd +0 -0
- mindspore/_c_expression.cp310-win_amd64.pyd +0 -0
- mindspore/_c_mindrecord.cp310-win_amd64.pyd +0 -0
- mindspore/_extends/parse/compile_config.py +24 -1
- mindspore/_extends/parse/deprecated/deprecated_tensor_method.py +6 -2
- mindspore/_extends/parse/resources.py +1 -1
- mindspore/_extends/parse/standard_method.py +8 -1
- mindspore/_extends/parse/trope.py +2 -1
- mindspore/_extends/pijit/pijit_func_white_list.py +7 -22
- mindspore/avcodec-59.dll +0 -0
- mindspore/avdevice-59.dll +0 -0
- mindspore/avfilter-8.dll +0 -0
- mindspore/avformat-59.dll +0 -0
- mindspore/avutil-57.dll +0 -0
- mindspore/boost/base.py +29 -2
- mindspore/common/_decorator.py +3 -2
- mindspore/common/_grad_function.py +3 -1
- mindspore/common/_tensor_cpp_method.py +1 -1
- mindspore/common/_tensor_docs.py +275 -64
- mindspore/common/_utils.py +0 -44
- mindspore/common/api.py +285 -35
- mindspore/common/dump.py +7 -108
- mindspore/common/dynamic_shape/auto_dynamic_shape.py +1 -3
- mindspore/common/hook_handle.py +60 -0
- mindspore/common/jit_config.py +5 -1
- mindspore/common/jit_trace.py +27 -12
- mindspore/common/lazy_inline.py +5 -3
- mindspore/common/parameter.py +13 -107
- mindspore/common/recompute.py +4 -11
- mindspore/common/tensor.py +16 -169
- mindspore/communication/_comm_helper.py +11 -1
- mindspore/communication/comm_func.py +138 -4
- mindspore/communication/management.py +85 -1
- mindspore/config/op_info.config +0 -15
- mindspore/context.py +5 -85
- mindspore/dataset/engine/datasets.py +8 -4
- mindspore/dataset/engine/datasets_vision.py +1 -1
- mindspore/dataset/engine/validators.py +1 -15
- mindspore/dnnl.dll +0 -0
- mindspore/{experimental/llm_boost/ascend_native → graph}/__init__.py +7 -7
- mindspore/graph/custom_pass.py +55 -0
- mindspore/include/dataset/execute.h +2 -2
- mindspore/jpeg62.dll +0 -0
- mindspore/mindrecord/__init__.py +3 -3
- mindspore/mindrecord/common/exceptions.py +1 -0
- mindspore/mindrecord/config.py +1 -1
- mindspore/{parallel/mpi → mindrecord/core}/__init__.py +4 -1
- mindspore/mindrecord/{shardheader.py → core/shardheader.py} +2 -1
- mindspore/mindrecord/{shardindexgenerator.py → core/shardindexgenerator.py} +1 -1
- mindspore/mindrecord/{shardreader.py → core/shardreader.py} +2 -1
- mindspore/mindrecord/{shardsegment.py → core/shardsegment.py} +2 -2
- mindspore/mindrecord/{shardutils.py → core/shardutils.py} +1 -1
- mindspore/mindrecord/{shardwriter.py → core/shardwriter.py} +1 -1
- mindspore/mindrecord/filereader.py +4 -4
- mindspore/mindrecord/filewriter.py +5 -5
- mindspore/mindrecord/mindpage.py +2 -2
- mindspore/mindrecord/tools/cifar10.py +1 -1
- mindspore/mindrecord/tools/cifar100.py +1 -1
- mindspore/mindrecord/tools/cifar100_to_mr.py +1 -1
- mindspore/mindrecord/tools/cifar10_to_mr.py +1 -1
- mindspore/mindrecord/tools/csv_to_mr.py +1 -1
- mindspore/mindrecord/tools/imagenet_to_mr.py +1 -1
- mindspore/mindrecord/tools/mnist_to_mr.py +1 -1
- mindspore/mindrecord/tools/tfrecord_to_mr.py +1 -1
- mindspore/mindspore_backend_common.dll +0 -0
- mindspore/mindspore_backend_manager.dll +0 -0
- mindspore/mindspore_cluster.dll +0 -0
- mindspore/mindspore_common.dll +0 -0
- mindspore/mindspore_core.dll +0 -0
- mindspore/mindspore_cpu.dll +0 -0
- mindspore/mindspore_dump.dll +0 -0
- mindspore/mindspore_frontend.dll +0 -0
- mindspore/mindspore_glog.dll +0 -0
- mindspore/mindspore_hardware_abstract.dll +0 -0
- mindspore/mindspore_memory_pool.dll +0 -0
- mindspore/mindspore_ms_backend.dll +0 -0
- mindspore/mindspore_ops.dll +0 -0
- mindspore/{mindspore_ops_host.dll → mindspore_ops_cpu.dll} +0 -0
- mindspore/mindspore_profiler.dll +0 -0
- mindspore/mindspore_pyboost.dll +0 -0
- mindspore/mindspore_pynative.dll +0 -0
- mindspore/mindspore_runtime_pipeline.dll +0 -0
- mindspore/mindspore_runtime_utils.dll +0 -0
- mindspore/mindspore_tools.dll +0 -0
- mindspore/mint/__init__.py +15 -10
- mindspore/mint/distributed/distributed.py +182 -62
- mindspore/mint/nn/__init__.py +2 -16
- mindspore/mint/nn/functional.py +4 -110
- mindspore/mint/nn/layer/__init__.py +0 -2
- mindspore/mint/nn/layer/activation.py +0 -6
- mindspore/mint/nn/layer/basic.py +0 -47
- mindspore/mint/nn/layer/conv.py +4 -4
- mindspore/mint/nn/layer/normalization.py +8 -13
- mindspore/mint/nn/layer/pooling.py +0 -4
- mindspore/nn/__init__.py +1 -3
- mindspore/nn/cell.py +16 -66
- mindspore/nn/layer/basic.py +49 -1
- mindspore/nn/layer/container.py +16 -0
- mindspore/nn/layer/embedding.py +4 -169
- mindspore/nn/layer/normalization.py +2 -1
- mindspore/nn/layer/thor_layer.py +4 -85
- mindspore/nn/optim/ada_grad.py +0 -1
- mindspore/nn/optim/adafactor.py +0 -1
- mindspore/nn/optim/adam.py +31 -124
- mindspore/nn/optim/adamax.py +0 -1
- mindspore/nn/optim/asgd.py +0 -1
- mindspore/nn/optim/ftrl.py +8 -102
- mindspore/nn/optim/lamb.py +0 -1
- mindspore/nn/optim/lars.py +0 -3
- mindspore/nn/optim/lazyadam.py +25 -218
- mindspore/nn/optim/momentum.py +5 -43
- mindspore/nn/optim/optimizer.py +6 -55
- mindspore/nn/optim/proximal_ada_grad.py +0 -1
- mindspore/nn/optim/rmsprop.py +0 -1
- mindspore/nn/optim/rprop.py +0 -1
- mindspore/nn/optim/sgd.py +0 -1
- mindspore/nn/optim/tft_wrapper.py +0 -1
- mindspore/nn/optim/thor.py +0 -2
- mindspore/nn/probability/bijector/bijector.py +7 -8
- mindspore/nn/probability/bijector/gumbel_cdf.py +2 -2
- mindspore/nn/probability/bijector/power_transform.py +20 -21
- mindspore/nn/probability/bijector/scalar_affine.py +5 -5
- mindspore/nn/probability/bijector/softplus.py +13 -14
- mindspore/nn/wrap/grad_reducer.py +4 -74
- mindspore/numpy/array_creations.py +2 -2
- mindspore/numpy/fft.py +9 -9
- mindspore/{nn/reinforcement → onnx}/__init__.py +5 -8
- mindspore/onnx/onnx_export.py +137 -0
- mindspore/opencv_core4110.dll +0 -0
- mindspore/opencv_imgcodecs4110.dll +0 -0
- mindspore/{opencv_imgproc452.dll → opencv_imgproc4110.dll} +0 -0
- mindspore/ops/__init__.py +2 -0
- mindspore/ops/_grad_experimental/grad_comm_ops.py +38 -2
- mindspore/ops/_op_impl/aicpu/__init__.py +0 -10
- mindspore/ops/_op_impl/cpu/__init__.py +0 -5
- mindspore/ops/auto_generate/cpp_create_prim_instance_helper.py +16 -22
- mindspore/ops/auto_generate/gen_extend_func.py +2 -7
- mindspore/ops/auto_generate/gen_ops_def.py +98 -141
- mindspore/ops/auto_generate/gen_ops_prim.py +12708 -12686
- mindspore/ops/communication.py +97 -0
- mindspore/ops/composite/__init__.py +5 -2
- mindspore/ops/composite/base.py +15 -1
- mindspore/ops/composite/multitype_ops/__init__.py +3 -1
- mindspore/ops/composite/multitype_ops/_compile_utils.py +150 -8
- mindspore/ops/composite/multitype_ops/add_impl.py +7 -0
- mindspore/ops/composite/multitype_ops/mod_impl.py +27 -0
- mindspore/ops/function/__init__.py +1 -0
- mindspore/ops/function/array_func.py +14 -12
- mindspore/ops/function/comm_func.py +3883 -0
- mindspore/ops/function/debug_func.py +3 -4
- mindspore/ops/function/math_func.py +45 -54
- mindspore/ops/function/nn_func.py +75 -294
- mindspore/ops/function/random_func.py +9 -18
- mindspore/ops/functional.py +2 -0
- mindspore/ops/functional_overload.py +354 -18
- mindspore/ops/operations/__init__.py +2 -5
- mindspore/ops/operations/_custom_ops_utils.py +7 -9
- mindspore/ops/operations/_inner_ops.py +1 -38
- mindspore/ops/operations/_rl_inner_ops.py +0 -933
- mindspore/ops/operations/array_ops.py +1 -0
- mindspore/ops/operations/comm_ops.py +94 -2
- mindspore/ops/operations/custom_ops.py +228 -19
- mindspore/ops/operations/debug_ops.py +27 -29
- mindspore/ops/operations/manually_defined/ops_def.py +27 -306
- mindspore/ops/operations/nn_ops.py +2 -2
- mindspore/ops/operations/sparse_ops.py +0 -83
- mindspore/ops/primitive.py +1 -17
- mindspore/ops/tensor_method.py +72 -3
- mindspore/ops_generate/aclnn/aclnn_kernel_register_auto_cc_generator.py +5 -5
- mindspore/ops_generate/aclnn/gen_aclnn_implement.py +8 -8
- mindspore/ops_generate/api/functions_cc_generator.py +53 -4
- mindspore/ops_generate/api/tensor_func_reg_cpp_generator.py +25 -11
- mindspore/ops_generate/common/gen_constants.py +11 -10
- mindspore/ops_generate/common/op_proto.py +18 -1
- mindspore/ops_generate/common/template.py +102 -245
- mindspore/ops_generate/common/template_utils.py +212 -0
- mindspore/ops_generate/gen_custom_ops.py +69 -0
- mindspore/ops_generate/op_def/ops_def_cc_generator.py +78 -7
- mindspore/ops_generate/op_def_py/base_op_prim_py_generator.py +360 -0
- mindspore/ops_generate/op_def_py/custom_op_prim_py_generator.py +140 -0
- mindspore/ops_generate/op_def_py/op_def_py_generator.py +54 -7
- mindspore/ops_generate/op_def_py/op_prim_py_generator.py +5 -312
- mindspore/ops_generate/pyboost/auto_grad_impl_cc_generator.py +74 -17
- mindspore/ops_generate/pyboost/auto_grad_reg_cc_generator.py +22 -5
- mindspore/ops_generate/pyboost/op_template_parser.py +3 -2
- mindspore/ops_generate/pyboost/pyboost_functions_cpp_generator.py +21 -5
- mindspore/ops_generate/pyboost/pyboost_functions_h_generator.py +2 -2
- mindspore/ops_generate/pyboost/pyboost_functions_impl_cpp_generator.py +30 -10
- mindspore/ops_generate/pyboost/pyboost_grad_function_cpp_generator.py +10 -3
- mindspore/ops_generate/pyboost/pyboost_internal_kernel_info_adapter_generator.py +1 -1
- mindspore/ops_generate/pyboost/pyboost_native_grad_functions_generator.py +19 -9
- mindspore/ops_generate/pyboost/pyboost_op_cpp_code_generator.py +71 -28
- mindspore/ops_generate/pyboost/pyboost_overload_functions_cpp_generator.py +10 -9
- mindspore/ops_generate/pyboost/pyboost_utils.py +27 -16
- mindspore/ops_generate/resources/yaml_loader.py +13 -0
- mindspore/ops_generate/tensor_py_cc_generator.py +2 -2
- mindspore/parallel/_cell_wrapper.py +1 -1
- mindspore/parallel/_parallel_serialization.py +1 -4
- mindspore/parallel/_utils.py +29 -6
- mindspore/parallel/checkpoint_transform.py +18 -2
- mindspore/parallel/cluster/process_entity/_api.py +24 -32
- mindspore/parallel/cluster/process_entity/_utils.py +9 -5
- mindspore/{experimental/llm_boost/atb → parallel/distributed}/__init__.py +21 -23
- mindspore/parallel/distributed/distributed_data_parallel.py +393 -0
- mindspore/parallel/distributed/flatten_grad_buffer.py +295 -0
- mindspore/parallel/strategy.py +336 -0
- mindspore/parallel/transform_safetensors.py +117 -16
- mindspore/profiler/analysis/viewer/ascend_kernel_details_viewer.py +3 -0
- mindspore/profiler/analysis/viewer/ms_minddata_viewer.py +1 -1
- mindspore/profiler/common/constant.py +5 -0
- mindspore/profiler/common/file_manager.py +9 -0
- mindspore/profiler/common/msprof_cmd_tool.py +38 -2
- mindspore/profiler/common/path_manager.py +56 -24
- mindspore/profiler/common/profiler_context.py +2 -12
- mindspore/profiler/common/profiler_info.py +3 -3
- mindspore/profiler/common/profiler_path_manager.py +13 -0
- mindspore/profiler/common/util.py +30 -3
- mindspore/profiler/experimental_config.py +2 -1
- mindspore/profiler/platform/npu_profiler.py +33 -6
- mindspore/run_check/_check_version.py +108 -24
- mindspore/runtime/__init__.py +3 -2
- mindspore/runtime/executor.py +11 -3
- mindspore/runtime/memory.py +112 -0
- mindspore/swresample-4.dll +0 -0
- mindspore/swscale-6.dll +0 -0
- mindspore/tinyxml2.dll +0 -0
- mindspore/{experimental/llm_boost → tools}/__init__.py +5 -5
- mindspore/tools/data_dump.py +130 -0
- mindspore/tools/sdc_detect.py +91 -0
- mindspore/tools/stress_detect.py +63 -0
- mindspore/train/__init__.py +6 -6
- mindspore/train/_utils.py +5 -18
- mindspore/train/amp.py +6 -4
- mindspore/train/callback/_checkpoint.py +0 -9
- mindspore/train/callback/_train_fault_tolerance.py +69 -18
- mindspore/train/data_sink.py +1 -5
- mindspore/train/model.py +38 -211
- mindspore/train/serialization.py +126 -387
- mindspore/turbojpeg.dll +0 -0
- mindspore/utils/__init__.py +6 -3
- mindspore/utils/dlpack.py +92 -0
- mindspore/utils/dryrun.py +1 -1
- mindspore/utils/runtime_execution_order_check.py +10 -0
- mindspore/utils/sdc_detect.py +14 -12
- mindspore/utils/stress_detect.py +43 -0
- mindspore/utils/utils.py +144 -8
- mindspore/version.py +1 -1
- {mindspore-2.7.0.dist-info → mindspore-2.7.1.dist-info}/METADATA +3 -2
- {mindspore-2.7.0.dist-info → mindspore-2.7.1.dist-info}/RECORD +254 -267
- mindspore/experimental/llm_boost/ascend_native/llama_boost_ascend_native.py +0 -210
- mindspore/experimental/llm_boost/ascend_native/llm_boost.py +0 -52
- mindspore/experimental/llm_boost/atb/boost_base.py +0 -385
- mindspore/experimental/llm_boost/atb/llama_boost.py +0 -137
- mindspore/experimental/llm_boost/atb/qwen_boost.py +0 -124
- mindspore/experimental/llm_boost/register.py +0 -130
- mindspore/experimental/llm_boost/utils.py +0 -31
- mindspore/include/OWNERS +0 -7
- mindspore/mindspore_cpu_res_manager.dll +0 -0
- mindspore/mindspore_ops_kernel_common.dll +0 -0
- mindspore/mindspore_res_manager.dll +0 -0
- mindspore/nn/optim/_dist_optimizer_registry.py +0 -111
- mindspore/nn/reinforcement/_batch_read_write.py +0 -142
- mindspore/nn/reinforcement/_tensors_queue.py +0 -152
- mindspore/nn/reinforcement/tensor_array.py +0 -145
- mindspore/opencv_core452.dll +0 -0
- mindspore/opencv_imgcodecs452.dll +0 -0
- mindspore/ops/_op_impl/aicpu/priority_replay_buffer.py +0 -113
- mindspore/ops/_op_impl/aicpu/reservoir_replay_buffer.py +0 -96
- mindspore/ops/_op_impl/aicpu/sparse_cross.py +0 -42
- mindspore/ops/_op_impl/cpu/buffer_append.py +0 -28
- mindspore/ops/_op_impl/cpu/buffer_get.py +0 -28
- mindspore/ops/_op_impl/cpu/buffer_sample.py +0 -28
- mindspore/ops/_op_impl/cpu/priority_replay_buffer.py +0 -42
- mindspore/ops/operations/_tensor_array.py +0 -359
- mindspore/ops/operations/rl_ops.py +0 -288
- mindspore/parallel/_offload_context.py +0 -275
- mindspore/parallel/_recovery_context.py +0 -115
- mindspore/parallel/_transformer/__init__.py +0 -35
- mindspore/parallel/_transformer/layers.py +0 -765
- mindspore/parallel/_transformer/loss.py +0 -251
- mindspore/parallel/_transformer/moe.py +0 -693
- mindspore/parallel/_transformer/op_parallel_config.py +0 -222
- mindspore/parallel/_transformer/transformer.py +0 -3124
- mindspore/parallel/mpi/_mpi_config.py +0 -116
- mindspore/train/memory_profiling_pb2.py +0 -298
- {mindspore-2.7.0.dist-info → mindspore-2.7.1.dist-info}/WHEEL +0 -0
- {mindspore-2.7.0.dist-info → mindspore-2.7.1.dist-info}/entry_points.txt +0 -0
- {mindspore-2.7.0.dist-info → mindspore-2.7.1.dist-info}/top_level.txt +0 -0
mindspore/context.py
CHANGED
|
@@ -33,13 +33,12 @@ from mindspore.parallel._auto_parallel_context import _set_auto_parallel_context
|
|
|
33
33
|
_reset_auto_parallel_context
|
|
34
34
|
from mindspore.parallel._ps_context import _set_ps_context, _get_ps_context, _reset_ps_context, \
|
|
35
35
|
_need_reset_device_target_for_ps
|
|
36
|
-
from mindspore.parallel._offload_context import _set_offload_context, _get_offload_context
|
|
37
36
|
from mindspore.hal.device import is_initialized
|
|
38
37
|
from mindspore.common import api
|
|
39
38
|
|
|
40
39
|
__all__ = ['GRAPH_MODE', 'PYNATIVE_MODE', 'STRICT', 'COMPATIBLE', 'LAX', 'set_context', 'get_context',
|
|
41
40
|
'set_auto_parallel_context', 'get_auto_parallel_context', 'reset_auto_parallel_context', 'ParallelMode',
|
|
42
|
-
'set_ps_context', 'get_ps_context'
|
|
41
|
+
'set_ps_context', 'get_ps_context']
|
|
43
42
|
|
|
44
43
|
GRAPH_MODE = 0
|
|
45
44
|
PYNATIVE_MODE = 1
|
|
@@ -255,22 +254,6 @@ class _Context:
|
|
|
255
254
|
f"{exec_order_modes}, but got {exec_order}.")
|
|
256
255
|
self.set_param(ms_ctx_param.exec_order, exec_order)
|
|
257
256
|
|
|
258
|
-
def set_memory_offload(self, memory_offload):
|
|
259
|
-
"""
|
|
260
|
-
Enable memory offload or not, support "ON", "OFF".
|
|
261
|
-
|
|
262
|
-
Args:
|
|
263
|
-
memory_offload (str): "ON", "OFF"
|
|
264
|
-
"""
|
|
265
|
-
memory_offload_options = ["ON", "OFF"]
|
|
266
|
-
if memory_offload not in memory_offload_options:
|
|
267
|
-
raise ValueError(f"For 'context.set_context', the argument 'memory_offload' must be one of "
|
|
268
|
-
f"{memory_offload_options}, but got {memory_offload}.")
|
|
269
|
-
if memory_offload == "ON":
|
|
270
|
-
self.set_param(ms_ctx_param.memory_offload, True)
|
|
271
|
-
else:
|
|
272
|
-
self.set_param(ms_ctx_param.memory_offload, False)
|
|
273
|
-
|
|
274
257
|
def set_deterministic(self, deterministic):
|
|
275
258
|
"""
|
|
276
259
|
Enable model run in deterministic, and support the values "ON" and "OFF".
|
|
@@ -686,7 +669,6 @@ class _Context:
|
|
|
686
669
|
'memory_optimize_level': set_memory_optimize_level,
|
|
687
670
|
'exec_order': set_exec_order,
|
|
688
671
|
'op_timeout': set_op_timeout,
|
|
689
|
-
'memory_offload': set_memory_offload,
|
|
690
672
|
'deterministic': set_deterministic,
|
|
691
673
|
'ascend_config': set_ascend_config,
|
|
692
674
|
'jit_syntax_level': set_jit_syntax_level,
|
|
@@ -1266,65 +1248,6 @@ def reset_auto_parallel_context():
|
|
|
1266
1248
|
api.ms_compile_cache.clear()
|
|
1267
1249
|
|
|
1268
1250
|
|
|
1269
|
-
@args_type_check(offload_config=dict)
|
|
1270
|
-
def set_offload_context(offload_config):
|
|
1271
|
-
r"""
|
|
1272
|
-
Configure heterogeneous training detailed parameters to adjust the offload strategy, this api will be deprecated
|
|
1273
|
-
and removed in future versions.
|
|
1274
|
-
|
|
1275
|
-
Note:
|
|
1276
|
-
The offload configuration is only used if the memory offload feature is enabled
|
|
1277
|
-
via mindspore.set_context(memory_offload="ON"), and the memory_optimize_level must be set to O0. On the Ascend
|
|
1278
|
-
hardware platform, the graph compilation level must be O0.
|
|
1279
|
-
|
|
1280
|
-
Args:
|
|
1281
|
-
offload_config (dict): A dict contains the keys and values for setting the offload context
|
|
1282
|
-
configure.It supports the following keys.
|
|
1283
|
-
|
|
1284
|
-
- offload_path (str): The path of offload, relative paths are supported. Default: ``"./offload"``.
|
|
1285
|
-
- offload_cpu_size (str): The cpu memory size for offload. The format is "xxGB".
|
|
1286
|
-
- offload_disk_size (str): The disk size for offload. The format is "xxGB"
|
|
1287
|
-
- hbm_ratio (float): The ratio that can be used based on the maximum device memory.
|
|
1288
|
-
The range is (0,1], Default: ``1.0``.
|
|
1289
|
-
- cpu_ratio (float): The ratio that can be used based on the maximum host memory.
|
|
1290
|
-
The range is (0,1], Default: ``1.0``.
|
|
1291
|
-
- enable_pinned_mem (bool): The flag of whether enabling Pinned Memory. Default: ``True``.
|
|
1292
|
-
- enable_aio (bool): The flag of whether enabling aio. Default: ``True``.
|
|
1293
|
-
- aio_block_size (str): The size of aio block. The format is "xxGB".
|
|
1294
|
-
- aio_queue_depth (int): The depth of aio queue.
|
|
1295
|
-
- offload_param (str): The param for offload destination, cpu or disk, Default: ``""``.
|
|
1296
|
-
- offload_checkpoint (str): The checkpoint for offload destination, only valid if recompute is turned on,
|
|
1297
|
-
cpu or disk, Default: ``""``.
|
|
1298
|
-
- auto_offload (bool): The flag of whether auto offload. Default: ``True``.
|
|
1299
|
-
- host_mem_block_size (str): The memory block size of host memory pool. The format is "xxGB"
|
|
1300
|
-
|
|
1301
|
-
Raises:
|
|
1302
|
-
ValueError: If input key is not attribute in auto parallel context.
|
|
1303
|
-
|
|
1304
|
-
Examples:
|
|
1305
|
-
>>> from mindspore import context
|
|
1306
|
-
>>> context.set_offload_context(offload_config={"offload_param":"cpu"})
|
|
1307
|
-
"""
|
|
1308
|
-
_set_offload_context(offload_config)
|
|
1309
|
-
|
|
1310
|
-
|
|
1311
|
-
def get_offload_context():
|
|
1312
|
-
"""
|
|
1313
|
-
Gets the offload configuration parameters, this api will be deprecated and removed in future versions.
|
|
1314
|
-
|
|
1315
|
-
Configure through interface mindspore.set_offload_context(). If the user is not set, the default configuration is
|
|
1316
|
-
obtained.
|
|
1317
|
-
|
|
1318
|
-
Returns:
|
|
1319
|
-
Dict, heterogeneous training offload detailed configuration parameters.
|
|
1320
|
-
|
|
1321
|
-
Examples:
|
|
1322
|
-
>>> from mindspore import context
|
|
1323
|
-
>>> offload_config = context.get_offload_context()
|
|
1324
|
-
"""
|
|
1325
|
-
return _get_offload_context()
|
|
1326
|
-
|
|
1327
|
-
|
|
1328
1251
|
def _check_target_specific_cfgs(device, arg_key):
|
|
1329
1252
|
"""Checking whether a config is suitable for a specified device"""
|
|
1330
1253
|
device_cfgs = {
|
|
@@ -1400,8 +1323,7 @@ def _check_context_deprecated(key):
|
|
|
1400
1323
|
mindspore.device_context.gpu.op_precision.conv_fprop_algo(),
|
|
1401
1324
|
mindspore.device_context.gpu.op_precision.conv_wgrad_algo(),
|
|
1402
1325
|
mindspore.device_context.gpu.op_precision.conv_dgrad_algo()''',
|
|
1403
|
-
'runtime_num_threads': 'api mindspore.device_context.cpu.op_tuning.threads_num()'
|
|
1404
|
-
'memory_offload': "`device` parameter of `mindspore.Parameter`"}
|
|
1326
|
+
'runtime_num_threads': 'api mindspore.device_context.cpu.op_tuning.threads_num()'}
|
|
1405
1327
|
invalid_context_dict = {
|
|
1406
1328
|
'exception_dump': {'version': '2.6', 'interface': 'device_context.ascend.op_debug.aclinit_config()'}
|
|
1407
1329
|
}
|
|
@@ -1468,9 +1390,6 @@ def set_context(**kwargs):
|
|
|
1468
1390
|
inter_op_parallel_num(int): The thread number of op parallel at the same time.
|
|
1469
1391
|
Default ``0`` . This parameter will be deprecated and removed in future versions.
|
|
1470
1392
|
Please use the api :func:`mindspore.runtime.dispatch_threads_num` instead.
|
|
1471
|
-
memory_offload (str): Whether to enable the memory offload function. Default ``"OFF"`` .
|
|
1472
|
-
This parameter will be deprecated and removed in future versions. Please use the api
|
|
1473
|
-
:func:`mindspore.nn.Cell.offload` instead.
|
|
1474
1393
|
disable_format_transform (bool): Whether to disable the automatic format transform function from NCHW
|
|
1475
1394
|
to NHWC. Default ``False`` . This parameter will be deprecated and removed in future versions. Please
|
|
1476
1395
|
use the related parameter of :func:`mindspore.jit` instead.
|
|
@@ -1540,7 +1459,9 @@ def set_context(**kwargs):
|
|
|
1540
1459
|
This parameter will be deprecated and removed in future versions. Please use the
|
|
1541
1460
|
api :func:`mindspore.parallel.auto_parallel.AutoParallel.transformer_opt` instead.
|
|
1542
1461
|
- hccl_watchdog (bool): Enable a thread to monitor the failure of collective communication.
|
|
1543
|
-
Default ``True`` .
|
|
1462
|
+
Default ``True`` . This parameter will be deprecated and removed in future versions. Please use the
|
|
1463
|
+
environment variable `MS_ENABLE_THM="{HCCL_WATCHDOG:1}"` instead.
|
|
1464
|
+
|
|
1544
1465
|
gpu_config (dict): Set the parameters specific to gpu hardware platform. It is not set by default.
|
|
1545
1466
|
|
|
1546
1467
|
- conv_fprop_algo (str): Specifies convolution forward algorithm. Default ``"normal"`` .
|
|
@@ -1595,7 +1516,6 @@ def set_context(**kwargs):
|
|
|
1595
1516
|
>>> ms.set_context(inter_op_parallel_num=4)
|
|
1596
1517
|
>>> ms.set_context(disable_format_transform=True)
|
|
1597
1518
|
>>> ms.set_context(memory_optimize_level='O0')
|
|
1598
|
-
>>> ms.set_context(memory_offload='ON')
|
|
1599
1519
|
>>> ms.set_context(deterministic='ON')
|
|
1600
1520
|
>>> ms.set_context(ascend_config={"precision_mode": "force_fp16", "jit_compile": True,
|
|
1601
1521
|
... "atomic_clean_policy": 1, "op_precision_mode": "./op_precision_config_file",
|
|
@@ -457,8 +457,10 @@ class Dataset:
|
|
|
457
457
|
each bucket. Must contain len(bucket_boundaries)+1 elements.
|
|
458
458
|
element_length_function (Callable, optional): A function that takes in
|
|
459
459
|
M arguments where M = len(column_names) and returns an integer. If no value
|
|
460
|
-
provided, parameter M the len(column_names) must be 1,
|
|
461
|
-
|
|
460
|
+
provided, parameter M the len(column_names) must be 1. At this time, the length of the data in this
|
|
461
|
+
column is determined based on its ndim. If ndim=0, the data length is 0, indicating a str, bool, int,
|
|
462
|
+
or float scalar; if it is an array with ndim > 0, the length of the data is array.shape[0].
|
|
463
|
+
Default: ``None`` , indicating this parameter is not specified.
|
|
462
464
|
pad_info (dict, optional): The information about how to batch each column. The key
|
|
463
465
|
corresponds to the column name, and the value must be a tuple of 2 elements.
|
|
464
466
|
The first element corresponds to the shape to pad to, and the second
|
|
@@ -819,8 +821,7 @@ class Dataset:
|
|
|
819
821
|
return dataset
|
|
820
822
|
|
|
821
823
|
@check_map
|
|
822
|
-
def map(self, operations, input_columns=None, output_columns=None,
|
|
823
|
-
num_parallel_workers=None, **kwargs):
|
|
824
|
+
def map(self, operations, input_columns=None, output_columns=None, num_parallel_workers=None, **kwargs):
|
|
824
825
|
"""
|
|
825
826
|
Apply each operation in operations to this dataset.
|
|
826
827
|
|
|
@@ -3073,6 +3074,9 @@ def _worker_loop(quit_signal, operations, worker_id, op_type, key, video_backend
|
|
|
3073
3074
|
and shared memory. This logic no longer uses the Python multi-process pool, in_queue, and out_queue for
|
|
3074
3075
|
data transferring.
|
|
3075
3076
|
"""
|
|
3077
|
+
# Release the lock which had been holded in map_op.cc::Launch()/batch_op.cc::Launch()
|
|
3078
|
+
cde.unlock_shm_id_and_msg_id_mutex()
|
|
3079
|
+
|
|
3076
3080
|
# Initialize C++ side signal handlers
|
|
3077
3081
|
cde.register_worker_handlers()
|
|
3078
3082
|
|
|
@@ -2179,7 +2179,7 @@ class ImageFolderDataset(MappableDataset, VisionBaseDataset):
|
|
|
2179
2179
|
<https://www.mindspore.cn/docs/en/master/api_python/samples/dataset/dataset_gallery.html>`_
|
|
2180
2180
|
|
|
2181
2181
|
Note:
|
|
2182
|
-
- The shape of the image column is [
|
|
2182
|
+
- The shape of the image column is [undecoded_image_size] if `decode` flag is ``False``, or [H,W,C] otherwise.
|
|
2183
2183
|
- The parameters `num_samples` , `shuffle` , `num_shards` , `shard_id` can be used to control the sampler
|
|
2184
2184
|
used in the dataset, and their effects when combined with parameter `sampler` are as follows.
|
|
2185
2185
|
|
|
@@ -1481,23 +1481,9 @@ def check_map(method):
|
|
|
1481
1481
|
@wraps(method)
|
|
1482
1482
|
def new_method(self, *args, **kwargs):
|
|
1483
1483
|
from mindspore.dataset.callback import DSCallback
|
|
1484
|
-
[operations, input_columns, output_columns,
|
|
1484
|
+
[operations, input_columns, output_columns, num_parallel_workers, param_dict], _ = \
|
|
1485
1485
|
parse_user_args(method, *args, **kwargs)
|
|
1486
1486
|
|
|
1487
|
-
if column_order is not None:
|
|
1488
|
-
raise ValueError("The parameter 'column_order' had been deleted in map operation. "
|
|
1489
|
-
"Please use '.project' operation instead.\n"
|
|
1490
|
-
">> # Usage of old api:\n"
|
|
1491
|
-
">> dataset = dataset.map(operations=PyFunc,\n"
|
|
1492
|
-
">> input_columns=[\"column_a\"],\n"
|
|
1493
|
-
">> output_columns=[\"column_b\", \"column_c\"],\n"
|
|
1494
|
-
">> column_order=[\"column_b\", \"column_c\"])\n"
|
|
1495
|
-
">> # Usage of new api:\n"
|
|
1496
|
-
">> dataset = dataset.map(operations=PyFunc,\n"
|
|
1497
|
-
">> input_columns=[\"column_a\"],\n"
|
|
1498
|
-
">> output_columns=[\"column_b\", \"column_c\"])\n"
|
|
1499
|
-
">> dataset = dataset.project([\"column_b\", \"column_c\"])")
|
|
1500
|
-
|
|
1501
1487
|
(python_multiprocessing, max_rowsize, cache, callbacks, offload) = get_map_kwargs_from_dict(param_dict)
|
|
1502
1488
|
|
|
1503
1489
|
# check whether network computing operator exist in input operations(python function)
|
mindspore/dnnl.dll
CHANGED
|
Binary file
|
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
# Copyright
|
|
1
|
+
# Copyright 2025 Huawei Technologies Co., Ltd
|
|
2
2
|
#
|
|
3
3
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
4
4
|
# you may not use this file except in compliance with the License.
|
|
@@ -12,11 +12,11 @@
|
|
|
12
12
|
# See the License for the specific language governing permissions and
|
|
13
13
|
# limitations under the License.
|
|
14
14
|
# ============================================================================
|
|
15
|
-
"""
|
|
16
|
-
Provide llm boost for inference, such as LlamaBoost.
|
|
17
|
-
"""
|
|
18
|
-
from __future__ import absolute_import
|
|
19
15
|
|
|
20
|
-
|
|
16
|
+
"""Graph optimization and transformation functionality."""
|
|
21
17
|
|
|
22
|
-
|
|
18
|
+
from mindspore.graph.custom_pass import register_custom_pass
|
|
19
|
+
|
|
20
|
+
__all__ = [
|
|
21
|
+
"register_custom_pass"
|
|
22
|
+
]
|
|
@@ -0,0 +1,55 @@
|
|
|
1
|
+
# Copyright 2025 Huawei Technologies Co., Ltd
|
|
2
|
+
#
|
|
3
|
+
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
4
|
+
# you may not use this file except in compliance with the License.
|
|
5
|
+
# You may obtain a copy of the License at
|
|
6
|
+
#
|
|
7
|
+
# http://www.apache.org/licenses/LICENSE-2.0
|
|
8
|
+
#
|
|
9
|
+
# Unless required by applicable law or agreed to in writing, software
|
|
10
|
+
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
11
|
+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
12
|
+
# See the License for the specific language governing permissions and
|
|
13
|
+
# limitations under the License.
|
|
14
|
+
# ============================================================================
|
|
15
|
+
|
|
16
|
+
"""
|
|
17
|
+
Experimental custom pass registration functionality.
|
|
18
|
+
|
|
19
|
+
This module provides experimental APIs for registering custom optimization passes.
|
|
20
|
+
These APIs are subject to change and should be used with caution in production code.
|
|
21
|
+
"""
|
|
22
|
+
|
|
23
|
+
import mindspore._c_expression as _c_expression
|
|
24
|
+
|
|
25
|
+
|
|
26
|
+
def register_custom_pass(pass_name: str, plugin_so_path: str, device: str = "all", stage: str = "") -> bool:
|
|
27
|
+
"""Register a custom pass plugin.
|
|
28
|
+
|
|
29
|
+
.. warning::
|
|
30
|
+
This is an experimental API that is subject to change or deletion.
|
|
31
|
+
|
|
32
|
+
Args:
|
|
33
|
+
pass_name (str): Name of the pass expected to be provided by the plugin.
|
|
34
|
+
plugin_so_path (str): Absolute path to the plugin shared library (.so file).
|
|
35
|
+
device (str): Target device for the pass. Supported values: "cpu", "gpu", "ascend", or "all".
|
|
36
|
+
Default: "all".
|
|
37
|
+
stage (str): Pass stage. Reserved field for future use. Default: "".
|
|
38
|
+
|
|
39
|
+
Returns:
|
|
40
|
+
bool: True if plugin loaded successfully, False otherwise.
|
|
41
|
+
|
|
42
|
+
Examples:
|
|
43
|
+
>>> import mindspore.graph as graph
|
|
44
|
+
>>> # Register a custom optimization pass
|
|
45
|
+
>>> success = graph.register_custom_pass(
|
|
46
|
+
... pass_name="my_fusion_pass",
|
|
47
|
+
... plugin_so_path="/path/to/my_plugin.so",
|
|
48
|
+
... device="ascend"
|
|
49
|
+
... )
|
|
50
|
+
>>> print(f"Registration successful: {success}")
|
|
51
|
+
"""
|
|
52
|
+
return _c_expression.register_custom_pass(pass_name, plugin_so_path, device, stage)
|
|
53
|
+
|
|
54
|
+
|
|
55
|
+
__all__ = ["register_custom_pass"]
|
|
@@ -27,8 +27,8 @@
|
|
|
27
27
|
#include "include/dataset/constants.h"
|
|
28
28
|
#include "include/dataset/transforms.h"
|
|
29
29
|
#if defined(ENABLE_D)
|
|
30
|
-
#include "runtime/
|
|
31
|
-
#include "runtime/
|
|
30
|
+
#include "runtime/hardware_abstract/device_context/device_context.h"
|
|
31
|
+
#include "runtime/hardware_abstract/device_context/device_context_manager.h"
|
|
32
32
|
#endif
|
|
33
33
|
|
|
34
34
|
namespace mindspore {
|
mindspore/jpeg62.dll
CHANGED
|
Binary file
|
mindspore/mindrecord/__init__.py
CHANGED
|
@@ -31,8 +31,8 @@ datasets using `mindspore.dataset.MindDataset <https://www.mindspore.cn/docs/en/
|
|
|
31
31
|
mindspore.dataset.MindDataset.html>`_ .
|
|
32
32
|
|
|
33
33
|
Users can also convert datasets from other formats to the MindRecord format.
|
|
34
|
-
For more details, please refer to
|
|
35
|
-
|
|
34
|
+
For more details, please refer to
|
|
35
|
+
`Converting Dataset to MindRecord <https://www.mindspore.cn/tutorials/en/master/dataset/record.html>`_ .
|
|
36
36
|
Additionally, MindRecord supports file encryption, decryption,
|
|
37
37
|
and integrity checks to ensure the security of MindRecord format datasets.
|
|
38
38
|
"""
|
|
@@ -41,7 +41,7 @@ from .filewriter import FileWriter
|
|
|
41
41
|
from .filereader import FileReader
|
|
42
42
|
from .mindpage import MindPage
|
|
43
43
|
from .common.exceptions import *
|
|
44
|
-
from .shardutils import SUCCESS, FAILED
|
|
44
|
+
from .core.shardutils import SUCCESS, FAILED
|
|
45
45
|
from .tools.cifar10_to_mr import Cifar10ToMR
|
|
46
46
|
from .tools.cifar100_to_mr import Cifar100ToMR
|
|
47
47
|
from .tools.csv_to_mr import CsvToMR
|
mindspore/mindrecord/config.py
CHANGED
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
# Copyright
|
|
1
|
+
# Copyright 2025 Huawei Technologies Co., Ltd
|
|
2
2
|
#
|
|
3
3
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
4
4
|
# you may not use this file except in compliance with the License.
|
|
@@ -12,3 +12,6 @@
|
|
|
12
12
|
# See the License for the specific language governing permissions and
|
|
13
13
|
# limitations under the License.
|
|
14
14
|
# ============================================================================
|
|
15
|
+
"""
|
|
16
|
+
init core
|
|
17
|
+
"""
|
|
@@ -17,7 +17,8 @@ This module is to write data into mindrecord.
|
|
|
17
17
|
"""
|
|
18
18
|
import mindspore._c_mindrecord as ms
|
|
19
19
|
from mindspore import log as logger
|
|
20
|
-
from .common.exceptions import MRMAddSchemaError, MRMAddIndexError, MRMBuildSchemaError,
|
|
20
|
+
from mindspore.mindrecord.common.exceptions import MRMAddSchemaError, MRMAddIndexError, MRMBuildSchemaError, \
|
|
21
|
+
MRMGetMetaError
|
|
21
22
|
|
|
22
23
|
__all__ = ['ShardHeader']
|
|
23
24
|
|
|
@@ -17,7 +17,7 @@ This module is to write data into mindrecord.
|
|
|
17
17
|
"""
|
|
18
18
|
import mindspore._c_mindrecord as ms
|
|
19
19
|
from mindspore import log as logger
|
|
20
|
-
from .common.exceptions import MRMIndexGeneratorError, MRMGenerateIndexError
|
|
20
|
+
from mindspore.mindrecord.common.exceptions import MRMIndexGeneratorError, MRMGenerateIndexError
|
|
21
21
|
|
|
22
22
|
__all__ = ['ShardIndexGenerator']
|
|
23
23
|
|
|
@@ -17,7 +17,8 @@ This module is to read data from mindrecord.
|
|
|
17
17
|
"""
|
|
18
18
|
import mindspore._c_mindrecord as ms
|
|
19
19
|
from mindspore import log as logger
|
|
20
|
-
from .common.exceptions import MRMOpenError, MRMLaunchError
|
|
20
|
+
from mindspore.mindrecord.common.exceptions import MRMOpenError, MRMLaunchError
|
|
21
|
+
|
|
21
22
|
__all__ = ['ShardReader']
|
|
22
23
|
|
|
23
24
|
|
|
@@ -17,8 +17,8 @@ This module is to read page from mindrecord.
|
|
|
17
17
|
"""
|
|
18
18
|
import mindspore._c_mindrecord as ms
|
|
19
19
|
from mindspore import log as logger
|
|
20
|
-
from .shardutils import populate_data, SUCCESS
|
|
21
|
-
from .shardheader import ShardHeader
|
|
20
|
+
from mindspore.mindrecord.core.shardutils import populate_data, SUCCESS
|
|
21
|
+
from mindspore.mindrecord.core.shardheader import ShardHeader
|
|
22
22
|
|
|
23
23
|
__all__ = ['ShardSegment']
|
|
24
24
|
|
|
@@ -25,7 +25,7 @@ from functools import wraps
|
|
|
25
25
|
|
|
26
26
|
import numpy as np
|
|
27
27
|
import mindspore._c_mindrecord as ms
|
|
28
|
-
from .common.exceptions import ParamValueError, MRMUnsupportedSchemaError
|
|
28
|
+
from mindspore.mindrecord.common.exceptions import ParamValueError, MRMUnsupportedSchemaError
|
|
29
29
|
|
|
30
30
|
SUCCESS = ms.MSRStatus.SUCCESS
|
|
31
31
|
FAILED = ms.MSRStatus.FAILED
|
|
@@ -18,7 +18,7 @@ This module is to write data into mindrecord.
|
|
|
18
18
|
import numpy as np
|
|
19
19
|
import mindspore._c_mindrecord as ms
|
|
20
20
|
from mindspore import log as logger
|
|
21
|
-
from .common.exceptions import MRMOpenError, MRMOpenForAppendError, MRMInvalidHeaderSizeError, \
|
|
21
|
+
from mindspore.mindrecord.common.exceptions import MRMOpenError, MRMOpenForAppendError, MRMInvalidHeaderSizeError, \
|
|
22
22
|
MRMInvalidPageSizeError, MRMSetHeaderError, MRMWriteDatasetError, MRMCommitError
|
|
23
23
|
|
|
24
24
|
__all__ = ['ShardWriter']
|
|
@@ -18,10 +18,10 @@ This module is to read data from MindRecord.
|
|
|
18
18
|
import platform
|
|
19
19
|
from mindspore import log as logger
|
|
20
20
|
|
|
21
|
-
from .shardreader import ShardReader
|
|
22
|
-
from .shardheader import ShardHeader
|
|
23
|
-
from .shardutils import populate_data
|
|
24
|
-
from .shardutils import check_parameter
|
|
21
|
+
from .core.shardreader import ShardReader
|
|
22
|
+
from .core.shardheader import ShardHeader
|
|
23
|
+
from .core.shardutils import populate_data
|
|
24
|
+
from .core.shardutils import check_parameter
|
|
25
25
|
from .common.exceptions import ParamTypeError
|
|
26
26
|
from .config import _get_enc_key, _get_dec_mode, decrypt
|
|
27
27
|
|
|
@@ -25,11 +25,11 @@ import time
|
|
|
25
25
|
import multiprocessing as mp
|
|
26
26
|
import numpy as np
|
|
27
27
|
from mindspore import log as logger
|
|
28
|
-
from .shardwriter import ShardWriter
|
|
29
|
-
from .shardreader import ShardReader
|
|
30
|
-
from .shardheader import ShardHeader
|
|
31
|
-
from .shardindexgenerator import ShardIndexGenerator
|
|
32
|
-
from .shardutils import MIN_SHARD_COUNT, MAX_SHARD_COUNT, VALID_ATTRIBUTES, VALID_ARRAY_ATTRIBUTES, \
|
|
28
|
+
from .core.shardwriter import ShardWriter
|
|
29
|
+
from .core.shardreader import ShardReader
|
|
30
|
+
from .core.shardheader import ShardHeader
|
|
31
|
+
from .core.shardindexgenerator import ShardIndexGenerator
|
|
32
|
+
from .core.shardutils import MIN_SHARD_COUNT, MAX_SHARD_COUNT, VALID_ATTRIBUTES, VALID_ARRAY_ATTRIBUTES, \
|
|
33
33
|
check_filename, VALUE_TYPE_MAP, SUCCESS
|
|
34
34
|
from .common.exceptions import ParamValueError, ParamTypeError, MRMInvalidSchemaError, MRMDefineIndexError
|
|
35
35
|
from .config import _get_enc_key, _get_enc_mode, _get_dec_mode, encrypt, decrypt
|
mindspore/mindrecord/mindpage.py
CHANGED
|
@@ -16,8 +16,8 @@
|
|
|
16
16
|
This module is to support reading page from MindRecord.
|
|
17
17
|
"""
|
|
18
18
|
|
|
19
|
-
from .shardsegment import ShardSegment
|
|
20
|
-
from .shardutils import check_parameter
|
|
19
|
+
from .core.shardsegment import ShardSegment
|
|
20
|
+
from .core.shardutils import check_parameter
|
|
21
21
|
from .common.exceptions import ParamValueError, ParamTypeError, MRMDefineCategoryError
|
|
22
22
|
from .config import _get_enc_key, _get_dec_mode, decrypt
|
|
23
23
|
|
|
@@ -23,7 +23,7 @@ from mindspore import log as logger
|
|
|
23
23
|
from .cifar100 import Cifar100
|
|
24
24
|
from ..common.exceptions import PathNotExistsError
|
|
25
25
|
from ..filewriter import FileWriter
|
|
26
|
-
from ..shardutils import check_filename, ExceptionThread, SUCCESS
|
|
26
|
+
from ..core.shardutils import check_filename, ExceptionThread, SUCCESS
|
|
27
27
|
|
|
28
28
|
|
|
29
29
|
__all__ = ['Cifar100ToMR']
|
|
@@ -23,7 +23,7 @@ from mindspore import log as logger
|
|
|
23
23
|
from .cifar10 import Cifar10
|
|
24
24
|
from ..common.exceptions import PathNotExistsError
|
|
25
25
|
from ..filewriter import FileWriter
|
|
26
|
-
from ..shardutils import check_filename, ExceptionThread, SUCCESS
|
|
26
|
+
from ..core.shardutils import check_filename, ExceptionThread, SUCCESS
|
|
27
27
|
|
|
28
28
|
|
|
29
29
|
__all__ = ['Cifar10ToMR']
|
|
@@ -21,7 +21,7 @@ import time
|
|
|
21
21
|
from mindspore import log as logger
|
|
22
22
|
from ..common.exceptions import PathNotExistsError
|
|
23
23
|
from ..filewriter import FileWriter
|
|
24
|
-
from ..shardutils import check_filename, ExceptionThread
|
|
24
|
+
from ..core.shardutils import check_filename, ExceptionThread
|
|
25
25
|
|
|
26
26
|
__all__ = ['ImageNetToMR']
|
|
27
27
|
|
|
@@ -23,7 +23,7 @@ import numpy as np
|
|
|
23
23
|
|
|
24
24
|
from mindspore import log as logger
|
|
25
25
|
from ..filewriter import FileWriter
|
|
26
|
-
from ..shardutils import check_filename, ExceptionThread, SUCCESS, FAILED
|
|
26
|
+
from ..core.shardutils import check_filename, ExceptionThread, SUCCESS, FAILED
|
|
27
27
|
|
|
28
28
|
|
|
29
29
|
__all__ = ['MnistToMR']
|
|
@@ -21,7 +21,7 @@ import numpy as np
|
|
|
21
21
|
|
|
22
22
|
from mindspore import log as logger
|
|
23
23
|
from ..filewriter import FileWriter
|
|
24
|
-
from ..shardutils import check_filename, ExceptionThread
|
|
24
|
+
from ..core.shardutils import check_filename, ExceptionThread
|
|
25
25
|
|
|
26
26
|
__all__ = ['TFRecordToMR']
|
|
27
27
|
|
|
Binary file
|
|
Binary file
|
|
Binary file
|
mindspore/mindspore_common.dll
CHANGED
|
Binary file
|
mindspore/mindspore_core.dll
CHANGED
|
Binary file
|
|
Binary file
|
mindspore/mindspore_dump.dll
CHANGED
|
Binary file
|
mindspore/mindspore_frontend.dll
CHANGED
|
Binary file
|
mindspore/mindspore_glog.dll
CHANGED
|
Binary file
|
|
Binary file
|
|
Binary file
|
|
Binary file
|
mindspore/mindspore_ops.dll
CHANGED
|
Binary file
|
|
Binary file
|
mindspore/mindspore_profiler.dll
CHANGED
|
Binary file
|
mindspore/mindspore_pyboost.dll
CHANGED
|
Binary file
|
mindspore/mindspore_pynative.dll
CHANGED
|
Binary file
|
|
Binary file
|
|
Binary file
|
|
Binary file
|