mindspore 2.5.0__cp311-cp311-win_amd64.whl → 2.6.0rc1__cp311-cp311-win_amd64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of mindspore might be problematic. Click here for more details.
- mindspore/.commit_id +1 -1
- mindspore/Microsoft.VisualStudio.Telemetry.dll +0 -0
- mindspore/Newtonsoft.Json.dll +0 -0
- mindspore/__init__.py +6 -4
- mindspore/_c_dataengine.cp311-win_amd64.pyd +0 -0
- mindspore/_c_expression.cp311-win_amd64.pyd +0 -0
- mindspore/_c_mindrecord.cp311-win_amd64.pyd +0 -0
- mindspore/_check_jit_forbidden_api.py +3 -0
- mindspore/_checkparam.py +3 -33
- mindspore/_deprecated/__init__.py +17 -0
- mindspore/_deprecated/jit.py +198 -0
- mindspore/_extends/builtin_operations.py +1 -1
- mindspore/_extends/parse/__init__.py +6 -7
- mindspore/_extends/parse/compile_config.py +19 -0
- mindspore/_extends/parse/deprecated/deprecated_tensor_method.py +22 -3
- mindspore/_extends/parse/jit_fallback_modules/__init__.py +0 -0
- mindspore/_extends/parse/jit_fallback_modules/check_utils.py +123 -0
- mindspore/_extends/parse/jit_fallback_modules/third_party_modules.py +50 -0
- mindspore/_extends/parse/parser.py +24 -193
- mindspore/_extends/parse/resources.py +1 -5
- mindspore/_extends/parse/standard_method.py +97 -74
- mindspore/_extends/pijit/__init__.py +2 -2
- mindspore/_extends/pijit/pijit_func_white_list.py +16 -11
- mindspore/_extends/pijit/tensor_func_list.py +27 -0
- mindspore/_extends/utils.py +1 -1
- mindspore/amp.py +4 -4
- mindspore/atlprov.dll +0 -0
- mindspore/avcodec-59.dll +0 -0
- mindspore/avdevice-59.dll +0 -0
- mindspore/avfilter-8.dll +0 -0
- mindspore/avformat-59.dll +0 -0
- mindspore/avutil-57.dll +0 -0
- mindspore/boost/__init__.py +2 -2
- mindspore/boost/base.py +3 -7
- mindspore/boost/boost_cell_wrapper.py +2 -2
- mindspore/c1.dll +0 -0
- mindspore/c1xx.dll +0 -0
- mindspore/c2.dll +0 -0
- mindspore/common/__init__.py +4 -3
- mindspore/common/_grad_function.py +56 -0
- mindspore/common/_pijit_context.py +14 -5
- mindspore/common/_register_for_tensor.py +1 -1
- mindspore/common/_stub_tensor.py +5 -10
- mindspore/common/_tensor_cpp_method.py +1 -1
- mindspore/common/_tensor_docs.py +1915 -3287
- mindspore/common/api.py +341 -354
- mindspore/common/auto_dynamic_shape.py +41 -44
- mindspore/common/dtype.py +5 -2
- mindspore/common/dump.py +7 -5
- mindspore/common/file_system.py +3 -0
- mindspore/common/hook_handle.py +5 -3
- mindspore/common/initializer.py +10 -6
- mindspore/common/jit_begin_end.py +94 -0
- mindspore/common/jit_config.py +6 -1
- mindspore/common/jit_context.py +76 -0
- mindspore/common/jit_trace.py +378 -0
- mindspore/common/lazy_inline.py +2 -2
- mindspore/common/mutable.py +5 -4
- mindspore/common/parameter.py +106 -39
- mindspore/common/seed.py +2 -2
- mindspore/common/sparse_tensor.py +23 -17
- mindspore/common/tensor.py +297 -714
- mindspore/communication/__init__.py +7 -5
- mindspore/communication/_comm_helper.py +47 -2
- mindspore/communication/comm_func.py +70 -53
- mindspore/communication/management.py +83 -17
- mindspore/context.py +214 -560
- mindspore/dataset/__init__.py +44 -20
- mindspore/dataset/audio/__init__.py +2 -8
- mindspore/dataset/audio/transforms.py +3 -17
- mindspore/dataset/core/config.py +3 -3
- mindspore/dataset/engine/cache_client.py +1 -1
- mindspore/dataset/engine/datasets.py +102 -120
- mindspore/dataset/engine/datasets_audio.py +22 -22
- mindspore/dataset/engine/datasets_standard_format.py +43 -24
- mindspore/dataset/engine/datasets_text.py +78 -85
- mindspore/dataset/engine/datasets_user_defined.py +108 -76
- mindspore/dataset/engine/datasets_vision.py +111 -108
- mindspore/dataset/engine/iterators.py +5 -3
- mindspore/dataset/engine/obs/obs_mindrecord_dataset.py +1 -1
- mindspore/dataset/engine/samplers.py +279 -57
- mindspore/dataset/engine/serializer_deserializer.py +2 -1
- mindspore/dataset/engine/validators.py +10 -0
- mindspore/dataset/text/__init__.py +7 -6
- mindspore/dataset/text/transforms.py +6 -5
- mindspore/dataset/text/utils.py +3 -3
- mindspore/dataset/transforms/__init__.py +0 -9
- mindspore/dataset/transforms/transforms.py +3 -3
- mindspore/dataset/utils/browse_dataset.py +1 -1
- mindspore/dataset/vision/__init__.py +2 -9
- mindspore/dataset/vision/transforms.py +202 -158
- mindspore/dataset/vision/utils.py +7 -5
- mindspore/device_context/ascend/op_debug.py +60 -1
- mindspore/device_context/ascend/op_tuning.py +0 -4
- mindspore/device_manager.py +39 -3
- mindspore/dnnl.dll +0 -0
- mindspore/dpcmi.dll +0 -0
- mindspore/experimental/es/embedding_service.py +35 -27
- mindspore/experimental/map_parameter.py +4 -4
- mindspore/experimental/optim/adadelta.py +22 -26
- mindspore/experimental/optim/adagrad.py +4 -4
- mindspore/experimental/optim/adam.py +4 -0
- mindspore/experimental/optim/adamax.py +4 -4
- mindspore/experimental/optim/adamw.py +4 -0
- mindspore/experimental/optim/asgd.py +1 -1
- mindspore/experimental/optim/lr_scheduler.py +40 -22
- mindspore/experimental/optim/radam.py +5 -5
- mindspore/experimental/optim/rprop.py +1 -1
- mindspore/experimental/optim/sgd.py +1 -1
- mindspore/hal/contiguous_tensors_handle.py +6 -10
- mindspore/hal/device.py +55 -81
- mindspore/hal/event.py +38 -55
- mindspore/hal/memory.py +93 -144
- mindspore/hal/stream.py +81 -125
- mindspore/include/dataset/constants.h +7 -4
- mindspore/include/dataset/execute.h +2 -2
- mindspore/jpeg62.dll +0 -0
- mindspore/log.py +40 -2
- mindspore/mindrecord/__init__.py +20 -7
- mindspore/mindspore_backend_common.dll +0 -0
- mindspore/mindspore_backend_manager.dll +0 -0
- mindspore/mindspore_common.dll +0 -0
- mindspore/mindspore_core.dll +0 -0
- mindspore/mindspore_dump.dll +0 -0
- mindspore/mindspore_frontend.dll +0 -0
- mindspore/mindspore_glog.dll +0 -0
- mindspore/mindspore_memory_pool.dll +0 -0
- mindspore/mindspore_ms_backend.dll +0 -0
- mindspore/mindspore_ops.dll +0 -0
- mindspore/{mindspore_backend.dll → mindspore_ops_host.dll} +0 -0
- mindspore/mindspore_ops_kernel_common.dll +0 -0
- mindspore/mindspore_profiler.dll +0 -0
- mindspore/mindspore_pyboost.dll +0 -0
- mindspore/mindspore_pynative.dll +0 -0
- mindspore/mindspore_res_manager.dll +0 -0
- mindspore/mindspore_runtime_pipeline.dll +0 -0
- mindspore/mint/__init__.py +131 -700
- mindspore/mint/distributed/__init__.py +5 -1
- mindspore/mint/distributed/distributed.py +194 -109
- mindspore/mint/linalg/__init__.py +2 -0
- mindspore/mint/nn/__init__.py +280 -18
- mindspore/mint/nn/functional.py +282 -64
- mindspore/mint/nn/layer/__init__.py +4 -0
- mindspore/mint/nn/layer/_functions.py +7 -3
- mindspore/mint/nn/layer/activation.py +120 -13
- mindspore/mint/nn/layer/conv.py +218 -24
- mindspore/mint/nn/layer/normalization.py +15 -16
- mindspore/mint/nn/layer/padding.py +1 -1
- mindspore/mint/nn/layer/pooling.py +66 -1
- mindspore/mint/optim/__init__.py +2 -1
- mindspore/mint/optim/sgd.py +171 -0
- mindspore/msobj140.dll +0 -0
- mindspore/mspdb140.dll +0 -0
- mindspore/mspdbcore.dll +0 -0
- mindspore/mspdbst.dll +0 -0
- mindspore/mspft140.dll +0 -0
- mindspore/msvcdis140.dll +0 -0
- mindspore/msvcp140_1.dll +0 -0
- mindspore/msvcp140_2.dll +0 -0
- mindspore/msvcp140_atomic_wait.dll +0 -0
- mindspore/msvcp140_codecvt_ids.dll +0 -0
- mindspore/nn/__init__.py +4 -1
- mindspore/nn/cell.py +1250 -176
- mindspore/nn/layer/activation.py +23 -21
- mindspore/nn/layer/basic.py +22 -16
- mindspore/nn/layer/container.py +1 -1
- mindspore/nn/layer/conv.py +22 -17
- mindspore/nn/layer/embedding.py +9 -8
- mindspore/nn/layer/normalization.py +48 -42
- mindspore/nn/layer/pooling.py +75 -31
- mindspore/nn/layer/transformer.py +11 -10
- mindspore/nn/learning_rate_schedule.py +4 -2
- mindspore/nn/loss/loss.py +27 -19
- mindspore/nn/optim/ada_grad.py +6 -5
- mindspore/nn/optim/adadelta.py +9 -7
- mindspore/nn/optim/adafactor.py +1 -1
- mindspore/nn/optim/adam.py +16 -12
- mindspore/nn/optim/adamax.py +8 -7
- mindspore/nn/optim/adasum.py +5 -5
- mindspore/nn/optim/asgd.py +1 -1
- mindspore/nn/optim/ftrl.py +11 -9
- mindspore/nn/optim/lamb.py +1 -1
- mindspore/nn/optim/lazyadam.py +12 -10
- mindspore/nn/optim/momentum.py +7 -6
- mindspore/nn/optim/optimizer.py +2 -2
- mindspore/nn/optim/proximal_ada_grad.py +12 -10
- mindspore/nn/optim/rmsprop.py +13 -12
- mindspore/nn/optim/rprop.py +9 -7
- mindspore/nn/optim/sgd.py +9 -6
- mindspore/nn/optim/tft_wrapper.py +5 -2
- mindspore/nn/probability/bijector/bijector.py +17 -11
- mindspore/nn/probability/bijector/gumbel_cdf.py +5 -5
- mindspore/nn/probability/bijector/invert.py +2 -2
- mindspore/nn/probability/bijector/scalar_affine.py +3 -3
- mindspore/nn/probability/bijector/softplus.py +3 -2
- mindspore/nn/probability/distribution/beta.py +3 -3
- mindspore/nn/probability/distribution/categorical.py +1 -1
- mindspore/nn/probability/distribution/cauchy.py +4 -2
- mindspore/nn/probability/distribution/exponential.py +6 -7
- mindspore/nn/probability/distribution/gamma.py +2 -2
- mindspore/nn/probability/distribution/gumbel.py +2 -2
- mindspore/nn/probability/distribution/half_normal.py +5 -3
- mindspore/nn/probability/distribution/logistic.py +5 -3
- mindspore/nn/probability/distribution/poisson.py +1 -1
- mindspore/nn/probability/distribution/uniform.py +5 -3
- mindspore/nn/reinforcement/_tensors_queue.py +1 -1
- mindspore/nn/reinforcement/tensor_array.py +1 -1
- mindspore/nn/wrap/__init__.py +6 -6
- mindspore/nn/wrap/cell_wrapper.py +178 -117
- mindspore/nn/wrap/grad_reducer.py +45 -36
- mindspore/nn/wrap/loss_scale.py +3 -3
- mindspore/numpy/array_creations.py +3 -3
- mindspore/numpy/array_ops.py +1 -1
- mindspore/numpy/math_ops.py +4 -4
- mindspore/numpy/utils.py +1 -2
- mindspore/numpy/utils_const.py +1 -2
- mindspore/opencv_core452.dll +0 -0
- mindspore/opencv_imgcodecs452.dll +0 -0
- mindspore/opencv_imgproc452.dll +0 -0
- mindspore/ops/__init__.py +3 -2
- mindspore/ops/_grad_experimental/grad_comm_ops.py +18 -3
- mindspore/ops/_grad_experimental/grad_debug_ops.py +8 -1
- mindspore/ops/_grad_experimental/taylor_rule.py +29 -0
- mindspore/ops/_register_for_op.py +0 -11
- mindspore/{ops_generate → ops/_utils}/arg_dtype_cast.py +123 -4
- mindspore/{ops_generate → ops/_utils}/arg_handler.py +3 -4
- mindspore/ops/_vmap/vmap_array_ops.py +7 -6
- mindspore/ops/_vmap/vmap_grad_nn_ops.py +2 -1
- mindspore/ops/_vmap/vmap_math_ops.py +4 -7
- mindspore/ops/_vmap/vmap_nn_ops.py +9 -8
- mindspore/ops/auto_generate/__init__.py +4 -3
- mindspore/ops/auto_generate/cpp_create_prim_instance_helper.py +102 -49
- mindspore/ops/auto_generate/gen_extend_func.py +281 -135
- mindspore/ops/auto_generate/gen_ops_def.py +2574 -2326
- mindspore/ops/auto_generate/gen_ops_prim.py +8566 -2755
- mindspore/ops/auto_generate/pyboost_inner_prim.py +106 -76
- mindspore/ops/composite/__init__.py +2 -1
- mindspore/ops/composite/base.py +19 -24
- mindspore/ops/composite/math_ops.py +6 -16
- mindspore/ops/composite/multitype_ops/__init__.py +5 -2
- mindspore/ops/composite/multitype_ops/_compile_utils.py +2 -3
- mindspore/ops/composite/multitype_ops/_constexpr_utils.py +1 -2
- mindspore/ops/composite/multitype_ops/add_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/bitwise_and_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/bitwise_or_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/bitwise_xor_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/div_impl.py +6 -4
- mindspore/ops/composite/multitype_ops/equal_impl.py +4 -3
- mindspore/ops/composite/multitype_ops/floordiv_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/getitem_impl.py +3 -2
- mindspore/ops/composite/multitype_ops/greater_equal_impl.py +4 -3
- mindspore/ops/composite/multitype_ops/greater_impl.py +4 -3
- mindspore/ops/composite/multitype_ops/in_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/invert_impl.py +50 -0
- mindspore/ops/composite/multitype_ops/left_shift_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/less_equal_impl.py +4 -3
- mindspore/ops/composite/multitype_ops/less_impl.py +4 -3
- mindspore/ops/composite/multitype_ops/logic_not_impl.py +3 -2
- mindspore/ops/composite/multitype_ops/logical_and_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/logical_or_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/mod_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/mul_impl.py +3 -2
- mindspore/ops/composite/multitype_ops/negative_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/not_equal_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/not_in_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/ones_like_impl.py +18 -0
- mindspore/ops/composite/multitype_ops/pow_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/right_shift_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/setitem_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/sub_impl.py +2 -1
- mindspore/ops/function/__init__.py +28 -2
- mindspore/ops/function/_add_attr_func.py +58 -0
- mindspore/ops/function/array_func.py +1629 -2345
- mindspore/ops/function/clip_func.py +38 -45
- mindspore/ops/function/debug_func.py +36 -44
- mindspore/ops/function/grad/__init__.py +1 -0
- mindspore/ops/function/grad/grad_func.py +104 -71
- mindspore/ops/function/image_func.py +1 -1
- mindspore/ops/function/linalg_func.py +46 -78
- mindspore/ops/function/math_func.py +3035 -3705
- mindspore/ops/function/nn_func.py +676 -241
- mindspore/ops/function/other_func.py +159 -1
- mindspore/ops/function/parameter_func.py +17 -30
- mindspore/ops/function/random_func.py +204 -361
- mindspore/ops/function/reshard_func.py +4 -70
- mindspore/ops/function/sparse_func.py +3 -3
- mindspore/ops/function/sparse_unary_func.py +5 -5
- mindspore/ops/function/spectral_func.py +25 -58
- mindspore/ops/function/vmap_func.py +24 -17
- mindspore/ops/functional.py +6 -4
- mindspore/ops/functional_overload.py +547 -4
- mindspore/ops/op_info_register.py +32 -244
- mindspore/ops/operations/__init__.py +10 -5
- mindspore/ops/operations/_custom_ops_utils.py +247 -0
- mindspore/ops/operations/_grad_ops.py +1 -10
- mindspore/ops/operations/_inner_ops.py +5 -76
- mindspore/ops/operations/_ms_kernel.py +4 -10
- mindspore/ops/operations/_rl_inner_ops.py +1 -1
- mindspore/ops/operations/_scalar_ops.py +3 -2
- mindspore/ops/operations/_sequence_ops.py +1 -1
- mindspore/ops/operations/_tensor_array.py +1 -1
- mindspore/ops/operations/array_ops.py +37 -22
- mindspore/ops/operations/comm_ops.py +150 -107
- mindspore/ops/operations/custom_ops.py +221 -23
- mindspore/ops/operations/debug_ops.py +115 -16
- mindspore/ops/operations/inner_ops.py +1 -1
- mindspore/ops/operations/linalg_ops.py +1 -58
- mindspore/ops/operations/manually_defined/_inner.py +1 -1
- mindspore/ops/operations/manually_defined/ops_def.py +746 -79
- mindspore/ops/operations/math_ops.py +21 -18
- mindspore/ops/operations/nn_ops.py +65 -191
- mindspore/ops/operations/other_ops.py +62 -9
- mindspore/ops/operations/random_ops.py +13 -7
- mindspore/ops/operations/reshard_ops.py +1 -1
- mindspore/ops/operations/sparse_ops.py +2 -2
- mindspore/ops/primitive.py +43 -32
- mindspore/ops/tensor_method.py +232 -13
- mindspore/ops_generate/__init__.py +0 -5
- mindspore/ops_generate/aclnn/__init__.py +0 -0
- mindspore/ops_generate/{aclnn_kernel_register_auto_cc_generator.py → aclnn/aclnn_kernel_register_auto_cc_generator.py} +43 -18
- mindspore/ops_generate/{gen_aclnn_implement.py → aclnn/gen_aclnn_implement.py} +49 -51
- mindspore/ops_generate/api/__init__.py +0 -0
- mindspore/ops_generate/{add_tensor_docs_generator.py → api/add_tensor_docs_generator.py} +9 -7
- mindspore/ops_generate/{cpp_create_prim_instance_helper_generator.py → api/cpp_create_prim_instance_helper_generator.py} +6 -9
- mindspore/ops_generate/{functional_map_cpp_generator.py → api/functional_map_cpp_generator.py} +25 -12
- mindspore/ops_generate/{functional_overload_py_generator.py → api/functional_overload_py_generator.py} +8 -6
- mindspore/ops_generate/{functions_cc_generator.py → api/functions_cc_generator.py} +14 -10
- mindspore/ops_generate/api/gen_api.py +103 -0
- mindspore/ops_generate/{op_api_proto.py → api/op_api_proto.py} +98 -69
- mindspore/ops_generate/{tensor_func_reg_cpp_generator.py → api/tensor_func_reg_cpp_generator.py} +82 -43
- mindspore/ops_generate/common/__init__.py +0 -0
- mindspore/ops_generate/common/gen_constants.py +91 -0
- mindspore/ops_generate/{gen_utils.py → common/gen_utils.py} +72 -19
- mindspore/ops_generate/{op_proto.py → common/op_proto.py} +64 -1
- mindspore/ops_generate/{template.py → common/template.py} +96 -84
- mindspore/ops_generate/gen_ops.py +23 -325
- mindspore/ops_generate/op_def/__init__.py +0 -0
- mindspore/ops_generate/op_def/gen_op_def.py +90 -0
- mindspore/ops_generate/{lite_ops_cpp_generator.py → op_def/lite_ops_cpp_generator.py} +47 -11
- mindspore/ops_generate/{ops_def_cc_generator.py → op_def/ops_def_cc_generator.py} +18 -7
- mindspore/ops_generate/{ops_def_h_generator.py → op_def/ops_def_h_generator.py} +5 -5
- mindspore/ops_generate/{ops_name_h_generator.py → op_def/ops_name_h_generator.py} +30 -15
- mindspore/ops_generate/op_def/ops_primitive_h_generator.py +125 -0
- mindspore/ops_generate/op_def_py/__init__.py +0 -0
- mindspore/ops_generate/op_def_py/gen_op_def_py.py +47 -0
- mindspore/ops_generate/{op_def_py_generator.py → op_def_py/op_def_py_generator.py} +6 -5
- mindspore/ops_generate/{op_prim_py_generator.py → op_def_py/op_prim_py_generator.py} +24 -15
- mindspore/ops_generate/pyboost/__init__.py +0 -0
- mindspore/ops_generate/{auto_grad_impl_cc_generator.py → pyboost/auto_grad_impl_cc_generator.py} +11 -7
- mindspore/ops_generate/{auto_grad_reg_cc_generator.py → pyboost/auto_grad_reg_cc_generator.py} +7 -7
- mindspore/ops_generate/{gen_pyboost_func.py → pyboost/gen_pyboost_func.py} +40 -16
- mindspore/ops_generate/{op_template_parser.py → pyboost/op_template_parser.py} +105 -24
- mindspore/ops_generate/{pyboost_functions_cpp_generator.py → pyboost/pyboost_functions_cpp_generator.py} +55 -18
- mindspore/ops_generate/{pyboost_functions_h_generator.py → pyboost/pyboost_functions_h_generator.py} +42 -10
- mindspore/ops_generate/{pyboost_functions_py_generator.py → pyboost/pyboost_functions_py_generator.py} +6 -6
- mindspore/ops_generate/{pyboost_grad_function_cpp_generator.py → pyboost/pyboost_grad_function_cpp_generator.py} +11 -10
- mindspore/ops_generate/{pyboost_inner_prim_generator.py → pyboost/pyboost_inner_prim_generator.py} +8 -7
- mindspore/ops_generate/{pyboost_native_grad_functions_generator.py → pyboost/pyboost_native_grad_functions_generator.py} +14 -10
- mindspore/ops_generate/{pyboost_op_cpp_code_generator.py → pyboost/pyboost_op_cpp_code_generator.py} +140 -53
- mindspore/ops_generate/{pyboost_overload_functions_cpp_generator.py → pyboost/pyboost_overload_functions_cpp_generator.py} +28 -15
- mindspore/ops_generate/{pyboost_utils.py → pyboost/pyboost_utils.py} +88 -4
- mindspore/ops_generate/resources/__init__.py +0 -0
- mindspore/ops_generate/resources/resource_list.py +30 -0
- mindspore/ops_generate/resources/resource_loader.py +36 -0
- mindspore/ops_generate/resources/resource_manager.py +64 -0
- mindspore/ops_generate/resources/yaml_loader.py +88 -0
- mindspore/ops_generate/tensor_py_cc_generator.py +122 -0
- mindspore/parallel/__init__.py +6 -2
- mindspore/parallel/_auto_parallel_context.py +133 -6
- mindspore/parallel/_cell_wrapper.py +130 -15
- mindspore/parallel/_parallel_serialization.py +95 -4
- mindspore/parallel/_ps_context.py +1 -1
- mindspore/parallel/_recovery_context.py +7 -2
- mindspore/parallel/_tensor.py +142 -18
- mindspore/parallel/_utils.py +198 -25
- mindspore/parallel/algo_parameter_config.py +3 -3
- mindspore/parallel/auto_parallel.py +732 -0
- mindspore/parallel/checkpoint_convert.py +159 -0
- mindspore/parallel/checkpoint_transform.py +656 -37
- mindspore/parallel/cluster/process_entity/_api.py +151 -19
- mindspore/parallel/cluster/run.py +1 -1
- mindspore/parallel/function/__init__.py +24 -0
- mindspore/parallel/function/reshard_func.py +259 -0
- mindspore/parallel/nn/__init__.py +25 -0
- mindspore/parallel/nn/parallel_cell_wrapper.py +263 -0
- mindspore/parallel/nn/parallel_grad_reducer.py +169 -0
- mindspore/parallel/parameter_broadcast.py +24 -13
- mindspore/parallel/shard.py +137 -61
- mindspore/parallel/transform_safetensors.py +287 -95
- mindspore/pgodb140.dll +0 -0
- mindspore/pgort140.dll +0 -0
- mindspore/profiler/__init__.py +9 -5
- mindspore/profiler/analysis/parser/ascend_cann_parser.py +6 -2
- mindspore/profiler/analysis/parser/ms_framework_parser.py +4 -4
- mindspore/profiler/analysis/parser/timeline_assembly_factory/ascend_timeline_assembler.py +7 -4
- mindspore/profiler/analysis/parser/timeline_assembly_factory/trace_view_container.py +22 -0
- mindspore/profiler/analysis/parser/timeline_creator/fwk_timeline_creator.py +3 -3
- mindspore/profiler/analysis/parser/timeline_event/fwk_event.py +241 -86
- mindspore/profiler/analysis/viewer/ascend_communication_viewer.py +41 -2
- mindspore/profiler/analysis/viewer/ascend_kernel_details_viewer.py +33 -35
- mindspore/profiler/analysis/viewer/ascend_memory_viewer.py +7 -0
- mindspore/profiler/analysis/viewer/ascend_op_memory_viewer.py +8 -3
- mindspore/profiler/analysis/viewer/ascend_step_trace_time_viewer.py +141 -30
- mindspore/profiler/analysis/viewer/ms_dataset_viewer.py +5 -6
- mindspore/profiler/common/ascend_msprof_exporter.py +5 -4
- mindspore/profiler/common/constant.py +12 -0
- mindspore/profiler/common/msprof_cmd_tool.py +42 -23
- mindspore/profiler/common/path_manager.py +24 -0
- mindspore/profiler/common/profiler_context.py +26 -2
- mindspore/profiler/common/profiler_meta_data.py +74 -0
- mindspore/profiler/common/profiler_parameters.py +59 -18
- mindspore/profiler/common/profiler_path_manager.py +66 -7
- mindspore/profiler/dynamic_profiler.py +112 -79
- mindspore/profiler/envprofiler.py +26 -1
- mindspore/profiler/experimental_config.py +197 -0
- mindspore/profiler/mstx.py +57 -14
- mindspore/profiler/platform/npu_profiler.py +33 -7
- mindspore/profiler/profiler.py +541 -45
- mindspore/profiler/profiler_action_controller.py +1 -1
- mindspore/profiler/profiler_interface.py +4 -0
- mindspore/profiler/schedule.py +57 -22
- mindspore/rewrite/api/node.py +15 -13
- mindspore/rewrite/api/symbol_tree.py +1 -1
- mindspore/run_check/_check_version.py +25 -14
- mindspore/run_check/run_check.py +1 -1
- mindspore/runtime/__init__.py +2 -2
- mindspore/runtime/executor.py +40 -11
- mindspore/runtime/memory.py +25 -8
- mindspore/safeguard/rewrite_obfuscation.py +12 -9
- mindspore/swresample-4.dll +0 -0
- mindspore/swscale-6.dll +0 -0
- mindspore/tbbmalloc.dll +0 -0
- mindspore/tinyxml2.dll +0 -0
- mindspore/train/__init__.py +8 -8
- mindspore/train/_utils.py +35 -7
- mindspore/train/amp.py +1 -1
- mindspore/train/callback/__init__.py +2 -2
- mindspore/train/callback/_callback.py +2 -16
- mindspore/train/callback/_checkpoint.py +24 -40
- mindspore/train/callback/_cluster_monitor.py +14 -18
- mindspore/train/callback/_flops_collector.py +2 -3
- mindspore/train/callback/_history.py +7 -4
- mindspore/train/callback/_lambda_callback.py +2 -2
- mindspore/train/callback/_landscape.py +0 -3
- mindspore/train/callback/_loss_monitor.py +2 -1
- mindspore/train/callback/_on_request_exit.py +6 -5
- mindspore/train/callback/_reduce_lr_on_plateau.py +11 -6
- mindspore/train/callback/_summary_collector.py +8 -13
- mindspore/train/callback/_time_monitor.py +2 -1
- mindspore/train/callback/{_tft_register.py → _train_fault_tolerance.py} +179 -103
- mindspore/train/data_sink.py +25 -2
- mindspore/train/dataset_helper.py +4 -5
- mindspore/train/loss_scale_manager.py +8 -7
- mindspore/train/metrics/accuracy.py +3 -3
- mindspore/train/metrics/confusion_matrix.py +9 -9
- mindspore/train/metrics/error.py +3 -3
- mindspore/train/metrics/hausdorff_distance.py +4 -4
- mindspore/train/metrics/mean_surface_distance.py +3 -3
- mindspore/train/metrics/metric.py +0 -12
- mindspore/train/metrics/occlusion_sensitivity.py +4 -2
- mindspore/train/metrics/precision.py +8 -6
- mindspore/train/metrics/recall.py +9 -9
- mindspore/train/metrics/root_mean_square_surface_distance.py +2 -2
- mindspore/train/mind_ir_pb2.py +19 -12
- mindspore/train/model.py +176 -103
- mindspore/train/serialization.py +246 -988
- mindspore/train/summary/_summary_adapter.py +2 -2
- mindspore/train/summary/summary_record.py +1 -1
- mindspore/turbojpeg.dll +0 -0
- mindspore/utils/__init__.py +3 -2
- mindspore/utils/dryrun.py +4 -2
- mindspore/utils/hooks.py +81 -0
- mindspore/utils/utils.py +138 -4
- mindspore/vcmeta.dll +0 -0
- mindspore/vcruntime140.dll +0 -0
- mindspore/vcruntime140_1.dll +0 -0
- mindspore/version.py +1 -1
- {mindspore-2.5.0.dist-info → mindspore-2.6.0rc1.dist-info}/METADATA +2 -1
- {mindspore-2.5.0.dist-info → mindspore-2.6.0rc1.dist-info}/RECORD +483 -438
- mindspore/_install_custom.py +0 -43
- mindspore/common/_register_for_adapter.py +0 -74
- mindspore/ops/auto_generate/gen_arg_dtype_cast.py +0 -252
- mindspore/ops/auto_generate/gen_arg_handler.py +0 -136
- mindspore/ops/operations/_opaque_predicate_registry.py +0 -41
- mindspore/ops_generate/gen_constants.py +0 -190
- mindspore/ops_generate/gen_ops_inner_prim.py +0 -131
- mindspore/ops_generate/ops_primitive_h_generator.py +0 -81
- /mindspore/ops_generate/{base_generator.py → common/base_generator.py} +0 -0
- {mindspore-2.5.0.dist-info → mindspore-2.6.0rc1.dist-info}/WHEEL +0 -0
- {mindspore-2.5.0.dist-info → mindspore-2.6.0rc1.dist-info}/entry_points.txt +0 -0
- {mindspore-2.5.0.dist-info → mindspore-2.6.0rc1.dist-info}/top_level.txt +0 -0
mindspore/hal/event.py
CHANGED
|
@@ -24,41 +24,37 @@ function_event_status = {'Event': False, 'wait': False}
|
|
|
24
24
|
|
|
25
25
|
class Event(Event_):
|
|
26
26
|
r"""
|
|
27
|
-
Wrapper around a device event
|
|
27
|
+
Wrapper around a device event, this api will be deprecated and removed in future versions, please use
|
|
28
|
+
the api :func:`mindspore.runtime.Event` instead.
|
|
28
29
|
|
|
29
30
|
Device events are synchronization markers that can be used to monitor the device’s progress,
|
|
30
31
|
to accurately measure timing, and to synchronize device streams.
|
|
31
32
|
|
|
32
33
|
The underlying device events are lazily initialized when the event is first recorded.
|
|
33
34
|
|
|
34
|
-
Note:
|
|
35
|
-
- The api will be deprecated, please use the api :func:`mindspore.runtime.Event` instead.
|
|
36
|
-
|
|
37
35
|
Args:
|
|
38
|
-
enable_timing (bool, optional): indicates if the event should measure time
|
|
39
|
-
blocking (bool, optional): if ``True``, `wait` will be blocking
|
|
36
|
+
enable_timing (bool, optional): indicates if the event should measure time. Default ``False``.
|
|
37
|
+
blocking (bool, optional): if ``True``, `wait` will be blocking. Default ``False``.
|
|
40
38
|
|
|
41
39
|
Examples:
|
|
42
|
-
>>> import mindspore
|
|
43
|
-
>>>
|
|
44
|
-
>>>
|
|
45
|
-
>>>
|
|
46
|
-
>>>
|
|
47
|
-
>>>
|
|
48
|
-
>>>
|
|
49
|
-
>>>
|
|
50
|
-
>>>
|
|
51
|
-
|
|
52
|
-
>>> with ms.hal.StreamCtx(s1):
|
|
53
|
-
... d = ops.matmul(a, b)
|
|
40
|
+
>>> import mindspore
|
|
41
|
+
>>> start = mindspore.hal.Event(enable_timing=True)
|
|
42
|
+
>>> end = mindspore.hal.Event(enable_timing=True)
|
|
43
|
+
>>> s1 = mindspore.hal.Stream()
|
|
44
|
+
>>> s2 = mindspore.hal.Stream()
|
|
45
|
+
>>> a = mindspore.tensor(mindspore.ops.ones([2, 2]), mindspore.float32)
|
|
46
|
+
>>> b = mindspore.tensor(mindspore.ops.ones([2, 2]), mindspore.float32)
|
|
47
|
+
>>> c = mindspore.tensor(mindspore.ops.ones([2, 2]), mindspore.float32)
|
|
48
|
+
>>> with mindspore.hal.StreamCtx(s1):
|
|
49
|
+
... d = mindspore.ops.matmul(a, b)
|
|
54
50
|
... start.record()
|
|
55
51
|
>>> c += 2
|
|
56
52
|
>>> end.record()
|
|
57
|
-
>>> with
|
|
53
|
+
>>> with mindspore.hal.StreamCtx(s2):
|
|
58
54
|
... start.synchronize()
|
|
59
55
|
... end.synchronize()
|
|
60
56
|
... e = c + d
|
|
61
|
-
>>>
|
|
57
|
+
>>> mindspore.hal.synchronize()
|
|
62
58
|
>>> print(e)
|
|
63
59
|
[[5. 5.]
|
|
64
60
|
[5. 5.]]
|
|
@@ -78,17 +74,14 @@ class Event(Event_):
|
|
|
78
74
|
|
|
79
75
|
def record(self, stream=None):
|
|
80
76
|
r"""
|
|
81
|
-
|
|
77
|
+
Record the event in a given stream.
|
|
82
78
|
|
|
83
79
|
Uses :func:`mindspore.hal.current_stream()` if no `stream` is specified. The
|
|
84
80
|
stream's device must match the event's device.
|
|
85
81
|
|
|
86
82
|
Args:
|
|
87
83
|
stream (Stream, optional): a stream to record. If this argument is ``None``,
|
|
88
|
-
current stream will be used. Default
|
|
89
|
-
|
|
90
|
-
Raises:
|
|
91
|
-
TypeError: If 'stream' is neither a :class:`mindspore.hal.Stream` nor a ``None``.
|
|
84
|
+
current stream will be used. Default ``None``.
|
|
92
85
|
"""
|
|
93
86
|
if stream is None:
|
|
94
87
|
stream = current_stream_()
|
|
@@ -99,33 +92,28 @@ class Event(Event_):
|
|
|
99
92
|
|
|
100
93
|
def wait(self, stream=None):
|
|
101
94
|
r"""
|
|
102
|
-
|
|
95
|
+
Make all future work submitted to the given stream wait for this
|
|
103
96
|
event.
|
|
104
97
|
|
|
105
98
|
Use :func:`mindspore.hal.current_stream()` if no `stream` is specified.
|
|
106
99
|
|
|
107
100
|
Args:
|
|
108
101
|
stream (Stream, optional): a stream to record. If this argument is ``None``,
|
|
109
|
-
current stream will be used. Default
|
|
110
|
-
|
|
111
|
-
Raises:
|
|
112
|
-
TypeError: If 'stream' is neither a :class:`mindspore.hal.Stream` nor a ``None``.
|
|
102
|
+
current stream will be used. Default ``None``.
|
|
113
103
|
|
|
114
104
|
Examples:
|
|
115
|
-
>>> import mindspore
|
|
116
|
-
>>>
|
|
117
|
-
>>>
|
|
118
|
-
>>>
|
|
119
|
-
>>>
|
|
120
|
-
>>>
|
|
121
|
-
>>>
|
|
122
|
-
|
|
123
|
-
>>> with ms.hal.StreamCtx(s1):
|
|
124
|
-
... c = ops.matmul(a, b)
|
|
105
|
+
>>> import mindspore
|
|
106
|
+
>>> event = mindspore.hal.Event()
|
|
107
|
+
>>> s1 = mindspore.hal.Stream()
|
|
108
|
+
>>> s2 = mindspore.hal.Stream()
|
|
109
|
+
>>> a = mindspore.tensor(mindspore.ops.ones([2, 2]), mindspore.float32)
|
|
110
|
+
>>> b = mindspore.tensor(mindspore.ops.ones([2, 2]), mindspore.float32)
|
|
111
|
+
>>> with mindspore.hal.StreamCtx(s1):
|
|
112
|
+
... c = mindspore.ops.matmul(a, b)
|
|
125
113
|
... event.record()
|
|
126
114
|
>>> event.wait()
|
|
127
115
|
>>> d = c + 2
|
|
128
|
-
>>>
|
|
116
|
+
>>> mindspore.hal.synchronize()
|
|
129
117
|
>>> print(d)
|
|
130
118
|
[[4. 4.]
|
|
131
119
|
[4. 4.]]
|
|
@@ -144,7 +132,7 @@ class Event(Event_):
|
|
|
144
132
|
|
|
145
133
|
def synchronize(self):
|
|
146
134
|
r"""
|
|
147
|
-
|
|
135
|
+
Wait for the event to complete.
|
|
148
136
|
|
|
149
137
|
Waits until the completion of all work currently captured in this event.
|
|
150
138
|
This prevents the CPU thread from proceeding until the event completes.
|
|
@@ -154,20 +142,18 @@ class Event(Event_):
|
|
|
154
142
|
|
|
155
143
|
def query(self):
|
|
156
144
|
r"""
|
|
157
|
-
|
|
145
|
+
Check if all work currently captured by event has completed.
|
|
158
146
|
|
|
159
147
|
Returns:
|
|
160
148
|
A boolean indicating if all work currently captured by event has completed.
|
|
161
149
|
|
|
162
150
|
Examples:
|
|
163
|
-
>>> import mindspore
|
|
164
|
-
>>>
|
|
165
|
-
>>>
|
|
166
|
-
>>>
|
|
167
|
-
>>>
|
|
168
|
-
|
|
169
|
-
>>> with ms.hal.StreamCtx(s1):
|
|
170
|
-
... c = ops.matmul(a, b)
|
|
151
|
+
>>> import mindspore
|
|
152
|
+
>>> a = mindspore.tensor(mindspore.ops.ones([1024, 2048]), mindspore.float32)
|
|
153
|
+
>>> b = mindspore.tensor(mindspore.ops.ones([2048, 4096]), mindspore.float32)
|
|
154
|
+
>>> s1 = mindspore.hal.Stream()
|
|
155
|
+
>>> with mindspore.hal.StreamCtx(s1):
|
|
156
|
+
... c = mindspore.ops.matmul(a, b)
|
|
171
157
|
... ev = s1.record_event()
|
|
172
158
|
>>> s1.synchronize()
|
|
173
159
|
>>> assert ev.query()
|
|
@@ -177,7 +163,7 @@ class Event(Event_):
|
|
|
177
163
|
|
|
178
164
|
def elapsed_time(self, end_event):
|
|
179
165
|
r"""
|
|
180
|
-
|
|
166
|
+
Return the time elapsed in milliseconds after the event was
|
|
181
167
|
recorded and before the end_event was recorded.
|
|
182
168
|
|
|
183
169
|
Args:
|
|
@@ -185,9 +171,6 @@ class Event(Event_):
|
|
|
185
171
|
|
|
186
172
|
Returns:
|
|
187
173
|
float, the time elapsed in milliseconds.
|
|
188
|
-
|
|
189
|
-
Raises:
|
|
190
|
-
TypeError: If 'end_event' is not a :class:`mindspore.hal.Event`.
|
|
191
174
|
"""
|
|
192
175
|
# pylint: disable=useless-super-delegation
|
|
193
176
|
if not isinstance(end_event, Event):
|
mindspore/hal/memory.py
CHANGED
|
@@ -27,30 +27,25 @@ function_memory_status = {'memory_stats': False, 'memory_reserved': False, 'max_
|
|
|
27
27
|
@_check_inputs_validation
|
|
28
28
|
def memory_stats(device_target=None):
|
|
29
29
|
"""
|
|
30
|
-
Returns status information queried from the memory pool
|
|
30
|
+
Returns status information queried from the memory pool, this api will be deprecated and removed in future
|
|
31
|
+
versions, please use the api :func:`mindspore.runtime.memory_stats` instead.
|
|
31
32
|
|
|
32
33
|
Note:
|
|
33
|
-
-
|
|
34
|
-
|
|
35
|
-
Note:
|
|
36
|
-
- If `device_target` is not specified, get the device capability of the current backend set by context.
|
|
37
|
-
- For the `CPU` backend, a dictionary with empty data is always returned.
|
|
34
|
+
- For the `CPU` device, a dictionary with empty data is always returned.
|
|
38
35
|
|
|
39
36
|
Args:
|
|
40
|
-
device_target (str, optional): The device
|
|
41
|
-
Default
|
|
37
|
+
device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
|
|
38
|
+
``"Ascend"`` . Default ``None`` , represents the current device set by context.
|
|
42
39
|
|
|
43
40
|
Returns:
|
|
44
41
|
dict, the queried memory information.
|
|
45
42
|
|
|
46
43
|
Examples:
|
|
47
|
-
>>> import mindspore
|
|
48
|
-
>>>
|
|
49
|
-
>>>
|
|
50
|
-
>>>
|
|
51
|
-
>>>
|
|
52
|
-
>>> c = ops.add(a, b).asnumpy()
|
|
53
|
-
>>> print(ms.hal.memory_stats())
|
|
44
|
+
>>> import mindspore
|
|
45
|
+
>>> a = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
46
|
+
>>> b = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
47
|
+
>>> c = mindspore.ops.add(a, b).asnumpy()
|
|
48
|
+
>>> print(mindspore.hal.memory_stats())
|
|
54
49
|
{'total_reserved_memory': 1073741824, 'total_allocated_memory': 1024, 'total_idle_memory': 1073740800,
|
|
55
50
|
'total_eager_free_memory': 0, 'max_reserved_memory': 1073741824, 'max_allocated_memory': 1536,
|
|
56
51
|
'common_mem_pool_stats': {'block_unit_size': 1073741824, 'block_counts': 1, 'blocks_info':
|
|
@@ -72,30 +67,25 @@ def memory_stats(device_target=None):
|
|
|
72
67
|
@_check_inputs_validation
|
|
73
68
|
def memory_reserved(device_target=None):
|
|
74
69
|
"""
|
|
75
|
-
Returns the total amount of memory currently managed by the memory pool
|
|
76
|
-
|
|
77
|
-
Note:
|
|
78
|
-
- The api will be deprecated, please use the api :func:`mindspore.runtime.memory_reserved` instead.
|
|
70
|
+
Returns the total amount of memory currently managed by the memory pool, this api will be deprecated and removed in
|
|
71
|
+
future versions, please use the api :func:`mindspore.runtime.memory_reserved` instead.
|
|
79
72
|
|
|
80
73
|
Note:
|
|
81
|
-
-
|
|
82
|
-
- For the `CPU` backend, 0 is always returned.
|
|
74
|
+
- For the `CPU` device, 0 is always returned.
|
|
83
75
|
|
|
84
76
|
Args:
|
|
85
|
-
device_target (str, optional): The device
|
|
86
|
-
Default
|
|
77
|
+
device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
|
|
78
|
+
``"Ascend"`` . Default ``None`` , represents the current device set by context.
|
|
87
79
|
|
|
88
80
|
Returns:
|
|
89
81
|
int, in Byte.
|
|
90
82
|
|
|
91
83
|
Examples:
|
|
92
|
-
>>> import mindspore
|
|
93
|
-
>>>
|
|
94
|
-
>>>
|
|
95
|
-
>>>
|
|
96
|
-
>>>
|
|
97
|
-
>>> c = ops.add(a, b).asnumpy()
|
|
98
|
-
>>> print(ms.hal.memory_reserved())
|
|
84
|
+
>>> import mindspore
|
|
85
|
+
>>> a = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
86
|
+
>>> b = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
87
|
+
>>> c = mindspore.ops.add(a, b).asnumpy()
|
|
88
|
+
>>> print(mindspore.hal.memory_reserved())
|
|
99
89
|
1073741824
|
|
100
90
|
"""
|
|
101
91
|
if not function_memory_status['memory_reserved']:
|
|
@@ -111,29 +101,25 @@ def memory_reserved(device_target=None):
|
|
|
111
101
|
def max_memory_reserved(device_target=None):
|
|
112
102
|
"""
|
|
113
103
|
Returns the peak value of the total memory managed by the memory pool since the process was started.
|
|
104
|
+
This api will be deprecated and removed in future versions, please use
|
|
105
|
+
the api :func:`mindspore.runtime.max_memory_reserved` instead.
|
|
114
106
|
|
|
115
107
|
Note:
|
|
116
|
-
-
|
|
117
|
-
|
|
118
|
-
Note:
|
|
119
|
-
- If `device_target` is not specified, get the device capability of the current backend set by context.
|
|
120
|
-
- For the `CPU` backend, 0 is always returned.
|
|
108
|
+
- For the `CPU` device, 0 is always returned.
|
|
121
109
|
|
|
122
110
|
Args:
|
|
123
|
-
device_target (str, optional): The device
|
|
124
|
-
Default
|
|
111
|
+
device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
|
|
112
|
+
``"Ascend"`` . Default ``None`` , represents the current device set by context.
|
|
125
113
|
|
|
126
114
|
Returns:
|
|
127
115
|
int, in Byte.
|
|
128
116
|
|
|
129
117
|
Examples:
|
|
130
|
-
>>> import mindspore
|
|
131
|
-
>>>
|
|
132
|
-
>>>
|
|
133
|
-
>>>
|
|
134
|
-
>>>
|
|
135
|
-
>>> c = ops.add(a, b).asnumpy()
|
|
136
|
-
>>> print(ms.hal.max_memory_reserved())
|
|
118
|
+
>>> import mindspore
|
|
119
|
+
>>> a = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
120
|
+
>>> b = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
121
|
+
>>> c = mindspore.ops.add(a, b).asnumpy()
|
|
122
|
+
>>> print(mindspore.hal.max_memory_reserved())
|
|
137
123
|
1073741824
|
|
138
124
|
"""
|
|
139
125
|
if not function_memory_status['max_memory_reserved']:
|
|
@@ -149,14 +135,12 @@ def max_memory_reserved(device_target=None):
|
|
|
149
135
|
def empty_cache():
|
|
150
136
|
"""
|
|
151
137
|
Release all memory fragments in the memory pool, so that memory arrangement
|
|
152
|
-
will be optimized
|
|
153
|
-
|
|
154
|
-
Note:
|
|
155
|
-
- The api will be deprecated, please use the api :func:`mindspore.runtime.empty_cache` instead.
|
|
138
|
+
will be optimized, this api will be deprecated and removed in future versions, please use
|
|
139
|
+
the api :func:`mindspore.runtime.empty_cache` instead.
|
|
156
140
|
|
|
157
141
|
Note:
|
|
158
142
|
Currently, the MindSpore memory pool does not have the function of releasing memory fragments.
|
|
159
|
-
This interface is reserved but implemented as an empty method and prompted in log mode.
|
|
143
|
+
This interface is reserved but implemented as an empty method and prompted in log mode when using.
|
|
160
144
|
"""
|
|
161
145
|
if not function_memory_status['empty_cache']:
|
|
162
146
|
function_memory_status['empty_cache'] = True
|
|
@@ -166,33 +150,26 @@ def empty_cache():
|
|
|
166
150
|
@_check_inputs_validation
|
|
167
151
|
def reset_peak_memory_stats(device_target=None):
|
|
168
152
|
"""
|
|
169
|
-
Reset the "peak" stats tracked by memory manager.
|
|
170
|
-
|
|
171
|
-
Note:
|
|
172
|
-
- The api will be deprecated, please use the api :func:`mindspore.runtime.reset_peak_memory_stats` instead.
|
|
173
|
-
|
|
174
|
-
Note:
|
|
175
|
-
If `device_target` is not specified, get the device capability of the current backend set by context.
|
|
153
|
+
Reset the "peak" stats tracked by memory manager, this api will be deprecated and removed in future versions.
|
|
154
|
+
Please use the api :func:`mindspore.runtime.reset_peak_memory_stats` instead.
|
|
176
155
|
|
|
177
156
|
Args:
|
|
178
|
-
device_target (str, optional): The device
|
|
179
|
-
Default
|
|
157
|
+
device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
|
|
158
|
+
``"Ascend"`` . Default ``None`` , represents the current device set by context.
|
|
180
159
|
|
|
181
160
|
Examples:
|
|
182
|
-
>>> import mindspore
|
|
183
|
-
>>>
|
|
184
|
-
>>>
|
|
185
|
-
>>>
|
|
186
|
-
>>>
|
|
187
|
-
>>> c = ops.add(a, b).asnumpy()
|
|
188
|
-
>>> print(ms.hal.max_memory_reserved())
|
|
161
|
+
>>> import mindspore
|
|
162
|
+
>>> a = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
163
|
+
>>> b = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
164
|
+
>>> c = mindspore.ops.add(a, b).asnumpy()
|
|
165
|
+
>>> print(mindspore.hal.max_memory_reserved())
|
|
189
166
|
1073741824
|
|
190
|
-
>>> print(
|
|
167
|
+
>>> print(mindspore.hal.max_memory_allocated())
|
|
191
168
|
1536
|
|
192
|
-
>>>
|
|
193
|
-
>>> print(
|
|
169
|
+
>>> mindspore.hal.reset_peak_memory_stats()
|
|
170
|
+
>>> print(mindspore.hal.max_memory_reserved())
|
|
194
171
|
0
|
|
195
|
-
>>> print(
|
|
172
|
+
>>> print(mindspore.hal.max_memory_allocated())
|
|
196
173
|
0
|
|
197
174
|
"""
|
|
198
175
|
if not function_memory_status['reset_peak_memory_stats']:
|
|
@@ -208,17 +185,12 @@ def reset_peak_memory_stats(device_target=None):
|
|
|
208
185
|
@_check_inputs_validation
|
|
209
186
|
def memory_summary(device_target=None):
|
|
210
187
|
"""
|
|
211
|
-
Returns readable memory pool status information.
|
|
212
|
-
|
|
213
|
-
Note:
|
|
214
|
-
- The api will be deprecated, please use the api :func:`mindspore.runtime.memory_summary` instead.
|
|
215
|
-
|
|
216
|
-
Note:
|
|
217
|
-
If `device_target` is not specified, get the device capability of the current backend set by context.
|
|
188
|
+
Returns readable memory pool status information, this api will be deprecated and removed in future versions.
|
|
189
|
+
Please use the api :func:`mindspore.runtime.memory_summary` instead.
|
|
218
190
|
|
|
219
191
|
Args:
|
|
220
|
-
device_target (str, optional): The device
|
|
221
|
-
Default
|
|
192
|
+
device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
|
|
193
|
+
``"Ascend"`` . Default ``None`` , represents the current device set by context.
|
|
222
194
|
|
|
223
195
|
Returns:
|
|
224
196
|
str, readable memory pool status information in tabular form.
|
|
@@ -270,30 +242,25 @@ def memory_summary(device_target=None):
|
|
|
270
242
|
@_check_inputs_validation
|
|
271
243
|
def memory_allocated(device_target=None):
|
|
272
244
|
"""
|
|
273
|
-
Returns the actual memory size currently occupied by Tensor
|
|
274
|
-
|
|
275
|
-
Note:
|
|
276
|
-
- The api will be deprecated, please use the api :func:`mindspore.runtime.memory_allocated` instead.
|
|
245
|
+
Returns the actual memory size currently occupied by Tensor, this api will be deprecated and removed in future
|
|
246
|
+
versions, please use the api :func:`mindspore.runtime.memory_allocated` instead.
|
|
277
247
|
|
|
278
248
|
Note:
|
|
279
|
-
-
|
|
280
|
-
- For the `CPU` backend, 0 is always returned.
|
|
249
|
+
- For the `CPU` device, 0 is always returned.
|
|
281
250
|
|
|
282
251
|
Args:
|
|
283
|
-
device_target (str, optional): The device
|
|
284
|
-
Default
|
|
252
|
+
device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
|
|
253
|
+
``"Ascend"`` . Default ``None`` , represents the current device set by context.
|
|
285
254
|
|
|
286
255
|
Returns:
|
|
287
256
|
int, in Byte.
|
|
288
257
|
|
|
289
258
|
Examples:
|
|
290
|
-
>>> import mindspore
|
|
291
|
-
>>>
|
|
292
|
-
>>>
|
|
293
|
-
>>>
|
|
294
|
-
>>>
|
|
295
|
-
>>> c = ops.add(a, b).asnumpy()
|
|
296
|
-
>>> print(ms.hal.memory_allocated())
|
|
259
|
+
>>> import mindspore
|
|
260
|
+
>>> a = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
261
|
+
>>> b = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
262
|
+
>>> c = mindspore.ops.add(a, b).asnumpy()
|
|
263
|
+
>>> print(mindspore.hal.memory_allocated())
|
|
297
264
|
1024
|
|
298
265
|
"""
|
|
299
266
|
if not function_memory_status['memory_allocated']:
|
|
@@ -308,30 +275,26 @@ def memory_allocated(device_target=None):
|
|
|
308
275
|
@_check_inputs_validation
|
|
309
276
|
def max_memory_allocated(device_target=None):
|
|
310
277
|
"""
|
|
311
|
-
|
|
278
|
+
Return the peak memory size of the memory pool actually occupied by Tensor since the process was started.
|
|
279
|
+
This api will be deprecated and removed in future versions, please use
|
|
280
|
+
the api :func:`mindspore.runtime.max_memory_allocated` instead.
|
|
312
281
|
|
|
313
282
|
Note:
|
|
314
|
-
-
|
|
315
|
-
|
|
316
|
-
Note:
|
|
317
|
-
- If `device_target` is not specified, get the device capability of the current backend set by context.
|
|
318
|
-
- For the `CPU` backend, 0 is always returned.
|
|
283
|
+
- For the `CPU` device, 0 is always returned.
|
|
319
284
|
|
|
320
285
|
Args:
|
|
321
|
-
device_target (str, optional): The device
|
|
322
|
-
Default
|
|
286
|
+
device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
|
|
287
|
+
``"Ascend"`` . Default ``None`` , represents the current device set by context.
|
|
323
288
|
|
|
324
289
|
Returns:
|
|
325
290
|
int, in Byte.
|
|
326
291
|
|
|
327
292
|
Examples:
|
|
328
|
-
>>> import mindspore
|
|
329
|
-
>>>
|
|
330
|
-
>>>
|
|
331
|
-
>>>
|
|
332
|
-
>>>
|
|
333
|
-
>>> c = ops.add(a, b).asnumpy()
|
|
334
|
-
>>> print(ms.hal.max_memory_allocated())
|
|
293
|
+
>>> import mindspore
|
|
294
|
+
>>> a = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
295
|
+
>>> b = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
296
|
+
>>> c = mindspore.ops.add(a, b).asnumpy()
|
|
297
|
+
>>> print(mindspore.hal.max_memory_allocated())
|
|
335
298
|
1536
|
|
336
299
|
"""
|
|
337
300
|
if not function_memory_status['max_memory_allocated']:
|
|
@@ -346,29 +309,22 @@ def max_memory_allocated(device_target=None):
|
|
|
346
309
|
@_check_inputs_validation
|
|
347
310
|
def reset_max_memory_reserved(device_target=None):
|
|
348
311
|
"""
|
|
349
|
-
Reset the peak memory size managed by the memory pool.
|
|
350
|
-
|
|
351
|
-
Note:
|
|
352
|
-
- The api will be deprecated, please use the api :func:`mindspore.runtime.reset_max_memory_reserved` instead.
|
|
353
|
-
|
|
354
|
-
Note:
|
|
355
|
-
If `device_target` is not specified, get the device capability of the current backend set by context.
|
|
312
|
+
Reset the peak memory size managed by the memory pool, this api will be deprecated and removed in future versions.
|
|
313
|
+
Please use the api :func:`mindspore.runtime.reset_max_memory_reserved` instead.
|
|
356
314
|
|
|
357
315
|
Args:
|
|
358
|
-
device_target (str, optional): The device
|
|
359
|
-
Default
|
|
316
|
+
device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
|
|
317
|
+
``"Ascend"`` . Default ``None`` , represents the current device set by context.
|
|
360
318
|
|
|
361
319
|
Examples:
|
|
362
|
-
>>> import mindspore
|
|
363
|
-
>>>
|
|
364
|
-
>>>
|
|
365
|
-
>>>
|
|
366
|
-
>>>
|
|
367
|
-
>>> c = ops.add(a, b).asnumpy()
|
|
368
|
-
>>> print(ms.hal.max_memory_reserved())
|
|
320
|
+
>>> import mindspore
|
|
321
|
+
>>> a = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
322
|
+
>>> b = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
323
|
+
>>> c = mindspore.ops.add(a, b).asnumpy()
|
|
324
|
+
>>> print(mindspore.hal.max_memory_reserved())
|
|
369
325
|
1073741824
|
|
370
|
-
>>>
|
|
371
|
-
>>> print(
|
|
326
|
+
>>> mindspore.hal.reset_max_memory_reserved()
|
|
327
|
+
>>> print(mindspore.hal.max_memory_reserved())
|
|
372
328
|
0
|
|
373
329
|
"""
|
|
374
330
|
if not function_memory_status['reset_max_memory_reserved']:
|
|
@@ -383,29 +339,22 @@ def reset_max_memory_reserved(device_target=None):
|
|
|
383
339
|
@_check_inputs_validation
|
|
384
340
|
def reset_max_memory_allocated(device_target=None):
|
|
385
341
|
"""
|
|
386
|
-
Reset the peak memory size of the memory pool actually occupied by Tensor
|
|
387
|
-
|
|
388
|
-
Note:
|
|
389
|
-
- The api will be deprecated, please use the api :func:`mindspore.runtime.reset_max_memory_allocated` instead.
|
|
390
|
-
|
|
391
|
-
Note:
|
|
392
|
-
If `device_target` is not specified, get the device capability of the current backend set by context.
|
|
342
|
+
Reset the peak memory size of the memory pool actually occupied by Tensor, this api will be deprecated and removed
|
|
343
|
+
in future versions, please use the api :func:`mindspore.runtime.reset_max_memory_allocated` instead.
|
|
393
344
|
|
|
394
345
|
Args:
|
|
395
|
-
device_target (str, optional): The device
|
|
396
|
-
Default
|
|
346
|
+
device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
|
|
347
|
+
``"Ascend"`` . Default ``None`` , represents the current device set by context.
|
|
397
348
|
|
|
398
349
|
Examples:
|
|
399
|
-
>>> import mindspore
|
|
400
|
-
>>>
|
|
401
|
-
>>>
|
|
402
|
-
>>>
|
|
403
|
-
>>>
|
|
404
|
-
>>> c = ops.add(a, b).asnumpy()
|
|
405
|
-
>>> print(ms.hal.max_memory_allocated())
|
|
350
|
+
>>> import mindspore
|
|
351
|
+
>>> a = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
352
|
+
>>> b = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
353
|
+
>>> c = mindspore.ops.add(a, b).asnumpy()
|
|
354
|
+
>>> print(mindspore.hal.max_memory_allocated())
|
|
406
355
|
1536
|
|
407
|
-
>>>
|
|
408
|
-
>>> print(
|
|
356
|
+
>>> mindspore.hal.reset_max_memory_allocated()
|
|
357
|
+
>>> print(mindspore.hal.max_memory_allocated())
|
|
409
358
|
0
|
|
410
359
|
"""
|
|
411
360
|
if not function_memory_status['reset_max_memory_allocated']:
|