mindspore 2.5.0__cp311-cp311-win_amd64.whl → 2.6.0__cp311-cp311-win_amd64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of mindspore might be problematic. Click here for more details.
- mindspore/.commit_id +1 -1
- mindspore/Microsoft.VisualStudio.Telemetry.dll +0 -0
- mindspore/Newtonsoft.Json.dll +0 -0
- mindspore/__init__.py +6 -4
- mindspore/_c_dataengine.cp311-win_amd64.pyd +0 -0
- mindspore/_c_expression.cp311-win_amd64.pyd +0 -0
- mindspore/_c_mindrecord.cp311-win_amd64.pyd +0 -0
- mindspore/_check_jit_forbidden_api.py +3 -0
- mindspore/_checkparam.py +3 -33
- mindspore/_deprecated/__init__.py +17 -0
- mindspore/_deprecated/jit.py +198 -0
- mindspore/_extends/builtin_operations.py +1 -1
- mindspore/_extends/parse/__init__.py +6 -7
- mindspore/_extends/parse/compile_config.py +19 -0
- mindspore/_extends/parse/deprecated/deprecated_tensor_method.py +22 -3
- mindspore/_extends/parse/jit_fallback_modules/__init__.py +0 -0
- mindspore/_extends/parse/jit_fallback_modules/check_utils.py +123 -0
- mindspore/_extends/parse/jit_fallback_modules/third_party_modules.py +50 -0
- mindspore/_extends/parse/parser.py +25 -194
- mindspore/_extends/parse/resources.py +1 -5
- mindspore/_extends/parse/standard_method.py +109 -75
- mindspore/_extends/pijit/__init__.py +2 -2
- mindspore/_extends/pijit/pijit_func_white_list.py +16 -11
- mindspore/_extends/pijit/tensor_func_list.py +27 -0
- mindspore/_extends/utils.py +1 -1
- mindspore/amp.py +4 -4
- mindspore/atlprov.dll +0 -0
- mindspore/avcodec-59.dll +0 -0
- mindspore/avdevice-59.dll +0 -0
- mindspore/avfilter-8.dll +0 -0
- mindspore/avformat-59.dll +0 -0
- mindspore/avutil-57.dll +0 -0
- mindspore/boost/__init__.py +2 -2
- mindspore/boost/base.py +3 -7
- mindspore/boost/boost_cell_wrapper.py +2 -2
- mindspore/c1.dll +0 -0
- mindspore/c1xx.dll +0 -0
- mindspore/c2.dll +0 -0
- mindspore/common/__init__.py +4 -3
- mindspore/common/_grad_function.py +56 -0
- mindspore/common/_pijit_context.py +14 -5
- mindspore/common/_register_for_tensor.py +1 -1
- mindspore/common/_stub_tensor.py +5 -10
- mindspore/common/_tensor_cpp_method.py +1 -1
- mindspore/common/_tensor_docs.py +2014 -3386
- mindspore/common/api.py +386 -355
- mindspore/common/auto_dynamic_shape.py +41 -44
- mindspore/common/dtype.py +5 -2
- mindspore/common/dump.py +7 -5
- mindspore/common/file_system.py +3 -0
- mindspore/common/generator.py +3 -0
- mindspore/common/hook_handle.py +5 -3
- mindspore/common/initializer.py +10 -6
- mindspore/common/jit_begin_end.py +94 -0
- mindspore/common/jit_config.py +6 -1
- mindspore/common/jit_context.py +76 -0
- mindspore/common/jit_trace.py +378 -0
- mindspore/common/lazy_inline.py +2 -2
- mindspore/common/mutable.py +5 -4
- mindspore/common/parameter.py +106 -39
- mindspore/common/seed.py +2 -2
- mindspore/common/sparse_tensor.py +23 -17
- mindspore/common/tensor.py +332 -714
- mindspore/communication/__init__.py +7 -5
- mindspore/communication/_comm_helper.py +47 -2
- mindspore/communication/comm_func.py +70 -53
- mindspore/communication/management.py +83 -17
- mindspore/context.py +228 -571
- mindspore/dataset/__init__.py +44 -20
- mindspore/dataset/audio/__init__.py +2 -8
- mindspore/dataset/audio/transforms.py +3 -17
- mindspore/dataset/core/config.py +3 -3
- mindspore/dataset/engine/cache_client.py +1 -1
- mindspore/dataset/engine/datasets.py +102 -120
- mindspore/dataset/engine/datasets_audio.py +22 -22
- mindspore/dataset/engine/datasets_standard_format.py +43 -24
- mindspore/dataset/engine/datasets_text.py +78 -85
- mindspore/dataset/engine/datasets_user_defined.py +109 -77
- mindspore/dataset/engine/datasets_vision.py +111 -108
- mindspore/dataset/engine/iterators.py +5 -3
- mindspore/dataset/engine/obs/obs_mindrecord_dataset.py +1 -1
- mindspore/dataset/engine/samplers.py +279 -57
- mindspore/dataset/engine/serializer_deserializer.py +2 -1
- mindspore/dataset/engine/validators.py +10 -0
- mindspore/dataset/text/__init__.py +7 -6
- mindspore/dataset/text/transforms.py +6 -5
- mindspore/dataset/text/utils.py +3 -3
- mindspore/dataset/transforms/__init__.py +0 -9
- mindspore/dataset/transforms/transforms.py +3 -3
- mindspore/dataset/utils/browse_dataset.py +1 -1
- mindspore/dataset/vision/__init__.py +2 -9
- mindspore/dataset/vision/transforms.py +202 -158
- mindspore/dataset/vision/utils.py +7 -5
- mindspore/device_context/ascend/op_debug.py +60 -1
- mindspore/device_context/ascend/op_tuning.py +0 -4
- mindspore/device_manager.py +39 -3
- mindspore/dnnl.dll +0 -0
- mindspore/dpcmi.dll +0 -0
- mindspore/experimental/es/embedding_service.py +35 -27
- mindspore/experimental/llm_boost/ascend_native/llama_boost_ascend_native.py +0 -2
- mindspore/experimental/map_parameter.py +4 -4
- mindspore/experimental/optim/adadelta.py +22 -26
- mindspore/experimental/optim/adagrad.py +4 -4
- mindspore/experimental/optim/adam.py +4 -0
- mindspore/experimental/optim/adamax.py +4 -4
- mindspore/experimental/optim/adamw.py +4 -0
- mindspore/experimental/optim/asgd.py +1 -1
- mindspore/experimental/optim/lr_scheduler.py +40 -22
- mindspore/experimental/optim/radam.py +5 -5
- mindspore/experimental/optim/rprop.py +1 -1
- mindspore/experimental/optim/sgd.py +1 -1
- mindspore/hal/contiguous_tensors_handle.py +6 -10
- mindspore/hal/device.py +55 -81
- mindspore/hal/event.py +38 -55
- mindspore/hal/memory.py +115 -147
- mindspore/hal/stream.py +81 -125
- mindspore/include/dataset/constants.h +7 -4
- mindspore/include/dataset/execute.h +2 -2
- mindspore/jpeg62.dll +0 -0
- mindspore/log.py +40 -2
- mindspore/mindrecord/__init__.py +20 -7
- mindspore/mindspore_backend_common.dll +0 -0
- mindspore/mindspore_backend_manager.dll +0 -0
- mindspore/mindspore_common.dll +0 -0
- mindspore/mindspore_core.dll +0 -0
- mindspore/mindspore_dump.dll +0 -0
- mindspore/mindspore_frontend.dll +0 -0
- mindspore/mindspore_glog.dll +0 -0
- mindspore/mindspore_memory_pool.dll +0 -0
- mindspore/mindspore_ms_backend.dll +0 -0
- mindspore/mindspore_ops.dll +0 -0
- mindspore/{mindspore_backend.dll → mindspore_ops_host.dll} +0 -0
- mindspore/mindspore_ops_kernel_common.dll +0 -0
- mindspore/mindspore_profiler.dll +0 -0
- mindspore/mindspore_pyboost.dll +0 -0
- mindspore/mindspore_pynative.dll +0 -0
- mindspore/mindspore_res_manager.dll +0 -0
- mindspore/mindspore_runtime_pipeline.dll +0 -0
- mindspore/mint/__init__.py +133 -702
- mindspore/mint/distributed/__init__.py +5 -1
- mindspore/mint/distributed/distributed.py +198 -113
- mindspore/mint/linalg/__init__.py +2 -0
- mindspore/mint/nn/__init__.py +280 -18
- mindspore/mint/nn/functional.py +282 -64
- mindspore/mint/nn/layer/__init__.py +4 -0
- mindspore/mint/nn/layer/_functions.py +7 -3
- mindspore/mint/nn/layer/activation.py +120 -13
- mindspore/mint/nn/layer/conv.py +234 -28
- mindspore/mint/nn/layer/normalization.py +15 -16
- mindspore/mint/nn/layer/padding.py +1 -1
- mindspore/mint/nn/layer/pooling.py +66 -1
- mindspore/mint/optim/__init__.py +2 -1
- mindspore/mint/optim/sgd.py +171 -0
- mindspore/msobj140.dll +0 -0
- mindspore/mspdb140.dll +0 -0
- mindspore/mspdbcore.dll +0 -0
- mindspore/mspdbst.dll +0 -0
- mindspore/mspft140.dll +0 -0
- mindspore/msvcdis140.dll +0 -0
- mindspore/msvcp140_1.dll +0 -0
- mindspore/msvcp140_2.dll +0 -0
- mindspore/msvcp140_atomic_wait.dll +0 -0
- mindspore/msvcp140_codecvt_ids.dll +0 -0
- mindspore/nn/__init__.py +4 -1
- mindspore/nn/cell.py +1253 -179
- mindspore/nn/layer/activation.py +23 -21
- mindspore/nn/layer/basic.py +22 -16
- mindspore/nn/layer/container.py +1 -1
- mindspore/nn/layer/conv.py +53 -42
- mindspore/nn/layer/embedding.py +9 -8
- mindspore/nn/layer/normalization.py +48 -42
- mindspore/nn/layer/pooling.py +75 -31
- mindspore/nn/layer/transformer.py +11 -10
- mindspore/nn/learning_rate_schedule.py +4 -2
- mindspore/nn/loss/loss.py +27 -19
- mindspore/nn/optim/ada_grad.py +6 -5
- mindspore/nn/optim/adadelta.py +9 -7
- mindspore/nn/optim/adafactor.py +1 -1
- mindspore/nn/optim/adam.py +18 -14
- mindspore/nn/optim/adamax.py +8 -7
- mindspore/nn/optim/adasum.py +5 -5
- mindspore/nn/optim/asgd.py +3 -1
- mindspore/nn/optim/ftrl.py +11 -9
- mindspore/nn/optim/lamb.py +1 -1
- mindspore/nn/optim/lazyadam.py +12 -10
- mindspore/nn/optim/momentum.py +7 -6
- mindspore/nn/optim/optimizer.py +2 -2
- mindspore/nn/optim/proximal_ada_grad.py +12 -10
- mindspore/nn/optim/rmsprop.py +13 -12
- mindspore/nn/optim/rprop.py +9 -7
- mindspore/nn/optim/sgd.py +9 -6
- mindspore/nn/optim/tft_wrapper.py +5 -2
- mindspore/nn/probability/bijector/bijector.py +17 -11
- mindspore/nn/probability/bijector/gumbel_cdf.py +5 -5
- mindspore/nn/probability/bijector/invert.py +2 -2
- mindspore/nn/probability/bijector/scalar_affine.py +3 -3
- mindspore/nn/probability/bijector/softplus.py +3 -2
- mindspore/nn/probability/distribution/beta.py +3 -3
- mindspore/nn/probability/distribution/categorical.py +1 -1
- mindspore/nn/probability/distribution/cauchy.py +4 -2
- mindspore/nn/probability/distribution/exponential.py +6 -7
- mindspore/nn/probability/distribution/gamma.py +2 -2
- mindspore/nn/probability/distribution/gumbel.py +2 -2
- mindspore/nn/probability/distribution/half_normal.py +5 -3
- mindspore/nn/probability/distribution/logistic.py +5 -3
- mindspore/nn/probability/distribution/poisson.py +1 -1
- mindspore/nn/probability/distribution/uniform.py +5 -3
- mindspore/nn/reinforcement/_tensors_queue.py +1 -1
- mindspore/nn/reinforcement/tensor_array.py +1 -1
- mindspore/nn/wrap/__init__.py +6 -6
- mindspore/nn/wrap/cell_wrapper.py +178 -117
- mindspore/nn/wrap/grad_reducer.py +45 -36
- mindspore/nn/wrap/loss_scale.py +3 -3
- mindspore/numpy/array_creations.py +3 -3
- mindspore/numpy/array_ops.py +1 -1
- mindspore/numpy/utils.py +1 -2
- mindspore/numpy/utils_const.py +1 -2
- mindspore/opencv_core452.dll +0 -0
- mindspore/opencv_imgcodecs452.dll +0 -0
- mindspore/opencv_imgproc452.dll +0 -0
- mindspore/ops/__init__.py +3 -2
- mindspore/ops/_grad_experimental/grad_comm_ops.py +18 -3
- mindspore/ops/_grad_experimental/grad_debug_ops.py +8 -1
- mindspore/ops/_grad_experimental/taylor_rule.py +29 -0
- mindspore/ops/_register_for_op.py +0 -11
- mindspore/{ops_generate → ops/_utils}/arg_dtype_cast.py +123 -4
- mindspore/{ops_generate → ops/_utils}/arg_handler.py +3 -4
- mindspore/ops/_vmap/vmap_array_ops.py +32 -6
- mindspore/ops/_vmap/vmap_grad_nn_ops.py +2 -1
- mindspore/ops/_vmap/vmap_math_ops.py +4 -7
- mindspore/ops/_vmap/vmap_nn_ops.py +9 -8
- mindspore/ops/auto_generate/__init__.py +4 -3
- mindspore/ops/auto_generate/cpp_create_prim_instance_helper.py +127 -52
- mindspore/ops/auto_generate/gen_extend_func.py +286 -208
- mindspore/ops/auto_generate/gen_ops_def.py +2783 -2335
- mindspore/ops/auto_generate/gen_ops_prim.py +8992 -2686
- mindspore/ops/auto_generate/pyboost_inner_prim.py +106 -76
- mindspore/ops/composite/__init__.py +2 -1
- mindspore/ops/composite/base.py +19 -24
- mindspore/ops/composite/math_ops.py +6 -16
- mindspore/ops/composite/multitype_ops/__init__.py +5 -2
- mindspore/ops/composite/multitype_ops/_compile_utils.py +4 -5
- mindspore/ops/composite/multitype_ops/_constexpr_utils.py +1 -2
- mindspore/ops/composite/multitype_ops/add_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/bitwise_and_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/bitwise_or_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/bitwise_xor_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/div_impl.py +6 -4
- mindspore/ops/composite/multitype_ops/equal_impl.py +4 -3
- mindspore/ops/composite/multitype_ops/floordiv_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/getitem_impl.py +3 -2
- mindspore/ops/composite/multitype_ops/greater_equal_impl.py +4 -3
- mindspore/ops/composite/multitype_ops/greater_impl.py +4 -3
- mindspore/ops/composite/multitype_ops/in_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/invert_impl.py +50 -0
- mindspore/ops/composite/multitype_ops/left_shift_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/less_equal_impl.py +4 -3
- mindspore/ops/composite/multitype_ops/less_impl.py +4 -3
- mindspore/ops/composite/multitype_ops/logic_not_impl.py +3 -2
- mindspore/ops/composite/multitype_ops/logical_and_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/logical_or_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/mod_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/mul_impl.py +3 -2
- mindspore/ops/composite/multitype_ops/negative_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/not_equal_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/not_in_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/ones_like_impl.py +18 -0
- mindspore/ops/composite/multitype_ops/pow_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/right_shift_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/setitem_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/sub_impl.py +2 -1
- mindspore/ops/function/__init__.py +28 -2
- mindspore/ops/function/_add_attr_func.py +58 -0
- mindspore/ops/function/array_func.py +1631 -2347
- mindspore/ops/function/clip_func.py +38 -45
- mindspore/ops/function/debug_func.py +36 -44
- mindspore/ops/function/grad/__init__.py +1 -0
- mindspore/ops/function/grad/grad_func.py +104 -71
- mindspore/ops/function/image_func.py +1 -1
- mindspore/ops/function/linalg_func.py +46 -78
- mindspore/ops/function/math_func.py +3024 -3855
- mindspore/ops/function/nn_func.py +678 -274
- mindspore/ops/function/other_func.py +159 -1
- mindspore/ops/function/parameter_func.py +17 -30
- mindspore/ops/function/random_func.py +216 -361
- mindspore/ops/function/reshard_func.py +4 -70
- mindspore/ops/function/sparse_func.py +3 -3
- mindspore/ops/function/sparse_unary_func.py +5 -5
- mindspore/ops/function/spectral_func.py +25 -58
- mindspore/ops/function/vmap_func.py +26 -18
- mindspore/ops/functional.py +8 -5
- mindspore/ops/functional_overload.py +655 -4
- mindspore/ops/op_info_register.py +32 -244
- mindspore/ops/operations/__init__.py +21 -14
- mindspore/ops/operations/_custom_ops_utils.py +235 -0
- mindspore/ops/operations/_grad_ops.py +1 -10
- mindspore/ops/operations/_inner_ops.py +5 -76
- mindspore/ops/operations/_ms_kernel.py +4 -10
- mindspore/ops/operations/_rl_inner_ops.py +1 -1
- mindspore/ops/operations/_scalar_ops.py +3 -2
- mindspore/ops/operations/_sequence_ops.py +1 -1
- mindspore/ops/operations/_tensor_array.py +1 -1
- mindspore/ops/operations/array_ops.py +39 -24
- mindspore/ops/operations/comm_ops.py +150 -107
- mindspore/ops/operations/custom_ops.py +287 -32
- mindspore/ops/operations/debug_ops.py +119 -16
- mindspore/ops/operations/inner_ops.py +1 -1
- mindspore/ops/operations/linalg_ops.py +1 -58
- mindspore/ops/operations/manually_defined/_inner.py +1 -1
- mindspore/ops/operations/manually_defined/ops_def.py +746 -79
- mindspore/ops/operations/math_ops.py +21 -18
- mindspore/ops/operations/nn_ops.py +67 -224
- mindspore/ops/operations/other_ops.py +62 -9
- mindspore/ops/operations/random_ops.py +13 -7
- mindspore/ops/operations/reshard_ops.py +1 -1
- mindspore/ops/operations/sparse_ops.py +2 -2
- mindspore/ops/primitive.py +43 -32
- mindspore/ops/tensor_method.py +243 -17
- mindspore/ops_generate/__init__.py +0 -5
- mindspore/ops_generate/aclnn/__init__.py +0 -0
- mindspore/ops_generate/{aclnn_kernel_register_auto_cc_generator.py → aclnn/aclnn_kernel_register_auto_cc_generator.py} +43 -18
- mindspore/ops_generate/{gen_aclnn_implement.py → aclnn/gen_aclnn_implement.py} +49 -51
- mindspore/ops_generate/api/__init__.py +0 -0
- mindspore/ops_generate/{add_tensor_docs_generator.py → api/add_tensor_docs_generator.py} +9 -7
- mindspore/ops_generate/{cpp_create_prim_instance_helper_generator.py → api/cpp_create_prim_instance_helper_generator.py} +6 -9
- mindspore/ops_generate/{functional_map_cpp_generator.py → api/functional_map_cpp_generator.py} +25 -12
- mindspore/ops_generate/{functional_overload_py_generator.py → api/functional_overload_py_generator.py} +8 -6
- mindspore/ops_generate/{functions_cc_generator.py → api/functions_cc_generator.py} +14 -10
- mindspore/ops_generate/api/gen_api.py +103 -0
- mindspore/ops_generate/{op_api_proto.py → api/op_api_proto.py} +98 -69
- mindspore/ops_generate/{tensor_func_reg_cpp_generator.py → api/tensor_func_reg_cpp_generator.py} +82 -43
- mindspore/ops_generate/common/__init__.py +0 -0
- mindspore/ops_generate/common/gen_constants.py +91 -0
- mindspore/ops_generate/{gen_utils.py → common/gen_utils.py} +72 -19
- mindspore/ops_generate/{op_proto.py → common/op_proto.py} +64 -1
- mindspore/ops_generate/{template.py → common/template.py} +96 -84
- mindspore/ops_generate/gen_ops.py +23 -325
- mindspore/ops_generate/op_def/__init__.py +0 -0
- mindspore/ops_generate/op_def/gen_op_def.py +90 -0
- mindspore/ops_generate/{lite_ops_cpp_generator.py → op_def/lite_ops_cpp_generator.py} +47 -11
- mindspore/ops_generate/{ops_def_cc_generator.py → op_def/ops_def_cc_generator.py} +18 -10
- mindspore/ops_generate/{ops_def_h_generator.py → op_def/ops_def_h_generator.py} +5 -5
- mindspore/ops_generate/{ops_name_h_generator.py → op_def/ops_name_h_generator.py} +30 -15
- mindspore/ops_generate/op_def/ops_primitive_h_generator.py +125 -0
- mindspore/ops_generate/op_def_py/__init__.py +0 -0
- mindspore/ops_generate/op_def_py/gen_op_def_py.py +47 -0
- mindspore/ops_generate/{op_def_py_generator.py → op_def_py/op_def_py_generator.py} +6 -5
- mindspore/ops_generate/{op_prim_py_generator.py → op_def_py/op_prim_py_generator.py} +24 -15
- mindspore/ops_generate/pyboost/__init__.py +0 -0
- mindspore/ops_generate/{auto_grad_impl_cc_generator.py → pyboost/auto_grad_impl_cc_generator.py} +11 -7
- mindspore/ops_generate/{auto_grad_reg_cc_generator.py → pyboost/auto_grad_reg_cc_generator.py} +7 -7
- mindspore/ops_generate/{gen_pyboost_func.py → pyboost/gen_pyboost_func.py} +40 -16
- mindspore/ops_generate/{op_template_parser.py → pyboost/op_template_parser.py} +105 -24
- mindspore/ops_generate/{pyboost_functions_cpp_generator.py → pyboost/pyboost_functions_cpp_generator.py} +55 -18
- mindspore/ops_generate/{pyboost_functions_h_generator.py → pyboost/pyboost_functions_h_generator.py} +42 -10
- mindspore/ops_generate/{pyboost_functions_py_generator.py → pyboost/pyboost_functions_py_generator.py} +6 -6
- mindspore/ops_generate/{pyboost_grad_function_cpp_generator.py → pyboost/pyboost_grad_function_cpp_generator.py} +11 -10
- mindspore/ops_generate/{pyboost_inner_prim_generator.py → pyboost/pyboost_inner_prim_generator.py} +8 -7
- mindspore/ops_generate/{pyboost_native_grad_functions_generator.py → pyboost/pyboost_native_grad_functions_generator.py} +14 -10
- mindspore/ops_generate/{pyboost_op_cpp_code_generator.py → pyboost/pyboost_op_cpp_code_generator.py} +140 -53
- mindspore/ops_generate/{pyboost_overload_functions_cpp_generator.py → pyboost/pyboost_overload_functions_cpp_generator.py} +28 -15
- mindspore/ops_generate/{pyboost_utils.py → pyboost/pyboost_utils.py} +88 -4
- mindspore/ops_generate/resources/__init__.py +0 -0
- mindspore/ops_generate/resources/resource_list.py +30 -0
- mindspore/ops_generate/resources/resource_loader.py +36 -0
- mindspore/ops_generate/resources/resource_manager.py +64 -0
- mindspore/ops_generate/resources/yaml_loader.py +88 -0
- mindspore/ops_generate/tensor_py_cc_generator.py +122 -0
- mindspore/parallel/__init__.py +6 -2
- mindspore/parallel/_auto_parallel_context.py +140 -12
- mindspore/parallel/_cell_wrapper.py +132 -15
- mindspore/parallel/_parallel_serialization.py +95 -4
- mindspore/parallel/_ps_context.py +1 -1
- mindspore/parallel/_recovery_context.py +7 -2
- mindspore/parallel/_tensor.py +142 -18
- mindspore/parallel/_utils.py +198 -25
- mindspore/parallel/algo_parameter_config.py +3 -3
- mindspore/parallel/auto_parallel.py +732 -0
- mindspore/parallel/checkpoint_convert.py +159 -0
- mindspore/parallel/checkpoint_transform.py +658 -37
- mindspore/parallel/cluster/process_entity/_api.py +151 -19
- mindspore/parallel/cluster/run.py +1 -1
- mindspore/parallel/function/__init__.py +24 -0
- mindspore/parallel/function/reshard_func.py +258 -0
- mindspore/parallel/nn/__init__.py +25 -0
- mindspore/parallel/nn/parallel_cell_wrapper.py +263 -0
- mindspore/parallel/nn/parallel_grad_reducer.py +169 -0
- mindspore/parallel/parameter_broadcast.py +24 -13
- mindspore/parallel/shard.py +137 -62
- mindspore/parallel/transform_safetensors.py +288 -95
- mindspore/pgodb140.dll +0 -0
- mindspore/pgort140.dll +0 -0
- mindspore/profiler/__init__.py +9 -5
- mindspore/profiler/analysis/parser/ascend_cann_parser.py +6 -2
- mindspore/profiler/analysis/parser/ms_framework_parser.py +4 -4
- mindspore/profiler/analysis/parser/timeline_assembly_factory/ascend_timeline_assembler.py +7 -4
- mindspore/profiler/analysis/parser/timeline_assembly_factory/trace_view_container.py +25 -0
- mindspore/profiler/analysis/parser/timeline_creator/fwk_timeline_creator.py +3 -3
- mindspore/profiler/analysis/parser/timeline_event/fwk_event.py +241 -86
- mindspore/profiler/analysis/viewer/ascend_communication_viewer.py +41 -2
- mindspore/profiler/analysis/viewer/ascend_kernel_details_viewer.py +33 -35
- mindspore/profiler/analysis/viewer/ascend_memory_viewer.py +7 -0
- mindspore/profiler/analysis/viewer/ascend_op_memory_viewer.py +8 -3
- mindspore/profiler/analysis/viewer/ascend_step_trace_time_viewer.py +141 -30
- mindspore/profiler/analysis/viewer/ms_dataset_viewer.py +5 -6
- mindspore/profiler/common/ascend_msprof_exporter.py +5 -4
- mindspore/profiler/common/constant.py +12 -0
- mindspore/profiler/common/msprof_cmd_tool.py +42 -23
- mindspore/profiler/common/path_manager.py +24 -0
- mindspore/profiler/common/profiler_context.py +26 -2
- mindspore/profiler/common/profiler_meta_data.py +74 -0
- mindspore/profiler/common/profiler_parameters.py +59 -18
- mindspore/profiler/common/profiler_path_manager.py +66 -7
- mindspore/profiler/dynamic_profiler.py +112 -79
- mindspore/profiler/envprofiler.py +26 -1
- mindspore/profiler/experimental_config.py +197 -0
- mindspore/profiler/mstx.py +57 -14
- mindspore/profiler/platform/npu_profiler.py +33 -7
- mindspore/profiler/profiler.py +541 -45
- mindspore/profiler/profiler_action_controller.py +1 -1
- mindspore/profiler/profiler_interface.py +4 -0
- mindspore/profiler/schedule.py +57 -22
- mindspore/rewrite/api/node.py +15 -13
- mindspore/rewrite/api/symbol_tree.py +1 -1
- mindspore/run_check/_check_version.py +25 -14
- mindspore/run_check/run_check.py +1 -1
- mindspore/runtime/__init__.py +2 -2
- mindspore/runtime/executor.py +40 -11
- mindspore/runtime/memory.py +37 -13
- mindspore/safeguard/rewrite_obfuscation.py +12 -9
- mindspore/swresample-4.dll +0 -0
- mindspore/swscale-6.dll +0 -0
- mindspore/tbbmalloc.dll +0 -0
- mindspore/tinyxml2.dll +0 -0
- mindspore/train/__init__.py +8 -8
- mindspore/train/_utils.py +43 -9
- mindspore/train/amp.py +1 -1
- mindspore/train/callback/__init__.py +2 -2
- mindspore/train/callback/_callback.py +2 -16
- mindspore/train/callback/_checkpoint.py +24 -40
- mindspore/train/callback/_cluster_monitor.py +14 -18
- mindspore/train/callback/_flops_collector.py +2 -3
- mindspore/train/callback/_history.py +7 -4
- mindspore/train/callback/_lambda_callback.py +2 -2
- mindspore/train/callback/_landscape.py +0 -3
- mindspore/train/callback/_loss_monitor.py +2 -1
- mindspore/train/callback/_on_request_exit.py +6 -5
- mindspore/train/callback/_reduce_lr_on_plateau.py +11 -6
- mindspore/train/callback/_summary_collector.py +8 -13
- mindspore/train/callback/_time_monitor.py +2 -1
- mindspore/train/callback/{_tft_register.py → _train_fault_tolerance.py} +204 -105
- mindspore/train/data_sink.py +25 -2
- mindspore/train/dataset_helper.py +4 -5
- mindspore/train/loss_scale_manager.py +8 -7
- mindspore/train/metrics/accuracy.py +3 -3
- mindspore/train/metrics/confusion_matrix.py +9 -9
- mindspore/train/metrics/error.py +3 -3
- mindspore/train/metrics/hausdorff_distance.py +4 -4
- mindspore/train/metrics/mean_surface_distance.py +3 -3
- mindspore/train/metrics/metric.py +0 -12
- mindspore/train/metrics/occlusion_sensitivity.py +4 -2
- mindspore/train/metrics/precision.py +8 -6
- mindspore/train/metrics/recall.py +9 -9
- mindspore/train/metrics/root_mean_square_surface_distance.py +2 -2
- mindspore/train/mind_ir_pb2.py +19 -12
- mindspore/train/model.py +262 -127
- mindspore/train/serialization.py +246 -988
- mindspore/train/summary/_summary_adapter.py +2 -2
- mindspore/train/summary/summary_record.py +1 -1
- mindspore/turbojpeg.dll +0 -0
- mindspore/utils/__init__.py +3 -2
- mindspore/utils/dryrun.py +4 -2
- mindspore/utils/hooks.py +81 -0
- mindspore/utils/runtime_execution_order_check.py +2 -0
- mindspore/utils/utils.py +138 -4
- mindspore/vcmeta.dll +0 -0
- mindspore/vcruntime140.dll +0 -0
- mindspore/vcruntime140_1.dll +0 -0
- mindspore/version.py +1 -1
- {mindspore-2.5.0.dist-info → mindspore-2.6.0.dist-info}/METADATA +2 -1
- {mindspore-2.5.0.dist-info → mindspore-2.6.0.dist-info}/RECORD +485 -440
- mindspore/_install_custom.py +0 -43
- mindspore/common/_register_for_adapter.py +0 -74
- mindspore/ops/auto_generate/gen_arg_dtype_cast.py +0 -252
- mindspore/ops/auto_generate/gen_arg_handler.py +0 -136
- mindspore/ops/operations/_opaque_predicate_registry.py +0 -41
- mindspore/ops_generate/gen_constants.py +0 -190
- mindspore/ops_generate/gen_ops_inner_prim.py +0 -131
- mindspore/ops_generate/ops_primitive_h_generator.py +0 -81
- /mindspore/ops_generate/{base_generator.py → common/base_generator.py} +0 -0
- {mindspore-2.5.0.dist-info → mindspore-2.6.0.dist-info}/WHEEL +0 -0
- {mindspore-2.5.0.dist-info → mindspore-2.6.0.dist-info}/entry_points.txt +0 -0
- {mindspore-2.5.0.dist-info → mindspore-2.6.0.dist-info}/top_level.txt +0 -0
mindspore/hal/event.py
CHANGED
|
@@ -24,41 +24,37 @@ function_event_status = {'Event': False, 'wait': False}
|
|
|
24
24
|
|
|
25
25
|
class Event(Event_):
|
|
26
26
|
r"""
|
|
27
|
-
Wrapper around a device event
|
|
27
|
+
Wrapper around a device event, this api will be deprecated and removed in future versions, please use
|
|
28
|
+
the api :func:`mindspore.runtime.Event` instead.
|
|
28
29
|
|
|
29
30
|
Device events are synchronization markers that can be used to monitor the device’s progress,
|
|
30
31
|
to accurately measure timing, and to synchronize device streams.
|
|
31
32
|
|
|
32
33
|
The underlying device events are lazily initialized when the event is first recorded.
|
|
33
34
|
|
|
34
|
-
Note:
|
|
35
|
-
- The api will be deprecated, please use the api :func:`mindspore.runtime.Event` instead.
|
|
36
|
-
|
|
37
35
|
Args:
|
|
38
|
-
enable_timing (bool, optional): indicates if the event should measure time
|
|
39
|
-
blocking (bool, optional): if ``True``, `wait` will be blocking
|
|
36
|
+
enable_timing (bool, optional): indicates if the event should measure time. Default ``False``.
|
|
37
|
+
blocking (bool, optional): if ``True``, `wait` will be blocking. Default ``False``.
|
|
40
38
|
|
|
41
39
|
Examples:
|
|
42
|
-
>>> import mindspore
|
|
43
|
-
>>>
|
|
44
|
-
>>>
|
|
45
|
-
>>>
|
|
46
|
-
>>>
|
|
47
|
-
>>>
|
|
48
|
-
>>>
|
|
49
|
-
>>>
|
|
50
|
-
>>>
|
|
51
|
-
|
|
52
|
-
>>> with ms.hal.StreamCtx(s1):
|
|
53
|
-
... d = ops.matmul(a, b)
|
|
40
|
+
>>> import mindspore
|
|
41
|
+
>>> start = mindspore.hal.Event(enable_timing=True)
|
|
42
|
+
>>> end = mindspore.hal.Event(enable_timing=True)
|
|
43
|
+
>>> s1 = mindspore.hal.Stream()
|
|
44
|
+
>>> s2 = mindspore.hal.Stream()
|
|
45
|
+
>>> a = mindspore.tensor(mindspore.ops.ones([2, 2]), mindspore.float32)
|
|
46
|
+
>>> b = mindspore.tensor(mindspore.ops.ones([2, 2]), mindspore.float32)
|
|
47
|
+
>>> c = mindspore.tensor(mindspore.ops.ones([2, 2]), mindspore.float32)
|
|
48
|
+
>>> with mindspore.hal.StreamCtx(s1):
|
|
49
|
+
... d = mindspore.ops.matmul(a, b)
|
|
54
50
|
... start.record()
|
|
55
51
|
>>> c += 2
|
|
56
52
|
>>> end.record()
|
|
57
|
-
>>> with
|
|
53
|
+
>>> with mindspore.hal.StreamCtx(s2):
|
|
58
54
|
... start.synchronize()
|
|
59
55
|
... end.synchronize()
|
|
60
56
|
... e = c + d
|
|
61
|
-
>>>
|
|
57
|
+
>>> mindspore.hal.synchronize()
|
|
62
58
|
>>> print(e)
|
|
63
59
|
[[5. 5.]
|
|
64
60
|
[5. 5.]]
|
|
@@ -78,17 +74,14 @@ class Event(Event_):
|
|
|
78
74
|
|
|
79
75
|
def record(self, stream=None):
|
|
80
76
|
r"""
|
|
81
|
-
|
|
77
|
+
Record the event in a given stream.
|
|
82
78
|
|
|
83
79
|
Uses :func:`mindspore.hal.current_stream()` if no `stream` is specified. The
|
|
84
80
|
stream's device must match the event's device.
|
|
85
81
|
|
|
86
82
|
Args:
|
|
87
83
|
stream (Stream, optional): a stream to record. If this argument is ``None``,
|
|
88
|
-
current stream will be used. Default
|
|
89
|
-
|
|
90
|
-
Raises:
|
|
91
|
-
TypeError: If 'stream' is neither a :class:`mindspore.hal.Stream` nor a ``None``.
|
|
84
|
+
current stream will be used. Default ``None``.
|
|
92
85
|
"""
|
|
93
86
|
if stream is None:
|
|
94
87
|
stream = current_stream_()
|
|
@@ -99,33 +92,28 @@ class Event(Event_):
|
|
|
99
92
|
|
|
100
93
|
def wait(self, stream=None):
|
|
101
94
|
r"""
|
|
102
|
-
|
|
95
|
+
Make all future work submitted to the given stream wait for this
|
|
103
96
|
event.
|
|
104
97
|
|
|
105
98
|
Use :func:`mindspore.hal.current_stream()` if no `stream` is specified.
|
|
106
99
|
|
|
107
100
|
Args:
|
|
108
101
|
stream (Stream, optional): a stream to record. If this argument is ``None``,
|
|
109
|
-
current stream will be used. Default
|
|
110
|
-
|
|
111
|
-
Raises:
|
|
112
|
-
TypeError: If 'stream' is neither a :class:`mindspore.hal.Stream` nor a ``None``.
|
|
102
|
+
current stream will be used. Default ``None``.
|
|
113
103
|
|
|
114
104
|
Examples:
|
|
115
|
-
>>> import mindspore
|
|
116
|
-
>>>
|
|
117
|
-
>>>
|
|
118
|
-
>>>
|
|
119
|
-
>>>
|
|
120
|
-
>>>
|
|
121
|
-
>>>
|
|
122
|
-
|
|
123
|
-
>>> with ms.hal.StreamCtx(s1):
|
|
124
|
-
... c = ops.matmul(a, b)
|
|
105
|
+
>>> import mindspore
|
|
106
|
+
>>> event = mindspore.hal.Event()
|
|
107
|
+
>>> s1 = mindspore.hal.Stream()
|
|
108
|
+
>>> s2 = mindspore.hal.Stream()
|
|
109
|
+
>>> a = mindspore.tensor(mindspore.ops.ones([2, 2]), mindspore.float32)
|
|
110
|
+
>>> b = mindspore.tensor(mindspore.ops.ones([2, 2]), mindspore.float32)
|
|
111
|
+
>>> with mindspore.hal.StreamCtx(s1):
|
|
112
|
+
... c = mindspore.ops.matmul(a, b)
|
|
125
113
|
... event.record()
|
|
126
114
|
>>> event.wait()
|
|
127
115
|
>>> d = c + 2
|
|
128
|
-
>>>
|
|
116
|
+
>>> mindspore.hal.synchronize()
|
|
129
117
|
>>> print(d)
|
|
130
118
|
[[4. 4.]
|
|
131
119
|
[4. 4.]]
|
|
@@ -144,7 +132,7 @@ class Event(Event_):
|
|
|
144
132
|
|
|
145
133
|
def synchronize(self):
|
|
146
134
|
r"""
|
|
147
|
-
|
|
135
|
+
Wait for the event to complete.
|
|
148
136
|
|
|
149
137
|
Waits until the completion of all work currently captured in this event.
|
|
150
138
|
This prevents the CPU thread from proceeding until the event completes.
|
|
@@ -154,20 +142,18 @@ class Event(Event_):
|
|
|
154
142
|
|
|
155
143
|
def query(self):
|
|
156
144
|
r"""
|
|
157
|
-
|
|
145
|
+
Check if all work currently captured by event has completed.
|
|
158
146
|
|
|
159
147
|
Returns:
|
|
160
148
|
A boolean indicating if all work currently captured by event has completed.
|
|
161
149
|
|
|
162
150
|
Examples:
|
|
163
|
-
>>> import mindspore
|
|
164
|
-
>>>
|
|
165
|
-
>>>
|
|
166
|
-
>>>
|
|
167
|
-
>>>
|
|
168
|
-
|
|
169
|
-
>>> with ms.hal.StreamCtx(s1):
|
|
170
|
-
... c = ops.matmul(a, b)
|
|
151
|
+
>>> import mindspore
|
|
152
|
+
>>> a = mindspore.tensor(mindspore.ops.ones([1024, 2048]), mindspore.float32)
|
|
153
|
+
>>> b = mindspore.tensor(mindspore.ops.ones([2048, 4096]), mindspore.float32)
|
|
154
|
+
>>> s1 = mindspore.hal.Stream()
|
|
155
|
+
>>> with mindspore.hal.StreamCtx(s1):
|
|
156
|
+
... c = mindspore.ops.matmul(a, b)
|
|
171
157
|
... ev = s1.record_event()
|
|
172
158
|
>>> s1.synchronize()
|
|
173
159
|
>>> assert ev.query()
|
|
@@ -177,7 +163,7 @@ class Event(Event_):
|
|
|
177
163
|
|
|
178
164
|
def elapsed_time(self, end_event):
|
|
179
165
|
r"""
|
|
180
|
-
|
|
166
|
+
Return the time elapsed in milliseconds after the event was
|
|
181
167
|
recorded and before the end_event was recorded.
|
|
182
168
|
|
|
183
169
|
Args:
|
|
@@ -185,9 +171,6 @@ class Event(Event_):
|
|
|
185
171
|
|
|
186
172
|
Returns:
|
|
187
173
|
float, the time elapsed in milliseconds.
|
|
188
|
-
|
|
189
|
-
Raises:
|
|
190
|
-
TypeError: If 'end_event' is not a :class:`mindspore.hal.Event`.
|
|
191
174
|
"""
|
|
192
175
|
# pylint: disable=useless-super-delegation
|
|
193
176
|
if not isinstance(end_event, Event):
|
mindspore/hal/memory.py
CHANGED
|
@@ -14,43 +14,42 @@
|
|
|
14
14
|
# ============================================================================
|
|
15
15
|
|
|
16
16
|
"""Hardware memory interfaces."""
|
|
17
|
-
from mindspore._c_expression import _memory_stats, _reset_max_mem_reserved, _reset_max_mem_allocated
|
|
17
|
+
from mindspore._c_expression import _memory_stats, _reset_max_mem_reserved, _reset_max_mem_allocated, _empty_cache, \
|
|
18
|
+
DeviceContextManager
|
|
18
19
|
from mindspore import log as logger
|
|
20
|
+
import mindspore as ms
|
|
19
21
|
from .device import _check_inputs_validation, is_initialized
|
|
20
22
|
|
|
23
|
+
|
|
21
24
|
function_memory_status = {'memory_stats': False, 'memory_reserved': False, 'max_memory_reserved': False,
|
|
22
25
|
'empty_cache': False, 'reset_peak_memory_stats': False, 'memory_summary': False,
|
|
23
26
|
'memory_allocated': False, 'max_memory_allocated': False,
|
|
24
27
|
'reset_max_memory_reserved': False, 'reset_max_memory_allocated': False}
|
|
28
|
+
_device_context_mgr = DeviceContextManager.get_instance()
|
|
25
29
|
|
|
26
30
|
|
|
27
31
|
@_check_inputs_validation
|
|
28
32
|
def memory_stats(device_target=None):
|
|
29
33
|
"""
|
|
30
|
-
Returns status information queried from the memory pool
|
|
31
|
-
|
|
32
|
-
Note:
|
|
33
|
-
- The api will be deprecated, please use the api :func:`mindspore.runtime.memory_stats` instead.
|
|
34
|
+
Returns status information queried from the memory pool, this api will be deprecated and removed in future
|
|
35
|
+
versions, please use the api :func:`mindspore.runtime.memory_stats` instead.
|
|
34
36
|
|
|
35
37
|
Note:
|
|
36
|
-
-
|
|
37
|
-
- For the `CPU` backend, a dictionary with empty data is always returned.
|
|
38
|
+
- For the `CPU` device, a dictionary with empty data is always returned.
|
|
38
39
|
|
|
39
40
|
Args:
|
|
40
|
-
device_target (str, optional): The device
|
|
41
|
-
Default
|
|
41
|
+
device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
|
|
42
|
+
``"Ascend"`` . Default ``None`` , represents the current device set by context.
|
|
42
43
|
|
|
43
44
|
Returns:
|
|
44
45
|
dict, the queried memory information.
|
|
45
46
|
|
|
46
47
|
Examples:
|
|
47
|
-
>>> import mindspore
|
|
48
|
-
>>>
|
|
49
|
-
>>>
|
|
50
|
-
>>>
|
|
51
|
-
>>>
|
|
52
|
-
>>> c = ops.add(a, b).asnumpy()
|
|
53
|
-
>>> print(ms.hal.memory_stats())
|
|
48
|
+
>>> import mindspore
|
|
49
|
+
>>> a = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
50
|
+
>>> b = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
51
|
+
>>> c = mindspore.ops.add(a, b).asnumpy()
|
|
52
|
+
>>> print(mindspore.hal.memory_stats())
|
|
54
53
|
{'total_reserved_memory': 1073741824, 'total_allocated_memory': 1024, 'total_idle_memory': 1073740800,
|
|
55
54
|
'total_eager_free_memory': 0, 'max_reserved_memory': 1073741824, 'max_allocated_memory': 1536,
|
|
56
55
|
'common_mem_pool_stats': {'block_unit_size': 1073741824, 'block_counts': 1, 'blocks_info':
|
|
@@ -72,30 +71,25 @@ def memory_stats(device_target=None):
|
|
|
72
71
|
@_check_inputs_validation
|
|
73
72
|
def memory_reserved(device_target=None):
|
|
74
73
|
"""
|
|
75
|
-
Returns the total amount of memory currently managed by the memory pool
|
|
76
|
-
|
|
77
|
-
Note:
|
|
78
|
-
- The api will be deprecated, please use the api :func:`mindspore.runtime.memory_reserved` instead.
|
|
74
|
+
Returns the total amount of memory currently managed by the memory pool, this api will be deprecated and removed in
|
|
75
|
+
future versions, please use the api :func:`mindspore.runtime.memory_reserved` instead.
|
|
79
76
|
|
|
80
77
|
Note:
|
|
81
|
-
-
|
|
82
|
-
- For the `CPU` backend, 0 is always returned.
|
|
78
|
+
- For the `CPU` device, 0 is always returned.
|
|
83
79
|
|
|
84
80
|
Args:
|
|
85
|
-
device_target (str, optional): The device
|
|
86
|
-
Default
|
|
81
|
+
device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
|
|
82
|
+
``"Ascend"`` . Default ``None`` , represents the current device set by context.
|
|
87
83
|
|
|
88
84
|
Returns:
|
|
89
85
|
int, in Byte.
|
|
90
86
|
|
|
91
87
|
Examples:
|
|
92
|
-
>>> import mindspore
|
|
93
|
-
>>>
|
|
94
|
-
>>>
|
|
95
|
-
>>>
|
|
96
|
-
>>>
|
|
97
|
-
>>> c = ops.add(a, b).asnumpy()
|
|
98
|
-
>>> print(ms.hal.memory_reserved())
|
|
88
|
+
>>> import mindspore
|
|
89
|
+
>>> a = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
90
|
+
>>> b = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
91
|
+
>>> c = mindspore.ops.add(a, b).asnumpy()
|
|
92
|
+
>>> print(mindspore.hal.memory_reserved())
|
|
99
93
|
1073741824
|
|
100
94
|
"""
|
|
101
95
|
if not function_memory_status['memory_reserved']:
|
|
@@ -111,29 +105,25 @@ def memory_reserved(device_target=None):
|
|
|
111
105
|
def max_memory_reserved(device_target=None):
|
|
112
106
|
"""
|
|
113
107
|
Returns the peak value of the total memory managed by the memory pool since the process was started.
|
|
108
|
+
This api will be deprecated and removed in future versions, please use
|
|
109
|
+
the api :func:`mindspore.runtime.max_memory_reserved` instead.
|
|
114
110
|
|
|
115
111
|
Note:
|
|
116
|
-
-
|
|
117
|
-
|
|
118
|
-
Note:
|
|
119
|
-
- If `device_target` is not specified, get the device capability of the current backend set by context.
|
|
120
|
-
- For the `CPU` backend, 0 is always returned.
|
|
112
|
+
- For the `CPU` device, 0 is always returned.
|
|
121
113
|
|
|
122
114
|
Args:
|
|
123
|
-
device_target (str, optional): The device
|
|
124
|
-
Default
|
|
115
|
+
device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
|
|
116
|
+
``"Ascend"`` . Default ``None`` , represents the current device set by context.
|
|
125
117
|
|
|
126
118
|
Returns:
|
|
127
119
|
int, in Byte.
|
|
128
120
|
|
|
129
121
|
Examples:
|
|
130
|
-
>>> import mindspore
|
|
131
|
-
>>>
|
|
132
|
-
>>>
|
|
133
|
-
>>>
|
|
134
|
-
>>>
|
|
135
|
-
>>> c = ops.add(a, b).asnumpy()
|
|
136
|
-
>>> print(ms.hal.max_memory_reserved())
|
|
122
|
+
>>> import mindspore
|
|
123
|
+
>>> a = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
124
|
+
>>> b = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
125
|
+
>>> c = mindspore.ops.add(a, b).asnumpy()
|
|
126
|
+
>>> print(mindspore.hal.max_memory_reserved())
|
|
137
127
|
1073741824
|
|
138
128
|
"""
|
|
139
129
|
if not function_memory_status['max_memory_reserved']:
|
|
@@ -145,54 +135,60 @@ def max_memory_reserved(device_target=None):
|
|
|
145
135
|
return _memory_stats(device_target).get("max_reserved_memory", 0)
|
|
146
136
|
|
|
147
137
|
|
|
138
|
+
def _is_initialized(device_target):
|
|
139
|
+
"""
|
|
140
|
+
Returns whether specified backend is initialized.
|
|
141
|
+
"""
|
|
142
|
+
_device_context = _device_context_mgr.get_device_context(device_target)
|
|
143
|
+
if _device_context is None:
|
|
144
|
+
return False
|
|
145
|
+
return _device_context.initialized()
|
|
146
|
+
|
|
148
147
|
@_check_inputs_validation
|
|
149
148
|
def empty_cache():
|
|
150
149
|
"""
|
|
151
|
-
|
|
152
|
-
|
|
150
|
+
Empty cache in the memory pool, this api will be deprecated and removed in future versions.
|
|
151
|
+
Please use the api :func:`mindspore.runtime.empty_cache` instead.
|
|
153
152
|
|
|
154
153
|
Note:
|
|
155
|
-
-
|
|
154
|
+
- Empty cache help reduce the fragmentation of device memory.
|
|
155
|
+
- Support Atlas A2 series products.
|
|
156
156
|
|
|
157
|
-
|
|
158
|
-
|
|
159
|
-
This interface is reserved but implemented as an empty method and prompted in log mode.
|
|
157
|
+
Supported Platforms:
|
|
158
|
+
``Ascend``
|
|
160
159
|
"""
|
|
161
160
|
if not function_memory_status['empty_cache']:
|
|
162
161
|
function_memory_status['empty_cache'] = True
|
|
163
|
-
|
|
162
|
+
device_target = ms.context.get_context("device_target")
|
|
163
|
+
if not _is_initialized(device_target):
|
|
164
|
+
logger.warning(f"Backend {device_target} is not initialized yet.")
|
|
165
|
+
return
|
|
166
|
+
_empty_cache(device_target)
|
|
164
167
|
|
|
165
168
|
|
|
166
169
|
@_check_inputs_validation
|
|
167
170
|
def reset_peak_memory_stats(device_target=None):
|
|
168
171
|
"""
|
|
169
|
-
Reset the "peak" stats tracked by memory manager.
|
|
170
|
-
|
|
171
|
-
Note:
|
|
172
|
-
- The api will be deprecated, please use the api :func:`mindspore.runtime.reset_peak_memory_stats` instead.
|
|
173
|
-
|
|
174
|
-
Note:
|
|
175
|
-
If `device_target` is not specified, get the device capability of the current backend set by context.
|
|
172
|
+
Reset the "peak" stats tracked by memory manager, this api will be deprecated and removed in future versions.
|
|
173
|
+
Please use the api :func:`mindspore.runtime.reset_peak_memory_stats` instead.
|
|
176
174
|
|
|
177
175
|
Args:
|
|
178
|
-
device_target (str, optional): The device
|
|
179
|
-
Default
|
|
176
|
+
device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
|
|
177
|
+
``"Ascend"`` . Default ``None`` , represents the current device set by context.
|
|
180
178
|
|
|
181
179
|
Examples:
|
|
182
|
-
>>> import mindspore
|
|
183
|
-
>>>
|
|
184
|
-
>>>
|
|
185
|
-
>>>
|
|
186
|
-
>>>
|
|
187
|
-
>>> c = ops.add(a, b).asnumpy()
|
|
188
|
-
>>> print(ms.hal.max_memory_reserved())
|
|
180
|
+
>>> import mindspore
|
|
181
|
+
>>> a = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
182
|
+
>>> b = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
183
|
+
>>> c = mindspore.ops.add(a, b).asnumpy()
|
|
184
|
+
>>> print(mindspore.hal.max_memory_reserved())
|
|
189
185
|
1073741824
|
|
190
|
-
>>> print(
|
|
186
|
+
>>> print(mindspore.hal.max_memory_allocated())
|
|
191
187
|
1536
|
|
192
|
-
>>>
|
|
193
|
-
>>> print(
|
|
188
|
+
>>> mindspore.hal.reset_peak_memory_stats()
|
|
189
|
+
>>> print(mindspore.hal.max_memory_reserved())
|
|
194
190
|
0
|
|
195
|
-
>>> print(
|
|
191
|
+
>>> print(mindspore.hal.max_memory_allocated())
|
|
196
192
|
0
|
|
197
193
|
"""
|
|
198
194
|
if not function_memory_status['reset_peak_memory_stats']:
|
|
@@ -208,17 +204,12 @@ def reset_peak_memory_stats(device_target=None):
|
|
|
208
204
|
@_check_inputs_validation
|
|
209
205
|
def memory_summary(device_target=None):
|
|
210
206
|
"""
|
|
211
|
-
Returns readable memory pool status information.
|
|
212
|
-
|
|
213
|
-
Note:
|
|
214
|
-
- The api will be deprecated, please use the api :func:`mindspore.runtime.memory_summary` instead.
|
|
215
|
-
|
|
216
|
-
Note:
|
|
217
|
-
If `device_target` is not specified, get the device capability of the current backend set by context.
|
|
207
|
+
Returns readable memory pool status information, this api will be deprecated and removed in future versions.
|
|
208
|
+
Please use the api :func:`mindspore.runtime.memory_summary` instead.
|
|
218
209
|
|
|
219
210
|
Args:
|
|
220
|
-
device_target (str, optional): The device
|
|
221
|
-
Default
|
|
211
|
+
device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
|
|
212
|
+
``"Ascend"`` . Default ``None`` , represents the current device set by context.
|
|
222
213
|
|
|
223
214
|
Returns:
|
|
224
215
|
str, readable memory pool status information in tabular form.
|
|
@@ -270,30 +261,25 @@ def memory_summary(device_target=None):
|
|
|
270
261
|
@_check_inputs_validation
|
|
271
262
|
def memory_allocated(device_target=None):
|
|
272
263
|
"""
|
|
273
|
-
Returns the actual memory size currently occupied by Tensor
|
|
274
|
-
|
|
275
|
-
Note:
|
|
276
|
-
- The api will be deprecated, please use the api :func:`mindspore.runtime.memory_allocated` instead.
|
|
264
|
+
Returns the actual memory size currently occupied by Tensor, this api will be deprecated and removed in future
|
|
265
|
+
versions, please use the api :func:`mindspore.runtime.memory_allocated` instead.
|
|
277
266
|
|
|
278
267
|
Note:
|
|
279
|
-
-
|
|
280
|
-
- For the `CPU` backend, 0 is always returned.
|
|
268
|
+
- For the `CPU` device, 0 is always returned.
|
|
281
269
|
|
|
282
270
|
Args:
|
|
283
|
-
device_target (str, optional): The device
|
|
284
|
-
Default
|
|
271
|
+
device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
|
|
272
|
+
``"Ascend"`` . Default ``None`` , represents the current device set by context.
|
|
285
273
|
|
|
286
274
|
Returns:
|
|
287
275
|
int, in Byte.
|
|
288
276
|
|
|
289
277
|
Examples:
|
|
290
|
-
>>> import mindspore
|
|
291
|
-
>>>
|
|
292
|
-
>>>
|
|
293
|
-
>>>
|
|
294
|
-
>>>
|
|
295
|
-
>>> c = ops.add(a, b).asnumpy()
|
|
296
|
-
>>> print(ms.hal.memory_allocated())
|
|
278
|
+
>>> import mindspore
|
|
279
|
+
>>> a = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
280
|
+
>>> b = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
281
|
+
>>> c = mindspore.ops.add(a, b).asnumpy()
|
|
282
|
+
>>> print(mindspore.hal.memory_allocated())
|
|
297
283
|
1024
|
|
298
284
|
"""
|
|
299
285
|
if not function_memory_status['memory_allocated']:
|
|
@@ -308,30 +294,26 @@ def memory_allocated(device_target=None):
|
|
|
308
294
|
@_check_inputs_validation
|
|
309
295
|
def max_memory_allocated(device_target=None):
|
|
310
296
|
"""
|
|
311
|
-
|
|
312
|
-
|
|
313
|
-
|
|
314
|
-
- The api will be deprecated, please use the api :func:`mindspore.runtime.max_memory_allocated` instead.
|
|
297
|
+
Return the peak memory size of the memory pool actually occupied by Tensor since the process was started.
|
|
298
|
+
This api will be deprecated and removed in future versions, please use
|
|
299
|
+
the api :func:`mindspore.runtime.max_memory_allocated` instead.
|
|
315
300
|
|
|
316
301
|
Note:
|
|
317
|
-
-
|
|
318
|
-
- For the `CPU` backend, 0 is always returned.
|
|
302
|
+
- For the `CPU` device, 0 is always returned.
|
|
319
303
|
|
|
320
304
|
Args:
|
|
321
|
-
device_target (str, optional): The device
|
|
322
|
-
Default
|
|
305
|
+
device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
|
|
306
|
+
``"Ascend"`` . Default ``None`` , represents the current device set by context.
|
|
323
307
|
|
|
324
308
|
Returns:
|
|
325
309
|
int, in Byte.
|
|
326
310
|
|
|
327
311
|
Examples:
|
|
328
|
-
>>> import mindspore
|
|
329
|
-
>>>
|
|
330
|
-
>>>
|
|
331
|
-
>>>
|
|
332
|
-
>>>
|
|
333
|
-
>>> c = ops.add(a, b).asnumpy()
|
|
334
|
-
>>> print(ms.hal.max_memory_allocated())
|
|
312
|
+
>>> import mindspore
|
|
313
|
+
>>> a = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
314
|
+
>>> b = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
315
|
+
>>> c = mindspore.ops.add(a, b).asnumpy()
|
|
316
|
+
>>> print(mindspore.hal.max_memory_allocated())
|
|
335
317
|
1536
|
|
336
318
|
"""
|
|
337
319
|
if not function_memory_status['max_memory_allocated']:
|
|
@@ -346,29 +328,22 @@ def max_memory_allocated(device_target=None):
|
|
|
346
328
|
@_check_inputs_validation
|
|
347
329
|
def reset_max_memory_reserved(device_target=None):
|
|
348
330
|
"""
|
|
349
|
-
Reset the peak memory size managed by the memory pool.
|
|
350
|
-
|
|
351
|
-
Note:
|
|
352
|
-
- The api will be deprecated, please use the api :func:`mindspore.runtime.reset_max_memory_reserved` instead.
|
|
353
|
-
|
|
354
|
-
Note:
|
|
355
|
-
If `device_target` is not specified, get the device capability of the current backend set by context.
|
|
331
|
+
Reset the peak memory size managed by the memory pool, this api will be deprecated and removed in future versions.
|
|
332
|
+
Please use the api :func:`mindspore.runtime.reset_max_memory_reserved` instead.
|
|
356
333
|
|
|
357
334
|
Args:
|
|
358
|
-
device_target (str, optional): The device
|
|
359
|
-
Default
|
|
335
|
+
device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
|
|
336
|
+
``"Ascend"`` . Default ``None`` , represents the current device set by context.
|
|
360
337
|
|
|
361
338
|
Examples:
|
|
362
|
-
>>> import mindspore
|
|
363
|
-
>>>
|
|
364
|
-
>>>
|
|
365
|
-
>>>
|
|
366
|
-
>>>
|
|
367
|
-
>>> c = ops.add(a, b).asnumpy()
|
|
368
|
-
>>> print(ms.hal.max_memory_reserved())
|
|
339
|
+
>>> import mindspore
|
|
340
|
+
>>> a = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
341
|
+
>>> b = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
342
|
+
>>> c = mindspore.ops.add(a, b).asnumpy()
|
|
343
|
+
>>> print(mindspore.hal.max_memory_reserved())
|
|
369
344
|
1073741824
|
|
370
|
-
>>>
|
|
371
|
-
>>> print(
|
|
345
|
+
>>> mindspore.hal.reset_max_memory_reserved()
|
|
346
|
+
>>> print(mindspore.hal.max_memory_reserved())
|
|
372
347
|
0
|
|
373
348
|
"""
|
|
374
349
|
if not function_memory_status['reset_max_memory_reserved']:
|
|
@@ -383,29 +358,22 @@ def reset_max_memory_reserved(device_target=None):
|
|
|
383
358
|
@_check_inputs_validation
|
|
384
359
|
def reset_max_memory_allocated(device_target=None):
|
|
385
360
|
"""
|
|
386
|
-
Reset the peak memory size of the memory pool actually occupied by Tensor
|
|
387
|
-
|
|
388
|
-
Note:
|
|
389
|
-
- The api will be deprecated, please use the api :func:`mindspore.runtime.reset_max_memory_allocated` instead.
|
|
390
|
-
|
|
391
|
-
Note:
|
|
392
|
-
If `device_target` is not specified, get the device capability of the current backend set by context.
|
|
361
|
+
Reset the peak memory size of the memory pool actually occupied by Tensor, this api will be deprecated and removed
|
|
362
|
+
in future versions, please use the api :func:`mindspore.runtime.reset_max_memory_allocated` instead.
|
|
393
363
|
|
|
394
364
|
Args:
|
|
395
|
-
device_target (str, optional): The device
|
|
396
|
-
Default
|
|
365
|
+
device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
|
|
366
|
+
``"Ascend"`` . Default ``None`` , represents the current device set by context.
|
|
397
367
|
|
|
398
368
|
Examples:
|
|
399
|
-
>>> import mindspore
|
|
400
|
-
>>>
|
|
401
|
-
>>>
|
|
402
|
-
>>>
|
|
403
|
-
>>>
|
|
404
|
-
>>> c = ops.add(a, b).asnumpy()
|
|
405
|
-
>>> print(ms.hal.max_memory_allocated())
|
|
369
|
+
>>> import mindspore
|
|
370
|
+
>>> a = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
371
|
+
>>> b = mindspore.tensor(mindspore.ops.ones([1, 2]), mindspore.float32)
|
|
372
|
+
>>> c = mindspore.ops.add(a, b).asnumpy()
|
|
373
|
+
>>> print(mindspore.hal.max_memory_allocated())
|
|
406
374
|
1536
|
|
407
|
-
>>>
|
|
408
|
-
>>> print(
|
|
375
|
+
>>> mindspore.hal.reset_max_memory_allocated()
|
|
376
|
+
>>> print(mindspore.hal.max_memory_allocated())
|
|
409
377
|
0
|
|
410
378
|
"""
|
|
411
379
|
if not function_memory_status['reset_max_memory_allocated']:
|