mindspore 2.5.0__cp311-cp311-win_amd64.whl → 2.6.0rc1__cp311-cp311-win_amd64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of mindspore might be problematic. Click here for more details.
- mindspore/.commit_id +1 -1
- mindspore/Microsoft.VisualStudio.Telemetry.dll +0 -0
- mindspore/Newtonsoft.Json.dll +0 -0
- mindspore/__init__.py +6 -4
- mindspore/_c_dataengine.cp311-win_amd64.pyd +0 -0
- mindspore/_c_expression.cp311-win_amd64.pyd +0 -0
- mindspore/_c_mindrecord.cp311-win_amd64.pyd +0 -0
- mindspore/_check_jit_forbidden_api.py +3 -0
- mindspore/_checkparam.py +3 -33
- mindspore/_deprecated/__init__.py +17 -0
- mindspore/_deprecated/jit.py +198 -0
- mindspore/_extends/builtin_operations.py +1 -1
- mindspore/_extends/parse/__init__.py +6 -7
- mindspore/_extends/parse/compile_config.py +19 -0
- mindspore/_extends/parse/deprecated/deprecated_tensor_method.py +22 -3
- mindspore/_extends/parse/jit_fallback_modules/__init__.py +0 -0
- mindspore/_extends/parse/jit_fallback_modules/check_utils.py +123 -0
- mindspore/_extends/parse/jit_fallback_modules/third_party_modules.py +50 -0
- mindspore/_extends/parse/parser.py +24 -193
- mindspore/_extends/parse/resources.py +1 -5
- mindspore/_extends/parse/standard_method.py +97 -74
- mindspore/_extends/pijit/__init__.py +2 -2
- mindspore/_extends/pijit/pijit_func_white_list.py +16 -11
- mindspore/_extends/pijit/tensor_func_list.py +27 -0
- mindspore/_extends/utils.py +1 -1
- mindspore/amp.py +4 -4
- mindspore/atlprov.dll +0 -0
- mindspore/avcodec-59.dll +0 -0
- mindspore/avdevice-59.dll +0 -0
- mindspore/avfilter-8.dll +0 -0
- mindspore/avformat-59.dll +0 -0
- mindspore/avutil-57.dll +0 -0
- mindspore/boost/__init__.py +2 -2
- mindspore/boost/base.py +3 -7
- mindspore/boost/boost_cell_wrapper.py +2 -2
- mindspore/c1.dll +0 -0
- mindspore/c1xx.dll +0 -0
- mindspore/c2.dll +0 -0
- mindspore/common/__init__.py +4 -3
- mindspore/common/_grad_function.py +56 -0
- mindspore/common/_pijit_context.py +14 -5
- mindspore/common/_register_for_tensor.py +1 -1
- mindspore/common/_stub_tensor.py +5 -10
- mindspore/common/_tensor_cpp_method.py +1 -1
- mindspore/common/_tensor_docs.py +1915 -3287
- mindspore/common/api.py +341 -354
- mindspore/common/auto_dynamic_shape.py +41 -44
- mindspore/common/dtype.py +5 -2
- mindspore/common/dump.py +7 -5
- mindspore/common/file_system.py +3 -0
- mindspore/common/hook_handle.py +5 -3
- mindspore/common/initializer.py +10 -6
- mindspore/common/jit_begin_end.py +94 -0
- mindspore/common/jit_config.py +6 -1
- mindspore/common/jit_context.py +76 -0
- mindspore/common/jit_trace.py +378 -0
- mindspore/common/lazy_inline.py +2 -2
- mindspore/common/mutable.py +5 -4
- mindspore/common/parameter.py +106 -39
- mindspore/common/seed.py +2 -2
- mindspore/common/sparse_tensor.py +23 -17
- mindspore/common/tensor.py +297 -714
- mindspore/communication/__init__.py +7 -5
- mindspore/communication/_comm_helper.py +47 -2
- mindspore/communication/comm_func.py +70 -53
- mindspore/communication/management.py +83 -17
- mindspore/context.py +214 -560
- mindspore/dataset/__init__.py +44 -20
- mindspore/dataset/audio/__init__.py +2 -8
- mindspore/dataset/audio/transforms.py +3 -17
- mindspore/dataset/core/config.py +3 -3
- mindspore/dataset/engine/cache_client.py +1 -1
- mindspore/dataset/engine/datasets.py +102 -120
- mindspore/dataset/engine/datasets_audio.py +22 -22
- mindspore/dataset/engine/datasets_standard_format.py +43 -24
- mindspore/dataset/engine/datasets_text.py +78 -85
- mindspore/dataset/engine/datasets_user_defined.py +108 -76
- mindspore/dataset/engine/datasets_vision.py +111 -108
- mindspore/dataset/engine/iterators.py +5 -3
- mindspore/dataset/engine/obs/obs_mindrecord_dataset.py +1 -1
- mindspore/dataset/engine/samplers.py +279 -57
- mindspore/dataset/engine/serializer_deserializer.py +2 -1
- mindspore/dataset/engine/validators.py +10 -0
- mindspore/dataset/text/__init__.py +7 -6
- mindspore/dataset/text/transforms.py +6 -5
- mindspore/dataset/text/utils.py +3 -3
- mindspore/dataset/transforms/__init__.py +0 -9
- mindspore/dataset/transforms/transforms.py +3 -3
- mindspore/dataset/utils/browse_dataset.py +1 -1
- mindspore/dataset/vision/__init__.py +2 -9
- mindspore/dataset/vision/transforms.py +202 -158
- mindspore/dataset/vision/utils.py +7 -5
- mindspore/device_context/ascend/op_debug.py +60 -1
- mindspore/device_context/ascend/op_tuning.py +0 -4
- mindspore/device_manager.py +39 -3
- mindspore/dnnl.dll +0 -0
- mindspore/dpcmi.dll +0 -0
- mindspore/experimental/es/embedding_service.py +35 -27
- mindspore/experimental/map_parameter.py +4 -4
- mindspore/experimental/optim/adadelta.py +22 -26
- mindspore/experimental/optim/adagrad.py +4 -4
- mindspore/experimental/optim/adam.py +4 -0
- mindspore/experimental/optim/adamax.py +4 -4
- mindspore/experimental/optim/adamw.py +4 -0
- mindspore/experimental/optim/asgd.py +1 -1
- mindspore/experimental/optim/lr_scheduler.py +40 -22
- mindspore/experimental/optim/radam.py +5 -5
- mindspore/experimental/optim/rprop.py +1 -1
- mindspore/experimental/optim/sgd.py +1 -1
- mindspore/hal/contiguous_tensors_handle.py +6 -10
- mindspore/hal/device.py +55 -81
- mindspore/hal/event.py +38 -55
- mindspore/hal/memory.py +93 -144
- mindspore/hal/stream.py +81 -125
- mindspore/include/dataset/constants.h +7 -4
- mindspore/include/dataset/execute.h +2 -2
- mindspore/jpeg62.dll +0 -0
- mindspore/log.py +40 -2
- mindspore/mindrecord/__init__.py +20 -7
- mindspore/mindspore_backend_common.dll +0 -0
- mindspore/mindspore_backend_manager.dll +0 -0
- mindspore/mindspore_common.dll +0 -0
- mindspore/mindspore_core.dll +0 -0
- mindspore/mindspore_dump.dll +0 -0
- mindspore/mindspore_frontend.dll +0 -0
- mindspore/mindspore_glog.dll +0 -0
- mindspore/mindspore_memory_pool.dll +0 -0
- mindspore/mindspore_ms_backend.dll +0 -0
- mindspore/mindspore_ops.dll +0 -0
- mindspore/{mindspore_backend.dll → mindspore_ops_host.dll} +0 -0
- mindspore/mindspore_ops_kernel_common.dll +0 -0
- mindspore/mindspore_profiler.dll +0 -0
- mindspore/mindspore_pyboost.dll +0 -0
- mindspore/mindspore_pynative.dll +0 -0
- mindspore/mindspore_res_manager.dll +0 -0
- mindspore/mindspore_runtime_pipeline.dll +0 -0
- mindspore/mint/__init__.py +131 -700
- mindspore/mint/distributed/__init__.py +5 -1
- mindspore/mint/distributed/distributed.py +194 -109
- mindspore/mint/linalg/__init__.py +2 -0
- mindspore/mint/nn/__init__.py +280 -18
- mindspore/mint/nn/functional.py +282 -64
- mindspore/mint/nn/layer/__init__.py +4 -0
- mindspore/mint/nn/layer/_functions.py +7 -3
- mindspore/mint/nn/layer/activation.py +120 -13
- mindspore/mint/nn/layer/conv.py +218 -24
- mindspore/mint/nn/layer/normalization.py +15 -16
- mindspore/mint/nn/layer/padding.py +1 -1
- mindspore/mint/nn/layer/pooling.py +66 -1
- mindspore/mint/optim/__init__.py +2 -1
- mindspore/mint/optim/sgd.py +171 -0
- mindspore/msobj140.dll +0 -0
- mindspore/mspdb140.dll +0 -0
- mindspore/mspdbcore.dll +0 -0
- mindspore/mspdbst.dll +0 -0
- mindspore/mspft140.dll +0 -0
- mindspore/msvcdis140.dll +0 -0
- mindspore/msvcp140_1.dll +0 -0
- mindspore/msvcp140_2.dll +0 -0
- mindspore/msvcp140_atomic_wait.dll +0 -0
- mindspore/msvcp140_codecvt_ids.dll +0 -0
- mindspore/nn/__init__.py +4 -1
- mindspore/nn/cell.py +1250 -176
- mindspore/nn/layer/activation.py +23 -21
- mindspore/nn/layer/basic.py +22 -16
- mindspore/nn/layer/container.py +1 -1
- mindspore/nn/layer/conv.py +22 -17
- mindspore/nn/layer/embedding.py +9 -8
- mindspore/nn/layer/normalization.py +48 -42
- mindspore/nn/layer/pooling.py +75 -31
- mindspore/nn/layer/transformer.py +11 -10
- mindspore/nn/learning_rate_schedule.py +4 -2
- mindspore/nn/loss/loss.py +27 -19
- mindspore/nn/optim/ada_grad.py +6 -5
- mindspore/nn/optim/adadelta.py +9 -7
- mindspore/nn/optim/adafactor.py +1 -1
- mindspore/nn/optim/adam.py +16 -12
- mindspore/nn/optim/adamax.py +8 -7
- mindspore/nn/optim/adasum.py +5 -5
- mindspore/nn/optim/asgd.py +1 -1
- mindspore/nn/optim/ftrl.py +11 -9
- mindspore/nn/optim/lamb.py +1 -1
- mindspore/nn/optim/lazyadam.py +12 -10
- mindspore/nn/optim/momentum.py +7 -6
- mindspore/nn/optim/optimizer.py +2 -2
- mindspore/nn/optim/proximal_ada_grad.py +12 -10
- mindspore/nn/optim/rmsprop.py +13 -12
- mindspore/nn/optim/rprop.py +9 -7
- mindspore/nn/optim/sgd.py +9 -6
- mindspore/nn/optim/tft_wrapper.py +5 -2
- mindspore/nn/probability/bijector/bijector.py +17 -11
- mindspore/nn/probability/bijector/gumbel_cdf.py +5 -5
- mindspore/nn/probability/bijector/invert.py +2 -2
- mindspore/nn/probability/bijector/scalar_affine.py +3 -3
- mindspore/nn/probability/bijector/softplus.py +3 -2
- mindspore/nn/probability/distribution/beta.py +3 -3
- mindspore/nn/probability/distribution/categorical.py +1 -1
- mindspore/nn/probability/distribution/cauchy.py +4 -2
- mindspore/nn/probability/distribution/exponential.py +6 -7
- mindspore/nn/probability/distribution/gamma.py +2 -2
- mindspore/nn/probability/distribution/gumbel.py +2 -2
- mindspore/nn/probability/distribution/half_normal.py +5 -3
- mindspore/nn/probability/distribution/logistic.py +5 -3
- mindspore/nn/probability/distribution/poisson.py +1 -1
- mindspore/nn/probability/distribution/uniform.py +5 -3
- mindspore/nn/reinforcement/_tensors_queue.py +1 -1
- mindspore/nn/reinforcement/tensor_array.py +1 -1
- mindspore/nn/wrap/__init__.py +6 -6
- mindspore/nn/wrap/cell_wrapper.py +178 -117
- mindspore/nn/wrap/grad_reducer.py +45 -36
- mindspore/nn/wrap/loss_scale.py +3 -3
- mindspore/numpy/array_creations.py +3 -3
- mindspore/numpy/array_ops.py +1 -1
- mindspore/numpy/math_ops.py +4 -4
- mindspore/numpy/utils.py +1 -2
- mindspore/numpy/utils_const.py +1 -2
- mindspore/opencv_core452.dll +0 -0
- mindspore/opencv_imgcodecs452.dll +0 -0
- mindspore/opencv_imgproc452.dll +0 -0
- mindspore/ops/__init__.py +3 -2
- mindspore/ops/_grad_experimental/grad_comm_ops.py +18 -3
- mindspore/ops/_grad_experimental/grad_debug_ops.py +8 -1
- mindspore/ops/_grad_experimental/taylor_rule.py +29 -0
- mindspore/ops/_register_for_op.py +0 -11
- mindspore/{ops_generate → ops/_utils}/arg_dtype_cast.py +123 -4
- mindspore/{ops_generate → ops/_utils}/arg_handler.py +3 -4
- mindspore/ops/_vmap/vmap_array_ops.py +7 -6
- mindspore/ops/_vmap/vmap_grad_nn_ops.py +2 -1
- mindspore/ops/_vmap/vmap_math_ops.py +4 -7
- mindspore/ops/_vmap/vmap_nn_ops.py +9 -8
- mindspore/ops/auto_generate/__init__.py +4 -3
- mindspore/ops/auto_generate/cpp_create_prim_instance_helper.py +102 -49
- mindspore/ops/auto_generate/gen_extend_func.py +281 -135
- mindspore/ops/auto_generate/gen_ops_def.py +2574 -2326
- mindspore/ops/auto_generate/gen_ops_prim.py +8566 -2755
- mindspore/ops/auto_generate/pyboost_inner_prim.py +106 -76
- mindspore/ops/composite/__init__.py +2 -1
- mindspore/ops/composite/base.py +19 -24
- mindspore/ops/composite/math_ops.py +6 -16
- mindspore/ops/composite/multitype_ops/__init__.py +5 -2
- mindspore/ops/composite/multitype_ops/_compile_utils.py +2 -3
- mindspore/ops/composite/multitype_ops/_constexpr_utils.py +1 -2
- mindspore/ops/composite/multitype_ops/add_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/bitwise_and_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/bitwise_or_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/bitwise_xor_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/div_impl.py +6 -4
- mindspore/ops/composite/multitype_ops/equal_impl.py +4 -3
- mindspore/ops/composite/multitype_ops/floordiv_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/getitem_impl.py +3 -2
- mindspore/ops/composite/multitype_ops/greater_equal_impl.py +4 -3
- mindspore/ops/composite/multitype_ops/greater_impl.py +4 -3
- mindspore/ops/composite/multitype_ops/in_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/invert_impl.py +50 -0
- mindspore/ops/composite/multitype_ops/left_shift_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/less_equal_impl.py +4 -3
- mindspore/ops/composite/multitype_ops/less_impl.py +4 -3
- mindspore/ops/composite/multitype_ops/logic_not_impl.py +3 -2
- mindspore/ops/composite/multitype_ops/logical_and_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/logical_or_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/mod_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/mul_impl.py +3 -2
- mindspore/ops/composite/multitype_ops/negative_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/not_equal_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/not_in_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/ones_like_impl.py +18 -0
- mindspore/ops/composite/multitype_ops/pow_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/right_shift_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/setitem_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/sub_impl.py +2 -1
- mindspore/ops/function/__init__.py +28 -2
- mindspore/ops/function/_add_attr_func.py +58 -0
- mindspore/ops/function/array_func.py +1629 -2345
- mindspore/ops/function/clip_func.py +38 -45
- mindspore/ops/function/debug_func.py +36 -44
- mindspore/ops/function/grad/__init__.py +1 -0
- mindspore/ops/function/grad/grad_func.py +104 -71
- mindspore/ops/function/image_func.py +1 -1
- mindspore/ops/function/linalg_func.py +46 -78
- mindspore/ops/function/math_func.py +3035 -3705
- mindspore/ops/function/nn_func.py +676 -241
- mindspore/ops/function/other_func.py +159 -1
- mindspore/ops/function/parameter_func.py +17 -30
- mindspore/ops/function/random_func.py +204 -361
- mindspore/ops/function/reshard_func.py +4 -70
- mindspore/ops/function/sparse_func.py +3 -3
- mindspore/ops/function/sparse_unary_func.py +5 -5
- mindspore/ops/function/spectral_func.py +25 -58
- mindspore/ops/function/vmap_func.py +24 -17
- mindspore/ops/functional.py +6 -4
- mindspore/ops/functional_overload.py +547 -4
- mindspore/ops/op_info_register.py +32 -244
- mindspore/ops/operations/__init__.py +10 -5
- mindspore/ops/operations/_custom_ops_utils.py +247 -0
- mindspore/ops/operations/_grad_ops.py +1 -10
- mindspore/ops/operations/_inner_ops.py +5 -76
- mindspore/ops/operations/_ms_kernel.py +4 -10
- mindspore/ops/operations/_rl_inner_ops.py +1 -1
- mindspore/ops/operations/_scalar_ops.py +3 -2
- mindspore/ops/operations/_sequence_ops.py +1 -1
- mindspore/ops/operations/_tensor_array.py +1 -1
- mindspore/ops/operations/array_ops.py +37 -22
- mindspore/ops/operations/comm_ops.py +150 -107
- mindspore/ops/operations/custom_ops.py +221 -23
- mindspore/ops/operations/debug_ops.py +115 -16
- mindspore/ops/operations/inner_ops.py +1 -1
- mindspore/ops/operations/linalg_ops.py +1 -58
- mindspore/ops/operations/manually_defined/_inner.py +1 -1
- mindspore/ops/operations/manually_defined/ops_def.py +746 -79
- mindspore/ops/operations/math_ops.py +21 -18
- mindspore/ops/operations/nn_ops.py +65 -191
- mindspore/ops/operations/other_ops.py +62 -9
- mindspore/ops/operations/random_ops.py +13 -7
- mindspore/ops/operations/reshard_ops.py +1 -1
- mindspore/ops/operations/sparse_ops.py +2 -2
- mindspore/ops/primitive.py +43 -32
- mindspore/ops/tensor_method.py +232 -13
- mindspore/ops_generate/__init__.py +0 -5
- mindspore/ops_generate/aclnn/__init__.py +0 -0
- mindspore/ops_generate/{aclnn_kernel_register_auto_cc_generator.py → aclnn/aclnn_kernel_register_auto_cc_generator.py} +43 -18
- mindspore/ops_generate/{gen_aclnn_implement.py → aclnn/gen_aclnn_implement.py} +49 -51
- mindspore/ops_generate/api/__init__.py +0 -0
- mindspore/ops_generate/{add_tensor_docs_generator.py → api/add_tensor_docs_generator.py} +9 -7
- mindspore/ops_generate/{cpp_create_prim_instance_helper_generator.py → api/cpp_create_prim_instance_helper_generator.py} +6 -9
- mindspore/ops_generate/{functional_map_cpp_generator.py → api/functional_map_cpp_generator.py} +25 -12
- mindspore/ops_generate/{functional_overload_py_generator.py → api/functional_overload_py_generator.py} +8 -6
- mindspore/ops_generate/{functions_cc_generator.py → api/functions_cc_generator.py} +14 -10
- mindspore/ops_generate/api/gen_api.py +103 -0
- mindspore/ops_generate/{op_api_proto.py → api/op_api_proto.py} +98 -69
- mindspore/ops_generate/{tensor_func_reg_cpp_generator.py → api/tensor_func_reg_cpp_generator.py} +82 -43
- mindspore/ops_generate/common/__init__.py +0 -0
- mindspore/ops_generate/common/gen_constants.py +91 -0
- mindspore/ops_generate/{gen_utils.py → common/gen_utils.py} +72 -19
- mindspore/ops_generate/{op_proto.py → common/op_proto.py} +64 -1
- mindspore/ops_generate/{template.py → common/template.py} +96 -84
- mindspore/ops_generate/gen_ops.py +23 -325
- mindspore/ops_generate/op_def/__init__.py +0 -0
- mindspore/ops_generate/op_def/gen_op_def.py +90 -0
- mindspore/ops_generate/{lite_ops_cpp_generator.py → op_def/lite_ops_cpp_generator.py} +47 -11
- mindspore/ops_generate/{ops_def_cc_generator.py → op_def/ops_def_cc_generator.py} +18 -7
- mindspore/ops_generate/{ops_def_h_generator.py → op_def/ops_def_h_generator.py} +5 -5
- mindspore/ops_generate/{ops_name_h_generator.py → op_def/ops_name_h_generator.py} +30 -15
- mindspore/ops_generate/op_def/ops_primitive_h_generator.py +125 -0
- mindspore/ops_generate/op_def_py/__init__.py +0 -0
- mindspore/ops_generate/op_def_py/gen_op_def_py.py +47 -0
- mindspore/ops_generate/{op_def_py_generator.py → op_def_py/op_def_py_generator.py} +6 -5
- mindspore/ops_generate/{op_prim_py_generator.py → op_def_py/op_prim_py_generator.py} +24 -15
- mindspore/ops_generate/pyboost/__init__.py +0 -0
- mindspore/ops_generate/{auto_grad_impl_cc_generator.py → pyboost/auto_grad_impl_cc_generator.py} +11 -7
- mindspore/ops_generate/{auto_grad_reg_cc_generator.py → pyboost/auto_grad_reg_cc_generator.py} +7 -7
- mindspore/ops_generate/{gen_pyboost_func.py → pyboost/gen_pyboost_func.py} +40 -16
- mindspore/ops_generate/{op_template_parser.py → pyboost/op_template_parser.py} +105 -24
- mindspore/ops_generate/{pyboost_functions_cpp_generator.py → pyboost/pyboost_functions_cpp_generator.py} +55 -18
- mindspore/ops_generate/{pyboost_functions_h_generator.py → pyboost/pyboost_functions_h_generator.py} +42 -10
- mindspore/ops_generate/{pyboost_functions_py_generator.py → pyboost/pyboost_functions_py_generator.py} +6 -6
- mindspore/ops_generate/{pyboost_grad_function_cpp_generator.py → pyboost/pyboost_grad_function_cpp_generator.py} +11 -10
- mindspore/ops_generate/{pyboost_inner_prim_generator.py → pyboost/pyboost_inner_prim_generator.py} +8 -7
- mindspore/ops_generate/{pyboost_native_grad_functions_generator.py → pyboost/pyboost_native_grad_functions_generator.py} +14 -10
- mindspore/ops_generate/{pyboost_op_cpp_code_generator.py → pyboost/pyboost_op_cpp_code_generator.py} +140 -53
- mindspore/ops_generate/{pyboost_overload_functions_cpp_generator.py → pyboost/pyboost_overload_functions_cpp_generator.py} +28 -15
- mindspore/ops_generate/{pyboost_utils.py → pyboost/pyboost_utils.py} +88 -4
- mindspore/ops_generate/resources/__init__.py +0 -0
- mindspore/ops_generate/resources/resource_list.py +30 -0
- mindspore/ops_generate/resources/resource_loader.py +36 -0
- mindspore/ops_generate/resources/resource_manager.py +64 -0
- mindspore/ops_generate/resources/yaml_loader.py +88 -0
- mindspore/ops_generate/tensor_py_cc_generator.py +122 -0
- mindspore/parallel/__init__.py +6 -2
- mindspore/parallel/_auto_parallel_context.py +133 -6
- mindspore/parallel/_cell_wrapper.py +130 -15
- mindspore/parallel/_parallel_serialization.py +95 -4
- mindspore/parallel/_ps_context.py +1 -1
- mindspore/parallel/_recovery_context.py +7 -2
- mindspore/parallel/_tensor.py +142 -18
- mindspore/parallel/_utils.py +198 -25
- mindspore/parallel/algo_parameter_config.py +3 -3
- mindspore/parallel/auto_parallel.py +732 -0
- mindspore/parallel/checkpoint_convert.py +159 -0
- mindspore/parallel/checkpoint_transform.py +656 -37
- mindspore/parallel/cluster/process_entity/_api.py +151 -19
- mindspore/parallel/cluster/run.py +1 -1
- mindspore/parallel/function/__init__.py +24 -0
- mindspore/parallel/function/reshard_func.py +259 -0
- mindspore/parallel/nn/__init__.py +25 -0
- mindspore/parallel/nn/parallel_cell_wrapper.py +263 -0
- mindspore/parallel/nn/parallel_grad_reducer.py +169 -0
- mindspore/parallel/parameter_broadcast.py +24 -13
- mindspore/parallel/shard.py +137 -61
- mindspore/parallel/transform_safetensors.py +287 -95
- mindspore/pgodb140.dll +0 -0
- mindspore/pgort140.dll +0 -0
- mindspore/profiler/__init__.py +9 -5
- mindspore/profiler/analysis/parser/ascend_cann_parser.py +6 -2
- mindspore/profiler/analysis/parser/ms_framework_parser.py +4 -4
- mindspore/profiler/analysis/parser/timeline_assembly_factory/ascend_timeline_assembler.py +7 -4
- mindspore/profiler/analysis/parser/timeline_assembly_factory/trace_view_container.py +22 -0
- mindspore/profiler/analysis/parser/timeline_creator/fwk_timeline_creator.py +3 -3
- mindspore/profiler/analysis/parser/timeline_event/fwk_event.py +241 -86
- mindspore/profiler/analysis/viewer/ascend_communication_viewer.py +41 -2
- mindspore/profiler/analysis/viewer/ascend_kernel_details_viewer.py +33 -35
- mindspore/profiler/analysis/viewer/ascend_memory_viewer.py +7 -0
- mindspore/profiler/analysis/viewer/ascend_op_memory_viewer.py +8 -3
- mindspore/profiler/analysis/viewer/ascend_step_trace_time_viewer.py +141 -30
- mindspore/profiler/analysis/viewer/ms_dataset_viewer.py +5 -6
- mindspore/profiler/common/ascend_msprof_exporter.py +5 -4
- mindspore/profiler/common/constant.py +12 -0
- mindspore/profiler/common/msprof_cmd_tool.py +42 -23
- mindspore/profiler/common/path_manager.py +24 -0
- mindspore/profiler/common/profiler_context.py +26 -2
- mindspore/profiler/common/profiler_meta_data.py +74 -0
- mindspore/profiler/common/profiler_parameters.py +59 -18
- mindspore/profiler/common/profiler_path_manager.py +66 -7
- mindspore/profiler/dynamic_profiler.py +112 -79
- mindspore/profiler/envprofiler.py +26 -1
- mindspore/profiler/experimental_config.py +197 -0
- mindspore/profiler/mstx.py +57 -14
- mindspore/profiler/platform/npu_profiler.py +33 -7
- mindspore/profiler/profiler.py +541 -45
- mindspore/profiler/profiler_action_controller.py +1 -1
- mindspore/profiler/profiler_interface.py +4 -0
- mindspore/profiler/schedule.py +57 -22
- mindspore/rewrite/api/node.py +15 -13
- mindspore/rewrite/api/symbol_tree.py +1 -1
- mindspore/run_check/_check_version.py +25 -14
- mindspore/run_check/run_check.py +1 -1
- mindspore/runtime/__init__.py +2 -2
- mindspore/runtime/executor.py +40 -11
- mindspore/runtime/memory.py +25 -8
- mindspore/safeguard/rewrite_obfuscation.py +12 -9
- mindspore/swresample-4.dll +0 -0
- mindspore/swscale-6.dll +0 -0
- mindspore/tbbmalloc.dll +0 -0
- mindspore/tinyxml2.dll +0 -0
- mindspore/train/__init__.py +8 -8
- mindspore/train/_utils.py +35 -7
- mindspore/train/amp.py +1 -1
- mindspore/train/callback/__init__.py +2 -2
- mindspore/train/callback/_callback.py +2 -16
- mindspore/train/callback/_checkpoint.py +24 -40
- mindspore/train/callback/_cluster_monitor.py +14 -18
- mindspore/train/callback/_flops_collector.py +2 -3
- mindspore/train/callback/_history.py +7 -4
- mindspore/train/callback/_lambda_callback.py +2 -2
- mindspore/train/callback/_landscape.py +0 -3
- mindspore/train/callback/_loss_monitor.py +2 -1
- mindspore/train/callback/_on_request_exit.py +6 -5
- mindspore/train/callback/_reduce_lr_on_plateau.py +11 -6
- mindspore/train/callback/_summary_collector.py +8 -13
- mindspore/train/callback/_time_monitor.py +2 -1
- mindspore/train/callback/{_tft_register.py → _train_fault_tolerance.py} +179 -103
- mindspore/train/data_sink.py +25 -2
- mindspore/train/dataset_helper.py +4 -5
- mindspore/train/loss_scale_manager.py +8 -7
- mindspore/train/metrics/accuracy.py +3 -3
- mindspore/train/metrics/confusion_matrix.py +9 -9
- mindspore/train/metrics/error.py +3 -3
- mindspore/train/metrics/hausdorff_distance.py +4 -4
- mindspore/train/metrics/mean_surface_distance.py +3 -3
- mindspore/train/metrics/metric.py +0 -12
- mindspore/train/metrics/occlusion_sensitivity.py +4 -2
- mindspore/train/metrics/precision.py +8 -6
- mindspore/train/metrics/recall.py +9 -9
- mindspore/train/metrics/root_mean_square_surface_distance.py +2 -2
- mindspore/train/mind_ir_pb2.py +19 -12
- mindspore/train/model.py +176 -103
- mindspore/train/serialization.py +246 -988
- mindspore/train/summary/_summary_adapter.py +2 -2
- mindspore/train/summary/summary_record.py +1 -1
- mindspore/turbojpeg.dll +0 -0
- mindspore/utils/__init__.py +3 -2
- mindspore/utils/dryrun.py +4 -2
- mindspore/utils/hooks.py +81 -0
- mindspore/utils/utils.py +138 -4
- mindspore/vcmeta.dll +0 -0
- mindspore/vcruntime140.dll +0 -0
- mindspore/vcruntime140_1.dll +0 -0
- mindspore/version.py +1 -1
- {mindspore-2.5.0.dist-info → mindspore-2.6.0rc1.dist-info}/METADATA +2 -1
- {mindspore-2.5.0.dist-info → mindspore-2.6.0rc1.dist-info}/RECORD +483 -438
- mindspore/_install_custom.py +0 -43
- mindspore/common/_register_for_adapter.py +0 -74
- mindspore/ops/auto_generate/gen_arg_dtype_cast.py +0 -252
- mindspore/ops/auto_generate/gen_arg_handler.py +0 -136
- mindspore/ops/operations/_opaque_predicate_registry.py +0 -41
- mindspore/ops_generate/gen_constants.py +0 -190
- mindspore/ops_generate/gen_ops_inner_prim.py +0 -131
- mindspore/ops_generate/ops_primitive_h_generator.py +0 -81
- /mindspore/ops_generate/{base_generator.py → common/base_generator.py} +0 -0
- {mindspore-2.5.0.dist-info → mindspore-2.6.0rc1.dist-info}/WHEEL +0 -0
- {mindspore-2.5.0.dist-info → mindspore-2.6.0rc1.dist-info}/entry_points.txt +0 -0
- {mindspore-2.5.0.dist-info → mindspore-2.6.0rc1.dist-info}/top_level.txt +0 -0
|
@@ -41,10 +41,11 @@ class LRScheduler:
|
|
|
41
41
|
|
|
42
42
|
Args:
|
|
43
43
|
optimizer (:class:`mindspore.experimental.optim.Optimizer`): The optimizer instance.
|
|
44
|
-
last_epoch (int, optional): The
|
|
44
|
+
last_epoch (int, optional): The number of times the `step()` method of
|
|
45
|
+
the current learning rate adjustment strategy has been executed. Default: ``-1``.
|
|
45
46
|
|
|
46
47
|
Raises:
|
|
47
|
-
TypeError: If `optimizer`
|
|
48
|
+
TypeError: If `optimizer` does not satisfy the type requirement.
|
|
48
49
|
KeyError: If `last_epoch` != -1 and ``'initial_lr'`` not in param groups.
|
|
49
50
|
ValueError: if `last_epoch` is not int.
|
|
50
51
|
ValueError: If `last_epoch` is not greater than -1.
|
|
@@ -700,9 +701,8 @@ class ConstantLR(LRScheduler):
|
|
|
700
701
|
@jit_class
|
|
701
702
|
class SequentialLR:
|
|
702
703
|
r"""
|
|
703
|
-
|
|
704
|
-
|
|
705
|
-
which scheduler is supposed to be called at a given epoch.
|
|
704
|
+
Concatenate multiple learning rate adjustment strategies in `schedulers` in sequence,
|
|
705
|
+
switching to the next learning rate adjustment strategy at `milestone`.
|
|
706
706
|
|
|
707
707
|
.. warning::
|
|
708
708
|
This is an experimental lr scheduler module that is subject to change.
|
|
@@ -713,8 +713,10 @@ class SequentialLR:
|
|
|
713
713
|
optimizer (:class:`mindspore.experimental.optim.Optimizer`): Wrapped optimizer.
|
|
714
714
|
schedulers (list[:class:`mindspore.experimental.optim.lr_scheduler.LRScheduler`]):
|
|
715
715
|
List of learning rate schedulers.
|
|
716
|
-
milestones (list): List of integers
|
|
717
|
-
|
|
716
|
+
milestones (list): List of integers of milestone points,
|
|
717
|
+
sets which learning rate adjustment strategy is invoked for each epoch.
|
|
718
|
+
last_epoch (int, optional): The number of times the `step()` method
|
|
719
|
+
of the current learning rate adjustment strategy has been executed. Default: ``-1``.
|
|
718
720
|
|
|
719
721
|
Raises:
|
|
720
722
|
ValueError: The optimizer in `schedulers` is different from the `optimizer` passed in.
|
|
@@ -805,9 +807,8 @@ class ReduceLROnPlateau:
|
|
|
805
807
|
"""
|
|
806
808
|
Reduce learning rate when a metric has stopped improving.
|
|
807
809
|
Models often benefit from reducing the learning rate by a factor
|
|
808
|
-
of 2-10 once learning stagnates.
|
|
809
|
-
|
|
810
|
-
of epochs, the learning rate is reduced.
|
|
810
|
+
of 2-10 once learning stagnates. The scheduler reads the metrics `metrics` during execution
|
|
811
|
+
and adjusts the learning rate via the `step` method if the metrics do not improve within `patience` cycles.
|
|
811
812
|
|
|
812
813
|
.. warning::
|
|
813
814
|
This is an experimental lr scheduler module that is subject to change.
|
|
@@ -816,7 +817,8 @@ class ReduceLROnPlateau:
|
|
|
816
817
|
|
|
817
818
|
Args:
|
|
818
819
|
optimizer (:class:`mindspore.experimental.optim.Optimizer`): Wrapped optimizer.
|
|
819
|
-
mode (str, optional):
|
|
820
|
+
mode (str, optional): Trigger mode that triggers a reduction in learning rate
|
|
821
|
+
when the monitoring metrics are at their `min` / `max` point. In `min` mode, lr will
|
|
820
822
|
be reduced when the quantity monitored has stopped
|
|
821
823
|
decreasing; in `max` mode it will be reduced when the
|
|
822
824
|
quantity monitored has stopped increasing. Default: ``'min'``.
|
|
@@ -830,12 +832,25 @@ class ReduceLROnPlateau:
|
|
|
830
832
|
Default: ``10``.
|
|
831
833
|
threshold (float, optional): Threshold for measuring the new optimum,
|
|
832
834
|
to only focus on significant changes. Default: ``1e-4``.
|
|
833
|
-
threshold_mode (str, optional):
|
|
834
|
-
|
|
835
|
-
|
|
836
|
-
|
|
837
|
-
|
|
838
|
-
|
|
835
|
+
threshold_mode (str, optional): A mode for measuring indicators of change for the better.
|
|
836
|
+
One of `rel`, `abs`. Default: ``'rel'``.
|
|
837
|
+
|
|
838
|
+
Assume that `best` represents the best value of the current performance metric.
|
|
839
|
+
|
|
840
|
+
- In ``'rel'`` mode, the indicator is compared to a `threshold` in proportional form:
|
|
841
|
+
|
|
842
|
+
- When `mode` is ``'max'``, the indicator is considered better if it exceeds best * ( 1 + threshold ).
|
|
843
|
+
|
|
844
|
+
- When `mode` is ``'min'``, the indicator is considered better
|
|
845
|
+
if it is lower than best * ( 1 - threshold ).
|
|
846
|
+
|
|
847
|
+
- In ``'abs'`` mode, the indicator is compared to `threshold` in absolute value form:
|
|
848
|
+
|
|
849
|
+
- When `mode` is ``'max'``, the indicator is considered better if it exceeds best + threshold.
|
|
850
|
+
|
|
851
|
+
- When `mode` is ``'min'``, the indicator is considered better
|
|
852
|
+
if it is lower than best - threshold.
|
|
853
|
+
|
|
839
854
|
cooldown (int, optional): Number of epochs to wait before resuming
|
|
840
855
|
normal operation after lr has been reduced. Default: ``0``.
|
|
841
856
|
min_lr (Union(float, list), optional): A scalar or a list of scalars. A
|
|
@@ -1164,15 +1179,17 @@ class CyclicLR(LRScheduler):
|
|
|
1164
1179
|
class CosineAnnealingWarmRestarts(LRScheduler):
|
|
1165
1180
|
r"""
|
|
1166
1181
|
Set the learning rate of each parameter group using a cosine annealing warm restarts
|
|
1167
|
-
schedule.
|
|
1168
|
-
for learning rate, :math:`\eta_{t}` is the current learning rate, :math:`T_{0}` is the number of iterations for the
|
|
1169
|
-
first restar, :math:`T_{i}` is the current number of iterations between two warm restarts in SGDR,
|
|
1170
|
-
:math:`T_{cur}` is the number of epochs since the last restart in SGDR.
|
|
1182
|
+
schedule.
|
|
1171
1183
|
|
|
1172
1184
|
.. math::
|
|
1173
1185
|
\eta_t = \eta_{min} + \frac{1}{2}(\eta_{max} - \eta_{min})\left(1 +
|
|
1174
1186
|
\cos\left(\frac{T_{cur}}{T_{i}}\pi\right)\right)
|
|
1175
1187
|
|
|
1188
|
+
Where :math:`\eta_{max}` is set to the initial lr, :math:`\eta_{min}` is the minimum value
|
|
1189
|
+
for learning rate, :math:`\eta_{t}` is the current learning rate, :math:`T_{0}` is the number of iterations for the
|
|
1190
|
+
first restar, :math:`T_{i}` is the current number of iterations between two warm restarts in SGDR,
|
|
1191
|
+
:math:`T_{cur}` is the number of epochs since the last restart in SGDR.
|
|
1192
|
+
|
|
1176
1193
|
When :math:`T_{cur}=T_{i}`, set :math:`\eta_t = \eta_{min}`.
|
|
1177
1194
|
When :math:`T_{cur}=0` after restart, set :math:`\eta_t=\eta_{max}`.
|
|
1178
1195
|
|
|
@@ -1189,7 +1206,8 @@ class CosineAnnealingWarmRestarts(LRScheduler):
|
|
|
1189
1206
|
T_0 (int): Number of iterations for the first restart.
|
|
1190
1207
|
T_mult (int, optional): A factor increases :math:`T_{i}` after a restart. Default: ``1``.
|
|
1191
1208
|
eta_min (Union(float, int), optional): Minimum learning rate. Default: ``0``.
|
|
1192
|
-
last_epoch (int, optional): The
|
|
1209
|
+
last_epoch (int, optional): The number of times the `step()` method of
|
|
1210
|
+
the current learning rate adjustment strategy has been executed. Default: ``-1``.
|
|
1193
1211
|
|
|
1194
1212
|
Raises:
|
|
1195
1213
|
ValueError: `T_0` is less than or equal than 0 or not an int.
|
|
@@ -56,7 +56,7 @@ class RAdam(Optimizer):
|
|
|
56
56
|
|
|
57
57
|
.. math::
|
|
58
58
|
\begin{align*}
|
|
59
|
-
&\rule{
|
|
59
|
+
&\rule{180mm}{0.4pt} \\
|
|
60
60
|
&\textbf{Input}:
|
|
61
61
|
\gamma \text{ (lr)}, \: \beta_1, \beta_2 \text{ (betas)}, \: \theta_0 \text{ (params)}, \:f(\theta)
|
|
62
62
|
\text{ (objective)}, \:
|
|
@@ -67,7 +67,7 @@ class RAdam(Optimizer):
|
|
|
67
67
|
v_0 \leftarrow 0 \text{ (second moment)} \\
|
|
68
68
|
\rho_{\infty} \xleftarrow{\text{def}} \dfrac{2}{1 - \beta_2} - 1
|
|
69
69
|
\end{cases} \\
|
|
70
|
-
&\rule{
|
|
70
|
+
&\rule{180mm}{0.4pt} \\
|
|
71
71
|
&\textbf{For } t = 1 \text{ to } \ldots \text{ do}: \\
|
|
72
72
|
&\quad g_t \leftarrow \nabla_{\theta} f_t(\theta_{t - 1}) \\
|
|
73
73
|
&\quad \text{If } \lambda \neq 0: \\
|
|
@@ -84,9 +84,9 @@ class RAdam(Optimizer):
|
|
|
84
84
|
&\quad\quad \theta_t \leftarrow \theta_{t - 1} - \gamma \widehat{m_t} r_t l_t \\
|
|
85
85
|
&\quad \text{Else}: \\
|
|
86
86
|
&\quad\quad \theta_t \leftarrow \theta_{t - 1} - \gamma \widehat{m_t} \\
|
|
87
|
-
&\rule{
|
|
87
|
+
&\rule{180mm}{0.4pt} \\
|
|
88
88
|
&\bf{Return}: \theta_t \\
|
|
89
|
-
&\rule{
|
|
89
|
+
&\rule{180mm}{0.4pt}
|
|
90
90
|
\end{align*}
|
|
91
91
|
|
|
92
92
|
.. warning::
|
|
@@ -157,7 +157,7 @@ class RAdam(Optimizer):
|
|
|
157
157
|
self.increase_tensor = Tensor(1, mstype.int32)
|
|
158
158
|
self.assignadd = P.AssignAdd()
|
|
159
159
|
|
|
160
|
-
@jit
|
|
160
|
+
@jit(backend="ms_backend")
|
|
161
161
|
def implementation(self, lr, beta1, beta2, weight_decay, eps, start_id, end_id, gradients):
|
|
162
162
|
"""Extract the common computing part for acceleration"""
|
|
163
163
|
params = self.parameters[start_id: end_id]
|
|
@@ -133,7 +133,7 @@ class Rprop(Optimizer):
|
|
|
133
133
|
self.increase_tensor = Tensor(1, mstype.int32)
|
|
134
134
|
self.op_cast = P.Cast()
|
|
135
135
|
|
|
136
|
-
@jit
|
|
136
|
+
@jit(backend="ms_backend")
|
|
137
137
|
def implementation(self, etaminus, etaplus, group_id, lr, gradients, maximize, step_size_min, step_size_max):
|
|
138
138
|
"""Extract the common computing part for acceleration"""
|
|
139
139
|
etaminus, etaplus = op_cast(etaminus, mstype.float32), op_cast(etaplus, mstype.float32)
|
|
@@ -51,7 +51,7 @@ class SGD(Optimizer):
|
|
|
51
51
|
|
|
52
52
|
To be noticed, for the first step, :math:`v_{t+1} = gradient`.
|
|
53
53
|
|
|
54
|
-
Here
|
|
54
|
+
Here: where :math:`p`, :math:`v` and :math:`u` denote the parameters, accum, and momentum respectively.
|
|
55
55
|
|
|
56
56
|
.. warning::
|
|
57
57
|
This is an experimental optimizer API that is subject to change.
|
|
@@ -19,7 +19,7 @@ from mindspore.common.tensor import Tensor
|
|
|
19
19
|
from mindspore.common.api import _convert_python_data
|
|
20
20
|
from mindspore.common.dtype import type_size_in_bytes
|
|
21
21
|
from mindspore._c_expression import slice_by_tensor_index, slice_by_padding_shape, \
|
|
22
|
-
combine_tensor_list_contiguous as combine_tensor_list,
|
|
22
|
+
combine_tensor_list_contiguous as combine_tensor_list, TensorPy as Tensor_
|
|
23
23
|
|
|
24
24
|
|
|
25
25
|
def combine_tensor_list_contiguous(tensor_list, enable_mem_align=True):
|
|
@@ -29,7 +29,7 @@ def combine_tensor_list_contiguous(tensor_list, enable_mem_align=True):
|
|
|
29
29
|
Args:
|
|
30
30
|
tensor_list (list[Tensor], Tuple[Tensor]): The tensor list to be stored.
|
|
31
31
|
enable_mem_align (bool, optional): Whether to enable the memory alignment function.
|
|
32
|
-
False is not supported. Default
|
|
32
|
+
False is not supported. Default ``True`` .
|
|
33
33
|
|
|
34
34
|
Returns:
|
|
35
35
|
ContiguousTensorsHandle, a manager with contiguous memory.
|
|
@@ -59,7 +59,7 @@ class ContiguousTensorsHandle:
|
|
|
59
59
|
Args:
|
|
60
60
|
tensor_list (list[Tensor], Tuple[Tensor]): The tensor list to be stored.
|
|
61
61
|
enable_mem_align (bool, optional): Whether to enable the memory alignment function.
|
|
62
|
-
False is not supported. Default
|
|
62
|
+
False is not supported. Default ``True`` .
|
|
63
63
|
|
|
64
64
|
Returns:
|
|
65
65
|
ContiguousTensorsHandle, a manager with contiguous memory.
|
|
@@ -137,15 +137,11 @@ class ContiguousTensorsHandle:
|
|
|
137
137
|
Return the tensor which is sliced by tensor index.
|
|
138
138
|
|
|
139
139
|
Args:
|
|
140
|
-
start(int, None): Starting position. Default
|
|
141
|
-
end(int, None): Deadline position. Default
|
|
140
|
+
start(int, None): Starting position. Default ``None``.
|
|
141
|
+
end(int, None): Deadline position. Default ``None``.
|
|
142
142
|
|
|
143
143
|
Returns:
|
|
144
|
-
Tensor
|
|
145
|
-
|
|
146
|
-
Raises:
|
|
147
|
-
TypeError: If `start` or `end`, is neither an 'int' nor a 'none'.
|
|
148
|
-
ValueError: If values of `start` or `end` are negative, or out of the list range, or start >= end.
|
|
144
|
+
Tensor
|
|
149
145
|
|
|
150
146
|
Examples:
|
|
151
147
|
>>> import numpy as np
|
mindspore/hal/device.py
CHANGED
|
@@ -120,23 +120,21 @@ def _check_device_id(fn):
|
|
|
120
120
|
|
|
121
121
|
def is_initialized(device_target):
|
|
122
122
|
"""
|
|
123
|
-
|
|
123
|
+
Return whether specified device is initialized, this api will be deprecated and removed in future versions.
|
|
124
124
|
|
|
125
125
|
Note:
|
|
126
|
-
|
|
126
|
+
MindSpore's devices "CPU", "GPU" and "Ascend" will be initialized in the following scenarios:
|
|
127
127
|
|
|
128
|
-
|
|
129
|
-
|
|
130
|
-
|
|
131
|
-
- For distributed job, backend will be initialized after `mindspore.communication.init` method is called.
|
|
132
|
-
- For standalone job, backend will be initialized after running
|
|
128
|
+
- For distributed job, device will be initialized after `mindspore.communication.init` method is called.
|
|
129
|
+
- For standalone job, device will be initialized after running
|
|
133
130
|
the first operator or calling creating stream/event interfaces.
|
|
134
131
|
|
|
135
132
|
Args:
|
|
136
|
-
device_target (str): The device
|
|
133
|
+
device_target (str): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
|
|
134
|
+
``"Ascend"`` .
|
|
137
135
|
|
|
138
136
|
Returns:
|
|
139
|
-
|
|
137
|
+
bool
|
|
140
138
|
|
|
141
139
|
Examples:
|
|
142
140
|
>>> import mindspore as ms
|
|
@@ -162,25 +160,21 @@ def is_initialized(device_target):
|
|
|
162
160
|
|
|
163
161
|
def is_available(device_target):
|
|
164
162
|
"""
|
|
165
|
-
|
|
166
|
-
|
|
167
|
-
|
|
168
|
-
Note:
|
|
169
|
-
- The api will be deprecated.
|
|
170
|
-
- CPU hardware, please use the interface :func:`mindspore.device_context.cpu.is_available`.
|
|
171
|
-
- GPU hardware, please use the interface :func:`mindspore.device_context.gpu.is_available`.
|
|
172
|
-
- Ascend hardware, please use the interface :func:`mindspore.device_context.ascend.is_available`.
|
|
163
|
+
Return whether specified device is available, this api will be deprecated and removed in future versions, please
|
|
164
|
+
use the api :func:`mindspore.device_context.cpu.is_available` , :func:`mindspore.device_context.gpu.is_available`,
|
|
165
|
+
:func:`mindspore.device_context.ascend.is_available` instead.
|
|
173
166
|
|
|
174
167
|
Args:
|
|
175
|
-
device_target (str): The device
|
|
168
|
+
device_target (str): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
|
|
169
|
+
``"Ascend"`` . Default ``None`` , represents the current device set by context.
|
|
176
170
|
|
|
177
171
|
Returns:
|
|
178
|
-
|
|
172
|
+
bool
|
|
179
173
|
|
|
180
174
|
Examples:
|
|
181
|
-
>>> import mindspore
|
|
182
|
-
>>> device_target =
|
|
183
|
-
>>> print(
|
|
175
|
+
>>> import mindspore
|
|
176
|
+
>>> device_target = mindspore.context.get_context("device_target")
|
|
177
|
+
>>> print(mindspore.hal.is_available(device_target))
|
|
184
178
|
True
|
|
185
179
|
"""
|
|
186
180
|
if device_target not in valid_targets:
|
|
@@ -201,29 +195,24 @@ def is_available(device_target):
|
|
|
201
195
|
@_check_inputs_validation
|
|
202
196
|
def device_count(device_target=None):
|
|
203
197
|
"""
|
|
204
|
-
|
|
198
|
+
Return device count of specified device, this api will be deprecated and removed in future versions, please use
|
|
199
|
+
the api :func:`mindspore.device_context.cpu.device_count` , :func:`mindspore.device_context.gpu.device_count`,
|
|
200
|
+
:func:`mindspore.device_context.ascend.device_count` instead.
|
|
205
201
|
|
|
206
202
|
Note:
|
|
207
|
-
|
|
208
|
-
- CPU hardware, please use the interface :func:`mindspore.device_context.cpu.device_count`.
|
|
209
|
-
- GPU hardware, please use the interface :func:`mindspore.device_context.gpu.device_count`.
|
|
210
|
-
- Ascend hardware, please use the interface :func:`mindspore.device_context.ascend.device_count`.
|
|
211
|
-
|
|
212
|
-
|
|
213
|
-
Note:
|
|
214
|
-
If `device_target` is not specified, get the device count of the current backend set by context.
|
|
215
|
-
For CPU backend, this method always returns 1.
|
|
203
|
+
For CPU device, this method always returns 1.
|
|
216
204
|
|
|
217
205
|
Args:
|
|
218
|
-
device_target (str, optional): The device
|
|
206
|
+
device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
|
|
207
|
+
``"Ascend"`` . Default ``None`` , represents the current device set by context.
|
|
219
208
|
|
|
220
209
|
Returns:
|
|
221
|
-
int
|
|
210
|
+
int
|
|
222
211
|
|
|
223
212
|
Examples:
|
|
224
|
-
>>> import mindspore
|
|
225
|
-
>>> device_target =
|
|
226
|
-
>>> print(
|
|
213
|
+
>>> import mindspore
|
|
214
|
+
>>> device_target = mindspore.context.get_context("device_target")
|
|
215
|
+
>>> print(mindspore.hal.device_count(device_target))
|
|
227
216
|
"""
|
|
228
217
|
hal_instance = hal_instances.get(device_target)
|
|
229
218
|
if hal_instance is None:
|
|
@@ -235,20 +224,15 @@ def device_count(device_target=None):
|
|
|
235
224
|
@_check_inputs_validation
|
|
236
225
|
def get_device_capability(device_id, device_target=None):
|
|
237
226
|
"""
|
|
238
|
-
Get specified device's capability.
|
|
239
|
-
|
|
240
|
-
Note:
|
|
241
|
-
- The api will be deprecated.
|
|
242
|
-
|
|
243
|
-
Note:
|
|
244
|
-
If `device_target` is not specified, get the device capability of the current backend set by context.
|
|
227
|
+
Get specified device's capability, this api will be deprecated and removed in future versions.
|
|
245
228
|
|
|
246
229
|
Args:
|
|
247
230
|
device_id (int): The device id of which the capability will be returned.
|
|
248
|
-
device_target (str, optional):
|
|
231
|
+
device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
|
|
232
|
+
``"Ascend"`` . Default ``None`` , represents the current device set by context.
|
|
249
233
|
|
|
250
234
|
Returns:
|
|
251
|
-
tuple(
|
|
235
|
+
tuple(param1, param2) for GPU.
|
|
252
236
|
|
|
253
237
|
- param1 - int, cuda major revision number.
|
|
254
238
|
- param2 - int, cuda minor revision number.
|
|
@@ -256,9 +240,9 @@ def get_device_capability(device_id, device_target=None):
|
|
|
256
240
|
None for Ascend and CPU.
|
|
257
241
|
|
|
258
242
|
Examples:
|
|
259
|
-
>>> import mindspore
|
|
260
|
-
>>> device_target =
|
|
261
|
-
>>> print(
|
|
243
|
+
>>> import mindspore
|
|
244
|
+
>>> device_target = mindspore.context.get_context("device_target")
|
|
245
|
+
>>> print(mindspore.hal.get_device_capability(0, device_target))
|
|
262
246
|
"""
|
|
263
247
|
hal_instance = hal_instances.get(device_target)
|
|
264
248
|
if hal_instance is None:
|
|
@@ -270,20 +254,17 @@ def get_device_capability(device_id, device_target=None):
|
|
|
270
254
|
@_check_inputs_validation
|
|
271
255
|
def get_device_properties(device_id, device_target=None):
|
|
272
256
|
"""
|
|
273
|
-
Get specified device's properties.
|
|
274
|
-
|
|
275
|
-
Note:
|
|
276
|
-
- The api will be deprecated.
|
|
257
|
+
Get specified device's properties, this api will be deprecated and removed in future versions.
|
|
277
258
|
|
|
278
259
|
Note:
|
|
279
|
-
|
|
280
|
-
For Ascend, backend must be initialized before calling this method,
|
|
260
|
+
For Ascend, device must be initialized before calling this method,
|
|
281
261
|
or `total_memory` and `free_memory` will be 0,
|
|
282
262
|
and `device_id` will be ignored since this method only returns current device's properties.
|
|
283
263
|
|
|
284
264
|
Args:
|
|
285
265
|
device_id (int): The device id of which the properties will be returned.
|
|
286
|
-
device_target (str, optional): The device
|
|
266
|
+
device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
|
|
267
|
+
``"Ascend"`` . Default ``None`` , represents the current device set by context.
|
|
287
268
|
|
|
288
269
|
Returns:
|
|
289
270
|
- `cudaDeviceProp` for GPU.
|
|
@@ -314,9 +295,9 @@ def get_device_properties(device_id, device_target=None):
|
|
|
314
295
|
- None for CPU.
|
|
315
296
|
|
|
316
297
|
Examples:
|
|
317
|
-
>>> import mindspore
|
|
318
|
-
>>> device_target =
|
|
319
|
-
>>> print(
|
|
298
|
+
>>> import mindspore
|
|
299
|
+
>>> device_target = mindspore.context.get_context("device_target")
|
|
300
|
+
>>> print(mindspore.hal.get_device_properties(0, device_target))
|
|
320
301
|
"""
|
|
321
302
|
hal_instance = hal_instances.get(device_target)
|
|
322
303
|
if hal_instance is None:
|
|
@@ -328,26 +309,23 @@ def get_device_properties(device_id, device_target=None):
|
|
|
328
309
|
@_check_inputs_validation
|
|
329
310
|
def get_device_name(device_id, device_target=None):
|
|
330
311
|
"""
|
|
331
|
-
Get specified device's name.
|
|
332
|
-
|
|
333
|
-
Note:
|
|
334
|
-
- The api will be deprecated.
|
|
312
|
+
Get specified device's name, this api will be deprecated and removed in future versions.
|
|
335
313
|
|
|
336
314
|
Note:
|
|
337
|
-
|
|
338
|
-
This method always returns "CPU" for CPU backend.
|
|
315
|
+
This method always returns "CPU" for CPU device.
|
|
339
316
|
|
|
340
317
|
Args:
|
|
341
318
|
device_id (int): The device id of which the name will be returned.
|
|
342
|
-
device_target (str, optional): The device
|
|
319
|
+
device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
|
|
320
|
+
``"Ascend"`` . Default ``None`` , represents the current device set by context.
|
|
343
321
|
|
|
344
322
|
Returns:
|
|
345
|
-
str
|
|
323
|
+
str
|
|
346
324
|
|
|
347
325
|
Examples:
|
|
348
|
-
>>> import mindspore
|
|
349
|
-
>>> device_target =
|
|
350
|
-
>>> print(
|
|
326
|
+
>>> import mindspore
|
|
327
|
+
>>> device_target = mindspore.context.get_context("device_target")
|
|
328
|
+
>>> print(mindspore.hal.get_device_name(0, device_target))
|
|
351
329
|
"""
|
|
352
330
|
hal_instance = hal_instances.get(device_target)
|
|
353
331
|
if hal_instance is None:
|
|
@@ -358,25 +336,21 @@ def get_device_name(device_id, device_target=None):
|
|
|
358
336
|
@_check_inputs_validation
|
|
359
337
|
def get_arch_list(device_target=None):
|
|
360
338
|
"""
|
|
361
|
-
Get the architecture list this MindSpore was compiled for
|
|
362
|
-
|
|
363
|
-
Note:
|
|
364
|
-
- The api will be deprecated.
|
|
365
|
-
|
|
366
|
-
Note:
|
|
367
|
-
If `device_target` is not specified, get the device name of the current backend set by context.
|
|
339
|
+
Get the architecture list this MindSpore was compiled for, this api will be deprecated and removed in future
|
|
340
|
+
versions.
|
|
368
341
|
|
|
369
342
|
Args:
|
|
370
|
-
device_target (str, optional): The device
|
|
343
|
+
device_target (str, optional): The target device specified, should be one of ``"CPU"`` , ``"GPU"`` and
|
|
344
|
+
``"Ascend"`` . Default ``None`` , represents the current device set by context.
|
|
371
345
|
|
|
372
346
|
Returns:
|
|
373
347
|
str for GPU.
|
|
374
348
|
None for Ascend and CPU.
|
|
375
349
|
|
|
376
350
|
Examples:
|
|
377
|
-
>>> import mindspore
|
|
378
|
-
>>> device_target =
|
|
379
|
-
>>> print(
|
|
351
|
+
>>> import mindspore
|
|
352
|
+
>>> device_target = mindspore.context.get_context("device_target")
|
|
353
|
+
>>> print(mindspore.hal.get_arch_list(device_target))
|
|
380
354
|
"""
|
|
381
355
|
hal_instance = hal_instances.get(device_target)
|
|
382
356
|
if hal_instance is None:
|