mindspore 2.5.0__cp311-cp311-win_amd64.whl → 2.6.0__cp311-cp311-win_amd64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of mindspore might be problematic. Click here for more details.
- mindspore/.commit_id +1 -1
- mindspore/Microsoft.VisualStudio.Telemetry.dll +0 -0
- mindspore/Newtonsoft.Json.dll +0 -0
- mindspore/__init__.py +6 -4
- mindspore/_c_dataengine.cp311-win_amd64.pyd +0 -0
- mindspore/_c_expression.cp311-win_amd64.pyd +0 -0
- mindspore/_c_mindrecord.cp311-win_amd64.pyd +0 -0
- mindspore/_check_jit_forbidden_api.py +3 -0
- mindspore/_checkparam.py +3 -33
- mindspore/_deprecated/__init__.py +17 -0
- mindspore/_deprecated/jit.py +198 -0
- mindspore/_extends/builtin_operations.py +1 -1
- mindspore/_extends/parse/__init__.py +6 -7
- mindspore/_extends/parse/compile_config.py +19 -0
- mindspore/_extends/parse/deprecated/deprecated_tensor_method.py +22 -3
- mindspore/_extends/parse/jit_fallback_modules/__init__.py +0 -0
- mindspore/_extends/parse/jit_fallback_modules/check_utils.py +123 -0
- mindspore/_extends/parse/jit_fallback_modules/third_party_modules.py +50 -0
- mindspore/_extends/parse/parser.py +25 -194
- mindspore/_extends/parse/resources.py +1 -5
- mindspore/_extends/parse/standard_method.py +109 -75
- mindspore/_extends/pijit/__init__.py +2 -2
- mindspore/_extends/pijit/pijit_func_white_list.py +16 -11
- mindspore/_extends/pijit/tensor_func_list.py +27 -0
- mindspore/_extends/utils.py +1 -1
- mindspore/amp.py +4 -4
- mindspore/atlprov.dll +0 -0
- mindspore/avcodec-59.dll +0 -0
- mindspore/avdevice-59.dll +0 -0
- mindspore/avfilter-8.dll +0 -0
- mindspore/avformat-59.dll +0 -0
- mindspore/avutil-57.dll +0 -0
- mindspore/boost/__init__.py +2 -2
- mindspore/boost/base.py +3 -7
- mindspore/boost/boost_cell_wrapper.py +2 -2
- mindspore/c1.dll +0 -0
- mindspore/c1xx.dll +0 -0
- mindspore/c2.dll +0 -0
- mindspore/common/__init__.py +4 -3
- mindspore/common/_grad_function.py +56 -0
- mindspore/common/_pijit_context.py +14 -5
- mindspore/common/_register_for_tensor.py +1 -1
- mindspore/common/_stub_tensor.py +5 -10
- mindspore/common/_tensor_cpp_method.py +1 -1
- mindspore/common/_tensor_docs.py +2014 -3386
- mindspore/common/api.py +386 -355
- mindspore/common/auto_dynamic_shape.py +41 -44
- mindspore/common/dtype.py +5 -2
- mindspore/common/dump.py +7 -5
- mindspore/common/file_system.py +3 -0
- mindspore/common/generator.py +3 -0
- mindspore/common/hook_handle.py +5 -3
- mindspore/common/initializer.py +10 -6
- mindspore/common/jit_begin_end.py +94 -0
- mindspore/common/jit_config.py +6 -1
- mindspore/common/jit_context.py +76 -0
- mindspore/common/jit_trace.py +378 -0
- mindspore/common/lazy_inline.py +2 -2
- mindspore/common/mutable.py +5 -4
- mindspore/common/parameter.py +106 -39
- mindspore/common/seed.py +2 -2
- mindspore/common/sparse_tensor.py +23 -17
- mindspore/common/tensor.py +332 -714
- mindspore/communication/__init__.py +7 -5
- mindspore/communication/_comm_helper.py +47 -2
- mindspore/communication/comm_func.py +70 -53
- mindspore/communication/management.py +83 -17
- mindspore/context.py +228 -571
- mindspore/dataset/__init__.py +44 -20
- mindspore/dataset/audio/__init__.py +2 -8
- mindspore/dataset/audio/transforms.py +3 -17
- mindspore/dataset/core/config.py +3 -3
- mindspore/dataset/engine/cache_client.py +1 -1
- mindspore/dataset/engine/datasets.py +102 -120
- mindspore/dataset/engine/datasets_audio.py +22 -22
- mindspore/dataset/engine/datasets_standard_format.py +43 -24
- mindspore/dataset/engine/datasets_text.py +78 -85
- mindspore/dataset/engine/datasets_user_defined.py +109 -77
- mindspore/dataset/engine/datasets_vision.py +111 -108
- mindspore/dataset/engine/iterators.py +5 -3
- mindspore/dataset/engine/obs/obs_mindrecord_dataset.py +1 -1
- mindspore/dataset/engine/samplers.py +279 -57
- mindspore/dataset/engine/serializer_deserializer.py +2 -1
- mindspore/dataset/engine/validators.py +10 -0
- mindspore/dataset/text/__init__.py +7 -6
- mindspore/dataset/text/transforms.py +6 -5
- mindspore/dataset/text/utils.py +3 -3
- mindspore/dataset/transforms/__init__.py +0 -9
- mindspore/dataset/transforms/transforms.py +3 -3
- mindspore/dataset/utils/browse_dataset.py +1 -1
- mindspore/dataset/vision/__init__.py +2 -9
- mindspore/dataset/vision/transforms.py +202 -158
- mindspore/dataset/vision/utils.py +7 -5
- mindspore/device_context/ascend/op_debug.py +60 -1
- mindspore/device_context/ascend/op_tuning.py +0 -4
- mindspore/device_manager.py +39 -3
- mindspore/dnnl.dll +0 -0
- mindspore/dpcmi.dll +0 -0
- mindspore/experimental/es/embedding_service.py +35 -27
- mindspore/experimental/llm_boost/ascend_native/llama_boost_ascend_native.py +0 -2
- mindspore/experimental/map_parameter.py +4 -4
- mindspore/experimental/optim/adadelta.py +22 -26
- mindspore/experimental/optim/adagrad.py +4 -4
- mindspore/experimental/optim/adam.py +4 -0
- mindspore/experimental/optim/adamax.py +4 -4
- mindspore/experimental/optim/adamw.py +4 -0
- mindspore/experimental/optim/asgd.py +1 -1
- mindspore/experimental/optim/lr_scheduler.py +40 -22
- mindspore/experimental/optim/radam.py +5 -5
- mindspore/experimental/optim/rprop.py +1 -1
- mindspore/experimental/optim/sgd.py +1 -1
- mindspore/hal/contiguous_tensors_handle.py +6 -10
- mindspore/hal/device.py +55 -81
- mindspore/hal/event.py +38 -55
- mindspore/hal/memory.py +115 -147
- mindspore/hal/stream.py +81 -125
- mindspore/include/dataset/constants.h +7 -4
- mindspore/include/dataset/execute.h +2 -2
- mindspore/jpeg62.dll +0 -0
- mindspore/log.py +40 -2
- mindspore/mindrecord/__init__.py +20 -7
- mindspore/mindspore_backend_common.dll +0 -0
- mindspore/mindspore_backend_manager.dll +0 -0
- mindspore/mindspore_common.dll +0 -0
- mindspore/mindspore_core.dll +0 -0
- mindspore/mindspore_dump.dll +0 -0
- mindspore/mindspore_frontend.dll +0 -0
- mindspore/mindspore_glog.dll +0 -0
- mindspore/mindspore_memory_pool.dll +0 -0
- mindspore/mindspore_ms_backend.dll +0 -0
- mindspore/mindspore_ops.dll +0 -0
- mindspore/{mindspore_backend.dll → mindspore_ops_host.dll} +0 -0
- mindspore/mindspore_ops_kernel_common.dll +0 -0
- mindspore/mindspore_profiler.dll +0 -0
- mindspore/mindspore_pyboost.dll +0 -0
- mindspore/mindspore_pynative.dll +0 -0
- mindspore/mindspore_res_manager.dll +0 -0
- mindspore/mindspore_runtime_pipeline.dll +0 -0
- mindspore/mint/__init__.py +133 -702
- mindspore/mint/distributed/__init__.py +5 -1
- mindspore/mint/distributed/distributed.py +198 -113
- mindspore/mint/linalg/__init__.py +2 -0
- mindspore/mint/nn/__init__.py +280 -18
- mindspore/mint/nn/functional.py +282 -64
- mindspore/mint/nn/layer/__init__.py +4 -0
- mindspore/mint/nn/layer/_functions.py +7 -3
- mindspore/mint/nn/layer/activation.py +120 -13
- mindspore/mint/nn/layer/conv.py +234 -28
- mindspore/mint/nn/layer/normalization.py +15 -16
- mindspore/mint/nn/layer/padding.py +1 -1
- mindspore/mint/nn/layer/pooling.py +66 -1
- mindspore/mint/optim/__init__.py +2 -1
- mindspore/mint/optim/sgd.py +171 -0
- mindspore/msobj140.dll +0 -0
- mindspore/mspdb140.dll +0 -0
- mindspore/mspdbcore.dll +0 -0
- mindspore/mspdbst.dll +0 -0
- mindspore/mspft140.dll +0 -0
- mindspore/msvcdis140.dll +0 -0
- mindspore/msvcp140_1.dll +0 -0
- mindspore/msvcp140_2.dll +0 -0
- mindspore/msvcp140_atomic_wait.dll +0 -0
- mindspore/msvcp140_codecvt_ids.dll +0 -0
- mindspore/nn/__init__.py +4 -1
- mindspore/nn/cell.py +1253 -179
- mindspore/nn/layer/activation.py +23 -21
- mindspore/nn/layer/basic.py +22 -16
- mindspore/nn/layer/container.py +1 -1
- mindspore/nn/layer/conv.py +53 -42
- mindspore/nn/layer/embedding.py +9 -8
- mindspore/nn/layer/normalization.py +48 -42
- mindspore/nn/layer/pooling.py +75 -31
- mindspore/nn/layer/transformer.py +11 -10
- mindspore/nn/learning_rate_schedule.py +4 -2
- mindspore/nn/loss/loss.py +27 -19
- mindspore/nn/optim/ada_grad.py +6 -5
- mindspore/nn/optim/adadelta.py +9 -7
- mindspore/nn/optim/adafactor.py +1 -1
- mindspore/nn/optim/adam.py +18 -14
- mindspore/nn/optim/adamax.py +8 -7
- mindspore/nn/optim/adasum.py +5 -5
- mindspore/nn/optim/asgd.py +3 -1
- mindspore/nn/optim/ftrl.py +11 -9
- mindspore/nn/optim/lamb.py +1 -1
- mindspore/nn/optim/lazyadam.py +12 -10
- mindspore/nn/optim/momentum.py +7 -6
- mindspore/nn/optim/optimizer.py +2 -2
- mindspore/nn/optim/proximal_ada_grad.py +12 -10
- mindspore/nn/optim/rmsprop.py +13 -12
- mindspore/nn/optim/rprop.py +9 -7
- mindspore/nn/optim/sgd.py +9 -6
- mindspore/nn/optim/tft_wrapper.py +5 -2
- mindspore/nn/probability/bijector/bijector.py +17 -11
- mindspore/nn/probability/bijector/gumbel_cdf.py +5 -5
- mindspore/nn/probability/bijector/invert.py +2 -2
- mindspore/nn/probability/bijector/scalar_affine.py +3 -3
- mindspore/nn/probability/bijector/softplus.py +3 -2
- mindspore/nn/probability/distribution/beta.py +3 -3
- mindspore/nn/probability/distribution/categorical.py +1 -1
- mindspore/nn/probability/distribution/cauchy.py +4 -2
- mindspore/nn/probability/distribution/exponential.py +6 -7
- mindspore/nn/probability/distribution/gamma.py +2 -2
- mindspore/nn/probability/distribution/gumbel.py +2 -2
- mindspore/nn/probability/distribution/half_normal.py +5 -3
- mindspore/nn/probability/distribution/logistic.py +5 -3
- mindspore/nn/probability/distribution/poisson.py +1 -1
- mindspore/nn/probability/distribution/uniform.py +5 -3
- mindspore/nn/reinforcement/_tensors_queue.py +1 -1
- mindspore/nn/reinforcement/tensor_array.py +1 -1
- mindspore/nn/wrap/__init__.py +6 -6
- mindspore/nn/wrap/cell_wrapper.py +178 -117
- mindspore/nn/wrap/grad_reducer.py +45 -36
- mindspore/nn/wrap/loss_scale.py +3 -3
- mindspore/numpy/array_creations.py +3 -3
- mindspore/numpy/array_ops.py +1 -1
- mindspore/numpy/utils.py +1 -2
- mindspore/numpy/utils_const.py +1 -2
- mindspore/opencv_core452.dll +0 -0
- mindspore/opencv_imgcodecs452.dll +0 -0
- mindspore/opencv_imgproc452.dll +0 -0
- mindspore/ops/__init__.py +3 -2
- mindspore/ops/_grad_experimental/grad_comm_ops.py +18 -3
- mindspore/ops/_grad_experimental/grad_debug_ops.py +8 -1
- mindspore/ops/_grad_experimental/taylor_rule.py +29 -0
- mindspore/ops/_register_for_op.py +0 -11
- mindspore/{ops_generate → ops/_utils}/arg_dtype_cast.py +123 -4
- mindspore/{ops_generate → ops/_utils}/arg_handler.py +3 -4
- mindspore/ops/_vmap/vmap_array_ops.py +32 -6
- mindspore/ops/_vmap/vmap_grad_nn_ops.py +2 -1
- mindspore/ops/_vmap/vmap_math_ops.py +4 -7
- mindspore/ops/_vmap/vmap_nn_ops.py +9 -8
- mindspore/ops/auto_generate/__init__.py +4 -3
- mindspore/ops/auto_generate/cpp_create_prim_instance_helper.py +127 -52
- mindspore/ops/auto_generate/gen_extend_func.py +286 -208
- mindspore/ops/auto_generate/gen_ops_def.py +2783 -2335
- mindspore/ops/auto_generate/gen_ops_prim.py +8992 -2686
- mindspore/ops/auto_generate/pyboost_inner_prim.py +106 -76
- mindspore/ops/composite/__init__.py +2 -1
- mindspore/ops/composite/base.py +19 -24
- mindspore/ops/composite/math_ops.py +6 -16
- mindspore/ops/composite/multitype_ops/__init__.py +5 -2
- mindspore/ops/composite/multitype_ops/_compile_utils.py +4 -5
- mindspore/ops/composite/multitype_ops/_constexpr_utils.py +1 -2
- mindspore/ops/composite/multitype_ops/add_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/bitwise_and_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/bitwise_or_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/bitwise_xor_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/div_impl.py +6 -4
- mindspore/ops/composite/multitype_ops/equal_impl.py +4 -3
- mindspore/ops/composite/multitype_ops/floordiv_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/getitem_impl.py +3 -2
- mindspore/ops/composite/multitype_ops/greater_equal_impl.py +4 -3
- mindspore/ops/composite/multitype_ops/greater_impl.py +4 -3
- mindspore/ops/composite/multitype_ops/in_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/invert_impl.py +50 -0
- mindspore/ops/composite/multitype_ops/left_shift_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/less_equal_impl.py +4 -3
- mindspore/ops/composite/multitype_ops/less_impl.py +4 -3
- mindspore/ops/composite/multitype_ops/logic_not_impl.py +3 -2
- mindspore/ops/composite/multitype_ops/logical_and_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/logical_or_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/mod_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/mul_impl.py +3 -2
- mindspore/ops/composite/multitype_ops/negative_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/not_equal_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/not_in_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/ones_like_impl.py +18 -0
- mindspore/ops/composite/multitype_ops/pow_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/right_shift_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/setitem_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/sub_impl.py +2 -1
- mindspore/ops/function/__init__.py +28 -2
- mindspore/ops/function/_add_attr_func.py +58 -0
- mindspore/ops/function/array_func.py +1631 -2347
- mindspore/ops/function/clip_func.py +38 -45
- mindspore/ops/function/debug_func.py +36 -44
- mindspore/ops/function/grad/__init__.py +1 -0
- mindspore/ops/function/grad/grad_func.py +104 -71
- mindspore/ops/function/image_func.py +1 -1
- mindspore/ops/function/linalg_func.py +46 -78
- mindspore/ops/function/math_func.py +3024 -3855
- mindspore/ops/function/nn_func.py +678 -274
- mindspore/ops/function/other_func.py +159 -1
- mindspore/ops/function/parameter_func.py +17 -30
- mindspore/ops/function/random_func.py +216 -361
- mindspore/ops/function/reshard_func.py +4 -70
- mindspore/ops/function/sparse_func.py +3 -3
- mindspore/ops/function/sparse_unary_func.py +5 -5
- mindspore/ops/function/spectral_func.py +25 -58
- mindspore/ops/function/vmap_func.py +26 -18
- mindspore/ops/functional.py +8 -5
- mindspore/ops/functional_overload.py +655 -4
- mindspore/ops/op_info_register.py +32 -244
- mindspore/ops/operations/__init__.py +21 -14
- mindspore/ops/operations/_custom_ops_utils.py +235 -0
- mindspore/ops/operations/_grad_ops.py +1 -10
- mindspore/ops/operations/_inner_ops.py +5 -76
- mindspore/ops/operations/_ms_kernel.py +4 -10
- mindspore/ops/operations/_rl_inner_ops.py +1 -1
- mindspore/ops/operations/_scalar_ops.py +3 -2
- mindspore/ops/operations/_sequence_ops.py +1 -1
- mindspore/ops/operations/_tensor_array.py +1 -1
- mindspore/ops/operations/array_ops.py +39 -24
- mindspore/ops/operations/comm_ops.py +150 -107
- mindspore/ops/operations/custom_ops.py +287 -32
- mindspore/ops/operations/debug_ops.py +119 -16
- mindspore/ops/operations/inner_ops.py +1 -1
- mindspore/ops/operations/linalg_ops.py +1 -58
- mindspore/ops/operations/manually_defined/_inner.py +1 -1
- mindspore/ops/operations/manually_defined/ops_def.py +746 -79
- mindspore/ops/operations/math_ops.py +21 -18
- mindspore/ops/operations/nn_ops.py +67 -224
- mindspore/ops/operations/other_ops.py +62 -9
- mindspore/ops/operations/random_ops.py +13 -7
- mindspore/ops/operations/reshard_ops.py +1 -1
- mindspore/ops/operations/sparse_ops.py +2 -2
- mindspore/ops/primitive.py +43 -32
- mindspore/ops/tensor_method.py +243 -17
- mindspore/ops_generate/__init__.py +0 -5
- mindspore/ops_generate/aclnn/__init__.py +0 -0
- mindspore/ops_generate/{aclnn_kernel_register_auto_cc_generator.py → aclnn/aclnn_kernel_register_auto_cc_generator.py} +43 -18
- mindspore/ops_generate/{gen_aclnn_implement.py → aclnn/gen_aclnn_implement.py} +49 -51
- mindspore/ops_generate/api/__init__.py +0 -0
- mindspore/ops_generate/{add_tensor_docs_generator.py → api/add_tensor_docs_generator.py} +9 -7
- mindspore/ops_generate/{cpp_create_prim_instance_helper_generator.py → api/cpp_create_prim_instance_helper_generator.py} +6 -9
- mindspore/ops_generate/{functional_map_cpp_generator.py → api/functional_map_cpp_generator.py} +25 -12
- mindspore/ops_generate/{functional_overload_py_generator.py → api/functional_overload_py_generator.py} +8 -6
- mindspore/ops_generate/{functions_cc_generator.py → api/functions_cc_generator.py} +14 -10
- mindspore/ops_generate/api/gen_api.py +103 -0
- mindspore/ops_generate/{op_api_proto.py → api/op_api_proto.py} +98 -69
- mindspore/ops_generate/{tensor_func_reg_cpp_generator.py → api/tensor_func_reg_cpp_generator.py} +82 -43
- mindspore/ops_generate/common/__init__.py +0 -0
- mindspore/ops_generate/common/gen_constants.py +91 -0
- mindspore/ops_generate/{gen_utils.py → common/gen_utils.py} +72 -19
- mindspore/ops_generate/{op_proto.py → common/op_proto.py} +64 -1
- mindspore/ops_generate/{template.py → common/template.py} +96 -84
- mindspore/ops_generate/gen_ops.py +23 -325
- mindspore/ops_generate/op_def/__init__.py +0 -0
- mindspore/ops_generate/op_def/gen_op_def.py +90 -0
- mindspore/ops_generate/{lite_ops_cpp_generator.py → op_def/lite_ops_cpp_generator.py} +47 -11
- mindspore/ops_generate/{ops_def_cc_generator.py → op_def/ops_def_cc_generator.py} +18 -10
- mindspore/ops_generate/{ops_def_h_generator.py → op_def/ops_def_h_generator.py} +5 -5
- mindspore/ops_generate/{ops_name_h_generator.py → op_def/ops_name_h_generator.py} +30 -15
- mindspore/ops_generate/op_def/ops_primitive_h_generator.py +125 -0
- mindspore/ops_generate/op_def_py/__init__.py +0 -0
- mindspore/ops_generate/op_def_py/gen_op_def_py.py +47 -0
- mindspore/ops_generate/{op_def_py_generator.py → op_def_py/op_def_py_generator.py} +6 -5
- mindspore/ops_generate/{op_prim_py_generator.py → op_def_py/op_prim_py_generator.py} +24 -15
- mindspore/ops_generate/pyboost/__init__.py +0 -0
- mindspore/ops_generate/{auto_grad_impl_cc_generator.py → pyboost/auto_grad_impl_cc_generator.py} +11 -7
- mindspore/ops_generate/{auto_grad_reg_cc_generator.py → pyboost/auto_grad_reg_cc_generator.py} +7 -7
- mindspore/ops_generate/{gen_pyboost_func.py → pyboost/gen_pyboost_func.py} +40 -16
- mindspore/ops_generate/{op_template_parser.py → pyboost/op_template_parser.py} +105 -24
- mindspore/ops_generate/{pyboost_functions_cpp_generator.py → pyboost/pyboost_functions_cpp_generator.py} +55 -18
- mindspore/ops_generate/{pyboost_functions_h_generator.py → pyboost/pyboost_functions_h_generator.py} +42 -10
- mindspore/ops_generate/{pyboost_functions_py_generator.py → pyboost/pyboost_functions_py_generator.py} +6 -6
- mindspore/ops_generate/{pyboost_grad_function_cpp_generator.py → pyboost/pyboost_grad_function_cpp_generator.py} +11 -10
- mindspore/ops_generate/{pyboost_inner_prim_generator.py → pyboost/pyboost_inner_prim_generator.py} +8 -7
- mindspore/ops_generate/{pyboost_native_grad_functions_generator.py → pyboost/pyboost_native_grad_functions_generator.py} +14 -10
- mindspore/ops_generate/{pyboost_op_cpp_code_generator.py → pyboost/pyboost_op_cpp_code_generator.py} +140 -53
- mindspore/ops_generate/{pyboost_overload_functions_cpp_generator.py → pyboost/pyboost_overload_functions_cpp_generator.py} +28 -15
- mindspore/ops_generate/{pyboost_utils.py → pyboost/pyboost_utils.py} +88 -4
- mindspore/ops_generate/resources/__init__.py +0 -0
- mindspore/ops_generate/resources/resource_list.py +30 -0
- mindspore/ops_generate/resources/resource_loader.py +36 -0
- mindspore/ops_generate/resources/resource_manager.py +64 -0
- mindspore/ops_generate/resources/yaml_loader.py +88 -0
- mindspore/ops_generate/tensor_py_cc_generator.py +122 -0
- mindspore/parallel/__init__.py +6 -2
- mindspore/parallel/_auto_parallel_context.py +140 -12
- mindspore/parallel/_cell_wrapper.py +132 -15
- mindspore/parallel/_parallel_serialization.py +95 -4
- mindspore/parallel/_ps_context.py +1 -1
- mindspore/parallel/_recovery_context.py +7 -2
- mindspore/parallel/_tensor.py +142 -18
- mindspore/parallel/_utils.py +198 -25
- mindspore/parallel/algo_parameter_config.py +3 -3
- mindspore/parallel/auto_parallel.py +732 -0
- mindspore/parallel/checkpoint_convert.py +159 -0
- mindspore/parallel/checkpoint_transform.py +658 -37
- mindspore/parallel/cluster/process_entity/_api.py +151 -19
- mindspore/parallel/cluster/run.py +1 -1
- mindspore/parallel/function/__init__.py +24 -0
- mindspore/parallel/function/reshard_func.py +258 -0
- mindspore/parallel/nn/__init__.py +25 -0
- mindspore/parallel/nn/parallel_cell_wrapper.py +263 -0
- mindspore/parallel/nn/parallel_grad_reducer.py +169 -0
- mindspore/parallel/parameter_broadcast.py +24 -13
- mindspore/parallel/shard.py +137 -62
- mindspore/parallel/transform_safetensors.py +288 -95
- mindspore/pgodb140.dll +0 -0
- mindspore/pgort140.dll +0 -0
- mindspore/profiler/__init__.py +9 -5
- mindspore/profiler/analysis/parser/ascend_cann_parser.py +6 -2
- mindspore/profiler/analysis/parser/ms_framework_parser.py +4 -4
- mindspore/profiler/analysis/parser/timeline_assembly_factory/ascend_timeline_assembler.py +7 -4
- mindspore/profiler/analysis/parser/timeline_assembly_factory/trace_view_container.py +25 -0
- mindspore/profiler/analysis/parser/timeline_creator/fwk_timeline_creator.py +3 -3
- mindspore/profiler/analysis/parser/timeline_event/fwk_event.py +241 -86
- mindspore/profiler/analysis/viewer/ascend_communication_viewer.py +41 -2
- mindspore/profiler/analysis/viewer/ascend_kernel_details_viewer.py +33 -35
- mindspore/profiler/analysis/viewer/ascend_memory_viewer.py +7 -0
- mindspore/profiler/analysis/viewer/ascend_op_memory_viewer.py +8 -3
- mindspore/profiler/analysis/viewer/ascend_step_trace_time_viewer.py +141 -30
- mindspore/profiler/analysis/viewer/ms_dataset_viewer.py +5 -6
- mindspore/profiler/common/ascend_msprof_exporter.py +5 -4
- mindspore/profiler/common/constant.py +12 -0
- mindspore/profiler/common/msprof_cmd_tool.py +42 -23
- mindspore/profiler/common/path_manager.py +24 -0
- mindspore/profiler/common/profiler_context.py +26 -2
- mindspore/profiler/common/profiler_meta_data.py +74 -0
- mindspore/profiler/common/profiler_parameters.py +59 -18
- mindspore/profiler/common/profiler_path_manager.py +66 -7
- mindspore/profiler/dynamic_profiler.py +112 -79
- mindspore/profiler/envprofiler.py +26 -1
- mindspore/profiler/experimental_config.py +197 -0
- mindspore/profiler/mstx.py +57 -14
- mindspore/profiler/platform/npu_profiler.py +33 -7
- mindspore/profiler/profiler.py +541 -45
- mindspore/profiler/profiler_action_controller.py +1 -1
- mindspore/profiler/profiler_interface.py +4 -0
- mindspore/profiler/schedule.py +57 -22
- mindspore/rewrite/api/node.py +15 -13
- mindspore/rewrite/api/symbol_tree.py +1 -1
- mindspore/run_check/_check_version.py +25 -14
- mindspore/run_check/run_check.py +1 -1
- mindspore/runtime/__init__.py +2 -2
- mindspore/runtime/executor.py +40 -11
- mindspore/runtime/memory.py +37 -13
- mindspore/safeguard/rewrite_obfuscation.py +12 -9
- mindspore/swresample-4.dll +0 -0
- mindspore/swscale-6.dll +0 -0
- mindspore/tbbmalloc.dll +0 -0
- mindspore/tinyxml2.dll +0 -0
- mindspore/train/__init__.py +8 -8
- mindspore/train/_utils.py +43 -9
- mindspore/train/amp.py +1 -1
- mindspore/train/callback/__init__.py +2 -2
- mindspore/train/callback/_callback.py +2 -16
- mindspore/train/callback/_checkpoint.py +24 -40
- mindspore/train/callback/_cluster_monitor.py +14 -18
- mindspore/train/callback/_flops_collector.py +2 -3
- mindspore/train/callback/_history.py +7 -4
- mindspore/train/callback/_lambda_callback.py +2 -2
- mindspore/train/callback/_landscape.py +0 -3
- mindspore/train/callback/_loss_monitor.py +2 -1
- mindspore/train/callback/_on_request_exit.py +6 -5
- mindspore/train/callback/_reduce_lr_on_plateau.py +11 -6
- mindspore/train/callback/_summary_collector.py +8 -13
- mindspore/train/callback/_time_monitor.py +2 -1
- mindspore/train/callback/{_tft_register.py → _train_fault_tolerance.py} +204 -105
- mindspore/train/data_sink.py +25 -2
- mindspore/train/dataset_helper.py +4 -5
- mindspore/train/loss_scale_manager.py +8 -7
- mindspore/train/metrics/accuracy.py +3 -3
- mindspore/train/metrics/confusion_matrix.py +9 -9
- mindspore/train/metrics/error.py +3 -3
- mindspore/train/metrics/hausdorff_distance.py +4 -4
- mindspore/train/metrics/mean_surface_distance.py +3 -3
- mindspore/train/metrics/metric.py +0 -12
- mindspore/train/metrics/occlusion_sensitivity.py +4 -2
- mindspore/train/metrics/precision.py +8 -6
- mindspore/train/metrics/recall.py +9 -9
- mindspore/train/metrics/root_mean_square_surface_distance.py +2 -2
- mindspore/train/mind_ir_pb2.py +19 -12
- mindspore/train/model.py +262 -127
- mindspore/train/serialization.py +246 -988
- mindspore/train/summary/_summary_adapter.py +2 -2
- mindspore/train/summary/summary_record.py +1 -1
- mindspore/turbojpeg.dll +0 -0
- mindspore/utils/__init__.py +3 -2
- mindspore/utils/dryrun.py +4 -2
- mindspore/utils/hooks.py +81 -0
- mindspore/utils/runtime_execution_order_check.py +2 -0
- mindspore/utils/utils.py +138 -4
- mindspore/vcmeta.dll +0 -0
- mindspore/vcruntime140.dll +0 -0
- mindspore/vcruntime140_1.dll +0 -0
- mindspore/version.py +1 -1
- {mindspore-2.5.0.dist-info → mindspore-2.6.0.dist-info}/METADATA +2 -1
- {mindspore-2.5.0.dist-info → mindspore-2.6.0.dist-info}/RECORD +485 -440
- mindspore/_install_custom.py +0 -43
- mindspore/common/_register_for_adapter.py +0 -74
- mindspore/ops/auto_generate/gen_arg_dtype_cast.py +0 -252
- mindspore/ops/auto_generate/gen_arg_handler.py +0 -136
- mindspore/ops/operations/_opaque_predicate_registry.py +0 -41
- mindspore/ops_generate/gen_constants.py +0 -190
- mindspore/ops_generate/gen_ops_inner_prim.py +0 -131
- mindspore/ops_generate/ops_primitive_h_generator.py +0 -81
- /mindspore/ops_generate/{base_generator.py → common/base_generator.py} +0 -0
- {mindspore-2.5.0.dist-info → mindspore-2.6.0.dist-info}/WHEEL +0 -0
- {mindspore-2.5.0.dist-info → mindspore-2.6.0.dist-info}/entry_points.txt +0 -0
- {mindspore-2.5.0.dist-info → mindspore-2.6.0.dist-info}/top_level.txt +0 -0
|
@@ -286,37 +286,40 @@ class BatchNorm2d(_BatchNorm):
|
|
|
286
286
|
Note that the formula for updating the :math:`moving\_mean` and :math:`moving\_var` is
|
|
287
287
|
|
|
288
288
|
.. math::
|
|
289
|
-
\text{moving_mean}=\text{moving_mean*momentum}+μ_β\text{*(1
|
|
290
|
-
\text{moving_var}=\text{moving_var*momentum}+σ^2_β\text{*(1
|
|
289
|
+
\text{moving_mean}=\text{moving_mean*momentum}+μ_β\text{*(1-momentum)}\\
|
|
290
|
+
\text{moving_var}=\text{moving_var*momentum}+σ^2_β\text{*(1-momentum)}
|
|
291
291
|
|
|
292
292
|
where :math:`moving\_mean` is the updated mean, :math:`moving\_var` is the updated variance,
|
|
293
|
-
:math:`μ_β, σ^2_β` are the observed value (mean and variance) of each batch of data.
|
|
293
|
+
:math:`μ_β, σ^2_β` are the observed value (mean and variance respectively) of each batch of data.
|
|
294
294
|
|
|
295
295
|
Args:
|
|
296
296
|
num_features (int): The number of channels of the input tensor. Expected input size is :math:`(N, C, H, W)`,
|
|
297
297
|
`C` represents the number of channels.
|
|
298
|
-
eps (float): :math:`\epsilon` added to the denominator for numerical stability. Default: ``1e-5`` .
|
|
299
|
-
momentum (float): A floating hyperparameter of the momentum for the
|
|
298
|
+
eps (float, optional): :math:`\epsilon` added to the denominator for numerical stability. Default: ``1e-5`` .
|
|
299
|
+
momentum (float, optional): A floating hyperparameter of the momentum for the
|
|
300
300
|
running_mean and running_var computation. Default: ``0.9`` .
|
|
301
|
-
affine (bool): A bool value. When set to ``True`` , :math:`\gamma` and :math:`\beta` can be learned.
|
|
301
|
+
affine (bool, optional): A bool value. When set to ``True`` , :math:`\gamma` and :math:`\beta` can be learned.
|
|
302
302
|
Default: ``True`` .
|
|
303
|
-
gamma_init (Union[Tensor, str, Initializer, numbers.Number]): Initializer for the
|
|
303
|
+
gamma_init (Union[Tensor, str, Initializer, numbers.Number], optional): Initializer for the
|
|
304
|
+
:math:`\gamma` weight.
|
|
304
305
|
The values of str refer to the function `mindspore.common.initializer
|
|
305
306
|
<https://www.mindspore.cn/docs/en/master/api_python/mindspore.common.initializer.html>`_
|
|
306
307
|
including ``'zeros'`` , ``'ones'`` , etc. Default: ``'ones'`` .
|
|
307
|
-
beta_init (Union[Tensor, str, Initializer, numbers.Number]): Initializer for the
|
|
308
|
+
beta_init (Union[Tensor, str, Initializer, numbers.Number], optional): Initializer for the
|
|
309
|
+
:math:`\beta` weight.
|
|
308
310
|
The values of str refer to the function `mindspore.common.initializer
|
|
309
311
|
<https://www.mindspore.cn/docs/en/master/api_python/mindspore.common.initializer.html>`_
|
|
310
312
|
including ``'zeros'`` , ``'ones'`` , etc. Default: ``'zeros'`` .
|
|
311
|
-
moving_mean_init (Union[Tensor, str, Initializer, numbers.Number]): Initializer for the moving mean.
|
|
313
|
+
moving_mean_init (Union[Tensor, str, Initializer, numbers.Number], optional): Initializer for the moving mean.
|
|
312
314
|
The values of str refer to the function `mindspore.common.initializer
|
|
313
315
|
<https://www.mindspore.cn/docs/en/master/api_python/mindspore.common.initializer.html>`_
|
|
314
316
|
including ``'zeros'`` , ``'ones'`` , etc. Default: ``'zeros'`` .
|
|
315
|
-
moving_var_init (Union[Tensor, str, Initializer, numbers.Number]): Initializer for
|
|
317
|
+
moving_var_init (Union[Tensor, str, Initializer, numbers.Number], optional): Initializer for
|
|
318
|
+
the moving variance.
|
|
316
319
|
The values of str refer to the function `mindspore.common.initializer
|
|
317
320
|
<https://www.mindspore.cn/docs/en/master/api_python/mindspore.common.initializer.html>`_
|
|
318
321
|
including ``'zeros'`` , ``'ones'`` , etc. Default: ``'ones'`` .
|
|
319
|
-
use_batch_statistics (bool): Default: ``None`` .
|
|
322
|
+
use_batch_statistics (bool, optional): Default: ``None`` .
|
|
320
323
|
|
|
321
324
|
- If ``true`` , use the mean value and variance value of current batch data and track running mean
|
|
322
325
|
and running variance.
|
|
@@ -325,9 +328,9 @@ class BatchNorm2d(_BatchNorm):
|
|
|
325
328
|
training and evaluation mode. During training, the parameter is set to true, and during evaluation, the
|
|
326
329
|
parameter is set to false.
|
|
327
330
|
|
|
328
|
-
data_format (str): The optional value for data format, is ``'NHWC'`` or ``'NCHW'`` .
|
|
331
|
+
data_format (str, optional): The optional value for data format, is ``'NHWC'`` or ``'NCHW'`` .
|
|
329
332
|
Default: ``'NCHW'`` .
|
|
330
|
-
dtype (:class:`mindspore.dtype
|
|
333
|
+
dtype (:class:`mindspore.dtype`, optional): Dtype of Parameters. Default: ``mstype.float32`` .
|
|
331
334
|
|
|
332
335
|
Inputs:
|
|
333
336
|
- **x** (Tensor) - Tensor of shape :math:`(N, C, H, W)`. Supported types: float16, float32.
|
|
@@ -340,7 +343,7 @@ class BatchNorm2d(_BatchNorm):
|
|
|
340
343
|
TypeError: If `eps` is not a float.
|
|
341
344
|
ValueError: If `num_features` is less than 1.
|
|
342
345
|
ValueError: If `momentum` is not in range [0, 1].
|
|
343
|
-
ValueError: If `data_format` is neither 'NHWC' not 'NCHW'
|
|
346
|
+
ValueError: If `data_format` is neither ``'NHWC'`` not ``'NCHW'``.
|
|
344
347
|
|
|
345
348
|
Supported Platforms:
|
|
346
349
|
``Ascend`` ``GPU`` ``CPU``
|
|
@@ -508,32 +511,34 @@ class SyncBatchNorm(_BatchNorm):
|
|
|
508
511
|
|
|
509
512
|
Args:
|
|
510
513
|
num_features (int): `C` from an expected input of size :math:`(N, C, H, W)`.
|
|
511
|
-
eps (float): :math:`\epsilon`, a value added to the denominator for numerical stability.
|
|
512
|
-
|
|
514
|
+
eps (float, optional): :math:`\epsilon`, a value added to the denominator for numerical stability.
|
|
515
|
+
Default: ``1e-5`` .
|
|
516
|
+
momentum (float, optional): A floating hyperparameter of the momentum for the
|
|
513
517
|
running_mean and running_var computation. Default: ``0.9`` .
|
|
514
|
-
affine (bool): A bool value. When set to ``True`` , :math:`\gamma` and :math:`\beta`
|
|
518
|
+
affine (bool, optional): A bool value. When set to ``True`` , :math:`\gamma` and :math:`\beta` are learnable
|
|
519
|
+
parameters. When set to ``False`` , :math:`\gamma` and :math:`\beta` are unlearnable parameters.
|
|
515
520
|
Default: ``True`` .
|
|
516
|
-
gamma_init (Union[Tensor, str, Initializer, numbers.Number]): Initializer for the :math:`\gamma`
|
|
517
|
-
The values of str refer to the function `initializer` including ``'zeros'`` , ``'ones'`` ,
|
|
521
|
+
gamma_init (Union[Tensor, str, Initializer, numbers.Number], optional): Initializer for the :math:`\gamma`
|
|
522
|
+
weight. The values of str refer to the function `initializer` including ``'zeros'`` , ``'ones'`` ,
|
|
518
523
|
``'xavier_uniform'`` , ``'he_uniform'`` , etc. Default: ``'ones'`` .
|
|
519
|
-
beta_init (Union[Tensor, str, Initializer, numbers.Number]): Initializer for the :math:`\beta` weight.
|
|
524
|
+
beta_init (Union[Tensor, str, Initializer, numbers.Number], optional): Initializer for the :math:`\beta` weight.
|
|
520
525
|
The values of str refer to the function `initializer` including ``'zeros'`` , ``'ones'`` ,
|
|
521
526
|
``'xavier_uniform'`` , ``'he_uniform'`` , etc. Default: ``'zeros'`` .
|
|
522
|
-
moving_mean_init (Union[Tensor, str, Initializer, numbers.Number]): Initializer for the moving mean.
|
|
527
|
+
moving_mean_init (Union[Tensor, str, Initializer, numbers.Number], optional): Initializer for the moving mean.
|
|
523
528
|
The values of str refer to the function `initializer` including ``'zeros'`` , ``'ones'`` ,
|
|
524
529
|
``'xavier_uniform'`` , ``'he_uniform'`` , etc. Default: ``'zeros'`` .
|
|
525
|
-
moving_var_init (Union[Tensor, str, Initializer, numbers.Number]): Initializer for the moving
|
|
526
|
-
The values of str refer to the function `initializer` including ``'zeros'`` , ``'ones'`` ,
|
|
530
|
+
moving_var_init (Union[Tensor, str, Initializer, numbers.Number], optional): Initializer for the moving
|
|
531
|
+
variance. The values of str refer to the function `initializer` including ``'zeros'`` , ``'ones'`` ,
|
|
527
532
|
``'xavier_uniform'`` , ``'he_uniform'`` , etc. Default: ``'ones'`` .
|
|
528
|
-
use_batch_statistics (bool): If ``true`` , use the mean value and variance value of current batch
|
|
529
|
-
``false`` , use the mean value and variance value of specified value. If ``None`` , training
|
|
530
|
-
use the mean and variance of current batch data and track the running mean and variance, eval
|
|
531
|
-
use the running mean and variance. Default: ``None`` .
|
|
532
|
-
process_groups (list): A list to divide devices into different sync groups, containing N subtraction
|
|
533
|
-
Each subtraction list contains int numbers identifying rank ids which need to be synchronized in the
|
|
534
|
-
group. All int values must be in [0, rank_size) and different from each other. Default: ``None`` ,
|
|
533
|
+
use_batch_statistics (bool, optional): If ``true`` , use the mean value and variance value of current batch
|
|
534
|
+
data. If ``false`` , use the mean value and variance value of specified value. If ``None`` , training
|
|
535
|
+
process will use the mean and variance of current batch data and track the running mean and variance, eval
|
|
536
|
+
process will use the running mean and variance. Default: ``None`` .
|
|
537
|
+
process_groups (list, optional): A list to divide devices into different sync groups, containing N subtraction
|
|
538
|
+
lists. Each subtraction list contains int numbers identifying rank ids which need to be synchronized in the
|
|
539
|
+
same group. All int values must be in [0, rank_size) and different from each other. Default: ``None`` ,
|
|
535
540
|
indicating synchronization across all devices.
|
|
536
|
-
dtype (:class:`mindspore.dtype
|
|
541
|
+
dtype (:class:`mindspore.dtype`, optional): Dtype of Parameters. Default: ``mstype.float32`` .
|
|
537
542
|
|
|
538
543
|
Inputs:
|
|
539
544
|
- **x** (Tensor) - Tensor of shape :math:`(N, C_{in}, H_{in}, W_{in})`.
|
|
@@ -558,14 +563,14 @@ class SyncBatchNorm(_BatchNorm):
|
|
|
558
563
|
|
|
559
564
|
For the Ascend devices, users need to prepare the rank table, set rank_id and device_id.
|
|
560
565
|
Please see the `Ascend tutorial
|
|
561
|
-
<https://www.mindspore.cn/
|
|
566
|
+
<https://www.mindspore.cn/tutorials/en/master/parallel/rank_table.html>`_
|
|
562
567
|
for more details.
|
|
563
568
|
|
|
564
569
|
For the GPU devices, users need to prepare the host file and mpi, please see the `mpirun Startup
|
|
565
|
-
<https://www.mindspore.cn/
|
|
570
|
+
<https://www.mindspore.cn/tutorials/en/master/parallel/mpirun.html>`_ .
|
|
566
571
|
|
|
567
572
|
For the CPU device, users need to write a dynamic cluster startup script, please see the `Dynamic Cluster
|
|
568
|
-
Startup <https://www.mindspore.cn/
|
|
573
|
+
Startup <https://www.mindspore.cn/tutorials/en/master/parallel/dynamic_cluster.html>`_ .
|
|
569
574
|
|
|
570
575
|
This example should be run with multiple devices.
|
|
571
576
|
|
|
@@ -987,8 +992,8 @@ class InstanceNorm1d(_InstanceNorm):
|
|
|
987
992
|
ValueError: If `num_features` is less than 1.
|
|
988
993
|
ValueError: If `momentum` is not in range [0, 1].
|
|
989
994
|
ValueError: If the shape of `gamma_init` / `beta_init` is not :math:`(C)`.
|
|
990
|
-
KeyError: If any of `gamma_init`/`beta_init` is str and
|
|
991
|
-
|
|
995
|
+
KeyError: If any of `gamma_init`/`beta_init` is str and
|
|
996
|
+
there is no homonymous class inheriting from `Initializer`.
|
|
992
997
|
|
|
993
998
|
Supported Platforms:
|
|
994
999
|
``GPU``
|
|
@@ -1065,8 +1070,8 @@ class InstanceNorm2d(_InstanceNorm):
|
|
|
1065
1070
|
ValueError: If `num_features` is less than 1.
|
|
1066
1071
|
ValueError: If `momentum` is not in range [0, 1].
|
|
1067
1072
|
ValueError: If the shape of `gamma_init` / `beta_init` is not :math:`(C)`.
|
|
1068
|
-
KeyError: If any of `gamma_init`/`beta_init` is str and
|
|
1069
|
-
|
|
1073
|
+
KeyError: If any of `gamma_init`/`beta_init` is str and
|
|
1074
|
+
there is no homonymous class inheriting from `Initializer`.
|
|
1070
1075
|
|
|
1071
1076
|
Supported Platforms:
|
|
1072
1077
|
``GPU``
|
|
@@ -1171,10 +1176,11 @@ class GroupNorm(Cell):
|
|
|
1171
1176
|
|
|
1172
1177
|
Group Normalization is widely used in recurrent neural networks. It applies
|
|
1173
1178
|
normalization on a mini-batch of inputs for each single training case as described
|
|
1174
|
-
in the paper `Group Normalization <https://arxiv.org/pdf/1803.08494.pdf>`_.
|
|
1175
|
-
|
|
1176
|
-
|
|
1177
|
-
and
|
|
1179
|
+
in the paper `Group Normalization <https://arxiv.org/pdf/1803.08494.pdf>`_.
|
|
1180
|
+
Group Normalization
|
|
1181
|
+
divides the channels into groups and computes within each group the mean and variance for normalization.
|
|
1182
|
+
:math:`\gamma` and :math:`\beta` are scale
|
|
1183
|
+
and shift values obtained by training learning.
|
|
1178
1184
|
It can be described using the following formula:
|
|
1179
1185
|
|
|
1180
1186
|
.. math::
|
mindspore/nn/layer/pooling.py
CHANGED
|
@@ -32,7 +32,7 @@ from mindspore.ops.auto_generate import avg_pool1d_ext
|
|
|
32
32
|
__all__ = ['AvgPool3d', 'MaxPool3d', 'AvgPool2d', 'MaxPool2d', 'AvgPool1d', 'MaxPool1d', 'FractionalMaxPool2d',
|
|
33
33
|
'FractionalMaxPool3d', 'AdaptiveAvgPool1d', 'AdaptiveMaxPool1d', 'AdaptiveMaxPool2d', 'AdaptiveMaxPool3d',
|
|
34
34
|
'AdaptiveAvgPool2d', 'AdaptiveAvgPool3d', 'MaxUnpool1d', 'MaxUnpool2d', 'MaxUnpool3d', 'LPPool1d',
|
|
35
|
-
'LPPool2d', 'AvgPool2dExt', 'MaxPool2dExt', 'AvgPool1dExt']
|
|
35
|
+
'LPPool2d', 'AvgPool2dExt', 'AvgPool3dExt', 'MaxPool2dExt', 'AvgPool1dExt']
|
|
36
36
|
|
|
37
37
|
|
|
38
38
|
class _PoolNd(Cell):
|
|
@@ -299,11 +299,12 @@ class MaxPool3d(_PoolNd):
|
|
|
299
299
|
For Atlas training series products, this interface is not supported.
|
|
300
300
|
|
|
301
301
|
Args:
|
|
302
|
-
kernel_size (Union[int, tuple[int]]): The size of kernel used to take the maximum value,
|
|
302
|
+
kernel_size (Union[int, tuple[int]], optional): The size of kernel used to take the maximum value,
|
|
303
303
|
is an int number or a single element tuple that represents depth, height and width of the kernel, or a tuple
|
|
304
304
|
of three int numbers that represent depth, height and width respectively.
|
|
305
305
|
The value must be a positive integer. Default: ``1`` .
|
|
306
|
-
stride (Union[int, tuple[int]]): The moving stride of pooling operation,
|
|
306
|
+
stride (Union[int, tuple[int]], optional): The moving stride of pooling operation,
|
|
307
|
+
an int number or a single element tuple
|
|
307
308
|
that represents the moving stride of pooling kernel in the directions of depth, height and the width,
|
|
308
309
|
or a tuple of three int numbers that represent depth, height and width of movement respectively.
|
|
309
310
|
The value must be a positive integer. If the value is None, the default value `kernel_size` is used.
|
|
@@ -324,18 +325,19 @@ class MaxPool3d(_PoolNd):
|
|
|
324
325
|
in the depth, height and width dimension is determined by the `padding` parameter.
|
|
325
326
|
If this mode is set, `padding` must be greater than or equal to 0.
|
|
326
327
|
|
|
327
|
-
padding (Union(int, tuple[int], list[int])): Pooling padding value. Default: ``0`` .
|
|
328
|
+
padding (Union(int, tuple[int], list[int]), optional): Pooling padding value. Default: ``0`` .
|
|
328
329
|
`padding` can only be an integer or a tuple/list containing one or three integers.
|
|
329
330
|
If `padding` is an integer or a tuple/list containing one integer, it will be padded in six directions of
|
|
330
331
|
front, back, top, bottom, left and right of the input. If `padding` is a tuple/list containing three
|
|
331
332
|
integers, it will be padded in front and back of the input `padding[0]` times, up and down `padding[1]`
|
|
332
333
|
times, and left and right of the input `padding[2]` times.
|
|
333
|
-
dilation (Union(int, tuple[int])): The spacing between the elements of the kernel in convolution,
|
|
334
|
+
dilation (Union(int, tuple[int]), optional): The spacing between the elements of the kernel in convolution,
|
|
334
335
|
used to increase the receptive field of the pooling operation. If it is a tuple, it must contain one or
|
|
335
336
|
three integers. Default: ``1`` .
|
|
336
|
-
return_indices (bool): If ``True`` , output is a Tuple of 2 Tensors,
|
|
337
|
+
return_indices (bool, optional): If ``True`` , output is a Tuple of 2 Tensors,
|
|
338
|
+
representing the maxpool result and where
|
|
337
339
|
the max values are generated. Otherwise, only the maxpool result is returned. Default: ``False`` .
|
|
338
|
-
ceil_mode (bool): If ``True``, use ceil to calculate output shape.
|
|
340
|
+
ceil_mode (bool, optional): If ``True``, use ceil to calculate output shape.
|
|
339
341
|
If ``False``, use ceil to calculate output shape. Default: ``False`` .
|
|
340
342
|
|
|
341
343
|
Inputs:
|
|
@@ -713,9 +715,9 @@ class MaxPool1d(_PoolNd):
|
|
|
713
715
|
\text{input}(N_i, C_j, s_0 \times l + n)
|
|
714
716
|
|
|
715
717
|
Args:
|
|
716
|
-
kernel_size (int): The size of kernel used to take the max value
|
|
717
|
-
stride (int): The distance of kernel moving, an int number that represents
|
|
718
|
-
the width of movement is stride
|
|
718
|
+
kernel_size (int, optional): The size of kernel used to take the max value. Default: ``1`` .
|
|
719
|
+
stride (int, optional): The distance of kernel moving, an int number that represents
|
|
720
|
+
the width of movement is stride. Default: ``1`` .
|
|
719
721
|
pad_mode (str, optional): Specifies the padding mode with a padding value of 0. It can be set to:
|
|
720
722
|
``"same"`` , ``"valid"`` or ``"pad"`` . Default: ``"valid"`` .
|
|
721
723
|
|
|
@@ -731,24 +733,25 @@ class MaxPool1d(_PoolNd):
|
|
|
731
733
|
at the begin and end is determined by the `padding` parameter.
|
|
732
734
|
If this mode is set, `padding` must be greater than or equal to 0.
|
|
733
735
|
|
|
734
|
-
padding (Union(int, tuple[int], list[int])): Padding value for the pooling. Default value is ``0``.
|
|
736
|
+
padding (Union(int, tuple[int], list[int]), optional): Padding value for the pooling. Default value is ``0``.
|
|
735
737
|
padding can only be an integer or a tuple/list containing a single integer, in which case padding times or
|
|
736
738
|
padding[0] times are padded on both sides of the input.
|
|
737
|
-
dilation (Union(int, tuple[int])): The spacing between the elements of the kernel in convolution,
|
|
739
|
+
dilation (Union(int, tuple[int]), optional): The spacing between the elements of the kernel in convolution,
|
|
738
740
|
used to increase the receptive field of the pooling operation. If it is a tuple, its length can only be 1.
|
|
739
741
|
Default: ``1`` .
|
|
740
|
-
return_indices (bool): If ``True`` , the function will return
|
|
742
|
+
return_indices (bool, optional): If ``True`` , the function will return
|
|
743
|
+
both the result of max pooling and the indices of
|
|
741
744
|
the max elements. Default: ``False`` .
|
|
742
|
-
ceil_mode (bool): If True, use ceil to compute the output shape instead of floor. Default: ``False`` .
|
|
745
|
+
ceil_mode (bool, optional): If True, use ceil to compute the output shape instead of floor. Default: ``False`` .
|
|
743
746
|
|
|
744
747
|
Inputs:
|
|
745
748
|
- **x** (Tensor) - Tensor of shape :math:`(N, C_{in}, L_{in})` or :math:`(C_{in}, L_{in})`.
|
|
746
749
|
|
|
747
750
|
Outputs:
|
|
748
|
-
If `return_indices` is False
|
|
751
|
+
If `return_indices` is ``False``, output is a Tensor, with shape :math:`(N, C_{out}, L_{out})` or
|
|
749
752
|
:math:`(C_{out}, L_{out})`. It has the same data type as `x`.
|
|
750
753
|
|
|
751
|
-
If `return_indices` is True
|
|
754
|
+
If `return_indices` is ``True``, output is a Tuple of 2 Tensors, representing the maxpool result and where
|
|
752
755
|
the max values are generated.
|
|
753
756
|
|
|
754
757
|
- **output** (Tensor) - Maxpooling result, with shape :math:`(N, C_{out}, L_{out})` or
|
|
@@ -1021,6 +1024,47 @@ class AvgPool3d(_PoolNd):
|
|
|
1021
1024
|
return out
|
|
1022
1025
|
|
|
1023
1026
|
|
|
1027
|
+
class AvgPool3dExt(Cell):
|
|
1028
|
+
r"""
|
|
1029
|
+
Applies a 3D average pooling over an input Tensor which can be regarded as
|
|
1030
|
+
a composition of 3D input planes.
|
|
1031
|
+
|
|
1032
|
+
.. warning::
|
|
1033
|
+
This is an experimental API that is subject to change or deletion.
|
|
1034
|
+
|
|
1035
|
+
For details, please refer to :func:`mindspore.mint.nn.functional.avg_pool3d`.
|
|
1036
|
+
|
|
1037
|
+
Supported Platforms:
|
|
1038
|
+
``Ascend``
|
|
1039
|
+
|
|
1040
|
+
Examples:
|
|
1041
|
+
>>> import mindspore as ms
|
|
1042
|
+
>>> pool = ms.nn.AvgPool3dExt(kernel_size=3, stride=1)
|
|
1043
|
+
>>> x = ms.ops.randn(1, 2, 4, 4, 5).astype(ms.float32)
|
|
1044
|
+
>>> output = pool(x)
|
|
1045
|
+
>>> print(output.shape)
|
|
1046
|
+
(1, 2, 2, 2, 3)
|
|
1047
|
+
>>> x1 = ms.ops.randn(6, 5, 7, 7, 5).astype(ms.float32)
|
|
1048
|
+
>>> pool2 = ms.nn.AvgPool3dExt(4, stride=2, padding=(2, 2, 1), divisor_override=10)
|
|
1049
|
+
>>> output2 = pool2(x1)
|
|
1050
|
+
>>> print(output2.shape)
|
|
1051
|
+
(6, 5, 4, 4, 2)
|
|
1052
|
+
"""
|
|
1053
|
+
def __init__(self, kernel_size, stride=None, padding=0, ceil_mode=False,
|
|
1054
|
+
count_include_pad=True, divisor_override=None):
|
|
1055
|
+
super(AvgPool3dExt, self).__init__()
|
|
1056
|
+
self.kernel_size = kernel_size
|
|
1057
|
+
self.stride = stride
|
|
1058
|
+
self.padding = padding
|
|
1059
|
+
self.ceil_mode = ceil_mode
|
|
1060
|
+
self.count_include_pad = count_include_pad
|
|
1061
|
+
self.divisor_override = divisor_override
|
|
1062
|
+
|
|
1063
|
+
def construct(self, input):
|
|
1064
|
+
return ops.function.nn_func.avg_pool3d_ext(input, self.kernel_size, self.stride, self.padding,
|
|
1065
|
+
self.ceil_mode, self.count_include_pad, self.divisor_override)
|
|
1066
|
+
|
|
1067
|
+
|
|
1024
1068
|
class AvgPool1dExt(Cell):
|
|
1025
1069
|
r"""
|
|
1026
1070
|
Applies a 1D average pooling over an input Tensor which can be regarded as
|
|
@@ -1270,8 +1314,8 @@ class AvgPool1d(_PoolNd):
|
|
|
1270
1314
|
This interface currently does not support Atlas A2 training series products.
|
|
1271
1315
|
|
|
1272
1316
|
Args:
|
|
1273
|
-
kernel_size (int): The size of kernel window used to take the average value, Default: ``1`` .
|
|
1274
|
-
stride (int): The distance of kernel moving, an int number that represents
|
|
1317
|
+
kernel_size (int, optional): The size of kernel window used to take the average value, Default: ``1`` .
|
|
1318
|
+
stride (int, optional): The distance of kernel moving, an int number that represents
|
|
1275
1319
|
the width of movement is strides, Default: ``1`` .
|
|
1276
1320
|
pad_mode (str, optional): Specifies the padding mode with a padding value of 0. It can be set to:
|
|
1277
1321
|
``"same"`` , ``"valid"`` or ``"pad"`` . Default: ``"valid"`` .
|
|
@@ -1282,17 +1326,20 @@ class AvgPool1d(_PoolNd):
|
|
|
1282
1326
|
uniformly distributed around the input, if it is odd, the excess padding is goes to the right side.
|
|
1283
1327
|
If this mode is set, `padding` must be 0.
|
|
1284
1328
|
- ``"valid"``: No padding is applied to the input, and the output returns the maximum
|
|
1285
|
-
possible length.
|
|
1286
|
-
|
|
1329
|
+
possible length. If a full stride cannot be formed, the extra pixels will be discarded.
|
|
1330
|
+
If this mode is set, `padding` must be 0.
|
|
1287
1331
|
- ``"pad"``: Pad the input with a specified amount. In this mode, the amount of padding
|
|
1288
1332
|
at the begin and end is determined by the `padding` parameter.
|
|
1289
1333
|
If this mode is set, `padding` must be greater than or equal to 0.
|
|
1290
1334
|
|
|
1291
|
-
padding (Union(int, tuple[int], list[int])): Pooling padding value,
|
|
1335
|
+
padding (Union(int, tuple[int], list[int]), optional): Pooling padding value,
|
|
1336
|
+
only ``"pad"`` mode can be set to non-zero.
|
|
1292
1337
|
Default: ``0`` . padding can only be an integer or a tuple/list containing a single integer, in which case
|
|
1293
1338
|
padding times or padding[0] times are padded on both sides of the input.
|
|
1294
|
-
ceil_mode (bool): If ``True`` , use ceil to compute the output shape instead of floor.
|
|
1295
|
-
|
|
1339
|
+
ceil_mode (bool, optional): If ``True`` , use ceil to compute the output shape instead of floor.
|
|
1340
|
+
Default: ``False`` .
|
|
1341
|
+
count_include_pad (bool, optional): If ``True`` , averaging calculation will include the zero-padding.
|
|
1342
|
+
Default: ``True`` .
|
|
1296
1343
|
|
|
1297
1344
|
Inputs:
|
|
1298
1345
|
- **x** (Tensor) - Tensor of shape :math:`(N, C_{in}, L_{in})` or :math:`(C_{in}, L_{in})`.
|
|
@@ -1728,13 +1775,14 @@ class AdaptiveMaxPool2d(Cell):
|
|
|
1728
1775
|
\end{align}
|
|
1729
1776
|
|
|
1730
1777
|
Note:
|
|
1731
|
-
|
|
1778
|
+
In KBK mode, `output_size` does not support mutable.
|
|
1732
1779
|
|
|
1733
1780
|
Args:
|
|
1734
1781
|
output_size (Union[int, tuple]): The target output size. `output_size` can be a tuple :math:`(H, W)`,
|
|
1735
1782
|
or an int H for :math:`(H, H)`. :math:`H` and :math:`W` can be int or None.
|
|
1736
1783
|
If it is None, it means the output size is the same as the input size.
|
|
1737
|
-
return_indices (bool):
|
|
1784
|
+
return_indices (bool, optional): Whether to output the index of the maximum value.
|
|
1785
|
+
If `return_indices` is ``True`` , the indices of max value would be output.
|
|
1738
1786
|
Default: ``False`` .
|
|
1739
1787
|
|
|
1740
1788
|
Inputs:
|
|
@@ -1797,15 +1845,11 @@ class AdaptiveMaxPool2d(Cell):
|
|
|
1797
1845
|
def __init__(self, output_size, return_indices=False):
|
|
1798
1846
|
"""Initialize AdaptiveMaxPool2d."""
|
|
1799
1847
|
super(AdaptiveMaxPool2d, self).__init__()
|
|
1800
|
-
|
|
1801
|
-
self.adaptive_max_pool2d = ops.AdaptiveMaxPool2D(output_size)
|
|
1848
|
+
self.output_size = output_size
|
|
1802
1849
|
self.return_indices = return_indices
|
|
1803
1850
|
|
|
1804
1851
|
def construct(self, input):
|
|
1805
|
-
|
|
1806
|
-
if self.return_indices:
|
|
1807
|
-
return output
|
|
1808
|
-
return output[0]
|
|
1852
|
+
return ops.adaptive_max_pool2d(input, self.output_size, self.return_indices)
|
|
1809
1853
|
|
|
1810
1854
|
|
|
1811
1855
|
class AdaptiveMaxPool3d(Cell):
|
|
@@ -54,16 +54,16 @@ class MultiheadAttention(Cell):
|
|
|
54
54
|
embed_dim (int): Total dimension of MultiheadAttention.
|
|
55
55
|
num_heads (int): Number of attention heads. Note that `embed_dim` will be split
|
|
56
56
|
across `num_heads` (i.e. each head will have dimension `embed_dim // num_heads`).
|
|
57
|
-
dropout (float): Dropout probability of `attn_output_weights`. Default: ``0.0``.
|
|
58
|
-
has_bias (bool): Whether adds bias to input / output projection layers. Default: ``True``.
|
|
59
|
-
add_bias_kv (bool): Whether adds bias to the key and value sequences at axis=0. Default: ``False``.
|
|
60
|
-
add_zero_attn (bool): Whether adds a new batch of zeros to the key and value sequences at axis=1.
|
|
57
|
+
dropout (float, optional): Dropout probability of `attn_output_weights`. Default: ``0.0``.
|
|
58
|
+
has_bias (bool, optional): Whether adds bias to input / output projection layers. Default: ``True``.
|
|
59
|
+
add_bias_kv (bool, optional): Whether adds bias to the key and value sequences at axis=0. Default: ``False``.
|
|
60
|
+
add_zero_attn (bool, optional): Whether adds a new batch of zeros to the key and value sequences at axis=1.
|
|
61
61
|
Default: ``False``.
|
|
62
|
-
kdim (int): Total number of features for keys. Default: ``None`` (`kdim=embed_dim`).
|
|
63
|
-
vdim (int): Total number of features for values. Default: ``None`` (`vdim=embed_dim`).
|
|
64
|
-
batch_first (bool): If ``True``, then the input and output shape are :math:`(batch, seq, feature)` ,
|
|
62
|
+
kdim (int, optional): Total number of features for keys. Default: ``None`` (`kdim=embed_dim`).
|
|
63
|
+
vdim (int, optional): Total number of features for values. Default: ``None`` (`vdim=embed_dim`).
|
|
64
|
+
batch_first (bool, optional): If ``True``, then the input and output shape are :math:`(batch, seq, feature)` ,
|
|
65
65
|
else :math:`(seq, batch, feature)` . Default: ``False``.
|
|
66
|
-
dtype (:class:`mindspore.dtype
|
|
66
|
+
dtype (:class:`mindspore.dtype`, optional): Data type of Parameter. Default: ``mstype.float32`` .
|
|
67
67
|
|
|
68
68
|
Inputs:
|
|
69
69
|
- **query** (Tensor) - The query embeddings. If `query` is unbatched, the shape is :math:`(L, E_q)`,
|
|
@@ -85,7 +85,7 @@ class MultiheadAttention(Cell):
|
|
|
85
85
|
For a binary mask, a ``True`` value indicates that the corresponding `key` value will be ignored for
|
|
86
86
|
the purpose of attention. For a float mask, it will be directly added to the corresponding `key` value.
|
|
87
87
|
Supported float types: float16, float32, float64. Default: ``None``.
|
|
88
|
-
- **need_weights** (bool) - Whether returns `attn_output_weights` in addition to `attn_outputs`.
|
|
88
|
+
- **need_weights** (bool, optional) - Whether returns `attn_output_weights` in addition to `attn_outputs`.
|
|
89
89
|
Default: ``True``.
|
|
90
90
|
- **attn_mask** (Tensor, optional) - If specified, a 2D or 3D mask preventing attention to certain positions.
|
|
91
91
|
Must be of shape :math:`(L, S)` or :math:`(N\cdot\text{num_heads}, L, S)`, where :math:`N` is the
|
|
@@ -94,7 +94,8 @@ class MultiheadAttention(Cell):
|
|
|
94
94
|
in the batch. For a binary mask, a ``True`` value indicates that the corresponding position is not allowed
|
|
95
95
|
to attend. For a float mask, the mask values will be added to the attention weight.
|
|
96
96
|
Supported float types: float16, float32, float64. Default: ``None``.
|
|
97
|
-
- **average_attn_weights** (bool) - If true, indicates that
|
|
97
|
+
- **average_attn_weights** (bool, optional) - If true, indicates that
|
|
98
|
+
the returned `attn_weights` should be averaged
|
|
98
99
|
across heads. Otherwise, `attn_weights` are provided separately per head. Note that this flag only
|
|
99
100
|
has an effect when `need_weights=True`. Default: ``True`` (i.e. average weights across heads)
|
|
100
101
|
|
|
@@ -80,7 +80,8 @@ class ExponentialDecayLR(LearningRateSchedule):
|
|
|
80
80
|
learning_rate (float): The initial value of learning rate.
|
|
81
81
|
decay_rate (float): The decay rate.
|
|
82
82
|
decay_steps (int): Number of steps to decay over.
|
|
83
|
-
is_stair (bool): If
|
|
83
|
+
is_stair (bool, optional): If ``True``, learning rate is decayed once every `decay_steps` time.
|
|
84
|
+
Default: ``False`` .
|
|
84
85
|
|
|
85
86
|
Inputs:
|
|
86
87
|
- **global_step** (Tensor) - The current step number. :math:`current\_step` in the above formula.
|
|
@@ -223,7 +224,8 @@ class InverseDecayLR(LearningRateSchedule):
|
|
|
223
224
|
learning_rate (float): The initial value of learning rate.
|
|
224
225
|
decay_rate (float): The decay rate.
|
|
225
226
|
decay_steps (int): Number of steps to decay over.
|
|
226
|
-
is_stair (bool): If true, learning rate decay once every `decay_steps` times.
|
|
227
|
+
is_stair (bool, optional): If true, learning rate decay once every `decay_steps` times.
|
|
228
|
+
If False, the learning rate
|
|
227
229
|
decays for every step. Default: ``False`` .
|
|
228
230
|
|
|
229
231
|
Inputs:
|
mindspore/nn/loss/loss.py
CHANGED
|
@@ -127,7 +127,8 @@ class LossBase(Cell):
|
|
|
127
127
|
Args:
|
|
128
128
|
x (Tensor): Tensor of shape :math:`(N, *)` where :math:`*` means, any number of
|
|
129
129
|
additional dimensions.
|
|
130
|
-
weights (Union[float, Tensor]):
|
|
130
|
+
weights (Union[float, Tensor], optional): Weights. When `weights` is a Tensor,
|
|
131
|
+
the rank is either 0, or the same rank as inputs,
|
|
131
132
|
and must be broadcastable to inputs (i.e., all dimensions must be either `1`,
|
|
132
133
|
or the same as the corresponding inputs dimension). Default: ``1.0`` .
|
|
133
134
|
|
|
@@ -617,7 +618,8 @@ class MarginRankingLoss(LossBase):
|
|
|
617
618
|
|
|
618
619
|
class SmoothL1Loss(LossBase):
|
|
619
620
|
r"""
|
|
620
|
-
SmoothL1 loss function
|
|
621
|
+
SmoothL1 loss function. Compare the error value element-wise and
|
|
622
|
+
if the absolute error between the predicted value and the target value
|
|
621
623
|
is less than the set threshold `beta`, the square term is used, otherwise the absolute error term is used.
|
|
622
624
|
|
|
623
625
|
Given two input :math:`x,\ y`, the SmoothL1Loss can be described as follows:
|
|
@@ -667,11 +669,13 @@ class SmoothL1Loss(LossBase):
|
|
|
667
669
|
|
|
668
670
|
- Ascend: float16, float32, bfloat16.
|
|
669
671
|
- CPU/GPU: float16, float32, float64.
|
|
672
|
+
|
|
670
673
|
- **labels** (Tensor) - Ground truth data.
|
|
671
674
|
|
|
672
675
|
- CPU/Ascend: has the same shape as the `logits`,
|
|
673
676
|
`logits` and `labels` comply with the implicit type conversion rules to make the data types consistent.
|
|
674
677
|
- GPU: has the same shape and dtype as the `logits`.
|
|
678
|
+
|
|
675
679
|
Outputs:
|
|
676
680
|
Tensor, if `reduction` is ``'none'``, then output is a tensor with the same shape as `logits`.
|
|
677
681
|
Otherwise the shape of output tensor is :math:`()`.
|
|
@@ -732,16 +736,19 @@ class SoftMarginLoss(LossBase):
|
|
|
732
736
|
- ``'sum'``: the output elements will be summed.
|
|
733
737
|
|
|
734
738
|
Inputs:
|
|
735
|
-
- **logits** (Tensor) - Predict data. Data type must be float16
|
|
736
|
-
|
|
739
|
+
- **logits** (Tensor) - Predict data. Data type must be float16, float32,
|
|
740
|
+
bfloat16 (Among them, the Atlas training series products do not support bfloat16).
|
|
741
|
+
- **labels** (Tensor) - Ground truth data, with the same shape as `logits`.
|
|
742
|
+
In GE mode, the data type should be the same as `logits`.
|
|
737
743
|
|
|
738
744
|
Outputs:
|
|
739
|
-
Tensor or Scalar, if `reduction` is ``
|
|
745
|
+
Tensor or Scalar, if `reduction` is ``'none'``, its shape is the same as `logits`.
|
|
740
746
|
Otherwise, a scalar value will be returned.
|
|
741
747
|
|
|
742
748
|
Raises:
|
|
743
749
|
TypeError: If `logits` or `labels` is not a Tensor.
|
|
744
|
-
TypeError: If dtype of `logits` or `labels` is
|
|
750
|
+
TypeError: If dtype of `logits` or `labels` is not float16, float32,
|
|
751
|
+
bfloat16 (Among them, the Atlas training series products do not support bfloat16).
|
|
745
752
|
ValueError: If shape of `logits` is not the same as `labels`.
|
|
746
753
|
ValueError: If `reduction` is not one of ``'none'``, ``'mean'``, ``'sum'``.
|
|
747
754
|
|
|
@@ -762,10 +769,10 @@ class SoftMarginLoss(LossBase):
|
|
|
762
769
|
|
|
763
770
|
def __init__(self, reduction='mean'):
|
|
764
771
|
super(SoftMarginLoss, self).__init__()
|
|
765
|
-
self.
|
|
772
|
+
self.reduction = reduction
|
|
766
773
|
|
|
767
774
|
def construct(self, logits, labels):
|
|
768
|
-
return
|
|
775
|
+
return F.soft_margin_loss(logits, labels, self.reduction)
|
|
769
776
|
|
|
770
777
|
|
|
771
778
|
class SoftmaxCrossEntropyWithLogits(LossBase):
|
|
@@ -813,8 +820,8 @@ class SoftmaxCrossEntropyWithLogits(LossBase):
|
|
|
813
820
|
|
|
814
821
|
Raises:
|
|
815
822
|
TypeError: If `sparse` is not a bool.
|
|
816
|
-
TypeError: If `sparse` is True and dtype of `labels` is neither int32 nor int64.
|
|
817
|
-
TypeError: If `sparse` is False and dtype of `labels` is neither float16 not float32.
|
|
823
|
+
TypeError: If `sparse` is ``True`` and dtype of `labels` is neither int32 nor int64.
|
|
824
|
+
TypeError: If `sparse` is ``False`` and dtype of `labels` is neither float16 not float32.
|
|
818
825
|
ValueError: If `reduction` is not one of ``'none'``, ``'mean'``, ``'sum'``.
|
|
819
826
|
|
|
820
827
|
Supported Platforms:
|
|
@@ -893,8 +900,8 @@ class DiceLoss(LossBase):
|
|
|
893
900
|
:math:`pred` represent `logits`, :math:`true` represent `labels` .
|
|
894
901
|
|
|
895
902
|
Args:
|
|
896
|
-
smooth (float): A term added to the denominator to improve numerical stability.
|
|
897
|
-
|
|
903
|
+
smooth (float, optional): A term added to the denominator to improve numerical stability.
|
|
904
|
+
Should be greater than 0. Default: ``1e-5`` .
|
|
898
905
|
|
|
899
906
|
Inputs:
|
|
900
907
|
- **logits** (Tensor) - Input predicted value. The data type must be float16 or float32.
|
|
@@ -938,11 +945,12 @@ class DiceLoss(LossBase):
|
|
|
938
945
|
if label.dtype == mstype.uint8:
|
|
939
946
|
raise TypeError(f"For '{self.cls_name}', the dtype of 'labels' can not be uint8.")
|
|
940
947
|
intersection = self.reduce_sum(self.mul(logits.view(-1), label.view(-1)))
|
|
941
|
-
|
|
942
|
-
|
|
948
|
+
unionset_part1 = self.reduce_sum(self.mul(logits.view(-1), logits.view(-1)))
|
|
949
|
+
unionset_part2 = self.reduce_sum(self.mul(label.view(-1), label.view(-1)))
|
|
950
|
+
unionset = ops.add(unionset_part1, unionset_part2)
|
|
943
951
|
|
|
944
|
-
single_dice_coeff = (2 * intersection) / (unionset
|
|
945
|
-
dice_loss = 1
|
|
952
|
+
single_dice_coeff = (2 * intersection) / ops.add(unionset, self.smooth)
|
|
953
|
+
dice_loss = ops.sub(1, single_dice_coeff)
|
|
946
954
|
|
|
947
955
|
return dice_loss
|
|
948
956
|
|
|
@@ -1058,7 +1066,7 @@ class MultiClassDiceLoss(LossBase):
|
|
|
1058
1066
|
dice_loss = self.binarydiceloss(logits[:, i], label[:, i])
|
|
1059
1067
|
if self.weights is not None:
|
|
1060
1068
|
_check_weights(self.weights.shape[0], label.shape[1], self.cls_name)
|
|
1061
|
-
dice_loss
|
|
1069
|
+
dice_loss = dice_loss * self.weights[i]
|
|
1062
1070
|
total_loss += dice_loss
|
|
1063
1071
|
|
|
1064
1072
|
return total_loss / label.shape[1]
|
|
@@ -2571,7 +2579,7 @@ class KLDivLoss(LossBase):
|
|
|
2571
2579
|
the updating formulas of KLDivLoss algorithm are as follows,
|
|
2572
2580
|
|
|
2573
2581
|
.. math::
|
|
2574
|
-
L(x, target) = target \cdot (\log target - x)
|
|
2582
|
+
L(x, target) = target \cdot (\log target - \log x)
|
|
2575
2583
|
|
|
2576
2584
|
Then,
|
|
2577
2585
|
|
|
@@ -2865,7 +2873,7 @@ class HingeEmbeddingLoss(LossBase):
|
|
|
2865
2873
|
where :math:`L = \{l_1,\dots,l_N\}^\top`.
|
|
2866
2874
|
|
|
2867
2875
|
Args:
|
|
2868
|
-
margin (float, int): Threshold defined by Hinge Embedding Loss :math:`margin`.
|
|
2876
|
+
margin (float, int, optional): Threshold defined by Hinge Embedding Loss :math:`margin`.
|
|
2869
2877
|
Represented as :math:`\Delta` in the formula. Default: ``1.0`` .
|
|
2870
2878
|
reduction (str, optional): Apply specific reduction method to the output: ``'none'`` , ``'mean'`` ,
|
|
2871
2879
|
``'sum'`` . Default: ``'mean'`` .
|
mindspore/nn/optim/ada_grad.py
CHANGED
|
@@ -113,8 +113,8 @@ class Adagrad(Optimizer):
|
|
|
113
113
|
If `order_params` in the keys, other keys will be ignored and the element of 'order_params' must be in
|
|
114
114
|
one group of `params`.
|
|
115
115
|
|
|
116
|
-
accum (float): The starting value for :math:`h`, must be zero or positive values. Default: ``0.1`` .
|
|
117
|
-
learning_rate (Union[float, int, Tensor, Iterable, LearningRateSchedule]): Default: ``0.001`` .
|
|
116
|
+
accum (float, optional): The starting value for :math:`h`, must be zero or positive values. Default: ``0.1`` .
|
|
117
|
+
learning_rate (Union[float, int, Tensor, Iterable, LearningRateSchedule], optional): Default: ``0.001`` .
|
|
118
118
|
|
|
119
119
|
- float: The fixed learning rate value. Must be equal to or greater than 0.
|
|
120
120
|
|
|
@@ -130,13 +130,14 @@ class Adagrad(Optimizer):
|
|
|
130
130
|
<https://www.mindspore.cn/docs/en/master/api_python/mindspore.nn.html#learningrateschedule-class>`_
|
|
131
131
|
with step as the input to get the learning rate of current step.
|
|
132
132
|
|
|
133
|
-
update_slots (bool): Whether the :math:`h` will be updated. Default: ``True`` .
|
|
134
|
-
loss_scale (float): Value for the loss scale. It must be greater than 0.0. In general,
|
|
133
|
+
update_slots (bool, optional): Whether the :math:`h` will be updated. Default: ``True`` .
|
|
134
|
+
loss_scale (float, optional): Value for the loss scale. It must be greater than 0.0. In general,
|
|
135
|
+
use the default value.
|
|
135
136
|
Only when `FixedLossScaleManager` is used for training and the `drop_overflow_update` in
|
|
136
137
|
`FixedLossScaleManager` is set to False, then this value needs to be the same as the `loss_scale` in
|
|
137
138
|
`FixedLossScaleManager`. Refer to class :class:`mindspore.amp.FixedLossScaleManager` for more details.
|
|
138
139
|
Default: ``1.0`` .
|
|
139
|
-
weight_decay (Union[float, int, Cell]): Weight decay (L2 penalty). Default: ``0.0`` .
|
|
140
|
+
weight_decay (Union[float, int, Cell], optional): Weight decay (L2 penalty). Default: ``0.0`` .
|
|
140
141
|
|
|
141
142
|
- float: The fixed weight decay value. Must be equal to or greater than 0.
|
|
142
143
|
|