mindspore 2.2.14__cp39-cp39-win_amd64.whl → 2.4.0__cp39-cp39-win_amd64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of mindspore might be problematic. Click here for more details.
- mindspore/.commit_id +1 -1
- mindspore/Microsoft.VisualStudio.Telemetry.dll +0 -0
- mindspore/Newtonsoft.Json.dll +0 -0
- mindspore/__init__.py +8 -5
- mindspore/_c_dataengine.cp39-win_amd64.pyd +0 -0
- mindspore/_c_expression.cp39-win_amd64.pyd +0 -0
- mindspore/_c_mindrecord.cp39-win_amd64.pyd +0 -0
- mindspore/_checkparam.py +124 -25
- mindspore/_extends/builtin_operations.py +2 -1
- mindspore/_extends/graph_kernel/model/graph_parallel.py +16 -6
- mindspore/_extends/parallel_compile/akg_compiler/akg_process.py +3 -16
- mindspore/_extends/parallel_compile/akg_compiler/build_tbe_kernel.py +16 -4
- mindspore/_extends/parallel_compile/akg_compiler/compiler.py +1 -0
- mindspore/_extends/parallel_compile/akg_compiler/gen_custom_op_files.py +96 -0
- mindspore/_extends/parallel_compile/akg_compiler/tbe_topi.py +2 -1
- mindspore/_extends/parallel_compile/akg_compiler/util.py +5 -2
- mindspore/_extends/parse/__init__.py +18 -14
- mindspore/_extends/parse/compile_config.py +299 -0
- mindspore/_extends/parse/namespace.py +2 -2
- mindspore/_extends/parse/parser.py +182 -68
- mindspore/_extends/parse/resources.py +45 -14
- mindspore/_extends/parse/standard_method.py +192 -252
- mindspore/{ops/_op_impl/tbe/atomic_addr_clean.py → _extends/pijit/__init__.py} +6 -16
- mindspore/_extends/pijit/pijit_func_white_list.py +669 -0
- mindspore/_extends/remote/kernel_build_server.py +2 -0
- mindspore/_profiler.py +30 -0
- mindspore/amp.py +67 -26
- mindspore/atlprov.dll +0 -0
- mindspore/avcodec-59.dll +0 -0
- mindspore/avdevice-59.dll +0 -0
- mindspore/avfilter-8.dll +0 -0
- mindspore/avformat-59.dll +0 -0
- mindspore/avutil-57.dll +0 -0
- mindspore/boost/adasum.py +1 -1
- mindspore/boost/base.py +1 -1
- mindspore/boost/boost_cell_wrapper.py +2 -2
- mindspore/boost/grad_freeze.py +2 -2
- mindspore/boost/group_loss_scale_manager.py +1 -1
- mindspore/boost/less_batch_normalization.py +9 -6
- mindspore/c1.dll +0 -0
- mindspore/c1xx.dll +0 -0
- mindspore/c2.dll +0 -0
- mindspore/common/__init__.py +20 -7
- mindspore/common/_jit_fallback_utils.py +2 -3
- mindspore/common/_pijit_context.py +190 -0
- mindspore/common/_register_for_adapter.py +7 -0
- mindspore/common/_register_for_recompute.py +48 -0
- mindspore/common/_register_for_tensor.py +10 -10
- mindspore/common/_stub_tensor.py +7 -1
- mindspore/common/_tensor_overload.py +139 -0
- mindspore/common/_utils.py +5 -17
- mindspore/common/api.py +449 -129
- mindspore/common/auto_dynamic_shape.py +27 -14
- mindspore/common/dtype.py +17 -10
- mindspore/common/dump.py +8 -11
- mindspore/common/file_system.py +48 -0
- mindspore/common/generator.py +254 -0
- mindspore/common/hook_handle.py +65 -30
- mindspore/common/initializer.py +1 -1
- mindspore/common/jit_config.py +34 -14
- mindspore/common/lazy_inline.py +72 -19
- mindspore/common/mindir_util.py +12 -2
- mindspore/common/mutable.py +79 -14
- mindspore/common/no_inline.py +54 -0
- mindspore/common/np_dtype.py +25 -0
- mindspore/common/parameter.py +73 -21
- mindspore/common/recompute.py +292 -0
- mindspore/common/seed.py +9 -9
- mindspore/common/sparse_tensor.py +276 -24
- mindspore/common/symbol.py +122 -0
- mindspore/common/tensor.py +668 -514
- mindspore/communication/__init__.py +6 -11
- mindspore/communication/_comm_helper.py +43 -3
- mindspore/communication/comm_func.py +1395 -0
- mindspore/communication/management.py +117 -104
- mindspore/config/op_info.config +22 -54
- mindspore/context.py +455 -71
- mindspore/dataset/__init__.py +5 -5
- mindspore/dataset/audio/__init__.py +6 -6
- mindspore/dataset/audio/transforms.py +711 -158
- mindspore/dataset/callback/ds_callback.py +2 -2
- mindspore/dataset/core/config.py +7 -0
- mindspore/dataset/core/validator_helpers.py +7 -0
- mindspore/dataset/engine/cache_client.py +2 -2
- mindspore/dataset/engine/datasets.py +201 -116
- mindspore/dataset/engine/datasets_audio.py +14 -14
- mindspore/dataset/engine/datasets_standard_format.py +83 -3
- mindspore/dataset/engine/datasets_text.py +39 -39
- mindspore/dataset/engine/datasets_user_defined.py +230 -141
- mindspore/dataset/engine/datasets_vision.py +78 -74
- mindspore/dataset/engine/iterators.py +29 -0
- mindspore/dataset/engine/obs/util.py +7 -0
- mindspore/dataset/engine/offload.py +5 -7
- mindspore/dataset/engine/queue.py +138 -66
- mindspore/dataset/engine/serializer_deserializer.py +2 -2
- mindspore/dataset/engine/validators.py +41 -15
- mindspore/dataset/text/__init__.py +2 -5
- mindspore/dataset/text/transforms.py +408 -121
- mindspore/dataset/text/utils.py +9 -9
- mindspore/dataset/transforms/__init__.py +0 -3
- mindspore/dataset/transforms/transforms.py +261 -76
- mindspore/dataset/utils/browse_dataset.py +9 -9
- mindspore/dataset/utils/line_reader.py +2 -0
- mindspore/dataset/vision/__init__.py +7 -10
- mindspore/dataset/vision/c_transforms.py +10 -10
- mindspore/dataset/vision/py_transforms_util.py +1 -1
- mindspore/dataset/vision/transforms.py +2844 -549
- mindspore/dataset/vision/utils.py +161 -10
- mindspore/dataset/vision/validators.py +16 -3
- mindspore/dnnl.dll +0 -0
- mindspore/dpcmi.dll +0 -0
- mindspore/{rewrite/ast_creator_register.py → experimental/es/__init__.py} +5 -20
- mindspore/experimental/es/embedding_service.py +883 -0
- mindspore/experimental/es/embedding_service_layer.py +581 -0
- mindspore/experimental/llm_boost/__init__.py +21 -0
- mindspore/experimental/llm_boost/atb/__init__.py +23 -0
- mindspore/experimental/llm_boost/atb/boost_base.py +211 -0
- mindspore/experimental/llm_boost/atb/llama_boost.py +115 -0
- mindspore/experimental/llm_boost/atb/qwen_boost.py +101 -0
- mindspore/experimental/llm_boost/register.py +129 -0
- mindspore/experimental/llm_boost/utils.py +31 -0
- mindspore/experimental/optim/__init__.py +12 -2
- mindspore/experimental/optim/adadelta.py +161 -0
- mindspore/experimental/optim/adagrad.py +168 -0
- mindspore/experimental/optim/adam.py +35 -34
- mindspore/experimental/optim/adamax.py +170 -0
- mindspore/experimental/optim/adamw.py +124 -15
- mindspore/experimental/optim/asgd.py +153 -0
- mindspore/experimental/optim/lr_scheduler.py +66 -121
- mindspore/experimental/optim/nadam.py +157 -0
- mindspore/experimental/optim/optimizer.py +18 -8
- mindspore/experimental/optim/radam.py +194 -0
- mindspore/experimental/optim/rmsprop.py +154 -0
- mindspore/experimental/optim/rprop.py +164 -0
- mindspore/experimental/optim/sgd.py +28 -19
- mindspore/hal/__init__.py +40 -0
- mindspore/hal/_ascend.py +57 -0
- mindspore/hal/_base.py +57 -0
- mindspore/hal/_cpu.py +56 -0
- mindspore/hal/_gpu.py +57 -0
- mindspore/hal/contiguous_tensors_handle.py +175 -0
- mindspore/hal/device.py +356 -0
- mindspore/hal/event.py +179 -0
- mindspore/hal/memory.py +326 -0
- mindspore/hal/stream.py +357 -0
- mindspore/include/api/data_type.h +2 -2
- mindspore/include/api/dual_abi_helper.h +16 -3
- mindspore/include/api/model.h +4 -3
- mindspore/include/api/model_group.h +13 -1
- mindspore/include/api/status.h +14 -0
- mindspore/include/api/types.h +10 -10
- mindspore/include/c_api/model_c.h +173 -0
- mindspore/include/c_api/types_c.h +19 -0
- mindspore/include/dataset/config.h +2 -2
- mindspore/include/dataset/constants.h +2 -2
- mindspore/include/dataset/execute.h +3 -5
- mindspore/include/dataset/vision.h +58 -2
- mindspore/jpeg62.dll +0 -0
- mindspore/log.py +3 -3
- mindspore/mindrecord/__init__.py +5 -1
- mindspore/mindrecord/config.py +809 -0
- mindspore/mindrecord/filereader.py +25 -0
- mindspore/mindrecord/filewriter.py +138 -103
- mindspore/mindrecord/mindpage.py +40 -6
- mindspore/mindrecord/shardutils.py +3 -2
- mindspore/mindrecord/shardwriter.py +7 -0
- mindspore/mindrecord/tools/cifar100_to_mr.py +8 -13
- mindspore/mindrecord/tools/cifar10_to_mr.py +9 -15
- mindspore/mindrecord/tools/csv_to_mr.py +4 -9
- mindspore/mindrecord/tools/imagenet_to_mr.py +3 -8
- mindspore/mindrecord/tools/mnist_to_mr.py +7 -12
- mindspore/mindrecord/tools/tfrecord_to_mr.py +1 -6
- mindspore/mindspore_backend.dll +0 -0
- mindspore/mindspore_common.dll +0 -0
- mindspore/mindspore_core.dll +0 -0
- mindspore/mindspore_glog.dll +0 -0
- mindspore/mindspore_np_dtype.dll +0 -0
- mindspore/mindspore_ops.dll +0 -0
- mindspore/mint/__init__.py +1586 -0
- mindspore/mint/distributed/__init__.py +31 -0
- mindspore/mint/distributed/distributed.py +254 -0
- mindspore/{rewrite/ast_transformers → mint/linalg}/__init__.py +9 -4
- mindspore/mint/nn/__init__.py +757 -0
- mindspore/mint/nn/functional.py +679 -0
- mindspore/mint/nn/layer/__init__.py +39 -0
- mindspore/mint/nn/layer/activation.py +133 -0
- mindspore/mint/nn/layer/normalization.py +477 -0
- mindspore/mint/nn/layer/pooling.py +110 -0
- mindspore/mint/optim/__init__.py +24 -0
- mindspore/mint/optim/adamw.py +206 -0
- mindspore/mint/special/__init__.py +63 -0
- mindspore/msobj140.dll +0 -0
- mindspore/mspdb140.dll +0 -0
- mindspore/mspdbcore.dll +0 -0
- mindspore/mspdbst.dll +0 -0
- mindspore/mspft140.dll +0 -0
- mindspore/msvcdis140.dll +0 -0
- mindspore/msvcp140_1.dll +0 -0
- mindspore/msvcp140_2.dll +0 -0
- mindspore/msvcp140_atomic_wait.dll +0 -0
- mindspore/msvcp140_codecvt_ids.dll +0 -0
- mindspore/multiprocessing/__init__.py +73 -0
- mindspore/nn/cell.py +461 -323
- mindspore/nn/dynamic_lr.py +2 -2
- mindspore/nn/layer/activation.py +292 -135
- mindspore/nn/layer/basic.py +288 -83
- mindspore/nn/layer/channel_shuffle.py +3 -16
- mindspore/nn/layer/container.py +3 -3
- mindspore/nn/layer/conv.py +75 -66
- mindspore/nn/layer/embedding.py +221 -45
- mindspore/nn/layer/image.py +4 -7
- mindspore/nn/layer/math.py +1 -1
- mindspore/nn/layer/normalization.py +150 -68
- mindspore/nn/layer/padding.py +64 -87
- mindspore/nn/layer/pooling.py +175 -12
- mindspore/nn/layer/rnn_cells.py +6 -16
- mindspore/nn/layer/rnns.py +6 -5
- mindspore/nn/layer/thor_layer.py +1 -2
- mindspore/nn/layer/timedistributed.py +1 -1
- mindspore/nn/layer/transformer.py +55 -53
- mindspore/nn/learning_rate_schedule.py +6 -5
- mindspore/nn/loss/__init__.py +2 -2
- mindspore/nn/loss/loss.py +145 -88
- mindspore/nn/optim/__init__.py +2 -1
- mindspore/nn/optim/ada_grad.py +4 -2
- mindspore/nn/optim/adadelta.py +4 -2
- mindspore/nn/optim/adafactor.py +1 -1
- mindspore/nn/optim/adam.py +102 -181
- mindspore/nn/optim/adamax.py +4 -2
- mindspore/nn/optim/adasum.py +3 -3
- mindspore/nn/optim/asgd.py +4 -2
- mindspore/nn/optim/ftrl.py +31 -61
- mindspore/nn/optim/lamb.py +5 -3
- mindspore/nn/optim/lars.py +2 -2
- mindspore/nn/optim/lazyadam.py +6 -4
- mindspore/nn/optim/momentum.py +13 -25
- mindspore/nn/optim/optimizer.py +6 -3
- mindspore/nn/optim/proximal_ada_grad.py +4 -2
- mindspore/nn/optim/rmsprop.py +9 -3
- mindspore/nn/optim/rprop.py +4 -2
- mindspore/nn/optim/sgd.py +5 -3
- mindspore/nn/optim/tft_wrapper.py +127 -0
- mindspore/nn/optim/thor.py +2 -2
- mindspore/nn/probability/distribution/_utils/custom_ops.py +2 -2
- mindspore/nn/probability/distribution/beta.py +2 -2
- mindspore/nn/probability/distribution/categorical.py +4 -6
- mindspore/nn/probability/distribution/cauchy.py +2 -2
- mindspore/nn/probability/distribution/exponential.py +2 -2
- mindspore/nn/probability/distribution/geometric.py +1 -1
- mindspore/nn/probability/distribution/gumbel.py +2 -2
- mindspore/nn/probability/distribution/logistic.py +1 -1
- mindspore/nn/probability/distribution/poisson.py +2 -2
- mindspore/nn/probability/distribution/uniform.py +2 -2
- mindspore/nn/reinforcement/_tensors_queue.py +13 -1
- mindspore/nn/wrap/__init__.py +2 -1
- mindspore/nn/wrap/cell_wrapper.py +46 -12
- mindspore/nn/wrap/grad_reducer.py +148 -8
- mindspore/nn/wrap/loss_scale.py +44 -7
- mindspore/numpy/__init__.py +2 -0
- mindspore/numpy/array_creations.py +67 -68
- mindspore/numpy/array_ops.py +70 -66
- mindspore/numpy/dtypes.py +3 -3
- mindspore/numpy/fft.py +966 -0
- mindspore/numpy/logic_ops.py +11 -10
- mindspore/numpy/math_ops.py +147 -152
- mindspore/numpy/utils.py +3 -0
- mindspore/numpy/utils_const.py +4 -4
- mindspore/opencv_core452.dll +0 -0
- mindspore/opencv_imgcodecs452.dll +0 -0
- mindspore/opencv_imgproc452.dll +0 -0
- mindspore/ops/__init__.py +9 -6
- mindspore/ops/_grad_experimental/grad_array_ops.py +4 -129
- mindspore/ops/_grad_experimental/grad_comm_ops.py +135 -36
- mindspore/ops/_grad_experimental/grad_math_ops.py +61 -298
- mindspore/ops/_grad_experimental/grad_nn_ops.py +0 -53
- mindspore/ops/_grad_experimental/grad_quant_ops.py +3 -3
- mindspore/ops/_grad_experimental/grad_sparse.py +1 -1
- mindspore/ops/_grad_experimental/grad_sparse_ops.py +3 -3
- mindspore/ops/_op_impl/__init__.py +0 -1
- mindspore/ops/_op_impl/aicpu/gamma.py +2 -0
- mindspore/ops/_op_impl/aicpu/generate_eod_mask.py +1 -1
- mindspore/ops/_op_impl/aicpu/log_uniform_candidate_sampler.py +1 -3
- mindspore/ops/_op_impl/aicpu/poisson.py +2 -0
- mindspore/ops/_op_impl/cpu/__init__.py +1 -3
- mindspore/ops/_op_impl/cpu/adam.py +2 -2
- mindspore/ops/_op_impl/cpu/adam_weight_decay.py +3 -2
- mindspore/ops/_op_impl/cpu/maximum_grad.py +16 -14
- mindspore/ops/_op_impl/cpu/minimum_grad.py +8 -0
- mindspore/ops/_vmap/vmap_array_ops.py +162 -101
- mindspore/ops/_vmap/vmap_base.py +8 -1
- mindspore/ops/_vmap/vmap_grad_math_ops.py +95 -9
- mindspore/ops/_vmap/vmap_grad_nn_ops.py +143 -58
- mindspore/ops/_vmap/vmap_image_ops.py +70 -13
- mindspore/ops/_vmap/vmap_math_ops.py +147 -59
- mindspore/ops/_vmap/vmap_nn_ops.py +292 -117
- mindspore/ops/_vmap/vmap_other_ops.py +1 -1
- mindspore/ops/auto_generate/__init__.py +31 -0
- mindspore/ops/auto_generate/cpp_create_prim_instance_helper.py +309 -0
- mindspore/ops/auto_generate/gen_arg_dtype_cast.py +252 -0
- mindspore/ops/auto_generate/gen_arg_handler.py +197 -0
- mindspore/ops/auto_generate/gen_extend_func.py +1701 -0
- mindspore/ops/auto_generate/gen_ops_def.py +8482 -0
- mindspore/ops/auto_generate/gen_ops_prim.py +16704 -0
- mindspore/ops/auto_generate/pyboost_inner_prim.py +549 -0
- mindspore/ops/composite/__init__.py +5 -2
- mindspore/ops/composite/base.py +201 -66
- mindspore/ops/composite/math_ops.py +10 -49
- mindspore/ops/composite/multitype_ops/_compile_utils.py +192 -618
- mindspore/ops/composite/multitype_ops/_constexpr_utils.py +25 -134
- mindspore/ops/composite/multitype_ops/add_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/bitwise_and_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/bitwise_or_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/bitwise_xor_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/div_impl.py +8 -0
- mindspore/ops/composite/multitype_ops/equal_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/floordiv_impl.py +8 -0
- mindspore/ops/composite/multitype_ops/getitem_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/greater_equal_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/greater_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/in_impl.py +8 -2
- mindspore/ops/composite/multitype_ops/left_shift_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/less_equal_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/less_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/logic_not_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/logical_and_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/logical_or_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/mod_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/mul_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/negative_impl.py +9 -3
- mindspore/ops/composite/multitype_ops/not_equal_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/not_in_impl.py +8 -3
- mindspore/ops/composite/multitype_ops/ones_like_impl.py +2 -2
- mindspore/ops/composite/multitype_ops/pow_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/right_shift_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/setitem_impl.py +32 -21
- mindspore/ops/composite/multitype_ops/sub_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/zeros_like_impl.py +6 -3
- mindspore/ops/deprecated.py +14 -3
- mindspore/ops/function/__init__.py +53 -11
- mindspore/ops/function/array_func.py +1269 -1821
- mindspore/ops/function/clip_func.py +19 -31
- mindspore/ops/function/debug_func.py +114 -5
- mindspore/ops/function/fft_func.py +44 -0
- mindspore/ops/function/grad/grad_func.py +30 -22
- mindspore/ops/function/image_func.py +27 -21
- mindspore/ops/function/linalg_func.py +35 -68
- mindspore/ops/function/math_func.py +1170 -2697
- mindspore/ops/function/nn_func.py +2116 -1128
- mindspore/ops/function/other_func.py +8 -8
- mindspore/ops/function/parameter_func.py +5 -93
- mindspore/ops/function/random_func.py +435 -113
- mindspore/ops/function/reshard_func.py +104 -0
- mindspore/ops/function/sparse_func.py +4 -4
- mindspore/ops/function/sparse_unary_func.py +9 -16
- mindspore/ops/function/spectral_func.py +1 -1
- mindspore/ops/function/vmap_func.py +16 -15
- mindspore/ops/functional.py +355 -346
- mindspore/ops/op_info_register.py +18 -45
- mindspore/ops/operations/__init__.py +38 -24
- mindspore/ops/operations/_grad_ops.py +21 -927
- mindspore/ops/operations/_infer_ops.py +19 -0
- mindspore/ops/operations/_inner_ops.py +173 -607
- mindspore/ops/operations/_rl_inner_ops.py +2 -2
- mindspore/ops/operations/_scalar_ops.py +5 -480
- mindspore/ops/operations/_sequence_ops.py +6 -36
- mindspore/ops/operations/_tensor_array.py +8 -8
- mindspore/ops/operations/array_ops.py +106 -2837
- mindspore/ops/operations/comm_ops.py +799 -127
- mindspore/ops/operations/custom_ops.py +124 -119
- mindspore/ops/operations/debug_ops.py +142 -41
- mindspore/ops/operations/image_ops.py +1 -217
- mindspore/ops/operations/inner_ops.py +5 -40
- mindspore/ops/operations/linalg_ops.py +1 -49
- mindspore/ops/operations/manually_defined/__init__.py +24 -0
- mindspore/ops/operations/manually_defined/_inner.py +73 -0
- mindspore/ops/operations/manually_defined/ops_def.py +2271 -0
- mindspore/ops/operations/math_ops.py +666 -4972
- mindspore/ops/operations/nn_ops.py +205 -2213
- mindspore/ops/operations/other_ops.py +60 -49
- mindspore/ops/operations/random_ops.py +50 -54
- mindspore/ops/operations/reshard_ops.py +53 -0
- mindspore/ops/operations/sparse_ops.py +4 -4
- mindspore/ops/primitive.py +216 -103
- mindspore/ops_generate/__init__.py +27 -0
- mindspore/ops_generate/arg_dtype_cast.py +252 -0
- mindspore/ops_generate/arg_handler.py +197 -0
- mindspore/ops_generate/gen_aclnn_implement.py +263 -0
- mindspore/ops_generate/gen_constants.py +36 -0
- mindspore/ops_generate/gen_ops.py +1099 -0
- mindspore/ops_generate/gen_ops_inner_prim.py +131 -0
- mindspore/ops_generate/gen_pyboost_func.py +1052 -0
- mindspore/ops_generate/gen_utils.py +209 -0
- mindspore/ops_generate/op_proto.py +145 -0
- mindspore/ops_generate/pyboost_utils.py +367 -0
- mindspore/ops_generate/template.py +261 -0
- mindspore/parallel/__init__.py +8 -4
- mindspore/parallel/_auto_parallel_context.py +100 -10
- mindspore/parallel/_cell_wrapper.py +99 -9
- mindspore/parallel/_cost_model_context.py +1 -1
- mindspore/parallel/_dp_allreduce_fusion.py +159 -159
- mindspore/parallel/_parallel_serialization.py +67 -23
- mindspore/parallel/_ps_context.py +1 -1
- mindspore/parallel/_recovery_context.py +1 -1
- mindspore/parallel/_tensor.py +99 -22
- mindspore/parallel/_transformer/__init__.py +1 -1
- mindspore/parallel/_transformer/layers.py +1 -1
- mindspore/parallel/_transformer/loss.py +1 -1
- mindspore/parallel/_transformer/moe.py +1 -1
- mindspore/parallel/_transformer/op_parallel_config.py +1 -1
- mindspore/parallel/_transformer/transformer.py +2 -2
- mindspore/parallel/_utils.py +173 -6
- mindspore/parallel/algo_parameter_config.py +8 -10
- mindspore/parallel/checkpoint_transform.py +204 -38
- mindspore/parallel/cluster/__init__.py +15 -0
- mindspore/parallel/cluster/process_entity/__init__.py +18 -0
- mindspore/parallel/cluster/process_entity/_api.py +352 -0
- mindspore/parallel/cluster/process_entity/_utils.py +101 -0
- mindspore/parallel/cluster/run.py +136 -0
- mindspore/parallel/mpi/__init__.py +1 -1
- mindspore/parallel/mpi/_mpi_config.py +1 -1
- mindspore/parallel/parameter_broadcast.py +151 -0
- mindspore/parallel/shard.py +279 -37
- mindspore/parallel/transform_safetensors.py +993 -0
- mindspore/pgodb140.dll +0 -0
- mindspore/pgort140.dll +0 -0
- mindspore/profiler/__init__.py +4 -2
- mindspore/profiler/common/constant.py +29 -0
- mindspore/profiler/common/process_pool.py +41 -0
- mindspore/profiler/common/registry.py +47 -0
- mindspore/profiler/common/singleton.py +28 -0
- mindspore/profiler/common/util.py +153 -0
- mindspore/profiler/dynamic_profiler.py +694 -0
- mindspore/profiler/envprofiling.py +18 -20
- mindspore/{_extends/parallel_compile/tbe_compiler → profiler/parser/ascend_analysis}/__init__.py +1 -1
- mindspore/profiler/parser/ascend_analysis/constant.py +71 -0
- mindspore/profiler/parser/ascend_analysis/file_manager.py +180 -0
- mindspore/profiler/parser/ascend_analysis/function_event.py +185 -0
- mindspore/profiler/parser/ascend_analysis/fwk_cann_parser.py +136 -0
- mindspore/profiler/parser/ascend_analysis/fwk_file_parser.py +131 -0
- mindspore/profiler/parser/ascend_analysis/msprof_timeline_parser.py +104 -0
- mindspore/profiler/parser/ascend_analysis/path_manager.py +313 -0
- mindspore/profiler/parser/ascend_analysis/profiler_info_parser.py +123 -0
- mindspore/profiler/parser/ascend_analysis/tlv_decoder.py +86 -0
- mindspore/profiler/parser/ascend_analysis/trace_event_manager.py +75 -0
- mindspore/profiler/parser/ascend_cluster_generator.py +14 -9
- mindspore/profiler/parser/ascend_communicate_generator.py +0 -1
- mindspore/profiler/parser/ascend_flops_generator.py +20 -4
- mindspore/profiler/parser/ascend_hccl_generator.py +29 -278
- mindspore/profiler/parser/ascend_integrate_generator.py +42 -0
- mindspore/profiler/parser/ascend_memory_generator.py +185 -0
- mindspore/profiler/parser/ascend_msprof_exporter.py +148 -146
- mindspore/profiler/parser/ascend_msprof_generator.py +73 -283
- mindspore/profiler/parser/ascend_op_generator.py +92 -42
- mindspore/profiler/parser/ascend_timeline_generator.py +298 -133
- mindspore/profiler/parser/base_timeline_generator.py +25 -25
- mindspore/profiler/parser/cpu_gpu_timeline_generator.py +25 -12
- mindspore/profiler/parser/framework_parser.py +4 -393
- mindspore/profiler/parser/gpu_analysis/__init__.py +14 -0
- mindspore/profiler/parser/gpu_analysis/function_event.py +44 -0
- mindspore/profiler/parser/gpu_analysis/fwk_file_parser.py +89 -0
- mindspore/profiler/parser/gpu_analysis/profiler_info_parser.py +72 -0
- mindspore/profiler/parser/integrator.py +3 -1
- mindspore/profiler/parser/memory_usage_parser.py +0 -154
- mindspore/profiler/parser/minddata_parser.py +72 -3
- mindspore/profiler/parser/profiler_info.py +94 -7
- mindspore/profiler/profiler.py +153 -0
- mindspore/profiler/profiling.py +631 -508
- mindspore/rewrite/__init__.py +2 -14
- mindspore/rewrite/api/node.py +122 -36
- mindspore/rewrite/api/pattern_engine.py +2 -3
- mindspore/rewrite/api/scoped_value.py +16 -15
- mindspore/rewrite/api/symbol_tree.py +45 -29
- mindspore/rewrite/ast_helpers/__init__.py +3 -6
- mindspore/rewrite/ast_helpers/ast_converter.py +143 -0
- mindspore/rewrite/ast_helpers/ast_finder.py +48 -0
- mindspore/rewrite/ast_helpers/ast_flattener.py +268 -0
- mindspore/rewrite/ast_helpers/ast_modifier.py +160 -92
- mindspore/rewrite/common/__init__.py +1 -2
- mindspore/rewrite/common/config.py +24 -0
- mindspore/rewrite/common/{rewrite_elog.py → error_log.py} +39 -39
- mindspore/rewrite/{namer.py → common/namer.py} +63 -18
- mindspore/rewrite/common/namespace.py +118 -0
- mindspore/rewrite/node/__init__.py +5 -5
- mindspore/rewrite/node/call_function.py +23 -7
- mindspore/rewrite/node/cell_container.py +7 -3
- mindspore/rewrite/node/control_flow.py +53 -28
- mindspore/rewrite/node/node.py +212 -196
- mindspore/rewrite/node/node_manager.py +51 -22
- mindspore/rewrite/node/node_topological_manager.py +3 -23
- mindspore/rewrite/parsers/__init__.py +12 -0
- mindspore/rewrite/parsers/arguments_parser.py +8 -9
- mindspore/rewrite/parsers/assign_parser.py +637 -413
- mindspore/rewrite/parsers/attribute_parser.py +3 -4
- mindspore/rewrite/parsers/class_def_parser.py +115 -148
- mindspore/rewrite/parsers/constant_parser.py +5 -5
- mindspore/rewrite/parsers/container_parser.py +4 -6
- mindspore/rewrite/parsers/expr_parser.py +55 -0
- mindspore/rewrite/parsers/for_parser.py +31 -98
- mindspore/rewrite/parsers/function_def_parser.py +13 -5
- mindspore/rewrite/parsers/if_parser.py +28 -10
- mindspore/rewrite/parsers/module_parser.py +8 -182
- mindspore/rewrite/parsers/parser.py +1 -5
- mindspore/rewrite/parsers/parser_register.py +1 -1
- mindspore/rewrite/parsers/return_parser.py +5 -10
- mindspore/rewrite/parsers/while_parser.py +59 -0
- mindspore/rewrite/sparsify/utils.py +1 -1
- mindspore/rewrite/symbol_tree/__init__.py +20 -0
- mindspore/rewrite/{symbol_tree.py → symbol_tree/symbol_tree.py} +705 -186
- mindspore/rewrite/{symbol_tree_builder.py → symbol_tree/symbol_tree_builder.py} +8 -8
- mindspore/rewrite/{symbol_tree_dumper.py → symbol_tree/symbol_tree_dumper.py} +4 -4
- mindspore/run_check/_check_version.py +40 -115
- mindspore/run_check/run_check.py +1 -1
- mindspore/safeguard/rewrite_obfuscation.py +597 -263
- mindspore/swresample-4.dll +0 -0
- mindspore/swscale-6.dll +0 -0
- mindspore/tbbmalloc.dll +0 -0
- mindspore/tinyxml2.dll +0 -0
- mindspore/train/__init__.py +7 -5
- mindspore/train/_utils.py +204 -4
- mindspore/train/amp.py +335 -295
- mindspore/train/anf_ir_pb2.py +14 -2
- mindspore/train/callback/__init__.py +5 -2
- mindspore/train/callback/_backup_and_restore.py +5 -5
- mindspore/train/callback/_callback.py +4 -4
- mindspore/train/callback/_checkpoint.py +220 -43
- mindspore/train/callback/_cluster_monitor.py +201 -0
- mindspore/train/callback/_early_stop.py +2 -2
- mindspore/train/callback/_flops_collector.py +239 -0
- mindspore/train/callback/_landscape.py +15 -9
- mindspore/train/callback/_loss_monitor.py +5 -5
- mindspore/train/callback/_on_request_exit.py +136 -33
- mindspore/train/callback/_reduce_lr_on_plateau.py +2 -2
- mindspore/train/callback/_summary_collector.py +12 -12
- mindspore/train/callback/_tft_register.py +352 -0
- mindspore/train/callback/_time_monitor.py +3 -3
- mindspore/train/data_sink.py +6 -5
- mindspore/train/dataset_helper.py +66 -23
- mindspore/train/loss_scale_manager.py +2 -2
- mindspore/train/metrics/accuracy.py +7 -7
- mindspore/train/metrics/confusion_matrix.py +8 -6
- mindspore/train/metrics/cosine_similarity.py +6 -4
- mindspore/train/metrics/error.py +2 -2
- mindspore/train/metrics/metric.py +3 -3
- mindspore/train/metrics/perplexity.py +2 -1
- mindspore/train/metrics/roc.py +4 -4
- mindspore/train/metrics/topk.py +2 -2
- mindspore/train/mind_ir_pb2.py +116 -37
- mindspore/train/model.py +382 -76
- mindspore/train/serialization.py +787 -288
- mindspore/train/summary/_summary_adapter.py +1 -1
- mindspore/train/summary/summary_record.py +51 -28
- mindspore/train/train_thor/convert_utils.py +3 -3
- mindspore/turbojpeg.dll +0 -0
- mindspore/utils/__init__.py +21 -0
- mindspore/utils/utils.py +60 -0
- mindspore/vcmeta.dll +0 -0
- mindspore/vcruntime140.dll +0 -0
- mindspore/vcruntime140_1.dll +0 -0
- mindspore/version.py +1 -1
- {mindspore-2.2.14.dist-info → mindspore-2.4.0.dist-info}/METADATA +8 -4
- mindspore-2.4.0.dist-info/RECORD +1406 -0
- {mindspore-2.2.14.dist-info → mindspore-2.4.0.dist-info}/entry_points.txt +1 -0
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_adapter.py +0 -662
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_helper.py +0 -377
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_job.py +0 -201
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_job_manager.py +0 -515
- mindspore/gen_ops.py +0 -273
- mindspore/include/c_api/ms/abstract.h +0 -67
- mindspore/include/c_api/ms/attribute.h +0 -197
- mindspore/include/c_api/ms/base/handle_types.h +0 -43
- mindspore/include/c_api/ms/base/macros.h +0 -32
- mindspore/include/c_api/ms/base/status.h +0 -33
- mindspore/include/c_api/ms/base/types.h +0 -282
- mindspore/include/c_api/ms/context.h +0 -102
- mindspore/include/c_api/ms/graph.h +0 -160
- mindspore/include/c_api/ms/node.h +0 -606
- mindspore/include/c_api/ms/tensor.h +0 -161
- mindspore/include/c_api/ms/value.h +0 -84
- mindspore/mindspore_shared_lib.dll +0 -0
- mindspore/nn/layer/flash_attention.py +0 -189
- mindspore/ops/_op_impl/aicpu/strided_slice_v2.py +0 -93
- mindspore/ops/_op_impl/aicpu/strided_slice_v2_grad.py +0 -66
- mindspore/ops/_op_impl/cpu/concat.py +0 -39
- mindspore/ops/_op_impl/cpu/tensor_shape.py +0 -42
- mindspore/ops/_op_impl/tbe/__init__.py +0 -47
- mindspore/ops/_op_impl/tbe/abs.py +0 -38
- mindspore/ops/_op_impl/tbe/abs_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/abs_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/abs_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/accumulate_n_v2.py +0 -41
- mindspore/ops/_op_impl/tbe/accumulate_n_v2_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/acos.py +0 -37
- mindspore/ops/_op_impl/tbe/acos_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/acos_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/acos_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/acosh.py +0 -37
- mindspore/ops/_op_impl/tbe/acosh_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/acosh_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/acosh_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/act_ulq_clamp_max_grad.py +0 -38
- mindspore/ops/_op_impl/tbe/act_ulq_clamp_min_grad.py +0 -38
- mindspore/ops/_op_impl/tbe/acts_ulq.py +0 -45
- mindspore/ops/_op_impl/tbe/acts_ulq_input_grad.py +0 -38
- mindspore/ops/_op_impl/tbe/adam_apply_one.py +0 -50
- mindspore/ops/_op_impl/tbe/adam_apply_one_assign.py +0 -53
- mindspore/ops/_op_impl/tbe/adam_apply_one_ds.py +0 -51
- mindspore/ops/_op_impl/tbe/adam_apply_one_with_decay.py +0 -54
- mindspore/ops/_op_impl/tbe/adam_apply_one_with_decay_assign.py +0 -54
- mindspore/ops/_op_impl/tbe/adam_apply_one_with_decay_ds.py +0 -55
- mindspore/ops/_op_impl/tbe/adaptive_max_pool2d.py +0 -37
- mindspore/ops/_op_impl/tbe/add.py +0 -42
- mindspore/ops/_op_impl/tbe/add_ds.py +0 -43
- mindspore/ops/_op_impl/tbe/add_n.py +0 -39
- mindspore/ops/_op_impl/tbe/add_n_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/addcdiv.py +0 -41
- mindspore/ops/_op_impl/tbe/addcdiv_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/addcmul.py +0 -43
- mindspore/ops/_op_impl/tbe/addcmul_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/apply_ada_max.py +0 -68
- mindspore/ops/_op_impl/tbe/apply_ada_max_ds.py +0 -69
- mindspore/ops/_op_impl/tbe/apply_adadelta.py +0 -66
- mindspore/ops/_op_impl/tbe/apply_adadelta_ds.py +0 -67
- mindspore/ops/_op_impl/tbe/apply_adagrad.py +0 -55
- mindspore/ops/_op_impl/tbe/apply_adagrad_d_a.py +0 -67
- mindspore/ops/_op_impl/tbe/apply_adagrad_ds.py +0 -56
- mindspore/ops/_op_impl/tbe/apply_adagrad_v2.py +0 -48
- mindspore/ops/_op_impl/tbe/apply_adagrad_v2_ds.py +0 -49
- mindspore/ops/_op_impl/tbe/apply_adam.py +0 -79
- mindspore/ops/_op_impl/tbe/apply_adam_ds.py +0 -80
- mindspore/ops/_op_impl/tbe/apply_adam_with_amsgrad.py +0 -60
- mindspore/ops/_op_impl/tbe/apply_adam_with_amsgrad_ds.py +0 -61
- mindspore/ops/_op_impl/tbe/apply_add_sign.py +0 -65
- mindspore/ops/_op_impl/tbe/apply_add_sign_ds.py +0 -66
- mindspore/ops/_op_impl/tbe/apply_centered_rms_prop.py +0 -77
- mindspore/ops/_op_impl/tbe/apply_centered_rms_prop_ds.py +0 -78
- mindspore/ops/_op_impl/tbe/apply_ftrl.py +0 -67
- mindspore/ops/_op_impl/tbe/apply_ftrl_ds.py +0 -68
- mindspore/ops/_op_impl/tbe/apply_gradient_descent.py +0 -44
- mindspore/ops/_op_impl/tbe/apply_gradient_descent_ds.py +0 -45
- mindspore/ops/_op_impl/tbe/apply_keras_momentum.py +0 -49
- mindspore/ops/_op_impl/tbe/apply_momentum.py +0 -64
- mindspore/ops/_op_impl/tbe/apply_momentum_ds.py +0 -65
- mindspore/ops/_op_impl/tbe/apply_power_sign.py +0 -65
- mindspore/ops/_op_impl/tbe/apply_power_sign_ds.py +0 -66
- mindspore/ops/_op_impl/tbe/apply_proximal_adagrad.py +0 -57
- mindspore/ops/_op_impl/tbe/apply_proximal_adagrad_ds.py +0 -58
- mindspore/ops/_op_impl/tbe/apply_proximal_gradient_descent.py +0 -54
- mindspore/ops/_op_impl/tbe/apply_proximal_gradient_descent_ds.py +0 -55
- mindspore/ops/_op_impl/tbe/apply_rms_prop.py +0 -52
- mindspore/ops/_op_impl/tbe/approximate_equal.py +0 -39
- mindspore/ops/_op_impl/tbe/approximate_equal_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/arg_max.py +0 -38
- mindspore/ops/_op_impl/tbe/arg_max_with_value.py +0 -38
- mindspore/ops/_op_impl/tbe/arg_max_with_value_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/arg_min.py +0 -38
- mindspore/ops/_op_impl/tbe/arg_min_v2_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/arg_min_with_value.py +0 -38
- mindspore/ops/_op_impl/tbe/arg_min_with_value_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/asin.py +0 -37
- mindspore/ops/_op_impl/tbe/asin_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/asin_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/asin_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/asinh.py +0 -37
- mindspore/ops/_op_impl/tbe/asinh_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/asinh_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/asinh_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/assign.py +0 -79
- mindspore/ops/_op_impl/tbe/assign_add.py +0 -59
- mindspore/ops/_op_impl/tbe/assign_add_ds.py +0 -60
- mindspore/ops/_op_impl/tbe/assign_ds.py +0 -80
- mindspore/ops/_op_impl/tbe/assign_sub.py +0 -55
- mindspore/ops/_op_impl/tbe/assign_sub_ds.py +0 -56
- mindspore/ops/_op_impl/tbe/atan.py +0 -37
- mindspore/ops/_op_impl/tbe/atan2.py +0 -38
- mindspore/ops/_op_impl/tbe/atan2_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/atan_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/atan_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/atan_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/atanh.py +0 -37
- mindspore/ops/_op_impl/tbe/atanh_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/avg_pool.py +0 -43
- mindspore/ops/_op_impl/tbe/avg_pool_3d.py +0 -44
- mindspore/ops/_op_impl/tbe/avg_pool_3d_grad.py +0 -45
- mindspore/ops/_op_impl/tbe/avg_pool_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/avg_pool_grad.py +0 -42
- mindspore/ops/_op_impl/tbe/avg_pool_grad_vm.py +0 -42
- mindspore/ops/_op_impl/tbe/basic_lstm_cell.py +0 -57
- mindspore/ops/_op_impl/tbe/basic_lstm_cell_c_state_grad.py +0 -50
- mindspore/ops/_op_impl/tbe/basic_lstm_cell_c_state_grad_v2.py +0 -51
- mindspore/ops/_op_impl/tbe/basic_lstm_cell_input_grad.py +0 -42
- mindspore/ops/_op_impl/tbe/basic_lstm_cell_weight_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/batch_matmul.py +0 -42
- mindspore/ops/_op_impl/tbe/batch_matmul_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/batch_matmul_v2.py +0 -47
- mindspore/ops/_op_impl/tbe/batch_to_space.py +0 -38
- mindspore/ops/_op_impl/tbe/batch_to_space_nd.py +0 -38
- mindspore/ops/_op_impl/tbe/batch_to_space_nd_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/batch_to_space_nd_v2.py +0 -41
- mindspore/ops/_op_impl/tbe/batchnorm.py +0 -58
- mindspore/ops/_op_impl/tbe/batchnorm_grad.py +0 -58
- mindspore/ops/_op_impl/tbe/bce_with_logits_loss.py +0 -42
- mindspore/ops/_op_impl/tbe/bessel_i0e.py +0 -37
- mindspore/ops/_op_impl/tbe/bessel_i0e_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/bessel_i1e.py +0 -37
- mindspore/ops/_op_impl/tbe/bessel_i1e_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/bias_add.py +0 -38
- mindspore/ops/_op_impl/tbe/bias_add_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/bias_add_grad.py +0 -53
- mindspore/ops/_op_impl/tbe/binary_cross_entropy.py +0 -39
- mindspore/ops/_op_impl/tbe/binary_cross_entropy_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/binary_cross_entropy_grad.py +0 -44
- mindspore/ops/_op_impl/tbe/binary_cross_entropy_grad_ds.py +0 -45
- mindspore/ops/_op_impl/tbe/bitwise_and.py +0 -39
- mindspore/ops/_op_impl/tbe/bitwise_and_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/bitwise_or.py +0 -39
- mindspore/ops/_op_impl/tbe/bitwise_or_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/bitwise_xor.py +0 -39
- mindspore/ops/_op_impl/tbe/bitwise_xor_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/bn_infer.py +0 -43
- mindspore/ops/_op_impl/tbe/bn_infer_ds.py +0 -45
- mindspore/ops/_op_impl/tbe/bn_infer_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/bn_infer_grad_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/bn_inference.py +0 -50
- mindspore/ops/_op_impl/tbe/bn_training_reduce.py +0 -38
- mindspore/ops/_op_impl/tbe/bn_training_reduce_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/bn_training_reduce_grad.py +0 -46
- mindspore/ops/_op_impl/tbe/bn_training_reduce_grad_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/bn_training_update.py +0 -52
- mindspore/ops/_op_impl/tbe/bn_training_update_ds.py +0 -53
- mindspore/ops/_op_impl/tbe/bn_training_update_grad.py +0 -44
- mindspore/ops/_op_impl/tbe/bn_training_update_grad_ds.py +0 -45
- mindspore/ops/_op_impl/tbe/bn_training_update_v2.py +0 -48
- mindspore/ops/_op_impl/tbe/bn_training_update_v3.py +0 -51
- mindspore/ops/_op_impl/tbe/bounding_box_decode.py +0 -41
- mindspore/ops/_op_impl/tbe/bounding_box_decode_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/bounding_box_encode.py +0 -38
- mindspore/ops/_op_impl/tbe/broadcast_to.py +0 -40
- mindspore/ops/_op_impl/tbe/broadcast_to_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/cast.py +0 -55
- mindspore/ops/_op_impl/tbe/cast_ds.py +0 -58
- mindspore/ops/_op_impl/tbe/cdist.py +0 -38
- mindspore/ops/_op_impl/tbe/cdist_grad.py +0 -42
- mindspore/ops/_op_impl/tbe/ceil.py +0 -37
- mindspore/ops/_op_impl/tbe/ceil_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/celu.py +0 -39
- mindspore/ops/_op_impl/tbe/centralization.py +0 -39
- mindspore/ops/_op_impl/tbe/check_valid.py +0 -38
- mindspore/ops/_op_impl/tbe/check_valid_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/clip_by_norm_no_div_sum.py +0 -41
- mindspore/ops/_op_impl/tbe/clip_by_norm_no_div_sum_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/clip_by_value.py +0 -41
- mindspore/ops/_op_impl/tbe/clip_by_value_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/concat.py +0 -40
- mindspore/ops/_op_impl/tbe/concat_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/confusion_matrix.py +0 -63
- mindspore/ops/_op_impl/tbe/confusion_mul_grad.py +0 -40
- mindspore/ops/_op_impl/tbe/confusion_softmax_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/confusion_transpose_d.py +0 -39
- mindspore/ops/_op_impl/tbe/conv2d.py +0 -47
- mindspore/ops/_op_impl/tbe/conv2d_backprop_filter.py +0 -42
- mindspore/ops/_op_impl/tbe/conv2d_backprop_filter_ds.py +0 -43
- mindspore/ops/_op_impl/tbe/conv2d_backprop_input.py +0 -42
- mindspore/ops/_op_impl/tbe/conv2d_backprop_input_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/conv2d_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/conv2d_transpose.py +0 -48
- mindspore/ops/_op_impl/tbe/conv3d.py +0 -45
- mindspore/ops/_op_impl/tbe/conv3d_backprop_filter.py +0 -42
- mindspore/ops/_op_impl/tbe/conv3d_backprop_input.py +0 -42
- mindspore/ops/_op_impl/tbe/conv3d_transpose.py +0 -47
- mindspore/ops/_op_impl/tbe/conv3d_transpose_ds.py +0 -48
- mindspore/ops/_op_impl/tbe/cos.py +0 -37
- mindspore/ops/_op_impl/tbe/cos_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/cosh.py +0 -37
- mindspore/ops/_op_impl/tbe/cosh_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/ctc_loss_v2.py +0 -42
- mindspore/ops/_op_impl/tbe/ctc_loss_v2_grad.py +0 -44
- mindspore/ops/_op_impl/tbe/cum_sum.py +0 -42
- mindspore/ops/_op_impl/tbe/cum_sum_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/cummin.py +0 -41
- mindspore/ops/_op_impl/tbe/cumprod.py +0 -42
- mindspore/ops/_op_impl/tbe/data_format_dim_map.py +0 -38
- mindspore/ops/_op_impl/tbe/data_format_dim_map_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/deformable_offsets.py +0 -45
- mindspore/ops/_op_impl/tbe/deformable_offsets_grad.py +0 -48
- mindspore/ops/_op_impl/tbe/depth_to_space_ds.py +0 -49
- mindspore/ops/_op_impl/tbe/depthwise_conv2d.py +0 -44
- mindspore/ops/_op_impl/tbe/depthwise_conv2d_backprop_filter.py +0 -41
- mindspore/ops/_op_impl/tbe/depthwise_conv2d_backprop_input.py +0 -41
- mindspore/ops/_op_impl/tbe/diag.py +0 -38
- mindspore/ops/_op_impl/tbe/diag_part.py +0 -38
- mindspore/ops/_op_impl/tbe/dilation.py +0 -40
- mindspore/ops/_op_impl/tbe/div.py +0 -41
- mindspore/ops/_op_impl/tbe/div_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/div_no_nan.py +0 -41
- mindspore/ops/_op_impl/tbe/div_no_nan_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/dropout_do_mask.py +0 -38
- mindspore/ops/_op_impl/tbe/dropout_do_mask_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/dropout_do_mask_v3.py +0 -39
- mindspore/ops/_op_impl/tbe/dynamic_atomic_addr_clean.py +0 -34
- mindspore/ops/_op_impl/tbe/dynamic_gru_v2.py +0 -95
- mindspore/ops/_op_impl/tbe/dynamic_rnn.py +0 -82
- mindspore/ops/_op_impl/tbe/elu.py +0 -38
- mindspore/ops/_op_impl/tbe/elu_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/elu_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/elu_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/equal.py +0 -42
- mindspore/ops/_op_impl/tbe/equal_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/erf.py +0 -37
- mindspore/ops/_op_impl/tbe/erf_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/erfc.py +0 -37
- mindspore/ops/_op_impl/tbe/erfc_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/erfinv.py +0 -36
- mindspore/ops/_op_impl/tbe/exp.py +0 -40
- mindspore/ops/_op_impl/tbe/exp_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/expand_dims.py +0 -38
- mindspore/ops/_op_impl/tbe/expm1.py +0 -37
- mindspore/ops/_op_impl/tbe/expm1_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/extract_image_patches.py +0 -41
- mindspore/ops/_op_impl/tbe/extract_volume_patches.py +0 -39
- mindspore/ops/_op_impl/tbe/fake_quant_with_min_max_vars.py +0 -39
- mindspore/ops/_op_impl/tbe/fake_quant_with_min_max_vars_gradient.py +0 -43
- mindspore/ops/_op_impl/tbe/fake_quant_with_min_max_vars_per_channel.py +0 -39
- mindspore/ops/_op_impl/tbe/fake_quant_with_min_max_vars_per_channel_gradient.py +0 -43
- mindspore/ops/_op_impl/tbe/fast_gelu.py +0 -37
- mindspore/ops/_op_impl/tbe/fast_gelu_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/fast_gelu_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/fast_gelu_grad_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/fill.py +0 -56
- mindspore/ops/_op_impl/tbe/fill_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/flatten.py +0 -48
- mindspore/ops/_op_impl/tbe/floor.py +0 -37
- mindspore/ops/_op_impl/tbe/floor_div.py +0 -41
- mindspore/ops/_op_impl/tbe/floor_div_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/floor_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/floor_mod.py +0 -39
- mindspore/ops/_op_impl/tbe/floor_mod_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/fused_dbn_dw.py +0 -52
- mindspore/ops/_op_impl/tbe/fused_mul_add.py +0 -38
- mindspore/ops/_op_impl/tbe/fused_mul_add_n.py +0 -48
- mindspore/ops/_op_impl/tbe/fused_mul_add_n_l2loss.py +0 -53
- mindspore/ops/_op_impl/tbe/fused_mul_apply_momentum.py +0 -57
- mindspore/ops/_op_impl/tbe/fused_mul_apply_momentum_extern.py +0 -67
- mindspore/ops/_op_impl/tbe/gather_nd.py +0 -52
- mindspore/ops/_op_impl/tbe/gather_nd_ds.py +0 -48
- mindspore/ops/_op_impl/tbe/gather_v2.py +0 -56
- mindspore/ops/_op_impl/tbe/gather_v2_ds.py +0 -68
- mindspore/ops/_op_impl/tbe/gelu.py +0 -37
- mindspore/ops/_op_impl/tbe/gelu_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/gelu_grad.py +0 -42
- mindspore/ops/_op_impl/tbe/gelu_grad_ds.py +0 -43
- mindspore/ops/_op_impl/tbe/ger.py +0 -43
- mindspore/ops/_op_impl/tbe/ger_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/greater.py +0 -43
- mindspore/ops/_op_impl/tbe/greater_equal.py +0 -41
- mindspore/ops/_op_impl/tbe/greater_equal_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/gru_v2_hidden_grad.py +0 -51
- mindspore/ops/_op_impl/tbe/gru_v2_hidden_grad_cell.py +0 -52
- mindspore/ops/_op_impl/tbe/hard_swish.py +0 -37
- mindspore/ops/_op_impl/tbe/hard_swish_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/hard_swish_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/hard_swish_grad_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/histogram_fixed_width.py +0 -40
- mindspore/ops/_op_impl/tbe/hshrink.py +0 -33
- mindspore/ops/_op_impl/tbe/hshrink_grad.py +0 -37
- mindspore/ops/_op_impl/tbe/hsigmoid.py +0 -45
- mindspore/ops/_op_impl/tbe/hsigmoid_grad.py +0 -39
- mindspore/ops/_op_impl/tbe/ifmr.py +0 -47
- mindspore/ops/_op_impl/tbe/ifmr_ds.py +0 -48
- mindspore/ops/_op_impl/tbe/im2col.py +0 -42
- mindspore/ops/_op_impl/tbe/in_top_k.py +0 -37
- mindspore/ops/_op_impl/tbe/inplace_add.py +0 -39
- mindspore/ops/_op_impl/tbe/inplace_index_add.py +0 -46
- mindspore/ops/_op_impl/tbe/inplace_sub.py +0 -39
- mindspore/ops/_op_impl/tbe/inplace_update.py +0 -39
- mindspore/ops/_op_impl/tbe/inplace_update_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/inv.py +0 -38
- mindspore/ops/_op_impl/tbe/inv_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/inv_grad.py +0 -40
- mindspore/ops/_op_impl/tbe/inv_grad_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/invert.py +0 -37
- mindspore/ops/_op_impl/tbe/invert_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/iou.py +0 -38
- mindspore/ops/_op_impl/tbe/iou_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/is_close.py +0 -40
- mindspore/ops/_op_impl/tbe/kl_div_loss.py +0 -38
- mindspore/ops/_op_impl/tbe/kl_div_loss_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/kl_div_loss_grad.py +0 -40
- mindspore/ops/_op_impl/tbe/l2_loss.py +0 -36
- mindspore/ops/_op_impl/tbe/l2_loss_ds.py +0 -37
- mindspore/ops/_op_impl/tbe/l2_normalize.py +0 -38
- mindspore/ops/_op_impl/tbe/l2_normalize_grad.py +0 -40
- mindspore/ops/_op_impl/tbe/lamb_apply_optimizer_assign.py +0 -55
- mindspore/ops/_op_impl/tbe/lamb_apply_weight_assign.py +0 -42
- mindspore/ops/_op_impl/tbe/lamb_next_mv.py +0 -59
- mindspore/ops/_op_impl/tbe/lamb_next_mv_with_decay.py +0 -59
- mindspore/ops/_op_impl/tbe/lamb_next_right.py +0 -44
- mindspore/ops/_op_impl/tbe/lamb_update_with_lr.py +0 -48
- mindspore/ops/_op_impl/tbe/lamb_update_with_lr_v2.py +0 -44
- mindspore/ops/_op_impl/tbe/lars_update.py +0 -50
- mindspore/ops/_op_impl/tbe/lars_update_ds.py +0 -51
- mindspore/ops/_op_impl/tbe/layer_norm.py +0 -46
- mindspore/ops/_op_impl/tbe/layer_norm_beta_gamma_backprop.py +0 -44
- mindspore/ops/_op_impl/tbe/layer_norm_beta_gamma_backprop_ds.py +0 -45
- mindspore/ops/_op_impl/tbe/layer_norm_beta_gamma_backprop_v2.py +0 -40
- mindspore/ops/_op_impl/tbe/layer_norm_beta_gamma_backprop_v2_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/layer_norm_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/layer_norm_grad.py +0 -48
- mindspore/ops/_op_impl/tbe/layer_norm_x_backprop.py +0 -43
- mindspore/ops/_op_impl/tbe/layer_norm_x_backprop_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/layer_norm_x_backprop_v2.py +0 -45
- mindspore/ops/_op_impl/tbe/layer_norm_x_backprop_v2_ds.py +0 -45
- mindspore/ops/_op_impl/tbe/lerp.py +0 -38
- mindspore/ops/_op_impl/tbe/less.py +0 -41
- mindspore/ops/_op_impl/tbe/less_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/less_equal.py +0 -41
- mindspore/ops/_op_impl/tbe/less_equal_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/log.py +0 -40
- mindspore/ops/_op_impl/tbe/log1p.py +0 -37
- mindspore/ops/_op_impl/tbe/log1p_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/log_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/logical_and.py +0 -37
- mindspore/ops/_op_impl/tbe/logical_and_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/logical_not.py +0 -36
- mindspore/ops/_op_impl/tbe/logical_not_ds.py +0 -37
- mindspore/ops/_op_impl/tbe/logical_or.py +0 -37
- mindspore/ops/_op_impl/tbe/logical_or_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/logsoftmax.py +0 -37
- mindspore/ops/_op_impl/tbe/logsoftmax_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/logsoftmax_grad.py +0 -38
- mindspore/ops/_op_impl/tbe/logsoftmax_grad_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/lp_norm.py +0 -40
- mindspore/ops/_op_impl/tbe/lp_norm_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/lrn.py +0 -41
- mindspore/ops/_op_impl/tbe/lrn_grad.py +0 -42
- mindspore/ops/_op_impl/tbe/lstm_input_grad.py +0 -51
- mindspore/ops/_op_impl/tbe/masked_fill.py +0 -40
- mindspore/ops/_op_impl/tbe/masked_fill_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/matmul.py +0 -53
- mindspore/ops/_op_impl/tbe/matmul_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/matmul_v2.py +0 -50
- mindspore/ops/_op_impl/tbe/matrix_diag.py +0 -45
- mindspore/ops/_op_impl/tbe/matrix_diag_part.py +0 -45
- mindspore/ops/_op_impl/tbe/matrix_set_diag.py +0 -46
- mindspore/ops/_op_impl/tbe/max_pool.py +0 -39
- mindspore/ops/_op_impl/tbe/max_pool3d.py +0 -44
- mindspore/ops/_op_impl/tbe/max_pool3d_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/max_pool3d_grad_grad.py +0 -44
- mindspore/ops/_op_impl/tbe/max_pool_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/max_pool_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/max_pool_grad_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/max_pool_grad_grad_with_argmax.py +0 -41
- mindspore/ops/_op_impl/tbe/max_pool_grad_with_argmax.py +0 -42
- mindspore/ops/_op_impl/tbe/max_pool_with_argmax.py +0 -40
- mindspore/ops/_op_impl/tbe/maximum.py +0 -39
- mindspore/ops/_op_impl/tbe/maximum_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/maximum_grad.py +0 -46
- mindspore/ops/_op_impl/tbe/maximum_grad_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/mem_set.py +0 -38
- mindspore/ops/_op_impl/tbe/minimum.py +0 -40
- mindspore/ops/_op_impl/tbe/minimum_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/minimum_grad.py +0 -46
- mindspore/ops/_op_impl/tbe/minimum_grad_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/mish.py +0 -37
- mindspore/ops/_op_impl/tbe/mod.py +0 -41
- mindspore/ops/_op_impl/tbe/mod_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/mul.py +0 -37
- mindspore/ops/_op_impl/tbe/mul_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/mul_no_nan.py +0 -39
- mindspore/ops/_op_impl/tbe/mul_no_nan_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/multilabel_margin_loss.py +0 -39
- mindspore/ops/_op_impl/tbe/neg.py +0 -39
- mindspore/ops/_op_impl/tbe/neg_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/new_im2col.py +0 -40
- mindspore/ops/_op_impl/tbe/nll_loss.py +0 -41
- mindspore/ops/_op_impl/tbe/nll_loss_grad.py +0 -44
- mindspore/ops/_op_impl/tbe/nms_with_mask.py +0 -39
- mindspore/ops/_op_impl/tbe/not_equal.py +0 -41
- mindspore/ops/_op_impl/tbe/not_equal_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/npu_alloc_float_status.py +0 -34
- mindspore/ops/_op_impl/tbe/npu_clear_float_status.py +0 -35
- mindspore/ops/_op_impl/tbe/npu_clear_float_status_v2.py +0 -35
- mindspore/ops/_op_impl/tbe/npu_get_float_status.py +0 -35
- mindspore/ops/_op_impl/tbe/npu_get_float_status_v2.py +0 -35
- mindspore/ops/_op_impl/tbe/one_hot.py +0 -48
- mindspore/ops/_op_impl/tbe/one_hot_ds.py +0 -45
- mindspore/ops/_op_impl/tbe/ones_like.py +0 -40
- mindspore/ops/_op_impl/tbe/ones_like_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/p_s_r_o_i_pooling.py +0 -40
- mindspore/ops/_op_impl/tbe/p_s_r_o_i_pooling_grad.py +0 -40
- mindspore/ops/_op_impl/tbe/pack.py +0 -58
- mindspore/ops/_op_impl/tbe/pack_ds.py +0 -59
- mindspore/ops/_op_impl/tbe/pad_d.py +0 -40
- mindspore/ops/_op_impl/tbe/pad_d_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/parallel_concat.py +0 -70
- mindspore/ops/_op_impl/tbe/parallel_resize_bilinear.py +0 -45
- mindspore/ops/_op_impl/tbe/parallel_resize_bilinear_grad.py +0 -44
- mindspore/ops/_op_impl/tbe/pdist.py +0 -36
- mindspore/ops/_op_impl/tbe/pooling.py +0 -46
- mindspore/ops/_op_impl/tbe/population_count.py +0 -38
- mindspore/ops/_op_impl/tbe/pow.py +0 -41
- mindspore/ops/_op_impl/tbe/pow_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/prelu.py +0 -37
- mindspore/ops/_op_impl/tbe/prelu_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/prelu_grad.py +0 -40
- mindspore/ops/_op_impl/tbe/range.py +0 -39
- mindspore/ops/_op_impl/tbe/real_div.py +0 -38
- mindspore/ops/_op_impl/tbe/real_div_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/reciprocal.py +0 -36
- mindspore/ops/_op_impl/tbe/reciprocal_ds.py +0 -37
- mindspore/ops/_op_impl/tbe/reciprocal_grad.py +0 -38
- mindspore/ops/_op_impl/tbe/reciprocal_grad_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/reduce_all.py +0 -38
- mindspore/ops/_op_impl/tbe/reduce_all_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/reduce_any.py +0 -38
- mindspore/ops/_op_impl/tbe/reduce_any_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/reduce_max.py +0 -43
- mindspore/ops/_op_impl/tbe/reduce_max_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/reduce_mean.py +0 -40
- mindspore/ops/_op_impl/tbe/reduce_mean_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/reduce_min.py +0 -41
- mindspore/ops/_op_impl/tbe/reduce_min_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/reduce_prod.py +0 -42
- mindspore/ops/_op_impl/tbe/reduce_prod_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/reduce_std.py +0 -44
- mindspore/ops/_op_impl/tbe/reduce_sum.py +0 -39
- mindspore/ops/_op_impl/tbe/reduce_sum_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/relu.py +0 -39
- mindspore/ops/_op_impl/tbe/relu6.py +0 -38
- mindspore/ops/_op_impl/tbe/relu6_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/relu6_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/relu6_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/relu_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/relu_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/relu_grad_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/relu_grad_v2.py +0 -40
- mindspore/ops/_op_impl/tbe/relu_grad_v2_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/relu_v2.py +0 -40
- mindspore/ops/_op_impl/tbe/relu_v2_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/renorm.py +0 -39
- mindspore/ops/_op_impl/tbe/resize_bilinear.py +0 -40
- mindspore/ops/_op_impl/tbe/resize_bilinear_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/resize_bilinear_v2.py +0 -43
- mindspore/ops/_op_impl/tbe/resize_nearest_neighbor.py +0 -40
- mindspore/ops/_op_impl/tbe/resize_nearest_neighbor_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/resize_nearest_neighbor_grad.py +0 -39
- mindspore/ops/_op_impl/tbe/resize_nearest_neighbor_grad_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/reverse_v2_d.py +0 -37
- mindspore/ops/_op_impl/tbe/rint.py +0 -37
- mindspore/ops/_op_impl/tbe/rint_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/roi_align.py +0 -43
- mindspore/ops/_op_impl/tbe/roi_align_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/roi_align_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/roi_align_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/roll.py +0 -42
- mindspore/ops/_op_impl/tbe/round.py +0 -38
- mindspore/ops/_op_impl/tbe/round_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/rsqrt.py +0 -37
- mindspore/ops/_op_impl/tbe/rsqrt_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/rsqrt_grad.py +0 -40
- mindspore/ops/_op_impl/tbe/rsqrt_grad_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/scatter_add.py +0 -44
- mindspore/ops/_op_impl/tbe/scatter_div.py +0 -46
- mindspore/ops/_op_impl/tbe/scatter_max.py +0 -45
- mindspore/ops/_op_impl/tbe/scatter_min.py +0 -45
- mindspore/ops/_op_impl/tbe/scatter_mul.py +0 -44
- mindspore/ops/_op_impl/tbe/scatter_nd.py +0 -41
- mindspore/ops/_op_impl/tbe/scatter_nd_add.py +0 -45
- mindspore/ops/_op_impl/tbe/scatter_nd_d.py +0 -41
- mindspore/ops/_op_impl/tbe/scatter_nd_ds.py +0 -49
- mindspore/ops/_op_impl/tbe/scatter_nd_sub.py +0 -47
- mindspore/ops/_op_impl/tbe/scatter_nd_sub_ds.py +0 -48
- mindspore/ops/_op_impl/tbe/scatter_nd_update.py +0 -47
- mindspore/ops/_op_impl/tbe/scatter_nd_update_ds.py +0 -48
- mindspore/ops/_op_impl/tbe/scatter_non_aliasing_add.py +0 -39
- mindspore/ops/_op_impl/tbe/scatter_non_aliasing_add_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/scatter_sub.py +0 -47
- mindspore/ops/_op_impl/tbe/scatter_sub_ds.py +0 -48
- mindspore/ops/_op_impl/tbe/scatter_update.py +0 -43
- mindspore/ops/_op_impl/tbe/select.py +0 -38
- mindspore/ops/_op_impl/tbe/select_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/selu.py +0 -39
- mindspore/ops/_op_impl/tbe/selu_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/sgd.py +0 -62
- mindspore/ops/_op_impl/tbe/sigmoid.py +0 -37
- mindspore/ops/_op_impl/tbe/sigmoid_cross_entropy_with_logits.py +0 -41
- mindspore/ops/_op_impl/tbe/sigmoid_cross_entropy_with_logits_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/sigmoid_cross_entropy_with_logits_grad.py +0 -42
- mindspore/ops/_op_impl/tbe/sigmoid_cross_entropy_with_logits_grad_ds.py +0 -43
- mindspore/ops/_op_impl/tbe/sigmoid_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/sigmoid_grad.py +0 -39
- mindspore/ops/_op_impl/tbe/sigmoid_grad_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/sign.py +0 -38
- mindspore/ops/_op_impl/tbe/sign_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/sin.py +0 -37
- mindspore/ops/_op_impl/tbe/sin_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/sinh.py +0 -37
- mindspore/ops/_op_impl/tbe/sinh_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/slice.py +0 -58
- mindspore/ops/_op_impl/tbe/smooth_l1_loss.py +0 -45
- mindspore/ops/_op_impl/tbe/smooth_l1_loss_ds.py +0 -46
- mindspore/ops/_op_impl/tbe/smooth_l1_loss_grad.py +0 -46
- mindspore/ops/_op_impl/tbe/smooth_l1_loss_grad_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/soft_margin_loss.py +0 -38
- mindspore/ops/_op_impl/tbe/soft_margin_loss_grad.py +0 -39
- mindspore/ops/_op_impl/tbe/soft_shrink.py +0 -36
- mindspore/ops/_op_impl/tbe/soft_shrink_grad.py +0 -38
- mindspore/ops/_op_impl/tbe/softmax.py +0 -37
- mindspore/ops/_op_impl/tbe/softmax_cross_entropy_with_logits.py +0 -38
- mindspore/ops/_op_impl/tbe/softmax_cross_entropy_with_logits_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/softmax_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/softmax_grad_ext.py +0 -42
- mindspore/ops/_op_impl/tbe/softmax_v2_with_dropout_do_mask_v3.py +0 -39
- mindspore/ops/_op_impl/tbe/softplus.py +0 -37
- mindspore/ops/_op_impl/tbe/softplus_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/softplus_grad.py +0 -38
- mindspore/ops/_op_impl/tbe/softplus_grad_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/softsign.py +0 -37
- mindspore/ops/_op_impl/tbe/softsign_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/sort.py +0 -38
- mindspore/ops/_op_impl/tbe/sort_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/space_to_batch.py +0 -38
- mindspore/ops/_op_impl/tbe/space_to_batch_nd.py +0 -38
- mindspore/ops/_op_impl/tbe/space_to_depth.py +0 -47
- mindspore/ops/_op_impl/tbe/sparse_apply_adadelta.py +0 -56
- mindspore/ops/_op_impl/tbe/sparse_apply_adagrad.py +0 -45
- mindspore/ops/_op_impl/tbe/sparse_apply_adagrad_ds.py +0 -46
- mindspore/ops/_op_impl/tbe/sparse_apply_adagrad_v2.py +0 -46
- mindspore/ops/_op_impl/tbe/sparse_apply_adagrad_v2_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/sparse_apply_ftrl_d.py +0 -53
- mindspore/ops/_op_impl/tbe/sparse_apply_ftrl_d_ds.py +0 -50
- mindspore/ops/_op_impl/tbe/sparse_apply_ftrl_v2.py +0 -50
- mindspore/ops/_op_impl/tbe/sparse_apply_proximal_adagrad.py +0 -66
- mindspore/ops/_op_impl/tbe/sparse_apply_proximal_adagrad_ds.py +0 -67
- mindspore/ops/_op_impl/tbe/sparse_apply_r_m_s_prop.py +0 -57
- mindspore/ops/_op_impl/tbe/sparse_apply_r_m_s_prop_ds.py +0 -58
- mindspore/ops/_op_impl/tbe/sparse_gather_v2.py +0 -56
- mindspore/ops/_op_impl/tbe/sparse_gather_v2_ds.py +0 -58
- mindspore/ops/_op_impl/tbe/split_d.py +0 -38
- mindspore/ops/_op_impl/tbe/split_d_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/split_v.py +0 -39
- mindspore/ops/_op_impl/tbe/splitv.py +0 -39
- mindspore/ops/_op_impl/tbe/sqrt.py +0 -37
- mindspore/ops/_op_impl/tbe/sqrt_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/sqrt_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/sqrt_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/square.py +0 -38
- mindspore/ops/_op_impl/tbe/square_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/square_sum_all.py +0 -40
- mindspore/ops/_op_impl/tbe/square_sum_all_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/square_sum_v1.py +0 -38
- mindspore/ops/_op_impl/tbe/square_sum_v1_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/square_sum_v2.py +0 -39
- mindspore/ops/_op_impl/tbe/squared_difference.py +0 -39
- mindspore/ops/_op_impl/tbe/squared_difference_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/squeeze.py +0 -37
- mindspore/ops/_op_impl/tbe/strided_read.py +0 -38
- mindspore/ops/_op_impl/tbe/strided_slice_d.py +0 -44
- mindspore/ops/_op_impl/tbe/strided_slice_ds.py +0 -71
- mindspore/ops/_op_impl/tbe/strided_slice_grad_d.py +0 -51
- mindspore/ops/_op_impl/tbe/strided_slice_grad_ds.py +0 -57
- mindspore/ops/_op_impl/tbe/strided_write.py +0 -38
- mindspore/ops/_op_impl/tbe/sub.py +0 -39
- mindspore/ops/_op_impl/tbe/sub_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/tan.py +0 -38
- mindspore/ops/_op_impl/tbe/tan_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/tanh.py +0 -37
- mindspore/ops/_op_impl/tbe/tanh_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/tanh_grad.py +0 -39
- mindspore/ops/_op_impl/tbe/tanh_grad_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/tensor_move.py +0 -49
- mindspore/ops/_op_impl/tbe/tensor_move_ds.py +0 -50
- mindspore/ops/_op_impl/tbe/tensor_scatter_update.py +0 -41
- mindspore/ops/_op_impl/tbe/tile.py +0 -37
- mindspore/ops/_op_impl/tbe/tile_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/top_k.py +0 -42
- mindspore/ops/_op_impl/tbe/top_k_ds.py +0 -43
- mindspore/ops/_op_impl/tbe/trans_data.py +0 -167
- mindspore/ops/_op_impl/tbe/trans_data_ds.py +0 -180
- mindspore/ops/_op_impl/tbe/trans_data_rnn.py +0 -44
- mindspore/ops/_op_impl/tbe/transpose.py +0 -60
- mindspore/ops/_op_impl/tbe/transpose_d.py +0 -47
- mindspore/ops/_op_impl/tbe/transpose_nod.py +0 -60
- mindspore/ops/_op_impl/tbe/trunc.py +0 -39
- mindspore/ops/_op_impl/tbe/truncate_div.py +0 -41
- mindspore/ops/_op_impl/tbe/truncate_div_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/truncate_mod.py +0 -41
- mindspore/ops/_op_impl/tbe/truncate_mod_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/unpack.py +0 -38
- mindspore/ops/_op_impl/tbe/unpack_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/unsorted_segment_max.py +0 -49
- mindspore/ops/_op_impl/tbe/unsorted_segment_max_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/unsorted_segment_min.py +0 -49
- mindspore/ops/_op_impl/tbe/unsorted_segment_min_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/unsorted_segment_prod.py +0 -49
- mindspore/ops/_op_impl/tbe/unsorted_segment_prod_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/unsorted_segment_sum.py +0 -38
- mindspore/ops/_op_impl/tbe/unsorted_segment_sum_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/wts_arq.py +0 -40
- mindspore/ops/_op_impl/tbe/xdivy.py +0 -38
- mindspore/ops/_op_impl/tbe/xdivy_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/xlogy.py +0 -38
- mindspore/ops/_op_impl/tbe/xlogy_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/zeros_like.py +0 -41
- mindspore/ops/_op_impl/tbe/zeros_like_ds.py +0 -42
- mindspore/ops/_tracefunc.py +0 -241
- mindspore/ops/arg_dtype_cast.py +0 -54
- mindspore/ops/silent_check.py +0 -162
- mindspore/profiler/parser/msadvisor_analyzer.py +0 -82
- mindspore/profiler/parser/msadvisor_parser.py +0 -240
- mindspore/rewrite/api/tree_node_helper.py +0 -60
- mindspore/rewrite/ast_helpers/ast_creator.py +0 -115
- mindspore/rewrite/ast_transformers/flatten_recursive_stmt.py +0 -267
- mindspore/rewrite/ast_transformers/remove_return_out_of_if.py +0 -228
- mindspore/rewrite/namespace.py +0 -53
- mindspore-2.2.14.dist-info/RECORD +0 -1924
- {mindspore-2.2.14.dist-info → mindspore-2.4.0.dist-info}/WHEEL +0 -0
- {mindspore-2.2.14.dist-info → mindspore-2.4.0.dist-info}/top_level.txt +0 -0
mindspore/nn/layer/activation.py
CHANGED
|
@@ -33,6 +33,7 @@ __all__ = ['Softmin',
|
|
|
33
33
|
'Softmax',
|
|
34
34
|
'Softmax2d',
|
|
35
35
|
'LogSoftmax',
|
|
36
|
+
'LogSoftmaxExt',
|
|
36
37
|
'ReLU',
|
|
37
38
|
'ReLU6',
|
|
38
39
|
'RReLU',
|
|
@@ -46,6 +47,7 @@ __all__ = ['Softmin',
|
|
|
46
47
|
'Sigmoid',
|
|
47
48
|
'Softsign',
|
|
48
49
|
'PReLU',
|
|
50
|
+
'PReLUExt',
|
|
49
51
|
'get_activation',
|
|
50
52
|
'LeakyReLU',
|
|
51
53
|
'HSigmoid',
|
|
@@ -64,7 +66,7 @@ __all__ = ['Softmin',
|
|
|
64
66
|
|
|
65
67
|
class CELU(Cell):
|
|
66
68
|
r"""
|
|
67
|
-
|
|
69
|
+
CELU Activation Operator.
|
|
68
70
|
|
|
69
71
|
Applies the continuously differentiable exponential linear units function element-wise.
|
|
70
72
|
|
|
@@ -72,9 +74,11 @@ class CELU(Cell):
|
|
|
72
74
|
|
|
73
75
|
\text{CELU}(x) = \max(0,x) + \min(0, \alpha * (\exp(x/\alpha) - 1))
|
|
74
76
|
|
|
77
|
+
For more details, refer to `CELU <https://arxiv.org/abs/1704.07483>`_ .
|
|
78
|
+
|
|
75
79
|
CELU Activation Function Graph:
|
|
76
80
|
|
|
77
|
-
.. image:: images/CELU.png
|
|
81
|
+
.. image:: ../images/CELU.png
|
|
78
82
|
:align: center
|
|
79
83
|
|
|
80
84
|
Args:
|
|
@@ -91,7 +95,7 @@ class CELU(Cell):
|
|
|
91
95
|
TypeError: If `alpha` is not a float.
|
|
92
96
|
ValueError: If `alpha` has the value of 0.
|
|
93
97
|
TypeError: If `x` is not a Tensor.
|
|
94
|
-
TypeError: If the dtype of
|
|
98
|
+
TypeError: If the dtype of `x` is neither float16 nor float32.
|
|
95
99
|
|
|
96
100
|
Supported Platforms:
|
|
97
101
|
``Ascend`` ``GPU`` ``CPU``
|
|
@@ -234,25 +238,24 @@ class Softmax(Cell):
|
|
|
234
238
|
Softmax is defined as:
|
|
235
239
|
|
|
236
240
|
.. math::
|
|
237
|
-
\text{softmax}(
|
|
241
|
+
\text{softmax}(input_{i}) = \frac{\exp(input_i)}{\sum_{j=0}^{n-1}\exp(input_j)},
|
|
238
242
|
|
|
239
|
-
where :math:`
|
|
243
|
+
where :math:`input_{i}` is the :math:`i`-th slice in the given dimension of the input Tensor.
|
|
240
244
|
|
|
241
245
|
Args:
|
|
242
|
-
axis (
|
|
243
|
-
the range of axis is `[-
|
|
246
|
+
axis (int, optional): The axis to apply Softmax operation, if the dimension of `input` is input.ndim,
|
|
247
|
+
the range of axis is `[-input.ndim, input.ndim)`, -1 means the last dimension. Default: ``-1`` .
|
|
244
248
|
|
|
245
249
|
Inputs:
|
|
246
|
-
- **
|
|
250
|
+
- **input** (Tensor) - The input of Softmax.
|
|
247
251
|
|
|
248
252
|
Outputs:
|
|
249
|
-
Tensor, which has the same type and shape as `
|
|
253
|
+
Tensor, which has the same type and shape as `input` with values in the range[0, 1].
|
|
250
254
|
|
|
251
255
|
Raises:
|
|
252
256
|
TypeError: If `axis` is neither an int nor a tuple.
|
|
253
|
-
TypeError: If dtype of `x` is neither float16 nor float32.
|
|
254
257
|
ValueError: If `axis` is a tuple whose length is less than 1.
|
|
255
|
-
ValueError: If `axis` is a tuple whose elements are not all in range [-
|
|
258
|
+
ValueError: If `axis` is a tuple whose elements are not all in range `[-input.ndim, input.ndim)`.
|
|
256
259
|
|
|
257
260
|
Supported Platforms:
|
|
258
261
|
``Ascend`` ``GPU`` ``CPU``
|
|
@@ -262,9 +265,9 @@ class Softmax(Cell):
|
|
|
262
265
|
>>> from mindspore import Tensor, nn
|
|
263
266
|
>>> import numpy as np
|
|
264
267
|
>>> # axis = -1(default), and the sum of return value is 1.0.
|
|
265
|
-
>>>
|
|
268
|
+
>>> input = Tensor(np.array([-1, -2, 0, 2, 1]), mindspore.float16)
|
|
266
269
|
>>> softmax = nn.Softmax()
|
|
267
|
-
>>> output = softmax(
|
|
270
|
+
>>> output = softmax(input)
|
|
268
271
|
>>> print(output)
|
|
269
272
|
[0.03168 0.01166 0.0861 0.636 0.2341 ]
|
|
270
273
|
"""
|
|
@@ -274,13 +277,42 @@ class Softmax(Cell):
|
|
|
274
277
|
super(Softmax, self).__init__()
|
|
275
278
|
self.softmax = P.Softmax(axis)
|
|
276
279
|
|
|
277
|
-
def construct(self,
|
|
278
|
-
return self.softmax(
|
|
280
|
+
def construct(self, input):
|
|
281
|
+
return self.softmax(input)
|
|
282
|
+
|
|
283
|
+
|
|
284
|
+
class SoftmaxExt(Cell):
|
|
285
|
+
r"""
|
|
286
|
+
Applies the Softmax function to an n-dimensional input Tensor.
|
|
287
|
+
|
|
288
|
+
For details, please refer to :func:`mindspore.mint.nn.functional.softmax`.
|
|
289
|
+
|
|
290
|
+
Supported Platforms:
|
|
291
|
+
``Ascend``
|
|
292
|
+
|
|
293
|
+
Examples:
|
|
294
|
+
>>> import mindspore
|
|
295
|
+
>>> from mindspore import Tensor, nn
|
|
296
|
+
>>> import numpy as np
|
|
297
|
+
>>> input = Tensor(np.array([-1, -2, 0, 2, 1]), mindspore.float16)
|
|
298
|
+
>>> softmax = nn.SoftmaxExt()
|
|
299
|
+
>>> output = softmax(input)
|
|
300
|
+
>>> print(output)
|
|
301
|
+
[0.03168 0.01166 0.0861 0.636 0.2341 ]
|
|
302
|
+
"""
|
|
303
|
+
|
|
304
|
+
def __init__(self, dim=None):
|
|
305
|
+
"""Initialize Softmax."""
|
|
306
|
+
super(SoftmaxExt, self).__init__()
|
|
307
|
+
self.dim = dim
|
|
308
|
+
|
|
309
|
+
def construct(self, input):
|
|
310
|
+
return ops.function.nn_func.softmax_ext(input, self.dim)
|
|
279
311
|
|
|
280
312
|
|
|
281
313
|
class LogSoftmax(Cell):
|
|
282
314
|
r"""
|
|
283
|
-
Applies the LogSoftmax function to n-dimensional input tensor.
|
|
315
|
+
Applies the LogSoftmax function to n-dimensional input tensor element-wise.
|
|
284
316
|
|
|
285
317
|
The input is transformed by the Softmax function and then by the log function to lie in range[-inf,0).
|
|
286
318
|
|
|
@@ -328,11 +360,55 @@ class LogSoftmax(Cell):
|
|
|
328
360
|
return self.log_softmax(x)
|
|
329
361
|
|
|
330
362
|
|
|
331
|
-
class
|
|
363
|
+
class LogSoftmaxExt(Cell):
|
|
332
364
|
r"""
|
|
333
|
-
|
|
365
|
+
Applies the Log Softmax function to the input tensor on the specified axis.
|
|
366
|
+
Supposes a slice in the given axis, :math:`x` for each element :math:`x_i`,
|
|
367
|
+
the Log Softmax function is shown as follows:
|
|
368
|
+
|
|
369
|
+
.. math::
|
|
370
|
+
\text{output}(x_i) = \log \left(\frac{\exp(x_i)} {\sum_{j = 0}^{N-1}\exp(x_j)}\right),
|
|
371
|
+
|
|
372
|
+
where :math:`N` is the length of the Tensor.
|
|
373
|
+
|
|
374
|
+
Args:
|
|
375
|
+
dim (int, optional): The axis to perform the Log softmax operation. Default: ``None`` .
|
|
376
|
+
|
|
377
|
+
Returns:
|
|
378
|
+
Tensor, with the same shape as the input.
|
|
379
|
+
|
|
380
|
+
Raises:
|
|
381
|
+
ValueError: If `dim` is not in range [-len(input.shape), len(input.shape)).
|
|
382
|
+
|
|
383
|
+
Supported Platforms:
|
|
384
|
+
``Ascend``
|
|
385
|
+
|
|
386
|
+
Examples:
|
|
387
|
+
>>> import mindspore
|
|
388
|
+
>>> from mindspore import Tensor, nn
|
|
389
|
+
>>> import numpy as np
|
|
390
|
+
>>> x = Tensor(np.array([[-1.0, 4.0, -8.0], [2.0, -5.0, 9.0]]), mindspore.float32)
|
|
391
|
+
>>> log_softmax = nn.LogSoftmaxExt(dim=-1)
|
|
392
|
+
>>> output = log_softmax(x)
|
|
393
|
+
>>> print(output)
|
|
394
|
+
[[-5.00672150e+00 -6.72150636e-03 -1.20067215e+01]
|
|
395
|
+
[-7.00091219e+00 -1.40009127e+01 -9.12250078e-04]]
|
|
396
|
+
"""
|
|
334
397
|
|
|
398
|
+
def __init__(self, dim=None):
|
|
399
|
+
"""Initialize LogSoftmaxExt."""
|
|
400
|
+
super(LogSoftmaxExt, self).__init__()
|
|
401
|
+
self.log_softmax = P.LogSoftmaxExt()
|
|
402
|
+
self.dim = dim
|
|
403
|
+
|
|
404
|
+
def construct(self, x):
|
|
405
|
+
return self.log_softmax(x, dim=self.dim)
|
|
406
|
+
|
|
407
|
+
|
|
408
|
+
class ELU(Cell):
|
|
409
|
+
r"""
|
|
335
410
|
Applies the exponential linear unit function element-wise.
|
|
411
|
+
|
|
336
412
|
The activation function is defined as:
|
|
337
413
|
|
|
338
414
|
.. math::
|
|
@@ -346,21 +422,22 @@ class ELU(Cell):
|
|
|
346
422
|
|
|
347
423
|
ELU Activation Function Graph:
|
|
348
424
|
|
|
349
|
-
.. image:: images/ELU.png
|
|
425
|
+
.. image:: ../images/ELU.png
|
|
350
426
|
:align: center
|
|
351
427
|
|
|
352
428
|
Args:
|
|
353
429
|
alpha (float): The alpha value of ELU, the data type is float. Default: ``1.0`` .
|
|
430
|
+
Only alpha equal to ``1.0`` is supported currently.
|
|
354
431
|
|
|
355
432
|
Inputs:
|
|
356
|
-
- **
|
|
433
|
+
- **input_x** (Tensor) - The input of ELU is a Tensor of any dimension with data type of float16 or float32.
|
|
357
434
|
|
|
358
435
|
Outputs:
|
|
359
|
-
Tensor, with the same type and shape as the `
|
|
436
|
+
Tensor, with the same type and shape as the `input_x`.
|
|
360
437
|
|
|
361
438
|
Raises:
|
|
362
439
|
TypeError: If `alpha` is not a float.
|
|
363
|
-
TypeError: If dtype of `
|
|
440
|
+
TypeError: If dtype of `input_x` is neither float16 nor float32.
|
|
364
441
|
ValueError: If `alpha` is not equal to 1.0.
|
|
365
442
|
|
|
366
443
|
Supported Platforms:
|
|
@@ -388,13 +465,13 @@ class ELU(Cell):
|
|
|
388
465
|
|
|
389
466
|
class ReLU(Cell):
|
|
390
467
|
r"""
|
|
391
|
-
Rectified Linear Unit activation function.
|
|
468
|
+
Applies ReLU (Rectified Linear Unit activation function) element-wise.
|
|
392
469
|
|
|
393
470
|
.. math::
|
|
394
471
|
|
|
395
|
-
\text{ReLU}(
|
|
472
|
+
\text{ReLU}(input) = (input)^+ = \max(0, input),
|
|
396
473
|
|
|
397
|
-
It returns element-wise :math:`\max(0,
|
|
474
|
+
It returns element-wise :math:`\max(0, input)`.
|
|
398
475
|
|
|
399
476
|
.. note::
|
|
400
477
|
The neurons with the negative output
|
|
@@ -402,29 +479,28 @@ class ReLU(Cell):
|
|
|
402
479
|
|
|
403
480
|
ReLU Activation Function Graph:
|
|
404
481
|
|
|
405
|
-
.. image:: images/ReLU.png
|
|
482
|
+
.. image:: ../images/ReLU.png
|
|
406
483
|
:align: center
|
|
407
484
|
|
|
408
485
|
Inputs:
|
|
409
|
-
- **
|
|
410
|
-
spore.cn/docs/en/r2.2/api_python/mindspore.html#mindspore.dtype>`_ .
|
|
486
|
+
- **input** (Tensor) - The input of ReLU is a Tensor of any dimension.
|
|
411
487
|
|
|
412
488
|
Outputs:
|
|
413
|
-
Tensor, with the same type and shape as the `
|
|
489
|
+
Tensor, with the same type and shape as the `input`.
|
|
414
490
|
|
|
415
491
|
Raises:
|
|
416
|
-
TypeError: If dtype of `
|
|
492
|
+
TypeError: If dtype of `input` is not supported.
|
|
417
493
|
|
|
418
494
|
Supported Platforms:
|
|
419
495
|
``Ascend`` ``GPU`` ``CPU``
|
|
420
496
|
|
|
421
497
|
Examples:
|
|
498
|
+
>>> import numpy as np
|
|
422
499
|
>>> import mindspore
|
|
423
500
|
>>> from mindspore import Tensor, nn
|
|
424
|
-
>>>
|
|
425
|
-
>>> x = Tensor(np.array([-1, 2, -3, 2, -1]), mindspore.float16)
|
|
501
|
+
>>> input = Tensor(np.array([-1, 2, -3, 2, -1]), mindspore.float16)
|
|
426
502
|
>>> relu = nn.ReLU()
|
|
427
|
-
>>> output = relu(
|
|
503
|
+
>>> output = relu(input)
|
|
428
504
|
>>> print(output)
|
|
429
505
|
[0. 2. 0. 2. 0.]
|
|
430
506
|
"""
|
|
@@ -434,13 +510,13 @@ class ReLU(Cell):
|
|
|
434
510
|
super(ReLU, self).__init__()
|
|
435
511
|
self.relu = P.ReLU()
|
|
436
512
|
|
|
437
|
-
def construct(self,
|
|
438
|
-
return self.relu(
|
|
513
|
+
def construct(self, input):
|
|
514
|
+
return self.relu(input)
|
|
439
515
|
|
|
440
516
|
|
|
441
517
|
class ReLU6(Cell):
|
|
442
518
|
r"""
|
|
443
|
-
Compute ReLU6 activation function.
|
|
519
|
+
Compute ReLU6 activation function element-wise.
|
|
444
520
|
|
|
445
521
|
ReLU6 is similar to ReLU with a upper limit of 6, which if the inputs are greater than 6, the outputs
|
|
446
522
|
will be suppressed to 6.
|
|
@@ -452,7 +528,7 @@ class ReLU6(Cell):
|
|
|
452
528
|
|
|
453
529
|
ReLU6 Activation Function Graph:
|
|
454
530
|
|
|
455
|
-
.. image:: images/ReLU6.png
|
|
531
|
+
.. image:: ../images/ReLU6.png
|
|
456
532
|
:align: center
|
|
457
533
|
|
|
458
534
|
Inputs:
|
|
@@ -505,7 +581,7 @@ class LeakyReLU(Cell):
|
|
|
505
581
|
|
|
506
582
|
LeakyReLU Activation Function Graph:
|
|
507
583
|
|
|
508
|
-
.. image:: images/LeakyReLU.png
|
|
584
|
+
.. image:: ../images/LeakyReLU.png
|
|
509
585
|
:align: center
|
|
510
586
|
|
|
511
587
|
Args:
|
|
@@ -547,8 +623,7 @@ class LeakyReLU(Cell):
|
|
|
547
623
|
|
|
548
624
|
class RReLU(Cell):
|
|
549
625
|
r"""
|
|
550
|
-
|
|
551
|
-
Randomized Leaky ReLU activation function.
|
|
626
|
+
Applies RReLU (Randomized Leaky ReLU activation function) element-wise.
|
|
552
627
|
|
|
553
628
|
The activation function is defined as:
|
|
554
629
|
|
|
@@ -575,7 +650,7 @@ class RReLU(Cell):
|
|
|
575
650
|
TypeError: If `lower` is not a float or an int.
|
|
576
651
|
TypeError: If `upper` is not a float or an int.
|
|
577
652
|
TypeError: If `x` is not a Tensor.
|
|
578
|
-
TypeError: If `x` is not a Tensor of mindspore.float16 or
|
|
653
|
+
TypeError: If `x` is not a Tensor of mindspore.float16 or mindspore.float32.
|
|
579
654
|
ValueError: If `lower` is greater than upper.
|
|
580
655
|
|
|
581
656
|
Supported Platforms:
|
|
@@ -621,11 +696,11 @@ class RReLU(Cell):
|
|
|
621
696
|
|
|
622
697
|
class SeLU(Cell):
|
|
623
698
|
r"""
|
|
624
|
-
|
|
699
|
+
Applies activation function SeLU (Scaled exponential Linear Unit) element-wise.
|
|
625
700
|
|
|
626
701
|
SeLU Activation Function Graph:
|
|
627
702
|
|
|
628
|
-
.. image:: images/SeLU.png
|
|
703
|
+
.. image:: ../images/SeLU.png
|
|
629
704
|
:align: center
|
|
630
705
|
|
|
631
706
|
Refer to :func:`mindspore.ops.selu` for more details.
|
|
@@ -656,15 +731,13 @@ class SeLU(Cell):
|
|
|
656
731
|
|
|
657
732
|
class SiLU(Cell):
|
|
658
733
|
r"""
|
|
659
|
-
Sigmoid Linear Unit activation function.
|
|
660
|
-
|
|
661
734
|
Applies the silu linear unit function element-wise.
|
|
662
735
|
|
|
663
736
|
.. math::
|
|
664
737
|
|
|
665
738
|
\text{SiLU}(x) = x * \sigma(x),
|
|
666
739
|
|
|
667
|
-
where :math:`x_i` is input, :math:`\sigma(x)` is Sigmoid function.
|
|
740
|
+
where :math:`x_i` is an element of the input, :math:`\sigma(x)` is Sigmoid function.
|
|
668
741
|
|
|
669
742
|
.. math::
|
|
670
743
|
|
|
@@ -672,17 +745,18 @@ class SiLU(Cell):
|
|
|
672
745
|
|
|
673
746
|
SiLU Activation Function Graph:
|
|
674
747
|
|
|
675
|
-
.. image:: images/SiLU.png
|
|
748
|
+
.. image:: ../images/SiLU.png
|
|
676
749
|
:align: center
|
|
677
750
|
|
|
678
751
|
Inputs:
|
|
679
|
-
- **
|
|
752
|
+
- **input** (Tensor) - `input` is :math:`x` in the preceding formula.
|
|
753
|
+
Input with the data type float16 or float32. Tensor of any dimension.
|
|
680
754
|
|
|
681
755
|
Outputs:
|
|
682
|
-
Tensor, with the same type and shape as the `
|
|
756
|
+
Tensor, with the same type and shape as the `input`.
|
|
683
757
|
|
|
684
758
|
Raises:
|
|
685
|
-
TypeError: If dtype of `
|
|
759
|
+
TypeError: If dtype of `input` is neither float16 nor float32.
|
|
686
760
|
|
|
687
761
|
Supported Platforms:
|
|
688
762
|
``Ascend`` ``GPU`` ``CPU``
|
|
@@ -691,9 +765,9 @@ class SiLU(Cell):
|
|
|
691
765
|
>>> import mindspore
|
|
692
766
|
>>> from mindspore import Tensor, nn
|
|
693
767
|
>>> import numpy as np
|
|
694
|
-
>>>
|
|
768
|
+
>>> input = Tensor(np.array([-1, 2, -3, 2, -1]), mindspore.float16)
|
|
695
769
|
>>> silu = nn.SiLU()
|
|
696
|
-
>>> output = silu(
|
|
770
|
+
>>> output = silu(input)
|
|
697
771
|
>>> print(output)
|
|
698
772
|
[-0.269 1.762 -0.1423 1.762 -0.269]
|
|
699
773
|
"""
|
|
@@ -720,7 +794,7 @@ class Tanh(Cell):
|
|
|
720
794
|
|
|
721
795
|
Tanh Activation Function Graph:
|
|
722
796
|
|
|
723
|
-
.. image:: images/Tanh.png
|
|
797
|
+
.. image:: ../images/Tanh.png
|
|
724
798
|
:align: center
|
|
725
799
|
|
|
726
800
|
Inputs:
|
|
@@ -757,9 +831,7 @@ class Tanh(Cell):
|
|
|
757
831
|
|
|
758
832
|
class Tanhshrink(Cell):
|
|
759
833
|
r"""
|
|
760
|
-
Tanhshrink activation function.
|
|
761
|
-
|
|
762
|
-
The tanhshrink function is evaluated by element and returns a new tensor.
|
|
834
|
+
Applies Tanhshrink activation function element-wise and returns a new tensor.
|
|
763
835
|
|
|
764
836
|
Tanh function is defined as:
|
|
765
837
|
|
|
@@ -815,7 +887,7 @@ class Hardtanh(Cell):
|
|
|
815
887
|
|
|
816
888
|
Hardtanh Activation Function Graph:
|
|
817
889
|
|
|
818
|
-
.. image:: images/Hardtanh.png
|
|
890
|
+
.. image:: ../images/Hardtanh.png
|
|
819
891
|
:align: center
|
|
820
892
|
|
|
821
893
|
Note:
|
|
@@ -868,8 +940,6 @@ class Hardtanh(Cell):
|
|
|
868
940
|
|
|
869
941
|
class GELU(Cell):
|
|
870
942
|
r"""
|
|
871
|
-
Gaussian error linear unit activation function.
|
|
872
|
-
|
|
873
943
|
Applies GELU function to each element of the input. The input is a Tensor with any valid shape.
|
|
874
944
|
|
|
875
945
|
GELU is defined as:
|
|
@@ -883,7 +953,7 @@ class GELU(Cell):
|
|
|
883
953
|
|
|
884
954
|
GELU Activation Function Graph:
|
|
885
955
|
|
|
886
|
-
.. image:: images/GELU.png
|
|
956
|
+
.. image:: ../images/GELU.png
|
|
887
957
|
:align: center
|
|
888
958
|
|
|
889
959
|
Args:
|
|
@@ -898,14 +968,21 @@ class GELU(Cell):
|
|
|
898
968
|
:math:`x * P(X <= x) = 0.5 * x * (1 + erf(x / \sqrt(2)))`, where P(X) ~ N(0, 1).
|
|
899
969
|
|
|
900
970
|
Inputs:
|
|
901
|
-
- **x** (Tensor) - The input of GELU with data type of float16 or
|
|
971
|
+
- **x** (Tensor) - The input of GELU with data type of float16, float32, or float64.
|
|
902
972
|
The shape is :math:`(N,*)` where :math:`*` means, any number of additional dimensions.
|
|
903
973
|
|
|
904
974
|
Outputs:
|
|
905
975
|
Tensor, with the same type and shape as the `x`.
|
|
906
976
|
|
|
977
|
+
Note:
|
|
978
|
+
when calculating the input gradient of GELU with an input value of infinity, there are differences
|
|
979
|
+
in the output of the backward between ``Ascend`` and ``GPU``.
|
|
980
|
+
when x is -inf, the computation result of ``Ascend`` is 0, and the computation result of ``GPU`` is Nan.
|
|
981
|
+
when x is inf, the computation result of ``Ascend`` is dy, and the computation result of ``GPU`` is Nan.
|
|
982
|
+
In mathematical terms, the result of Ascend has higher precision.
|
|
983
|
+
|
|
907
984
|
Raises:
|
|
908
|
-
TypeError: If dtype of `x` is
|
|
985
|
+
TypeError: If dtype of `x` is not one of float16, float32, or float64.
|
|
909
986
|
|
|
910
987
|
Supported Platforms:
|
|
911
988
|
``Ascend`` ``GPU`` ``CPU``
|
|
@@ -942,8 +1019,6 @@ class GELU(Cell):
|
|
|
942
1019
|
|
|
943
1020
|
class FastGelu(Cell):
|
|
944
1021
|
r"""
|
|
945
|
-
Fast Gaussian error linear unit activation function.
|
|
946
|
-
|
|
947
1022
|
Applies FastGelu function to each element of the input. The input is a Tensor with any valid shape.
|
|
948
1023
|
|
|
949
1024
|
FastGelu is defined as:
|
|
@@ -956,7 +1031,7 @@ class FastGelu(Cell):
|
|
|
956
1031
|
|
|
957
1032
|
FastGelu Activation Function Graph:
|
|
958
1033
|
|
|
959
|
-
.. image:: images/FastGelu.png
|
|
1034
|
+
.. image:: ../images/FastGelu.png
|
|
960
1035
|
:align: center
|
|
961
1036
|
|
|
962
1037
|
Inputs:
|
|
@@ -995,9 +1070,7 @@ class FastGelu(Cell):
|
|
|
995
1070
|
|
|
996
1071
|
class Sigmoid(Cell):
|
|
997
1072
|
r"""
|
|
998
|
-
|
|
999
|
-
|
|
1000
|
-
Applies sigmoid-type activation element-wise.
|
|
1073
|
+
Applies sigmoid activation function element-wise.
|
|
1001
1074
|
|
|
1002
1075
|
Sigmoid function is defined as:
|
|
1003
1076
|
|
|
@@ -1005,23 +1078,23 @@ class Sigmoid(Cell):
|
|
|
1005
1078
|
|
|
1006
1079
|
\text{sigmoid}(x_i) = \frac{1}{1 + \exp(-x_i)},
|
|
1007
1080
|
|
|
1008
|
-
where :math:`x_i` is the element of
|
|
1081
|
+
where :math:`x_i` is the element of `x`.
|
|
1009
1082
|
|
|
1010
1083
|
Sigmoid Activation Function Graph:
|
|
1011
1084
|
|
|
1012
|
-
.. image:: images/Sigmoid.png
|
|
1085
|
+
.. image:: ../images/Sigmoid.png
|
|
1013
1086
|
:align: center
|
|
1014
1087
|
|
|
1015
1088
|
Inputs:
|
|
1016
|
-
- **
|
|
1017
|
-
float16, float32, float64, complex64 or complex128.
|
|
1089
|
+
- **input** (Tensor) - `input` is :math:`x` in the preceding formula. Tensor of any dimension,
|
|
1090
|
+
the data type is float16, float32, float64, complex64 or complex128.
|
|
1018
1091
|
|
|
1019
1092
|
Outputs:
|
|
1020
|
-
Tensor, with the same type and shape as the `
|
|
1093
|
+
Tensor, with the same type and shape as the `input`.
|
|
1021
1094
|
|
|
1022
1095
|
Raises:
|
|
1023
|
-
TypeError: If dtype of `
|
|
1024
|
-
TypeError: If `
|
|
1096
|
+
TypeError: If dtype of `input` is not float16, float32, float64, complex64 or complex128.
|
|
1097
|
+
TypeError: If `input` is not a Tensor.
|
|
1025
1098
|
|
|
1026
1099
|
Supported Platforms:
|
|
1027
1100
|
``Ascend`` ``GPU`` ``CPU``
|
|
@@ -1030,9 +1103,9 @@ class Sigmoid(Cell):
|
|
|
1030
1103
|
>>> import mindspore
|
|
1031
1104
|
>>> from mindspore import Tensor, nn
|
|
1032
1105
|
>>> import numpy as np
|
|
1033
|
-
>>>
|
|
1106
|
+
>>> input = Tensor(np.array([-1, -2, 0, 2, 1]), mindspore.float16)
|
|
1034
1107
|
>>> sigmoid = nn.Sigmoid()
|
|
1035
|
-
>>> output = sigmoid(
|
|
1108
|
+
>>> output = sigmoid(input)
|
|
1036
1109
|
>>> print(output)
|
|
1037
1110
|
[0.2688 0.11914 0.5 0.881 0.7305 ]
|
|
1038
1111
|
"""
|
|
@@ -1048,11 +1121,11 @@ class Sigmoid(Cell):
|
|
|
1048
1121
|
|
|
1049
1122
|
class Softsign(Cell):
|
|
1050
1123
|
r"""
|
|
1051
|
-
|
|
1124
|
+
Applies softsign activation function element-wise.
|
|
1052
1125
|
|
|
1053
1126
|
Softsign Activation Function Graph:
|
|
1054
1127
|
|
|
1055
|
-
.. image:: images/Softsign.png
|
|
1128
|
+
.. image:: ../images/Softsign.png
|
|
1056
1129
|
:align: center
|
|
1057
1130
|
|
|
1058
1131
|
Refer to :func:`mindspore.ops.softsign` for more details.
|
|
@@ -1082,9 +1155,7 @@ class Softsign(Cell):
|
|
|
1082
1155
|
|
|
1083
1156
|
class PReLU(Cell):
|
|
1084
1157
|
r"""
|
|
1085
|
-
PReLU activation function.
|
|
1086
|
-
|
|
1087
|
-
Applies the PReLU function element-wise.
|
|
1158
|
+
Applies PReLU activation function element-wise.
|
|
1088
1159
|
|
|
1089
1160
|
PReLU is defined as:
|
|
1090
1161
|
|
|
@@ -1100,7 +1171,7 @@ class PReLU(Cell):
|
|
|
1100
1171
|
|
|
1101
1172
|
PReLU Activation Function Graph:
|
|
1102
1173
|
|
|
1103
|
-
.. image:: images/PReLU.png
|
|
1174
|
+
.. image:: ../images/PReLU.png
|
|
1104
1175
|
:align: center
|
|
1105
1176
|
|
|
1106
1177
|
Args:
|
|
@@ -1176,29 +1247,100 @@ class PReLU(Cell):
|
|
|
1176
1247
|
return self.prelu(x, F.cast(self.w, x.dtype))
|
|
1177
1248
|
|
|
1178
1249
|
|
|
1250
|
+
class PReLUExt(Cell):
|
|
1251
|
+
r"""
|
|
1252
|
+
Applies PReLU activation function element-wise.
|
|
1253
|
+
|
|
1254
|
+
PReLU is defined as:
|
|
1255
|
+
|
|
1256
|
+
.. math::
|
|
1257
|
+
|
|
1258
|
+
PReLU(x_i)= \max(0, x_i) + w * \min(0, x_i),
|
|
1259
|
+
|
|
1260
|
+
where :math:`x_i` is an element of an channel of the input.
|
|
1261
|
+
|
|
1262
|
+
Here :math:`w` is a learnable parameter with a default initial value 0.25.
|
|
1263
|
+
Parameter :math:`w` has dimensionality of the argument channel. If called without argument
|
|
1264
|
+
channel, a single parameter :math:`w` will be shared across all channels.
|
|
1265
|
+
|
|
1266
|
+
PReLU Activation Function Graph:
|
|
1267
|
+
|
|
1268
|
+
.. image:: ../images/PReLU2.png
|
|
1269
|
+
:align: center
|
|
1270
|
+
|
|
1271
|
+
.. note::
|
|
1272
|
+
Channel dim is the 2nd dim of input. When input has dims < 2, then there is
|
|
1273
|
+
no channel dim and the number of channels = 1.
|
|
1274
|
+
|
|
1275
|
+
Args:
|
|
1276
|
+
num_parameters (int): number of `w` to learn. Although it takes an int as input,
|
|
1277
|
+
there is only two legitimate values: 1, or the number of channels at Tensor `input`. Default: ``1`` .
|
|
1278
|
+
init (float): the initial value of `w`. Default: ``0.25`` .
|
|
1279
|
+
dtype (mindspore.dtype, optional): the type of `w`. Default: ``None`` . Supported data type
|
|
1280
|
+
is {float16, float32, bfloat16}.
|
|
1281
|
+
|
|
1282
|
+
Inputs:
|
|
1283
|
+
- **input** (Tensor) - The input of PReLU.
|
|
1284
|
+
|
|
1285
|
+
Outputs:
|
|
1286
|
+
Tensor, with the same dtype and shape as the `input`.
|
|
1287
|
+
|
|
1288
|
+
Supported Platforms:
|
|
1289
|
+
``Ascend``
|
|
1290
|
+
|
|
1291
|
+
Examples:
|
|
1292
|
+
>>> import mindspore
|
|
1293
|
+
>>> from mindspore import Tensor, nn
|
|
1294
|
+
>>> import numpy as np
|
|
1295
|
+
>>> x = Tensor(np.array([[[[0.1, 0.6], [0.9, 0.9]]]]), mindspore.float32)
|
|
1296
|
+
>>> prelu = nn.PReLUExt()
|
|
1297
|
+
>>> output = prelu(x)
|
|
1298
|
+
>>> print(output)
|
|
1299
|
+
[[[[0.1 0.6]
|
|
1300
|
+
[0.9 0.9]]]]
|
|
1301
|
+
|
|
1302
|
+
"""
|
|
1303
|
+
|
|
1304
|
+
def __init__(self, num_parameters=1, init=0.25, dtype=None):
|
|
1305
|
+
"""Initialize PReLUExt."""
|
|
1306
|
+
super(PReLUExt, self).__init__()
|
|
1307
|
+
tmp = np.empty((num_parameters,), dtype=np.float32)
|
|
1308
|
+
tmp.fill(init)
|
|
1309
|
+
w = Tensor(tmp, dtype=dtype)
|
|
1310
|
+
self.weight = Parameter(w, name='weight')
|
|
1311
|
+
|
|
1312
|
+
def construct(self, input):
|
|
1313
|
+
return ops.prelu(input, self.weight)
|
|
1314
|
+
|
|
1315
|
+
|
|
1179
1316
|
class HSwish(Cell):
|
|
1180
1317
|
r"""
|
|
1181
|
-
Applies
|
|
1318
|
+
Applies Hard Swish activation function element-wise.
|
|
1182
1319
|
|
|
1183
1320
|
Hard swish is defined as:
|
|
1184
1321
|
|
|
1185
1322
|
.. math::
|
|
1186
|
-
\text{
|
|
1323
|
+
\text{Hardswish}(input) =
|
|
1324
|
+
\begin{cases}
|
|
1325
|
+
0, & \text{ if } input \leq -3, \\
|
|
1326
|
+
input, & \text{ if } input \geq +3, \\
|
|
1327
|
+
input*(input + 3)/6, & \text{ otherwise }
|
|
1328
|
+
\end{cases}
|
|
1187
1329
|
|
|
1188
1330
|
HSwish Activation Function Graph:
|
|
1189
1331
|
|
|
1190
|
-
.. image:: images/HSwish.png
|
|
1332
|
+
.. image:: ../images/HSwish.png
|
|
1191
1333
|
:align: center
|
|
1192
1334
|
|
|
1193
1335
|
Inputs:
|
|
1194
|
-
- **
|
|
1195
|
-
The shape is :math:`(N,*)` where :math:`*` means, any number of additional dimensions.
|
|
1336
|
+
- **input** (Tensor) - The input of HSwish.
|
|
1196
1337
|
|
|
1197
1338
|
Outputs:
|
|
1198
|
-
Tensor, with the same type and shape as the `
|
|
1339
|
+
Tensor, with the same type and shape as the `input`.
|
|
1199
1340
|
|
|
1200
1341
|
Raises:
|
|
1201
|
-
TypeError: If
|
|
1342
|
+
TypeError: If `input` is not a tensor.
|
|
1343
|
+
TypeError: If `input` is neither int nor float.
|
|
1202
1344
|
|
|
1203
1345
|
Supported Platforms:
|
|
1204
1346
|
``Ascend`` ``GPU`` ``CPU``
|
|
@@ -1207,9 +1349,9 @@ class HSwish(Cell):
|
|
|
1207
1349
|
>>> import mindspore
|
|
1208
1350
|
>>> from mindspore import Tensor, nn
|
|
1209
1351
|
>>> import numpy as np
|
|
1210
|
-
>>>
|
|
1352
|
+
>>> input = Tensor(np.array([-1, -2, 0, 2, 1]), mindspore.float16)
|
|
1211
1353
|
>>> hswish = nn.HSwish()
|
|
1212
|
-
>>> result = hswish(
|
|
1354
|
+
>>> result = hswish(input)
|
|
1213
1355
|
>>> print(result)
|
|
1214
1356
|
[-0.3333 -0.3333 0. 1.667 0.6665]
|
|
1215
1357
|
"""
|
|
@@ -1219,32 +1361,38 @@ class HSwish(Cell):
|
|
|
1219
1361
|
super(HSwish, self).__init__()
|
|
1220
1362
|
self.hswish = P.HSwish()
|
|
1221
1363
|
|
|
1222
|
-
def construct(self,
|
|
1223
|
-
return self.hswish(
|
|
1364
|
+
def construct(self, input):
|
|
1365
|
+
return self.hswish(input)
|
|
1224
1366
|
|
|
1225
1367
|
|
|
1226
1368
|
class HSigmoid(Cell):
|
|
1227
1369
|
r"""
|
|
1228
|
-
Hard
|
|
1370
|
+
Applies Hard Sigmoid activation function element-wise.
|
|
1229
1371
|
|
|
1230
|
-
Hard
|
|
1372
|
+
Hard Sigmoid is defined as:
|
|
1231
1373
|
|
|
1232
1374
|
.. math::
|
|
1233
|
-
\text{
|
|
1375
|
+
\text{Hardsigmoid}(input) =
|
|
1376
|
+
\begin{cases}
|
|
1377
|
+
0, & \text{ if } input \leq -3, \\
|
|
1378
|
+
1, & \text{ if } input \geq +3, \\
|
|
1379
|
+
input/6 + 1/2, & \text{ otherwise }
|
|
1380
|
+
\end{cases}
|
|
1234
1381
|
|
|
1235
1382
|
HSigmoid Activation Function Graph:
|
|
1236
1383
|
|
|
1237
|
-
.. image:: images/HSigmoid.png
|
|
1384
|
+
.. image:: ../images/HSigmoid.png
|
|
1238
1385
|
:align: center
|
|
1239
1386
|
|
|
1240
1387
|
Inputs:
|
|
1241
|
-
- **
|
|
1388
|
+
- **input** (Tensor) - The input of HSigmoid.
|
|
1242
1389
|
|
|
1243
1390
|
Outputs:
|
|
1244
|
-
Tensor, with the same type and shape as the `
|
|
1391
|
+
Tensor, with the same type and shape as the `input`.
|
|
1245
1392
|
|
|
1246
1393
|
Raises:
|
|
1247
|
-
TypeError: If `
|
|
1394
|
+
TypeError: If `input` is not a Tensor.
|
|
1395
|
+
TypeError: If `input` is neither int nor float.
|
|
1248
1396
|
|
|
1249
1397
|
Supported Platforms:
|
|
1250
1398
|
``Ascend`` ``GPU`` ``CPU``
|
|
@@ -1253,9 +1401,9 @@ class HSigmoid(Cell):
|
|
|
1253
1401
|
>>> import mindspore
|
|
1254
1402
|
>>> from mindspore import Tensor, nn
|
|
1255
1403
|
>>> import numpy as np
|
|
1256
|
-
>>>
|
|
1404
|
+
>>> input = Tensor(np.array([-1, -2, 0, 2, 1]), mindspore.float16)
|
|
1257
1405
|
>>> hsigmoid = nn.HSigmoid()
|
|
1258
|
-
>>> result = hsigmoid(
|
|
1406
|
+
>>> result = hsigmoid(input)
|
|
1259
1407
|
>>> print(result)
|
|
1260
1408
|
[0.3333 0.1666 0.5 0.8335 0.6665]
|
|
1261
1409
|
"""
|
|
@@ -1265,8 +1413,8 @@ class HSigmoid(Cell):
|
|
|
1265
1413
|
super(HSigmoid, self).__init__()
|
|
1266
1414
|
self.hsigmoid = P.HSigmoid()
|
|
1267
1415
|
|
|
1268
|
-
def construct(self,
|
|
1269
|
-
return self.hsigmoid(
|
|
1416
|
+
def construct(self, input):
|
|
1417
|
+
return self.hsigmoid(input)
|
|
1270
1418
|
|
|
1271
1419
|
|
|
1272
1420
|
class LogSigmoid(Cell):
|
|
@@ -1282,7 +1430,7 @@ class LogSigmoid(Cell):
|
|
|
1282
1430
|
|
|
1283
1431
|
LogSigmoid Activation Function Graph:
|
|
1284
1432
|
|
|
1285
|
-
.. image:: images/LogSigmoid.png
|
|
1433
|
+
.. image:: ../images/LogSigmoid.png
|
|
1286
1434
|
:align: center
|
|
1287
1435
|
|
|
1288
1436
|
Inputs:
|
|
@@ -1378,25 +1526,26 @@ class SoftShrink(Cell):
|
|
|
1378
1526
|
|
|
1379
1527
|
SoftShrink Activation Function Graph:
|
|
1380
1528
|
|
|
1381
|
-
.. image:: images/Softshrink.png
|
|
1529
|
+
.. image:: ../images/Softshrink.png
|
|
1382
1530
|
:align: center
|
|
1383
1531
|
|
|
1384
1532
|
Args:
|
|
1385
|
-
lambd (
|
|
1386
|
-
|
|
1533
|
+
lambd (number, optional): The threshold :math:`\lambda` defined by the Soft Shrink formula.
|
|
1534
|
+
It should be greater than or equal to 0, default: ``0.5`` .
|
|
1387
1535
|
|
|
1388
1536
|
Inputs:
|
|
1389
|
-
- **
|
|
1390
|
-
|
|
1537
|
+
- **input** (Tensor) - The input of Soft Shrink. Supported dtypes:
|
|
1538
|
+
|
|
1539
|
+
- Ascend: float16, float32, bfloat16.
|
|
1540
|
+
- CPU/GPU: float16, float32.
|
|
1391
1541
|
|
|
1392
1542
|
Outputs:
|
|
1393
|
-
Tensor,
|
|
1543
|
+
Tensor, the same shape and data type as the input.
|
|
1394
1544
|
|
|
1395
1545
|
Raises:
|
|
1396
|
-
TypeError: If lambd is not a float.
|
|
1397
|
-
TypeError: If
|
|
1398
|
-
TypeError: If dtype of
|
|
1399
|
-
ValueError: If lambd is less than 0.
|
|
1546
|
+
TypeError: If `lambd` is not a float, int or bool.
|
|
1547
|
+
TypeError: If `input` is not a tensor.
|
|
1548
|
+
TypeError: If dtype of `input` is not float16, float32 or bfloat16.
|
|
1400
1549
|
|
|
1401
1550
|
Supported Platforms:
|
|
1402
1551
|
``Ascend`` ``GPU`` ``CPU``
|
|
@@ -1405,9 +1554,9 @@ class SoftShrink(Cell):
|
|
|
1405
1554
|
>>> import mindspore
|
|
1406
1555
|
>>> from mindspore import Tensor, nn
|
|
1407
1556
|
>>> import numpy as np
|
|
1408
|
-
>>>
|
|
1557
|
+
>>> input = Tensor(np.array([[ 0.5297, 0.7871, 1.1754], [ 0.7836, 0.6218, -1.1542]]), mindspore.float16)
|
|
1409
1558
|
>>> softshrink = nn.SoftShrink()
|
|
1410
|
-
>>> output = softshrink(
|
|
1559
|
+
>>> output = softshrink(input)
|
|
1411
1560
|
>>> print(output)
|
|
1412
1561
|
[[ 0.02979 0.287 0.676 ]
|
|
1413
1562
|
[ 0.2837 0.1216 -0.6543 ]]
|
|
@@ -1417,14 +1566,14 @@ class SoftShrink(Cell):
|
|
|
1417
1566
|
super(SoftShrink, self).__init__()
|
|
1418
1567
|
self.softshrink = P.SoftShrink(lambd)
|
|
1419
1568
|
|
|
1420
|
-
def construct(self,
|
|
1421
|
-
output = self.softshrink(
|
|
1569
|
+
def construct(self, input):
|
|
1570
|
+
output = self.softshrink(input)
|
|
1422
1571
|
return output
|
|
1423
1572
|
|
|
1424
1573
|
|
|
1425
1574
|
class HShrink(Cell):
|
|
1426
1575
|
r"""
|
|
1427
|
-
Hard Shrink activation function
|
|
1576
|
+
Applies Hard Shrink activation function element-wise.
|
|
1428
1577
|
|
|
1429
1578
|
The formula is defined as follows:
|
|
1430
1579
|
|
|
@@ -1438,21 +1587,25 @@ class HShrink(Cell):
|
|
|
1438
1587
|
|
|
1439
1588
|
HShrink Activation Function Graph:
|
|
1440
1589
|
|
|
1441
|
-
.. image:: images/HShrink.png
|
|
1590
|
+
.. image:: ../images/HShrink.png
|
|
1442
1591
|
:align: center
|
|
1443
1592
|
|
|
1444
1593
|
Args:
|
|
1445
|
-
lambd (
|
|
1594
|
+
lambd (number, optional): The threshold :math:`\lambda` defined by the Hard Shrink formula. Default: ``0.5`` .
|
|
1446
1595
|
|
|
1447
1596
|
Inputs:
|
|
1448
|
-
- **
|
|
1597
|
+
- **input** (Tensor) - The input of Hard Shrink. Supported dtypes:
|
|
1598
|
+
|
|
1599
|
+
- Ascend: float16, float32, bfloat16.
|
|
1600
|
+
- CPU/GPU: float16, float32.
|
|
1449
1601
|
|
|
1450
1602
|
Outputs:
|
|
1451
1603
|
Tensor, the same shape and data type as the input.
|
|
1452
1604
|
|
|
1453
1605
|
Raises:
|
|
1454
|
-
TypeError: If `lambd` is not a float.
|
|
1455
|
-
TypeError: If
|
|
1606
|
+
TypeError: If `lambd` is not a float, int or bool.
|
|
1607
|
+
TypeError: If `input` is not a tensor.
|
|
1608
|
+
TypeError: If dtype of `input` is not float16, float32 or bfloat16.
|
|
1456
1609
|
|
|
1457
1610
|
Supported Platforms:
|
|
1458
1611
|
``Ascend`` ``GPU`` ``CPU``
|
|
@@ -1461,24 +1614,25 @@ class HShrink(Cell):
|
|
|
1461
1614
|
>>> import mindspore
|
|
1462
1615
|
>>> from mindspore import Tensor, nn
|
|
1463
1616
|
>>> import numpy as np
|
|
1464
|
-
>>>
|
|
1617
|
+
>>> input = Tensor(np.array([[0.5, 1, 2.0], [0.0533, 0.0776, -2.1233]]), mindspore.float32)
|
|
1465
1618
|
>>> hshrink = nn.HShrink()
|
|
1466
|
-
>>> output = hshrink(
|
|
1619
|
+
>>> output = hshrink(input)
|
|
1467
1620
|
>>> print(output)
|
|
1468
1621
|
[[ 0. 1. 2. ]
|
|
1469
|
-
|
|
1622
|
+
[ 0. 0. -2.1233]]
|
|
1470
1623
|
"""
|
|
1471
1624
|
|
|
1472
1625
|
def __init__(self, lambd=0.5):
|
|
1473
1626
|
super(HShrink, self).__init__()
|
|
1474
1627
|
self.hshrink = P.HShrink(lambd)
|
|
1475
1628
|
|
|
1476
|
-
def construct(self,
|
|
1477
|
-
return self.hshrink(
|
|
1629
|
+
def construct(self, input):
|
|
1630
|
+
return self.hshrink(input)
|
|
1478
1631
|
|
|
1479
1632
|
|
|
1480
1633
|
class Threshold(Cell):
|
|
1481
|
-
r"""
|
|
1634
|
+
r"""
|
|
1635
|
+
Thresholds each element of the input Tensor.
|
|
1482
1636
|
|
|
1483
1637
|
The formula is defined as follows:
|
|
1484
1638
|
|
|
@@ -1497,7 +1651,7 @@ class Threshold(Cell):
|
|
|
1497
1651
|
- **input_x** (Tensor) - The input of Threshold with data type of float16 or float32.
|
|
1498
1652
|
|
|
1499
1653
|
Outputs:
|
|
1500
|
-
Tensor, the same shape and data type as the
|
|
1654
|
+
Tensor, the same shape and data type as the `input_x`.
|
|
1501
1655
|
|
|
1502
1656
|
Raises:
|
|
1503
1657
|
TypeError: If `threshold` is not a float or an int.
|
|
@@ -1528,13 +1682,14 @@ class Threshold(Cell):
|
|
|
1528
1682
|
|
|
1529
1683
|
class Mish(Cell):
|
|
1530
1684
|
r"""
|
|
1531
|
-
Computes MISH(A Self Regularized Non-Monotonic Neural Activation Function)
|
|
1685
|
+
Computes MISH (A Self Regularized Non-Monotonic Neural Activation Function)
|
|
1686
|
+
of input tensors element-wise.
|
|
1532
1687
|
|
|
1533
1688
|
Refer to :func:`mindspore.ops.mish` for more details.
|
|
1534
1689
|
|
|
1535
1690
|
Mish Activation Function Graph:
|
|
1536
1691
|
|
|
1537
|
-
.. image:: images/Mish.png
|
|
1692
|
+
.. image:: ../images/Mish.png
|
|
1538
1693
|
:align: center
|
|
1539
1694
|
|
|
1540
1695
|
Supported Platforms:
|
|
@@ -1612,6 +1767,7 @@ _activation = {
|
|
|
1612
1767
|
'softmax': Softmax,
|
|
1613
1768
|
'softmax2d': Softmax2d,
|
|
1614
1769
|
'logsoftmax': LogSoftmax,
|
|
1770
|
+
'logsoftmaxExt': LogSoftmaxExt,
|
|
1615
1771
|
'relu': ReLU,
|
|
1616
1772
|
'relu6': ReLU6,
|
|
1617
1773
|
'rrelu': RReLU,
|
|
@@ -1625,6 +1781,7 @@ _activation = {
|
|
|
1625
1781
|
'sigmoid': Sigmoid,
|
|
1626
1782
|
'softsign': Softsign,
|
|
1627
1783
|
'prelu': PReLU,
|
|
1784
|
+
'preluExt': PReLUExt,
|
|
1628
1785
|
'leakyrelu': LeakyReLU,
|
|
1629
1786
|
'hswish': HSwish,
|
|
1630
1787
|
'hsigmoid': HSigmoid,
|