mindspore 2.2.11__cp39-cp39-win_amd64.whl → 2.3.0__cp39-cp39-win_amd64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of mindspore might be problematic. Click here for more details.
- mindspore/.commit_id +1 -1
- mindspore/__init__.py +7 -5
- mindspore/_c_dataengine.cp39-win_amd64.pyd +0 -0
- mindspore/_c_expression.cp39-win_amd64.pyd +0 -0
- mindspore/_c_mindrecord.cp39-win_amd64.pyd +0 -0
- mindspore/_checkparam.py +76 -18
- mindspore/_extends/builtin_operations.py +2 -1
- mindspore/_extends/graph_kernel/model/graph_parallel.py +16 -6
- mindspore/_extends/parallel_compile/akg_compiler/akg_process.py +3 -16
- mindspore/_extends/parallel_compile/akg_compiler/build_tbe_kernel.py +16 -4
- mindspore/_extends/parallel_compile/akg_compiler/compiler.py +1 -0
- mindspore/_extends/parallel_compile/akg_compiler/gen_custom_op_files.py +96 -0
- mindspore/_extends/parallel_compile/akg_compiler/tbe_topi.py +2 -1
- mindspore/_extends/parallel_compile/akg_compiler/util.py +5 -2
- mindspore/_extends/parse/__init__.py +18 -14
- mindspore/_extends/parse/compile_config.py +258 -0
- mindspore/_extends/parse/namespace.py +2 -2
- mindspore/_extends/parse/parser.py +174 -62
- mindspore/_extends/parse/resources.py +45 -14
- mindspore/_extends/parse/standard_method.py +142 -240
- mindspore/{ops/_op_impl/tbe/atomic_addr_clean.py → _extends/pijit/__init__.py} +6 -16
- mindspore/_extends/pijit/pijit_func_white_list.py +343 -0
- mindspore/_extends/remote/kernel_build_server.py +2 -0
- mindspore/_profiler.py +30 -0
- mindspore/amp.py +51 -24
- mindspore/avcodec-59.dll +0 -0
- mindspore/avdevice-59.dll +0 -0
- mindspore/avfilter-8.dll +0 -0
- mindspore/avformat-59.dll +0 -0
- mindspore/avutil-57.dll +0 -0
- mindspore/boost/adasum.py +1 -1
- mindspore/boost/base.py +1 -1
- mindspore/boost/boost_cell_wrapper.py +2 -2
- mindspore/boost/grad_freeze.py +2 -2
- mindspore/boost/group_loss_scale_manager.py +1 -1
- mindspore/boost/less_batch_normalization.py +9 -6
- mindspore/common/__init__.py +15 -4
- mindspore/common/_jit_fallback_utils.py +2 -3
- mindspore/common/_register_for_adapter.py +7 -0
- mindspore/common/_register_for_recompute.py +48 -0
- mindspore/common/_register_for_tensor.py +8 -9
- mindspore/common/_stub_tensor.py +7 -1
- mindspore/common/_utils.py +5 -17
- mindspore/common/api.py +411 -106
- mindspore/common/auto_dynamic_shape.py +27 -14
- mindspore/common/dtype.py +17 -10
- mindspore/common/dump.py +6 -8
- mindspore/common/file_system.py +48 -0
- mindspore/common/generator.py +260 -0
- mindspore/common/hook_handle.py +51 -4
- mindspore/common/initializer.py +1 -1
- mindspore/common/jit_config.py +34 -14
- mindspore/common/lazy_inline.py +72 -19
- mindspore/common/mindir_util.py +12 -2
- mindspore/common/mutable.py +79 -14
- mindspore/common/no_inline.py +54 -0
- mindspore/common/np_dtype.py +25 -0
- mindspore/common/parameter.py +30 -11
- mindspore/common/recompute.py +262 -0
- mindspore/common/seed.py +9 -9
- mindspore/common/sparse_tensor.py +272 -24
- mindspore/common/symbol.py +122 -0
- mindspore/common/tensor.py +468 -496
- mindspore/communication/__init__.py +6 -11
- mindspore/communication/_comm_helper.py +5 -0
- mindspore/communication/comm_func.py +1140 -0
- mindspore/communication/management.py +118 -102
- mindspore/config/op_info.config +22 -54
- mindspore/context.py +378 -65
- mindspore/dataset/__init__.py +5 -5
- mindspore/dataset/audio/__init__.py +6 -6
- mindspore/dataset/audio/transforms.py +711 -158
- mindspore/dataset/callback/ds_callback.py +2 -2
- mindspore/dataset/engine/cache_client.py +2 -2
- mindspore/dataset/engine/datasets.py +163 -83
- mindspore/dataset/engine/datasets_audio.py +14 -14
- mindspore/dataset/engine/datasets_standard_format.py +33 -3
- mindspore/dataset/engine/datasets_text.py +38 -38
- mindspore/dataset/engine/datasets_user_defined.py +78 -59
- mindspore/dataset/engine/datasets_vision.py +77 -73
- mindspore/dataset/engine/offload.py +5 -7
- mindspore/dataset/engine/queue.py +56 -38
- mindspore/dataset/engine/validators.py +11 -5
- mindspore/dataset/text/__init__.py +3 -3
- mindspore/dataset/text/transforms.py +408 -121
- mindspore/dataset/text/utils.py +9 -9
- mindspore/dataset/transforms/__init__.py +1 -1
- mindspore/dataset/transforms/transforms.py +261 -76
- mindspore/dataset/utils/browse_dataset.py +9 -9
- mindspore/dataset/vision/__init__.py +8 -8
- mindspore/dataset/vision/c_transforms.py +10 -10
- mindspore/dataset/vision/py_transforms_util.py +3 -3
- mindspore/dataset/vision/transforms.py +2844 -549
- mindspore/dataset/vision/utils.py +161 -10
- mindspore/dataset/vision/validators.py +14 -2
- mindspore/dnnl.dll +0 -0
- mindspore/experimental/optim/__init__.py +12 -2
- mindspore/experimental/optim/adadelta.py +161 -0
- mindspore/experimental/optim/adagrad.py +168 -0
- mindspore/experimental/optim/adam.py +35 -34
- mindspore/experimental/optim/adamax.py +170 -0
- mindspore/experimental/optim/adamw.py +40 -16
- mindspore/experimental/optim/asgd.py +153 -0
- mindspore/experimental/optim/lr_scheduler.py +71 -127
- mindspore/experimental/optim/nadam.py +157 -0
- mindspore/experimental/optim/optimizer.py +15 -8
- mindspore/experimental/optim/radam.py +194 -0
- mindspore/experimental/optim/rmsprop.py +154 -0
- mindspore/experimental/optim/rprop.py +164 -0
- mindspore/experimental/optim/sgd.py +28 -19
- mindspore/hal/__init__.py +40 -0
- mindspore/hal/_ascend.py +57 -0
- mindspore/hal/_base.py +57 -0
- mindspore/hal/_cpu.py +56 -0
- mindspore/hal/_gpu.py +57 -0
- mindspore/hal/device.py +356 -0
- mindspore/hal/event.py +179 -0
- mindspore/hal/memory.py +326 -0
- mindspore/hal/stream.py +339 -0
- mindspore/include/api/data_type.h +2 -2
- mindspore/include/api/dual_abi_helper.h +16 -3
- mindspore/include/api/model.h +4 -3
- mindspore/include/api/status.h +14 -0
- mindspore/include/c_api/model_c.h +173 -0
- mindspore/include/c_api/ms/base/types.h +1 -0
- mindspore/include/c_api/types_c.h +19 -0
- mindspore/include/dataset/execute.h +1 -3
- mindspore/include/dataset/vision.h +54 -2
- mindspore/jpeg62.dll +0 -0
- mindspore/log.py +2 -2
- mindspore/mindrecord/__init__.py +5 -1
- mindspore/mindrecord/config.py +809 -0
- mindspore/mindrecord/filereader.py +25 -0
- mindspore/mindrecord/filewriter.py +76 -58
- mindspore/mindrecord/mindpage.py +40 -6
- mindspore/mindrecord/shardutils.py +3 -2
- mindspore/mindrecord/shardwriter.py +7 -0
- mindspore/mindrecord/tools/cifar100_to_mr.py +53 -66
- mindspore/mindrecord/tools/cifar10_to_mr.py +48 -63
- mindspore/mindrecord/tools/csv_to_mr.py +7 -17
- mindspore/mindrecord/tools/imagenet_to_mr.py +3 -8
- mindspore/mindrecord/tools/mnist_to_mr.py +11 -21
- mindspore/mindrecord/tools/tfrecord_to_mr.py +2 -10
- mindspore/mindspore_backend.dll +0 -0
- mindspore/mindspore_common.dll +0 -0
- mindspore/mindspore_core.dll +0 -0
- mindspore/mindspore_glog.dll +0 -0
- mindspore/mindspore_np_dtype.dll +0 -0
- mindspore/mindspore_shared_lib.dll +0 -0
- mindspore/mint/__init__.py +1137 -0
- mindspore/{rewrite/ast_transformers → mint/linalg}/__init__.py +9 -4
- mindspore/mint/nn/__init__.py +512 -0
- mindspore/mint/nn/functional.py +573 -0
- mindspore/mint/optim/__init__.py +24 -0
- mindspore/mint/optim/adamw.py +185 -0
- mindspore/multiprocessing/__init__.py +72 -0
- mindspore/nn/__init__.py +1 -0
- mindspore/nn/cell.py +213 -257
- mindspore/nn/dynamic_lr.py +2 -2
- mindspore/nn/extend/__init__.py +29 -0
- mindspore/nn/extend/basic.py +140 -0
- mindspore/nn/extend/embedding.py +143 -0
- mindspore/{rewrite/ast_creator_register.py → nn/extend/layer/__init__.py} +9 -19
- mindspore/nn/extend/layer/normalization.py +109 -0
- mindspore/nn/extend/pooling.py +117 -0
- mindspore/nn/layer/activation.py +84 -94
- mindspore/nn/layer/basic.py +177 -82
- mindspore/nn/layer/channel_shuffle.py +3 -16
- mindspore/nn/layer/container.py +3 -3
- mindspore/nn/layer/conv.py +75 -66
- mindspore/nn/layer/embedding.py +103 -45
- mindspore/nn/layer/embedding_service.py +531 -0
- mindspore/nn/layer/embedding_service_layer.py +393 -0
- mindspore/nn/layer/image.py +4 -7
- mindspore/nn/layer/math.py +1 -1
- mindspore/nn/layer/normalization.py +52 -66
- mindspore/nn/layer/padding.py +30 -39
- mindspore/nn/layer/pooling.py +18 -9
- mindspore/nn/layer/rnn_cells.py +6 -16
- mindspore/nn/layer/rnns.py +6 -5
- mindspore/nn/layer/thor_layer.py +1 -2
- mindspore/nn/layer/timedistributed.py +1 -1
- mindspore/nn/layer/transformer.py +52 -50
- mindspore/nn/learning_rate_schedule.py +6 -5
- mindspore/nn/loss/loss.py +63 -84
- mindspore/nn/optim/ada_grad.py +6 -4
- mindspore/nn/optim/adadelta.py +3 -1
- mindspore/nn/optim/adafactor.py +1 -1
- mindspore/nn/optim/adam.py +102 -181
- mindspore/nn/optim/adamax.py +4 -2
- mindspore/nn/optim/adasum.py +3 -3
- mindspore/nn/optim/asgd.py +4 -2
- mindspore/nn/optim/ftrl.py +31 -61
- mindspore/nn/optim/lamb.py +5 -3
- mindspore/nn/optim/lars.py +2 -2
- mindspore/nn/optim/lazyadam.py +6 -4
- mindspore/nn/optim/momentum.py +13 -25
- mindspore/nn/optim/optimizer.py +6 -3
- mindspore/nn/optim/proximal_ada_grad.py +4 -2
- mindspore/nn/optim/rmsprop.py +9 -3
- mindspore/nn/optim/rprop.py +4 -2
- mindspore/nn/optim/sgd.py +7 -4
- mindspore/nn/optim/thor.py +2 -2
- mindspore/nn/probability/distribution/_utils/custom_ops.py +2 -2
- mindspore/nn/probability/distribution/beta.py +2 -2
- mindspore/nn/probability/distribution/categorical.py +4 -6
- mindspore/nn/probability/distribution/cauchy.py +2 -2
- mindspore/nn/probability/distribution/exponential.py +2 -2
- mindspore/nn/probability/distribution/geometric.py +1 -1
- mindspore/nn/probability/distribution/gumbel.py +2 -2
- mindspore/nn/probability/distribution/logistic.py +1 -1
- mindspore/nn/probability/distribution/poisson.py +2 -2
- mindspore/nn/probability/distribution/uniform.py +2 -2
- mindspore/nn/reinforcement/_tensors_queue.py +13 -1
- mindspore/nn/wrap/__init__.py +2 -1
- mindspore/nn/wrap/cell_wrapper.py +58 -13
- mindspore/nn/wrap/grad_reducer.py +148 -8
- mindspore/nn/wrap/loss_scale.py +32 -9
- mindspore/numpy/__init__.py +2 -0
- mindspore/numpy/array_creations.py +2 -0
- mindspore/numpy/array_ops.py +6 -6
- mindspore/numpy/dtypes.py +3 -3
- mindspore/numpy/fft.py +431 -0
- mindspore/numpy/math_ops.py +61 -67
- mindspore/numpy/utils.py +3 -0
- mindspore/opencv_core452.dll +0 -0
- mindspore/opencv_imgcodecs452.dll +0 -0
- mindspore/opencv_imgproc452.dll +0 -0
- mindspore/ops/__init__.py +8 -4
- mindspore/ops/_grad_experimental/grad_array_ops.py +4 -160
- mindspore/ops/_grad_experimental/grad_comm_ops.py +93 -36
- mindspore/ops/_grad_experimental/grad_inner_ops.py +8 -0
- mindspore/ops/_grad_experimental/grad_math_ops.py +92 -287
- mindspore/ops/_grad_experimental/grad_nn_ops.py +0 -53
- mindspore/ops/_grad_experimental/grad_quant_ops.py +3 -3
- mindspore/ops/_grad_experimental/grad_sparse.py +1 -1
- mindspore/ops/_grad_experimental/grad_sparse_ops.py +3 -3
- mindspore/ops/_op_impl/__init__.py +0 -1
- mindspore/ops/_op_impl/aicpu/__init__.py +1 -0
- mindspore/ops/_op_impl/aicpu/gamma.py +2 -0
- mindspore/ops/_op_impl/{cpu/concat.py → aicpu/generate_eod_mask.py} +16 -17
- mindspore/ops/_op_impl/aicpu/log_uniform_candidate_sampler.py +1 -3
- mindspore/ops/_op_impl/aicpu/poisson.py +2 -0
- mindspore/ops/_op_impl/cpu/__init__.py +1 -3
- mindspore/ops/_op_impl/cpu/adam.py +2 -2
- mindspore/ops/_op_impl/cpu/adam_weight_decay.py +3 -2
- mindspore/ops/_op_impl/cpu/maximum_grad.py +16 -14
- mindspore/ops/_op_impl/cpu/minimum_grad.py +8 -0
- mindspore/ops/_vmap/vmap_array_ops.py +164 -101
- mindspore/ops/_vmap/vmap_base.py +8 -1
- mindspore/ops/_vmap/vmap_grad_math_ops.py +95 -9
- mindspore/ops/_vmap/vmap_grad_nn_ops.py +143 -58
- mindspore/ops/_vmap/vmap_image_ops.py +70 -13
- mindspore/ops/_vmap/vmap_math_ops.py +130 -58
- mindspore/ops/_vmap/vmap_nn_ops.py +249 -115
- mindspore/ops/_vmap/vmap_other_ops.py +1 -1
- mindspore/ops/auto_generate/__init__.py +31 -0
- mindspore/ops/auto_generate/cpp_create_prim_instance_helper.py +231 -0
- mindspore/ops/auto_generate/gen_arg_dtype_cast.py +250 -0
- mindspore/ops/auto_generate/gen_arg_handler.py +197 -0
- mindspore/ops/auto_generate/gen_extend_func.py +980 -0
- mindspore/ops/auto_generate/gen_ops_def.py +6443 -0
- mindspore/ops/auto_generate/gen_ops_prim.py +13167 -0
- mindspore/ops/auto_generate/pyboost_inner_prim.py +429 -0
- mindspore/ops/composite/__init__.py +5 -2
- mindspore/ops/composite/base.py +121 -23
- mindspore/ops/composite/math_ops.py +10 -49
- mindspore/ops/composite/multitype_ops/_compile_utils.py +191 -618
- mindspore/ops/composite/multitype_ops/_constexpr_utils.py +25 -134
- mindspore/ops/composite/multitype_ops/add_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/bitwise_and_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/bitwise_or_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/bitwise_xor_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/div_impl.py +8 -0
- mindspore/ops/composite/multitype_ops/equal_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/floordiv_impl.py +8 -0
- mindspore/ops/composite/multitype_ops/getitem_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/greater_equal_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/greater_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/in_impl.py +8 -2
- mindspore/ops/composite/multitype_ops/left_shift_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/less_equal_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/less_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/logic_not_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/logical_and_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/logical_or_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/mod_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/mul_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/negative_impl.py +9 -3
- mindspore/ops/composite/multitype_ops/not_equal_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/not_in_impl.py +6 -1
- mindspore/ops/composite/multitype_ops/ones_like_impl.py +2 -2
- mindspore/ops/composite/multitype_ops/pow_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/right_shift_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/setitem_impl.py +32 -21
- mindspore/ops/composite/multitype_ops/sub_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/zeros_like_impl.py +6 -3
- mindspore/ops/deprecated.py +14 -3
- mindspore/ops/extend/__init__.py +53 -0
- mindspore/ops/extend/array_func.py +218 -0
- mindspore/ops/extend/math_func.py +76 -0
- mindspore/ops/extend/nn_func.py +308 -0
- mindspore/ops/function/__init__.py +31 -11
- mindspore/ops/function/array_func.py +848 -1736
- mindspore/ops/function/clip_func.py +19 -31
- mindspore/ops/function/debug_func.py +2 -5
- mindspore/ops/function/fft_func.py +31 -0
- mindspore/ops/function/grad/grad_func.py +27 -20
- mindspore/ops/function/image_func.py +27 -21
- mindspore/ops/function/linalg_func.py +30 -53
- mindspore/ops/function/math_func.py +916 -2791
- mindspore/ops/function/nn_func.py +1445 -889
- mindspore/ops/function/other_func.py +6 -7
- mindspore/ops/function/parameter_func.py +6 -92
- mindspore/ops/function/random_func.py +254 -108
- mindspore/ops/function/reshard_func.py +102 -0
- mindspore/ops/function/sparse_func.py +4 -4
- mindspore/ops/function/sparse_unary_func.py +11 -18
- mindspore/ops/function/spectral_func.py +1 -1
- mindspore/ops/function/vmap_func.py +15 -14
- mindspore/ops/functional.py +342 -343
- mindspore/ops/op_info_register.py +16 -43
- mindspore/ops/operations/__init__.py +32 -23
- mindspore/ops/operations/_embedding_cache_ops.py +1 -1
- mindspore/ops/operations/_grad_ops.py +21 -853
- mindspore/ops/operations/_infer_ops.py +19 -0
- mindspore/ops/operations/_inner_ops.py +155 -511
- mindspore/ops/operations/_quant_ops.py +4 -4
- mindspore/ops/operations/_rl_inner_ops.py +3 -3
- mindspore/ops/operations/_scalar_ops.py +5 -480
- mindspore/ops/operations/_sequence_ops.py +6 -36
- mindspore/ops/operations/_tensor_array.py +8 -8
- mindspore/ops/operations/array_ops.py +112 -2698
- mindspore/ops/operations/comm_ops.py +801 -118
- mindspore/ops/operations/custom_ops.py +62 -121
- mindspore/ops/operations/debug_ops.py +105 -36
- mindspore/ops/operations/image_ops.py +3 -219
- mindspore/ops/operations/inner_ops.py +54 -40
- mindspore/ops/operations/linalg_ops.py +1 -49
- mindspore/ops/operations/manually_defined/__init__.py +24 -0
- mindspore/ops/operations/manually_defined/_inner.py +61 -0
- mindspore/ops/operations/manually_defined/ops_def.py +2016 -0
- mindspore/ops/operations/math_ops.py +621 -4654
- mindspore/ops/operations/nn_ops.py +316 -2226
- mindspore/ops/operations/other_ops.py +53 -45
- mindspore/ops/operations/random_ops.py +4 -51
- mindspore/ops/operations/reshard_ops.py +53 -0
- mindspore/ops/operations/sparse_ops.py +8 -8
- mindspore/ops/primitive.py +204 -103
- mindspore/ops/silent_check.py +162 -0
- mindspore/ops_generate/__init__.py +27 -0
- mindspore/ops_generate/arg_dtype_cast.py +250 -0
- mindspore/ops_generate/arg_handler.py +197 -0
- mindspore/ops_generate/gen_aclnn_implement.py +263 -0
- mindspore/ops_generate/gen_ops.py +1084 -0
- mindspore/ops_generate/gen_ops_inner_prim.py +131 -0
- mindspore/ops_generate/gen_pyboost_func.py +968 -0
- mindspore/ops_generate/gen_utils.py +209 -0
- mindspore/ops_generate/op_proto.py +138 -0
- mindspore/ops_generate/pyboost_utils.py +354 -0
- mindspore/ops_generate/template.py +239 -0
- mindspore/parallel/__init__.py +7 -4
- mindspore/parallel/_auto_parallel_context.py +155 -6
- mindspore/parallel/_cell_wrapper.py +16 -9
- mindspore/parallel/_cost_model_context.py +1 -1
- mindspore/parallel/_dp_allreduce_fusion.py +159 -159
- mindspore/parallel/_parallel_serialization.py +62 -14
- mindspore/parallel/_ps_context.py +1 -1
- mindspore/parallel/_recovery_context.py +1 -1
- mindspore/parallel/_tensor.py +18 -9
- mindspore/parallel/_transformer/__init__.py +1 -1
- mindspore/parallel/_transformer/layers.py +1 -1
- mindspore/parallel/_transformer/loss.py +1 -1
- mindspore/parallel/_transformer/moe.py +1 -1
- mindspore/parallel/_transformer/op_parallel_config.py +1 -1
- mindspore/parallel/_transformer/transformer.py +10 -10
- mindspore/parallel/_utils.py +161 -6
- mindspore/parallel/algo_parameter_config.py +6 -8
- mindspore/parallel/checkpoint_transform.py +369 -64
- mindspore/parallel/cluster/__init__.py +15 -0
- mindspore/parallel/cluster/process_entity/__init__.py +18 -0
- mindspore/parallel/cluster/process_entity/_api.py +344 -0
- mindspore/parallel/cluster/process_entity/_utils.py +126 -0
- mindspore/parallel/cluster/run.py +136 -0
- mindspore/parallel/mpi/__init__.py +1 -1
- mindspore/parallel/mpi/_mpi_config.py +1 -1
- mindspore/parallel/parameter_broadcast.py +152 -0
- mindspore/parallel/shard.py +128 -17
- mindspore/profiler/__init__.py +3 -2
- mindspore/profiler/common/process_pool.py +41 -0
- mindspore/profiler/common/singleton.py +28 -0
- mindspore/profiler/common/util.py +125 -0
- mindspore/profiler/envprofiling.py +2 -2
- mindspore/{_extends/parallel_compile/tbe_compiler → profiler/parser/ascend_analysis}/__init__.py +1 -1
- mindspore/profiler/parser/ascend_analysis/constant.py +53 -0
- mindspore/profiler/parser/ascend_analysis/file_manager.py +159 -0
- mindspore/profiler/parser/ascend_analysis/function_event.py +161 -0
- mindspore/profiler/parser/ascend_analysis/fwk_cann_parser.py +131 -0
- mindspore/profiler/parser/ascend_analysis/fwk_file_parser.py +85 -0
- mindspore/profiler/parser/ascend_analysis/msprof_timeline_parser.py +57 -0
- mindspore/profiler/parser/ascend_analysis/profiler_info_parser.py +116 -0
- mindspore/profiler/parser/ascend_analysis/tlv_decoder.py +86 -0
- mindspore/profiler/parser/ascend_analysis/trace_event_manager.py +68 -0
- mindspore/profiler/parser/ascend_cluster_generator.py +116 -0
- mindspore/profiler/parser/ascend_communicate_generator.py +314 -0
- mindspore/profiler/parser/ascend_flops_generator.py +27 -5
- mindspore/profiler/parser/ascend_fpbp_generator.py +8 -2
- mindspore/profiler/parser/ascend_hccl_generator.py +31 -280
- mindspore/profiler/parser/ascend_integrate_generator.py +42 -0
- mindspore/profiler/parser/ascend_memory_generator.py +185 -0
- mindspore/profiler/parser/ascend_msprof_exporter.py +151 -126
- mindspore/profiler/parser/ascend_msprof_generator.py +75 -274
- mindspore/profiler/parser/ascend_op_generator.py +94 -36
- mindspore/profiler/parser/ascend_timeline_generator.py +297 -131
- mindspore/profiler/parser/base_timeline_generator.py +17 -3
- mindspore/profiler/parser/cpu_gpu_timeline_generator.py +2 -1
- mindspore/profiler/parser/framework_parser.py +11 -4
- mindspore/profiler/parser/integrator.py +3 -1
- mindspore/profiler/parser/memory_usage_parser.py +8 -2
- mindspore/profiler/parser/minddata_analyzer.py +8 -2
- mindspore/profiler/parser/minddata_parser.py +73 -4
- mindspore/profiler/parser/msadvisor_analyzer.py +5 -3
- mindspore/profiler/parser/msadvisor_parser.py +10 -4
- mindspore/profiler/parser/profiler_info.py +16 -1
- mindspore/profiler/profiling.py +522 -195
- mindspore/rewrite/__init__.py +2 -13
- mindspore/rewrite/api/node.py +123 -37
- mindspore/rewrite/api/pattern_engine.py +2 -3
- mindspore/rewrite/api/scoped_value.py +16 -15
- mindspore/rewrite/api/symbol_tree.py +46 -30
- mindspore/rewrite/ast_helpers/__init__.py +3 -6
- mindspore/rewrite/ast_helpers/ast_converter.py +143 -0
- mindspore/rewrite/ast_helpers/ast_finder.py +48 -0
- mindspore/rewrite/ast_helpers/ast_flattener.py +268 -0
- mindspore/rewrite/ast_helpers/ast_modifier.py +160 -92
- mindspore/rewrite/common/__init__.py +1 -2
- mindspore/rewrite/common/config.py +24 -0
- mindspore/rewrite/common/{rewrite_elog.py → error_log.py} +39 -39
- mindspore/rewrite/{namer.py → common/namer.py} +63 -18
- mindspore/rewrite/common/namespace.py +118 -0
- mindspore/rewrite/node/__init__.py +5 -5
- mindspore/rewrite/node/call_function.py +23 -7
- mindspore/rewrite/node/cell_container.py +7 -3
- mindspore/rewrite/node/control_flow.py +53 -28
- mindspore/rewrite/node/node.py +212 -196
- mindspore/rewrite/node/node_manager.py +51 -22
- mindspore/rewrite/node/node_topological_manager.py +3 -23
- mindspore/rewrite/parsers/__init__.py +12 -0
- mindspore/rewrite/parsers/arguments_parser.py +8 -9
- mindspore/rewrite/parsers/assign_parser.py +637 -413
- mindspore/rewrite/parsers/attribute_parser.py +3 -4
- mindspore/rewrite/parsers/class_def_parser.py +115 -148
- mindspore/rewrite/parsers/constant_parser.py +5 -5
- mindspore/rewrite/parsers/container_parser.py +4 -6
- mindspore/rewrite/parsers/expr_parser.py +55 -0
- mindspore/rewrite/parsers/for_parser.py +31 -98
- mindspore/rewrite/parsers/function_def_parser.py +13 -5
- mindspore/rewrite/parsers/if_parser.py +28 -10
- mindspore/rewrite/parsers/module_parser.py +8 -182
- mindspore/rewrite/parsers/parser.py +1 -5
- mindspore/rewrite/parsers/parser_register.py +1 -1
- mindspore/rewrite/parsers/return_parser.py +5 -10
- mindspore/rewrite/parsers/while_parser.py +59 -0
- mindspore/rewrite/sparsify/utils.py +1 -1
- mindspore/rewrite/symbol_tree/__init__.py +20 -0
- mindspore/rewrite/{symbol_tree.py → symbol_tree/symbol_tree.py} +704 -185
- mindspore/rewrite/{symbol_tree_builder.py → symbol_tree/symbol_tree_builder.py} +8 -8
- mindspore/rewrite/{symbol_tree_dumper.py → symbol_tree/symbol_tree_dumper.py} +4 -4
- mindspore/run_check/_check_version.py +6 -14
- mindspore/run_check/run_check.py +1 -1
- mindspore/safeguard/rewrite_obfuscation.py +9 -19
- mindspore/swresample-4.dll +0 -0
- mindspore/swscale-6.dll +0 -0
- mindspore/tinyxml2.dll +0 -0
- mindspore/train/__init__.py +6 -5
- mindspore/train/_utils.py +178 -4
- mindspore/train/amp.py +167 -245
- mindspore/train/anf_ir_pb2.py +14 -2
- mindspore/train/callback/__init__.py +5 -2
- mindspore/train/callback/_backup_and_restore.py +5 -5
- mindspore/train/callback/_callback.py +4 -4
- mindspore/train/callback/_checkpoint.py +151 -37
- mindspore/train/callback/_cluster_monitor.py +201 -0
- mindspore/train/callback/_early_stop.py +2 -2
- mindspore/train/callback/_flops_collector.py +238 -0
- mindspore/train/callback/_landscape.py +16 -11
- mindspore/train/callback/_loss_monitor.py +2 -2
- mindspore/train/callback/_mindio_ttp.py +443 -0
- mindspore/train/callback/_on_request_exit.py +2 -2
- mindspore/train/callback/_reduce_lr_on_plateau.py +2 -2
- mindspore/train/callback/_summary_collector.py +13 -14
- mindspore/train/callback/_time_monitor.py +3 -3
- mindspore/train/data_sink.py +6 -5
- mindspore/train/dataset_helper.py +66 -21
- mindspore/train/loss_scale_manager.py +2 -2
- mindspore/train/metrics/accuracy.py +7 -7
- mindspore/train/metrics/confusion_matrix.py +8 -6
- mindspore/train/metrics/cosine_similarity.py +6 -4
- mindspore/train/metrics/error.py +2 -2
- mindspore/train/metrics/metric.py +3 -3
- mindspore/train/metrics/perplexity.py +2 -1
- mindspore/train/metrics/topk.py +2 -2
- mindspore/train/mind_ir_pb2.py +89 -15
- mindspore/train/model.py +298 -56
- mindspore/train/serialization.py +501 -221
- mindspore/train/summary/_summary_adapter.py +1 -1
- mindspore/train/summary/_writer_pool.py +1 -1
- mindspore/train/summary/summary_record.py +56 -34
- mindspore/train/train_thor/convert_utils.py +3 -3
- mindspore/turbojpeg.dll +0 -0
- mindspore/version.py +1 -1
- {mindspore-2.2.11.dist-info → mindspore-2.3.0.dist-info}/METADATA +3 -3
- mindspore-2.3.0.dist-info/RECORD +1400 -0
- {mindspore-2.2.11.dist-info → mindspore-2.3.0.dist-info}/entry_points.txt +1 -0
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_adapter.py +0 -662
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_helper.py +0 -377
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_job.py +0 -201
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_job_manager.py +0 -515
- mindspore/gen_ops.py +0 -273
- mindspore/nn/layer/flash_attention.py +0 -189
- mindspore/ops/_op_impl/cpu/tensor_shape.py +0 -42
- mindspore/ops/_op_impl/tbe/__init__.py +0 -47
- mindspore/ops/_op_impl/tbe/abs.py +0 -38
- mindspore/ops/_op_impl/tbe/abs_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/abs_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/abs_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/accumulate_n_v2.py +0 -41
- mindspore/ops/_op_impl/tbe/accumulate_n_v2_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/acos.py +0 -37
- mindspore/ops/_op_impl/tbe/acos_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/acos_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/acos_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/acosh.py +0 -37
- mindspore/ops/_op_impl/tbe/acosh_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/acosh_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/acosh_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/act_ulq_clamp_max_grad.py +0 -38
- mindspore/ops/_op_impl/tbe/act_ulq_clamp_min_grad.py +0 -38
- mindspore/ops/_op_impl/tbe/acts_ulq.py +0 -45
- mindspore/ops/_op_impl/tbe/acts_ulq_input_grad.py +0 -38
- mindspore/ops/_op_impl/tbe/adam_apply_one.py +0 -50
- mindspore/ops/_op_impl/tbe/adam_apply_one_assign.py +0 -53
- mindspore/ops/_op_impl/tbe/adam_apply_one_ds.py +0 -51
- mindspore/ops/_op_impl/tbe/adam_apply_one_with_decay.py +0 -54
- mindspore/ops/_op_impl/tbe/adam_apply_one_with_decay_assign.py +0 -54
- mindspore/ops/_op_impl/tbe/adam_apply_one_with_decay_ds.py +0 -55
- mindspore/ops/_op_impl/tbe/adaptive_max_pool2d.py +0 -37
- mindspore/ops/_op_impl/tbe/add.py +0 -42
- mindspore/ops/_op_impl/tbe/add_ds.py +0 -43
- mindspore/ops/_op_impl/tbe/add_n.py +0 -39
- mindspore/ops/_op_impl/tbe/add_n_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/addcdiv.py +0 -41
- mindspore/ops/_op_impl/tbe/addcdiv_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/addcmul.py +0 -43
- mindspore/ops/_op_impl/tbe/addcmul_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/apply_ada_max.py +0 -68
- mindspore/ops/_op_impl/tbe/apply_ada_max_ds.py +0 -69
- mindspore/ops/_op_impl/tbe/apply_adadelta.py +0 -66
- mindspore/ops/_op_impl/tbe/apply_adadelta_ds.py +0 -67
- mindspore/ops/_op_impl/tbe/apply_adagrad.py +0 -55
- mindspore/ops/_op_impl/tbe/apply_adagrad_d_a.py +0 -67
- mindspore/ops/_op_impl/tbe/apply_adagrad_ds.py +0 -56
- mindspore/ops/_op_impl/tbe/apply_adagrad_v2.py +0 -48
- mindspore/ops/_op_impl/tbe/apply_adagrad_v2_ds.py +0 -49
- mindspore/ops/_op_impl/tbe/apply_adam.py +0 -79
- mindspore/ops/_op_impl/tbe/apply_adam_ds.py +0 -80
- mindspore/ops/_op_impl/tbe/apply_adam_with_amsgrad.py +0 -60
- mindspore/ops/_op_impl/tbe/apply_adam_with_amsgrad_ds.py +0 -61
- mindspore/ops/_op_impl/tbe/apply_add_sign.py +0 -65
- mindspore/ops/_op_impl/tbe/apply_add_sign_ds.py +0 -66
- mindspore/ops/_op_impl/tbe/apply_centered_rms_prop.py +0 -77
- mindspore/ops/_op_impl/tbe/apply_centered_rms_prop_ds.py +0 -78
- mindspore/ops/_op_impl/tbe/apply_ftrl.py +0 -67
- mindspore/ops/_op_impl/tbe/apply_ftrl_ds.py +0 -68
- mindspore/ops/_op_impl/tbe/apply_gradient_descent.py +0 -44
- mindspore/ops/_op_impl/tbe/apply_gradient_descent_ds.py +0 -45
- mindspore/ops/_op_impl/tbe/apply_keras_momentum.py +0 -49
- mindspore/ops/_op_impl/tbe/apply_momentum.py +0 -64
- mindspore/ops/_op_impl/tbe/apply_momentum_ds.py +0 -65
- mindspore/ops/_op_impl/tbe/apply_power_sign.py +0 -65
- mindspore/ops/_op_impl/tbe/apply_power_sign_ds.py +0 -66
- mindspore/ops/_op_impl/tbe/apply_proximal_adagrad.py +0 -57
- mindspore/ops/_op_impl/tbe/apply_proximal_adagrad_ds.py +0 -58
- mindspore/ops/_op_impl/tbe/apply_proximal_gradient_descent.py +0 -54
- mindspore/ops/_op_impl/tbe/apply_proximal_gradient_descent_ds.py +0 -55
- mindspore/ops/_op_impl/tbe/apply_rms_prop.py +0 -52
- mindspore/ops/_op_impl/tbe/approximate_equal.py +0 -39
- mindspore/ops/_op_impl/tbe/approximate_equal_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/arg_max.py +0 -38
- mindspore/ops/_op_impl/tbe/arg_max_with_value.py +0 -38
- mindspore/ops/_op_impl/tbe/arg_max_with_value_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/arg_min.py +0 -38
- mindspore/ops/_op_impl/tbe/arg_min_v2_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/arg_min_with_value.py +0 -38
- mindspore/ops/_op_impl/tbe/arg_min_with_value_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/asin.py +0 -37
- mindspore/ops/_op_impl/tbe/asin_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/asin_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/asin_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/asinh.py +0 -37
- mindspore/ops/_op_impl/tbe/asinh_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/asinh_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/asinh_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/assign.py +0 -79
- mindspore/ops/_op_impl/tbe/assign_add.py +0 -59
- mindspore/ops/_op_impl/tbe/assign_add_ds.py +0 -60
- mindspore/ops/_op_impl/tbe/assign_ds.py +0 -80
- mindspore/ops/_op_impl/tbe/assign_sub.py +0 -55
- mindspore/ops/_op_impl/tbe/assign_sub_ds.py +0 -56
- mindspore/ops/_op_impl/tbe/atan.py +0 -37
- mindspore/ops/_op_impl/tbe/atan2.py +0 -38
- mindspore/ops/_op_impl/tbe/atan2_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/atan_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/atan_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/atan_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/atanh.py +0 -37
- mindspore/ops/_op_impl/tbe/atanh_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/avg_pool.py +0 -43
- mindspore/ops/_op_impl/tbe/avg_pool_3d.py +0 -44
- mindspore/ops/_op_impl/tbe/avg_pool_3d_grad.py +0 -45
- mindspore/ops/_op_impl/tbe/avg_pool_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/avg_pool_grad.py +0 -42
- mindspore/ops/_op_impl/tbe/avg_pool_grad_vm.py +0 -42
- mindspore/ops/_op_impl/tbe/basic_lstm_cell.py +0 -57
- mindspore/ops/_op_impl/tbe/basic_lstm_cell_c_state_grad.py +0 -50
- mindspore/ops/_op_impl/tbe/basic_lstm_cell_c_state_grad_v2.py +0 -51
- mindspore/ops/_op_impl/tbe/basic_lstm_cell_input_grad.py +0 -42
- mindspore/ops/_op_impl/tbe/basic_lstm_cell_weight_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/batch_matmul.py +0 -42
- mindspore/ops/_op_impl/tbe/batch_matmul_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/batch_matmul_v2.py +0 -47
- mindspore/ops/_op_impl/tbe/batch_to_space.py +0 -38
- mindspore/ops/_op_impl/tbe/batch_to_space_nd.py +0 -38
- mindspore/ops/_op_impl/tbe/batch_to_space_nd_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/batch_to_space_nd_v2.py +0 -41
- mindspore/ops/_op_impl/tbe/batchnorm.py +0 -58
- mindspore/ops/_op_impl/tbe/batchnorm_grad.py +0 -58
- mindspore/ops/_op_impl/tbe/bce_with_logits_loss.py +0 -42
- mindspore/ops/_op_impl/tbe/bessel_i0e.py +0 -37
- mindspore/ops/_op_impl/tbe/bessel_i0e_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/bessel_i1e.py +0 -37
- mindspore/ops/_op_impl/tbe/bessel_i1e_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/bias_add.py +0 -38
- mindspore/ops/_op_impl/tbe/bias_add_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/bias_add_grad.py +0 -53
- mindspore/ops/_op_impl/tbe/binary_cross_entropy.py +0 -39
- mindspore/ops/_op_impl/tbe/binary_cross_entropy_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/binary_cross_entropy_grad.py +0 -44
- mindspore/ops/_op_impl/tbe/binary_cross_entropy_grad_ds.py +0 -45
- mindspore/ops/_op_impl/tbe/bitwise_and.py +0 -39
- mindspore/ops/_op_impl/tbe/bitwise_and_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/bitwise_or.py +0 -39
- mindspore/ops/_op_impl/tbe/bitwise_or_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/bitwise_xor.py +0 -39
- mindspore/ops/_op_impl/tbe/bitwise_xor_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/bn_infer.py +0 -43
- mindspore/ops/_op_impl/tbe/bn_infer_ds.py +0 -45
- mindspore/ops/_op_impl/tbe/bn_infer_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/bn_infer_grad_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/bn_inference.py +0 -50
- mindspore/ops/_op_impl/tbe/bn_training_reduce.py +0 -38
- mindspore/ops/_op_impl/tbe/bn_training_reduce_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/bn_training_reduce_grad.py +0 -46
- mindspore/ops/_op_impl/tbe/bn_training_reduce_grad_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/bn_training_update.py +0 -52
- mindspore/ops/_op_impl/tbe/bn_training_update_ds.py +0 -53
- mindspore/ops/_op_impl/tbe/bn_training_update_grad.py +0 -44
- mindspore/ops/_op_impl/tbe/bn_training_update_grad_ds.py +0 -45
- mindspore/ops/_op_impl/tbe/bn_training_update_v2.py +0 -48
- mindspore/ops/_op_impl/tbe/bn_training_update_v3.py +0 -51
- mindspore/ops/_op_impl/tbe/bounding_box_decode.py +0 -41
- mindspore/ops/_op_impl/tbe/bounding_box_decode_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/bounding_box_encode.py +0 -38
- mindspore/ops/_op_impl/tbe/broadcast_to.py +0 -40
- mindspore/ops/_op_impl/tbe/broadcast_to_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/cast.py +0 -55
- mindspore/ops/_op_impl/tbe/cast_ds.py +0 -58
- mindspore/ops/_op_impl/tbe/cdist.py +0 -38
- mindspore/ops/_op_impl/tbe/cdist_grad.py +0 -42
- mindspore/ops/_op_impl/tbe/ceil.py +0 -37
- mindspore/ops/_op_impl/tbe/ceil_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/celu.py +0 -39
- mindspore/ops/_op_impl/tbe/centralization.py +0 -39
- mindspore/ops/_op_impl/tbe/check_valid.py +0 -38
- mindspore/ops/_op_impl/tbe/check_valid_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/clip_by_norm_no_div_sum.py +0 -41
- mindspore/ops/_op_impl/tbe/clip_by_norm_no_div_sum_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/clip_by_value.py +0 -41
- mindspore/ops/_op_impl/tbe/clip_by_value_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/concat.py +0 -40
- mindspore/ops/_op_impl/tbe/concat_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/confusion_matrix.py +0 -63
- mindspore/ops/_op_impl/tbe/confusion_mul_grad.py +0 -40
- mindspore/ops/_op_impl/tbe/confusion_softmax_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/confusion_transpose_d.py +0 -39
- mindspore/ops/_op_impl/tbe/conv2d.py +0 -47
- mindspore/ops/_op_impl/tbe/conv2d_backprop_filter.py +0 -42
- mindspore/ops/_op_impl/tbe/conv2d_backprop_filter_ds.py +0 -43
- mindspore/ops/_op_impl/tbe/conv2d_backprop_input.py +0 -42
- mindspore/ops/_op_impl/tbe/conv2d_backprop_input_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/conv2d_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/conv2d_transpose.py +0 -48
- mindspore/ops/_op_impl/tbe/conv3d.py +0 -45
- mindspore/ops/_op_impl/tbe/conv3d_backprop_filter.py +0 -42
- mindspore/ops/_op_impl/tbe/conv3d_backprop_input.py +0 -42
- mindspore/ops/_op_impl/tbe/conv3d_transpose.py +0 -47
- mindspore/ops/_op_impl/tbe/conv3d_transpose_ds.py +0 -48
- mindspore/ops/_op_impl/tbe/cos.py +0 -37
- mindspore/ops/_op_impl/tbe/cos_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/cosh.py +0 -37
- mindspore/ops/_op_impl/tbe/cosh_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/ctc_loss_v2.py +0 -42
- mindspore/ops/_op_impl/tbe/ctc_loss_v2_grad.py +0 -44
- mindspore/ops/_op_impl/tbe/cum_sum.py +0 -42
- mindspore/ops/_op_impl/tbe/cum_sum_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/cummin.py +0 -41
- mindspore/ops/_op_impl/tbe/cumprod.py +0 -42
- mindspore/ops/_op_impl/tbe/data_format_dim_map.py +0 -38
- mindspore/ops/_op_impl/tbe/data_format_dim_map_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/deformable_offsets.py +0 -45
- mindspore/ops/_op_impl/tbe/deformable_offsets_grad.py +0 -48
- mindspore/ops/_op_impl/tbe/depth_to_space_ds.py +0 -49
- mindspore/ops/_op_impl/tbe/depthwise_conv2d.py +0 -44
- mindspore/ops/_op_impl/tbe/depthwise_conv2d_backprop_filter.py +0 -41
- mindspore/ops/_op_impl/tbe/depthwise_conv2d_backprop_input.py +0 -41
- mindspore/ops/_op_impl/tbe/diag.py +0 -38
- mindspore/ops/_op_impl/tbe/diag_part.py +0 -38
- mindspore/ops/_op_impl/tbe/dilation.py +0 -40
- mindspore/ops/_op_impl/tbe/div.py +0 -41
- mindspore/ops/_op_impl/tbe/div_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/div_no_nan.py +0 -41
- mindspore/ops/_op_impl/tbe/div_no_nan_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/dropout_do_mask.py +0 -38
- mindspore/ops/_op_impl/tbe/dropout_do_mask_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/dropout_do_mask_v3.py +0 -39
- mindspore/ops/_op_impl/tbe/dynamic_atomic_addr_clean.py +0 -34
- mindspore/ops/_op_impl/tbe/dynamic_gru_v2.py +0 -95
- mindspore/ops/_op_impl/tbe/dynamic_rnn.py +0 -82
- mindspore/ops/_op_impl/tbe/elu.py +0 -38
- mindspore/ops/_op_impl/tbe/elu_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/elu_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/elu_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/equal.py +0 -42
- mindspore/ops/_op_impl/tbe/equal_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/erf.py +0 -37
- mindspore/ops/_op_impl/tbe/erf_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/erfc.py +0 -37
- mindspore/ops/_op_impl/tbe/erfc_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/erfinv.py +0 -36
- mindspore/ops/_op_impl/tbe/exp.py +0 -40
- mindspore/ops/_op_impl/tbe/exp_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/expand_dims.py +0 -38
- mindspore/ops/_op_impl/tbe/expm1.py +0 -37
- mindspore/ops/_op_impl/tbe/expm1_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/extract_image_patches.py +0 -41
- mindspore/ops/_op_impl/tbe/extract_volume_patches.py +0 -39
- mindspore/ops/_op_impl/tbe/fake_quant_with_min_max_vars.py +0 -39
- mindspore/ops/_op_impl/tbe/fake_quant_with_min_max_vars_gradient.py +0 -43
- mindspore/ops/_op_impl/tbe/fake_quant_with_min_max_vars_per_channel.py +0 -39
- mindspore/ops/_op_impl/tbe/fake_quant_with_min_max_vars_per_channel_gradient.py +0 -43
- mindspore/ops/_op_impl/tbe/fast_gelu.py +0 -37
- mindspore/ops/_op_impl/tbe/fast_gelu_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/fast_gelu_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/fast_gelu_grad_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/fill.py +0 -56
- mindspore/ops/_op_impl/tbe/fill_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/flatten.py +0 -48
- mindspore/ops/_op_impl/tbe/floor.py +0 -37
- mindspore/ops/_op_impl/tbe/floor_div.py +0 -41
- mindspore/ops/_op_impl/tbe/floor_div_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/floor_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/floor_mod.py +0 -39
- mindspore/ops/_op_impl/tbe/floor_mod_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/fused_dbn_dw.py +0 -52
- mindspore/ops/_op_impl/tbe/fused_mul_add.py +0 -38
- mindspore/ops/_op_impl/tbe/fused_mul_add_n.py +0 -48
- mindspore/ops/_op_impl/tbe/fused_mul_add_n_l2loss.py +0 -53
- mindspore/ops/_op_impl/tbe/fused_mul_apply_momentum.py +0 -57
- mindspore/ops/_op_impl/tbe/fused_mul_apply_momentum_extern.py +0 -67
- mindspore/ops/_op_impl/tbe/gather_nd.py +0 -52
- mindspore/ops/_op_impl/tbe/gather_nd_ds.py +0 -48
- mindspore/ops/_op_impl/tbe/gather_v2.py +0 -56
- mindspore/ops/_op_impl/tbe/gather_v2_ds.py +0 -68
- mindspore/ops/_op_impl/tbe/gelu.py +0 -37
- mindspore/ops/_op_impl/tbe/gelu_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/gelu_grad.py +0 -42
- mindspore/ops/_op_impl/tbe/gelu_grad_ds.py +0 -43
- mindspore/ops/_op_impl/tbe/ger.py +0 -43
- mindspore/ops/_op_impl/tbe/ger_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/greater.py +0 -43
- mindspore/ops/_op_impl/tbe/greater_equal.py +0 -41
- mindspore/ops/_op_impl/tbe/greater_equal_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/gru_v2_hidden_grad.py +0 -51
- mindspore/ops/_op_impl/tbe/gru_v2_hidden_grad_cell.py +0 -52
- mindspore/ops/_op_impl/tbe/hard_swish.py +0 -37
- mindspore/ops/_op_impl/tbe/hard_swish_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/hard_swish_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/hard_swish_grad_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/histogram_fixed_width.py +0 -40
- mindspore/ops/_op_impl/tbe/hshrink.py +0 -33
- mindspore/ops/_op_impl/tbe/hshrink_grad.py +0 -37
- mindspore/ops/_op_impl/tbe/hsigmoid.py +0 -45
- mindspore/ops/_op_impl/tbe/hsigmoid_grad.py +0 -39
- mindspore/ops/_op_impl/tbe/ifmr.py +0 -47
- mindspore/ops/_op_impl/tbe/ifmr_ds.py +0 -48
- mindspore/ops/_op_impl/tbe/im2col.py +0 -42
- mindspore/ops/_op_impl/tbe/in_top_k.py +0 -37
- mindspore/ops/_op_impl/tbe/inplace_add.py +0 -39
- mindspore/ops/_op_impl/tbe/inplace_index_add.py +0 -46
- mindspore/ops/_op_impl/tbe/inplace_sub.py +0 -39
- mindspore/ops/_op_impl/tbe/inplace_update.py +0 -39
- mindspore/ops/_op_impl/tbe/inplace_update_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/inv.py +0 -38
- mindspore/ops/_op_impl/tbe/inv_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/inv_grad.py +0 -40
- mindspore/ops/_op_impl/tbe/inv_grad_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/invert.py +0 -37
- mindspore/ops/_op_impl/tbe/invert_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/iou.py +0 -38
- mindspore/ops/_op_impl/tbe/iou_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/is_close.py +0 -40
- mindspore/ops/_op_impl/tbe/kl_div_loss.py +0 -38
- mindspore/ops/_op_impl/tbe/kl_div_loss_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/kl_div_loss_grad.py +0 -40
- mindspore/ops/_op_impl/tbe/l2_loss.py +0 -36
- mindspore/ops/_op_impl/tbe/l2_loss_ds.py +0 -37
- mindspore/ops/_op_impl/tbe/l2_normalize.py +0 -38
- mindspore/ops/_op_impl/tbe/l2_normalize_grad.py +0 -40
- mindspore/ops/_op_impl/tbe/lamb_apply_optimizer_assign.py +0 -55
- mindspore/ops/_op_impl/tbe/lamb_apply_weight_assign.py +0 -42
- mindspore/ops/_op_impl/tbe/lamb_next_mv.py +0 -59
- mindspore/ops/_op_impl/tbe/lamb_next_mv_with_decay.py +0 -59
- mindspore/ops/_op_impl/tbe/lamb_next_right.py +0 -44
- mindspore/ops/_op_impl/tbe/lamb_update_with_lr.py +0 -48
- mindspore/ops/_op_impl/tbe/lamb_update_with_lr_v2.py +0 -44
- mindspore/ops/_op_impl/tbe/lars_update.py +0 -50
- mindspore/ops/_op_impl/tbe/lars_update_ds.py +0 -51
- mindspore/ops/_op_impl/tbe/layer_norm.py +0 -46
- mindspore/ops/_op_impl/tbe/layer_norm_beta_gamma_backprop.py +0 -44
- mindspore/ops/_op_impl/tbe/layer_norm_beta_gamma_backprop_ds.py +0 -45
- mindspore/ops/_op_impl/tbe/layer_norm_beta_gamma_backprop_v2.py +0 -40
- mindspore/ops/_op_impl/tbe/layer_norm_beta_gamma_backprop_v2_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/layer_norm_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/layer_norm_grad.py +0 -48
- mindspore/ops/_op_impl/tbe/layer_norm_x_backprop.py +0 -43
- mindspore/ops/_op_impl/tbe/layer_norm_x_backprop_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/layer_norm_x_backprop_v2.py +0 -45
- mindspore/ops/_op_impl/tbe/layer_norm_x_backprop_v2_ds.py +0 -45
- mindspore/ops/_op_impl/tbe/lerp.py +0 -38
- mindspore/ops/_op_impl/tbe/less.py +0 -41
- mindspore/ops/_op_impl/tbe/less_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/less_equal.py +0 -41
- mindspore/ops/_op_impl/tbe/less_equal_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/log.py +0 -40
- mindspore/ops/_op_impl/tbe/log1p.py +0 -37
- mindspore/ops/_op_impl/tbe/log1p_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/log_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/logical_and.py +0 -37
- mindspore/ops/_op_impl/tbe/logical_and_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/logical_not.py +0 -36
- mindspore/ops/_op_impl/tbe/logical_not_ds.py +0 -37
- mindspore/ops/_op_impl/tbe/logical_or.py +0 -37
- mindspore/ops/_op_impl/tbe/logical_or_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/logsoftmax.py +0 -37
- mindspore/ops/_op_impl/tbe/logsoftmax_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/logsoftmax_grad.py +0 -38
- mindspore/ops/_op_impl/tbe/logsoftmax_grad_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/lp_norm.py +0 -40
- mindspore/ops/_op_impl/tbe/lp_norm_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/lrn.py +0 -41
- mindspore/ops/_op_impl/tbe/lrn_grad.py +0 -42
- mindspore/ops/_op_impl/tbe/lstm_input_grad.py +0 -51
- mindspore/ops/_op_impl/tbe/masked_fill.py +0 -40
- mindspore/ops/_op_impl/tbe/masked_fill_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/matmul.py +0 -53
- mindspore/ops/_op_impl/tbe/matmul_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/matmul_v2.py +0 -50
- mindspore/ops/_op_impl/tbe/matrix_diag.py +0 -45
- mindspore/ops/_op_impl/tbe/matrix_diag_part.py +0 -45
- mindspore/ops/_op_impl/tbe/matrix_set_diag.py +0 -46
- mindspore/ops/_op_impl/tbe/max_pool.py +0 -39
- mindspore/ops/_op_impl/tbe/max_pool3d.py +0 -44
- mindspore/ops/_op_impl/tbe/max_pool3d_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/max_pool3d_grad_grad.py +0 -44
- mindspore/ops/_op_impl/tbe/max_pool_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/max_pool_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/max_pool_grad_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/max_pool_grad_grad_with_argmax.py +0 -41
- mindspore/ops/_op_impl/tbe/max_pool_grad_with_argmax.py +0 -42
- mindspore/ops/_op_impl/tbe/max_pool_with_argmax.py +0 -40
- mindspore/ops/_op_impl/tbe/maximum.py +0 -39
- mindspore/ops/_op_impl/tbe/maximum_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/maximum_grad.py +0 -46
- mindspore/ops/_op_impl/tbe/maximum_grad_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/mem_set.py +0 -38
- mindspore/ops/_op_impl/tbe/minimum.py +0 -40
- mindspore/ops/_op_impl/tbe/minimum_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/minimum_grad.py +0 -46
- mindspore/ops/_op_impl/tbe/minimum_grad_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/mish.py +0 -37
- mindspore/ops/_op_impl/tbe/mod.py +0 -41
- mindspore/ops/_op_impl/tbe/mod_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/mul.py +0 -37
- mindspore/ops/_op_impl/tbe/mul_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/mul_no_nan.py +0 -39
- mindspore/ops/_op_impl/tbe/mul_no_nan_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/multilabel_margin_loss.py +0 -39
- mindspore/ops/_op_impl/tbe/neg.py +0 -39
- mindspore/ops/_op_impl/tbe/neg_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/new_im2col.py +0 -40
- mindspore/ops/_op_impl/tbe/nll_loss.py +0 -41
- mindspore/ops/_op_impl/tbe/nll_loss_grad.py +0 -44
- mindspore/ops/_op_impl/tbe/nms_with_mask.py +0 -39
- mindspore/ops/_op_impl/tbe/not_equal.py +0 -41
- mindspore/ops/_op_impl/tbe/not_equal_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/npu_alloc_float_status.py +0 -34
- mindspore/ops/_op_impl/tbe/npu_clear_float_status.py +0 -35
- mindspore/ops/_op_impl/tbe/npu_clear_float_status_v2.py +0 -35
- mindspore/ops/_op_impl/tbe/npu_get_float_status.py +0 -35
- mindspore/ops/_op_impl/tbe/npu_get_float_status_v2.py +0 -35
- mindspore/ops/_op_impl/tbe/one_hot.py +0 -48
- mindspore/ops/_op_impl/tbe/one_hot_ds.py +0 -45
- mindspore/ops/_op_impl/tbe/ones_like.py +0 -40
- mindspore/ops/_op_impl/tbe/ones_like_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/p_s_r_o_i_pooling.py +0 -40
- mindspore/ops/_op_impl/tbe/p_s_r_o_i_pooling_grad.py +0 -40
- mindspore/ops/_op_impl/tbe/pack.py +0 -58
- mindspore/ops/_op_impl/tbe/pack_ds.py +0 -59
- mindspore/ops/_op_impl/tbe/pad_d.py +0 -40
- mindspore/ops/_op_impl/tbe/pad_d_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/parallel_concat.py +0 -70
- mindspore/ops/_op_impl/tbe/parallel_resize_bilinear.py +0 -45
- mindspore/ops/_op_impl/tbe/parallel_resize_bilinear_grad.py +0 -44
- mindspore/ops/_op_impl/tbe/pdist.py +0 -36
- mindspore/ops/_op_impl/tbe/pooling.py +0 -46
- mindspore/ops/_op_impl/tbe/population_count.py +0 -38
- mindspore/ops/_op_impl/tbe/pow.py +0 -41
- mindspore/ops/_op_impl/tbe/pow_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/prelu.py +0 -37
- mindspore/ops/_op_impl/tbe/prelu_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/prelu_grad.py +0 -40
- mindspore/ops/_op_impl/tbe/range.py +0 -39
- mindspore/ops/_op_impl/tbe/real_div.py +0 -38
- mindspore/ops/_op_impl/tbe/real_div_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/reciprocal.py +0 -36
- mindspore/ops/_op_impl/tbe/reciprocal_ds.py +0 -37
- mindspore/ops/_op_impl/tbe/reciprocal_grad.py +0 -38
- mindspore/ops/_op_impl/tbe/reciprocal_grad_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/reduce_all.py +0 -38
- mindspore/ops/_op_impl/tbe/reduce_all_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/reduce_any.py +0 -38
- mindspore/ops/_op_impl/tbe/reduce_any_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/reduce_max.py +0 -43
- mindspore/ops/_op_impl/tbe/reduce_max_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/reduce_mean.py +0 -40
- mindspore/ops/_op_impl/tbe/reduce_mean_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/reduce_min.py +0 -41
- mindspore/ops/_op_impl/tbe/reduce_min_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/reduce_prod.py +0 -42
- mindspore/ops/_op_impl/tbe/reduce_prod_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/reduce_std.py +0 -44
- mindspore/ops/_op_impl/tbe/reduce_sum.py +0 -39
- mindspore/ops/_op_impl/tbe/reduce_sum_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/relu.py +0 -39
- mindspore/ops/_op_impl/tbe/relu6.py +0 -38
- mindspore/ops/_op_impl/tbe/relu6_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/relu6_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/relu6_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/relu_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/relu_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/relu_grad_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/relu_grad_v2.py +0 -40
- mindspore/ops/_op_impl/tbe/relu_grad_v2_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/relu_v2.py +0 -40
- mindspore/ops/_op_impl/tbe/relu_v2_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/renorm.py +0 -39
- mindspore/ops/_op_impl/tbe/resize_bilinear.py +0 -40
- mindspore/ops/_op_impl/tbe/resize_bilinear_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/resize_bilinear_v2.py +0 -43
- mindspore/ops/_op_impl/tbe/resize_nearest_neighbor.py +0 -40
- mindspore/ops/_op_impl/tbe/resize_nearest_neighbor_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/resize_nearest_neighbor_grad.py +0 -39
- mindspore/ops/_op_impl/tbe/resize_nearest_neighbor_grad_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/reverse_v2_d.py +0 -37
- mindspore/ops/_op_impl/tbe/rint.py +0 -37
- mindspore/ops/_op_impl/tbe/rint_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/roi_align.py +0 -43
- mindspore/ops/_op_impl/tbe/roi_align_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/roi_align_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/roi_align_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/roll.py +0 -42
- mindspore/ops/_op_impl/tbe/round.py +0 -38
- mindspore/ops/_op_impl/tbe/round_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/rsqrt.py +0 -37
- mindspore/ops/_op_impl/tbe/rsqrt_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/rsqrt_grad.py +0 -40
- mindspore/ops/_op_impl/tbe/rsqrt_grad_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/scatter_add.py +0 -44
- mindspore/ops/_op_impl/tbe/scatter_div.py +0 -46
- mindspore/ops/_op_impl/tbe/scatter_max.py +0 -45
- mindspore/ops/_op_impl/tbe/scatter_min.py +0 -45
- mindspore/ops/_op_impl/tbe/scatter_mul.py +0 -44
- mindspore/ops/_op_impl/tbe/scatter_nd.py +0 -41
- mindspore/ops/_op_impl/tbe/scatter_nd_add.py +0 -45
- mindspore/ops/_op_impl/tbe/scatter_nd_d.py +0 -41
- mindspore/ops/_op_impl/tbe/scatter_nd_ds.py +0 -49
- mindspore/ops/_op_impl/tbe/scatter_nd_sub.py +0 -47
- mindspore/ops/_op_impl/tbe/scatter_nd_sub_ds.py +0 -48
- mindspore/ops/_op_impl/tbe/scatter_nd_update.py +0 -47
- mindspore/ops/_op_impl/tbe/scatter_nd_update_ds.py +0 -48
- mindspore/ops/_op_impl/tbe/scatter_non_aliasing_add.py +0 -39
- mindspore/ops/_op_impl/tbe/scatter_non_aliasing_add_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/scatter_sub.py +0 -47
- mindspore/ops/_op_impl/tbe/scatter_sub_ds.py +0 -48
- mindspore/ops/_op_impl/tbe/scatter_update.py +0 -43
- mindspore/ops/_op_impl/tbe/select.py +0 -38
- mindspore/ops/_op_impl/tbe/select_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/selu.py +0 -39
- mindspore/ops/_op_impl/tbe/selu_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/sgd.py +0 -62
- mindspore/ops/_op_impl/tbe/sigmoid.py +0 -37
- mindspore/ops/_op_impl/tbe/sigmoid_cross_entropy_with_logits.py +0 -41
- mindspore/ops/_op_impl/tbe/sigmoid_cross_entropy_with_logits_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/sigmoid_cross_entropy_with_logits_grad.py +0 -42
- mindspore/ops/_op_impl/tbe/sigmoid_cross_entropy_with_logits_grad_ds.py +0 -43
- mindspore/ops/_op_impl/tbe/sigmoid_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/sigmoid_grad.py +0 -39
- mindspore/ops/_op_impl/tbe/sigmoid_grad_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/sign.py +0 -38
- mindspore/ops/_op_impl/tbe/sign_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/sin.py +0 -37
- mindspore/ops/_op_impl/tbe/sin_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/sinh.py +0 -37
- mindspore/ops/_op_impl/tbe/sinh_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/slice.py +0 -58
- mindspore/ops/_op_impl/tbe/smooth_l1_loss.py +0 -45
- mindspore/ops/_op_impl/tbe/smooth_l1_loss_ds.py +0 -46
- mindspore/ops/_op_impl/tbe/smooth_l1_loss_grad.py +0 -46
- mindspore/ops/_op_impl/tbe/smooth_l1_loss_grad_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/soft_margin_loss.py +0 -38
- mindspore/ops/_op_impl/tbe/soft_margin_loss_grad.py +0 -39
- mindspore/ops/_op_impl/tbe/soft_shrink.py +0 -36
- mindspore/ops/_op_impl/tbe/soft_shrink_grad.py +0 -38
- mindspore/ops/_op_impl/tbe/softmax.py +0 -37
- mindspore/ops/_op_impl/tbe/softmax_cross_entropy_with_logits.py +0 -38
- mindspore/ops/_op_impl/tbe/softmax_cross_entropy_with_logits_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/softmax_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/softmax_grad_ext.py +0 -42
- mindspore/ops/_op_impl/tbe/softmax_v2_with_dropout_do_mask_v3.py +0 -39
- mindspore/ops/_op_impl/tbe/softplus.py +0 -37
- mindspore/ops/_op_impl/tbe/softplus_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/softplus_grad.py +0 -38
- mindspore/ops/_op_impl/tbe/softplus_grad_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/softsign.py +0 -37
- mindspore/ops/_op_impl/tbe/softsign_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/sort.py +0 -38
- mindspore/ops/_op_impl/tbe/sort_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/space_to_batch.py +0 -38
- mindspore/ops/_op_impl/tbe/space_to_batch_nd.py +0 -38
- mindspore/ops/_op_impl/tbe/space_to_depth.py +0 -47
- mindspore/ops/_op_impl/tbe/sparse_apply_adadelta.py +0 -56
- mindspore/ops/_op_impl/tbe/sparse_apply_adagrad.py +0 -45
- mindspore/ops/_op_impl/tbe/sparse_apply_adagrad_ds.py +0 -46
- mindspore/ops/_op_impl/tbe/sparse_apply_adagrad_v2.py +0 -46
- mindspore/ops/_op_impl/tbe/sparse_apply_adagrad_v2_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/sparse_apply_ftrl_d.py +0 -53
- mindspore/ops/_op_impl/tbe/sparse_apply_ftrl_d_ds.py +0 -50
- mindspore/ops/_op_impl/tbe/sparse_apply_ftrl_v2.py +0 -50
- mindspore/ops/_op_impl/tbe/sparse_apply_proximal_adagrad.py +0 -66
- mindspore/ops/_op_impl/tbe/sparse_apply_proximal_adagrad_ds.py +0 -67
- mindspore/ops/_op_impl/tbe/sparse_apply_r_m_s_prop.py +0 -57
- mindspore/ops/_op_impl/tbe/sparse_apply_r_m_s_prop_ds.py +0 -58
- mindspore/ops/_op_impl/tbe/sparse_gather_v2.py +0 -56
- mindspore/ops/_op_impl/tbe/sparse_gather_v2_ds.py +0 -58
- mindspore/ops/_op_impl/tbe/split_d.py +0 -38
- mindspore/ops/_op_impl/tbe/split_d_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/split_v.py +0 -39
- mindspore/ops/_op_impl/tbe/splitv.py +0 -39
- mindspore/ops/_op_impl/tbe/sqrt.py +0 -37
- mindspore/ops/_op_impl/tbe/sqrt_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/sqrt_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/sqrt_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/square.py +0 -38
- mindspore/ops/_op_impl/tbe/square_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/square_sum_all.py +0 -40
- mindspore/ops/_op_impl/tbe/square_sum_all_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/square_sum_v1.py +0 -38
- mindspore/ops/_op_impl/tbe/square_sum_v1_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/square_sum_v2.py +0 -39
- mindspore/ops/_op_impl/tbe/squared_difference.py +0 -39
- mindspore/ops/_op_impl/tbe/squared_difference_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/squeeze.py +0 -37
- mindspore/ops/_op_impl/tbe/strided_read.py +0 -38
- mindspore/ops/_op_impl/tbe/strided_slice_d.py +0 -44
- mindspore/ops/_op_impl/tbe/strided_slice_ds.py +0 -71
- mindspore/ops/_op_impl/tbe/strided_slice_grad_d.py +0 -51
- mindspore/ops/_op_impl/tbe/strided_slice_grad_ds.py +0 -57
- mindspore/ops/_op_impl/tbe/strided_write.py +0 -38
- mindspore/ops/_op_impl/tbe/sub.py +0 -39
- mindspore/ops/_op_impl/tbe/sub_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/tan.py +0 -38
- mindspore/ops/_op_impl/tbe/tan_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/tanh.py +0 -37
- mindspore/ops/_op_impl/tbe/tanh_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/tanh_grad.py +0 -39
- mindspore/ops/_op_impl/tbe/tanh_grad_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/tensor_move.py +0 -49
- mindspore/ops/_op_impl/tbe/tensor_move_ds.py +0 -50
- mindspore/ops/_op_impl/tbe/tensor_scatter_update.py +0 -41
- mindspore/ops/_op_impl/tbe/tile.py +0 -37
- mindspore/ops/_op_impl/tbe/tile_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/top_k.py +0 -42
- mindspore/ops/_op_impl/tbe/top_k_ds.py +0 -43
- mindspore/ops/_op_impl/tbe/trans_data.py +0 -167
- mindspore/ops/_op_impl/tbe/trans_data_ds.py +0 -180
- mindspore/ops/_op_impl/tbe/trans_data_rnn.py +0 -44
- mindspore/ops/_op_impl/tbe/transpose.py +0 -60
- mindspore/ops/_op_impl/tbe/transpose_d.py +0 -47
- mindspore/ops/_op_impl/tbe/transpose_nod.py +0 -60
- mindspore/ops/_op_impl/tbe/trunc.py +0 -39
- mindspore/ops/_op_impl/tbe/truncate_div.py +0 -41
- mindspore/ops/_op_impl/tbe/truncate_div_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/truncate_mod.py +0 -41
- mindspore/ops/_op_impl/tbe/truncate_mod_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/unpack.py +0 -38
- mindspore/ops/_op_impl/tbe/unpack_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/unsorted_segment_max.py +0 -49
- mindspore/ops/_op_impl/tbe/unsorted_segment_max_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/unsorted_segment_min.py +0 -49
- mindspore/ops/_op_impl/tbe/unsorted_segment_min_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/unsorted_segment_prod.py +0 -49
- mindspore/ops/_op_impl/tbe/unsorted_segment_prod_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/unsorted_segment_sum.py +0 -38
- mindspore/ops/_op_impl/tbe/unsorted_segment_sum_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/wts_arq.py +0 -40
- mindspore/ops/_op_impl/tbe/xdivy.py +0 -38
- mindspore/ops/_op_impl/tbe/xdivy_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/xlogy.py +0 -38
- mindspore/ops/_op_impl/tbe/xlogy_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/zeros_like.py +0 -41
- mindspore/ops/_op_impl/tbe/zeros_like_ds.py +0 -42
- mindspore/ops/_tracefunc.py +0 -241
- mindspore/ops/arg_dtype_cast.py +0 -54
- mindspore/rewrite/api/tree_node_helper.py +0 -60
- mindspore/rewrite/ast_helpers/ast_creator.py +0 -115
- mindspore/rewrite/ast_transformers/flatten_recursive_stmt.py +0 -267
- mindspore/rewrite/ast_transformers/remove_return_out_of_if.py +0 -228
- mindspore/rewrite/namespace.py +0 -53
- mindspore-2.2.11.dist-info/RECORD +0 -1920
- {mindspore-2.2.11.dist-info → mindspore-2.3.0.dist-info}/WHEEL +0 -0
- {mindspore-2.2.11.dist-info → mindspore-2.3.0.dist-info}/top_level.txt +0 -0
|
@@ -21,7 +21,6 @@ import numbers
|
|
|
21
21
|
import hashlib
|
|
22
22
|
|
|
23
23
|
from mindspore.ops import operations as P
|
|
24
|
-
from mindspore.ops import functional as F
|
|
25
24
|
from mindspore.ops.operations import _inner_ops as inner
|
|
26
25
|
from mindspore.common.parameter import Parameter
|
|
27
26
|
from mindspore.common.initializer import initializer, Initializer
|
|
@@ -36,6 +35,7 @@ from mindspore.common import dtype as mstype
|
|
|
36
35
|
from mindspore.parallel._utils import _is_in_auto_parallel_mode
|
|
37
36
|
from mindspore.nn.cell import Cell
|
|
38
37
|
from mindspore import log as logger
|
|
38
|
+
from mindspore.ops import group_norm
|
|
39
39
|
|
|
40
40
|
__all__ = ['BatchNorm1d', 'BatchNorm2d', 'BatchNorm3d', 'LayerNorm', 'GroupNorm',
|
|
41
41
|
'SyncBatchNorm', 'InstanceNorm1d', 'InstanceNorm2d', 'InstanceNorm3d']
|
|
@@ -128,15 +128,19 @@ class _BatchNorm(Cell):
|
|
|
128
128
|
self.assign_sub_mean = P.AssignSub().shard(data_parallel_strategy)
|
|
129
129
|
self.assign_sub_var = P.AssignSub().shard(data_parallel_strategy)
|
|
130
130
|
|
|
131
|
-
|
|
132
131
|
@staticmethod
|
|
133
132
|
@_primexpr
|
|
134
133
|
def _check_input_dim(shape, cls_name):
|
|
135
134
|
raise NotImplementedError
|
|
136
135
|
|
|
137
|
-
|
|
138
136
|
def construct(self, x):
|
|
139
137
|
self._check_input_dim(self.shape(x), self.cls_name)
|
|
138
|
+
x_shape = self.shape(x)
|
|
139
|
+
reshaped_x = x
|
|
140
|
+
if len(x_shape) == 2:
|
|
141
|
+
reshaped_x = self.reshape(x, (x_shape[0], x_shape[1], 1, 1))
|
|
142
|
+
elif len(x_shape) == 3:
|
|
143
|
+
reshaped_x = self.reshape(x, (x_shape[0], x_shape[1], x_shape[2], 1))
|
|
140
144
|
if self.use_batch_statistics is None:
|
|
141
145
|
if self.training:
|
|
142
146
|
return self.bn_train(x,
|
|
@@ -145,11 +149,14 @@ class _BatchNorm(Cell):
|
|
|
145
149
|
self.moving_mean,
|
|
146
150
|
self.moving_variance)[0]
|
|
147
151
|
if not self.training:
|
|
148
|
-
|
|
149
|
-
|
|
150
|
-
|
|
151
|
-
|
|
152
|
-
|
|
152
|
+
bn_out = self.bn_infer(reshaped_x,
|
|
153
|
+
self.gamma,
|
|
154
|
+
self.beta,
|
|
155
|
+
self.moving_mean,
|
|
156
|
+
self.moving_variance)[0]
|
|
157
|
+
if len(x_shape) < 4:
|
|
158
|
+
bn_out = self.reshape(bn_out, x_shape)
|
|
159
|
+
return bn_out
|
|
153
160
|
|
|
154
161
|
if self.use_batch_statistics:
|
|
155
162
|
return self.bn_train(x,
|
|
@@ -158,11 +165,14 @@ class _BatchNorm(Cell):
|
|
|
158
165
|
self.moving_mean,
|
|
159
166
|
self.moving_variance)[0]
|
|
160
167
|
|
|
161
|
-
|
|
162
|
-
|
|
163
|
-
|
|
164
|
-
|
|
165
|
-
|
|
168
|
+
bn_out = self.bn_infer(reshaped_x,
|
|
169
|
+
self.gamma,
|
|
170
|
+
self.beta,
|
|
171
|
+
self.moving_mean,
|
|
172
|
+
self.moving_variance)[0]
|
|
173
|
+
if len(x_shape) < 4:
|
|
174
|
+
bn_out = self.reshape(bn_out, x_shape)
|
|
175
|
+
return bn_out
|
|
166
176
|
|
|
167
177
|
def extend_repr(self):
|
|
168
178
|
return 'num_features={}, eps={}, momentum={}, gamma={}, beta={}, moving_mean={}, moving_variance={}'.format(
|
|
@@ -196,19 +206,19 @@ class BatchNorm1d(_BatchNorm):
|
|
|
196
206
|
Default: ``True`` .
|
|
197
207
|
gamma_init (Union[Tensor, str, Initializer, numbers.Number]): Initializer for the :math:`\gamma` weight.
|
|
198
208
|
The values of str refer to the function `mindspore.common.initializer
|
|
199
|
-
<https://www.mindspore.cn/docs/en/
|
|
209
|
+
<https://www.mindspore.cn/docs/en/master/api_python/mindspore.common.initializer.html>`_
|
|
200
210
|
including ``'zeros'`` , ``'ones'`` , etc. Default: ``'ones'`` .
|
|
201
211
|
beta_init (Union[Tensor, str, Initializer, numbers.Number]): Initializer for the :math:`\beta` weight.
|
|
202
212
|
The values of str refer to the function `mindspore.common.initializer
|
|
203
|
-
<https://www.mindspore.cn/docs/en/
|
|
213
|
+
<https://www.mindspore.cn/docs/en/master/api_python/mindspore.common.initializer.html>`_
|
|
204
214
|
including ``'zeros'`` , ``'ones'``, etc. Default: ``'zeros'`` .
|
|
205
215
|
moving_mean_init (Union[Tensor, str, Initializer, numbers.Number]): Initializer for the moving mean.
|
|
206
216
|
The values of str refer to the function `mindspore.common.initializer
|
|
207
|
-
<https://www.mindspore.cn/docs/en/
|
|
217
|
+
<https://www.mindspore.cn/docs/en/master/api_python/mindspore.common.initializer.html>`_
|
|
208
218
|
including ``'zeros'`` , ``'ones'`` , etc. Default: ``'zeros'`` .
|
|
209
219
|
moving_var_init (Union[Tensor, str, Initializer, numbers.Number]): Initializer for the moving variance.
|
|
210
220
|
The values of str refer to the function `mindspore.common.initializer
|
|
211
|
-
<https://www.mindspore.cn/docs/en/
|
|
221
|
+
<https://www.mindspore.cn/docs/en/master/api_python/mindspore.common.initializer.html>`_
|
|
212
222
|
including ``'zeros'`` , ``'ones'`` , etc. Default: ``'ones'`` .
|
|
213
223
|
use_batch_statistics (bool): If ``true`` , use the mean value and variance value of current batch data. If
|
|
214
224
|
``false`` , use the mean value and variance value of specified value. If ``None`` , the training process
|
|
@@ -292,19 +302,19 @@ class BatchNorm2d(_BatchNorm):
|
|
|
292
302
|
Default: ``True`` .
|
|
293
303
|
gamma_init (Union[Tensor, str, Initializer, numbers.Number]): Initializer for the :math:`\gamma` weight.
|
|
294
304
|
The values of str refer to the function `mindspore.common.initializer
|
|
295
|
-
<https://www.mindspore.cn/docs/en/
|
|
305
|
+
<https://www.mindspore.cn/docs/en/master/api_python/mindspore.common.initializer.html>`_
|
|
296
306
|
including ``'zeros'`` , ``'ones'`` , etc. Default: ``'ones'`` .
|
|
297
307
|
beta_init (Union[Tensor, str, Initializer, numbers.Number]): Initializer for the :math:`\beta` weight.
|
|
298
308
|
The values of str refer to the function `mindspore.common.initializer
|
|
299
|
-
<https://www.mindspore.cn/docs/en/
|
|
309
|
+
<https://www.mindspore.cn/docs/en/master/api_python/mindspore.common.initializer.html>`_
|
|
300
310
|
including ``'zeros'`` , ``'ones'`` , etc. Default: ``'zeros'`` .
|
|
301
311
|
moving_mean_init (Union[Tensor, str, Initializer, numbers.Number]): Initializer for the moving mean.
|
|
302
312
|
The values of str refer to the function `mindspore.common.initializer
|
|
303
|
-
<https://www.mindspore.cn/docs/en/
|
|
313
|
+
<https://www.mindspore.cn/docs/en/master/api_python/mindspore.common.initializer.html>`_
|
|
304
314
|
including ``'zeros'`` , ``'ones'`` , etc. Default: ``'zeros'`` .
|
|
305
315
|
moving_var_init (Union[Tensor, str, Initializer, numbers.Number]): Initializer for the moving variance.
|
|
306
316
|
The values of str refer to the function `mindspore.common.initializer
|
|
307
|
-
<https://www.mindspore.cn/docs/en/
|
|
317
|
+
<https://www.mindspore.cn/docs/en/master/api_python/mindspore.common.initializer.html>`_
|
|
308
318
|
including ``'zeros'`` , ``'ones'`` , etc. Default: ``'ones'`` .
|
|
309
319
|
use_batch_statistics (bool): Default: ``None`` .
|
|
310
320
|
|
|
@@ -381,19 +391,19 @@ class BatchNorm3d(Cell):
|
|
|
381
391
|
affine (bool): A bool value. When set to ``True`` , gamma and beta can be learned. Default: ``True`` .
|
|
382
392
|
gamma_init (Union[Tensor, str, Initializer, numbers.Number]): Initializer for the gamma weight.
|
|
383
393
|
The values of str refer to the function `mindspore.common.initializer
|
|
384
|
-
<https://www.mindspore.cn/docs/en/
|
|
394
|
+
<https://www.mindspore.cn/docs/en/master/api_python/mindspore.common.initializer.html>`_
|
|
385
395
|
including ``'zeros'`` , ``'ones'`` , etc. Default: ``'ones'`` .
|
|
386
396
|
beta_init (Union[Tensor, str, Initializer, numbers.Number]): Initializer for the beta weight.
|
|
387
397
|
The values of str refer to the function `mindspore.common.initializer
|
|
388
|
-
<https://www.mindspore.cn/docs/en/
|
|
398
|
+
<https://www.mindspore.cn/docs/en/master/api_python/mindspore.common.initializer.html>`_
|
|
389
399
|
including ``'zeros'`` , ``'ones'`` , etc. Default: ``'zeros'`` .
|
|
390
400
|
moving_mean_init (Union[Tensor, str, Initializer, numbers.Number]): Initializer for the moving mean.
|
|
391
401
|
The values of str refer to the function `mindspore.common.initializer
|
|
392
|
-
<https://www.mindspore.cn/docs/en/
|
|
402
|
+
<https://www.mindspore.cn/docs/en/master/api_python/mindspore.common.initializer.html>`_
|
|
393
403
|
including ``'zeros'`` , ``'ones'`` , etc. Default: ``'zeros'`` .
|
|
394
404
|
moving_var_init (Union[Tensor, str, Initializer, numbers.Number]): Initializer for the moving variance.
|
|
395
405
|
The values of str refer to the function `mindspore.common.initializer
|
|
396
|
-
<https://www.mindspore.cn/docs/en/
|
|
406
|
+
<https://www.mindspore.cn/docs/en/master/api_python/mindspore.common.initializer.html>`_
|
|
397
407
|
including ``'zeros'`` , ``'ones'`` , etc. Default: ``'ones'`` .
|
|
398
408
|
use_batch_statistics (bool): If true, use the mean value and variance value of current batch data. If
|
|
399
409
|
``false``, use the mean value and variance value of specified value. If ``None`` , the training process
|
|
@@ -460,7 +470,6 @@ class BatchNorm3d(Cell):
|
|
|
460
470
|
dim = len(shape)
|
|
461
471
|
_check_dim(dim, 5, cls_name)
|
|
462
472
|
|
|
463
|
-
|
|
464
473
|
def construct(self, x):
|
|
465
474
|
x_shape = self.shape(x)
|
|
466
475
|
self._check_input_dim(x_shape, self.cls_name)
|
|
@@ -549,14 +558,14 @@ class SyncBatchNorm(_BatchNorm):
|
|
|
549
558
|
|
|
550
559
|
For the Ascend devices, users need to prepare the rank table, set rank_id and device_id.
|
|
551
560
|
Please see the `Ascend tutorial
|
|
552
|
-
<https://www.mindspore.cn/tutorials/experts/en/
|
|
561
|
+
<https://www.mindspore.cn/tutorials/experts/en/master/parallel/rank_table.html>`_
|
|
553
562
|
for more details.
|
|
554
563
|
|
|
555
564
|
For the GPU devices, users need to prepare the host file and mpi, please see the `mpirun Startup
|
|
556
|
-
<https://www.mindspore.cn/tutorials/experts/en/
|
|
565
|
+
<https://www.mindspore.cn/tutorials/experts/en/master/parallel/mpirun.html>`_ .
|
|
557
566
|
|
|
558
567
|
For the CPU device, users need to write a dynamic cluster startup script, please see the `Dynamic Cluster
|
|
559
|
-
Startup <https://www.mindspore.cn/tutorials/experts/en/
|
|
568
|
+
Startup <https://www.mindspore.cn/tutorials/experts/en/master/parallel/dynamic_cluster.html>`_ .
|
|
560
569
|
|
|
561
570
|
This example should be run with multiple devices.
|
|
562
571
|
|
|
@@ -649,16 +658,6 @@ class SyncBatchNorm(_BatchNorm):
|
|
|
649
658
|
self.group_name = group_dict[rank_list_name]
|
|
650
659
|
logger.info("the group for {} already exists, no need to create".format(rank_list_name))
|
|
651
660
|
|
|
652
|
-
@staticmethod
|
|
653
|
-
@_primexpr
|
|
654
|
-
def _check_input_dim(shape, cls_name):
|
|
655
|
-
def _check(dim):
|
|
656
|
-
if dim not in (2, 4):
|
|
657
|
-
raise ValueError(f"For '{cls_name}', the must have 2 dims or 4 dims, but got {dim}.")
|
|
658
|
-
dim = len(shape)
|
|
659
|
-
_check(dim)
|
|
660
|
-
|
|
661
|
-
|
|
662
661
|
def _check_rank_ids(self, process_groups, rank_size):
|
|
663
662
|
seen = set()
|
|
664
663
|
for rid in itertools.chain(*process_groups):
|
|
@@ -668,6 +667,15 @@ class SyncBatchNorm(_BatchNorm):
|
|
|
668
667
|
f"but got {process_groups}.")
|
|
669
668
|
seen.add(rid)
|
|
670
669
|
|
|
670
|
+
@staticmethod
|
|
671
|
+
@_primexpr
|
|
672
|
+
def _check_input_dim(shape, cls_name):
|
|
673
|
+
def _check(dim):
|
|
674
|
+
if dim not in (2, 4):
|
|
675
|
+
raise ValueError(f"For '{cls_name}', the must have 2 dims or 4 dims, but got {dim}.")
|
|
676
|
+
dim = len(shape)
|
|
677
|
+
_check(dim)
|
|
678
|
+
|
|
671
679
|
|
|
672
680
|
class LayerNorm(Cell):
|
|
673
681
|
r"""
|
|
@@ -1092,7 +1100,8 @@ class GroupNorm(Cell):
|
|
|
1092
1100
|
dtype (:class:`mindspore.dtype`): Dtype of Parameters. Default: ``mstype.float32`` .
|
|
1093
1101
|
|
|
1094
1102
|
Inputs:
|
|
1095
|
-
- **x** (Tensor) - The input feature with shape :math:`(N, C,
|
|
1103
|
+
- **x** (Tensor) - The input feature with shape :math:`(N, C, *)`, where :math:`*` means, any number of
|
|
1104
|
+
additional dimensions.
|
|
1096
1105
|
|
|
1097
1106
|
Outputs:
|
|
1098
1107
|
Tensor, the normalized and scaled offset tensor, has the same shape and data type as the `x`.
|
|
@@ -1137,34 +1146,13 @@ class GroupNorm(Cell):
|
|
|
1137
1146
|
self.affine = validator.check_bool(affine, arg_name="affine", prim_name=self.cls_name)
|
|
1138
1147
|
|
|
1139
1148
|
self.gamma = Parameter(initializer(
|
|
1140
|
-
gamma_init, num_channels, dtype=dtype), name="gamma", requires_grad=affine)
|
|
1149
|
+
gamma_init, self.num_channels, dtype=dtype), name="gamma", requires_grad=affine)
|
|
1141
1150
|
self.beta = Parameter(initializer(
|
|
1142
|
-
beta_init, num_channels, dtype=dtype), name="beta", requires_grad=affine)
|
|
1143
|
-
self.reduce_mean = P.ReduceMean(keep_dims=True)
|
|
1144
|
-
self.reduce_sum = P.ReduceSum(keep_dims=True)
|
|
1145
|
-
self.shape = F.shape
|
|
1146
|
-
self.reshape = F.reshape
|
|
1147
|
-
self.square = F.square
|
|
1148
|
-
self.sqrt = P.Sqrt()
|
|
1151
|
+
beta_init, self.num_channels, dtype=dtype), name="beta", requires_grad=affine)
|
|
1149
1152
|
|
|
1150
1153
|
def _cal_output(self, x):
|
|
1151
1154
|
"""calculate groupnorm output"""
|
|
1152
|
-
|
|
1153
|
-
self._channel_check(channel, self.num_channels, self.cls_name)
|
|
1154
|
-
x = F.reshape(x, (batch, self.num_groups, -1))
|
|
1155
|
-
mean = self.reduce_mean(x, 2)
|
|
1156
|
-
var = F.div(self.reduce_sum(F.square(F.sub(x, mean)), 2), (channel * height * width / self.num_groups))
|
|
1157
|
-
std = self.sqrt(var + self.eps)
|
|
1158
|
-
x = F.div(F.sub(x, mean), std)
|
|
1159
|
-
x = F.reshape(x, (batch, channel, height, width))
|
|
1160
|
-
output = F.add(x * F.reshape(self.gamma, (-1, 1, 1)), F.reshape(self.beta, (-1, 1, 1)))
|
|
1161
|
-
return output
|
|
1162
|
-
|
|
1163
|
-
@staticmethod
|
|
1164
|
-
@_primexpr
|
|
1165
|
-
def _check_input_dim(shape, cls_name):
|
|
1166
|
-
dim = len(shape)
|
|
1167
|
-
_check_dim(dim, 4, cls_name)
|
|
1155
|
+
return group_norm(x, self.num_groups, self.gamma.to(x.dtype), self.beta.to(x.dtype), self.eps)
|
|
1168
1156
|
|
|
1169
1157
|
@staticmethod
|
|
1170
1158
|
@_primexpr
|
|
@@ -1185,7 +1173,5 @@ class GroupNorm(Cell):
|
|
|
1185
1173
|
return 'num_groups={}, num_channels={}'.format(self.num_groups, self.num_channels)
|
|
1186
1174
|
|
|
1187
1175
|
def construct(self, x):
|
|
1188
|
-
self._check_input_dim(F.shape(x), self.cls_name)
|
|
1189
|
-
self._check_dtype(x.dtype, [mstype.float16, mstype.float32], self.cls_name)
|
|
1190
1176
|
output = self._cal_output(x)
|
|
1191
1177
|
return output
|
mindspore/nn/layer/padding.py
CHANGED
|
@@ -208,24 +208,11 @@ class _ConstantPadNd(Cell):
|
|
|
208
208
|
raise TypeError(msg)
|
|
209
209
|
|
|
210
210
|
self.value = value
|
|
211
|
-
self.padding =
|
|
212
|
-
self._name = name
|
|
211
|
+
self.padding = padding
|
|
213
212
|
|
|
214
213
|
def construct(self, x):
|
|
215
214
|
"""Construct the pad net."""
|
|
216
|
-
|
|
217
|
-
padding = _check(input_shape, self.padding, self._name)
|
|
218
|
-
new_padding, start, end = _get_new_padding(padding)
|
|
219
|
-
mask = ops.OnesLike()(x)
|
|
220
|
-
output = ops.Pad(new_padding)(x)
|
|
221
|
-
mask = ops.Pad(new_padding)(mask)
|
|
222
|
-
ones = ops.OnesLike()(output)
|
|
223
|
-
value = ops.fill(output.dtype, output.shape, self.value)
|
|
224
|
-
output = ops.Add()(ops.Mul()(mask, output), ops.Mul()(ops.Sub()(ones, mask), value))
|
|
225
|
-
slice_op = ops.Slice()
|
|
226
|
-
begin, size = _get_begin_size(output.shape, start, end)
|
|
227
|
-
output = slice_op(output, begin, size)
|
|
228
|
-
return output
|
|
215
|
+
return ops.pad(x, padding=self.padding, mode='constant', value=self.value)
|
|
229
216
|
|
|
230
217
|
|
|
231
218
|
class ConstantPad1d(_ConstantPadNd):
|
|
@@ -238,8 +225,13 @@ class ConstantPad1d(_ConstantPadNd):
|
|
|
238
225
|
If a 2-tuple, uses (padding_0, padding_1) to pad. If the input is `x`, the size of last
|
|
239
226
|
dimension of output is :math:`padding\_0 + x.shape[-1] + padding\_1`. The remaining dimensions
|
|
240
227
|
of the output are consistent with those of the input.
|
|
228
|
+
Only support non-negative value while running in Ascend.
|
|
241
229
|
value (Union[int, float]): Padding value.
|
|
242
230
|
|
|
231
|
+
Inputs:
|
|
232
|
+
- **x** (Tensor) - shape is :math:`(N, *)`, where :math:`*` means, any number of additional dimensions.
|
|
233
|
+
It is not supported that the size of dimensions is greater than 5 while running on Ascend.
|
|
234
|
+
|
|
243
235
|
Returns:
|
|
244
236
|
Tensor, the tensor after padding.
|
|
245
237
|
|
|
@@ -248,6 +240,8 @@ class ConstantPad1d(_ConstantPadNd):
|
|
|
248
240
|
TypeError: If `value` is not int or float.
|
|
249
241
|
ValueError: If the length of `padding` with tuple type is not equal to 2.
|
|
250
242
|
ValueError: If the output shape after padding is not positive.
|
|
243
|
+
ValueError: If the rank of 'x' is more than 5 while running in Ascend.
|
|
244
|
+
ValueError: If `padding` contains negative value while running in Ascend.
|
|
251
245
|
|
|
252
246
|
Supported Platforms:
|
|
253
247
|
``Ascend`` ``GPU`` ``CPU``
|
|
@@ -316,8 +310,13 @@ class ConstantPad2d(_ConstantPadNd):
|
|
|
316
310
|
If the input is `x`, the size of last dimension of output is :math:`padding\_0 + x.shape[-1] + padding\_1`.
|
|
317
311
|
The size of penultimate dimension of output is :math:`padding\_2 + x.shape[-2] + padding\_3`.
|
|
318
312
|
The remaining dimensions of the output are consistent with those of the input.
|
|
313
|
+
Only support non-negative value while running in Ascend.
|
|
319
314
|
value (Union[int, float]): Padding value.
|
|
320
315
|
|
|
316
|
+
Inputs:
|
|
317
|
+
- **x** (Tensor) - shape is :math:`(N, *)`, where :math:`*` means, any number of additional dimensions.
|
|
318
|
+
It is not supported that the size of dimensions is greater than 5 while running on Ascend.
|
|
319
|
+
|
|
321
320
|
Returns:
|
|
322
321
|
Tensor, the tensor after padding.
|
|
323
322
|
|
|
@@ -326,6 +325,8 @@ class ConstantPad2d(_ConstantPadNd):
|
|
|
326
325
|
TypeError: If `value` is not int or float.
|
|
327
326
|
ValueError: If the length of `padding` is more than 4 or not a multiple of 2.
|
|
328
327
|
ValueError: If the output shape after padding is not positive.
|
|
328
|
+
ValueError: If the rank of 'x' is more than 5 while running in Ascend.
|
|
329
|
+
ValueError: If `padding` contains negative value while running in Ascend.
|
|
329
330
|
|
|
330
331
|
Supported Platforms:
|
|
331
332
|
``Ascend`` ``GPU`` ``CPU``
|
|
@@ -369,8 +370,13 @@ class ConstantPad3d(_ConstantPadNd):
|
|
|
369
370
|
The size of penultimate dimension of output is :math:`padding\_2 + x.shape[-2] + padding\_3`.
|
|
370
371
|
The size of 3rd to last dimension of output is :math:`padding\_4 + x.shape[-3] + padding\_5`.
|
|
371
372
|
The remaining dimensions of the output are consistent with those of the input.
|
|
373
|
+
Only support non-negative value while running in Ascend.
|
|
372
374
|
value (Union[int, float]): Padding value.
|
|
373
375
|
|
|
376
|
+
Inputs:
|
|
377
|
+
- **x** (Tensor) - shape is :math:`(N, *)`, where :math:`*` means, any number of additional dimensions.
|
|
378
|
+
It is not supported that the size of dimensions is greater than 5 while running on Ascend.
|
|
379
|
+
|
|
374
380
|
Returns:
|
|
375
381
|
Tensor, the tensor after padding.
|
|
376
382
|
|
|
@@ -379,6 +385,8 @@ class ConstantPad3d(_ConstantPadNd):
|
|
|
379
385
|
TypeError: If `value` is not int or float.
|
|
380
386
|
ValueError: If the length of `padding` is more than 6 or not a multiple of 2.
|
|
381
387
|
ValueError: If the output shape after padding is not positive.
|
|
388
|
+
ValueError: If the rank of 'x' is more than 5 while running in Ascend.
|
|
389
|
+
ValueError: If `padding` contains negative value while running in Ascend.
|
|
382
390
|
|
|
383
391
|
Supported Platforms:
|
|
384
392
|
``Ascend`` ``GPU`` ``CPU``
|
|
@@ -620,6 +628,11 @@ class ZeroPad2d(_ConstantPadNd):
|
|
|
620
628
|
If the input is `x`, the size of last dimension of output is :math:`padding\_0 + x.shape[-1] + padding\_1`.
|
|
621
629
|
The size of penultimate dimension of output is :math:`padding\_2 + x.shape[-2] + padding\_3`.
|
|
622
630
|
The remaining dimensions of the output are consistent with those of the input.
|
|
631
|
+
Only support non-negative value while running in Ascend.
|
|
632
|
+
|
|
633
|
+
Inputs:
|
|
634
|
+
- **x** (Tensor) - shape is :math:`(N, *)`, where :math:`*` means, any number of additional dimensions.
|
|
635
|
+
It is not supported that the size of dimensions is greater than 5 while running in Ascend.
|
|
623
636
|
|
|
624
637
|
Returns:
|
|
625
638
|
Tensor, the tensor after padding.
|
|
@@ -628,6 +641,8 @@ class ZeroPad2d(_ConstantPadNd):
|
|
|
628
641
|
TypeError: If `padding` is not a tuple or int.
|
|
629
642
|
ValueError: If the length of `padding` is more than 4 or not a multiple of 2.
|
|
630
643
|
ValueError: If the output shape after padding is not positive.
|
|
644
|
+
ValueError: If the rank of 'x' is more than 5 while running in Ascend.
|
|
645
|
+
ValueError: If `padding` contains negative value while running in Ascend.
|
|
631
646
|
|
|
632
647
|
Supported Platforms:
|
|
633
648
|
``Ascend`` ``GPU`` ``CPU``
|
|
@@ -679,18 +694,12 @@ class _ReplicationPadNd(Cell):
|
|
|
679
694
|
self.padding = padding
|
|
680
695
|
self.padv3 = nn_ops.PadV3(mode="edge")
|
|
681
696
|
|
|
682
|
-
@staticmethod
|
|
683
|
-
@_primexpr
|
|
684
|
-
def _check_input_dim(shape, cls_name):
|
|
685
|
-
raise NotImplementedError
|
|
686
|
-
|
|
687
697
|
@staticmethod
|
|
688
698
|
@constexpr
|
|
689
699
|
def _need_expend_dim(x):
|
|
690
700
|
raise NotImplementedError
|
|
691
701
|
|
|
692
702
|
def construct(self, x):
|
|
693
|
-
self._check_input_dim(x.shape, self.name)
|
|
694
703
|
need_expend_dims = self._need_expend_dim(x)
|
|
695
704
|
if need_expend_dims:
|
|
696
705
|
x = x.expand_dims(0)
|
|
@@ -751,12 +760,6 @@ class ReplicationPad1d(_ReplicationPadNd):
|
|
|
751
760
|
padding = (padding, padding)
|
|
752
761
|
super(ReplicationPad1d, self).__init__(padding, name="ReplicationPad1d")
|
|
753
762
|
|
|
754
|
-
@staticmethod
|
|
755
|
-
@_primexpr
|
|
756
|
-
def _check_input_dim(shape, cls_name):
|
|
757
|
-
dim = len(shape)
|
|
758
|
-
_check_dim(dim, 2, 3, cls_name)
|
|
759
|
-
|
|
760
763
|
def _need_expend_dim(self, x):
|
|
761
764
|
input_shape = x.shape
|
|
762
765
|
return 1 if len(input_shape) == 2 else 0
|
|
@@ -821,12 +824,6 @@ class ReplicationPad2d(_ReplicationPadNd):
|
|
|
821
824
|
padding = (padding, padding, padding, padding)
|
|
822
825
|
super(ReplicationPad2d, self).__init__(padding, name="ReplicationPad2d")
|
|
823
826
|
|
|
824
|
-
@staticmethod
|
|
825
|
-
@_primexpr
|
|
826
|
-
def _check_input_dim(shape, cls_name):
|
|
827
|
-
dim = len(shape)
|
|
828
|
-
_check_dim(dim, 3, 4, cls_name)
|
|
829
|
-
|
|
830
827
|
def _need_expend_dim(self, x):
|
|
831
828
|
input_shape = x.shape
|
|
832
829
|
return 1 if len(input_shape) == 3 else 0
|
|
@@ -892,12 +889,6 @@ class ReplicationPad3d(_ReplicationPadNd):
|
|
|
892
889
|
padding = (padding, padding, padding, padding, padding, padding)
|
|
893
890
|
super(ReplicationPad3d, self).__init__(padding, name="ReplicationPad3d")
|
|
894
891
|
|
|
895
|
-
@staticmethod
|
|
896
|
-
@_primexpr
|
|
897
|
-
def _check_input_dim(shape, cls_name):
|
|
898
|
-
dim = len(shape)
|
|
899
|
-
_check_dim(dim, 4, 5, cls_name)
|
|
900
|
-
|
|
901
892
|
def _need_expend_dim(self, x):
|
|
902
893
|
input_shape = x.shape
|
|
903
894
|
return 1 if len(input_shape) == 4 else 0
|
mindspore/nn/layer/pooling.py
CHANGED
|
@@ -27,6 +27,7 @@ from mindspore.common import dtype as mstype
|
|
|
27
27
|
from mindspore.ops.operations.nn_ops import AdaptiveMaxPool2D
|
|
28
28
|
from mindspore.ops.operations.nn_ops import AdaptiveMaxPool3D, AdaptiveAvgPool3D
|
|
29
29
|
from mindspore.nn.cell import Cell
|
|
30
|
+
from mindspore._c_expression import MSContext
|
|
30
31
|
|
|
31
32
|
__all__ = ['AvgPool3d', 'MaxPool3d', 'AvgPool2d', 'MaxPool2d', 'AvgPool1d', 'MaxPool1d', 'FractionalMaxPool2d',
|
|
32
33
|
'FractionalMaxPool3d', 'AdaptiveAvgPool1d', 'AdaptiveMaxPool1d', 'AdaptiveMaxPool2d', 'AdaptiveMaxPool3d',
|
|
@@ -105,7 +106,8 @@ class LPPool1d(Cell):
|
|
|
105
106
|
kernel_size (int): The size of kernel window.
|
|
106
107
|
stride (int): The distance of kernel moving, an int number that represents the width of movement is stride,
|
|
107
108
|
if the value is None, the default value `kernel_size` is used. Default: ``None`` .
|
|
108
|
-
ceil_mode (bool):
|
|
109
|
+
ceil_mode (bool): If ``True``, use ceil to calculate output shape.
|
|
110
|
+
If ``False``, use ceil to calculate output shape. Default: ``False`` .
|
|
109
111
|
|
|
110
112
|
Inputs:
|
|
111
113
|
- **x** (Tensor) - Tensor of shape :math:`(N_{in}, C_{in}, L_{in})` or :math:`(C_{in}, L_{in})`.
|
|
@@ -119,7 +121,7 @@ class LPPool1d(Cell):
|
|
|
119
121
|
|
|
120
122
|
|
|
121
123
|
Raises:
|
|
122
|
-
TypeError: If `x` is not
|
|
124
|
+
TypeError: If `x` is not a Tensor.
|
|
123
125
|
TypeError: If `kernel_size` or `stride` is not an int.
|
|
124
126
|
TypeError: If `ceil_mode` is not a bool.
|
|
125
127
|
TypeError: If `norm_type` is neither float nor int.
|
|
@@ -168,7 +170,7 @@ class LPPool2d(Cell):
|
|
|
168
170
|
f(X) = \sqrt[p]{\sum_{x \in X} x^{p}}
|
|
169
171
|
|
|
170
172
|
Args:
|
|
171
|
-
norm_type(Union[int, float])
|
|
173
|
+
norm_type(Union[int, float]): Type of normalization, represents :math:`p` in the formula, can not be 0.
|
|
172
174
|
|
|
173
175
|
- if p = 1, the result is the sum of the elements within the pooling kernel(proportional to average
|
|
174
176
|
pooling).
|
|
@@ -179,7 +181,7 @@ class LPPool2d(Cell):
|
|
|
179
181
|
or a tuple of two int numbers that represent height and width respectively.
|
|
180
182
|
stride(Union[int, tuple[int]]): The distance of kernel moving, an int number that represents
|
|
181
183
|
the height and width of movement are both stride, or a tuple of two int numbers that
|
|
182
|
-
represent height and width of movement respectively, if the value is None
|
|
184
|
+
represent height and width of movement respectively, if the value is ``None``,
|
|
183
185
|
the default value `kernel_size` is used. Default: ``None`` .
|
|
184
186
|
ceil_mode(bool): Whether to use ceil or floor to calculate output shape. Default: ``False`` .
|
|
185
187
|
|
|
@@ -197,7 +199,7 @@ class LPPool2d(Cell):
|
|
|
197
199
|
W_{out} = \left\lfloor\frac{W_{in} - \text{kernel_size}[1]}{\text{stride}[1]} + 1\right\rfloor
|
|
198
200
|
|
|
199
201
|
Raises:
|
|
200
|
-
TypeError: If `x` is not
|
|
202
|
+
TypeError: If `x` is not a Tensor.
|
|
201
203
|
TypeError: If `kernel_size` or `stride` is neither int nor tuple.
|
|
202
204
|
TypeError: If `ceil_mode` is not a bool.
|
|
203
205
|
TypeError: If `norm_type` is neither float nor int.
|
|
@@ -324,7 +326,8 @@ class MaxPool3d(_PoolNd):
|
|
|
324
326
|
three integers. Default: ``1`` .
|
|
325
327
|
return_indices (bool): If ``True`` , output is a Tuple of 2 Tensors, representing the maxpool result and where
|
|
326
328
|
the max values are generated. Otherwise, only the maxpool result is returned. Default: ``False`` .
|
|
327
|
-
ceil_mode (bool):
|
|
329
|
+
ceil_mode (bool): If ``True``, use ceil to calculate output shape.
|
|
330
|
+
If ``False``, use ceil to calculate output shape. Default: ``False`` .
|
|
328
331
|
|
|
329
332
|
Inputs:
|
|
330
333
|
- **x** (Tensor) - Tensor of shape :math:`(N_{in}, C_{in}, D_{in}, H_{in}, W_{in})` or
|
|
@@ -620,7 +623,7 @@ class MaxPool1d(_PoolNd):
|
|
|
620
623
|
at the begin and end is determined by the `padding` parameter.
|
|
621
624
|
If this mode is set, `padding` must be greater than or equal to 0.
|
|
622
625
|
|
|
623
|
-
padding (Union(int, tuple[int], list[int])): Padding value for the pooling. Default value is 0
|
|
626
|
+
padding (Union(int, tuple[int], list[int])): Padding value for the pooling. Default value is ``0``.
|
|
624
627
|
padding can only be an integer or a tuple/list containing a single integer, in which case padding times or
|
|
625
628
|
padding[0] times are padded on both sides of the input.
|
|
626
629
|
dilation (Union(int, tuple[int])): The spacing between the elements of the kernel in convolution,
|
|
@@ -1012,8 +1015,12 @@ class AvgPool2d(_PoolNd):
|
|
|
1012
1015
|
data_format="NCHW"):
|
|
1013
1016
|
"""Initialize AvgPool2d."""
|
|
1014
1017
|
super(AvgPool2d, self).__init__(kernel_size, stride, pad_mode, data_format)
|
|
1018
|
+
self.ascend_910bc_target = (MSContext.get_instance().get_ascend_soc_version() in ['ascend910b', 'ascend910c'])
|
|
1015
1019
|
if pad_mode.upper() == 'PAD' or padding != 0 or ceil_mode or not count_include_pad \
|
|
1016
1020
|
or divisor_override is not None:
|
|
1021
|
+
if self.ascend_910bc_target:
|
|
1022
|
+
raise ValueError(f"For '{self.cls_name}, the pad_mod 'PAD' is not support in 910B now, "
|
|
1023
|
+
f"it will be supported in the future.")
|
|
1017
1024
|
if self.format == "NHWC":
|
|
1018
1025
|
raise ValueError(f"For '{self.cls_name}, the 'NHWC' format are not support when 'pad_mode' is 'pad' or "
|
|
1019
1026
|
f"'padding' is not 0 or 'ceil_mode' is not False or 'count_include_pad' is not True"
|
|
@@ -1660,6 +1667,8 @@ class AdaptiveMaxPool3d(Cell):
|
|
|
1660
1667
|
def __init__(self, output_size, return_indices=False):
|
|
1661
1668
|
"""Initialize AdaptiveMaxPool3d."""
|
|
1662
1669
|
super(AdaptiveMaxPool3d, self).__init__()
|
|
1670
|
+
if isinstance(output_size, int):
|
|
1671
|
+
output_size = (output_size, output_size, output_size)
|
|
1663
1672
|
self.output_size = Tensor(output_size, dtype=mstype.int32)
|
|
1664
1673
|
self.return_indices = return_indices
|
|
1665
1674
|
self.adaptive_max_pool3d = AdaptiveMaxPool3D()
|
|
@@ -1975,7 +1984,7 @@ class MaxUnpool2d(Cell):
|
|
|
1975
1984
|
stride (Union[int, tuple[int]]): The distance of kernel moving, an int number that represents
|
|
1976
1985
|
the height and width of movement are both stride, or a tuple of two int numbers that
|
|
1977
1986
|
represent height and width of movement respectively.
|
|
1978
|
-
If stride is None
|
|
1987
|
+
If stride is ``None``, then stride equal to kernel_size. Default: ``None`` .
|
|
1979
1988
|
padding (Union[int, tuple[int]]): The pad value to be filled. Default: ``0`` . If `padding` is an integer,
|
|
1980
1989
|
the paddings of height and width are the same, equal to padding. If `padding` is a tuple of two
|
|
1981
1990
|
integers, the padding of height and width equal to padding[0] and padding[1] correspondingly.
|
|
@@ -2068,7 +2077,7 @@ class MaxUnpool3d(Cell):
|
|
|
2068
2077
|
stride (Union[int, tuple[int]]): The distance of kernel moving, an int number that represents
|
|
2069
2078
|
the depth, height and width of movement are both stride, or a tuple of three int numbers that
|
|
2070
2079
|
represent depth, height and width of movement respectively.
|
|
2071
|
-
If stride is None
|
|
2080
|
+
If stride is ``None``, then stride equal to kernel_size. Default: ``None`` .
|
|
2072
2081
|
padding (Union[int, tuple[int]]): The pad value to be filled. Default: ``0`` . If `padding` is an integer,
|
|
2073
2082
|
the paddings of depth, height and width are the same, equal to padding. If `padding` is a tuple of three
|
|
2074
2083
|
integers, the padding of depth, height and width equal to padding[0], padding[1] and padding[2]
|
mindspore/nn/layer/rnn_cells.py
CHANGED
|
@@ -25,7 +25,7 @@ from mindspore import log as logger
|
|
|
25
25
|
from mindspore.common.tensor import Tensor
|
|
26
26
|
from mindspore.common.parameter import Parameter
|
|
27
27
|
from mindspore.common.initializer import initializer, Uniform
|
|
28
|
-
from mindspore.ops.primitive import constexpr
|
|
28
|
+
from mindspore.ops.primitive import constexpr
|
|
29
29
|
from mindspore.nn.cell import Cell
|
|
30
30
|
from mindspore import _checkparam as validator
|
|
31
31
|
|
|
@@ -61,13 +61,6 @@ def _check_tuple_length(param_name, input_data, length, cls_name):
|
|
|
61
61
|
f"but got '{len(input_data)}'")
|
|
62
62
|
|
|
63
63
|
|
|
64
|
-
@_primexpr
|
|
65
|
-
def _check_batch_size_equal(batch_size_x, batch_size_hx, cls_name):
|
|
66
|
-
if batch_size_x != batch_size_hx:
|
|
67
|
-
raise ValueError(f"For '{cls_name}' batch size of x and hx must be equal, but got {batch_size_x} of x "
|
|
68
|
-
f"and {batch_size_hx} of hx.")
|
|
69
|
-
|
|
70
|
-
|
|
71
64
|
def _check_lstmcell_init(func):
|
|
72
65
|
"""Internal function, used to check init args."""
|
|
73
66
|
@wraps(func)
|
|
@@ -185,7 +178,7 @@ class RNNCell(RNNCellBase):
|
|
|
185
178
|
Args:
|
|
186
179
|
input_size (int): Number of features of input.
|
|
187
180
|
hidden_size (int): Number of features of hidden layer.
|
|
188
|
-
has_bias (bool): Whether the cell has bias `
|
|
181
|
+
has_bias (bool): Whether the cell has bias :math:`b_{ih}` and :math:`b_{hh}`. Default: ``True`` .
|
|
189
182
|
nonlinearity (str): The non-linearity to use. Can be either ``"tanh"`` or ``"relu"`` .
|
|
190
183
|
Default: ``"tanh"`` .
|
|
191
184
|
dtype (:class:`mindspore.dtype`): Dtype of Parameters. Default: ``mstype.float32`` .
|
|
@@ -232,7 +225,6 @@ class RNNCell(RNNCellBase):
|
|
|
232
225
|
_check_is_tensor('hx', hx, self.cls_name)
|
|
233
226
|
_check_input_dtype(x.dtype, "x", [mstype.float32, mstype.float16], self.cls_name)
|
|
234
227
|
_check_input_dtype(hx.dtype, "hx", [mstype.float32, mstype.float16], self.cls_name)
|
|
235
|
-
_check_batch_size_equal(x.shape[0], hx.shape[0], self.cls_name)
|
|
236
228
|
|
|
237
229
|
if self.nonlinearity == "tanh":
|
|
238
230
|
ret = _rnn_tanh_cell(x, hx, self.weight_ih, self.weight_hh, self.bias_ih, self.bias_hh)
|
|
@@ -271,7 +263,7 @@ class LSTMCell(RNNCellBase):
|
|
|
271
263
|
Args:
|
|
272
264
|
input_size (int): Number of features of input.
|
|
273
265
|
hidden_size (int): Number of features of hidden layer.
|
|
274
|
-
has_bias (bool): Whether the cell has bias `
|
|
266
|
+
has_bias (bool): Whether the cell has bias `b_{ih}` and `b_{hh}`. Default: ``True`` .
|
|
275
267
|
dtype (:class:`mindspore.dtype`): Dtype of Parameters. Default: ``mstype.float32`` .
|
|
276
268
|
|
|
277
269
|
Inputs:
|
|
@@ -318,8 +310,6 @@ class LSTMCell(RNNCellBase):
|
|
|
318
310
|
_check_input_dtype(x.dtype, "x", [mstype.float32, mstype.float16], self.cls_name)
|
|
319
311
|
_check_input_dtype(hx[0].dtype, "hx[0]", [mstype.float32, mstype.float16], self.cls_name)
|
|
320
312
|
_check_input_dtype(hx[1].dtype, "hx[1]", [mstype.float32, mstype.float16], self.cls_name)
|
|
321
|
-
_check_batch_size_equal(x.shape[0], hx[0].shape[0], self.cls_name)
|
|
322
|
-
_check_batch_size_equal(x.shape[0], hx[1].shape[0], self.cls_name)
|
|
323
313
|
return _lstm_cell(x, hx, self.weight_ih, self.weight_hh, self.bias_ih, self.bias_hh)
|
|
324
314
|
|
|
325
315
|
def _check_construct_args(self, *inputs, **kwargs):
|
|
@@ -346,7 +336,8 @@ class GRUCell(RNNCellBase):
|
|
|
346
336
|
\end{array}
|
|
347
337
|
|
|
348
338
|
Here :math:`\sigma` is the sigmoid function, and :math:`*` is the Hadamard product. :math:`W, b`
|
|
349
|
-
are learnable weights between the output and the input in the formula.
|
|
339
|
+
are learnable weights between the output and the input in the formula. :math:`h` is hidden state.
|
|
340
|
+
:math:`r` is reset gate. :math:`z` is update gate. :math:`n` is n-th layer. For instance,
|
|
350
341
|
:math:`W_{ir}, b_{ir}` are the weight and bias used to transform from input :math:`x` to :math:`r`.
|
|
351
342
|
Details can be found in paper
|
|
352
343
|
`Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation
|
|
@@ -355,7 +346,7 @@ class GRUCell(RNNCellBase):
|
|
|
355
346
|
Args:
|
|
356
347
|
input_size (int): Number of features of input.
|
|
357
348
|
hidden_size (int): Number of features of hidden layer.
|
|
358
|
-
has_bias (bool): Whether the cell has bias `
|
|
349
|
+
has_bias (bool): Whether the cell has bias :math:`b_{in}` and :math:`b_{hn}`. Default: ``True`` .
|
|
359
350
|
dtype (:class:`mindspore.dtype`): Dtype of Parameters. Default: ``mstype.float32`` .
|
|
360
351
|
|
|
361
352
|
Inputs:
|
|
@@ -394,5 +385,4 @@ class GRUCell(RNNCellBase):
|
|
|
394
385
|
_check_is_tensor('hx', hx, self.cls_name)
|
|
395
386
|
_check_input_dtype(x.dtype, "x", [mstype.float32, mstype.float16], self.cls_name)
|
|
396
387
|
_check_input_dtype(hx.dtype, "hx", [mstype.float32, mstype.float16], self.cls_name)
|
|
397
|
-
_check_batch_size_equal(x.shape[0], hx.shape[0], self.cls_name)
|
|
398
388
|
return _gru_cell(x, hx, self.weight_ih, self.weight_hh, self.bias_ih, self.bias_hh)
|