mindspore 2.2.11__cp39-cp39-win_amd64.whl → 2.3.0__cp39-cp39-win_amd64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of mindspore might be problematic. Click here for more details.
- mindspore/.commit_id +1 -1
- mindspore/__init__.py +7 -5
- mindspore/_c_dataengine.cp39-win_amd64.pyd +0 -0
- mindspore/_c_expression.cp39-win_amd64.pyd +0 -0
- mindspore/_c_mindrecord.cp39-win_amd64.pyd +0 -0
- mindspore/_checkparam.py +76 -18
- mindspore/_extends/builtin_operations.py +2 -1
- mindspore/_extends/graph_kernel/model/graph_parallel.py +16 -6
- mindspore/_extends/parallel_compile/akg_compiler/akg_process.py +3 -16
- mindspore/_extends/parallel_compile/akg_compiler/build_tbe_kernel.py +16 -4
- mindspore/_extends/parallel_compile/akg_compiler/compiler.py +1 -0
- mindspore/_extends/parallel_compile/akg_compiler/gen_custom_op_files.py +96 -0
- mindspore/_extends/parallel_compile/akg_compiler/tbe_topi.py +2 -1
- mindspore/_extends/parallel_compile/akg_compiler/util.py +5 -2
- mindspore/_extends/parse/__init__.py +18 -14
- mindspore/_extends/parse/compile_config.py +258 -0
- mindspore/_extends/parse/namespace.py +2 -2
- mindspore/_extends/parse/parser.py +174 -62
- mindspore/_extends/parse/resources.py +45 -14
- mindspore/_extends/parse/standard_method.py +142 -240
- mindspore/{ops/_op_impl/tbe/atomic_addr_clean.py → _extends/pijit/__init__.py} +6 -16
- mindspore/_extends/pijit/pijit_func_white_list.py +343 -0
- mindspore/_extends/remote/kernel_build_server.py +2 -0
- mindspore/_profiler.py +30 -0
- mindspore/amp.py +51 -24
- mindspore/avcodec-59.dll +0 -0
- mindspore/avdevice-59.dll +0 -0
- mindspore/avfilter-8.dll +0 -0
- mindspore/avformat-59.dll +0 -0
- mindspore/avutil-57.dll +0 -0
- mindspore/boost/adasum.py +1 -1
- mindspore/boost/base.py +1 -1
- mindspore/boost/boost_cell_wrapper.py +2 -2
- mindspore/boost/grad_freeze.py +2 -2
- mindspore/boost/group_loss_scale_manager.py +1 -1
- mindspore/boost/less_batch_normalization.py +9 -6
- mindspore/common/__init__.py +15 -4
- mindspore/common/_jit_fallback_utils.py +2 -3
- mindspore/common/_register_for_adapter.py +7 -0
- mindspore/common/_register_for_recompute.py +48 -0
- mindspore/common/_register_for_tensor.py +8 -9
- mindspore/common/_stub_tensor.py +7 -1
- mindspore/common/_utils.py +5 -17
- mindspore/common/api.py +411 -106
- mindspore/common/auto_dynamic_shape.py +27 -14
- mindspore/common/dtype.py +17 -10
- mindspore/common/dump.py +6 -8
- mindspore/common/file_system.py +48 -0
- mindspore/common/generator.py +260 -0
- mindspore/common/hook_handle.py +51 -4
- mindspore/common/initializer.py +1 -1
- mindspore/common/jit_config.py +34 -14
- mindspore/common/lazy_inline.py +72 -19
- mindspore/common/mindir_util.py +12 -2
- mindspore/common/mutable.py +79 -14
- mindspore/common/no_inline.py +54 -0
- mindspore/common/np_dtype.py +25 -0
- mindspore/common/parameter.py +30 -11
- mindspore/common/recompute.py +262 -0
- mindspore/common/seed.py +9 -9
- mindspore/common/sparse_tensor.py +272 -24
- mindspore/common/symbol.py +122 -0
- mindspore/common/tensor.py +468 -496
- mindspore/communication/__init__.py +6 -11
- mindspore/communication/_comm_helper.py +5 -0
- mindspore/communication/comm_func.py +1140 -0
- mindspore/communication/management.py +118 -102
- mindspore/config/op_info.config +22 -54
- mindspore/context.py +378 -65
- mindspore/dataset/__init__.py +5 -5
- mindspore/dataset/audio/__init__.py +6 -6
- mindspore/dataset/audio/transforms.py +711 -158
- mindspore/dataset/callback/ds_callback.py +2 -2
- mindspore/dataset/engine/cache_client.py +2 -2
- mindspore/dataset/engine/datasets.py +163 -83
- mindspore/dataset/engine/datasets_audio.py +14 -14
- mindspore/dataset/engine/datasets_standard_format.py +33 -3
- mindspore/dataset/engine/datasets_text.py +38 -38
- mindspore/dataset/engine/datasets_user_defined.py +78 -59
- mindspore/dataset/engine/datasets_vision.py +77 -73
- mindspore/dataset/engine/offload.py +5 -7
- mindspore/dataset/engine/queue.py +56 -38
- mindspore/dataset/engine/validators.py +11 -5
- mindspore/dataset/text/__init__.py +3 -3
- mindspore/dataset/text/transforms.py +408 -121
- mindspore/dataset/text/utils.py +9 -9
- mindspore/dataset/transforms/__init__.py +1 -1
- mindspore/dataset/transforms/transforms.py +261 -76
- mindspore/dataset/utils/browse_dataset.py +9 -9
- mindspore/dataset/vision/__init__.py +8 -8
- mindspore/dataset/vision/c_transforms.py +10 -10
- mindspore/dataset/vision/py_transforms_util.py +3 -3
- mindspore/dataset/vision/transforms.py +2844 -549
- mindspore/dataset/vision/utils.py +161 -10
- mindspore/dataset/vision/validators.py +14 -2
- mindspore/dnnl.dll +0 -0
- mindspore/experimental/optim/__init__.py +12 -2
- mindspore/experimental/optim/adadelta.py +161 -0
- mindspore/experimental/optim/adagrad.py +168 -0
- mindspore/experimental/optim/adam.py +35 -34
- mindspore/experimental/optim/adamax.py +170 -0
- mindspore/experimental/optim/adamw.py +40 -16
- mindspore/experimental/optim/asgd.py +153 -0
- mindspore/experimental/optim/lr_scheduler.py +71 -127
- mindspore/experimental/optim/nadam.py +157 -0
- mindspore/experimental/optim/optimizer.py +15 -8
- mindspore/experimental/optim/radam.py +194 -0
- mindspore/experimental/optim/rmsprop.py +154 -0
- mindspore/experimental/optim/rprop.py +164 -0
- mindspore/experimental/optim/sgd.py +28 -19
- mindspore/hal/__init__.py +40 -0
- mindspore/hal/_ascend.py +57 -0
- mindspore/hal/_base.py +57 -0
- mindspore/hal/_cpu.py +56 -0
- mindspore/hal/_gpu.py +57 -0
- mindspore/hal/device.py +356 -0
- mindspore/hal/event.py +179 -0
- mindspore/hal/memory.py +326 -0
- mindspore/hal/stream.py +339 -0
- mindspore/include/api/data_type.h +2 -2
- mindspore/include/api/dual_abi_helper.h +16 -3
- mindspore/include/api/model.h +4 -3
- mindspore/include/api/status.h +14 -0
- mindspore/include/c_api/model_c.h +173 -0
- mindspore/include/c_api/ms/base/types.h +1 -0
- mindspore/include/c_api/types_c.h +19 -0
- mindspore/include/dataset/execute.h +1 -3
- mindspore/include/dataset/vision.h +54 -2
- mindspore/jpeg62.dll +0 -0
- mindspore/log.py +2 -2
- mindspore/mindrecord/__init__.py +5 -1
- mindspore/mindrecord/config.py +809 -0
- mindspore/mindrecord/filereader.py +25 -0
- mindspore/mindrecord/filewriter.py +76 -58
- mindspore/mindrecord/mindpage.py +40 -6
- mindspore/mindrecord/shardutils.py +3 -2
- mindspore/mindrecord/shardwriter.py +7 -0
- mindspore/mindrecord/tools/cifar100_to_mr.py +53 -66
- mindspore/mindrecord/tools/cifar10_to_mr.py +48 -63
- mindspore/mindrecord/tools/csv_to_mr.py +7 -17
- mindspore/mindrecord/tools/imagenet_to_mr.py +3 -8
- mindspore/mindrecord/tools/mnist_to_mr.py +11 -21
- mindspore/mindrecord/tools/tfrecord_to_mr.py +2 -10
- mindspore/mindspore_backend.dll +0 -0
- mindspore/mindspore_common.dll +0 -0
- mindspore/mindspore_core.dll +0 -0
- mindspore/mindspore_glog.dll +0 -0
- mindspore/mindspore_np_dtype.dll +0 -0
- mindspore/mindspore_shared_lib.dll +0 -0
- mindspore/mint/__init__.py +1137 -0
- mindspore/{rewrite/ast_transformers → mint/linalg}/__init__.py +9 -4
- mindspore/mint/nn/__init__.py +512 -0
- mindspore/mint/nn/functional.py +573 -0
- mindspore/mint/optim/__init__.py +24 -0
- mindspore/mint/optim/adamw.py +185 -0
- mindspore/multiprocessing/__init__.py +72 -0
- mindspore/nn/__init__.py +1 -0
- mindspore/nn/cell.py +213 -257
- mindspore/nn/dynamic_lr.py +2 -2
- mindspore/nn/extend/__init__.py +29 -0
- mindspore/nn/extend/basic.py +140 -0
- mindspore/nn/extend/embedding.py +143 -0
- mindspore/{rewrite/ast_creator_register.py → nn/extend/layer/__init__.py} +9 -19
- mindspore/nn/extend/layer/normalization.py +109 -0
- mindspore/nn/extend/pooling.py +117 -0
- mindspore/nn/layer/activation.py +84 -94
- mindspore/nn/layer/basic.py +177 -82
- mindspore/nn/layer/channel_shuffle.py +3 -16
- mindspore/nn/layer/container.py +3 -3
- mindspore/nn/layer/conv.py +75 -66
- mindspore/nn/layer/embedding.py +103 -45
- mindspore/nn/layer/embedding_service.py +531 -0
- mindspore/nn/layer/embedding_service_layer.py +393 -0
- mindspore/nn/layer/image.py +4 -7
- mindspore/nn/layer/math.py +1 -1
- mindspore/nn/layer/normalization.py +52 -66
- mindspore/nn/layer/padding.py +30 -39
- mindspore/nn/layer/pooling.py +18 -9
- mindspore/nn/layer/rnn_cells.py +6 -16
- mindspore/nn/layer/rnns.py +6 -5
- mindspore/nn/layer/thor_layer.py +1 -2
- mindspore/nn/layer/timedistributed.py +1 -1
- mindspore/nn/layer/transformer.py +52 -50
- mindspore/nn/learning_rate_schedule.py +6 -5
- mindspore/nn/loss/loss.py +63 -84
- mindspore/nn/optim/ada_grad.py +6 -4
- mindspore/nn/optim/adadelta.py +3 -1
- mindspore/nn/optim/adafactor.py +1 -1
- mindspore/nn/optim/adam.py +102 -181
- mindspore/nn/optim/adamax.py +4 -2
- mindspore/nn/optim/adasum.py +3 -3
- mindspore/nn/optim/asgd.py +4 -2
- mindspore/nn/optim/ftrl.py +31 -61
- mindspore/nn/optim/lamb.py +5 -3
- mindspore/nn/optim/lars.py +2 -2
- mindspore/nn/optim/lazyadam.py +6 -4
- mindspore/nn/optim/momentum.py +13 -25
- mindspore/nn/optim/optimizer.py +6 -3
- mindspore/nn/optim/proximal_ada_grad.py +4 -2
- mindspore/nn/optim/rmsprop.py +9 -3
- mindspore/nn/optim/rprop.py +4 -2
- mindspore/nn/optim/sgd.py +7 -4
- mindspore/nn/optim/thor.py +2 -2
- mindspore/nn/probability/distribution/_utils/custom_ops.py +2 -2
- mindspore/nn/probability/distribution/beta.py +2 -2
- mindspore/nn/probability/distribution/categorical.py +4 -6
- mindspore/nn/probability/distribution/cauchy.py +2 -2
- mindspore/nn/probability/distribution/exponential.py +2 -2
- mindspore/nn/probability/distribution/geometric.py +1 -1
- mindspore/nn/probability/distribution/gumbel.py +2 -2
- mindspore/nn/probability/distribution/logistic.py +1 -1
- mindspore/nn/probability/distribution/poisson.py +2 -2
- mindspore/nn/probability/distribution/uniform.py +2 -2
- mindspore/nn/reinforcement/_tensors_queue.py +13 -1
- mindspore/nn/wrap/__init__.py +2 -1
- mindspore/nn/wrap/cell_wrapper.py +58 -13
- mindspore/nn/wrap/grad_reducer.py +148 -8
- mindspore/nn/wrap/loss_scale.py +32 -9
- mindspore/numpy/__init__.py +2 -0
- mindspore/numpy/array_creations.py +2 -0
- mindspore/numpy/array_ops.py +6 -6
- mindspore/numpy/dtypes.py +3 -3
- mindspore/numpy/fft.py +431 -0
- mindspore/numpy/math_ops.py +61 -67
- mindspore/numpy/utils.py +3 -0
- mindspore/opencv_core452.dll +0 -0
- mindspore/opencv_imgcodecs452.dll +0 -0
- mindspore/opencv_imgproc452.dll +0 -0
- mindspore/ops/__init__.py +8 -4
- mindspore/ops/_grad_experimental/grad_array_ops.py +4 -160
- mindspore/ops/_grad_experimental/grad_comm_ops.py +93 -36
- mindspore/ops/_grad_experimental/grad_inner_ops.py +8 -0
- mindspore/ops/_grad_experimental/grad_math_ops.py +92 -287
- mindspore/ops/_grad_experimental/grad_nn_ops.py +0 -53
- mindspore/ops/_grad_experimental/grad_quant_ops.py +3 -3
- mindspore/ops/_grad_experimental/grad_sparse.py +1 -1
- mindspore/ops/_grad_experimental/grad_sparse_ops.py +3 -3
- mindspore/ops/_op_impl/__init__.py +0 -1
- mindspore/ops/_op_impl/aicpu/__init__.py +1 -0
- mindspore/ops/_op_impl/aicpu/gamma.py +2 -0
- mindspore/ops/_op_impl/{cpu/concat.py → aicpu/generate_eod_mask.py} +16 -17
- mindspore/ops/_op_impl/aicpu/log_uniform_candidate_sampler.py +1 -3
- mindspore/ops/_op_impl/aicpu/poisson.py +2 -0
- mindspore/ops/_op_impl/cpu/__init__.py +1 -3
- mindspore/ops/_op_impl/cpu/adam.py +2 -2
- mindspore/ops/_op_impl/cpu/adam_weight_decay.py +3 -2
- mindspore/ops/_op_impl/cpu/maximum_grad.py +16 -14
- mindspore/ops/_op_impl/cpu/minimum_grad.py +8 -0
- mindspore/ops/_vmap/vmap_array_ops.py +164 -101
- mindspore/ops/_vmap/vmap_base.py +8 -1
- mindspore/ops/_vmap/vmap_grad_math_ops.py +95 -9
- mindspore/ops/_vmap/vmap_grad_nn_ops.py +143 -58
- mindspore/ops/_vmap/vmap_image_ops.py +70 -13
- mindspore/ops/_vmap/vmap_math_ops.py +130 -58
- mindspore/ops/_vmap/vmap_nn_ops.py +249 -115
- mindspore/ops/_vmap/vmap_other_ops.py +1 -1
- mindspore/ops/auto_generate/__init__.py +31 -0
- mindspore/ops/auto_generate/cpp_create_prim_instance_helper.py +231 -0
- mindspore/ops/auto_generate/gen_arg_dtype_cast.py +250 -0
- mindspore/ops/auto_generate/gen_arg_handler.py +197 -0
- mindspore/ops/auto_generate/gen_extend_func.py +980 -0
- mindspore/ops/auto_generate/gen_ops_def.py +6443 -0
- mindspore/ops/auto_generate/gen_ops_prim.py +13167 -0
- mindspore/ops/auto_generate/pyboost_inner_prim.py +429 -0
- mindspore/ops/composite/__init__.py +5 -2
- mindspore/ops/composite/base.py +121 -23
- mindspore/ops/composite/math_ops.py +10 -49
- mindspore/ops/composite/multitype_ops/_compile_utils.py +191 -618
- mindspore/ops/composite/multitype_ops/_constexpr_utils.py +25 -134
- mindspore/ops/composite/multitype_ops/add_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/bitwise_and_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/bitwise_or_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/bitwise_xor_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/div_impl.py +8 -0
- mindspore/ops/composite/multitype_ops/equal_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/floordiv_impl.py +8 -0
- mindspore/ops/composite/multitype_ops/getitem_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/greater_equal_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/greater_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/in_impl.py +8 -2
- mindspore/ops/composite/multitype_ops/left_shift_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/less_equal_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/less_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/logic_not_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/logical_and_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/logical_or_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/mod_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/mul_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/negative_impl.py +9 -3
- mindspore/ops/composite/multitype_ops/not_equal_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/not_in_impl.py +6 -1
- mindspore/ops/composite/multitype_ops/ones_like_impl.py +2 -2
- mindspore/ops/composite/multitype_ops/pow_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/right_shift_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/setitem_impl.py +32 -21
- mindspore/ops/composite/multitype_ops/sub_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/zeros_like_impl.py +6 -3
- mindspore/ops/deprecated.py +14 -3
- mindspore/ops/extend/__init__.py +53 -0
- mindspore/ops/extend/array_func.py +218 -0
- mindspore/ops/extend/math_func.py +76 -0
- mindspore/ops/extend/nn_func.py +308 -0
- mindspore/ops/function/__init__.py +31 -11
- mindspore/ops/function/array_func.py +848 -1736
- mindspore/ops/function/clip_func.py +19 -31
- mindspore/ops/function/debug_func.py +2 -5
- mindspore/ops/function/fft_func.py +31 -0
- mindspore/ops/function/grad/grad_func.py +27 -20
- mindspore/ops/function/image_func.py +27 -21
- mindspore/ops/function/linalg_func.py +30 -53
- mindspore/ops/function/math_func.py +916 -2791
- mindspore/ops/function/nn_func.py +1445 -889
- mindspore/ops/function/other_func.py +6 -7
- mindspore/ops/function/parameter_func.py +6 -92
- mindspore/ops/function/random_func.py +254 -108
- mindspore/ops/function/reshard_func.py +102 -0
- mindspore/ops/function/sparse_func.py +4 -4
- mindspore/ops/function/sparse_unary_func.py +11 -18
- mindspore/ops/function/spectral_func.py +1 -1
- mindspore/ops/function/vmap_func.py +15 -14
- mindspore/ops/functional.py +342 -343
- mindspore/ops/op_info_register.py +16 -43
- mindspore/ops/operations/__init__.py +32 -23
- mindspore/ops/operations/_embedding_cache_ops.py +1 -1
- mindspore/ops/operations/_grad_ops.py +21 -853
- mindspore/ops/operations/_infer_ops.py +19 -0
- mindspore/ops/operations/_inner_ops.py +155 -511
- mindspore/ops/operations/_quant_ops.py +4 -4
- mindspore/ops/operations/_rl_inner_ops.py +3 -3
- mindspore/ops/operations/_scalar_ops.py +5 -480
- mindspore/ops/operations/_sequence_ops.py +6 -36
- mindspore/ops/operations/_tensor_array.py +8 -8
- mindspore/ops/operations/array_ops.py +112 -2698
- mindspore/ops/operations/comm_ops.py +801 -118
- mindspore/ops/operations/custom_ops.py +62 -121
- mindspore/ops/operations/debug_ops.py +105 -36
- mindspore/ops/operations/image_ops.py +3 -219
- mindspore/ops/operations/inner_ops.py +54 -40
- mindspore/ops/operations/linalg_ops.py +1 -49
- mindspore/ops/operations/manually_defined/__init__.py +24 -0
- mindspore/ops/operations/manually_defined/_inner.py +61 -0
- mindspore/ops/operations/manually_defined/ops_def.py +2016 -0
- mindspore/ops/operations/math_ops.py +621 -4654
- mindspore/ops/operations/nn_ops.py +316 -2226
- mindspore/ops/operations/other_ops.py +53 -45
- mindspore/ops/operations/random_ops.py +4 -51
- mindspore/ops/operations/reshard_ops.py +53 -0
- mindspore/ops/operations/sparse_ops.py +8 -8
- mindspore/ops/primitive.py +204 -103
- mindspore/ops/silent_check.py +162 -0
- mindspore/ops_generate/__init__.py +27 -0
- mindspore/ops_generate/arg_dtype_cast.py +250 -0
- mindspore/ops_generate/arg_handler.py +197 -0
- mindspore/ops_generate/gen_aclnn_implement.py +263 -0
- mindspore/ops_generate/gen_ops.py +1084 -0
- mindspore/ops_generate/gen_ops_inner_prim.py +131 -0
- mindspore/ops_generate/gen_pyboost_func.py +968 -0
- mindspore/ops_generate/gen_utils.py +209 -0
- mindspore/ops_generate/op_proto.py +138 -0
- mindspore/ops_generate/pyboost_utils.py +354 -0
- mindspore/ops_generate/template.py +239 -0
- mindspore/parallel/__init__.py +7 -4
- mindspore/parallel/_auto_parallel_context.py +155 -6
- mindspore/parallel/_cell_wrapper.py +16 -9
- mindspore/parallel/_cost_model_context.py +1 -1
- mindspore/parallel/_dp_allreduce_fusion.py +159 -159
- mindspore/parallel/_parallel_serialization.py +62 -14
- mindspore/parallel/_ps_context.py +1 -1
- mindspore/parallel/_recovery_context.py +1 -1
- mindspore/parallel/_tensor.py +18 -9
- mindspore/parallel/_transformer/__init__.py +1 -1
- mindspore/parallel/_transformer/layers.py +1 -1
- mindspore/parallel/_transformer/loss.py +1 -1
- mindspore/parallel/_transformer/moe.py +1 -1
- mindspore/parallel/_transformer/op_parallel_config.py +1 -1
- mindspore/parallel/_transformer/transformer.py +10 -10
- mindspore/parallel/_utils.py +161 -6
- mindspore/parallel/algo_parameter_config.py +6 -8
- mindspore/parallel/checkpoint_transform.py +369 -64
- mindspore/parallel/cluster/__init__.py +15 -0
- mindspore/parallel/cluster/process_entity/__init__.py +18 -0
- mindspore/parallel/cluster/process_entity/_api.py +344 -0
- mindspore/parallel/cluster/process_entity/_utils.py +126 -0
- mindspore/parallel/cluster/run.py +136 -0
- mindspore/parallel/mpi/__init__.py +1 -1
- mindspore/parallel/mpi/_mpi_config.py +1 -1
- mindspore/parallel/parameter_broadcast.py +152 -0
- mindspore/parallel/shard.py +128 -17
- mindspore/profiler/__init__.py +3 -2
- mindspore/profiler/common/process_pool.py +41 -0
- mindspore/profiler/common/singleton.py +28 -0
- mindspore/profiler/common/util.py +125 -0
- mindspore/profiler/envprofiling.py +2 -2
- mindspore/{_extends/parallel_compile/tbe_compiler → profiler/parser/ascend_analysis}/__init__.py +1 -1
- mindspore/profiler/parser/ascend_analysis/constant.py +53 -0
- mindspore/profiler/parser/ascend_analysis/file_manager.py +159 -0
- mindspore/profiler/parser/ascend_analysis/function_event.py +161 -0
- mindspore/profiler/parser/ascend_analysis/fwk_cann_parser.py +131 -0
- mindspore/profiler/parser/ascend_analysis/fwk_file_parser.py +85 -0
- mindspore/profiler/parser/ascend_analysis/msprof_timeline_parser.py +57 -0
- mindspore/profiler/parser/ascend_analysis/profiler_info_parser.py +116 -0
- mindspore/profiler/parser/ascend_analysis/tlv_decoder.py +86 -0
- mindspore/profiler/parser/ascend_analysis/trace_event_manager.py +68 -0
- mindspore/profiler/parser/ascend_cluster_generator.py +116 -0
- mindspore/profiler/parser/ascend_communicate_generator.py +314 -0
- mindspore/profiler/parser/ascend_flops_generator.py +27 -5
- mindspore/profiler/parser/ascend_fpbp_generator.py +8 -2
- mindspore/profiler/parser/ascend_hccl_generator.py +31 -280
- mindspore/profiler/parser/ascend_integrate_generator.py +42 -0
- mindspore/profiler/parser/ascend_memory_generator.py +185 -0
- mindspore/profiler/parser/ascend_msprof_exporter.py +151 -126
- mindspore/profiler/parser/ascend_msprof_generator.py +75 -274
- mindspore/profiler/parser/ascend_op_generator.py +94 -36
- mindspore/profiler/parser/ascend_timeline_generator.py +297 -131
- mindspore/profiler/parser/base_timeline_generator.py +17 -3
- mindspore/profiler/parser/cpu_gpu_timeline_generator.py +2 -1
- mindspore/profiler/parser/framework_parser.py +11 -4
- mindspore/profiler/parser/integrator.py +3 -1
- mindspore/profiler/parser/memory_usage_parser.py +8 -2
- mindspore/profiler/parser/minddata_analyzer.py +8 -2
- mindspore/profiler/parser/minddata_parser.py +73 -4
- mindspore/profiler/parser/msadvisor_analyzer.py +5 -3
- mindspore/profiler/parser/msadvisor_parser.py +10 -4
- mindspore/profiler/parser/profiler_info.py +16 -1
- mindspore/profiler/profiling.py +522 -195
- mindspore/rewrite/__init__.py +2 -13
- mindspore/rewrite/api/node.py +123 -37
- mindspore/rewrite/api/pattern_engine.py +2 -3
- mindspore/rewrite/api/scoped_value.py +16 -15
- mindspore/rewrite/api/symbol_tree.py +46 -30
- mindspore/rewrite/ast_helpers/__init__.py +3 -6
- mindspore/rewrite/ast_helpers/ast_converter.py +143 -0
- mindspore/rewrite/ast_helpers/ast_finder.py +48 -0
- mindspore/rewrite/ast_helpers/ast_flattener.py +268 -0
- mindspore/rewrite/ast_helpers/ast_modifier.py +160 -92
- mindspore/rewrite/common/__init__.py +1 -2
- mindspore/rewrite/common/config.py +24 -0
- mindspore/rewrite/common/{rewrite_elog.py → error_log.py} +39 -39
- mindspore/rewrite/{namer.py → common/namer.py} +63 -18
- mindspore/rewrite/common/namespace.py +118 -0
- mindspore/rewrite/node/__init__.py +5 -5
- mindspore/rewrite/node/call_function.py +23 -7
- mindspore/rewrite/node/cell_container.py +7 -3
- mindspore/rewrite/node/control_flow.py +53 -28
- mindspore/rewrite/node/node.py +212 -196
- mindspore/rewrite/node/node_manager.py +51 -22
- mindspore/rewrite/node/node_topological_manager.py +3 -23
- mindspore/rewrite/parsers/__init__.py +12 -0
- mindspore/rewrite/parsers/arguments_parser.py +8 -9
- mindspore/rewrite/parsers/assign_parser.py +637 -413
- mindspore/rewrite/parsers/attribute_parser.py +3 -4
- mindspore/rewrite/parsers/class_def_parser.py +115 -148
- mindspore/rewrite/parsers/constant_parser.py +5 -5
- mindspore/rewrite/parsers/container_parser.py +4 -6
- mindspore/rewrite/parsers/expr_parser.py +55 -0
- mindspore/rewrite/parsers/for_parser.py +31 -98
- mindspore/rewrite/parsers/function_def_parser.py +13 -5
- mindspore/rewrite/parsers/if_parser.py +28 -10
- mindspore/rewrite/parsers/module_parser.py +8 -182
- mindspore/rewrite/parsers/parser.py +1 -5
- mindspore/rewrite/parsers/parser_register.py +1 -1
- mindspore/rewrite/parsers/return_parser.py +5 -10
- mindspore/rewrite/parsers/while_parser.py +59 -0
- mindspore/rewrite/sparsify/utils.py +1 -1
- mindspore/rewrite/symbol_tree/__init__.py +20 -0
- mindspore/rewrite/{symbol_tree.py → symbol_tree/symbol_tree.py} +704 -185
- mindspore/rewrite/{symbol_tree_builder.py → symbol_tree/symbol_tree_builder.py} +8 -8
- mindspore/rewrite/{symbol_tree_dumper.py → symbol_tree/symbol_tree_dumper.py} +4 -4
- mindspore/run_check/_check_version.py +6 -14
- mindspore/run_check/run_check.py +1 -1
- mindspore/safeguard/rewrite_obfuscation.py +9 -19
- mindspore/swresample-4.dll +0 -0
- mindspore/swscale-6.dll +0 -0
- mindspore/tinyxml2.dll +0 -0
- mindspore/train/__init__.py +6 -5
- mindspore/train/_utils.py +178 -4
- mindspore/train/amp.py +167 -245
- mindspore/train/anf_ir_pb2.py +14 -2
- mindspore/train/callback/__init__.py +5 -2
- mindspore/train/callback/_backup_and_restore.py +5 -5
- mindspore/train/callback/_callback.py +4 -4
- mindspore/train/callback/_checkpoint.py +151 -37
- mindspore/train/callback/_cluster_monitor.py +201 -0
- mindspore/train/callback/_early_stop.py +2 -2
- mindspore/train/callback/_flops_collector.py +238 -0
- mindspore/train/callback/_landscape.py +16 -11
- mindspore/train/callback/_loss_monitor.py +2 -2
- mindspore/train/callback/_mindio_ttp.py +443 -0
- mindspore/train/callback/_on_request_exit.py +2 -2
- mindspore/train/callback/_reduce_lr_on_plateau.py +2 -2
- mindspore/train/callback/_summary_collector.py +13 -14
- mindspore/train/callback/_time_monitor.py +3 -3
- mindspore/train/data_sink.py +6 -5
- mindspore/train/dataset_helper.py +66 -21
- mindspore/train/loss_scale_manager.py +2 -2
- mindspore/train/metrics/accuracy.py +7 -7
- mindspore/train/metrics/confusion_matrix.py +8 -6
- mindspore/train/metrics/cosine_similarity.py +6 -4
- mindspore/train/metrics/error.py +2 -2
- mindspore/train/metrics/metric.py +3 -3
- mindspore/train/metrics/perplexity.py +2 -1
- mindspore/train/metrics/topk.py +2 -2
- mindspore/train/mind_ir_pb2.py +89 -15
- mindspore/train/model.py +298 -56
- mindspore/train/serialization.py +501 -221
- mindspore/train/summary/_summary_adapter.py +1 -1
- mindspore/train/summary/_writer_pool.py +1 -1
- mindspore/train/summary/summary_record.py +56 -34
- mindspore/train/train_thor/convert_utils.py +3 -3
- mindspore/turbojpeg.dll +0 -0
- mindspore/version.py +1 -1
- {mindspore-2.2.11.dist-info → mindspore-2.3.0.dist-info}/METADATA +3 -3
- mindspore-2.3.0.dist-info/RECORD +1400 -0
- {mindspore-2.2.11.dist-info → mindspore-2.3.0.dist-info}/entry_points.txt +1 -0
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_adapter.py +0 -662
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_helper.py +0 -377
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_job.py +0 -201
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_job_manager.py +0 -515
- mindspore/gen_ops.py +0 -273
- mindspore/nn/layer/flash_attention.py +0 -189
- mindspore/ops/_op_impl/cpu/tensor_shape.py +0 -42
- mindspore/ops/_op_impl/tbe/__init__.py +0 -47
- mindspore/ops/_op_impl/tbe/abs.py +0 -38
- mindspore/ops/_op_impl/tbe/abs_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/abs_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/abs_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/accumulate_n_v2.py +0 -41
- mindspore/ops/_op_impl/tbe/accumulate_n_v2_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/acos.py +0 -37
- mindspore/ops/_op_impl/tbe/acos_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/acos_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/acos_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/acosh.py +0 -37
- mindspore/ops/_op_impl/tbe/acosh_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/acosh_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/acosh_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/act_ulq_clamp_max_grad.py +0 -38
- mindspore/ops/_op_impl/tbe/act_ulq_clamp_min_grad.py +0 -38
- mindspore/ops/_op_impl/tbe/acts_ulq.py +0 -45
- mindspore/ops/_op_impl/tbe/acts_ulq_input_grad.py +0 -38
- mindspore/ops/_op_impl/tbe/adam_apply_one.py +0 -50
- mindspore/ops/_op_impl/tbe/adam_apply_one_assign.py +0 -53
- mindspore/ops/_op_impl/tbe/adam_apply_one_ds.py +0 -51
- mindspore/ops/_op_impl/tbe/adam_apply_one_with_decay.py +0 -54
- mindspore/ops/_op_impl/tbe/adam_apply_one_with_decay_assign.py +0 -54
- mindspore/ops/_op_impl/tbe/adam_apply_one_with_decay_ds.py +0 -55
- mindspore/ops/_op_impl/tbe/adaptive_max_pool2d.py +0 -37
- mindspore/ops/_op_impl/tbe/add.py +0 -42
- mindspore/ops/_op_impl/tbe/add_ds.py +0 -43
- mindspore/ops/_op_impl/tbe/add_n.py +0 -39
- mindspore/ops/_op_impl/tbe/add_n_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/addcdiv.py +0 -41
- mindspore/ops/_op_impl/tbe/addcdiv_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/addcmul.py +0 -43
- mindspore/ops/_op_impl/tbe/addcmul_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/apply_ada_max.py +0 -68
- mindspore/ops/_op_impl/tbe/apply_ada_max_ds.py +0 -69
- mindspore/ops/_op_impl/tbe/apply_adadelta.py +0 -66
- mindspore/ops/_op_impl/tbe/apply_adadelta_ds.py +0 -67
- mindspore/ops/_op_impl/tbe/apply_adagrad.py +0 -55
- mindspore/ops/_op_impl/tbe/apply_adagrad_d_a.py +0 -67
- mindspore/ops/_op_impl/tbe/apply_adagrad_ds.py +0 -56
- mindspore/ops/_op_impl/tbe/apply_adagrad_v2.py +0 -48
- mindspore/ops/_op_impl/tbe/apply_adagrad_v2_ds.py +0 -49
- mindspore/ops/_op_impl/tbe/apply_adam.py +0 -79
- mindspore/ops/_op_impl/tbe/apply_adam_ds.py +0 -80
- mindspore/ops/_op_impl/tbe/apply_adam_with_amsgrad.py +0 -60
- mindspore/ops/_op_impl/tbe/apply_adam_with_amsgrad_ds.py +0 -61
- mindspore/ops/_op_impl/tbe/apply_add_sign.py +0 -65
- mindspore/ops/_op_impl/tbe/apply_add_sign_ds.py +0 -66
- mindspore/ops/_op_impl/tbe/apply_centered_rms_prop.py +0 -77
- mindspore/ops/_op_impl/tbe/apply_centered_rms_prop_ds.py +0 -78
- mindspore/ops/_op_impl/tbe/apply_ftrl.py +0 -67
- mindspore/ops/_op_impl/tbe/apply_ftrl_ds.py +0 -68
- mindspore/ops/_op_impl/tbe/apply_gradient_descent.py +0 -44
- mindspore/ops/_op_impl/tbe/apply_gradient_descent_ds.py +0 -45
- mindspore/ops/_op_impl/tbe/apply_keras_momentum.py +0 -49
- mindspore/ops/_op_impl/tbe/apply_momentum.py +0 -64
- mindspore/ops/_op_impl/tbe/apply_momentum_ds.py +0 -65
- mindspore/ops/_op_impl/tbe/apply_power_sign.py +0 -65
- mindspore/ops/_op_impl/tbe/apply_power_sign_ds.py +0 -66
- mindspore/ops/_op_impl/tbe/apply_proximal_adagrad.py +0 -57
- mindspore/ops/_op_impl/tbe/apply_proximal_adagrad_ds.py +0 -58
- mindspore/ops/_op_impl/tbe/apply_proximal_gradient_descent.py +0 -54
- mindspore/ops/_op_impl/tbe/apply_proximal_gradient_descent_ds.py +0 -55
- mindspore/ops/_op_impl/tbe/apply_rms_prop.py +0 -52
- mindspore/ops/_op_impl/tbe/approximate_equal.py +0 -39
- mindspore/ops/_op_impl/tbe/approximate_equal_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/arg_max.py +0 -38
- mindspore/ops/_op_impl/tbe/arg_max_with_value.py +0 -38
- mindspore/ops/_op_impl/tbe/arg_max_with_value_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/arg_min.py +0 -38
- mindspore/ops/_op_impl/tbe/arg_min_v2_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/arg_min_with_value.py +0 -38
- mindspore/ops/_op_impl/tbe/arg_min_with_value_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/asin.py +0 -37
- mindspore/ops/_op_impl/tbe/asin_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/asin_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/asin_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/asinh.py +0 -37
- mindspore/ops/_op_impl/tbe/asinh_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/asinh_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/asinh_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/assign.py +0 -79
- mindspore/ops/_op_impl/tbe/assign_add.py +0 -59
- mindspore/ops/_op_impl/tbe/assign_add_ds.py +0 -60
- mindspore/ops/_op_impl/tbe/assign_ds.py +0 -80
- mindspore/ops/_op_impl/tbe/assign_sub.py +0 -55
- mindspore/ops/_op_impl/tbe/assign_sub_ds.py +0 -56
- mindspore/ops/_op_impl/tbe/atan.py +0 -37
- mindspore/ops/_op_impl/tbe/atan2.py +0 -38
- mindspore/ops/_op_impl/tbe/atan2_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/atan_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/atan_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/atan_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/atanh.py +0 -37
- mindspore/ops/_op_impl/tbe/atanh_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/avg_pool.py +0 -43
- mindspore/ops/_op_impl/tbe/avg_pool_3d.py +0 -44
- mindspore/ops/_op_impl/tbe/avg_pool_3d_grad.py +0 -45
- mindspore/ops/_op_impl/tbe/avg_pool_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/avg_pool_grad.py +0 -42
- mindspore/ops/_op_impl/tbe/avg_pool_grad_vm.py +0 -42
- mindspore/ops/_op_impl/tbe/basic_lstm_cell.py +0 -57
- mindspore/ops/_op_impl/tbe/basic_lstm_cell_c_state_grad.py +0 -50
- mindspore/ops/_op_impl/tbe/basic_lstm_cell_c_state_grad_v2.py +0 -51
- mindspore/ops/_op_impl/tbe/basic_lstm_cell_input_grad.py +0 -42
- mindspore/ops/_op_impl/tbe/basic_lstm_cell_weight_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/batch_matmul.py +0 -42
- mindspore/ops/_op_impl/tbe/batch_matmul_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/batch_matmul_v2.py +0 -47
- mindspore/ops/_op_impl/tbe/batch_to_space.py +0 -38
- mindspore/ops/_op_impl/tbe/batch_to_space_nd.py +0 -38
- mindspore/ops/_op_impl/tbe/batch_to_space_nd_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/batch_to_space_nd_v2.py +0 -41
- mindspore/ops/_op_impl/tbe/batchnorm.py +0 -58
- mindspore/ops/_op_impl/tbe/batchnorm_grad.py +0 -58
- mindspore/ops/_op_impl/tbe/bce_with_logits_loss.py +0 -42
- mindspore/ops/_op_impl/tbe/bessel_i0e.py +0 -37
- mindspore/ops/_op_impl/tbe/bessel_i0e_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/bessel_i1e.py +0 -37
- mindspore/ops/_op_impl/tbe/bessel_i1e_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/bias_add.py +0 -38
- mindspore/ops/_op_impl/tbe/bias_add_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/bias_add_grad.py +0 -53
- mindspore/ops/_op_impl/tbe/binary_cross_entropy.py +0 -39
- mindspore/ops/_op_impl/tbe/binary_cross_entropy_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/binary_cross_entropy_grad.py +0 -44
- mindspore/ops/_op_impl/tbe/binary_cross_entropy_grad_ds.py +0 -45
- mindspore/ops/_op_impl/tbe/bitwise_and.py +0 -39
- mindspore/ops/_op_impl/tbe/bitwise_and_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/bitwise_or.py +0 -39
- mindspore/ops/_op_impl/tbe/bitwise_or_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/bitwise_xor.py +0 -39
- mindspore/ops/_op_impl/tbe/bitwise_xor_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/bn_infer.py +0 -43
- mindspore/ops/_op_impl/tbe/bn_infer_ds.py +0 -45
- mindspore/ops/_op_impl/tbe/bn_infer_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/bn_infer_grad_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/bn_inference.py +0 -50
- mindspore/ops/_op_impl/tbe/bn_training_reduce.py +0 -38
- mindspore/ops/_op_impl/tbe/bn_training_reduce_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/bn_training_reduce_grad.py +0 -46
- mindspore/ops/_op_impl/tbe/bn_training_reduce_grad_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/bn_training_update.py +0 -52
- mindspore/ops/_op_impl/tbe/bn_training_update_ds.py +0 -53
- mindspore/ops/_op_impl/tbe/bn_training_update_grad.py +0 -44
- mindspore/ops/_op_impl/tbe/bn_training_update_grad_ds.py +0 -45
- mindspore/ops/_op_impl/tbe/bn_training_update_v2.py +0 -48
- mindspore/ops/_op_impl/tbe/bn_training_update_v3.py +0 -51
- mindspore/ops/_op_impl/tbe/bounding_box_decode.py +0 -41
- mindspore/ops/_op_impl/tbe/bounding_box_decode_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/bounding_box_encode.py +0 -38
- mindspore/ops/_op_impl/tbe/broadcast_to.py +0 -40
- mindspore/ops/_op_impl/tbe/broadcast_to_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/cast.py +0 -55
- mindspore/ops/_op_impl/tbe/cast_ds.py +0 -58
- mindspore/ops/_op_impl/tbe/cdist.py +0 -38
- mindspore/ops/_op_impl/tbe/cdist_grad.py +0 -42
- mindspore/ops/_op_impl/tbe/ceil.py +0 -37
- mindspore/ops/_op_impl/tbe/ceil_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/celu.py +0 -39
- mindspore/ops/_op_impl/tbe/centralization.py +0 -39
- mindspore/ops/_op_impl/tbe/check_valid.py +0 -38
- mindspore/ops/_op_impl/tbe/check_valid_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/clip_by_norm_no_div_sum.py +0 -41
- mindspore/ops/_op_impl/tbe/clip_by_norm_no_div_sum_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/clip_by_value.py +0 -41
- mindspore/ops/_op_impl/tbe/clip_by_value_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/concat.py +0 -40
- mindspore/ops/_op_impl/tbe/concat_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/confusion_matrix.py +0 -63
- mindspore/ops/_op_impl/tbe/confusion_mul_grad.py +0 -40
- mindspore/ops/_op_impl/tbe/confusion_softmax_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/confusion_transpose_d.py +0 -39
- mindspore/ops/_op_impl/tbe/conv2d.py +0 -47
- mindspore/ops/_op_impl/tbe/conv2d_backprop_filter.py +0 -42
- mindspore/ops/_op_impl/tbe/conv2d_backprop_filter_ds.py +0 -43
- mindspore/ops/_op_impl/tbe/conv2d_backprop_input.py +0 -42
- mindspore/ops/_op_impl/tbe/conv2d_backprop_input_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/conv2d_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/conv2d_transpose.py +0 -48
- mindspore/ops/_op_impl/tbe/conv3d.py +0 -45
- mindspore/ops/_op_impl/tbe/conv3d_backprop_filter.py +0 -42
- mindspore/ops/_op_impl/tbe/conv3d_backprop_input.py +0 -42
- mindspore/ops/_op_impl/tbe/conv3d_transpose.py +0 -47
- mindspore/ops/_op_impl/tbe/conv3d_transpose_ds.py +0 -48
- mindspore/ops/_op_impl/tbe/cos.py +0 -37
- mindspore/ops/_op_impl/tbe/cos_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/cosh.py +0 -37
- mindspore/ops/_op_impl/tbe/cosh_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/ctc_loss_v2.py +0 -42
- mindspore/ops/_op_impl/tbe/ctc_loss_v2_grad.py +0 -44
- mindspore/ops/_op_impl/tbe/cum_sum.py +0 -42
- mindspore/ops/_op_impl/tbe/cum_sum_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/cummin.py +0 -41
- mindspore/ops/_op_impl/tbe/cumprod.py +0 -42
- mindspore/ops/_op_impl/tbe/data_format_dim_map.py +0 -38
- mindspore/ops/_op_impl/tbe/data_format_dim_map_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/deformable_offsets.py +0 -45
- mindspore/ops/_op_impl/tbe/deformable_offsets_grad.py +0 -48
- mindspore/ops/_op_impl/tbe/depth_to_space_ds.py +0 -49
- mindspore/ops/_op_impl/tbe/depthwise_conv2d.py +0 -44
- mindspore/ops/_op_impl/tbe/depthwise_conv2d_backprop_filter.py +0 -41
- mindspore/ops/_op_impl/tbe/depthwise_conv2d_backprop_input.py +0 -41
- mindspore/ops/_op_impl/tbe/diag.py +0 -38
- mindspore/ops/_op_impl/tbe/diag_part.py +0 -38
- mindspore/ops/_op_impl/tbe/dilation.py +0 -40
- mindspore/ops/_op_impl/tbe/div.py +0 -41
- mindspore/ops/_op_impl/tbe/div_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/div_no_nan.py +0 -41
- mindspore/ops/_op_impl/tbe/div_no_nan_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/dropout_do_mask.py +0 -38
- mindspore/ops/_op_impl/tbe/dropout_do_mask_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/dropout_do_mask_v3.py +0 -39
- mindspore/ops/_op_impl/tbe/dynamic_atomic_addr_clean.py +0 -34
- mindspore/ops/_op_impl/tbe/dynamic_gru_v2.py +0 -95
- mindspore/ops/_op_impl/tbe/dynamic_rnn.py +0 -82
- mindspore/ops/_op_impl/tbe/elu.py +0 -38
- mindspore/ops/_op_impl/tbe/elu_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/elu_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/elu_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/equal.py +0 -42
- mindspore/ops/_op_impl/tbe/equal_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/erf.py +0 -37
- mindspore/ops/_op_impl/tbe/erf_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/erfc.py +0 -37
- mindspore/ops/_op_impl/tbe/erfc_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/erfinv.py +0 -36
- mindspore/ops/_op_impl/tbe/exp.py +0 -40
- mindspore/ops/_op_impl/tbe/exp_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/expand_dims.py +0 -38
- mindspore/ops/_op_impl/tbe/expm1.py +0 -37
- mindspore/ops/_op_impl/tbe/expm1_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/extract_image_patches.py +0 -41
- mindspore/ops/_op_impl/tbe/extract_volume_patches.py +0 -39
- mindspore/ops/_op_impl/tbe/fake_quant_with_min_max_vars.py +0 -39
- mindspore/ops/_op_impl/tbe/fake_quant_with_min_max_vars_gradient.py +0 -43
- mindspore/ops/_op_impl/tbe/fake_quant_with_min_max_vars_per_channel.py +0 -39
- mindspore/ops/_op_impl/tbe/fake_quant_with_min_max_vars_per_channel_gradient.py +0 -43
- mindspore/ops/_op_impl/tbe/fast_gelu.py +0 -37
- mindspore/ops/_op_impl/tbe/fast_gelu_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/fast_gelu_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/fast_gelu_grad_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/fill.py +0 -56
- mindspore/ops/_op_impl/tbe/fill_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/flatten.py +0 -48
- mindspore/ops/_op_impl/tbe/floor.py +0 -37
- mindspore/ops/_op_impl/tbe/floor_div.py +0 -41
- mindspore/ops/_op_impl/tbe/floor_div_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/floor_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/floor_mod.py +0 -39
- mindspore/ops/_op_impl/tbe/floor_mod_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/fused_dbn_dw.py +0 -52
- mindspore/ops/_op_impl/tbe/fused_mul_add.py +0 -38
- mindspore/ops/_op_impl/tbe/fused_mul_add_n.py +0 -48
- mindspore/ops/_op_impl/tbe/fused_mul_add_n_l2loss.py +0 -53
- mindspore/ops/_op_impl/tbe/fused_mul_apply_momentum.py +0 -57
- mindspore/ops/_op_impl/tbe/fused_mul_apply_momentum_extern.py +0 -67
- mindspore/ops/_op_impl/tbe/gather_nd.py +0 -52
- mindspore/ops/_op_impl/tbe/gather_nd_ds.py +0 -48
- mindspore/ops/_op_impl/tbe/gather_v2.py +0 -56
- mindspore/ops/_op_impl/tbe/gather_v2_ds.py +0 -68
- mindspore/ops/_op_impl/tbe/gelu.py +0 -37
- mindspore/ops/_op_impl/tbe/gelu_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/gelu_grad.py +0 -42
- mindspore/ops/_op_impl/tbe/gelu_grad_ds.py +0 -43
- mindspore/ops/_op_impl/tbe/ger.py +0 -43
- mindspore/ops/_op_impl/tbe/ger_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/greater.py +0 -43
- mindspore/ops/_op_impl/tbe/greater_equal.py +0 -41
- mindspore/ops/_op_impl/tbe/greater_equal_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/gru_v2_hidden_grad.py +0 -51
- mindspore/ops/_op_impl/tbe/gru_v2_hidden_grad_cell.py +0 -52
- mindspore/ops/_op_impl/tbe/hard_swish.py +0 -37
- mindspore/ops/_op_impl/tbe/hard_swish_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/hard_swish_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/hard_swish_grad_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/histogram_fixed_width.py +0 -40
- mindspore/ops/_op_impl/tbe/hshrink.py +0 -33
- mindspore/ops/_op_impl/tbe/hshrink_grad.py +0 -37
- mindspore/ops/_op_impl/tbe/hsigmoid.py +0 -45
- mindspore/ops/_op_impl/tbe/hsigmoid_grad.py +0 -39
- mindspore/ops/_op_impl/tbe/ifmr.py +0 -47
- mindspore/ops/_op_impl/tbe/ifmr_ds.py +0 -48
- mindspore/ops/_op_impl/tbe/im2col.py +0 -42
- mindspore/ops/_op_impl/tbe/in_top_k.py +0 -37
- mindspore/ops/_op_impl/tbe/inplace_add.py +0 -39
- mindspore/ops/_op_impl/tbe/inplace_index_add.py +0 -46
- mindspore/ops/_op_impl/tbe/inplace_sub.py +0 -39
- mindspore/ops/_op_impl/tbe/inplace_update.py +0 -39
- mindspore/ops/_op_impl/tbe/inplace_update_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/inv.py +0 -38
- mindspore/ops/_op_impl/tbe/inv_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/inv_grad.py +0 -40
- mindspore/ops/_op_impl/tbe/inv_grad_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/invert.py +0 -37
- mindspore/ops/_op_impl/tbe/invert_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/iou.py +0 -38
- mindspore/ops/_op_impl/tbe/iou_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/is_close.py +0 -40
- mindspore/ops/_op_impl/tbe/kl_div_loss.py +0 -38
- mindspore/ops/_op_impl/tbe/kl_div_loss_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/kl_div_loss_grad.py +0 -40
- mindspore/ops/_op_impl/tbe/l2_loss.py +0 -36
- mindspore/ops/_op_impl/tbe/l2_loss_ds.py +0 -37
- mindspore/ops/_op_impl/tbe/l2_normalize.py +0 -38
- mindspore/ops/_op_impl/tbe/l2_normalize_grad.py +0 -40
- mindspore/ops/_op_impl/tbe/lamb_apply_optimizer_assign.py +0 -55
- mindspore/ops/_op_impl/tbe/lamb_apply_weight_assign.py +0 -42
- mindspore/ops/_op_impl/tbe/lamb_next_mv.py +0 -59
- mindspore/ops/_op_impl/tbe/lamb_next_mv_with_decay.py +0 -59
- mindspore/ops/_op_impl/tbe/lamb_next_right.py +0 -44
- mindspore/ops/_op_impl/tbe/lamb_update_with_lr.py +0 -48
- mindspore/ops/_op_impl/tbe/lamb_update_with_lr_v2.py +0 -44
- mindspore/ops/_op_impl/tbe/lars_update.py +0 -50
- mindspore/ops/_op_impl/tbe/lars_update_ds.py +0 -51
- mindspore/ops/_op_impl/tbe/layer_norm.py +0 -46
- mindspore/ops/_op_impl/tbe/layer_norm_beta_gamma_backprop.py +0 -44
- mindspore/ops/_op_impl/tbe/layer_norm_beta_gamma_backprop_ds.py +0 -45
- mindspore/ops/_op_impl/tbe/layer_norm_beta_gamma_backprop_v2.py +0 -40
- mindspore/ops/_op_impl/tbe/layer_norm_beta_gamma_backprop_v2_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/layer_norm_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/layer_norm_grad.py +0 -48
- mindspore/ops/_op_impl/tbe/layer_norm_x_backprop.py +0 -43
- mindspore/ops/_op_impl/tbe/layer_norm_x_backprop_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/layer_norm_x_backprop_v2.py +0 -45
- mindspore/ops/_op_impl/tbe/layer_norm_x_backprop_v2_ds.py +0 -45
- mindspore/ops/_op_impl/tbe/lerp.py +0 -38
- mindspore/ops/_op_impl/tbe/less.py +0 -41
- mindspore/ops/_op_impl/tbe/less_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/less_equal.py +0 -41
- mindspore/ops/_op_impl/tbe/less_equal_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/log.py +0 -40
- mindspore/ops/_op_impl/tbe/log1p.py +0 -37
- mindspore/ops/_op_impl/tbe/log1p_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/log_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/logical_and.py +0 -37
- mindspore/ops/_op_impl/tbe/logical_and_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/logical_not.py +0 -36
- mindspore/ops/_op_impl/tbe/logical_not_ds.py +0 -37
- mindspore/ops/_op_impl/tbe/logical_or.py +0 -37
- mindspore/ops/_op_impl/tbe/logical_or_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/logsoftmax.py +0 -37
- mindspore/ops/_op_impl/tbe/logsoftmax_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/logsoftmax_grad.py +0 -38
- mindspore/ops/_op_impl/tbe/logsoftmax_grad_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/lp_norm.py +0 -40
- mindspore/ops/_op_impl/tbe/lp_norm_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/lrn.py +0 -41
- mindspore/ops/_op_impl/tbe/lrn_grad.py +0 -42
- mindspore/ops/_op_impl/tbe/lstm_input_grad.py +0 -51
- mindspore/ops/_op_impl/tbe/masked_fill.py +0 -40
- mindspore/ops/_op_impl/tbe/masked_fill_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/matmul.py +0 -53
- mindspore/ops/_op_impl/tbe/matmul_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/matmul_v2.py +0 -50
- mindspore/ops/_op_impl/tbe/matrix_diag.py +0 -45
- mindspore/ops/_op_impl/tbe/matrix_diag_part.py +0 -45
- mindspore/ops/_op_impl/tbe/matrix_set_diag.py +0 -46
- mindspore/ops/_op_impl/tbe/max_pool.py +0 -39
- mindspore/ops/_op_impl/tbe/max_pool3d.py +0 -44
- mindspore/ops/_op_impl/tbe/max_pool3d_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/max_pool3d_grad_grad.py +0 -44
- mindspore/ops/_op_impl/tbe/max_pool_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/max_pool_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/max_pool_grad_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/max_pool_grad_grad_with_argmax.py +0 -41
- mindspore/ops/_op_impl/tbe/max_pool_grad_with_argmax.py +0 -42
- mindspore/ops/_op_impl/tbe/max_pool_with_argmax.py +0 -40
- mindspore/ops/_op_impl/tbe/maximum.py +0 -39
- mindspore/ops/_op_impl/tbe/maximum_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/maximum_grad.py +0 -46
- mindspore/ops/_op_impl/tbe/maximum_grad_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/mem_set.py +0 -38
- mindspore/ops/_op_impl/tbe/minimum.py +0 -40
- mindspore/ops/_op_impl/tbe/minimum_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/minimum_grad.py +0 -46
- mindspore/ops/_op_impl/tbe/minimum_grad_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/mish.py +0 -37
- mindspore/ops/_op_impl/tbe/mod.py +0 -41
- mindspore/ops/_op_impl/tbe/mod_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/mul.py +0 -37
- mindspore/ops/_op_impl/tbe/mul_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/mul_no_nan.py +0 -39
- mindspore/ops/_op_impl/tbe/mul_no_nan_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/multilabel_margin_loss.py +0 -39
- mindspore/ops/_op_impl/tbe/neg.py +0 -39
- mindspore/ops/_op_impl/tbe/neg_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/new_im2col.py +0 -40
- mindspore/ops/_op_impl/tbe/nll_loss.py +0 -41
- mindspore/ops/_op_impl/tbe/nll_loss_grad.py +0 -44
- mindspore/ops/_op_impl/tbe/nms_with_mask.py +0 -39
- mindspore/ops/_op_impl/tbe/not_equal.py +0 -41
- mindspore/ops/_op_impl/tbe/not_equal_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/npu_alloc_float_status.py +0 -34
- mindspore/ops/_op_impl/tbe/npu_clear_float_status.py +0 -35
- mindspore/ops/_op_impl/tbe/npu_clear_float_status_v2.py +0 -35
- mindspore/ops/_op_impl/tbe/npu_get_float_status.py +0 -35
- mindspore/ops/_op_impl/tbe/npu_get_float_status_v2.py +0 -35
- mindspore/ops/_op_impl/tbe/one_hot.py +0 -48
- mindspore/ops/_op_impl/tbe/one_hot_ds.py +0 -45
- mindspore/ops/_op_impl/tbe/ones_like.py +0 -40
- mindspore/ops/_op_impl/tbe/ones_like_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/p_s_r_o_i_pooling.py +0 -40
- mindspore/ops/_op_impl/tbe/p_s_r_o_i_pooling_grad.py +0 -40
- mindspore/ops/_op_impl/tbe/pack.py +0 -58
- mindspore/ops/_op_impl/tbe/pack_ds.py +0 -59
- mindspore/ops/_op_impl/tbe/pad_d.py +0 -40
- mindspore/ops/_op_impl/tbe/pad_d_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/parallel_concat.py +0 -70
- mindspore/ops/_op_impl/tbe/parallel_resize_bilinear.py +0 -45
- mindspore/ops/_op_impl/tbe/parallel_resize_bilinear_grad.py +0 -44
- mindspore/ops/_op_impl/tbe/pdist.py +0 -36
- mindspore/ops/_op_impl/tbe/pooling.py +0 -46
- mindspore/ops/_op_impl/tbe/population_count.py +0 -38
- mindspore/ops/_op_impl/tbe/pow.py +0 -41
- mindspore/ops/_op_impl/tbe/pow_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/prelu.py +0 -37
- mindspore/ops/_op_impl/tbe/prelu_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/prelu_grad.py +0 -40
- mindspore/ops/_op_impl/tbe/range.py +0 -39
- mindspore/ops/_op_impl/tbe/real_div.py +0 -38
- mindspore/ops/_op_impl/tbe/real_div_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/reciprocal.py +0 -36
- mindspore/ops/_op_impl/tbe/reciprocal_ds.py +0 -37
- mindspore/ops/_op_impl/tbe/reciprocal_grad.py +0 -38
- mindspore/ops/_op_impl/tbe/reciprocal_grad_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/reduce_all.py +0 -38
- mindspore/ops/_op_impl/tbe/reduce_all_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/reduce_any.py +0 -38
- mindspore/ops/_op_impl/tbe/reduce_any_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/reduce_max.py +0 -43
- mindspore/ops/_op_impl/tbe/reduce_max_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/reduce_mean.py +0 -40
- mindspore/ops/_op_impl/tbe/reduce_mean_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/reduce_min.py +0 -41
- mindspore/ops/_op_impl/tbe/reduce_min_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/reduce_prod.py +0 -42
- mindspore/ops/_op_impl/tbe/reduce_prod_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/reduce_std.py +0 -44
- mindspore/ops/_op_impl/tbe/reduce_sum.py +0 -39
- mindspore/ops/_op_impl/tbe/reduce_sum_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/relu.py +0 -39
- mindspore/ops/_op_impl/tbe/relu6.py +0 -38
- mindspore/ops/_op_impl/tbe/relu6_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/relu6_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/relu6_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/relu_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/relu_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/relu_grad_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/relu_grad_v2.py +0 -40
- mindspore/ops/_op_impl/tbe/relu_grad_v2_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/relu_v2.py +0 -40
- mindspore/ops/_op_impl/tbe/relu_v2_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/renorm.py +0 -39
- mindspore/ops/_op_impl/tbe/resize_bilinear.py +0 -40
- mindspore/ops/_op_impl/tbe/resize_bilinear_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/resize_bilinear_v2.py +0 -43
- mindspore/ops/_op_impl/tbe/resize_nearest_neighbor.py +0 -40
- mindspore/ops/_op_impl/tbe/resize_nearest_neighbor_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/resize_nearest_neighbor_grad.py +0 -39
- mindspore/ops/_op_impl/tbe/resize_nearest_neighbor_grad_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/reverse_v2_d.py +0 -37
- mindspore/ops/_op_impl/tbe/rint.py +0 -37
- mindspore/ops/_op_impl/tbe/rint_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/roi_align.py +0 -43
- mindspore/ops/_op_impl/tbe/roi_align_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/roi_align_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/roi_align_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/roll.py +0 -42
- mindspore/ops/_op_impl/tbe/round.py +0 -38
- mindspore/ops/_op_impl/tbe/round_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/rsqrt.py +0 -37
- mindspore/ops/_op_impl/tbe/rsqrt_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/rsqrt_grad.py +0 -40
- mindspore/ops/_op_impl/tbe/rsqrt_grad_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/scatter_add.py +0 -44
- mindspore/ops/_op_impl/tbe/scatter_div.py +0 -46
- mindspore/ops/_op_impl/tbe/scatter_max.py +0 -45
- mindspore/ops/_op_impl/tbe/scatter_min.py +0 -45
- mindspore/ops/_op_impl/tbe/scatter_mul.py +0 -44
- mindspore/ops/_op_impl/tbe/scatter_nd.py +0 -41
- mindspore/ops/_op_impl/tbe/scatter_nd_add.py +0 -45
- mindspore/ops/_op_impl/tbe/scatter_nd_d.py +0 -41
- mindspore/ops/_op_impl/tbe/scatter_nd_ds.py +0 -49
- mindspore/ops/_op_impl/tbe/scatter_nd_sub.py +0 -47
- mindspore/ops/_op_impl/tbe/scatter_nd_sub_ds.py +0 -48
- mindspore/ops/_op_impl/tbe/scatter_nd_update.py +0 -47
- mindspore/ops/_op_impl/tbe/scatter_nd_update_ds.py +0 -48
- mindspore/ops/_op_impl/tbe/scatter_non_aliasing_add.py +0 -39
- mindspore/ops/_op_impl/tbe/scatter_non_aliasing_add_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/scatter_sub.py +0 -47
- mindspore/ops/_op_impl/tbe/scatter_sub_ds.py +0 -48
- mindspore/ops/_op_impl/tbe/scatter_update.py +0 -43
- mindspore/ops/_op_impl/tbe/select.py +0 -38
- mindspore/ops/_op_impl/tbe/select_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/selu.py +0 -39
- mindspore/ops/_op_impl/tbe/selu_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/sgd.py +0 -62
- mindspore/ops/_op_impl/tbe/sigmoid.py +0 -37
- mindspore/ops/_op_impl/tbe/sigmoid_cross_entropy_with_logits.py +0 -41
- mindspore/ops/_op_impl/tbe/sigmoid_cross_entropy_with_logits_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/sigmoid_cross_entropy_with_logits_grad.py +0 -42
- mindspore/ops/_op_impl/tbe/sigmoid_cross_entropy_with_logits_grad_ds.py +0 -43
- mindspore/ops/_op_impl/tbe/sigmoid_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/sigmoid_grad.py +0 -39
- mindspore/ops/_op_impl/tbe/sigmoid_grad_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/sign.py +0 -38
- mindspore/ops/_op_impl/tbe/sign_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/sin.py +0 -37
- mindspore/ops/_op_impl/tbe/sin_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/sinh.py +0 -37
- mindspore/ops/_op_impl/tbe/sinh_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/slice.py +0 -58
- mindspore/ops/_op_impl/tbe/smooth_l1_loss.py +0 -45
- mindspore/ops/_op_impl/tbe/smooth_l1_loss_ds.py +0 -46
- mindspore/ops/_op_impl/tbe/smooth_l1_loss_grad.py +0 -46
- mindspore/ops/_op_impl/tbe/smooth_l1_loss_grad_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/soft_margin_loss.py +0 -38
- mindspore/ops/_op_impl/tbe/soft_margin_loss_grad.py +0 -39
- mindspore/ops/_op_impl/tbe/soft_shrink.py +0 -36
- mindspore/ops/_op_impl/tbe/soft_shrink_grad.py +0 -38
- mindspore/ops/_op_impl/tbe/softmax.py +0 -37
- mindspore/ops/_op_impl/tbe/softmax_cross_entropy_with_logits.py +0 -38
- mindspore/ops/_op_impl/tbe/softmax_cross_entropy_with_logits_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/softmax_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/softmax_grad_ext.py +0 -42
- mindspore/ops/_op_impl/tbe/softmax_v2_with_dropout_do_mask_v3.py +0 -39
- mindspore/ops/_op_impl/tbe/softplus.py +0 -37
- mindspore/ops/_op_impl/tbe/softplus_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/softplus_grad.py +0 -38
- mindspore/ops/_op_impl/tbe/softplus_grad_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/softsign.py +0 -37
- mindspore/ops/_op_impl/tbe/softsign_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/sort.py +0 -38
- mindspore/ops/_op_impl/tbe/sort_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/space_to_batch.py +0 -38
- mindspore/ops/_op_impl/tbe/space_to_batch_nd.py +0 -38
- mindspore/ops/_op_impl/tbe/space_to_depth.py +0 -47
- mindspore/ops/_op_impl/tbe/sparse_apply_adadelta.py +0 -56
- mindspore/ops/_op_impl/tbe/sparse_apply_adagrad.py +0 -45
- mindspore/ops/_op_impl/tbe/sparse_apply_adagrad_ds.py +0 -46
- mindspore/ops/_op_impl/tbe/sparse_apply_adagrad_v2.py +0 -46
- mindspore/ops/_op_impl/tbe/sparse_apply_adagrad_v2_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/sparse_apply_ftrl_d.py +0 -53
- mindspore/ops/_op_impl/tbe/sparse_apply_ftrl_d_ds.py +0 -50
- mindspore/ops/_op_impl/tbe/sparse_apply_ftrl_v2.py +0 -50
- mindspore/ops/_op_impl/tbe/sparse_apply_proximal_adagrad.py +0 -66
- mindspore/ops/_op_impl/tbe/sparse_apply_proximal_adagrad_ds.py +0 -67
- mindspore/ops/_op_impl/tbe/sparse_apply_r_m_s_prop.py +0 -57
- mindspore/ops/_op_impl/tbe/sparse_apply_r_m_s_prop_ds.py +0 -58
- mindspore/ops/_op_impl/tbe/sparse_gather_v2.py +0 -56
- mindspore/ops/_op_impl/tbe/sparse_gather_v2_ds.py +0 -58
- mindspore/ops/_op_impl/tbe/split_d.py +0 -38
- mindspore/ops/_op_impl/tbe/split_d_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/split_v.py +0 -39
- mindspore/ops/_op_impl/tbe/splitv.py +0 -39
- mindspore/ops/_op_impl/tbe/sqrt.py +0 -37
- mindspore/ops/_op_impl/tbe/sqrt_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/sqrt_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/sqrt_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/square.py +0 -38
- mindspore/ops/_op_impl/tbe/square_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/square_sum_all.py +0 -40
- mindspore/ops/_op_impl/tbe/square_sum_all_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/square_sum_v1.py +0 -38
- mindspore/ops/_op_impl/tbe/square_sum_v1_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/square_sum_v2.py +0 -39
- mindspore/ops/_op_impl/tbe/squared_difference.py +0 -39
- mindspore/ops/_op_impl/tbe/squared_difference_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/squeeze.py +0 -37
- mindspore/ops/_op_impl/tbe/strided_read.py +0 -38
- mindspore/ops/_op_impl/tbe/strided_slice_d.py +0 -44
- mindspore/ops/_op_impl/tbe/strided_slice_ds.py +0 -71
- mindspore/ops/_op_impl/tbe/strided_slice_grad_d.py +0 -51
- mindspore/ops/_op_impl/tbe/strided_slice_grad_ds.py +0 -57
- mindspore/ops/_op_impl/tbe/strided_write.py +0 -38
- mindspore/ops/_op_impl/tbe/sub.py +0 -39
- mindspore/ops/_op_impl/tbe/sub_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/tan.py +0 -38
- mindspore/ops/_op_impl/tbe/tan_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/tanh.py +0 -37
- mindspore/ops/_op_impl/tbe/tanh_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/tanh_grad.py +0 -39
- mindspore/ops/_op_impl/tbe/tanh_grad_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/tensor_move.py +0 -49
- mindspore/ops/_op_impl/tbe/tensor_move_ds.py +0 -50
- mindspore/ops/_op_impl/tbe/tensor_scatter_update.py +0 -41
- mindspore/ops/_op_impl/tbe/tile.py +0 -37
- mindspore/ops/_op_impl/tbe/tile_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/top_k.py +0 -42
- mindspore/ops/_op_impl/tbe/top_k_ds.py +0 -43
- mindspore/ops/_op_impl/tbe/trans_data.py +0 -167
- mindspore/ops/_op_impl/tbe/trans_data_ds.py +0 -180
- mindspore/ops/_op_impl/tbe/trans_data_rnn.py +0 -44
- mindspore/ops/_op_impl/tbe/transpose.py +0 -60
- mindspore/ops/_op_impl/tbe/transpose_d.py +0 -47
- mindspore/ops/_op_impl/tbe/transpose_nod.py +0 -60
- mindspore/ops/_op_impl/tbe/trunc.py +0 -39
- mindspore/ops/_op_impl/tbe/truncate_div.py +0 -41
- mindspore/ops/_op_impl/tbe/truncate_div_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/truncate_mod.py +0 -41
- mindspore/ops/_op_impl/tbe/truncate_mod_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/unpack.py +0 -38
- mindspore/ops/_op_impl/tbe/unpack_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/unsorted_segment_max.py +0 -49
- mindspore/ops/_op_impl/tbe/unsorted_segment_max_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/unsorted_segment_min.py +0 -49
- mindspore/ops/_op_impl/tbe/unsorted_segment_min_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/unsorted_segment_prod.py +0 -49
- mindspore/ops/_op_impl/tbe/unsorted_segment_prod_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/unsorted_segment_sum.py +0 -38
- mindspore/ops/_op_impl/tbe/unsorted_segment_sum_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/wts_arq.py +0 -40
- mindspore/ops/_op_impl/tbe/xdivy.py +0 -38
- mindspore/ops/_op_impl/tbe/xdivy_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/xlogy.py +0 -38
- mindspore/ops/_op_impl/tbe/xlogy_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/zeros_like.py +0 -41
- mindspore/ops/_op_impl/tbe/zeros_like_ds.py +0 -42
- mindspore/ops/_tracefunc.py +0 -241
- mindspore/ops/arg_dtype_cast.py +0 -54
- mindspore/rewrite/api/tree_node_helper.py +0 -60
- mindspore/rewrite/ast_helpers/ast_creator.py +0 -115
- mindspore/rewrite/ast_transformers/flatten_recursive_stmt.py +0 -267
- mindspore/rewrite/ast_transformers/remove_return_out_of_if.py +0 -228
- mindspore/rewrite/namespace.py +0 -53
- mindspore-2.2.11.dist-info/RECORD +0 -1920
- {mindspore-2.2.11.dist-info → mindspore-2.3.0.dist-info}/WHEEL +0 -0
- {mindspore-2.2.11.dist-info → mindspore-2.3.0.dist-info}/top_level.txt +0 -0
|
@@ -16,6 +16,7 @@
|
|
|
16
16
|
"""Inner operators."""
|
|
17
17
|
from types import FunctionType, MethodType
|
|
18
18
|
from collections.abc import Iterable
|
|
19
|
+
import os
|
|
19
20
|
import numpy as np
|
|
20
21
|
|
|
21
22
|
from mindspore.common import Tensor
|
|
@@ -23,7 +24,6 @@ from mindspore.common._stub_tensor import StubTensor
|
|
|
23
24
|
from mindspore.ops import composite as C
|
|
24
25
|
from mindspore.ops.operations.array_ops import Cast
|
|
25
26
|
from mindspore.ops.operations._scalar_ops import bit_or, bit_and
|
|
26
|
-
from mindspore.ops.operations.comm_ops import ReduceOp
|
|
27
27
|
from mindspore.ops import signature as sig
|
|
28
28
|
from mindspore.ops.operations.math_ops import _infer_shape_reduce
|
|
29
29
|
from mindspore.ops.primitive import PrimitiveWithCheck, PrimitiveWithInfer, prim_attr_register, Primitive, \
|
|
@@ -33,10 +33,12 @@ from mindspore._c_expression import typing
|
|
|
33
33
|
from mindspore import _checkparam as validator
|
|
34
34
|
from mindspore.common import dtype as mstype
|
|
35
35
|
from mindspore.common.parameter import Parameter
|
|
36
|
-
from mindspore.communication.management import GlobalComm, get_rank
|
|
36
|
+
from mindspore.communication.management import GlobalComm, get_rank, _get_group, get_group_size
|
|
37
37
|
from mindspore.common.api import _pynative_executor
|
|
38
38
|
from mindspore.common._register_for_adapter import ms_adapter_registry
|
|
39
39
|
from mindspore import ops
|
|
40
|
+
from ..auto_generate import TensorCopySlices, SiLU, Cummin, TopKRouter, ExtractImagePatches, DecoderKVCache, \
|
|
41
|
+
PromptKVCache, ApplyCamePart1, ApplyCamePart2, ApplyCamePart3, ApplyCamePart4
|
|
40
42
|
|
|
41
43
|
# Bit operation
|
|
42
44
|
bit_and = bit_and()
|
|
@@ -55,73 +57,28 @@ string_mul = Primitive("string_mul")
|
|
|
55
57
|
string_getitem = Primitive("string_getitem")
|
|
56
58
|
|
|
57
59
|
|
|
58
|
-
class
|
|
60
|
+
class Generator(Primitive):
|
|
59
61
|
r"""
|
|
60
|
-
|
|
61
|
-
The input tensor must be a 4-D tensor and the data format is NCHW.
|
|
62
|
-
|
|
63
|
-
Args:
|
|
64
|
-
ksizes (Union[tuple[int], list[int]]): The size of sliding window, must be a tuple or a list of integers,
|
|
65
|
-
and the format is [1, 1, ksize_row, ksize_col].
|
|
66
|
-
strides (Union[tuple[int], list[int]]): Distance between the centers of the two consecutive patches,
|
|
67
|
-
must be a tuple or list of int, and the format is [1, 1, stride_row, stride_col].
|
|
68
|
-
rates (Union[tuple[int], list[int]]): In each extracted patch, the gap between the corresponding dimension
|
|
69
|
-
pixel positions, must be a tuple or a list of integers, and the format is [1, 1, rate_row, rate_col].
|
|
70
|
-
padding (str): The type of padding algorithm, is a string whose value is "same" or "valid",
|
|
71
|
-
not case sensitive. Default: "valid".
|
|
72
|
-
|
|
73
|
-
- same: Means that the patch can take the part beyond the original image, and this part is filled with 0.
|
|
74
|
-
|
|
75
|
-
- valid: Means that the taken patch area must be completely covered in the original image.
|
|
62
|
+
Manage the state of random number generation.
|
|
76
63
|
|
|
77
64
|
Inputs:
|
|
78
|
-
- **
|
|
65
|
+
- **cmd** (int) : operation to be executed.
|
|
66
|
+
- **inputs** (tuple[tensor]) : inputs for the operation.
|
|
79
67
|
|
|
80
68
|
Outputs:
|
|
81
|
-
|
|
82
|
-
|
|
83
|
-
|
|
84
|
-
|
|
85
|
-
.. math::
|
|
86
|
-
out_depth=ksize\_row * ksize\_col * in\_depth
|
|
87
|
-
|
|
88
|
-
and
|
|
89
|
-
if 'padding' is "valid":
|
|
90
|
-
|
|
91
|
-
.. math::
|
|
92
|
-
out\_row=floor((in\_row - (ksize\_row + (ksize\_row - 1) * (rate\_row - 1))) / stride\_row) + 1
|
|
93
|
-
out\_col=floor((in\_col - (ksize\_col + (ksize\_col - 1) * (rate\_col - 1))) / stride\_col) + 1
|
|
94
|
-
|
|
95
|
-
if 'padding' is "same":
|
|
96
|
-
|
|
97
|
-
.. math::
|
|
98
|
-
out\_row=floor((in\_row - 1) / stride\_row) + 1
|
|
99
|
-
out\_col=floor((in\_col - 1) / stride\_col) + 1
|
|
100
|
-
|
|
101
|
-
Supported Platforms:
|
|
102
|
-
``Ascend`` ``GPU``
|
|
69
|
+
- **seed** (Tensor): Seed for the random number generation algorithm.
|
|
70
|
+
- **offset** (Tensor): Offset of the random number sequence.
|
|
71
|
+
- **state** (Tensor): State tensor, can be used to restore current state.
|
|
103
72
|
"""
|
|
104
73
|
|
|
105
74
|
@prim_attr_register
|
|
106
|
-
def __init__(self
|
|
107
|
-
""
|
|
108
|
-
|
|
109
|
-
|
|
110
|
-
|
|
111
|
-
|
|
112
|
-
|
|
113
|
-
f"{arg_name}_col, 1], but got {arg_val}.")
|
|
114
|
-
if not isinstance(arg_val[2], int) or not isinstance(arg_val[3], int) or arg_val[2] < 1 or arg_val[3] < 1:
|
|
115
|
-
raise ValueError(f"For '{prim_name}' the {arg_name}_row and {arg_name}_col in {arg_name}s must be "
|
|
116
|
-
f"an positive integer number, but got {arg_name}_row is {arg_val[2]}, "
|
|
117
|
-
f"{arg_name}_col is {arg_val[3]}")
|
|
118
|
-
|
|
119
|
-
_check_tuple_or_list("ksize", ksizes, self.name)
|
|
120
|
-
_check_tuple_or_list("stride", strides, self.name)
|
|
121
|
-
_check_tuple_or_list("rate", rates, self.name)
|
|
122
|
-
validator.check_value_type('padding', padding, [str], self.name)
|
|
123
|
-
self.padding = validator.check_string(padding.upper(), ['VALID', 'SAME'], 'padding', self.name)
|
|
124
|
-
self.add_prim_attr("padding", self.padding)
|
|
75
|
+
def __init__(self):
|
|
76
|
+
self.add_prim_attr("side_effect_mem", True)
|
|
77
|
+
|
|
78
|
+
def __call__(self, cmd, inputs):
|
|
79
|
+
if cmd == 0: # step cmd
|
|
80
|
+
return inputs[0], inputs[1]
|
|
81
|
+
return super().__call__(cmd, inputs)
|
|
125
82
|
|
|
126
83
|
|
|
127
84
|
class Quant(PrimitiveWithInfer):
|
|
@@ -139,7 +96,7 @@ class Quant(PrimitiveWithInfer):
|
|
|
139
96
|
y = round(scale * x * scale + offset)
|
|
140
97
|
|
|
141
98
|
Note:
|
|
142
|
-
This operation only support
|
|
99
|
+
This operation only support Atlas 200/300/500 inference product.
|
|
143
100
|
|
|
144
101
|
Args:
|
|
145
102
|
scale (float) : Specifies the scaling ratio.
|
|
@@ -252,7 +209,7 @@ class Dequant(PrimitiveWithInfer):
|
|
|
252
209
|
y = x * deq\_scale * deq\_scale
|
|
253
210
|
|
|
254
211
|
Note:
|
|
255
|
-
This operation only support
|
|
212
|
+
This operation only support Atlas 200/300/500 inference product.
|
|
256
213
|
|
|
257
214
|
Args:
|
|
258
215
|
sqrt_mode (bool) : Specifies whether to perform square root on `scale`. Default: ``False``.
|
|
@@ -273,10 +230,10 @@ class Dequant(PrimitiveWithInfer):
|
|
|
273
230
|
"""
|
|
274
231
|
|
|
275
232
|
@prim_attr_register
|
|
276
|
-
def __init__(self, sqrt_mode=False, relu_flag=False):
|
|
233
|
+
def __init__(self, sqrt_mode=False, relu_flag=False, dtype=mstype.float16):
|
|
277
234
|
self.sqrt_mode = validator.check_value_type("sqrt_mode", sqrt_mode, [bool], self.name)
|
|
278
235
|
self.relu_flag = validator.check_value_type("relu_flag", relu_flag, [bool], self.name)
|
|
279
|
-
self.
|
|
236
|
+
self.dtype = dtype
|
|
280
237
|
|
|
281
238
|
def infer_shape(self, x_shape, deq_scale_shape):
|
|
282
239
|
return x_shape
|
|
@@ -288,6 +245,53 @@ class Dequant(PrimitiveWithInfer):
|
|
|
288
245
|
return mstype.float16
|
|
289
246
|
|
|
290
247
|
|
|
248
|
+
class AntiQuant(Primitive):
|
|
249
|
+
r"""
|
|
250
|
+
Returns the antiquantized value of input_x.
|
|
251
|
+
|
|
252
|
+
If `sqrt_mode` is False:
|
|
253
|
+
|
|
254
|
+
.. math::
|
|
255
|
+
y = scale * (x + offset)
|
|
256
|
+
|
|
257
|
+
If `sqrt_mode` is True:
|
|
258
|
+
|
|
259
|
+
.. math::
|
|
260
|
+
y = scale * scale * (x + offset)
|
|
261
|
+
|
|
262
|
+
Note:
|
|
263
|
+
This operation only support Atlas 200/300/500 inference product.
|
|
264
|
+
|
|
265
|
+
Args:
|
|
266
|
+
scale (float) : Specifies the scaling ratio.
|
|
267
|
+
offset (float): Specifies the offset.
|
|
268
|
+
sqrt_mode (bool) : Specifies whether to perform square root on `scale`. Default: ``False``.
|
|
269
|
+
|
|
270
|
+
Inputs:
|
|
271
|
+
- **input_x** (Tensor) : Input tensor. Must be mindspore.int8.
|
|
272
|
+
|
|
273
|
+
Outputs:
|
|
274
|
+
- Tensor: The antiquantized output tensor of type mindspore.float32.
|
|
275
|
+
|
|
276
|
+
Examples:
|
|
277
|
+
>>> from mindspore.ops.operations._inner_ops import AntiQuant
|
|
278
|
+
>>> input_x = Tensor([50.0, 20.0], mstype.int8)
|
|
279
|
+
>>> antiquant = AntiQuant(2.0, 1.0, False)
|
|
280
|
+
>>> y = antiquant(input_x)
|
|
281
|
+
>>> print(y)
|
|
282
|
+
[102. 42.]
|
|
283
|
+
"""
|
|
284
|
+
|
|
285
|
+
@prim_attr_register
|
|
286
|
+
def __init__(self, sqrt_mode=False, dtype=mstype.float16):
|
|
287
|
+
super().__init__("AntiQuant")
|
|
288
|
+
self.sqrt_mode = validator.check_value_type("sqrt_mode", sqrt_mode, [bool], self.name)
|
|
289
|
+
self.dtype = dtype
|
|
290
|
+
|
|
291
|
+
self.init_prim_io_names(inputs=['x', 'scale', 'offset'],
|
|
292
|
+
outputs=['y'])
|
|
293
|
+
|
|
294
|
+
|
|
291
295
|
class MatrixDiag(PrimitiveWithInfer):
|
|
292
296
|
"""
|
|
293
297
|
Returns a batched diagonal tensor with a given batched diagonal values.
|
|
@@ -385,227 +389,6 @@ class MatrixDiagPart(PrimitiveWithInfer):
|
|
|
385
389
|
return out_shape
|
|
386
390
|
|
|
387
391
|
|
|
388
|
-
class Send(PrimitiveWithInfer):
|
|
389
|
-
"""
|
|
390
|
-
Send tensors from src_rank to the specified dest_rank.
|
|
391
|
-
|
|
392
|
-
Note:
|
|
393
|
-
Send and Receive must be used in combination and have same sr_tag.
|
|
394
|
-
Send must be used between servers.
|
|
395
|
-
|
|
396
|
-
Args:
|
|
397
|
-
sr_tag (int): A required integer identifying the send/recv message tag. The message will
|
|
398
|
-
will be received by the Receive op with the same "sr_tag".
|
|
399
|
-
dest_rank (int): A required integer identifying the destination rank.
|
|
400
|
-
group (str): The communication group to work on. Default: "hccl_world_group/nccl_world_group".
|
|
401
|
-
|
|
402
|
-
Inputs:
|
|
403
|
-
- **input_x** (Tensor) - The shape of tensor is :math:`(x_1, x_2, ..., x_R)`.
|
|
404
|
-
|
|
405
|
-
Examples:
|
|
406
|
-
>>> import mindspore.ops as ops
|
|
407
|
-
>>> import mindspore.nn as nn
|
|
408
|
-
>>> from mindspore.communication import init
|
|
409
|
-
>>> from mindspore import Tensor
|
|
410
|
-
>>> import numpy as np
|
|
411
|
-
>>>
|
|
412
|
-
>>> init()
|
|
413
|
-
>>> class Net(nn.Cell):
|
|
414
|
-
>>> def __init__(self):
|
|
415
|
-
>>> super(Net, self).__init__()
|
|
416
|
-
>>> self.depend = ops.Depend()
|
|
417
|
-
>>> self.send = ops.Send(st_tag=0, dest_rank=8, group="hccl_world_group")
|
|
418
|
-
>>>
|
|
419
|
-
>>> def construct(self, x):
|
|
420
|
-
>>> out = self.depend(x, self.send(x))
|
|
421
|
-
>>> return out
|
|
422
|
-
>>>
|
|
423
|
-
>>> input_ = Tensor(np.ones([2, 8]).astype(np.float32))
|
|
424
|
-
>>> net = Net()
|
|
425
|
-
>>> output = net(input_)
|
|
426
|
-
"""
|
|
427
|
-
|
|
428
|
-
@prim_attr_register
|
|
429
|
-
def __init__(self, sr_tag, dest_rank, group=GlobalComm.WORLD_COMM_GROUP, group_back=GlobalComm.WORLD_COMM_GROUP):
|
|
430
|
-
self.rank = dest_rank
|
|
431
|
-
self.sr_tag = sr_tag
|
|
432
|
-
self.group = group
|
|
433
|
-
self.add_prim_attr("no_eliminate", True)
|
|
434
|
-
|
|
435
|
-
def infer_shape(self, x_shape):
|
|
436
|
-
self.add_prim_attr("shape", x_shape)
|
|
437
|
-
return x_shape
|
|
438
|
-
|
|
439
|
-
def infer_dtype(self, x_dtype):
|
|
440
|
-
return x_dtype
|
|
441
|
-
|
|
442
|
-
|
|
443
|
-
class Receive(PrimitiveWithInfer):
|
|
444
|
-
"""
|
|
445
|
-
Receive tensors from src_rank.
|
|
446
|
-
|
|
447
|
-
Note:
|
|
448
|
-
Send and Receive must be used in combination and have same sr_tag.
|
|
449
|
-
Receive must be used between servers.
|
|
450
|
-
|
|
451
|
-
Args:
|
|
452
|
-
sr_tag (int): A required integer identifying the send/recv message tag. The message will
|
|
453
|
-
will be send by the Send op with the same "sr_tag".
|
|
454
|
-
src_rank (int): A required integer identifying the source rank.
|
|
455
|
-
shape (list[int]): A required list identifying the shape of the tensor to be received.
|
|
456
|
-
dtype (Type): A required Type identifying the type of the tensor to be received. The supported types:
|
|
457
|
-
int8, int16, int32, float16, float32.
|
|
458
|
-
group (str, optional): The communication group to work on.
|
|
459
|
-
Default: "hccl_world_group" on Ascend, "nccl_world_group" on GPU.
|
|
460
|
-
|
|
461
|
-
Inputs:
|
|
462
|
-
- **input_x** (Tensor) - The shape of tensor is :math:`(x_1, x_2, ..., x_R)`.
|
|
463
|
-
|
|
464
|
-
Examples:
|
|
465
|
-
>>> import mindspore.ops as ops
|
|
466
|
-
>>> import mindspore.nn as nn
|
|
467
|
-
>>> from mindspore.communication import init
|
|
468
|
-
>>> from mindspore import Tensor
|
|
469
|
-
>>> import numpy as np
|
|
470
|
-
>>>
|
|
471
|
-
>>> init()
|
|
472
|
-
>>> class Net(nn.Cell):
|
|
473
|
-
>>> def __init__(self):
|
|
474
|
-
>>> super(Net, self).__init__()
|
|
475
|
-
>>> self.recv = ops.Receive(st_tag=0, src_rank=0, shape=[2, 8], dtype=np.float32,
|
|
476
|
-
>>> group="hccl_world_group")
|
|
477
|
-
>>>
|
|
478
|
-
>>> def construct(self):
|
|
479
|
-
>>> out = self.recv()
|
|
480
|
-
>>> return out
|
|
481
|
-
>>>
|
|
482
|
-
>>> net = Net()
|
|
483
|
-
>>> output = net()
|
|
484
|
-
"""
|
|
485
|
-
|
|
486
|
-
@prim_attr_register
|
|
487
|
-
def __init__(self, sr_tag, src_rank, shape, dtype, group=GlobalComm.WORLD_COMM_GROUP,
|
|
488
|
-
group_back=GlobalComm.WORLD_COMM_GROUP):
|
|
489
|
-
self.rank = src_rank
|
|
490
|
-
self.tag = sr_tag
|
|
491
|
-
self.shape = shape
|
|
492
|
-
self.dtype = dtype
|
|
493
|
-
self.group = group
|
|
494
|
-
self.add_prim_attr("no_eliminate", True)
|
|
495
|
-
valid_type = [mstype.float16, mstype.bfloat16, mstype.float32, mstype.int32, mstype.int8, mstype.uint8]
|
|
496
|
-
args = {"dtype": dtype}
|
|
497
|
-
validator.check_scalar_or_tensor_types_same(args, valid_type, self.name)
|
|
498
|
-
|
|
499
|
-
def infer_shape(self, x_shape=None):
|
|
500
|
-
return self.get_attr_dict()['shape']
|
|
501
|
-
|
|
502
|
-
def infer_dtype(self, x_dtype=None):
|
|
503
|
-
return self.get_attr_dict()['dtype']
|
|
504
|
-
|
|
505
|
-
|
|
506
|
-
class Reduce(PrimitiveWithInfer):
|
|
507
|
-
"""
|
|
508
|
-
Reduces tensor across the processes in the specified communication group.
|
|
509
|
-
|
|
510
|
-
Note:
|
|
511
|
-
Only process with destination rank receives the reduced output.
|
|
512
|
-
Other processes only get a tensor with shape [1], which has no mathematical meaning.
|
|
513
|
-
|
|
514
|
-
Args:
|
|
515
|
-
dest_rank (int): Specifies the rank of the process that receives the reduced output.
|
|
516
|
-
op (str, optional): Specifies an operation used for element-wise reductions, like sum, prod, max, and min.
|
|
517
|
-
On the CPU, only 'sum' is supported. Default: ``ReduceOp.SUM`` .
|
|
518
|
-
group (str, optional): The communication group to work on.
|
|
519
|
-
Default: "hccl_world_group" on Ascend, "nccl_world_group" on GPU.
|
|
520
|
-
|
|
521
|
-
Inputs:
|
|
522
|
-
- **input_x** (Tensor) - The shape of tensor is :math:`(x_1, x_2, ..., x_R)`.
|
|
523
|
-
|
|
524
|
-
Examples:
|
|
525
|
-
>>> import mindspore.ops as ops
|
|
526
|
-
>>> import mindspore.nn as nn
|
|
527
|
-
>>> from mindspore.communication import init
|
|
528
|
-
>>> from mindspore import Tensor
|
|
529
|
-
>>> import numpy as np
|
|
530
|
-
>>> # Launch 4 processes.
|
|
531
|
-
>>> init()
|
|
532
|
-
>>> class ReduceNet(nn.Cell):
|
|
533
|
-
>>> def __init__(self):
|
|
534
|
-
>>> super(Net, self).__init__()
|
|
535
|
-
>>> self.reduce = ops.Reduce(dest_rank=1)
|
|
536
|
-
>>>
|
|
537
|
-
>>> def construct(self, x):
|
|
538
|
-
>>> out = self.reduce(x)
|
|
539
|
-
>>> return out
|
|
540
|
-
>>> input = Tensor(np.ones([2, 8]).astype(np.float32))
|
|
541
|
-
>>> net = ReduceNet()
|
|
542
|
-
>>> output = net(input)
|
|
543
|
-
>>> print(output)
|
|
544
|
-
Process with rank 1: [[4. 4. 4. 4. 4. 4. 4. 4.]
|
|
545
|
-
[4. 4. 4. 4. 4. 4. 4. 4.]],
|
|
546
|
-
Other proesses: [0.].
|
|
547
|
-
"""
|
|
548
|
-
|
|
549
|
-
@prim_attr_register
|
|
550
|
-
def __init__(self, dest_rank, op=ReduceOp.SUM, group=GlobalComm.WORLD_COMM_GROUP):
|
|
551
|
-
self.dest_rank = dest_rank
|
|
552
|
-
self.op = op
|
|
553
|
-
self.group = group
|
|
554
|
-
|
|
555
|
-
def infer_shape(self, x_shape):
|
|
556
|
-
# The process with dest_rank returns the reduced output.
|
|
557
|
-
# Other processes only gets a tensor with shape [1], which has no mathematical meaning.
|
|
558
|
-
if self.dest_rank == get_rank():
|
|
559
|
-
return x_shape
|
|
560
|
-
return [1]
|
|
561
|
-
|
|
562
|
-
def infer_dtype(self, x_dtype):
|
|
563
|
-
return x_dtype
|
|
564
|
-
|
|
565
|
-
|
|
566
|
-
class Barrier(PrimitiveWithInfer):
|
|
567
|
-
"""
|
|
568
|
-
Synchronizes all processes in the specified group.
|
|
569
|
-
|
|
570
|
-
Note:
|
|
571
|
-
After calling this collective operator,
|
|
572
|
-
this process will be blocked until all other processes in the group call this operator.
|
|
573
|
-
|
|
574
|
-
Args:
|
|
575
|
-
group (str, optional): The communication group to work on.
|
|
576
|
-
Default: "hccl_world_group" on Ascend, "nccl_world_group" on GPU.
|
|
577
|
-
|
|
578
|
-
Examples:
|
|
579
|
-
>>> import mindspore.ops as ops
|
|
580
|
-
>>> import mindspore.nn as nn
|
|
581
|
-
>>> from mindspore.communication import init
|
|
582
|
-
>>> from mindspore import Tensor
|
|
583
|
-
>>> import numpy as np
|
|
584
|
-
>>> # Launch 4 processes.
|
|
585
|
-
>>> init()
|
|
586
|
-
>>> class BarrierNet(nn.Cell):
|
|
587
|
-
>>> def __init__(self):
|
|
588
|
-
>>> super(Net, self).__init__()
|
|
589
|
-
>>> self.barrier = ops.Barrier()
|
|
590
|
-
>>>
|
|
591
|
-
>>> def construct(self):
|
|
592
|
-
>>> self.barrier()
|
|
593
|
-
>>> net = BarrierNet()
|
|
594
|
-
>>> net()
|
|
595
|
-
"""
|
|
596
|
-
|
|
597
|
-
@prim_attr_register
|
|
598
|
-
def __init__(self, group=GlobalComm.WORLD_COMM_GROUP):
|
|
599
|
-
self.group = group
|
|
600
|
-
self.add_prim_attr("side_effect_mem", True)
|
|
601
|
-
|
|
602
|
-
def infer_shape(self):
|
|
603
|
-
return [1]
|
|
604
|
-
|
|
605
|
-
def infer_dtype(self):
|
|
606
|
-
return mstype.float32
|
|
607
|
-
|
|
608
|
-
|
|
609
392
|
class MatrixSetDiag(PrimitiveWithInfer):
|
|
610
393
|
r"""
|
|
611
394
|
Modifies the batched diagonal part of a batched tensor.
|
|
@@ -1008,7 +791,7 @@ class Centralization(PrimitiveWithInfer):
|
|
|
1008
791
|
|
|
1009
792
|
Inputs:
|
|
1010
793
|
- **input_x** (Tensor) - The input tensor. The data type mast be float16 or float32.
|
|
1011
|
-
- **axis** (Union[
|
|
794
|
+
- **axis** (Union[int, Tuple(int), List(int)]) - The dimensions to reduce. Default: (), reduce all dimensions.
|
|
1012
795
|
Only constant value is allowed. Must be in the range [-rank(input_x), rank(input_x)).
|
|
1013
796
|
|
|
1014
797
|
Outputs:
|
|
@@ -1328,45 +1111,6 @@ class DynamicBroadcastGradientArgs(Primitive):
|
|
|
1328
1111
|
"""Init BroadcastGradientArgs"""
|
|
1329
1112
|
|
|
1330
1113
|
|
|
1331
|
-
class TensorCopySlices(Primitive):
|
|
1332
|
-
"""
|
|
1333
|
-
Copy continues memory.
|
|
1334
|
-
|
|
1335
|
-
Inputs:
|
|
1336
|
-
- **x** (Tensor) - The target Tensor.
|
|
1337
|
-
- **value** (Tensor) - The tensor to update x.
|
|
1338
|
-
- **begin** (tuple[int]) - A tuple which represents the location where to start. Only
|
|
1339
|
-
constant value is allowed.
|
|
1340
|
-
- **end** (tuple[int]) - A tuple or which represents the maximum location where to end.
|
|
1341
|
-
Only constant value is allowed.
|
|
1342
|
-
- **strides** (tuple[int]) - A tuple which represents the stride is continuously added
|
|
1343
|
-
before reaching the maximum location. Only constant value is allowed.
|
|
1344
|
-
|
|
1345
|
-
Outputs:
|
|
1346
|
-
- **y** (Tensor), has the same shape and data type of x.
|
|
1347
|
-
|
|
1348
|
-
Examples:
|
|
1349
|
-
>>> import numpy as np
|
|
1350
|
-
>>> from mindspore.ops.operations import _inner_ops
|
|
1351
|
-
>>> copy_slices = _inner_ops.TensorCopySlices()
|
|
1352
|
-
>>> out = copy_slices(Tensor(np.zeros((5, 5))), Tensor(np.ones((2, 5))), (3, 0), (5, 5), (1, 1))
|
|
1353
|
-
>>> print(out)
|
|
1354
|
-
[[1., 1., 1., 1., 1.],
|
|
1355
|
-
[1., 1., 1., 1., 1.],
|
|
1356
|
-
[1., 1., 1., 1., 1.],
|
|
1357
|
-
[0., 0., 0., 0., 0.],
|
|
1358
|
-
[0., 0., 0., 0., 0.]]
|
|
1359
|
-
|
|
1360
|
-
Supported Platforms:
|
|
1361
|
-
``Ascend`` ``GPU`` ``CPU``
|
|
1362
|
-
"""
|
|
1363
|
-
|
|
1364
|
-
@prim_attr_register
|
|
1365
|
-
def __init__(self):
|
|
1366
|
-
"""Initialize TensorScatterUpdate"""
|
|
1367
|
-
self.init_prim_io_names(inputs=['x', 'value', 'begin', 'end', 'strides'], outputs=['y'])
|
|
1368
|
-
|
|
1369
|
-
|
|
1370
1114
|
class DSDMatmul(PrimitiveWithInfer):
|
|
1371
1115
|
"""
|
|
1372
1116
|
The definition of the CusSquare primitive.
|
|
@@ -1591,46 +1335,6 @@ class DynamicBroadcastTo(Primitive):
|
|
|
1591
1335
|
self.init_prim_io_names(inputs=['x', 'shape'], outputs=['y'])
|
|
1592
1336
|
|
|
1593
1337
|
|
|
1594
|
-
class Cummin(Primitive):
|
|
1595
|
-
r"""
|
|
1596
|
-
Returns the cumulative minimum of elements and the index.
|
|
1597
|
-
|
|
1598
|
-
.. warning::
|
|
1599
|
-
This is an experimental API that is subject to change or deletion.
|
|
1600
|
-
|
|
1601
|
-
Refer to :func:`mindspore.ops.cummin` for more detail.
|
|
1602
|
-
|
|
1603
|
-
Args:
|
|
1604
|
-
axis (int): The axis to accumulate the tensor's value. Must be in the range [-rank(input), rank(input)).
|
|
1605
|
-
|
|
1606
|
-
Inputs:
|
|
1607
|
-
- **input** (Tensor) - The input tensor.
|
|
1608
|
-
|
|
1609
|
-
Outputs:
|
|
1610
|
-
A tuple of 2 Tensors(values, indices), containing the cumulative minimum of elements and the index,
|
|
1611
|
-
The shape of each output tensor is the same as input `input`.
|
|
1612
|
-
|
|
1613
|
-
Supported Platforms:
|
|
1614
|
-
``Ascend`` ``GPU`` ``CPU``
|
|
1615
|
-
|
|
1616
|
-
Examples:
|
|
1617
|
-
>>> from mindspore import Tensor, ops
|
|
1618
|
-
>>> import mindspore
|
|
1619
|
-
>>> a = Tensor([-0.2284, -0.6628, 0.0975, 0.2680, -1.3298, -0.4220], mindspore.float32)
|
|
1620
|
-
>>> func = ops.Cummin(axis=0)
|
|
1621
|
-
>>> output = func(a)
|
|
1622
|
-
>>> print(output[0])
|
|
1623
|
-
[-0.2284 -0.6628 -0.6628 -0.6628 -1.3298 -1.3298]
|
|
1624
|
-
>>> print(output[1])
|
|
1625
|
-
[0 1 1 1 4 4]
|
|
1626
|
-
"""
|
|
1627
|
-
|
|
1628
|
-
@prim_attr_register
|
|
1629
|
-
def __init__(self, axis):
|
|
1630
|
-
"""Initialize Cummin"""
|
|
1631
|
-
validator.check_value_type('axis', axis, [int], self.name)
|
|
1632
|
-
|
|
1633
|
-
|
|
1634
1338
|
class DynamicResizeNearestNeighbor(Primitive):
|
|
1635
1339
|
r"""
|
|
1636
1340
|
Resizes the input tensor by using the nearest neighbor algorithm.
|
|
@@ -1947,7 +1651,6 @@ class Format(PrimitiveWithInfer):
|
|
|
1947
1651
|
def __init__(self):
|
|
1948
1652
|
self.init_prim_io_names(inputs=['string', 'args'], outputs=['string'])
|
|
1949
1653
|
|
|
1950
|
-
|
|
1951
1654
|
def __infer__(self, str_, *var):
|
|
1952
1655
|
def check_variable(str_, var):
|
|
1953
1656
|
if _check_contains_variable(str_['dtype'], str_['value']):
|
|
@@ -1958,11 +1661,9 @@ class Format(PrimitiveWithInfer):
|
|
|
1958
1661
|
return True
|
|
1959
1662
|
return False
|
|
1960
1663
|
|
|
1961
|
-
|
|
1962
1664
|
if check_variable(str_, var):
|
|
1963
1665
|
return {'dtype': mstype.string, 'shape': [], 'value': None}
|
|
1964
1666
|
|
|
1965
|
-
|
|
1966
1667
|
str_value = str_['value']
|
|
1967
1668
|
kwargs = dict()
|
|
1968
1669
|
var_value = list()
|
|
@@ -2147,14 +1848,13 @@ class ClipByNorm(PrimitiveWithInfer):
|
|
|
2147
1848
|
@prim_attr_register
|
|
2148
1849
|
def __init__(self, axis=None):
|
|
2149
1850
|
"""Initialize ClipByNorm"""
|
|
2150
|
-
self.axis_str = 'axis'
|
|
2151
1851
|
self.axis = () if axis is None else axis
|
|
2152
|
-
validator.check_value_type(
|
|
1852
|
+
validator.check_value_type('axis', self.axis, [int, tuple, list], self.name)
|
|
2153
1853
|
axis_check = self.axis if isinstance(self.axis, Iterable) else (self.axis,)
|
|
2154
1854
|
for i, value in enumerate(axis_check):
|
|
2155
1855
|
validator.check_value_type('axis[%d]' % i, value, [int], self.name)
|
|
2156
|
-
self.init_attrs[
|
|
2157
|
-
self.add_prim_attr(
|
|
1856
|
+
self.init_attrs['axis'] = self.axis
|
|
1857
|
+
self.add_prim_attr('axis', self.axis)
|
|
2158
1858
|
self.init_prim_io_names(inputs=['x', 'clip_norm'], outputs=['output'])
|
|
2159
1859
|
|
|
2160
1860
|
def infer_shape(self, x_shape, clip_norm_shape):
|
|
@@ -2247,7 +1947,8 @@ class MixedPrecisionCast(Primitive):
|
|
|
2247
1947
|
|
|
2248
1948
|
def __call__(self, dst_dtype, x):
|
|
2249
1949
|
def cast_inner(data):
|
|
2250
|
-
if isinstance(data, Tensor) and data.dtype in (mstype.float16, mstype.float32,
|
|
1950
|
+
if isinstance(data, Tensor) and data.dtype in (mstype.float16, mstype.float32,
|
|
1951
|
+
mstype.float64, mstype.bfloat16):
|
|
2251
1952
|
return self.cast(data, dst_dtype)
|
|
2252
1953
|
return data
|
|
2253
1954
|
|
|
@@ -2558,7 +2259,7 @@ class ConvertToMsTensor(Primitive):
|
|
|
2558
2259
|
"""Run in PyNative mode"""
|
|
2559
2260
|
if isinstance(x, StubTensor):
|
|
2560
2261
|
return StubTensor(stub=x.stub, tensor=x.tensor)
|
|
2561
|
-
return ops.deepcopy(x)
|
|
2262
|
+
return ops.auto_generate.deepcopy(x)
|
|
2562
2263
|
|
|
2563
2264
|
|
|
2564
2265
|
convert_to_ms_tensor = ConvertToMsTensor()
|
|
@@ -2620,28 +2321,6 @@ class IsParameter(PrimitiveWithInfer):
|
|
|
2620
2321
|
'value': isinstance(x['dtype'], mstype.RefType)}
|
|
2621
2322
|
|
|
2622
2323
|
|
|
2623
|
-
class SiLU(Primitive):
|
|
2624
|
-
r"""
|
|
2625
|
-
Computes SiLU (Sigmoid Linear Unit activation function) of input tensors element-wise.
|
|
2626
|
-
|
|
2627
|
-
Refer to :func:`mindspore.ops.silu` for more details.
|
|
2628
|
-
|
|
2629
|
-
Supported Platforms:
|
|
2630
|
-
``Ascend`` ``GPU`` ``CPU``
|
|
2631
|
-
|
|
2632
|
-
Examples:
|
|
2633
|
-
>>> x = Tensor(np.array([-1, 2, -3, 2, -1]), mindspore.float16)
|
|
2634
|
-
>>> output = ops.SiLU(x)
|
|
2635
|
-
>>> print(output)
|
|
2636
|
-
[-0.269 1.762 -0.1423 1.762 -0.269]
|
|
2637
|
-
"""
|
|
2638
|
-
|
|
2639
|
-
@prim_attr_register
|
|
2640
|
-
def __init__(self):
|
|
2641
|
-
"""Initialize SiLU"""
|
|
2642
|
-
self.init_prim_io_names(inputs=['x'], outputs=['output'])
|
|
2643
|
-
|
|
2644
|
-
|
|
2645
2324
|
class TileSize(Primitive):
|
|
2646
2325
|
r"""
|
|
2647
2326
|
Tile size for matmul
|
|
@@ -2725,6 +2404,7 @@ class CopyWithSlice(Primitive):
|
|
|
2725
2404
|
r"""
|
|
2726
2405
|
Copy data to discontinuous tensor
|
|
2727
2406
|
"""
|
|
2407
|
+
|
|
2728
2408
|
@prim_attr_register
|
|
2729
2409
|
def __init__(self):
|
|
2730
2410
|
self.add_prim_attr('side_effect_mem', True)
|
|
@@ -2774,10 +2454,10 @@ class FFN(Primitive):
|
|
|
2774
2454
|
>>> h = 1024
|
|
2775
2455
|
>>> h_f = 4 * h
|
|
2776
2456
|
>>> e = 16
|
|
2777
|
-
>>> x = Tensor(np.random.randn(
|
|
2457
|
+
>>> x = Tensor(np.random.randn(s, h).astype(np.float16))
|
|
2778
2458
|
>>> w1 = Tensor(np.random.randn(e, h, h_f).astype(np.float16))
|
|
2779
2459
|
>>> w2 = Tensor(np.random.randn(e, h_f, h).astype(np.float16))
|
|
2780
|
-
>>> expert_tokens = Tensor(np.
|
|
2460
|
+
>>> expert_tokens = Tensor(np.full(e, 8))
|
|
2781
2461
|
>>> bias1 = Tensor(np.random.randn(e, h_f).astype(np.float16))
|
|
2782
2462
|
>>> bias2 = Tensor(np.random.randn(e, h).astype(np.float16))
|
|
2783
2463
|
>>> ffn = _inner_ops.FFN("fastgelu", 1)
|
|
@@ -2789,135 +2469,99 @@ class FFN(Primitive):
|
|
|
2789
2469
|
def __init__(self, activation, inner_precise):
|
|
2790
2470
|
"""Initialize FFN."""
|
|
2791
2471
|
self.init_prim_io_names(inputs=["x", "weight1", "weight2", "expert_tokens", "bias1",
|
|
2792
|
-
"bias2", "scale", "offset", "deq_scale1", "deq_scale2"
|
|
2472
|
+
"bias2", "scale", "offset", "deq_scale1", "deq_scale2",
|
|
2473
|
+
"antiquant_scale1", "antiquant_scale2",
|
|
2474
|
+
"antiquant_offset1", "antiquant_offset2"],
|
|
2793
2475
|
outputs=["y"])
|
|
2794
2476
|
cls_name = self.name
|
|
2795
2477
|
validator.check_value_type("activation", activation, [str], cls_name)
|
|
2796
2478
|
validator.check_value_type("inner_precise", inner_precise, [int], cls_name)
|
|
2797
2479
|
|
|
2798
2480
|
|
|
2799
|
-
class
|
|
2800
|
-
|
|
2801
|
-
The
|
|
2481
|
+
class _MirrorSilentCheck(PrimitiveWithInfer):
|
|
2482
|
+
"""
|
|
2483
|
+
The operator _MirrorSilentCheck implements accuracy-sensitive detection on the tensor input in backpropagator.
|
|
2484
|
+
Call _MirrorSilentCheck in method __call__ of derived class to implement accuracy-sensitive detection.
|
|
2802
2485
|
|
|
2803
|
-
|
|
2804
|
-
|
|
2805
|
-
|
|
2806
|
-
|
|
2807
|
-
|
|
2808
|
-
|
|
2809
|
-
|
|
2810
|
-
|
|
2811
|
-
|
|
2812
|
-
|
|
2813
|
-
|
|
2814
|
-
|
|
2815
|
-
Valid_seq_len tensor of shape :math:`(batch\_size)`.
|
|
2816
|
-
batch_index (Tensor): The batch_index tensor with data type of int64.
|
|
2817
|
-
Batch_index tensor of shape :math:`(1)`. Indicate that which batch of cache tensor is going to be update.
|
|
2818
|
-
seq_len_axis (int64): The seq_len_axis indicate which axis is seq_eln, set to '1' or '2'. Default: "2".
|
|
2819
|
-
new_max_seq_len (Tensor): The new_max_seq_len tensor with data type of int64.
|
|
2820
|
-
New_max_seq_len tensor of shape :math:`(1)`.
|
|
2821
|
-
Indicate that user want to change the shape of cache tensor from
|
|
2822
|
-
:math:`(batch\_size, num_head, max\_seq\_length, hidden\_size)` to
|
|
2823
|
-
:math:
|
|
2824
|
-
`(batch\_size * max\_seq\_length / new\_max\_seq\_length, num_head, new\_max\_seq\_length, hidden\_size)`
|
|
2825
|
-
to update the cache tensor. This will not real change the shape of `cache` tensor. Not able for now.
|
|
2826
|
-
cur_max_seq_len (Tensor): The new_max_seq_len tensor with data type of int64.
|
|
2827
|
-
Cur_max_seq_len tensor of shape :math:`(1)`. Keep the current seq_len of cache tensor. Not abel for now.
|
|
2486
|
+
Inputs:
|
|
2487
|
+
- **input** (Tensor) : The tensor used for detection.
|
|
2488
|
+
Its data type must be mindspore.float16, mindspore.float32 or mindspore.bfloat16.
|
|
2489
|
+
- **pre_val** (Parameter(Tensor)) : Support parameter in accuracy-sensitive detection.
|
|
2490
|
+
Please only generated by method generate_params() of ASDBase.
|
|
2491
|
+
- **min_val** (Parameter(Tensor)) : Support parameter in accuracy-sensitive detection.
|
|
2492
|
+
Please only generated by method generate_params() of ASDBase.
|
|
2493
|
+
- **max_val** (Parameter(Tensor)) : Support parameter in accuracy-sensitive detection.
|
|
2494
|
+
Please only generated by method generate_params() of ASDBase.
|
|
2495
|
+
- **cnt** (Parameter(Tensor)) : Support parameter in accuracy-sensitive detection.
|
|
2496
|
+
Please only generated by method generate_params() of ASDBase.
|
|
2497
|
+
After each invocation of _MirrorSilentCheck, increment the value of cnt by one.
|
|
2828
2498
|
|
|
2829
2499
|
Outputs:
|
|
2830
|
-
|
|
2831
|
-
|
|
2832
|
-
Supported Platforms:
|
|
2833
|
-
``Ascend``
|
|
2834
|
-
|
|
2835
|
-
Examples:
|
|
2836
|
-
>>> from mindspore.ops.operations import _inner_ops
|
|
2837
|
-
>>> b = 4
|
|
2838
|
-
>>> h = 40
|
|
2839
|
-
>>> max_s = 1024
|
|
2840
|
-
>>> s = 1
|
|
2841
|
-
>>> d = 128
|
|
2842
|
-
>>> cache = Tensor(np.random.randn(b, h, max_s, d).astype(np.float16))
|
|
2843
|
-
>>> update = Tensor(np.random.randn(b, h, s, d).astype(np.float16))
|
|
2844
|
-
>>> valid_seq_len = Tensor(np.random.randn(b).astype(np.int64))
|
|
2845
|
-
>>> batch_index = Tensor(np.random.randn(1).astype(np.int64))
|
|
2846
|
-
>>> new_max_seq_len = Tensor(np.random.randn(1).astype(np.int64))
|
|
2847
|
-
>>> cur_max_seq_len = Tensor(np.random.randn(1).astype(np.int64))
|
|
2848
|
-
>>> decoder_kv_cache = _inner_ops.DecoderKVCache()
|
|
2849
|
-
>>> output = decoder_kv_cache(cache, update, valid_seq_len, batch_index, 2, new_max_seq_len, cur_max_seq_len)
|
|
2850
|
-
>>> print(cache)
|
|
2500
|
+
- **output** (Tensor) - Same shape, type and value as `input`.
|
|
2851
2501
|
"""
|
|
2852
2502
|
@prim_attr_register
|
|
2853
|
-
def __init__(self):
|
|
2854
|
-
|
|
2855
|
-
self.
|
|
2856
|
-
|
|
2857
|
-
|
|
2503
|
+
def __init__(self, min_steps=8):
|
|
2504
|
+
upper_thresh, sigma_thresh = self.get_thresh()
|
|
2505
|
+
self.min_steps = min_steps
|
|
2506
|
+
self.thresh_l1 = upper_thresh[0]
|
|
2507
|
+
self.coeff_l1 = sigma_thresh[0]
|
|
2508
|
+
self.thresh_l2 = upper_thresh[1]
|
|
2509
|
+
self.coeff_l2 = sigma_thresh[1]
|
|
2858
2510
|
self.add_prim_attr('side_effect_mem', True)
|
|
2859
2511
|
|
|
2512
|
+
def parse_thresh(self, env_var_name, default_value, min_value):
|
|
2513
|
+
env_var = os.environ.get(env_var_name, default=default_value)
|
|
2514
|
+
thresh = [value.strip() for value in env_var.split(",")]
|
|
2515
|
+
if len(thresh) != 2 or not all(value.isdigit() for value in thresh):
|
|
2516
|
+
thresh = default_value.split(",")
|
|
2517
|
+
thresh = [float(max(int(value), min_value)) for value in thresh]
|
|
2518
|
+
if thresh[0] <= thresh[1]:
|
|
2519
|
+
thresh = [float(value) for value in default_value.split(",")]
|
|
2860
2520
|
|
|
2861
|
-
|
|
2862
|
-
r"""
|
|
2863
|
-
The PromptKVCache is used for prefill the KVCache of transformer network.
|
|
2521
|
+
return thresh
|
|
2864
2522
|
|
|
2865
|
-
|
|
2866
|
-
|
|
2867
|
-
|
|
2868
|
-
|
|
2869
|
-
When seq_len_axis is 1, cache tensor of shape
|
|
2870
|
-
:math:`(batch\_size, max\_seq\_length, num_head, hidden\_size)`.
|
|
2871
|
-
update (Tensor]): The tensor which is used to update the cache tensor. Same data type as cache tensor.
|
|
2872
|
-
When seq_len_axis is 2, update tensor of shape
|
|
2873
|
-
:math:`(batch\_size, num_head, update\_seq\_length, hidden\_size)`.
|
|
2874
|
-
When seq_len_axis is 1, update tensor of shape
|
|
2875
|
-
:math:`(batch\_size, update\_seq\_length, num_head, hidden\_size)`.
|
|
2876
|
-
valid_seq_len (Tensor): The valid_seq_len tensor with data type of int64.
|
|
2877
|
-
Valid_seq_len tensor of shape :math:`(batch\_size)`.
|
|
2878
|
-
batch_index (Tensor): The batch_index tensor with data type of int64.
|
|
2879
|
-
Batch_index tensor of shape :math:`(1)`. Indicate that which batch of cache tensor is going to be update.
|
|
2880
|
-
seq_len_axis (int64): The seq_len_axis indicate which axis is seq_eln, set to '1' or '2'. Default: "2".
|
|
2881
|
-
new_max_seq_len (Tensor): The new_max_seq_len tensor with data type of int64.
|
|
2882
|
-
New_max_seq_len tensor of shape :math:`(1)`.
|
|
2883
|
-
Indicate that user want to change the shape of cache tensor from
|
|
2884
|
-
:math:`(batch\_size, num_head, max\_seq\_length, hidden\_size)` to
|
|
2885
|
-
:math:
|
|
2886
|
-
`(batch\_size * max\_seq\_length / new\_max\_seq\_length, num_head, new\_max\_seq\_length, hidden\_size)`
|
|
2887
|
-
to update the cache tensor. This will not real change the shape of `cache` tensor. Not able for now.
|
|
2888
|
-
cur_max_seq_len (Tensor): The new_max_seq_len tensor with data type of int64.
|
|
2889
|
-
Cur_max_seq_len tensor of shape :math:`(1)`. Keep the current seq_len of cache tensor. Not abel for now.
|
|
2890
|
-
align_mode (int64): indicate which axis is seq_eln, 0 is 'right', 1 is 'left'. Default: 0.
|
|
2523
|
+
def get_thresh(self):
|
|
2524
|
+
upper_thresh = self.parse_thresh("NPU_ASD_UPPER_THRESH", "1000000,10000", 3)
|
|
2525
|
+
sigma_thresh = self.parse_thresh("NPU_ASD_SIGMA_THRESH", "100000,5000", 3)
|
|
2526
|
+
return upper_thresh, sigma_thresh
|
|
2891
2527
|
|
|
2892
|
-
|
|
2893
|
-
|
|
2528
|
+
def infer_shape(self, x_shape, pre_shape, min_shape, max_shape, n_step, loss_scale_shape):
|
|
2529
|
+
return x_shape
|
|
2894
2530
|
|
|
2895
|
-
|
|
2896
|
-
|
|
2531
|
+
def infer_dtype(self, x_dtype, pre_dtype, min_dtype, max_dtype, n_dtype, loss_scale_dtype):
|
|
2532
|
+
return x_dtype
|
|
2897
2533
|
|
|
2898
|
-
|
|
2899
|
-
|
|
2900
|
-
|
|
2901
|
-
|
|
2902
|
-
>>> h = 40
|
|
2903
|
-
>>> max_s = 1024
|
|
2904
|
-
>>> s = 256
|
|
2905
|
-
>>> d = 128
|
|
2906
|
-
>>> cache = Tensor(np.random.randn(b, h, max_s, d).astype(np.float16))
|
|
2907
|
-
>>> update = Tensor(np.random.randn(b, h, s, d).astype(np.float16))
|
|
2908
|
-
>>> valid_seq_len = Tensor(np.random.randn(b).astype(np.int64))
|
|
2909
|
-
>>> batch_index = Tensor(np.random.randn(1).astype(np.int64))
|
|
2910
|
-
>>> new_max_seq_len = Tensor(np.random.randn(1).astype(np.int64))
|
|
2911
|
-
>>> cur_max_seq_len = Tensor(np.random.randn(1).astype(np.int64))
|
|
2912
|
-
>>> prompt_kv_cache = _inner_ops.PromptKVCache(0)
|
|
2913
|
-
>>> output = prompt_kv_cache(cache, update, valid_seq_len, batch_index, 2, new_max_seq_len, cur_max_seq_len)
|
|
2914
|
-
>>> print(cache)
|
|
2534
|
+
|
|
2535
|
+
class _VirtualConverterEnd(PrimitiveWithInfer):
|
|
2536
|
+
"""
|
|
2537
|
+
Auto parallel virtual operator.
|
|
2915
2538
|
"""
|
|
2539
|
+
|
|
2916
2540
|
@prim_attr_register
|
|
2917
|
-
def __init__(self,
|
|
2918
|
-
"""Initialize
|
|
2919
|
-
self.
|
|
2920
|
-
|
|
2921
|
-
|
|
2922
|
-
self.
|
|
2923
|
-
|
|
2541
|
+
def __init__(self, input_nums):
|
|
2542
|
+
"""Initialize _VirtualConverterEnd."""
|
|
2543
|
+
self.input_nums = input_nums
|
|
2544
|
+
|
|
2545
|
+
def infer_shape(self, *args):
|
|
2546
|
+
return (args[0][0] * self.input_nums,) + tuple(args[0][1:])
|
|
2547
|
+
|
|
2548
|
+
def infer_dtype(self, *args):
|
|
2549
|
+
return args[0]
|
|
2550
|
+
|
|
2551
|
+
|
|
2552
|
+
class _VirtualConverterBegin(PrimitiveWithInfer):
|
|
2553
|
+
"""
|
|
2554
|
+
Auto parallel virtual operator.
|
|
2555
|
+
"""
|
|
2556
|
+
|
|
2557
|
+
@prim_attr_register
|
|
2558
|
+
def __init__(self, output_nums):
|
|
2559
|
+
"""Initialize _VirtualConverterBegin."""
|
|
2560
|
+
self.output_nums = output_nums
|
|
2561
|
+
|
|
2562
|
+
def infer_shape(self, arg):
|
|
2563
|
+
new_arg = (arg[0] / self.output_nums,) + tuple(arg[1:])
|
|
2564
|
+
return (new_arg,) * self.output_nums
|
|
2565
|
+
|
|
2566
|
+
def infer_dtype(self, arg):
|
|
2567
|
+
return (arg,) * self.output_nums
|