mindspore 2.2.11__cp39-cp39-win_amd64.whl → 2.3.0__cp39-cp39-win_amd64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of mindspore might be problematic. Click here for more details.
- mindspore/.commit_id +1 -1
- mindspore/__init__.py +7 -5
- mindspore/_c_dataengine.cp39-win_amd64.pyd +0 -0
- mindspore/_c_expression.cp39-win_amd64.pyd +0 -0
- mindspore/_c_mindrecord.cp39-win_amd64.pyd +0 -0
- mindspore/_checkparam.py +76 -18
- mindspore/_extends/builtin_operations.py +2 -1
- mindspore/_extends/graph_kernel/model/graph_parallel.py +16 -6
- mindspore/_extends/parallel_compile/akg_compiler/akg_process.py +3 -16
- mindspore/_extends/parallel_compile/akg_compiler/build_tbe_kernel.py +16 -4
- mindspore/_extends/parallel_compile/akg_compiler/compiler.py +1 -0
- mindspore/_extends/parallel_compile/akg_compiler/gen_custom_op_files.py +96 -0
- mindspore/_extends/parallel_compile/akg_compiler/tbe_topi.py +2 -1
- mindspore/_extends/parallel_compile/akg_compiler/util.py +5 -2
- mindspore/_extends/parse/__init__.py +18 -14
- mindspore/_extends/parse/compile_config.py +258 -0
- mindspore/_extends/parse/namespace.py +2 -2
- mindspore/_extends/parse/parser.py +174 -62
- mindspore/_extends/parse/resources.py +45 -14
- mindspore/_extends/parse/standard_method.py +142 -240
- mindspore/{ops/_op_impl/tbe/atomic_addr_clean.py → _extends/pijit/__init__.py} +6 -16
- mindspore/_extends/pijit/pijit_func_white_list.py +343 -0
- mindspore/_extends/remote/kernel_build_server.py +2 -0
- mindspore/_profiler.py +30 -0
- mindspore/amp.py +51 -24
- mindspore/avcodec-59.dll +0 -0
- mindspore/avdevice-59.dll +0 -0
- mindspore/avfilter-8.dll +0 -0
- mindspore/avformat-59.dll +0 -0
- mindspore/avutil-57.dll +0 -0
- mindspore/boost/adasum.py +1 -1
- mindspore/boost/base.py +1 -1
- mindspore/boost/boost_cell_wrapper.py +2 -2
- mindspore/boost/grad_freeze.py +2 -2
- mindspore/boost/group_loss_scale_manager.py +1 -1
- mindspore/boost/less_batch_normalization.py +9 -6
- mindspore/common/__init__.py +15 -4
- mindspore/common/_jit_fallback_utils.py +2 -3
- mindspore/common/_register_for_adapter.py +7 -0
- mindspore/common/_register_for_recompute.py +48 -0
- mindspore/common/_register_for_tensor.py +8 -9
- mindspore/common/_stub_tensor.py +7 -1
- mindspore/common/_utils.py +5 -17
- mindspore/common/api.py +411 -106
- mindspore/common/auto_dynamic_shape.py +27 -14
- mindspore/common/dtype.py +17 -10
- mindspore/common/dump.py +6 -8
- mindspore/common/file_system.py +48 -0
- mindspore/common/generator.py +260 -0
- mindspore/common/hook_handle.py +51 -4
- mindspore/common/initializer.py +1 -1
- mindspore/common/jit_config.py +34 -14
- mindspore/common/lazy_inline.py +72 -19
- mindspore/common/mindir_util.py +12 -2
- mindspore/common/mutable.py +79 -14
- mindspore/common/no_inline.py +54 -0
- mindspore/common/np_dtype.py +25 -0
- mindspore/common/parameter.py +30 -11
- mindspore/common/recompute.py +262 -0
- mindspore/common/seed.py +9 -9
- mindspore/common/sparse_tensor.py +272 -24
- mindspore/common/symbol.py +122 -0
- mindspore/common/tensor.py +468 -496
- mindspore/communication/__init__.py +6 -11
- mindspore/communication/_comm_helper.py +5 -0
- mindspore/communication/comm_func.py +1140 -0
- mindspore/communication/management.py +118 -102
- mindspore/config/op_info.config +22 -54
- mindspore/context.py +378 -65
- mindspore/dataset/__init__.py +5 -5
- mindspore/dataset/audio/__init__.py +6 -6
- mindspore/dataset/audio/transforms.py +711 -158
- mindspore/dataset/callback/ds_callback.py +2 -2
- mindspore/dataset/engine/cache_client.py +2 -2
- mindspore/dataset/engine/datasets.py +163 -83
- mindspore/dataset/engine/datasets_audio.py +14 -14
- mindspore/dataset/engine/datasets_standard_format.py +33 -3
- mindspore/dataset/engine/datasets_text.py +38 -38
- mindspore/dataset/engine/datasets_user_defined.py +78 -59
- mindspore/dataset/engine/datasets_vision.py +77 -73
- mindspore/dataset/engine/offload.py +5 -7
- mindspore/dataset/engine/queue.py +56 -38
- mindspore/dataset/engine/validators.py +11 -5
- mindspore/dataset/text/__init__.py +3 -3
- mindspore/dataset/text/transforms.py +408 -121
- mindspore/dataset/text/utils.py +9 -9
- mindspore/dataset/transforms/__init__.py +1 -1
- mindspore/dataset/transforms/transforms.py +261 -76
- mindspore/dataset/utils/browse_dataset.py +9 -9
- mindspore/dataset/vision/__init__.py +8 -8
- mindspore/dataset/vision/c_transforms.py +10 -10
- mindspore/dataset/vision/py_transforms_util.py +3 -3
- mindspore/dataset/vision/transforms.py +2844 -549
- mindspore/dataset/vision/utils.py +161 -10
- mindspore/dataset/vision/validators.py +14 -2
- mindspore/dnnl.dll +0 -0
- mindspore/experimental/optim/__init__.py +12 -2
- mindspore/experimental/optim/adadelta.py +161 -0
- mindspore/experimental/optim/adagrad.py +168 -0
- mindspore/experimental/optim/adam.py +35 -34
- mindspore/experimental/optim/adamax.py +170 -0
- mindspore/experimental/optim/adamw.py +40 -16
- mindspore/experimental/optim/asgd.py +153 -0
- mindspore/experimental/optim/lr_scheduler.py +71 -127
- mindspore/experimental/optim/nadam.py +157 -0
- mindspore/experimental/optim/optimizer.py +15 -8
- mindspore/experimental/optim/radam.py +194 -0
- mindspore/experimental/optim/rmsprop.py +154 -0
- mindspore/experimental/optim/rprop.py +164 -0
- mindspore/experimental/optim/sgd.py +28 -19
- mindspore/hal/__init__.py +40 -0
- mindspore/hal/_ascend.py +57 -0
- mindspore/hal/_base.py +57 -0
- mindspore/hal/_cpu.py +56 -0
- mindspore/hal/_gpu.py +57 -0
- mindspore/hal/device.py +356 -0
- mindspore/hal/event.py +179 -0
- mindspore/hal/memory.py +326 -0
- mindspore/hal/stream.py +339 -0
- mindspore/include/api/data_type.h +2 -2
- mindspore/include/api/dual_abi_helper.h +16 -3
- mindspore/include/api/model.h +4 -3
- mindspore/include/api/status.h +14 -0
- mindspore/include/c_api/model_c.h +173 -0
- mindspore/include/c_api/ms/base/types.h +1 -0
- mindspore/include/c_api/types_c.h +19 -0
- mindspore/include/dataset/execute.h +1 -3
- mindspore/include/dataset/vision.h +54 -2
- mindspore/jpeg62.dll +0 -0
- mindspore/log.py +2 -2
- mindspore/mindrecord/__init__.py +5 -1
- mindspore/mindrecord/config.py +809 -0
- mindspore/mindrecord/filereader.py +25 -0
- mindspore/mindrecord/filewriter.py +76 -58
- mindspore/mindrecord/mindpage.py +40 -6
- mindspore/mindrecord/shardutils.py +3 -2
- mindspore/mindrecord/shardwriter.py +7 -0
- mindspore/mindrecord/tools/cifar100_to_mr.py +53 -66
- mindspore/mindrecord/tools/cifar10_to_mr.py +48 -63
- mindspore/mindrecord/tools/csv_to_mr.py +7 -17
- mindspore/mindrecord/tools/imagenet_to_mr.py +3 -8
- mindspore/mindrecord/tools/mnist_to_mr.py +11 -21
- mindspore/mindrecord/tools/tfrecord_to_mr.py +2 -10
- mindspore/mindspore_backend.dll +0 -0
- mindspore/mindspore_common.dll +0 -0
- mindspore/mindspore_core.dll +0 -0
- mindspore/mindspore_glog.dll +0 -0
- mindspore/mindspore_np_dtype.dll +0 -0
- mindspore/mindspore_shared_lib.dll +0 -0
- mindspore/mint/__init__.py +1137 -0
- mindspore/{rewrite/ast_transformers → mint/linalg}/__init__.py +9 -4
- mindspore/mint/nn/__init__.py +512 -0
- mindspore/mint/nn/functional.py +573 -0
- mindspore/mint/optim/__init__.py +24 -0
- mindspore/mint/optim/adamw.py +185 -0
- mindspore/multiprocessing/__init__.py +72 -0
- mindspore/nn/__init__.py +1 -0
- mindspore/nn/cell.py +213 -257
- mindspore/nn/dynamic_lr.py +2 -2
- mindspore/nn/extend/__init__.py +29 -0
- mindspore/nn/extend/basic.py +140 -0
- mindspore/nn/extend/embedding.py +143 -0
- mindspore/{rewrite/ast_creator_register.py → nn/extend/layer/__init__.py} +9 -19
- mindspore/nn/extend/layer/normalization.py +109 -0
- mindspore/nn/extend/pooling.py +117 -0
- mindspore/nn/layer/activation.py +84 -94
- mindspore/nn/layer/basic.py +177 -82
- mindspore/nn/layer/channel_shuffle.py +3 -16
- mindspore/nn/layer/container.py +3 -3
- mindspore/nn/layer/conv.py +75 -66
- mindspore/nn/layer/embedding.py +103 -45
- mindspore/nn/layer/embedding_service.py +531 -0
- mindspore/nn/layer/embedding_service_layer.py +393 -0
- mindspore/nn/layer/image.py +4 -7
- mindspore/nn/layer/math.py +1 -1
- mindspore/nn/layer/normalization.py +52 -66
- mindspore/nn/layer/padding.py +30 -39
- mindspore/nn/layer/pooling.py +18 -9
- mindspore/nn/layer/rnn_cells.py +6 -16
- mindspore/nn/layer/rnns.py +6 -5
- mindspore/nn/layer/thor_layer.py +1 -2
- mindspore/nn/layer/timedistributed.py +1 -1
- mindspore/nn/layer/transformer.py +52 -50
- mindspore/nn/learning_rate_schedule.py +6 -5
- mindspore/nn/loss/loss.py +63 -84
- mindspore/nn/optim/ada_grad.py +6 -4
- mindspore/nn/optim/adadelta.py +3 -1
- mindspore/nn/optim/adafactor.py +1 -1
- mindspore/nn/optim/adam.py +102 -181
- mindspore/nn/optim/adamax.py +4 -2
- mindspore/nn/optim/adasum.py +3 -3
- mindspore/nn/optim/asgd.py +4 -2
- mindspore/nn/optim/ftrl.py +31 -61
- mindspore/nn/optim/lamb.py +5 -3
- mindspore/nn/optim/lars.py +2 -2
- mindspore/nn/optim/lazyadam.py +6 -4
- mindspore/nn/optim/momentum.py +13 -25
- mindspore/nn/optim/optimizer.py +6 -3
- mindspore/nn/optim/proximal_ada_grad.py +4 -2
- mindspore/nn/optim/rmsprop.py +9 -3
- mindspore/nn/optim/rprop.py +4 -2
- mindspore/nn/optim/sgd.py +7 -4
- mindspore/nn/optim/thor.py +2 -2
- mindspore/nn/probability/distribution/_utils/custom_ops.py +2 -2
- mindspore/nn/probability/distribution/beta.py +2 -2
- mindspore/nn/probability/distribution/categorical.py +4 -6
- mindspore/nn/probability/distribution/cauchy.py +2 -2
- mindspore/nn/probability/distribution/exponential.py +2 -2
- mindspore/nn/probability/distribution/geometric.py +1 -1
- mindspore/nn/probability/distribution/gumbel.py +2 -2
- mindspore/nn/probability/distribution/logistic.py +1 -1
- mindspore/nn/probability/distribution/poisson.py +2 -2
- mindspore/nn/probability/distribution/uniform.py +2 -2
- mindspore/nn/reinforcement/_tensors_queue.py +13 -1
- mindspore/nn/wrap/__init__.py +2 -1
- mindspore/nn/wrap/cell_wrapper.py +58 -13
- mindspore/nn/wrap/grad_reducer.py +148 -8
- mindspore/nn/wrap/loss_scale.py +32 -9
- mindspore/numpy/__init__.py +2 -0
- mindspore/numpy/array_creations.py +2 -0
- mindspore/numpy/array_ops.py +6 -6
- mindspore/numpy/dtypes.py +3 -3
- mindspore/numpy/fft.py +431 -0
- mindspore/numpy/math_ops.py +61 -67
- mindspore/numpy/utils.py +3 -0
- mindspore/opencv_core452.dll +0 -0
- mindspore/opencv_imgcodecs452.dll +0 -0
- mindspore/opencv_imgproc452.dll +0 -0
- mindspore/ops/__init__.py +8 -4
- mindspore/ops/_grad_experimental/grad_array_ops.py +4 -160
- mindspore/ops/_grad_experimental/grad_comm_ops.py +93 -36
- mindspore/ops/_grad_experimental/grad_inner_ops.py +8 -0
- mindspore/ops/_grad_experimental/grad_math_ops.py +92 -287
- mindspore/ops/_grad_experimental/grad_nn_ops.py +0 -53
- mindspore/ops/_grad_experimental/grad_quant_ops.py +3 -3
- mindspore/ops/_grad_experimental/grad_sparse.py +1 -1
- mindspore/ops/_grad_experimental/grad_sparse_ops.py +3 -3
- mindspore/ops/_op_impl/__init__.py +0 -1
- mindspore/ops/_op_impl/aicpu/__init__.py +1 -0
- mindspore/ops/_op_impl/aicpu/gamma.py +2 -0
- mindspore/ops/_op_impl/{cpu/concat.py → aicpu/generate_eod_mask.py} +16 -17
- mindspore/ops/_op_impl/aicpu/log_uniform_candidate_sampler.py +1 -3
- mindspore/ops/_op_impl/aicpu/poisson.py +2 -0
- mindspore/ops/_op_impl/cpu/__init__.py +1 -3
- mindspore/ops/_op_impl/cpu/adam.py +2 -2
- mindspore/ops/_op_impl/cpu/adam_weight_decay.py +3 -2
- mindspore/ops/_op_impl/cpu/maximum_grad.py +16 -14
- mindspore/ops/_op_impl/cpu/minimum_grad.py +8 -0
- mindspore/ops/_vmap/vmap_array_ops.py +164 -101
- mindspore/ops/_vmap/vmap_base.py +8 -1
- mindspore/ops/_vmap/vmap_grad_math_ops.py +95 -9
- mindspore/ops/_vmap/vmap_grad_nn_ops.py +143 -58
- mindspore/ops/_vmap/vmap_image_ops.py +70 -13
- mindspore/ops/_vmap/vmap_math_ops.py +130 -58
- mindspore/ops/_vmap/vmap_nn_ops.py +249 -115
- mindspore/ops/_vmap/vmap_other_ops.py +1 -1
- mindspore/ops/auto_generate/__init__.py +31 -0
- mindspore/ops/auto_generate/cpp_create_prim_instance_helper.py +231 -0
- mindspore/ops/auto_generate/gen_arg_dtype_cast.py +250 -0
- mindspore/ops/auto_generate/gen_arg_handler.py +197 -0
- mindspore/ops/auto_generate/gen_extend_func.py +980 -0
- mindspore/ops/auto_generate/gen_ops_def.py +6443 -0
- mindspore/ops/auto_generate/gen_ops_prim.py +13167 -0
- mindspore/ops/auto_generate/pyboost_inner_prim.py +429 -0
- mindspore/ops/composite/__init__.py +5 -2
- mindspore/ops/composite/base.py +121 -23
- mindspore/ops/composite/math_ops.py +10 -49
- mindspore/ops/composite/multitype_ops/_compile_utils.py +191 -618
- mindspore/ops/composite/multitype_ops/_constexpr_utils.py +25 -134
- mindspore/ops/composite/multitype_ops/add_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/bitwise_and_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/bitwise_or_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/bitwise_xor_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/div_impl.py +8 -0
- mindspore/ops/composite/multitype_ops/equal_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/floordiv_impl.py +8 -0
- mindspore/ops/composite/multitype_ops/getitem_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/greater_equal_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/greater_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/in_impl.py +8 -2
- mindspore/ops/composite/multitype_ops/left_shift_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/less_equal_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/less_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/logic_not_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/logical_and_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/logical_or_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/mod_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/mul_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/negative_impl.py +9 -3
- mindspore/ops/composite/multitype_ops/not_equal_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/not_in_impl.py +6 -1
- mindspore/ops/composite/multitype_ops/ones_like_impl.py +2 -2
- mindspore/ops/composite/multitype_ops/pow_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/right_shift_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/setitem_impl.py +32 -21
- mindspore/ops/composite/multitype_ops/sub_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/zeros_like_impl.py +6 -3
- mindspore/ops/deprecated.py +14 -3
- mindspore/ops/extend/__init__.py +53 -0
- mindspore/ops/extend/array_func.py +218 -0
- mindspore/ops/extend/math_func.py +76 -0
- mindspore/ops/extend/nn_func.py +308 -0
- mindspore/ops/function/__init__.py +31 -11
- mindspore/ops/function/array_func.py +848 -1736
- mindspore/ops/function/clip_func.py +19 -31
- mindspore/ops/function/debug_func.py +2 -5
- mindspore/ops/function/fft_func.py +31 -0
- mindspore/ops/function/grad/grad_func.py +27 -20
- mindspore/ops/function/image_func.py +27 -21
- mindspore/ops/function/linalg_func.py +30 -53
- mindspore/ops/function/math_func.py +916 -2791
- mindspore/ops/function/nn_func.py +1445 -889
- mindspore/ops/function/other_func.py +6 -7
- mindspore/ops/function/parameter_func.py +6 -92
- mindspore/ops/function/random_func.py +254 -108
- mindspore/ops/function/reshard_func.py +102 -0
- mindspore/ops/function/sparse_func.py +4 -4
- mindspore/ops/function/sparse_unary_func.py +11 -18
- mindspore/ops/function/spectral_func.py +1 -1
- mindspore/ops/function/vmap_func.py +15 -14
- mindspore/ops/functional.py +342 -343
- mindspore/ops/op_info_register.py +16 -43
- mindspore/ops/operations/__init__.py +32 -23
- mindspore/ops/operations/_embedding_cache_ops.py +1 -1
- mindspore/ops/operations/_grad_ops.py +21 -853
- mindspore/ops/operations/_infer_ops.py +19 -0
- mindspore/ops/operations/_inner_ops.py +155 -511
- mindspore/ops/operations/_quant_ops.py +4 -4
- mindspore/ops/operations/_rl_inner_ops.py +3 -3
- mindspore/ops/operations/_scalar_ops.py +5 -480
- mindspore/ops/operations/_sequence_ops.py +6 -36
- mindspore/ops/operations/_tensor_array.py +8 -8
- mindspore/ops/operations/array_ops.py +112 -2698
- mindspore/ops/operations/comm_ops.py +801 -118
- mindspore/ops/operations/custom_ops.py +62 -121
- mindspore/ops/operations/debug_ops.py +105 -36
- mindspore/ops/operations/image_ops.py +3 -219
- mindspore/ops/operations/inner_ops.py +54 -40
- mindspore/ops/operations/linalg_ops.py +1 -49
- mindspore/ops/operations/manually_defined/__init__.py +24 -0
- mindspore/ops/operations/manually_defined/_inner.py +61 -0
- mindspore/ops/operations/manually_defined/ops_def.py +2016 -0
- mindspore/ops/operations/math_ops.py +621 -4654
- mindspore/ops/operations/nn_ops.py +316 -2226
- mindspore/ops/operations/other_ops.py +53 -45
- mindspore/ops/operations/random_ops.py +4 -51
- mindspore/ops/operations/reshard_ops.py +53 -0
- mindspore/ops/operations/sparse_ops.py +8 -8
- mindspore/ops/primitive.py +204 -103
- mindspore/ops/silent_check.py +162 -0
- mindspore/ops_generate/__init__.py +27 -0
- mindspore/ops_generate/arg_dtype_cast.py +250 -0
- mindspore/ops_generate/arg_handler.py +197 -0
- mindspore/ops_generate/gen_aclnn_implement.py +263 -0
- mindspore/ops_generate/gen_ops.py +1084 -0
- mindspore/ops_generate/gen_ops_inner_prim.py +131 -0
- mindspore/ops_generate/gen_pyboost_func.py +968 -0
- mindspore/ops_generate/gen_utils.py +209 -0
- mindspore/ops_generate/op_proto.py +138 -0
- mindspore/ops_generate/pyboost_utils.py +354 -0
- mindspore/ops_generate/template.py +239 -0
- mindspore/parallel/__init__.py +7 -4
- mindspore/parallel/_auto_parallel_context.py +155 -6
- mindspore/parallel/_cell_wrapper.py +16 -9
- mindspore/parallel/_cost_model_context.py +1 -1
- mindspore/parallel/_dp_allreduce_fusion.py +159 -159
- mindspore/parallel/_parallel_serialization.py +62 -14
- mindspore/parallel/_ps_context.py +1 -1
- mindspore/parallel/_recovery_context.py +1 -1
- mindspore/parallel/_tensor.py +18 -9
- mindspore/parallel/_transformer/__init__.py +1 -1
- mindspore/parallel/_transformer/layers.py +1 -1
- mindspore/parallel/_transformer/loss.py +1 -1
- mindspore/parallel/_transformer/moe.py +1 -1
- mindspore/parallel/_transformer/op_parallel_config.py +1 -1
- mindspore/parallel/_transformer/transformer.py +10 -10
- mindspore/parallel/_utils.py +161 -6
- mindspore/parallel/algo_parameter_config.py +6 -8
- mindspore/parallel/checkpoint_transform.py +369 -64
- mindspore/parallel/cluster/__init__.py +15 -0
- mindspore/parallel/cluster/process_entity/__init__.py +18 -0
- mindspore/parallel/cluster/process_entity/_api.py +344 -0
- mindspore/parallel/cluster/process_entity/_utils.py +126 -0
- mindspore/parallel/cluster/run.py +136 -0
- mindspore/parallel/mpi/__init__.py +1 -1
- mindspore/parallel/mpi/_mpi_config.py +1 -1
- mindspore/parallel/parameter_broadcast.py +152 -0
- mindspore/parallel/shard.py +128 -17
- mindspore/profiler/__init__.py +3 -2
- mindspore/profiler/common/process_pool.py +41 -0
- mindspore/profiler/common/singleton.py +28 -0
- mindspore/profiler/common/util.py +125 -0
- mindspore/profiler/envprofiling.py +2 -2
- mindspore/{_extends/parallel_compile/tbe_compiler → profiler/parser/ascend_analysis}/__init__.py +1 -1
- mindspore/profiler/parser/ascend_analysis/constant.py +53 -0
- mindspore/profiler/parser/ascend_analysis/file_manager.py +159 -0
- mindspore/profiler/parser/ascend_analysis/function_event.py +161 -0
- mindspore/profiler/parser/ascend_analysis/fwk_cann_parser.py +131 -0
- mindspore/profiler/parser/ascend_analysis/fwk_file_parser.py +85 -0
- mindspore/profiler/parser/ascend_analysis/msprof_timeline_parser.py +57 -0
- mindspore/profiler/parser/ascend_analysis/profiler_info_parser.py +116 -0
- mindspore/profiler/parser/ascend_analysis/tlv_decoder.py +86 -0
- mindspore/profiler/parser/ascend_analysis/trace_event_manager.py +68 -0
- mindspore/profiler/parser/ascend_cluster_generator.py +116 -0
- mindspore/profiler/parser/ascend_communicate_generator.py +314 -0
- mindspore/profiler/parser/ascend_flops_generator.py +27 -5
- mindspore/profiler/parser/ascend_fpbp_generator.py +8 -2
- mindspore/profiler/parser/ascend_hccl_generator.py +31 -280
- mindspore/profiler/parser/ascend_integrate_generator.py +42 -0
- mindspore/profiler/parser/ascend_memory_generator.py +185 -0
- mindspore/profiler/parser/ascend_msprof_exporter.py +151 -126
- mindspore/profiler/parser/ascend_msprof_generator.py +75 -274
- mindspore/profiler/parser/ascend_op_generator.py +94 -36
- mindspore/profiler/parser/ascend_timeline_generator.py +297 -131
- mindspore/profiler/parser/base_timeline_generator.py +17 -3
- mindspore/profiler/parser/cpu_gpu_timeline_generator.py +2 -1
- mindspore/profiler/parser/framework_parser.py +11 -4
- mindspore/profiler/parser/integrator.py +3 -1
- mindspore/profiler/parser/memory_usage_parser.py +8 -2
- mindspore/profiler/parser/minddata_analyzer.py +8 -2
- mindspore/profiler/parser/minddata_parser.py +73 -4
- mindspore/profiler/parser/msadvisor_analyzer.py +5 -3
- mindspore/profiler/parser/msadvisor_parser.py +10 -4
- mindspore/profiler/parser/profiler_info.py +16 -1
- mindspore/profiler/profiling.py +522 -195
- mindspore/rewrite/__init__.py +2 -13
- mindspore/rewrite/api/node.py +123 -37
- mindspore/rewrite/api/pattern_engine.py +2 -3
- mindspore/rewrite/api/scoped_value.py +16 -15
- mindspore/rewrite/api/symbol_tree.py +46 -30
- mindspore/rewrite/ast_helpers/__init__.py +3 -6
- mindspore/rewrite/ast_helpers/ast_converter.py +143 -0
- mindspore/rewrite/ast_helpers/ast_finder.py +48 -0
- mindspore/rewrite/ast_helpers/ast_flattener.py +268 -0
- mindspore/rewrite/ast_helpers/ast_modifier.py +160 -92
- mindspore/rewrite/common/__init__.py +1 -2
- mindspore/rewrite/common/config.py +24 -0
- mindspore/rewrite/common/{rewrite_elog.py → error_log.py} +39 -39
- mindspore/rewrite/{namer.py → common/namer.py} +63 -18
- mindspore/rewrite/common/namespace.py +118 -0
- mindspore/rewrite/node/__init__.py +5 -5
- mindspore/rewrite/node/call_function.py +23 -7
- mindspore/rewrite/node/cell_container.py +7 -3
- mindspore/rewrite/node/control_flow.py +53 -28
- mindspore/rewrite/node/node.py +212 -196
- mindspore/rewrite/node/node_manager.py +51 -22
- mindspore/rewrite/node/node_topological_manager.py +3 -23
- mindspore/rewrite/parsers/__init__.py +12 -0
- mindspore/rewrite/parsers/arguments_parser.py +8 -9
- mindspore/rewrite/parsers/assign_parser.py +637 -413
- mindspore/rewrite/parsers/attribute_parser.py +3 -4
- mindspore/rewrite/parsers/class_def_parser.py +115 -148
- mindspore/rewrite/parsers/constant_parser.py +5 -5
- mindspore/rewrite/parsers/container_parser.py +4 -6
- mindspore/rewrite/parsers/expr_parser.py +55 -0
- mindspore/rewrite/parsers/for_parser.py +31 -98
- mindspore/rewrite/parsers/function_def_parser.py +13 -5
- mindspore/rewrite/parsers/if_parser.py +28 -10
- mindspore/rewrite/parsers/module_parser.py +8 -182
- mindspore/rewrite/parsers/parser.py +1 -5
- mindspore/rewrite/parsers/parser_register.py +1 -1
- mindspore/rewrite/parsers/return_parser.py +5 -10
- mindspore/rewrite/parsers/while_parser.py +59 -0
- mindspore/rewrite/sparsify/utils.py +1 -1
- mindspore/rewrite/symbol_tree/__init__.py +20 -0
- mindspore/rewrite/{symbol_tree.py → symbol_tree/symbol_tree.py} +704 -185
- mindspore/rewrite/{symbol_tree_builder.py → symbol_tree/symbol_tree_builder.py} +8 -8
- mindspore/rewrite/{symbol_tree_dumper.py → symbol_tree/symbol_tree_dumper.py} +4 -4
- mindspore/run_check/_check_version.py +6 -14
- mindspore/run_check/run_check.py +1 -1
- mindspore/safeguard/rewrite_obfuscation.py +9 -19
- mindspore/swresample-4.dll +0 -0
- mindspore/swscale-6.dll +0 -0
- mindspore/tinyxml2.dll +0 -0
- mindspore/train/__init__.py +6 -5
- mindspore/train/_utils.py +178 -4
- mindspore/train/amp.py +167 -245
- mindspore/train/anf_ir_pb2.py +14 -2
- mindspore/train/callback/__init__.py +5 -2
- mindspore/train/callback/_backup_and_restore.py +5 -5
- mindspore/train/callback/_callback.py +4 -4
- mindspore/train/callback/_checkpoint.py +151 -37
- mindspore/train/callback/_cluster_monitor.py +201 -0
- mindspore/train/callback/_early_stop.py +2 -2
- mindspore/train/callback/_flops_collector.py +238 -0
- mindspore/train/callback/_landscape.py +16 -11
- mindspore/train/callback/_loss_monitor.py +2 -2
- mindspore/train/callback/_mindio_ttp.py +443 -0
- mindspore/train/callback/_on_request_exit.py +2 -2
- mindspore/train/callback/_reduce_lr_on_plateau.py +2 -2
- mindspore/train/callback/_summary_collector.py +13 -14
- mindspore/train/callback/_time_monitor.py +3 -3
- mindspore/train/data_sink.py +6 -5
- mindspore/train/dataset_helper.py +66 -21
- mindspore/train/loss_scale_manager.py +2 -2
- mindspore/train/metrics/accuracy.py +7 -7
- mindspore/train/metrics/confusion_matrix.py +8 -6
- mindspore/train/metrics/cosine_similarity.py +6 -4
- mindspore/train/metrics/error.py +2 -2
- mindspore/train/metrics/metric.py +3 -3
- mindspore/train/metrics/perplexity.py +2 -1
- mindspore/train/metrics/topk.py +2 -2
- mindspore/train/mind_ir_pb2.py +89 -15
- mindspore/train/model.py +298 -56
- mindspore/train/serialization.py +501 -221
- mindspore/train/summary/_summary_adapter.py +1 -1
- mindspore/train/summary/_writer_pool.py +1 -1
- mindspore/train/summary/summary_record.py +56 -34
- mindspore/train/train_thor/convert_utils.py +3 -3
- mindspore/turbojpeg.dll +0 -0
- mindspore/version.py +1 -1
- {mindspore-2.2.11.dist-info → mindspore-2.3.0.dist-info}/METADATA +3 -3
- mindspore-2.3.0.dist-info/RECORD +1400 -0
- {mindspore-2.2.11.dist-info → mindspore-2.3.0.dist-info}/entry_points.txt +1 -0
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_adapter.py +0 -662
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_helper.py +0 -377
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_job.py +0 -201
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_job_manager.py +0 -515
- mindspore/gen_ops.py +0 -273
- mindspore/nn/layer/flash_attention.py +0 -189
- mindspore/ops/_op_impl/cpu/tensor_shape.py +0 -42
- mindspore/ops/_op_impl/tbe/__init__.py +0 -47
- mindspore/ops/_op_impl/tbe/abs.py +0 -38
- mindspore/ops/_op_impl/tbe/abs_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/abs_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/abs_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/accumulate_n_v2.py +0 -41
- mindspore/ops/_op_impl/tbe/accumulate_n_v2_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/acos.py +0 -37
- mindspore/ops/_op_impl/tbe/acos_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/acos_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/acos_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/acosh.py +0 -37
- mindspore/ops/_op_impl/tbe/acosh_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/acosh_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/acosh_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/act_ulq_clamp_max_grad.py +0 -38
- mindspore/ops/_op_impl/tbe/act_ulq_clamp_min_grad.py +0 -38
- mindspore/ops/_op_impl/tbe/acts_ulq.py +0 -45
- mindspore/ops/_op_impl/tbe/acts_ulq_input_grad.py +0 -38
- mindspore/ops/_op_impl/tbe/adam_apply_one.py +0 -50
- mindspore/ops/_op_impl/tbe/adam_apply_one_assign.py +0 -53
- mindspore/ops/_op_impl/tbe/adam_apply_one_ds.py +0 -51
- mindspore/ops/_op_impl/tbe/adam_apply_one_with_decay.py +0 -54
- mindspore/ops/_op_impl/tbe/adam_apply_one_with_decay_assign.py +0 -54
- mindspore/ops/_op_impl/tbe/adam_apply_one_with_decay_ds.py +0 -55
- mindspore/ops/_op_impl/tbe/adaptive_max_pool2d.py +0 -37
- mindspore/ops/_op_impl/tbe/add.py +0 -42
- mindspore/ops/_op_impl/tbe/add_ds.py +0 -43
- mindspore/ops/_op_impl/tbe/add_n.py +0 -39
- mindspore/ops/_op_impl/tbe/add_n_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/addcdiv.py +0 -41
- mindspore/ops/_op_impl/tbe/addcdiv_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/addcmul.py +0 -43
- mindspore/ops/_op_impl/tbe/addcmul_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/apply_ada_max.py +0 -68
- mindspore/ops/_op_impl/tbe/apply_ada_max_ds.py +0 -69
- mindspore/ops/_op_impl/tbe/apply_adadelta.py +0 -66
- mindspore/ops/_op_impl/tbe/apply_adadelta_ds.py +0 -67
- mindspore/ops/_op_impl/tbe/apply_adagrad.py +0 -55
- mindspore/ops/_op_impl/tbe/apply_adagrad_d_a.py +0 -67
- mindspore/ops/_op_impl/tbe/apply_adagrad_ds.py +0 -56
- mindspore/ops/_op_impl/tbe/apply_adagrad_v2.py +0 -48
- mindspore/ops/_op_impl/tbe/apply_adagrad_v2_ds.py +0 -49
- mindspore/ops/_op_impl/tbe/apply_adam.py +0 -79
- mindspore/ops/_op_impl/tbe/apply_adam_ds.py +0 -80
- mindspore/ops/_op_impl/tbe/apply_adam_with_amsgrad.py +0 -60
- mindspore/ops/_op_impl/tbe/apply_adam_with_amsgrad_ds.py +0 -61
- mindspore/ops/_op_impl/tbe/apply_add_sign.py +0 -65
- mindspore/ops/_op_impl/tbe/apply_add_sign_ds.py +0 -66
- mindspore/ops/_op_impl/tbe/apply_centered_rms_prop.py +0 -77
- mindspore/ops/_op_impl/tbe/apply_centered_rms_prop_ds.py +0 -78
- mindspore/ops/_op_impl/tbe/apply_ftrl.py +0 -67
- mindspore/ops/_op_impl/tbe/apply_ftrl_ds.py +0 -68
- mindspore/ops/_op_impl/tbe/apply_gradient_descent.py +0 -44
- mindspore/ops/_op_impl/tbe/apply_gradient_descent_ds.py +0 -45
- mindspore/ops/_op_impl/tbe/apply_keras_momentum.py +0 -49
- mindspore/ops/_op_impl/tbe/apply_momentum.py +0 -64
- mindspore/ops/_op_impl/tbe/apply_momentum_ds.py +0 -65
- mindspore/ops/_op_impl/tbe/apply_power_sign.py +0 -65
- mindspore/ops/_op_impl/tbe/apply_power_sign_ds.py +0 -66
- mindspore/ops/_op_impl/tbe/apply_proximal_adagrad.py +0 -57
- mindspore/ops/_op_impl/tbe/apply_proximal_adagrad_ds.py +0 -58
- mindspore/ops/_op_impl/tbe/apply_proximal_gradient_descent.py +0 -54
- mindspore/ops/_op_impl/tbe/apply_proximal_gradient_descent_ds.py +0 -55
- mindspore/ops/_op_impl/tbe/apply_rms_prop.py +0 -52
- mindspore/ops/_op_impl/tbe/approximate_equal.py +0 -39
- mindspore/ops/_op_impl/tbe/approximate_equal_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/arg_max.py +0 -38
- mindspore/ops/_op_impl/tbe/arg_max_with_value.py +0 -38
- mindspore/ops/_op_impl/tbe/arg_max_with_value_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/arg_min.py +0 -38
- mindspore/ops/_op_impl/tbe/arg_min_v2_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/arg_min_with_value.py +0 -38
- mindspore/ops/_op_impl/tbe/arg_min_with_value_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/asin.py +0 -37
- mindspore/ops/_op_impl/tbe/asin_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/asin_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/asin_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/asinh.py +0 -37
- mindspore/ops/_op_impl/tbe/asinh_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/asinh_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/asinh_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/assign.py +0 -79
- mindspore/ops/_op_impl/tbe/assign_add.py +0 -59
- mindspore/ops/_op_impl/tbe/assign_add_ds.py +0 -60
- mindspore/ops/_op_impl/tbe/assign_ds.py +0 -80
- mindspore/ops/_op_impl/tbe/assign_sub.py +0 -55
- mindspore/ops/_op_impl/tbe/assign_sub_ds.py +0 -56
- mindspore/ops/_op_impl/tbe/atan.py +0 -37
- mindspore/ops/_op_impl/tbe/atan2.py +0 -38
- mindspore/ops/_op_impl/tbe/atan2_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/atan_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/atan_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/atan_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/atanh.py +0 -37
- mindspore/ops/_op_impl/tbe/atanh_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/avg_pool.py +0 -43
- mindspore/ops/_op_impl/tbe/avg_pool_3d.py +0 -44
- mindspore/ops/_op_impl/tbe/avg_pool_3d_grad.py +0 -45
- mindspore/ops/_op_impl/tbe/avg_pool_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/avg_pool_grad.py +0 -42
- mindspore/ops/_op_impl/tbe/avg_pool_grad_vm.py +0 -42
- mindspore/ops/_op_impl/tbe/basic_lstm_cell.py +0 -57
- mindspore/ops/_op_impl/tbe/basic_lstm_cell_c_state_grad.py +0 -50
- mindspore/ops/_op_impl/tbe/basic_lstm_cell_c_state_grad_v2.py +0 -51
- mindspore/ops/_op_impl/tbe/basic_lstm_cell_input_grad.py +0 -42
- mindspore/ops/_op_impl/tbe/basic_lstm_cell_weight_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/batch_matmul.py +0 -42
- mindspore/ops/_op_impl/tbe/batch_matmul_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/batch_matmul_v2.py +0 -47
- mindspore/ops/_op_impl/tbe/batch_to_space.py +0 -38
- mindspore/ops/_op_impl/tbe/batch_to_space_nd.py +0 -38
- mindspore/ops/_op_impl/tbe/batch_to_space_nd_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/batch_to_space_nd_v2.py +0 -41
- mindspore/ops/_op_impl/tbe/batchnorm.py +0 -58
- mindspore/ops/_op_impl/tbe/batchnorm_grad.py +0 -58
- mindspore/ops/_op_impl/tbe/bce_with_logits_loss.py +0 -42
- mindspore/ops/_op_impl/tbe/bessel_i0e.py +0 -37
- mindspore/ops/_op_impl/tbe/bessel_i0e_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/bessel_i1e.py +0 -37
- mindspore/ops/_op_impl/tbe/bessel_i1e_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/bias_add.py +0 -38
- mindspore/ops/_op_impl/tbe/bias_add_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/bias_add_grad.py +0 -53
- mindspore/ops/_op_impl/tbe/binary_cross_entropy.py +0 -39
- mindspore/ops/_op_impl/tbe/binary_cross_entropy_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/binary_cross_entropy_grad.py +0 -44
- mindspore/ops/_op_impl/tbe/binary_cross_entropy_grad_ds.py +0 -45
- mindspore/ops/_op_impl/tbe/bitwise_and.py +0 -39
- mindspore/ops/_op_impl/tbe/bitwise_and_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/bitwise_or.py +0 -39
- mindspore/ops/_op_impl/tbe/bitwise_or_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/bitwise_xor.py +0 -39
- mindspore/ops/_op_impl/tbe/bitwise_xor_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/bn_infer.py +0 -43
- mindspore/ops/_op_impl/tbe/bn_infer_ds.py +0 -45
- mindspore/ops/_op_impl/tbe/bn_infer_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/bn_infer_grad_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/bn_inference.py +0 -50
- mindspore/ops/_op_impl/tbe/bn_training_reduce.py +0 -38
- mindspore/ops/_op_impl/tbe/bn_training_reduce_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/bn_training_reduce_grad.py +0 -46
- mindspore/ops/_op_impl/tbe/bn_training_reduce_grad_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/bn_training_update.py +0 -52
- mindspore/ops/_op_impl/tbe/bn_training_update_ds.py +0 -53
- mindspore/ops/_op_impl/tbe/bn_training_update_grad.py +0 -44
- mindspore/ops/_op_impl/tbe/bn_training_update_grad_ds.py +0 -45
- mindspore/ops/_op_impl/tbe/bn_training_update_v2.py +0 -48
- mindspore/ops/_op_impl/tbe/bn_training_update_v3.py +0 -51
- mindspore/ops/_op_impl/tbe/bounding_box_decode.py +0 -41
- mindspore/ops/_op_impl/tbe/bounding_box_decode_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/bounding_box_encode.py +0 -38
- mindspore/ops/_op_impl/tbe/broadcast_to.py +0 -40
- mindspore/ops/_op_impl/tbe/broadcast_to_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/cast.py +0 -55
- mindspore/ops/_op_impl/tbe/cast_ds.py +0 -58
- mindspore/ops/_op_impl/tbe/cdist.py +0 -38
- mindspore/ops/_op_impl/tbe/cdist_grad.py +0 -42
- mindspore/ops/_op_impl/tbe/ceil.py +0 -37
- mindspore/ops/_op_impl/tbe/ceil_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/celu.py +0 -39
- mindspore/ops/_op_impl/tbe/centralization.py +0 -39
- mindspore/ops/_op_impl/tbe/check_valid.py +0 -38
- mindspore/ops/_op_impl/tbe/check_valid_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/clip_by_norm_no_div_sum.py +0 -41
- mindspore/ops/_op_impl/tbe/clip_by_norm_no_div_sum_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/clip_by_value.py +0 -41
- mindspore/ops/_op_impl/tbe/clip_by_value_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/concat.py +0 -40
- mindspore/ops/_op_impl/tbe/concat_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/confusion_matrix.py +0 -63
- mindspore/ops/_op_impl/tbe/confusion_mul_grad.py +0 -40
- mindspore/ops/_op_impl/tbe/confusion_softmax_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/confusion_transpose_d.py +0 -39
- mindspore/ops/_op_impl/tbe/conv2d.py +0 -47
- mindspore/ops/_op_impl/tbe/conv2d_backprop_filter.py +0 -42
- mindspore/ops/_op_impl/tbe/conv2d_backprop_filter_ds.py +0 -43
- mindspore/ops/_op_impl/tbe/conv2d_backprop_input.py +0 -42
- mindspore/ops/_op_impl/tbe/conv2d_backprop_input_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/conv2d_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/conv2d_transpose.py +0 -48
- mindspore/ops/_op_impl/tbe/conv3d.py +0 -45
- mindspore/ops/_op_impl/tbe/conv3d_backprop_filter.py +0 -42
- mindspore/ops/_op_impl/tbe/conv3d_backprop_input.py +0 -42
- mindspore/ops/_op_impl/tbe/conv3d_transpose.py +0 -47
- mindspore/ops/_op_impl/tbe/conv3d_transpose_ds.py +0 -48
- mindspore/ops/_op_impl/tbe/cos.py +0 -37
- mindspore/ops/_op_impl/tbe/cos_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/cosh.py +0 -37
- mindspore/ops/_op_impl/tbe/cosh_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/ctc_loss_v2.py +0 -42
- mindspore/ops/_op_impl/tbe/ctc_loss_v2_grad.py +0 -44
- mindspore/ops/_op_impl/tbe/cum_sum.py +0 -42
- mindspore/ops/_op_impl/tbe/cum_sum_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/cummin.py +0 -41
- mindspore/ops/_op_impl/tbe/cumprod.py +0 -42
- mindspore/ops/_op_impl/tbe/data_format_dim_map.py +0 -38
- mindspore/ops/_op_impl/tbe/data_format_dim_map_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/deformable_offsets.py +0 -45
- mindspore/ops/_op_impl/tbe/deformable_offsets_grad.py +0 -48
- mindspore/ops/_op_impl/tbe/depth_to_space_ds.py +0 -49
- mindspore/ops/_op_impl/tbe/depthwise_conv2d.py +0 -44
- mindspore/ops/_op_impl/tbe/depthwise_conv2d_backprop_filter.py +0 -41
- mindspore/ops/_op_impl/tbe/depthwise_conv2d_backprop_input.py +0 -41
- mindspore/ops/_op_impl/tbe/diag.py +0 -38
- mindspore/ops/_op_impl/tbe/diag_part.py +0 -38
- mindspore/ops/_op_impl/tbe/dilation.py +0 -40
- mindspore/ops/_op_impl/tbe/div.py +0 -41
- mindspore/ops/_op_impl/tbe/div_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/div_no_nan.py +0 -41
- mindspore/ops/_op_impl/tbe/div_no_nan_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/dropout_do_mask.py +0 -38
- mindspore/ops/_op_impl/tbe/dropout_do_mask_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/dropout_do_mask_v3.py +0 -39
- mindspore/ops/_op_impl/tbe/dynamic_atomic_addr_clean.py +0 -34
- mindspore/ops/_op_impl/tbe/dynamic_gru_v2.py +0 -95
- mindspore/ops/_op_impl/tbe/dynamic_rnn.py +0 -82
- mindspore/ops/_op_impl/tbe/elu.py +0 -38
- mindspore/ops/_op_impl/tbe/elu_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/elu_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/elu_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/equal.py +0 -42
- mindspore/ops/_op_impl/tbe/equal_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/erf.py +0 -37
- mindspore/ops/_op_impl/tbe/erf_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/erfc.py +0 -37
- mindspore/ops/_op_impl/tbe/erfc_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/erfinv.py +0 -36
- mindspore/ops/_op_impl/tbe/exp.py +0 -40
- mindspore/ops/_op_impl/tbe/exp_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/expand_dims.py +0 -38
- mindspore/ops/_op_impl/tbe/expm1.py +0 -37
- mindspore/ops/_op_impl/tbe/expm1_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/extract_image_patches.py +0 -41
- mindspore/ops/_op_impl/tbe/extract_volume_patches.py +0 -39
- mindspore/ops/_op_impl/tbe/fake_quant_with_min_max_vars.py +0 -39
- mindspore/ops/_op_impl/tbe/fake_quant_with_min_max_vars_gradient.py +0 -43
- mindspore/ops/_op_impl/tbe/fake_quant_with_min_max_vars_per_channel.py +0 -39
- mindspore/ops/_op_impl/tbe/fake_quant_with_min_max_vars_per_channel_gradient.py +0 -43
- mindspore/ops/_op_impl/tbe/fast_gelu.py +0 -37
- mindspore/ops/_op_impl/tbe/fast_gelu_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/fast_gelu_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/fast_gelu_grad_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/fill.py +0 -56
- mindspore/ops/_op_impl/tbe/fill_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/flatten.py +0 -48
- mindspore/ops/_op_impl/tbe/floor.py +0 -37
- mindspore/ops/_op_impl/tbe/floor_div.py +0 -41
- mindspore/ops/_op_impl/tbe/floor_div_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/floor_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/floor_mod.py +0 -39
- mindspore/ops/_op_impl/tbe/floor_mod_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/fused_dbn_dw.py +0 -52
- mindspore/ops/_op_impl/tbe/fused_mul_add.py +0 -38
- mindspore/ops/_op_impl/tbe/fused_mul_add_n.py +0 -48
- mindspore/ops/_op_impl/tbe/fused_mul_add_n_l2loss.py +0 -53
- mindspore/ops/_op_impl/tbe/fused_mul_apply_momentum.py +0 -57
- mindspore/ops/_op_impl/tbe/fused_mul_apply_momentum_extern.py +0 -67
- mindspore/ops/_op_impl/tbe/gather_nd.py +0 -52
- mindspore/ops/_op_impl/tbe/gather_nd_ds.py +0 -48
- mindspore/ops/_op_impl/tbe/gather_v2.py +0 -56
- mindspore/ops/_op_impl/tbe/gather_v2_ds.py +0 -68
- mindspore/ops/_op_impl/tbe/gelu.py +0 -37
- mindspore/ops/_op_impl/tbe/gelu_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/gelu_grad.py +0 -42
- mindspore/ops/_op_impl/tbe/gelu_grad_ds.py +0 -43
- mindspore/ops/_op_impl/tbe/ger.py +0 -43
- mindspore/ops/_op_impl/tbe/ger_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/greater.py +0 -43
- mindspore/ops/_op_impl/tbe/greater_equal.py +0 -41
- mindspore/ops/_op_impl/tbe/greater_equal_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/gru_v2_hidden_grad.py +0 -51
- mindspore/ops/_op_impl/tbe/gru_v2_hidden_grad_cell.py +0 -52
- mindspore/ops/_op_impl/tbe/hard_swish.py +0 -37
- mindspore/ops/_op_impl/tbe/hard_swish_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/hard_swish_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/hard_swish_grad_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/histogram_fixed_width.py +0 -40
- mindspore/ops/_op_impl/tbe/hshrink.py +0 -33
- mindspore/ops/_op_impl/tbe/hshrink_grad.py +0 -37
- mindspore/ops/_op_impl/tbe/hsigmoid.py +0 -45
- mindspore/ops/_op_impl/tbe/hsigmoid_grad.py +0 -39
- mindspore/ops/_op_impl/tbe/ifmr.py +0 -47
- mindspore/ops/_op_impl/tbe/ifmr_ds.py +0 -48
- mindspore/ops/_op_impl/tbe/im2col.py +0 -42
- mindspore/ops/_op_impl/tbe/in_top_k.py +0 -37
- mindspore/ops/_op_impl/tbe/inplace_add.py +0 -39
- mindspore/ops/_op_impl/tbe/inplace_index_add.py +0 -46
- mindspore/ops/_op_impl/tbe/inplace_sub.py +0 -39
- mindspore/ops/_op_impl/tbe/inplace_update.py +0 -39
- mindspore/ops/_op_impl/tbe/inplace_update_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/inv.py +0 -38
- mindspore/ops/_op_impl/tbe/inv_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/inv_grad.py +0 -40
- mindspore/ops/_op_impl/tbe/inv_grad_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/invert.py +0 -37
- mindspore/ops/_op_impl/tbe/invert_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/iou.py +0 -38
- mindspore/ops/_op_impl/tbe/iou_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/is_close.py +0 -40
- mindspore/ops/_op_impl/tbe/kl_div_loss.py +0 -38
- mindspore/ops/_op_impl/tbe/kl_div_loss_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/kl_div_loss_grad.py +0 -40
- mindspore/ops/_op_impl/tbe/l2_loss.py +0 -36
- mindspore/ops/_op_impl/tbe/l2_loss_ds.py +0 -37
- mindspore/ops/_op_impl/tbe/l2_normalize.py +0 -38
- mindspore/ops/_op_impl/tbe/l2_normalize_grad.py +0 -40
- mindspore/ops/_op_impl/tbe/lamb_apply_optimizer_assign.py +0 -55
- mindspore/ops/_op_impl/tbe/lamb_apply_weight_assign.py +0 -42
- mindspore/ops/_op_impl/tbe/lamb_next_mv.py +0 -59
- mindspore/ops/_op_impl/tbe/lamb_next_mv_with_decay.py +0 -59
- mindspore/ops/_op_impl/tbe/lamb_next_right.py +0 -44
- mindspore/ops/_op_impl/tbe/lamb_update_with_lr.py +0 -48
- mindspore/ops/_op_impl/tbe/lamb_update_with_lr_v2.py +0 -44
- mindspore/ops/_op_impl/tbe/lars_update.py +0 -50
- mindspore/ops/_op_impl/tbe/lars_update_ds.py +0 -51
- mindspore/ops/_op_impl/tbe/layer_norm.py +0 -46
- mindspore/ops/_op_impl/tbe/layer_norm_beta_gamma_backprop.py +0 -44
- mindspore/ops/_op_impl/tbe/layer_norm_beta_gamma_backprop_ds.py +0 -45
- mindspore/ops/_op_impl/tbe/layer_norm_beta_gamma_backprop_v2.py +0 -40
- mindspore/ops/_op_impl/tbe/layer_norm_beta_gamma_backprop_v2_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/layer_norm_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/layer_norm_grad.py +0 -48
- mindspore/ops/_op_impl/tbe/layer_norm_x_backprop.py +0 -43
- mindspore/ops/_op_impl/tbe/layer_norm_x_backprop_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/layer_norm_x_backprop_v2.py +0 -45
- mindspore/ops/_op_impl/tbe/layer_norm_x_backprop_v2_ds.py +0 -45
- mindspore/ops/_op_impl/tbe/lerp.py +0 -38
- mindspore/ops/_op_impl/tbe/less.py +0 -41
- mindspore/ops/_op_impl/tbe/less_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/less_equal.py +0 -41
- mindspore/ops/_op_impl/tbe/less_equal_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/log.py +0 -40
- mindspore/ops/_op_impl/tbe/log1p.py +0 -37
- mindspore/ops/_op_impl/tbe/log1p_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/log_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/logical_and.py +0 -37
- mindspore/ops/_op_impl/tbe/logical_and_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/logical_not.py +0 -36
- mindspore/ops/_op_impl/tbe/logical_not_ds.py +0 -37
- mindspore/ops/_op_impl/tbe/logical_or.py +0 -37
- mindspore/ops/_op_impl/tbe/logical_or_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/logsoftmax.py +0 -37
- mindspore/ops/_op_impl/tbe/logsoftmax_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/logsoftmax_grad.py +0 -38
- mindspore/ops/_op_impl/tbe/logsoftmax_grad_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/lp_norm.py +0 -40
- mindspore/ops/_op_impl/tbe/lp_norm_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/lrn.py +0 -41
- mindspore/ops/_op_impl/tbe/lrn_grad.py +0 -42
- mindspore/ops/_op_impl/tbe/lstm_input_grad.py +0 -51
- mindspore/ops/_op_impl/tbe/masked_fill.py +0 -40
- mindspore/ops/_op_impl/tbe/masked_fill_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/matmul.py +0 -53
- mindspore/ops/_op_impl/tbe/matmul_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/matmul_v2.py +0 -50
- mindspore/ops/_op_impl/tbe/matrix_diag.py +0 -45
- mindspore/ops/_op_impl/tbe/matrix_diag_part.py +0 -45
- mindspore/ops/_op_impl/tbe/matrix_set_diag.py +0 -46
- mindspore/ops/_op_impl/tbe/max_pool.py +0 -39
- mindspore/ops/_op_impl/tbe/max_pool3d.py +0 -44
- mindspore/ops/_op_impl/tbe/max_pool3d_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/max_pool3d_grad_grad.py +0 -44
- mindspore/ops/_op_impl/tbe/max_pool_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/max_pool_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/max_pool_grad_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/max_pool_grad_grad_with_argmax.py +0 -41
- mindspore/ops/_op_impl/tbe/max_pool_grad_with_argmax.py +0 -42
- mindspore/ops/_op_impl/tbe/max_pool_with_argmax.py +0 -40
- mindspore/ops/_op_impl/tbe/maximum.py +0 -39
- mindspore/ops/_op_impl/tbe/maximum_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/maximum_grad.py +0 -46
- mindspore/ops/_op_impl/tbe/maximum_grad_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/mem_set.py +0 -38
- mindspore/ops/_op_impl/tbe/minimum.py +0 -40
- mindspore/ops/_op_impl/tbe/minimum_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/minimum_grad.py +0 -46
- mindspore/ops/_op_impl/tbe/minimum_grad_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/mish.py +0 -37
- mindspore/ops/_op_impl/tbe/mod.py +0 -41
- mindspore/ops/_op_impl/tbe/mod_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/mul.py +0 -37
- mindspore/ops/_op_impl/tbe/mul_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/mul_no_nan.py +0 -39
- mindspore/ops/_op_impl/tbe/mul_no_nan_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/multilabel_margin_loss.py +0 -39
- mindspore/ops/_op_impl/tbe/neg.py +0 -39
- mindspore/ops/_op_impl/tbe/neg_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/new_im2col.py +0 -40
- mindspore/ops/_op_impl/tbe/nll_loss.py +0 -41
- mindspore/ops/_op_impl/tbe/nll_loss_grad.py +0 -44
- mindspore/ops/_op_impl/tbe/nms_with_mask.py +0 -39
- mindspore/ops/_op_impl/tbe/not_equal.py +0 -41
- mindspore/ops/_op_impl/tbe/not_equal_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/npu_alloc_float_status.py +0 -34
- mindspore/ops/_op_impl/tbe/npu_clear_float_status.py +0 -35
- mindspore/ops/_op_impl/tbe/npu_clear_float_status_v2.py +0 -35
- mindspore/ops/_op_impl/tbe/npu_get_float_status.py +0 -35
- mindspore/ops/_op_impl/tbe/npu_get_float_status_v2.py +0 -35
- mindspore/ops/_op_impl/tbe/one_hot.py +0 -48
- mindspore/ops/_op_impl/tbe/one_hot_ds.py +0 -45
- mindspore/ops/_op_impl/tbe/ones_like.py +0 -40
- mindspore/ops/_op_impl/tbe/ones_like_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/p_s_r_o_i_pooling.py +0 -40
- mindspore/ops/_op_impl/tbe/p_s_r_o_i_pooling_grad.py +0 -40
- mindspore/ops/_op_impl/tbe/pack.py +0 -58
- mindspore/ops/_op_impl/tbe/pack_ds.py +0 -59
- mindspore/ops/_op_impl/tbe/pad_d.py +0 -40
- mindspore/ops/_op_impl/tbe/pad_d_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/parallel_concat.py +0 -70
- mindspore/ops/_op_impl/tbe/parallel_resize_bilinear.py +0 -45
- mindspore/ops/_op_impl/tbe/parallel_resize_bilinear_grad.py +0 -44
- mindspore/ops/_op_impl/tbe/pdist.py +0 -36
- mindspore/ops/_op_impl/tbe/pooling.py +0 -46
- mindspore/ops/_op_impl/tbe/population_count.py +0 -38
- mindspore/ops/_op_impl/tbe/pow.py +0 -41
- mindspore/ops/_op_impl/tbe/pow_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/prelu.py +0 -37
- mindspore/ops/_op_impl/tbe/prelu_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/prelu_grad.py +0 -40
- mindspore/ops/_op_impl/tbe/range.py +0 -39
- mindspore/ops/_op_impl/tbe/real_div.py +0 -38
- mindspore/ops/_op_impl/tbe/real_div_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/reciprocal.py +0 -36
- mindspore/ops/_op_impl/tbe/reciprocal_ds.py +0 -37
- mindspore/ops/_op_impl/tbe/reciprocal_grad.py +0 -38
- mindspore/ops/_op_impl/tbe/reciprocal_grad_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/reduce_all.py +0 -38
- mindspore/ops/_op_impl/tbe/reduce_all_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/reduce_any.py +0 -38
- mindspore/ops/_op_impl/tbe/reduce_any_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/reduce_max.py +0 -43
- mindspore/ops/_op_impl/tbe/reduce_max_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/reduce_mean.py +0 -40
- mindspore/ops/_op_impl/tbe/reduce_mean_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/reduce_min.py +0 -41
- mindspore/ops/_op_impl/tbe/reduce_min_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/reduce_prod.py +0 -42
- mindspore/ops/_op_impl/tbe/reduce_prod_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/reduce_std.py +0 -44
- mindspore/ops/_op_impl/tbe/reduce_sum.py +0 -39
- mindspore/ops/_op_impl/tbe/reduce_sum_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/relu.py +0 -39
- mindspore/ops/_op_impl/tbe/relu6.py +0 -38
- mindspore/ops/_op_impl/tbe/relu6_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/relu6_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/relu6_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/relu_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/relu_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/relu_grad_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/relu_grad_v2.py +0 -40
- mindspore/ops/_op_impl/tbe/relu_grad_v2_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/relu_v2.py +0 -40
- mindspore/ops/_op_impl/tbe/relu_v2_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/renorm.py +0 -39
- mindspore/ops/_op_impl/tbe/resize_bilinear.py +0 -40
- mindspore/ops/_op_impl/tbe/resize_bilinear_grad.py +0 -41
- mindspore/ops/_op_impl/tbe/resize_bilinear_v2.py +0 -43
- mindspore/ops/_op_impl/tbe/resize_nearest_neighbor.py +0 -40
- mindspore/ops/_op_impl/tbe/resize_nearest_neighbor_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/resize_nearest_neighbor_grad.py +0 -39
- mindspore/ops/_op_impl/tbe/resize_nearest_neighbor_grad_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/reverse_v2_d.py +0 -37
- mindspore/ops/_op_impl/tbe/rint.py +0 -37
- mindspore/ops/_op_impl/tbe/rint_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/roi_align.py +0 -43
- mindspore/ops/_op_impl/tbe/roi_align_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/roi_align_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/roi_align_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/roll.py +0 -42
- mindspore/ops/_op_impl/tbe/round.py +0 -38
- mindspore/ops/_op_impl/tbe/round_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/rsqrt.py +0 -37
- mindspore/ops/_op_impl/tbe/rsqrt_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/rsqrt_grad.py +0 -40
- mindspore/ops/_op_impl/tbe/rsqrt_grad_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/scatter_add.py +0 -44
- mindspore/ops/_op_impl/tbe/scatter_div.py +0 -46
- mindspore/ops/_op_impl/tbe/scatter_max.py +0 -45
- mindspore/ops/_op_impl/tbe/scatter_min.py +0 -45
- mindspore/ops/_op_impl/tbe/scatter_mul.py +0 -44
- mindspore/ops/_op_impl/tbe/scatter_nd.py +0 -41
- mindspore/ops/_op_impl/tbe/scatter_nd_add.py +0 -45
- mindspore/ops/_op_impl/tbe/scatter_nd_d.py +0 -41
- mindspore/ops/_op_impl/tbe/scatter_nd_ds.py +0 -49
- mindspore/ops/_op_impl/tbe/scatter_nd_sub.py +0 -47
- mindspore/ops/_op_impl/tbe/scatter_nd_sub_ds.py +0 -48
- mindspore/ops/_op_impl/tbe/scatter_nd_update.py +0 -47
- mindspore/ops/_op_impl/tbe/scatter_nd_update_ds.py +0 -48
- mindspore/ops/_op_impl/tbe/scatter_non_aliasing_add.py +0 -39
- mindspore/ops/_op_impl/tbe/scatter_non_aliasing_add_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/scatter_sub.py +0 -47
- mindspore/ops/_op_impl/tbe/scatter_sub_ds.py +0 -48
- mindspore/ops/_op_impl/tbe/scatter_update.py +0 -43
- mindspore/ops/_op_impl/tbe/select.py +0 -38
- mindspore/ops/_op_impl/tbe/select_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/selu.py +0 -39
- mindspore/ops/_op_impl/tbe/selu_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/sgd.py +0 -62
- mindspore/ops/_op_impl/tbe/sigmoid.py +0 -37
- mindspore/ops/_op_impl/tbe/sigmoid_cross_entropy_with_logits.py +0 -41
- mindspore/ops/_op_impl/tbe/sigmoid_cross_entropy_with_logits_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/sigmoid_cross_entropy_with_logits_grad.py +0 -42
- mindspore/ops/_op_impl/tbe/sigmoid_cross_entropy_with_logits_grad_ds.py +0 -43
- mindspore/ops/_op_impl/tbe/sigmoid_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/sigmoid_grad.py +0 -39
- mindspore/ops/_op_impl/tbe/sigmoid_grad_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/sign.py +0 -38
- mindspore/ops/_op_impl/tbe/sign_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/sin.py +0 -37
- mindspore/ops/_op_impl/tbe/sin_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/sinh.py +0 -37
- mindspore/ops/_op_impl/tbe/sinh_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/slice.py +0 -58
- mindspore/ops/_op_impl/tbe/smooth_l1_loss.py +0 -45
- mindspore/ops/_op_impl/tbe/smooth_l1_loss_ds.py +0 -46
- mindspore/ops/_op_impl/tbe/smooth_l1_loss_grad.py +0 -46
- mindspore/ops/_op_impl/tbe/smooth_l1_loss_grad_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/soft_margin_loss.py +0 -38
- mindspore/ops/_op_impl/tbe/soft_margin_loss_grad.py +0 -39
- mindspore/ops/_op_impl/tbe/soft_shrink.py +0 -36
- mindspore/ops/_op_impl/tbe/soft_shrink_grad.py +0 -38
- mindspore/ops/_op_impl/tbe/softmax.py +0 -37
- mindspore/ops/_op_impl/tbe/softmax_cross_entropy_with_logits.py +0 -38
- mindspore/ops/_op_impl/tbe/softmax_cross_entropy_with_logits_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/softmax_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/softmax_grad_ext.py +0 -42
- mindspore/ops/_op_impl/tbe/softmax_v2_with_dropout_do_mask_v3.py +0 -39
- mindspore/ops/_op_impl/tbe/softplus.py +0 -37
- mindspore/ops/_op_impl/tbe/softplus_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/softplus_grad.py +0 -38
- mindspore/ops/_op_impl/tbe/softplus_grad_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/softsign.py +0 -37
- mindspore/ops/_op_impl/tbe/softsign_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/sort.py +0 -38
- mindspore/ops/_op_impl/tbe/sort_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/space_to_batch.py +0 -38
- mindspore/ops/_op_impl/tbe/space_to_batch_nd.py +0 -38
- mindspore/ops/_op_impl/tbe/space_to_depth.py +0 -47
- mindspore/ops/_op_impl/tbe/sparse_apply_adadelta.py +0 -56
- mindspore/ops/_op_impl/tbe/sparse_apply_adagrad.py +0 -45
- mindspore/ops/_op_impl/tbe/sparse_apply_adagrad_ds.py +0 -46
- mindspore/ops/_op_impl/tbe/sparse_apply_adagrad_v2.py +0 -46
- mindspore/ops/_op_impl/tbe/sparse_apply_adagrad_v2_ds.py +0 -47
- mindspore/ops/_op_impl/tbe/sparse_apply_ftrl_d.py +0 -53
- mindspore/ops/_op_impl/tbe/sparse_apply_ftrl_d_ds.py +0 -50
- mindspore/ops/_op_impl/tbe/sparse_apply_ftrl_v2.py +0 -50
- mindspore/ops/_op_impl/tbe/sparse_apply_proximal_adagrad.py +0 -66
- mindspore/ops/_op_impl/tbe/sparse_apply_proximal_adagrad_ds.py +0 -67
- mindspore/ops/_op_impl/tbe/sparse_apply_r_m_s_prop.py +0 -57
- mindspore/ops/_op_impl/tbe/sparse_apply_r_m_s_prop_ds.py +0 -58
- mindspore/ops/_op_impl/tbe/sparse_gather_v2.py +0 -56
- mindspore/ops/_op_impl/tbe/sparse_gather_v2_ds.py +0 -58
- mindspore/ops/_op_impl/tbe/split_d.py +0 -38
- mindspore/ops/_op_impl/tbe/split_d_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/split_v.py +0 -39
- mindspore/ops/_op_impl/tbe/splitv.py +0 -39
- mindspore/ops/_op_impl/tbe/sqrt.py +0 -37
- mindspore/ops/_op_impl/tbe/sqrt_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/sqrt_grad.py +0 -43
- mindspore/ops/_op_impl/tbe/sqrt_grad_ds.py +0 -44
- mindspore/ops/_op_impl/tbe/square.py +0 -38
- mindspore/ops/_op_impl/tbe/square_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/square_sum_all.py +0 -40
- mindspore/ops/_op_impl/tbe/square_sum_all_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/square_sum_v1.py +0 -38
- mindspore/ops/_op_impl/tbe/square_sum_v1_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/square_sum_v2.py +0 -39
- mindspore/ops/_op_impl/tbe/squared_difference.py +0 -39
- mindspore/ops/_op_impl/tbe/squared_difference_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/squeeze.py +0 -37
- mindspore/ops/_op_impl/tbe/strided_read.py +0 -38
- mindspore/ops/_op_impl/tbe/strided_slice_d.py +0 -44
- mindspore/ops/_op_impl/tbe/strided_slice_ds.py +0 -71
- mindspore/ops/_op_impl/tbe/strided_slice_grad_d.py +0 -51
- mindspore/ops/_op_impl/tbe/strided_slice_grad_ds.py +0 -57
- mindspore/ops/_op_impl/tbe/strided_write.py +0 -38
- mindspore/ops/_op_impl/tbe/sub.py +0 -39
- mindspore/ops/_op_impl/tbe/sub_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/tan.py +0 -38
- mindspore/ops/_op_impl/tbe/tan_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/tanh.py +0 -37
- mindspore/ops/_op_impl/tbe/tanh_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/tanh_grad.py +0 -39
- mindspore/ops/_op_impl/tbe/tanh_grad_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/tensor_move.py +0 -49
- mindspore/ops/_op_impl/tbe/tensor_move_ds.py +0 -50
- mindspore/ops/_op_impl/tbe/tensor_scatter_update.py +0 -41
- mindspore/ops/_op_impl/tbe/tile.py +0 -37
- mindspore/ops/_op_impl/tbe/tile_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/top_k.py +0 -42
- mindspore/ops/_op_impl/tbe/top_k_ds.py +0 -43
- mindspore/ops/_op_impl/tbe/trans_data.py +0 -167
- mindspore/ops/_op_impl/tbe/trans_data_ds.py +0 -180
- mindspore/ops/_op_impl/tbe/trans_data_rnn.py +0 -44
- mindspore/ops/_op_impl/tbe/transpose.py +0 -60
- mindspore/ops/_op_impl/tbe/transpose_d.py +0 -47
- mindspore/ops/_op_impl/tbe/transpose_nod.py +0 -60
- mindspore/ops/_op_impl/tbe/trunc.py +0 -39
- mindspore/ops/_op_impl/tbe/truncate_div.py +0 -41
- mindspore/ops/_op_impl/tbe/truncate_div_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/truncate_mod.py +0 -41
- mindspore/ops/_op_impl/tbe/truncate_mod_ds.py +0 -42
- mindspore/ops/_op_impl/tbe/unpack.py +0 -38
- mindspore/ops/_op_impl/tbe/unpack_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/unsorted_segment_max.py +0 -49
- mindspore/ops/_op_impl/tbe/unsorted_segment_max_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/unsorted_segment_min.py +0 -49
- mindspore/ops/_op_impl/tbe/unsorted_segment_min_ds.py +0 -40
- mindspore/ops/_op_impl/tbe/unsorted_segment_prod.py +0 -49
- mindspore/ops/_op_impl/tbe/unsorted_segment_prod_ds.py +0 -38
- mindspore/ops/_op_impl/tbe/unsorted_segment_sum.py +0 -38
- mindspore/ops/_op_impl/tbe/unsorted_segment_sum_ds.py +0 -41
- mindspore/ops/_op_impl/tbe/wts_arq.py +0 -40
- mindspore/ops/_op_impl/tbe/xdivy.py +0 -38
- mindspore/ops/_op_impl/tbe/xdivy_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/xlogy.py +0 -38
- mindspore/ops/_op_impl/tbe/xlogy_ds.py +0 -39
- mindspore/ops/_op_impl/tbe/zeros_like.py +0 -41
- mindspore/ops/_op_impl/tbe/zeros_like_ds.py +0 -42
- mindspore/ops/_tracefunc.py +0 -241
- mindspore/ops/arg_dtype_cast.py +0 -54
- mindspore/rewrite/api/tree_node_helper.py +0 -60
- mindspore/rewrite/ast_helpers/ast_creator.py +0 -115
- mindspore/rewrite/ast_transformers/flatten_recursive_stmt.py +0 -267
- mindspore/rewrite/ast_transformers/remove_return_out_of_if.py +0 -228
- mindspore/rewrite/namespace.py +0 -53
- mindspore-2.2.11.dist-info/RECORD +0 -1920
- {mindspore-2.2.11.dist-info → mindspore-2.3.0.dist-info}/WHEEL +0 -0
- {mindspore-2.2.11.dist-info → mindspore-2.3.0.dist-info}/top_level.txt +0 -0
mindspore/context.py
CHANGED
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
# Copyright 2020-
|
|
1
|
+
# Copyright 2020-2024 Huawei Technologies Co., Ltd
|
|
2
2
|
#
|
|
3
3
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
4
4
|
# you may not use this file except in compliance with the License.
|
|
@@ -34,6 +34,7 @@ from mindspore.parallel._auto_parallel_context import _set_auto_parallel_context
|
|
|
34
34
|
from mindspore.parallel._ps_context import _set_ps_context, _get_ps_context, _reset_ps_context, \
|
|
35
35
|
_need_reset_device_target_for_ps
|
|
36
36
|
from mindspore.parallel._offload_context import _set_offload_context, _get_offload_context
|
|
37
|
+
from mindspore.hal.device import is_initialized
|
|
37
38
|
|
|
38
39
|
__all__ = ['GRAPH_MODE', 'PYNATIVE_MODE', 'STRICT', 'COMPATIBLE', 'LAX', 'set_context', 'get_context',
|
|
39
40
|
'set_auto_parallel_context', 'get_auto_parallel_context', 'reset_auto_parallel_context', 'ParallelMode',
|
|
@@ -50,6 +51,10 @@ STRICT = 0
|
|
|
50
51
|
COMPATIBLE = 1
|
|
51
52
|
LAX = 2
|
|
52
53
|
|
|
54
|
+
# Enumerate for the property 'debug_level'.
|
|
55
|
+
RELEASE = 0
|
|
56
|
+
DEBUG = 1
|
|
57
|
+
|
|
53
58
|
|
|
54
59
|
def _make_directory(path):
|
|
55
60
|
"""Make directory."""
|
|
@@ -161,6 +166,9 @@ class _Context:
|
|
|
161
166
|
self._context_switches = _ContextSwitchInfo(False)
|
|
162
167
|
self._context_handle = MSContext.get_instance()
|
|
163
168
|
self._support_binary = False
|
|
169
|
+
self.enable_compile_cache = None
|
|
170
|
+
self._mode = PYNATIVE_MODE
|
|
171
|
+
self._jit_config = {}
|
|
164
172
|
|
|
165
173
|
def __getattribute__(self, attr):
|
|
166
174
|
value = object.__getattribute__(self, attr)
|
|
@@ -176,7 +184,11 @@ class _Context:
|
|
|
176
184
|
|
|
177
185
|
def get_mode(self):
|
|
178
186
|
"""Get current mode."""
|
|
179
|
-
return self.
|
|
187
|
+
return self._mode
|
|
188
|
+
|
|
189
|
+
def get_jit_config(self):
|
|
190
|
+
"""Get current jit_config."""
|
|
191
|
+
return self._jit_config
|
|
180
192
|
|
|
181
193
|
def set_mode(self, mode):
|
|
182
194
|
"""
|
|
@@ -204,6 +216,7 @@ class _Context:
|
|
|
204
216
|
raise ValueError(f"For 'context.set_context', the argument 'mode' should be context.GRAPH_MODE (0) "
|
|
205
217
|
f"or context.PYNATIVE_MODE (1), but got {mode}.")
|
|
206
218
|
self.set_param(ms_ctx_param.mode, mode)
|
|
219
|
+
self._mode = mode
|
|
207
220
|
|
|
208
221
|
def set_jit_syntax_level(self, level):
|
|
209
222
|
""""Set the JIT syntax level for graph compiling"""
|
|
@@ -212,6 +225,13 @@ class _Context:
|
|
|
212
225
|
f"or context.LAX, but got {level}.")
|
|
213
226
|
self.set_param(ms_ctx_param.jit_syntax_level, level)
|
|
214
227
|
|
|
228
|
+
def set_debug_level(self, level):
|
|
229
|
+
""""Set the debug level for graph compiling"""
|
|
230
|
+
if level != RELEASE and level != DEBUG:
|
|
231
|
+
raise ValueError(f"For 'context.set_debug_level', the argument 'level' should be context.RELEASE "
|
|
232
|
+
f"or context.DEBUG, but got {level}.")
|
|
233
|
+
self.set_param(ms_ctx_param.debug_level, level)
|
|
234
|
+
|
|
215
235
|
def set_memory_optimize_level(self, memory_optimize_level):
|
|
216
236
|
"""
|
|
217
237
|
The memory optimize level, support "O0", "O1".
|
|
@@ -268,11 +288,16 @@ class _Context:
|
|
|
268
288
|
"allow_mix_precision_fp16" and "allow_mix_precision_bf16".
|
|
269
289
|
- jit_compile (bool): ``False`` and ``True``.
|
|
270
290
|
- atomic_clean_policy (int): ``0`` and ``1``. Default: ``1`` .
|
|
291
|
+
- op_precision_mode (str): precision mode config file path.
|
|
292
|
+
- op_debug_option (str): Enable debugging options for Ascend operators,
|
|
293
|
+
default not enabled, only supports ``"oom"`` currently.
|
|
294
|
+
``"oom"``: Detect memory out of bounds.
|
|
295
|
+
- ge_options (dict): Global or session CANN options.
|
|
271
296
|
- exception_dump (str): Enable exception dump for Ascend operators. ``"0"`` , ``"1"`` and ``"2"``.
|
|
272
297
|
Default: ``"2"`` .
|
|
273
|
-
- op_precision_mode (str): config file path.
|
|
274
298
|
- parallel_speed_up_json_path(Union[str, None]): The path to the parallel speed up json file.
|
|
275
299
|
If its value is None or '', it does not take effect. Default None.
|
|
300
|
+
- host_scheduling_max_threshold(int): The host scheduling max threshold.
|
|
276
301
|
"""
|
|
277
302
|
ascend_cfg_modes = {
|
|
278
303
|
'precision_mode': ["force_fp16", "allow_fp32_to_fp16", "allow_mix_precision", "must_keep_origin_dtype",
|
|
@@ -284,7 +309,15 @@ class _Context:
|
|
|
284
309
|
'conv_allow_hf32': [True, False],
|
|
285
310
|
'exception_dump': ["0", "1", "2"],
|
|
286
311
|
'op_precision_mode': (str,),
|
|
287
|
-
'
|
|
312
|
+
'ge_options': (dict,),
|
|
313
|
+
'parallel_speed_up_json_path': (str, None),
|
|
314
|
+
'host_scheduling_max_threshold': (int,),
|
|
315
|
+
'cur_step_num': (int,),
|
|
316
|
+
'save_checkpoint_steps': (int,),
|
|
317
|
+
'need_ckpt': (bool,),
|
|
318
|
+
'last_triggered_step': (int,),
|
|
319
|
+
'topo_order': (dict,),
|
|
320
|
+
'op_debug_option': (str, None),
|
|
288
321
|
}
|
|
289
322
|
ascend_cfg_setters = {
|
|
290
323
|
'precision_mode': self._get_ascend_config_setter('precision_mode'),
|
|
@@ -293,8 +326,16 @@ class _Context:
|
|
|
293
326
|
'matmul_allow_hf32': self._get_ascend_config_setter('matmul_allow_hf32', lambda v: "1" if v else "0"),
|
|
294
327
|
'conv_allow_hf32': self._get_ascend_config_setter('conv_allow_hf32', lambda v: "1" if v else "0"),
|
|
295
328
|
'exception_dump': self._get_ascend_config_setter('exception_dump'),
|
|
329
|
+
'op_debug_option': self._set_op_debug_option,
|
|
296
330
|
'op_precision_mode': self._set_op_precision_mode,
|
|
297
|
-
'
|
|
331
|
+
'ge_options': self._set_ge_options,
|
|
332
|
+
'parallel_speed_up_json_path': self._set_speedup_config_path,
|
|
333
|
+
'host_scheduling_max_threshold': self._get_ascend_config_setter('host_scheduling_max_threshold', str),
|
|
334
|
+
'cur_step_num': self._set_cur_step_num,
|
|
335
|
+
'save_checkpoint_steps': self._set_save_checkpoint_steps,
|
|
336
|
+
'need_ckpt': self._set_need_ckpt,
|
|
337
|
+
'last_triggered_step': self._set_last_triggered_step,
|
|
338
|
+
'topo_order': self._set_topo_order
|
|
298
339
|
}
|
|
299
340
|
ascend_cfg_set = tuple(ascend_cfg_modes.keys())
|
|
300
341
|
for ascend_key, ascend_value in ascend_config.items():
|
|
@@ -352,6 +393,31 @@ class _Context:
|
|
|
352
393
|
if gpu_key == 'matmul_allow_tf32':
|
|
353
394
|
self.set_param(ms_ctx_param.matmul_allow_tf32, gpu_config[gpu_key])
|
|
354
395
|
|
|
396
|
+
def set_jit_config(self, jit_config):
|
|
397
|
+
"""
|
|
398
|
+
Enable jit config.
|
|
399
|
+
|
|
400
|
+
Args:
|
|
401
|
+
jit_config (dict):
|
|
402
|
+
|
|
403
|
+
- jit_level (str): "O0", "O1" or "O2" to control the compilation optimization level.
|
|
404
|
+
"""
|
|
405
|
+
jit_cfgs = {'jit_level': ["O0", "O1", "O2"], 'infer_boost': ["on", "off"]}
|
|
406
|
+
key_args_map = {'jit_level': ms_ctx_param.jit_level, 'infer_boost': ms_ctx_param.infer_boost}
|
|
407
|
+
for jit_key in jit_config:
|
|
408
|
+
if jit_key not in jit_cfgs:
|
|
409
|
+
raise ValueError(f"For 'context.set_context', the key of argument 'jit_config' must be one of "
|
|
410
|
+
f"{jit_cfgs}, but got {jit_key}.")
|
|
411
|
+
supported_value = jit_cfgs.get(jit_key)
|
|
412
|
+
if jit_config[jit_key] not in supported_value:
|
|
413
|
+
raise ValueError(f"For 'jit_cfgs', the value of argument {jit_key} must be one of "
|
|
414
|
+
f"{supported_value}, but got {jit_config[jit_key]}.")
|
|
415
|
+
self._jit_config = jit_config
|
|
416
|
+
self.set_param(key_args_map[jit_key], jit_config[jit_key])
|
|
417
|
+
|
|
418
|
+
if 'infer_boost' in jit_config and jit_config['infer_boost'] == "on" and jit_config['jit_level'] != "O0":
|
|
419
|
+
raise ValueError(f"Only jit_level set O0 can set infer_boost to on.")
|
|
420
|
+
|
|
355
421
|
def set_backend_policy(self, policy):
|
|
356
422
|
success = self._context_handle.set_backend_policy(policy)
|
|
357
423
|
if not success:
|
|
@@ -473,9 +539,13 @@ class _Context:
|
|
|
473
539
|
|
|
474
540
|
def set_mempool_block_size(self, mempool_block_size):
|
|
475
541
|
"""Set the block size of memory pool."""
|
|
476
|
-
|
|
542
|
+
global_jit_config = get_jit_config()
|
|
543
|
+
is_force_kbk = False
|
|
544
|
+
if global_jit_config:
|
|
545
|
+
is_force_kbk = global_jit_config.get('jit_level') == "O0" or global_jit_config.get('jit_level') == "O1"
|
|
546
|
+
if _get_mode() == GRAPH_MODE and not is_force_kbk:
|
|
477
547
|
logger.warning("Graph mode doesn't support to set parameter 'mempool_block_size' of context currently, "
|
|
478
|
-
"you can use context.set_context to set pynative mode.")
|
|
548
|
+
"you can use context.set_context to set pynative mode or set jit_level=O0/O1.")
|
|
479
549
|
return
|
|
480
550
|
if not Validator.check_str_by_regular(mempool_block_size, _RE_PATTERN):
|
|
481
551
|
raise ValueError("For 'context.set_context', the argument 'mempool_block_size' should be in "
|
|
@@ -561,8 +631,10 @@ class _Context:
|
|
|
561
631
|
'deterministic': set_deterministic,
|
|
562
632
|
'ascend_config': set_ascend_config,
|
|
563
633
|
'jit_syntax_level': set_jit_syntax_level,
|
|
634
|
+
'debug_level': set_debug_level,
|
|
564
635
|
'gpu_config': set_gpu_config,
|
|
565
636
|
'aoe_config': set_aoe_config,
|
|
637
|
+
'jit_config': set_jit_config,
|
|
566
638
|
}
|
|
567
639
|
|
|
568
640
|
@property
|
|
@@ -610,6 +682,16 @@ class _Context:
|
|
|
610
682
|
trans_fn = lambda x: x
|
|
611
683
|
return _config_setter
|
|
612
684
|
|
|
685
|
+
def _set_op_debug_option(self, option_value):
|
|
686
|
+
valid_order = {'oom'}
|
|
687
|
+
if not isinstance(option_value, str):
|
|
688
|
+
raise TypeError(f"For 'ascend_config', the type of 'op_debug_option' must be str, "
|
|
689
|
+
f"but got {type(option_value)}.")
|
|
690
|
+
if option_value not in valid_order:
|
|
691
|
+
raise ValueError(f"For 'ascend_config', the 'op_debug_option' supports being set to 'oom' currently, "
|
|
692
|
+
f"but got {option_value}.")
|
|
693
|
+
self.set_param(ms_ctx_param.op_debug_option, option_value)
|
|
694
|
+
|
|
613
695
|
def _set_op_precision_mode(self, ascend_value):
|
|
614
696
|
op_precision_path = ascend_value
|
|
615
697
|
real_path = os.path.realpath(op_precision_path)
|
|
@@ -618,6 +700,74 @@ class _Context:
|
|
|
618
700
|
f"got '{op_precision_path}'.")
|
|
619
701
|
self.set_param(ms_ctx_param.op_precision_mode, ascend_value)
|
|
620
702
|
|
|
703
|
+
def _set_ge_options(self, ge_options):
|
|
704
|
+
"""Set ge options."""
|
|
705
|
+
for level, options in ge_options.items():
|
|
706
|
+
if level not in ['global', 'session']:
|
|
707
|
+
raise ValueError(f"For 'ascend_config', the key of ge_options must be one of "
|
|
708
|
+
f"('global', 'session'), but got {level}.")
|
|
709
|
+
|
|
710
|
+
if not isinstance(options, dict):
|
|
711
|
+
raise TypeError(f"For 'ge_options', the type of {level} options must be dict, "
|
|
712
|
+
f"but got {type(options)}. The error options: {options}.")
|
|
713
|
+
|
|
714
|
+
for key, value in options.items():
|
|
715
|
+
if not isinstance(key, str):
|
|
716
|
+
raise TypeError(f"For 'ge_options', the type of key and value must be str, "
|
|
717
|
+
f"but got {type(key)}. The error key is {key}.")
|
|
718
|
+
if not isinstance(value, str):
|
|
719
|
+
raise TypeError(f"For 'ge_options', the type of key and value must be str, "
|
|
720
|
+
f"but got {type(value)}. The error value is {value}")
|
|
721
|
+
|
|
722
|
+
options_str = json.dumps(ge_options)
|
|
723
|
+
self.set_param(ms_ctx_param.ge_options, options_str)
|
|
724
|
+
|
|
725
|
+
def _set_topo_order(self, topo_order):
|
|
726
|
+
"""
|
|
727
|
+
Set topo order.
|
|
728
|
+
|
|
729
|
+
Args:
|
|
730
|
+
topo_order (dict):
|
|
731
|
+
key: str, the name of the graph.
|
|
732
|
+
value: str, the topo order of the graph, should be one of 'dfs', 'bfs', 'rdfs'.
|
|
733
|
+
"""
|
|
734
|
+
valid_order = {'dfs', 'bfs', 'rdfs'}
|
|
735
|
+
if not isinstance(topo_order, dict):
|
|
736
|
+
raise TypeError(f"For 'ascend_config', the 'topo_order' should be a dict, "
|
|
737
|
+
f"got '{type(topo_order)}'.")
|
|
738
|
+
for k, v in topo_order.items():
|
|
739
|
+
if not isinstance(k, str):
|
|
740
|
+
raise TypeError("key {} is not a str".format(k))
|
|
741
|
+
if v not in valid_order:
|
|
742
|
+
raise ValueError("value {} should be one of {}.".format(v, valid_order))
|
|
743
|
+
|
|
744
|
+
options_str = json.dumps(topo_order)
|
|
745
|
+
self.set_param(ms_ctx_param.topo_order, options_str)
|
|
746
|
+
|
|
747
|
+
def _set_need_ckpt(self, need_ckpt):
|
|
748
|
+
"""Set need ckpt flag"""
|
|
749
|
+
if not isinstance(need_ckpt, bool):
|
|
750
|
+
raise TypeError(f"For step num, the value type should be int, but got {type(need_ckpt)}, {need_ckpt}")
|
|
751
|
+
self.set_param(ms_ctx_param.need_ckpt, need_ckpt)
|
|
752
|
+
|
|
753
|
+
def _set_cur_step_num(self, step_num):
|
|
754
|
+
"""set current step num at every step begin"""
|
|
755
|
+
if not isinstance(step_num, int):
|
|
756
|
+
raise TypeError(f"For step num, the value type should be int, but got {type(step_num)}, {step_num}")
|
|
757
|
+
self.set_param(ms_ctx_param.cur_step_num, step_num)
|
|
758
|
+
|
|
759
|
+
def _set_save_checkpoint_steps(self, steps):
|
|
760
|
+
"""set save checkpoint steps before run"""
|
|
761
|
+
if not isinstance(steps, int):
|
|
762
|
+
raise TypeError(f"For step num, the value type should be int, but got {type(steps)}, {steps}")
|
|
763
|
+
self.set_param(ms_ctx_param.save_checkpoint_steps, steps)
|
|
764
|
+
|
|
765
|
+
def _set_last_triggered_step(self, step):
|
|
766
|
+
"""set last triggered save ckpt steps before run"""
|
|
767
|
+
if not isinstance(step, int):
|
|
768
|
+
raise TypeError(f"For step num, the value type should be int, but got {type(step)}, {step}")
|
|
769
|
+
self.set_param(ms_ctx_param.last_triggered_step, step)
|
|
770
|
+
|
|
621
771
|
def _set_speedup_config_path(self, speedup_config_path):
|
|
622
772
|
""""Check and set speedup config for auto parallel."""
|
|
623
773
|
if speedup_config_path is None or speedup_config_path == "":
|
|
@@ -628,23 +778,34 @@ class _Context:
|
|
|
628
778
|
f"{speedup_config_real_path} does not exist, please check whether the "
|
|
629
779
|
f"'parallel_speed_up_json_path' is correct.")
|
|
630
780
|
try:
|
|
631
|
-
valid_option = {"recompute_comm_overlap": ms_ctx_param.recompute_comm_overlap,
|
|
632
|
-
"matmul_grad_comm_overlap": ms_ctx_param.matmul_grad_comm_overlap,
|
|
633
|
-
"enable_task_opt": ms_ctx_param.enable_task_opt,
|
|
634
|
-
"enable_grad_comm_opt": ms_ctx_param.enable_grad_comm_opt,
|
|
635
|
-
"
|
|
636
|
-
|
|
637
|
-
"
|
|
781
|
+
valid_option = {"recompute_comm_overlap": (ms_ctx_param.recompute_comm_overlap, bool),
|
|
782
|
+
"matmul_grad_comm_overlap": (ms_ctx_param.matmul_grad_comm_overlap, bool),
|
|
783
|
+
"enable_task_opt": (ms_ctx_param.enable_task_opt, bool),
|
|
784
|
+
"enable_grad_comm_opt": (ms_ctx_param.enable_grad_comm_opt, bool),
|
|
785
|
+
"recompute_allgather_overlap_fagrad":
|
|
786
|
+
(ms_ctx_param.recompute_allgather_overlap_fagrad, bool),
|
|
787
|
+
"interleaved_matmul_comm": (ms_ctx_param.interleaved_matmul_comm, bool),
|
|
788
|
+
"bias_add_comm_swap": (ms_ctx_param.bias_add_comm_swap, bool),
|
|
789
|
+
"enable_opt_shard_comm_opt": (ms_ctx_param.enable_opt_shard_comm_opt, bool),
|
|
790
|
+
"enable_begin_end_inline_opt": (ms_ctx_param.enable_begin_end_inline_opt, bool),
|
|
791
|
+
"enable_concat_eliminate_opt": (ms_ctx_param.enable_concat_eliminate_opt, bool),
|
|
792
|
+
"interleaved_layernorm_comm": (ms_ctx_param.interleaved_layernorm_comm, bool),
|
|
793
|
+
"compute_communicate_fusion_level":
|
|
794
|
+
(ms_ctx_param.compute_communicate_fusion_level, int),
|
|
795
|
+
"enable_flash_attention_load_balance":
|
|
796
|
+
(ms_ctx_param.enable_flash_attention_load_balance, bool)}
|
|
638
797
|
with open(speedup_config_real_path, 'r') as f:
|
|
639
798
|
speedup_config = json.load(f)
|
|
640
|
-
for
|
|
641
|
-
if not isinstance(
|
|
642
|
-
raise TypeError("key {} is not a str".format(
|
|
643
|
-
if
|
|
644
|
-
raise ValueError("key {} should be one of {}.".format(
|
|
645
|
-
|
|
646
|
-
|
|
647
|
-
|
|
799
|
+
for key, value in speedup_config.items():
|
|
800
|
+
if not isinstance(key, str):
|
|
801
|
+
raise TypeError("key {} is not a str".format(key))
|
|
802
|
+
if key not in valid_option:
|
|
803
|
+
raise ValueError("key {} should be one of {}.".format(key, valid_option.keys()))
|
|
804
|
+
set_func, valid_type = valid_option.get(key)
|
|
805
|
+
if not isinstance(value, valid_type):
|
|
806
|
+
raise TypeError(f"The value type of {key} must be {valid_type}, "
|
|
807
|
+
f"but got value is {value} and type is {type(value)}.")
|
|
808
|
+
self.set_param(set_func, value)
|
|
648
809
|
except (TypeError, ValueError) as exo:
|
|
649
810
|
raise ValueError(str(exo) + "\nFor 'context.set_context', "
|
|
650
811
|
"open or load the 'speedup_config_path' file {} "
|
|
@@ -681,8 +842,9 @@ def _context():
|
|
|
681
842
|
auto_parallel_search_mode=str, search_mode=str, parameter_broadcast=bool, strategy_ckpt_load_file=str,
|
|
682
843
|
strategy_ckpt_save_file=str, full_batch=bool, enable_parallel_optimizer=bool, enable_alltoall=bool,
|
|
683
844
|
all_reduce_fusion_config=list, pipeline_stages=int, pipeline_segments=int,
|
|
684
|
-
parallel_optimizer_config=dict,
|
|
685
|
-
|
|
845
|
+
pipeline_result_broadcast=bool, parallel_optimizer_config=dict,
|
|
846
|
+
pipeline_config=dict,
|
|
847
|
+
comm_fusion=dict, strategy_ckpt_config=dict, force_fp32_communication=bool)
|
|
686
848
|
def set_auto_parallel_context(**kwargs):
|
|
687
849
|
r"""
|
|
688
850
|
Set auto parallel context, only data parallel supported on CPU.
|
|
@@ -708,9 +870,12 @@ def set_auto_parallel_context(**kwargs):
|
|
|
708
870
|
enable_parallel_optimizer strategy_ckpt_save_file
|
|
709
871
|
parallel_optimizer_config dataset_strategy
|
|
710
872
|
enable_alltoall pipeline_stages
|
|
711
|
-
|
|
873
|
+
pipeline_config auto_parallel_search_mode
|
|
874
|
+
force_fp32_communication pipeline_result_broadcast
|
|
712
875
|
\ comm_fusion
|
|
713
876
|
\ strategy_ckpt_config
|
|
877
|
+
\ group_ckpt_save_file
|
|
878
|
+
\ auto_pipeline
|
|
714
879
|
=========================== ===========================
|
|
715
880
|
|
|
716
881
|
Args:
|
|
@@ -720,6 +885,8 @@ def set_auto_parallel_context(**kwargs):
|
|
|
720
885
|
"stand_alone" do not support gradients_mean. Default: ``False`` .
|
|
721
886
|
gradient_fp32_sync (bool): Run allreduce of gradients in fp32. "stand_alone", "data_parallel"
|
|
722
887
|
and "hybrid_parallel" do not support gradient_fp32_sync. Default: ``True`` .
|
|
888
|
+
loss_repeated_mean (bool) - Indicates whether the mean operator is executed backwards when the
|
|
889
|
+
calculation is repeated. Default: ``True`` .
|
|
723
890
|
parallel_mode (str): There are five kinds of parallel modes, ``"stand_alone"`` , ``"data_parallel"`` ,
|
|
724
891
|
``"hybrid_parallel"`` , ``"semi_auto_parallel"`` and ``"auto_parallel"`` . Note the pynative mode
|
|
725
892
|
only supports the ``"stand_alone"`` and ``"data_parallel"`` mode. Default: ``"stand_alone"`` .
|
|
@@ -734,15 +901,16 @@ def set_auto_parallel_context(**kwargs):
|
|
|
734
901
|
|
|
735
902
|
- auto_parallel: Achieving parallelism automatically.
|
|
736
903
|
search_mode (str): There are three kinds of shard strategy search modes: ``"recursive_programming"`` ,
|
|
737
|
-
``"
|
|
904
|
+
``"sharding_propagation"`` and ``"dynamic_programming"`` (Not recommended).
|
|
905
|
+
Default: ``"recursive_programming"`` .
|
|
738
906
|
|
|
739
907
|
- recursive_programming: Recursive programming search mode. In order to obtain optimal performance,
|
|
740
908
|
it is recommended that users set the batch size to be greater than or equal to the product of
|
|
741
909
|
the number of devices and the number of multi-copy parallelism.
|
|
742
910
|
|
|
743
|
-
- dynamic_programming: Dynamic programming search mode.
|
|
744
|
-
|
|
745
911
|
- sharding_propagation: Propagate shardings from configured ops to non-configured ops.
|
|
912
|
+
|
|
913
|
+
- dynamic_programming: Dynamic programming search mode.
|
|
746
914
|
auto_parallel_search_mode (str): This is the old version of 'search_mode'. Here, remaining this attribute is
|
|
747
915
|
for forward compatibility, and this attribute will be deleted in a future MindSpore version.
|
|
748
916
|
parameter_broadcast (bool): Whether to broadcast parameters before training. Before training, in order to have
|
|
@@ -768,6 +936,9 @@ def set_auto_parallel_context(**kwargs):
|
|
|
768
936
|
data parallel training in the benefit of time and memory saving. Currently, auto and semi auto
|
|
769
937
|
parallel mode support all optimizers in both Ascend and GPU. Data parallel mode only supports
|
|
770
938
|
`Lamb` and `AdamWeightDecay` in Ascend . Default: ``False`` .
|
|
939
|
+
force_fp32_communication (bool): A switch that determines whether reduce operators (AllReduce, ReduceScatter)
|
|
940
|
+
are forced to use the fp32 data type for communication during communication. True is the enable
|
|
941
|
+
switch. Default: ``False`` .
|
|
771
942
|
enable_alltoall (bool): A switch that allows AllToAll operators to be generated during communication. If its
|
|
772
943
|
value is ``False`` , there will be a combination of operators such as AllGather, Split and
|
|
773
944
|
Concat instead of AllToAll. Default: ``False`` .
|
|
@@ -777,6 +948,14 @@ def set_auto_parallel_context(**kwargs):
|
|
|
777
948
|
distributed alone in the pipeline. The total devices will be divided into 'pipeline_stags'
|
|
778
949
|
stages.
|
|
779
950
|
Default: ``1`` .
|
|
951
|
+
pipeline_result_broadcast (bool): A switch that broadcast the last stage result to all other stage in pipeline
|
|
952
|
+
parallel inference. Default: ``False`` .
|
|
953
|
+
pipeline_config (dict): A dict contains the keys and values for setting the pipeline parallelism configuration.
|
|
954
|
+
It supports the following keys:
|
|
955
|
+
|
|
956
|
+
- pipeline_interleave(bool): Indicates whether to enable the interleaved execution mode.
|
|
957
|
+
- pipeline_scheduler(str): Indicates the scheduling mode for pipeline parallelism. Only support
|
|
958
|
+
``gpipe/1f1b``.
|
|
780
959
|
parallel_optimizer_config (dict): A dict contains the keys and values for setting the parallel optimizer
|
|
781
960
|
configure. The configure provides more detailed behavior control about parallel training
|
|
782
961
|
when parallel optimizer is enabled. The configure will be effective when we use
|
|
@@ -836,14 +1015,18 @@ def set_auto_parallel_context(**kwargs):
|
|
|
836
1015
|
- load_file (str): The path to load parallel strategy checkpoint. If the file name extension is
|
|
837
1016
|
`.json`, the file is loaded in JSON format. Otherwise, the file is loaded in ProtoBuf
|
|
838
1017
|
format.
|
|
839
|
-
Default: ''
|
|
1018
|
+
Default: ``''``
|
|
840
1019
|
|
|
841
1020
|
- save_file (str): The path to save parallel strategy checkpoint. If the file name extension is
|
|
842
1021
|
`.json`, the file is saved in JSON format. Otherwise, the file is saved in ProtoBuf format.
|
|
843
|
-
Default: ''
|
|
1022
|
+
Default: ``''``
|
|
844
1023
|
|
|
845
1024
|
- only_trainable_params (bool): Only save/load the strategy information for trainable parameter.
|
|
846
1025
|
Default: ``True`` .
|
|
1026
|
+
group_ckpt_save_file (str): The path to save parallel group checkpoint.
|
|
1027
|
+
auto_pipeline (bool): Set the pipeline stage number to automatic. Its value will be selected between 1 and the
|
|
1028
|
+
parameter `pipeline_stages`. This option requires the `parallel_mode` to be ``auto_parallel``
|
|
1029
|
+
and the `search_mode` to be ``recursive_programming``. Default: ``False`` .
|
|
847
1030
|
|
|
848
1031
|
Raises:
|
|
849
1032
|
ValueError: If input key is not attribute in auto parallel context.
|
|
@@ -855,8 +1038,8 @@ def set_auto_parallel_context(**kwargs):
|
|
|
855
1038
|
>>> ms.set_auto_parallel_context(gradients_mean=True)
|
|
856
1039
|
>>> ms.set_auto_parallel_context(gradient_fp32_sync=False)
|
|
857
1040
|
>>> ms.set_auto_parallel_context(parallel_mode="auto_parallel")
|
|
858
|
-
>>> ms.set_auto_parallel_context(search_mode="
|
|
859
|
-
>>> ms.set_auto_parallel_context(auto_parallel_search_mode="
|
|
1041
|
+
>>> ms.set_auto_parallel_context(search_mode="recursive_programming")
|
|
1042
|
+
>>> ms.set_auto_parallel_context(auto_parallel_search_mode="recursive_programming")
|
|
860
1043
|
>>> ms.set_auto_parallel_context(parameter_broadcast=False)
|
|
861
1044
|
>>> ms.set_auto_parallel_context(strategy_ckpt_load_file="./strategy_stage1.ckpt")
|
|
862
1045
|
>>> ms.set_auto_parallel_context(strategy_ckpt_save_file="./strategy_stage1.ckpt")
|
|
@@ -865,6 +1048,7 @@ def set_auto_parallel_context(**kwargs):
|
|
|
865
1048
|
>>> ms.set_auto_parallel_context(enable_alltoall=False)
|
|
866
1049
|
>>> ms.set_auto_parallel_context(all_reduce_fusion_config=[8, 160])
|
|
867
1050
|
>>> ms.set_auto_parallel_context(pipeline_stages=2)
|
|
1051
|
+
>>> ms.set_auto_parallel_context(pipeline_stages=2, pipeline_result_broadcast=True)
|
|
868
1052
|
>>> parallel_config = {"gradient_accumulation_shard": True, "parallel_optimizer_threshold": 24,
|
|
869
1053
|
... "optimizer_weight_shard_size": 2}
|
|
870
1054
|
>>> ms.set_auto_parallel_context(parallel_optimizer_config=parallel_config, enable_parallel_optimizer=True)
|
|
@@ -913,9 +1097,12 @@ def reset_auto_parallel_context():
|
|
|
913
1097
|
- strategy_ckpt_save_file: ''.
|
|
914
1098
|
- full_batch: False.
|
|
915
1099
|
- enable_parallel_optimizer: False.
|
|
1100
|
+
- force_fp32_communication: False
|
|
916
1101
|
- enable_alltoall: False.
|
|
917
1102
|
- pipeline_stages: 1.
|
|
1103
|
+
- pipeline_result_broadcast: False.
|
|
918
1104
|
- fusion_threshold: 64.
|
|
1105
|
+
- auto_pipeline: False.
|
|
919
1106
|
|
|
920
1107
|
Examples:
|
|
921
1108
|
>>> import mindspore as ms
|
|
@@ -1005,6 +1192,23 @@ def _check_target_specific_cfgs(device, arg_key):
|
|
|
1005
1192
|
return False
|
|
1006
1193
|
|
|
1007
1194
|
|
|
1195
|
+
def _check_ascend_device_context_initialized(device_target, settings):
|
|
1196
|
+
if device_target == 'Ascend' and is_initialized(device_target):
|
|
1197
|
+
for key, _ in settings.items():
|
|
1198
|
+
if key in ('ascend_config', 'deterministic', 'jit_compile', 'exception_dump', 'device_id'):
|
|
1199
|
+
logger.warning(f"For 'context.set_context' in Ascend backend, the backend is already initialized, "
|
|
1200
|
+
"please set it before the definition of any Tensor and Parameter, and the "
|
|
1201
|
+
"instantiation and execution of any operation and net, otherwise the settings may not "
|
|
1202
|
+
"take effect. ")
|
|
1203
|
+
break
|
|
1204
|
+
|
|
1205
|
+
|
|
1206
|
+
def _check_key(key):
|
|
1207
|
+
if key in ('precision_mode', 'jit_compile', 'atomic_clean_policy', 'matmul_allow_hf32', 'conv_allow_hf32',
|
|
1208
|
+
'op_precision_mode', 'host_scheduling_max_threshold', 'ge_options', 'op_debug_option'):
|
|
1209
|
+
raise ValueError(f"Please set '{key}' through parameter ascend_config")
|
|
1210
|
+
|
|
1211
|
+
|
|
1008
1212
|
@args_type_check(mode=int, precompile_only=bool, device_target=str, device_id=int, save_graphs=(bool, int),
|
|
1009
1213
|
save_graphs_path=str, enable_dump=bool, aoe_tune_mode=str, aoe_config=dict,
|
|
1010
1214
|
save_dump_path=str, enable_reduce_precision=bool, variable_memory_max_size=str,
|
|
@@ -1013,8 +1217,8 @@ def _check_target_specific_cfgs(device, arg_key):
|
|
|
1013
1217
|
max_device_memory=str, print_file_path=str, max_call_depth=int, env_config_path=str,
|
|
1014
1218
|
graph_kernel_flags=str, save_compile_cache=bool, runtime_num_threads=int, load_compile_cache=bool,
|
|
1015
1219
|
grad_for_scalar=bool, pynative_synchronize=bool, mempool_block_size=str, disable_format_transform=bool,
|
|
1016
|
-
op_timeout=int, deterministic=str, ascend_config=dict, jit_syntax_level=int,
|
|
1017
|
-
jit_enable_inplace_ops=bool, gpu_config=dict)
|
|
1220
|
+
op_timeout=int, deterministic=str, ascend_config=dict, jit_syntax_level=int, debug_level=int,
|
|
1221
|
+
jit_enable_inplace_ops=bool, gpu_config=dict, jit_config=dict, enable_compile_cache=bool)
|
|
1018
1222
|
def set_context(**kwargs):
|
|
1019
1223
|
"""
|
|
1020
1224
|
Set context for running environment.
|
|
@@ -1063,6 +1267,8 @@ def set_context(**kwargs):
|
|
|
1063
1267
|
| | reserve_class_name_in_scope | CPU/GPU/Ascend |
|
|
1064
1268
|
| +------------------------------+----------------------------+
|
|
1065
1269
|
| | pynative_synchronize | CPU/GPU/Ascend |
|
|
1270
|
+
| +------------------------------+----------------------------+
|
|
1271
|
+
| | debug_level | CPU/GPU/Ascend |
|
|
1066
1272
|
+-------------------------+------------------------------+----------------------------+
|
|
1067
1273
|
| Executive Control | mode | CPU/GPU/Ascend |
|
|
1068
1274
|
| +------------------------------+----------------------------+
|
|
@@ -1103,6 +1309,8 @@ def set_context(**kwargs):
|
|
|
1103
1309
|
| | jit_syntax_level | CPU/GPU/Ascend |
|
|
1104
1310
|
| +------------------------------+----------------------------+
|
|
1105
1311
|
| | gpu_config | GPU |
|
|
1312
|
+
| +------------------------------+----------------------------+
|
|
1313
|
+
| | jit_config | CPU/GPU/Ascend |
|
|
1106
1314
|
+-------------------------+------------------------------+----------------------------+
|
|
1107
1315
|
|
|
1108
1316
|
Args:
|
|
@@ -1115,12 +1323,16 @@ def set_context(**kwargs):
|
|
|
1115
1323
|
and max_device_memory. 'max_device_memory' should be set before the program runs.
|
|
1116
1324
|
variable_memory_max_size (str): This parameter is deprecated, and will be removed in a future version.
|
|
1117
1325
|
Please use parameter 'max_device_memory' instead.
|
|
1118
|
-
mempool_block_size (str): Set the size of the memory pool block in PyNative mode
|
|
1119
|
-
The format is "xxGB". Default: ``"1GB"`` . Minimum size is "1G". The actual used memory block
|
|
1120
|
-
minimum of the available memory of the device and mempool_block_size.
|
|
1326
|
+
mempool_block_size (str): Set the size of the memory pool block in PyNative mode or jit level is 'O0'/'O1'
|
|
1327
|
+
for devices. The format is "xxGB". Default: ``"1GB"`` . Minimum size is "1G". The actual used memory block
|
|
1328
|
+
size is the minimum of the available memory of the device and mempool_block_size.
|
|
1121
1329
|
op_timeout (int): Set the maximum duration of executing an operator in seconds.
|
|
1122
|
-
If the execution time exceeds this value, system will terminate the task.
|
|
1123
|
-
|
|
1330
|
+
If the execution time exceeds this value, system will terminate the task.
|
|
1331
|
+
0 means endless wait. The defaults for AI Core and AICPU operators vary on different hardware.
|
|
1332
|
+
For more information,
|
|
1333
|
+
please refer to `Ascend Community document about aclrtSetOpExecuteTimeOut
|
|
1334
|
+
<https://www.hiascend.com/document/detail/en/CANNCommunityEdition/600alphaX/infacldevg/aclcppdevg/aclcppdevg_03_0069.html>`_.
|
|
1335
|
+
Default: ``900`` .
|
|
1124
1336
|
save_graphs (bool or int): Whether to save intermediate compilation graphs. Default: ``0`` .
|
|
1125
1337
|
Available values are:
|
|
1126
1338
|
|
|
@@ -1129,10 +1341,13 @@ def set_context(**kwargs):
|
|
|
1129
1341
|
- True or 2: Generate more ir files related to backend process.
|
|
1130
1342
|
- 3: Generate visualization computing graphs and detailed frontend ir graphs.
|
|
1131
1343
|
|
|
1344
|
+
When the network structure is complex, setting `save_graphs` attribute to ``2`` or ``3`` may take too long.
|
|
1345
|
+
If you need quick problem locating, you can switch to ``1`` first.
|
|
1346
|
+
|
|
1132
1347
|
When the `save_graphs` attribute is set as ``True`` , ``1`` , ``2`` or ``3`` , attribute of
|
|
1133
1348
|
`save_graphs_path` is used to set the intermediate compilation graph storage path. By default, the graphs
|
|
1134
1349
|
are saved in the current directory.
|
|
1135
|
-
save_graphs_path (str): Path to save graphs. Default: "."
|
|
1350
|
+
save_graphs_path (str): Path to save graphs. Default: ``"."``.
|
|
1136
1351
|
If the specified directory does not exist, the system will automatically create the directory.
|
|
1137
1352
|
During distributed training, graphs will be saved to the directory of
|
|
1138
1353
|
`save_graphs_path/rank_${rank_id}/`. `rank_id` is the ID of the current device in the cluster.
|
|
@@ -1152,6 +1367,8 @@ def set_context(**kwargs):
|
|
|
1152
1367
|
If the saved file already exists, the timestamp suffix will be added to the file. Saving data to a file
|
|
1153
1368
|
solves the problem of data loss in screen printing when a large amount of data is generated.
|
|
1154
1369
|
If it is not set, an error will be reported: prompt to set the upper absolute path.
|
|
1370
|
+
When print data to file, the total output bytes of single print must be less then 2GB(limited by
|
|
1371
|
+
protobuf).
|
|
1155
1372
|
env_config_path (str): Config path for DFX.
|
|
1156
1373
|
Through mindspore.set_context(env_config_path="./mindspore_config.json")
|
|
1157
1374
|
|
|
@@ -1196,7 +1413,7 @@ def set_context(**kwargs):
|
|
|
1196
1413
|
If enable_graph_kernel is set to ``True`` , acceleration can be enabled.
|
|
1197
1414
|
For details of graph kernel fusion, please check
|
|
1198
1415
|
`Enabling Graph Kernel Fusion
|
|
1199
|
-
<https://www.mindspore.cn/tutorials/experts/en/
|
|
1416
|
+
<https://www.mindspore.cn/tutorials/experts/en/master/optimize/graph_fusion_engine.html>`_.
|
|
1200
1417
|
graph_kernel_flags (str):
|
|
1201
1418
|
Optimization options of graph kernel fusion, and the priority is higher when it conflicts
|
|
1202
1419
|
with enable_graph_kernel. Only for experienced users.
|
|
@@ -1252,7 +1469,7 @@ def set_context(**kwargs):
|
|
|
1252
1469
|
the compile cache is loaded. Note that only limited automatic detection for the changes of
|
|
1253
1470
|
python scripts is supported by now, which means that there is a correctness risk. Default: ``False`` .
|
|
1254
1471
|
This is an experimental prototype that is subject to change and/or deletion.
|
|
1255
|
-
compile_cache_path (str): Path to save the compile cache. Default: "."
|
|
1472
|
+
compile_cache_path (str): Path to save the compile cache. Default: ``"."``.
|
|
1256
1473
|
If the specified directory does not exist, the system will automatically create the directory.
|
|
1257
1474
|
The cache will be saved to the directory of `compile_cache_path/rank_${rank_id}/`. The `rank_id` is
|
|
1258
1475
|
the ID of the current device in the cluster.
|
|
@@ -1269,16 +1486,18 @@ def set_context(**kwargs):
|
|
|
1269
1486
|
of the interfaces would be compiled by MindSpore to the interfaces definition .py file that should be
|
|
1270
1487
|
guaranteed to be writable. Then compile the .py file to the .pyc or .so file, and could run in Graph mode.
|
|
1271
1488
|
memory_optimize_level (str): The memory optimize level.
|
|
1272
|
-
|
|
1489
|
+
On Ascend hardware platform, default: ``O1``, on other hardware platforms, default: ``O0``.
|
|
1490
|
+
The value must be in ['O0', 'O1'].
|
|
1273
1491
|
|
|
1274
|
-
- O0: priority performance option, disable SOMAS (Safe Optimized Memory Allocation Solver)
|
|
1275
|
-
|
|
1492
|
+
- O0: priority performance option, disable SOMAS (Safe Optimized Memory Allocation Solver)
|
|
1493
|
+
and some other memory optimizations.
|
|
1494
|
+
- O1: priority memory option, enable SOMAS and some other memory optimizations.
|
|
1276
1495
|
memory_offload (str): Whether to enable the memory offload function. When it is enabled, the idle data will be
|
|
1277
1496
|
temporarily copied to the host side in the case of insufficient device memory. The value must be in the
|
|
1278
1497
|
range of ['ON', 'OFF'], and the default value is ``'OFF'`` .
|
|
1279
1498
|
|
|
1280
1499
|
- ON: Enable the memory Offload function. On Ascend hardware platform, this parameter does not take effect
|
|
1281
|
-
when the
|
|
1500
|
+
when the graph compilation level is not 'O0'; This parameter does not take effect when
|
|
1282
1501
|
memory_optimize_level is set 'O1'.
|
|
1283
1502
|
- OFF: Turn off the memory Offload function.
|
|
1284
1503
|
ascend_config (dict): Set the parameters specific to Ascend hardware platform. It is not set by default.
|
|
@@ -1289,22 +1508,27 @@ def set_context(**kwargs):
|
|
|
1289
1508
|
is ``force_fp16`` . The value range is as follows:
|
|
1290
1509
|
|
|
1291
1510
|
- force_fp16: When the operator supports both float16 and float32, select float16 directly.
|
|
1292
|
-
- allow_fp32_to_fp16:
|
|
1293
|
-
the
|
|
1511
|
+
- allow_fp32_to_fp16: For cube operators, use the float16. For vector operators,
|
|
1512
|
+
prefer to keep the origin dtype, if the operator in model can support float32,
|
|
1513
|
+
it will keep original dtype, otherwise it will reduce to float16.
|
|
1294
1514
|
- allow_mix_precision: Automatic mixing precision, facing the whole network operator, according
|
|
1295
1515
|
to the built-in optimization strategy, automatically reduces the precision of some operators
|
|
1296
1516
|
to float16 or bfloat16.
|
|
1297
1517
|
- must_keep_origin_dtype: Keep the accuracy of the original drawing.
|
|
1298
1518
|
- force_fp32: When the input of the matrix calculation operator is float16 and the output supports
|
|
1299
1519
|
float16 and float32, output is forced to float32.
|
|
1300
|
-
- allow_fp32_to_bf16:
|
|
1301
|
-
the
|
|
1520
|
+
- allow_fp32_to_bf16: For cube operators, use the bfloat16. For vector operators,
|
|
1521
|
+
prefer to keep the origin dtype, if the operator in model can support float32,
|
|
1522
|
+
it will keep original dtype, otherwise it will reduce to bfloat16.
|
|
1302
1523
|
- allow_mix_precision_fp16: Automatic mixing precision, facing the whole network operator, automatically
|
|
1303
1524
|
reduces the precision of some operators to float16 according to the built-in optimization strategy.
|
|
1304
1525
|
- allow_mix_precision_bf16: Automatic mixing precision, facing the whole network operator, according to
|
|
1305
1526
|
the built-in optimization strategy, automatically reduces the precision of some operators to bfloat16.
|
|
1306
1527
|
|
|
1307
|
-
- jit_compile (bool): Whether to select online compilation.
|
|
1528
|
+
- jit_compile (bool): Whether to select online compilation. When set to 'True', online compilation is
|
|
1529
|
+
prioritized. When set to 'False', compiled operator binary files are prioritized to improve compilation
|
|
1530
|
+
performance. The default settings are online compilation for static shape, and compiled operator binary
|
|
1531
|
+
files for dynamic shape.
|
|
1308
1532
|
- atomic_clean_policy (int): The policy for cleaning memory occupied by atomic operators in the network.
|
|
1309
1533
|
Default: ``1`` .
|
|
1310
1534
|
|
|
@@ -1320,24 +1544,66 @@ def set_context(**kwargs):
|
|
|
1320
1544
|
For detailed information, please refer to `Ascend community <https://www.hiascend.com/>`_ .
|
|
1321
1545
|
- exception_dump (str): Enable exception dump for Ascend operators, providing the input and output data for
|
|
1322
1546
|
failing Ascend operators. The value can be ``"0"`` , ``"1"`` and ``"2"``. For ``"0"`` , exception dump is
|
|
1323
|
-
turned off; for ``"1"``, all inputs and outputs will be dumped for AICore
|
|
1324
|
-
for ``"2"``, inputs will be dumped for AICore exception operators
|
|
1547
|
+
turned off; for ``"1"``, all inputs and outputs will be dumped for AICore exception operators;
|
|
1548
|
+
for ``"2"``, inputs will be dumped for AICore exception operators, reducing the saved information
|
|
1549
|
+
but improving performance. Default: ``"2"`` .
|
|
1325
1550
|
- op_precision_mode (str): Path to config file of op precision mode. For detailed information, please refer
|
|
1326
1551
|
to `Ascend community <https://www.hiascend.com/>`_ .
|
|
1552
|
+
- op_debug_option (str): Enable debugging options for Ascend operators, default not enabled.
|
|
1553
|
+
The value currently only supports being set to ``"oom"``.
|
|
1554
|
+
|
|
1555
|
+
- ``"oom"``: When there is a memory out of bounds during the execution of an operator,
|
|
1556
|
+
AscendCL will return an error code of ``EZ9999``.
|
|
1557
|
+
|
|
1558
|
+
- ge_options (dict): Set options for CANN. The options are divided into two categories: global and session.
|
|
1559
|
+
This is an experimental prototype that is subject to change and/or deletion.
|
|
1560
|
+
For detailed information, please refer to `Ascend community <https://www.hiascend.com/document/detail/zh/canncommercial/70RC1/inferapplicationdev/graphdevg/atlasgeapi_07_0119.html>`_ .
|
|
1561
|
+
The configuration options in `ge_options` may be duplicated with the options in `ascend_config`. If the
|
|
1562
|
+
same configuration options are set in both `ascend_config` and `ge_options`, the one set in `ge_options`
|
|
1563
|
+
shall prevail.
|
|
1564
|
+
|
|
1565
|
+
- global (dict): Set global options.
|
|
1566
|
+
- session (dict): Set session options.
|
|
1567
|
+
|
|
1327
1568
|
- parallel_speed_up_json_path(Union[str, None]): The path to the parallel speed up json file, configuration
|
|
1328
1569
|
can refer to `parallel_speed_up.json
|
|
1329
|
-
<https://gitee.com/mindspore/mindspore/blob/
|
|
1570
|
+
<https://gitee.com/mindspore/mindspore/blob/master/config/parallel_speed_up.json>`_ .
|
|
1330
1571
|
If its value is None or '', it does not take effect. Default None.
|
|
1331
1572
|
|
|
1332
1573
|
- recompute_comm_overlap (bool): Enable overlap between recompute ops and communication ops if True.
|
|
1333
1574
|
Default: False.
|
|
1334
|
-
- matmul_grad_comm_overlap (bool): Enable overlap between
|
|
1575
|
+
- matmul_grad_comm_overlap (bool): Enable overlap between dw matmul and
|
|
1576
|
+
tensor parallel communication ops if True. Default: False.
|
|
1577
|
+
- recompute_allgather_overlap_fagrad (bool): Enable overlap between duplicated allgather by recomputing
|
|
1578
|
+
in sequence parallel and flashattentionscoregrad ops if True. Default: False.
|
|
1579
|
+
- enable_task_opt (bool): Enable communication fusion to optimize the number of communication operator
|
|
1580
|
+
tasks if True.
|
|
1335
1581
|
Default: False.
|
|
1336
|
-
-
|
|
1582
|
+
- enable_grad_comm_opt (bool): Enable overlap between dx ops and data parallel communication ops if True.
|
|
1583
|
+
Currently, do not support
|
|
1584
|
+
`LazyInline <https://www.mindspore.cn/docs/en/master/api_python/mindspore/mindspore.lazy_inline.html>`
|
|
1337
1585
|
Default: False.
|
|
1338
|
-
-
|
|
1339
|
-
|
|
1586
|
+
- enable_opt_shard_comm_opt (bool): Enable overlap between forward ops
|
|
1587
|
+
and optimizer parallel allgather communication if True. Currently, do not support
|
|
1588
|
+
`LazyInline <https://www.mindspore.cn/docs/en/master/api_python/mindspore/mindspore.lazy_inline.html>`
|
|
1340
1589
|
Default: False.
|
|
1590
|
+
- compute_communicate_fusion_level (int): Enable the fusion between compute and communicate.
|
|
1591
|
+
Default: ``0``.
|
|
1592
|
+
|
|
1593
|
+
- 0: Disable fusion.
|
|
1594
|
+
|
|
1595
|
+
- 1: Apply fusion to forward nodes.
|
|
1596
|
+
|
|
1597
|
+
- 2: Apply fusion to backward nodes.
|
|
1598
|
+
|
|
1599
|
+
- 3: Apply fusion to all nodes.
|
|
1600
|
+
- bias_add_comm_swap (bool): Enable node execution order swap communication operators and add operators
|
|
1601
|
+
if ``True``. Only 1-dimension bias node is supported. Default: ``False``.
|
|
1602
|
+
- host_scheduling_max_threshold(int): The max threshold to control whether the dynamic shape process is
|
|
1603
|
+
used when run the static graph, the default value is 0. When the number of operations in the static graph
|
|
1604
|
+
is less than the max threshold, this graph will be executed in dynamic shape process. In large model
|
|
1605
|
+
scenarios, this approach can save stream resources. If the number of operations in the static graph is
|
|
1606
|
+
greater than the maximum threshold, this graph will be executed in original static process.
|
|
1341
1607
|
|
|
1342
1608
|
jit_syntax_level (int): Set JIT syntax level for graph compiling, triggered by GRAPH_MODE and @jit decorator.
|
|
1343
1609
|
The value must be ``STRICT`` or ``LAX`` . Default: ``LAX`` . All levels support all backends.
|
|
@@ -1348,6 +1614,12 @@ def set_context(**kwargs):
|
|
|
1348
1614
|
affected and not optimal. Cannot be used for MindIR load and export due to some syntax that may not be
|
|
1349
1615
|
able to be exported.
|
|
1350
1616
|
|
|
1617
|
+
debug_level (int): Set config for debugging. Default value: ``RELEASE``.
|
|
1618
|
+
|
|
1619
|
+
- ``RELEASE``: Used for normally running, and some debug information will be discard to get a better
|
|
1620
|
+
compiling performance.
|
|
1621
|
+
- ``DEBUG``: Used for debugging when errors occur, more information will be record in compiling process.
|
|
1622
|
+
|
|
1351
1623
|
gpu_config (dict): Set the parameters specific to gpu hardware platform. It is not set by default.
|
|
1352
1624
|
Currently, only setting `conv_fprop_algo` and `conv_dgrad_algo` and `conv_wgrad_algo` and `conv_allow_tf32`
|
|
1353
1625
|
and `matmul_allow_tf32` are supported on GPU hardware platform.
|
|
@@ -1419,6 +1691,26 @@ def set_context(**kwargs):
|
|
|
1419
1691
|
- matmul_allow_tf32 (bool): The flag below controls to allow Tensor core TF32 computation on CUBLAS and the
|
|
1420
1692
|
default value is ``False``.
|
|
1421
1693
|
|
|
1694
|
+
jit_config (dict): Set the global jit config for compile, take effect in network defined in Cell or jit
|
|
1695
|
+
decorators. It is not set by default.
|
|
1696
|
+
The setting in context is the global jit config, while JitConfig is the local network's jit config.
|
|
1697
|
+
When both exist simultaneously, the global jit config will not overwrite the local network's jit config.
|
|
1698
|
+
|
|
1699
|
+
- jit_level (str): Used to control the compilation optimization level. Default: ``""`` , The framework
|
|
1700
|
+
automatically selects the execution method based on product, Altas training product is O2, and all other
|
|
1701
|
+
products are O0. The value range is as follows:
|
|
1702
|
+
|
|
1703
|
+
- ``"O0"``: Except for optimizations that may affect functionality, all other optimizations are turned
|
|
1704
|
+
off, adopt KernelByKernel execution mode.
|
|
1705
|
+
- ``"O1"``: Using commonly used optimizations and automatic operator fusion optimizations,
|
|
1706
|
+
adopt KernelByKernel execution mode.
|
|
1707
|
+
- ``"O2"``: Ultimate performance optimization, adopt Sink execution mode.
|
|
1708
|
+
|
|
1709
|
+
- infer_boost (str): Used to control the infer mode. Default: ``"off"`` . The value range is as follows:
|
|
1710
|
+
|
|
1711
|
+
- ``"on"``: Enable infer mode, get better infer performance.
|
|
1712
|
+
- ``"off"``: Disable infer mode, use forward to infer, performance is not good.
|
|
1713
|
+
|
|
1422
1714
|
Raises:
|
|
1423
1715
|
ValueError: If input key is not an attribute in context.
|
|
1424
1716
|
|
|
@@ -1452,16 +1744,23 @@ def set_context(**kwargs):
|
|
|
1452
1744
|
>>> ms.set_context(memory_offload='ON')
|
|
1453
1745
|
>>> ms.set_context(deterministic='ON')
|
|
1454
1746
|
>>> ms.set_context(ascend_config={"precision_mode": "force_fp16", "jit_compile": True,
|
|
1455
|
-
... "atomic_clean_policy": 1, "op_precision_mode": "./op_precision_config_file"
|
|
1747
|
+
... "atomic_clean_policy": 1, "op_precision_mode": "./op_precision_config_file",
|
|
1748
|
+
... "op_debug_option": "oom",
|
|
1749
|
+
... "ge_options": {"global": {"ge.opSelectImplmode": "high_precision"},
|
|
1750
|
+
... "session": {"ge.exec.atomicCleanPolicy": "0"}}})
|
|
1456
1751
|
>>> ms.set_context(jit_syntax_level=ms.STRICT)
|
|
1752
|
+
>>> ms.set_context(debug_level=ms.context.DEBUG)
|
|
1457
1753
|
>>> ms.set_context(gpu_config={"conv_fprop_algo": "performance", "conv_allow_tf32": True,
|
|
1458
1754
|
... "matmul_allow_tf32": True})
|
|
1755
|
+
>>> ms.set_context(jit_config={"jit_level": "O0"})
|
|
1459
1756
|
"""
|
|
1460
1757
|
ctx = _context()
|
|
1461
1758
|
# set device target first
|
|
1462
1759
|
if 'device_target' in kwargs:
|
|
1463
1760
|
ctx.set_device_target(kwargs['device_target'])
|
|
1464
1761
|
device = ctx.get_param(ms_ctx_param.device_target)
|
|
1762
|
+
_check_ascend_device_context_initialized(device, kwargs)
|
|
1763
|
+
|
|
1465
1764
|
for key, value in kwargs.items():
|
|
1466
1765
|
if key in ('enable_sparse', 'auto_tune_mode'):
|
|
1467
1766
|
logger.warning(f"For 'context.set_context', '{key}' parameter is deprecated, "
|
|
@@ -1471,9 +1770,7 @@ def set_context(**kwargs):
|
|
|
1471
1770
|
logger.warning(f"For 'context.set_context', '{key}' parameter is deprecated. "
|
|
1472
1771
|
"For details, please see the interface parameter API comments")
|
|
1473
1772
|
continue
|
|
1474
|
-
|
|
1475
|
-
'op_precision_mode'):
|
|
1476
|
-
raise ValueError(f"Please set '{key}' through parameter ascend_config")
|
|
1773
|
+
_check_key(key)
|
|
1477
1774
|
if key == 'save_graphs':
|
|
1478
1775
|
if value is True:
|
|
1479
1776
|
value = 2
|
|
@@ -1484,6 +1781,13 @@ def set_context(**kwargs):
|
|
|
1484
1781
|
if key == 'jit_syntax_level' and value not in (STRICT, COMPATIBLE, LAX):
|
|
1485
1782
|
raise ValueError(f"For 'jit_syntax_level', the value should be context.STRICT"
|
|
1486
1783
|
f" or context.LAX, but got {value}.")
|
|
1784
|
+
if key == 'debug_level' and value not in (RELEASE, DEBUG):
|
|
1785
|
+
raise ValueError(f"For 'debug_level', the value should be context.DEBUG"
|
|
1786
|
+
f" or context.RELEASE, but got {value}.")
|
|
1787
|
+
if key == 'enable_compile_cache':
|
|
1788
|
+
setattr(ctx, key, value)
|
|
1789
|
+
ctx.set_param(ms_ctx_param.__members__[key], int(value))
|
|
1790
|
+
continue
|
|
1487
1791
|
if not _check_target_specific_cfgs(device, key):
|
|
1488
1792
|
continue
|
|
1489
1793
|
if hasattr(ctx, key):
|
|
@@ -1541,6 +1845,17 @@ def _get_mode():
|
|
|
1541
1845
|
return ctx.get_mode()
|
|
1542
1846
|
|
|
1543
1847
|
|
|
1848
|
+
def get_jit_config():
|
|
1849
|
+
"""
|
|
1850
|
+
Get global jit config.
|
|
1851
|
+
|
|
1852
|
+
Returns:
|
|
1853
|
+
Object: The Value of jit config.
|
|
1854
|
+
"""
|
|
1855
|
+
ctx = _context()
|
|
1856
|
+
return ctx.get_jit_config()
|
|
1857
|
+
|
|
1858
|
+
|
|
1544
1859
|
class ParallelMode:
|
|
1545
1860
|
"""
|
|
1546
1861
|
Parallel mode options.
|
|
@@ -1638,9 +1953,7 @@ def get_ps_context(attr_key):
|
|
|
1638
1953
|
|
|
1639
1954
|
def reset_ps_context():
|
|
1640
1955
|
"""
|
|
1641
|
-
Reset parameter server training mode context attributes to the default values
|
|
1642
|
-
|
|
1643
|
-
- enable_ps: False.
|
|
1956
|
+
Reset parameter server training mode context attributes to the default values.
|
|
1644
1957
|
|
|
1645
1958
|
Meaning of each field and its default value refer to :func:`mindspore.set_ps_context`.
|
|
1646
1959
|
|