mindspore 1.10.0__cp38-cp38-win_amd64.whl → 2.0.0rc1__cp38-cp38-win_amd64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of mindspore might be problematic. Click here for more details.
- mindspore/.commit_id +1 -1
- mindspore/ConcurrencyCheck.dll +0 -0
- mindspore/CppBuildInsights.dll +0 -0
- mindspore/CppCoreCheck.dll +0 -0
- mindspore/EnumIndex.dll +0 -0
- mindspore/EspXEngine.dll +0 -0
- mindspore/HResultCheck.dll +0 -0
- mindspore/KernelTraceControl.dll +0 -0
- mindspore/LocalESPC.dll +0 -0
- mindspore/Microsoft.Diagnostics.Tracing.EventSource.dll +0 -0
- mindspore/Microsoft.VisualStudio.RemoteControl.dll +0 -0
- mindspore/Microsoft.VisualStudio.Telemetry.dll +0 -0
- mindspore/Microsoft.VisualStudio.Utilities.Internal.dll +0 -0
- mindspore/Newtonsoft.Json.dll +0 -0
- mindspore/System.Runtime.CompilerServices.Unsafe.dll +0 -0
- mindspore/VariantClear.dll +0 -0
- mindspore/__init__.py +9 -4
- mindspore/_c_dataengine.cp38-win_amd64.pyd +0 -0
- mindspore/_c_expression.cp38-win_amd64.pyd +0 -0
- mindspore/_c_mindrecord.cp38-win_amd64.pyd +0 -0
- mindspore/_check_jit_forbidden_api.py +102 -0
- mindspore/_checkparam.py +1066 -1001
- mindspore/_extends/builtin_operations.py +32 -4
- mindspore/_extends/graph_kernel/model/graph_split.py +66 -222
- mindspore/_extends/parallel_compile/akg_compiler/akg_process.py +12 -9
- mindspore/_extends/parallel_compile/akg_compiler/build_tbe_kernel.py +119 -26
- mindspore/_extends/parallel_compile/akg_compiler/tbe_topi.py +50 -50
- mindspore/_extends/parallel_compile/akg_compiler/util.py +9 -6
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_adapter.py +4 -25
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_helper.py +9 -4
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_job_manager.py +1 -27
- mindspore/_extends/parse/__init__.py +5 -3
- mindspore/_extends/parse/namespace.py +17 -2
- mindspore/_extends/parse/parser.py +193 -34
- mindspore/_extends/parse/resources.py +7 -8
- mindspore/_extends/parse/standard_method.py +1780 -435
- mindspore/_extends/parse/trope.py +3 -1
- mindspore/amp.py +53 -58
- mindspore/atlprov.dll +0 -0
- mindspore/boost/adasum.py +3 -2
- mindspore/boost/boost.py +2 -2
- mindspore/boost/boost_cell_wrapper.py +46 -26
- mindspore/boost/dim_reduce.py +6 -5
- mindspore/boost/grad_accumulation.py +2 -1
- mindspore/boost/group_loss_scale_manager.py +1 -1
- mindspore/c1.dll +0 -0
- mindspore/c1xx.dll +0 -0
- mindspore/c2.dll +0 -0
- mindspore/cfgpersist.dll +0 -0
- mindspore/clang_rt.asan_dbg_dynamic-x86_64.dll +0 -0
- mindspore/clang_rt.asan_dynamic-x86_64.dll +0 -0
- mindspore/common/__init__.py +11 -10
- mindspore/common/_decorator.py +2 -0
- mindspore/common/_register_for_adapter.py +55 -0
- mindspore/common/_stub_tensor.py +201 -0
- mindspore/common/_utils.py +57 -0
- mindspore/common/api.py +582 -297
- mindspore/common/dtype.py +66 -18
- mindspore/common/dump.py +2 -2
- mindspore/common/initializer.py +38 -1
- mindspore/common/jit_config.py +25 -13
- mindspore/common/mutable.py +53 -24
- mindspore/common/parameter.py +60 -37
- mindspore/common/seed.py +8 -24
- mindspore/common/sparse_tensor.py +927 -0
- mindspore/common/tensor.py +1627 -3900
- mindspore/communication/__init__.py +10 -5
- mindspore/communication/_comm_helper.py +78 -214
- mindspore/communication/_hccl_management.py +2 -1
- mindspore/communication/management.py +136 -47
- mindspore/config/op_info.config +501 -1008
- mindspore/context.py +291 -56
- mindspore/d3dcompiler_47.dll +0 -0
- mindspore/dataset/__init__.py +12 -8
- mindspore/dataset/audio/__init__.py +9 -9
- mindspore/dataset/audio/transforms.py +1090 -228
- mindspore/dataset/audio/utils.py +87 -39
- mindspore/dataset/audio/validators.py +223 -1
- mindspore/dataset/callback/ds_callback.py +17 -15
- mindspore/dataset/core/config.py +246 -17
- mindspore/dataset/core/py_util_helpers.py +4 -3
- mindspore/dataset/core/validator_helpers.py +10 -10
- mindspore/{parallel/nn/layers.py → dataset/debug/__init__.py} +7 -8
- mindspore/dataset/debug/debug_hook.py +65 -0
- mindspore/dataset/debug/pre_defined_hook.py +67 -0
- mindspore/dataset/engine/__init__.py +7 -3
- mindspore/dataset/engine/cache_client.py +9 -9
- mindspore/dataset/engine/datasets.py +648 -477
- mindspore/dataset/engine/datasets_audio.py +165 -167
- mindspore/dataset/engine/datasets_standard_format.py +93 -67
- mindspore/dataset/engine/datasets_text.py +492 -342
- mindspore/dataset/engine/datasets_user_defined.py +85 -50
- mindspore/dataset/engine/datasets_vision.py +1224 -699
- mindspore/dataset/engine/graphdata.py +134 -69
- mindspore/dataset/engine/iterators.py +50 -9
- mindspore/dataset/engine/offload.py +52 -31
- mindspore/dataset/engine/samplers.py +27 -24
- mindspore/dataset/engine/serializer_deserializer.py +14 -15
- mindspore/dataset/engine/validators.py +213 -52
- mindspore/dataset/text/__init__.py +10 -8
- mindspore/dataset/text/transforms.py +152 -57
- mindspore/dataset/text/utils.py +98 -49
- mindspore/dataset/text/validators.py +25 -0
- mindspore/dataset/transforms/__init__.py +4 -2
- mindspore/dataset/transforms/c_transforms.py +11 -13
- mindspore/dataset/transforms/py_transforms.py +2 -2
- mindspore/dataset/transforms/py_transforms_util.py +10 -0
- mindspore/dataset/transforms/transforms.py +13 -15
- mindspore/dataset/transforms/validators.py +7 -7
- mindspore/dataset/utils/__init__.py +2 -1
- mindspore/dataset/utils/browse_dataset.py +13 -13
- mindspore/dataset/utils/line_reader.py +121 -0
- mindspore/dataset/vision/__init__.py +8 -7
- mindspore/dataset/vision/c_transforms.py +125 -126
- mindspore/dataset/vision/py_transforms.py +37 -37
- mindspore/dataset/vision/py_transforms_util.py +23 -20
- mindspore/dataset/vision/transforms.py +316 -315
- mindspore/dataset/vision/utils.py +313 -17
- mindspore/dataset/vision/validators.py +6 -6
- mindspore/default_config.py +0 -1
- mindspore/dpcmi.dll +0 -0
- mindspore/{compression → experimental}/__init__.py +6 -5
- mindspore/experimental/map_parameter.py +275 -0
- mindspore/include/OWNERS +0 -1
- mindspore/include/api/callback/callback.h +9 -13
- mindspore/include/api/callback/ckpt_saver.h +2 -2
- mindspore/include/api/callback/loss_monitor.h +2 -2
- mindspore/include/api/callback/lr_scheduler.h +5 -5
- mindspore/include/api/callback/time_monitor.h +2 -2
- mindspore/include/api/callback/train_accuracy.h +4 -6
- mindspore/include/api/cfg.h +19 -6
- mindspore/include/api/context.h +70 -9
- mindspore/include/api/delegate.h +8 -1
- mindspore/include/api/dual_abi_helper.h +8 -24
- mindspore/include/api/metrics/accuracy.h +2 -2
- mindspore/include/api/metrics/metrics.h +4 -3
- mindspore/include/api/model.h +9 -4
- mindspore/include/api/model_group.h +68 -0
- mindspore/include/api/model_parallel_runner.h +17 -17
- mindspore/include/api/net.h +12 -11
- mindspore/include/api/serialization.h +20 -4
- mindspore/include/api/status.h +7 -1
- mindspore/include/api/types.h +25 -21
- mindspore/include/api/visible.h +4 -0
- mindspore/include/c_api/model_c.h +5 -0
- mindspore/include/c_api/status_c.h +1 -1
- mindspore/include/dataset/config.h +1 -1
- mindspore/include/dataset/constants.h +14 -0
- mindspore/include/dataset/text.h +59 -0
- mindspore/include/dataset/vision.h +56 -117
- mindspore/include/dataset/vision_lite.h +102 -0
- mindspore/jpeg62.dll +0 -0
- mindspore/log.py +28 -28
- mindspore/mindrecord/common/exceptions.py +2 -4
- mindspore/mindrecord/filereader.py +19 -1
- mindspore/mindrecord/filewriter.py +250 -88
- mindspore/mindrecord/mindpage.py +13 -13
- mindspore/mindrecord/shardheader.py +15 -15
- mindspore/mindrecord/shardreader.py +9 -0
- mindspore/mindrecord/shardwriter.py +29 -29
- mindspore/mindrecord/tools/cifar100_to_mr.py +9 -9
- mindspore/mindrecord/tools/cifar10_to_mr.py +9 -9
- mindspore/mindrecord/tools/csv_to_mr.py +4 -4
- mindspore/mindrecord/tools/imagenet_to_mr.py +70 -65
- mindspore/mindrecord/tools/mnist_to_mr.py +41 -41
- mindspore/mindrecord/tools/tfrecord_to_mr.py +6 -6
- mindspore/{libmindspore_backend.dll → mindspore_backend.dll} +0 -0
- mindspore/mindspore_common.dll +0 -0
- mindspore/mindspore_core.dll +0 -0
- mindspore/mindspore_glog.dll +0 -0
- mindspore/mindspore_shared_lib.dll +0 -0
- mindspore/msobj140.dll +0 -0
- mindspore/mspdb140.dll +0 -0
- mindspore/mspdbcore.dll +0 -0
- mindspore/mspdbst.dll +0 -0
- mindspore/mspft140.dll +0 -0
- mindspore/msvcdis140.dll +0 -0
- mindspore/msvcp140_1.dll +0 -0
- mindspore/msvcp140_2.dll +0 -0
- mindspore/msvcp140_atomic_wait.dll +0 -0
- mindspore/msvcp140_codecvt_ids.dll +0 -0
- mindspore/nn/__init__.py +1 -5
- mindspore/nn/cell.py +297 -234
- mindspore/nn/dynamic_lr.py +1 -1
- mindspore/nn/grad/cell_grad.py +17 -42
- mindspore/nn/layer/__init__.py +7 -4
- mindspore/nn/layer/activation.py +131 -88
- mindspore/nn/layer/basic.py +313 -613
- mindspore/nn/layer/channel_shuffle.py +103 -0
- mindspore/nn/layer/combined.py +1 -1
- mindspore/nn/layer/container.py +52 -6
- mindspore/nn/layer/conv.py +112 -43
- mindspore/nn/layer/dense.py +10 -9
- mindspore/nn/layer/embedding.py +36 -34
- mindspore/nn/layer/image.py +123 -27
- mindspore/nn/layer/math.py +108 -107
- mindspore/nn/layer/normalization.py +212 -366
- mindspore/nn/layer/padding.py +370 -42
- mindspore/nn/layer/pooling.py +1443 -219
- mindspore/nn/layer/rnn_cells.py +11 -16
- mindspore/nn/layer/rnns.py +38 -39
- mindspore/nn/layer/thor_layer.py +24 -25
- mindspore/nn/layer/timedistributed.py +5 -5
- mindspore/nn/layer/transformer.py +701 -0
- mindspore/nn/learning_rate_schedule.py +8 -8
- mindspore/nn/loss/__init__.py +9 -6
- mindspore/nn/loss/loss.py +678 -142
- mindspore/nn/metrics.py +53 -0
- mindspore/nn/optim/_dist_optimizer_registry.py +2 -2
- mindspore/nn/optim/ada_grad.py +8 -8
- mindspore/nn/optim/adadelta.py +2 -3
- mindspore/nn/optim/adafactor.py +18 -14
- mindspore/nn/optim/adam.py +429 -87
- mindspore/nn/optim/adamax.py +5 -6
- mindspore/nn/optim/adasum.py +10 -8
- mindspore/nn/optim/asgd.py +7 -7
- mindspore/nn/optim/ftrl.py +81 -11
- mindspore/nn/optim/lamb.py +7 -8
- mindspore/nn/optim/lars.py +4 -4
- mindspore/nn/optim/lazyadam.py +82 -7
- mindspore/nn/optim/momentum.py +8 -7
- mindspore/nn/optim/optimizer.py +19 -10
- mindspore/nn/optim/proximal_ada_grad.py +6 -5
- mindspore/nn/optim/rmsprop.py +3 -3
- mindspore/nn/optim/rprop.py +20 -16
- mindspore/nn/optim/sgd.py +21 -15
- mindspore/nn/optim/thor.py +23 -21
- mindspore/nn/probability/__init__.py +0 -2
- mindspore/nn/probability/bijector/bijector.py +7 -6
- mindspore/nn/probability/bijector/invert.py +4 -2
- mindspore/nn/probability/bijector/softplus.py +2 -2
- mindspore/nn/probability/bnn_layers/dense_variational.py +1 -1
- mindspore/nn/probability/bnn_layers/layer_distribution.py +2 -2
- mindspore/nn/probability/distribution/__init__.py +6 -0
- mindspore/nn/probability/distribution/_utils/custom_ops.py +3 -2
- mindspore/nn/probability/distribution/_utils/utils.py +11 -17
- mindspore/nn/probability/distribution/bernoulli.py +6 -6
- mindspore/nn/probability/distribution/beta.py +1 -1
- mindspore/nn/probability/distribution/categorical.py +9 -9
- mindspore/nn/probability/distribution/cauchy.py +8 -8
- mindspore/nn/probability/distribution/distribution.py +12 -6
- mindspore/nn/probability/distribution/exponential.py +5 -5
- mindspore/nn/probability/distribution/gamma.py +3 -3
- mindspore/nn/probability/distribution/geometric.py +6 -5
- mindspore/nn/probability/distribution/gumbel.py +5 -5
- mindspore/nn/probability/distribution/half_normal.py +133 -0
- mindspore/nn/probability/distribution/laplace.py +128 -0
- mindspore/nn/probability/distribution/log_normal.py +0 -1
- mindspore/nn/probability/distribution/logistic.py +4 -5
- mindspore/nn/probability/distribution/normal.py +11 -15
- mindspore/nn/probability/distribution/poisson.py +6 -2
- mindspore/nn/probability/distribution/student_t.py +150 -0
- mindspore/nn/probability/distribution/transformed_distribution.py +4 -4
- mindspore/nn/probability/distribution/uniform.py +5 -5
- mindspore/nn/reinforcement/_tensors_queue.py +3 -3
- mindspore/nn/reinforcement/tensor_array.py +2 -2
- mindspore/nn/sparse/sparse.py +8 -1
- mindspore/nn/wrap/cell_wrapper.py +55 -27
- mindspore/nn/wrap/grad_reducer.py +20 -11
- mindspore/nn/wrap/loss_scale.py +47 -30
- mindspore/numpy/array_creations.py +33 -22
- mindspore/numpy/array_ops.py +46 -42
- mindspore/numpy/logic_ops.py +6 -27
- mindspore/numpy/math_ops.py +26 -19
- mindspore/numpy/utils.py +1 -8
- mindspore/numpy/utils_const.py +112 -62
- mindspore/opencv_core452.dll +0 -0
- mindspore/opencv_imgcodecs452.dll +0 -0
- mindspore/opencv_imgproc452.dll +0 -0
- mindspore/ops/__init__.py +6 -3
- mindspore/ops/_constants.py +0 -6
- mindspore/ops/_grad/__init__.py +2 -1
- mindspore/ops/_grad/grad_array_ops.py +209 -152
- mindspore/ops/_grad/grad_base.py +55 -17
- mindspore/ops/_grad/grad_clip_ops.py +11 -3
- mindspore/ops/_grad/grad_comm_ops.py +58 -47
- mindspore/ops/_grad/grad_implementations.py +21 -61
- mindspore/ops/_grad/grad_inner_ops.py +48 -6
- mindspore/ops/_grad/grad_math_ops.py +306 -161
- mindspore/ops/_grad/grad_nn_ops.py +192 -181
- mindspore/ops/_grad/grad_other_ops.py +1 -1
- mindspore/ops/_grad/grad_quant_ops.py +5 -5
- mindspore/ops/_grad/grad_sequence_ops.py +296 -0
- mindspore/ops/_grad/grad_sparse.py +15 -9
- mindspore/ops/_grad_experimental/__init__.py +1 -0
- mindspore/ops/_grad_experimental/grad_array_ops.py +441 -55
- mindspore/ops/_grad_experimental/grad_image_ops.py +25 -7
- mindspore/ops/_grad_experimental/grad_inner_ops.py +3 -44
- mindspore/ops/_grad_experimental/grad_linalg_ops.py +16 -21
- mindspore/ops/_grad_experimental/grad_math_ops.py +979 -49
- mindspore/ops/_grad_experimental/grad_nn_ops.py +78 -8
- mindspore/ops/_grad_experimental/grad_scalar_ops.py +112 -0
- mindspore/ops/_grad_experimental/grad_sparse_ops.py +197 -13
- mindspore/ops/_op_impl/__init__.py +3 -3
- mindspore/ops/_op_impl/_custom_op/__init__.py +0 -1
- mindspore/ops/_op_impl/_custom_op/_basic.py +0 -1
- mindspore/ops/_op_impl/_custom_op/batch_matmul_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/batchnorm_fold.py +4 -2
- mindspore/ops/_op_impl/_custom_op/batchnorm_fold2.py +2 -2
- mindspore/ops/_op_impl/_custom_op/batchnorm_fold2_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/batchnorm_fold2_grad_reduce.py +5 -5
- mindspore/ops/_op_impl/_custom_op/batchnorm_fold_grad.py +3 -3
- mindspore/ops/_op_impl/_custom_op/cholesky_trsm_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/correction_mul.py +3 -3
- mindspore/ops/_op_impl/_custom_op/correction_mul_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/dsd_back_impl.py +4 -8
- mindspore/ops/_op_impl/_custom_op/dsd_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perchannel.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perchannel_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perchannel_grad_reduce.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perlayer.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perlayer_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perlayer_grad_reduce.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_quant_perchannel.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_quant_perchannel_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_quant_perlayer.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_quant_perlayer_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fused_abs_max1_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/img2col_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/matmul_cube_dense_left_impl.py +2 -2
- mindspore/ops/_op_impl/_custom_op/matmul_cube_dense_right_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/matmul_cube_fracz_left_cast_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/matmul_cube_fracz_right_mul_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/matmul_cube_impl.py +2 -2
- mindspore/ops/_op_impl/_custom_op/matmul_dds_grad_impl.py +0 -1
- mindspore/ops/_op_impl/_custom_op/matmul_dds_impl.py +0 -1
- mindspore/ops/_op_impl/_custom_op/matrix_combine_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/minmax_update_perchannel.py +2 -2
- mindspore/ops/_op_impl/_custom_op/minmax_update_perlayer.py +2 -2
- mindspore/ops/_op_impl/_custom_op/transpose02314_impl.py +1 -1
- mindspore/ops/_op_impl/aicpu/__init__.py +238 -3
- mindspore/ops/_op_impl/aicpu/abs.py +36 -0
- mindspore/ops/_op_impl/aicpu/adaptive_avg_pool_2d.py +34 -0
- mindspore/ops/_op_impl/aicpu/adaptive_avg_pool_2d_grad.py +34 -0
- mindspore/ops/_op_impl/aicpu/adaptive_avg_pool_3d.py +39 -0
- mindspore/ops/_op_impl/aicpu/adaptive_avg_pool_3d_grad.py +39 -0
- mindspore/ops/_op_impl/aicpu/adaptive_max_pool_2d_grad.py +37 -0
- mindspore/ops/_op_impl/aicpu/adaptive_max_pool_3d.py +42 -0
- mindspore/ops/_op_impl/aicpu/adaptive_max_pool_3d_grad.py +152 -0
- mindspore/ops/_op_impl/aicpu/add.py +43 -0
- mindspore/ops/_op_impl/aicpu/addcdiv.py +0 -32
- mindspore/ops/_op_impl/aicpu/addcmul.py +0 -84
- mindspore/ops/_op_impl/aicpu/affine_grid_grad.py +35 -0
- mindspore/ops/_op_impl/aicpu/arg_max.py +75 -0
- mindspore/ops/_op_impl/aicpu/arg_min.py +75 -0
- mindspore/ops/_op_impl/aicpu/argmin_with_value.py +43 -0
- mindspore/ops/_op_impl/aicpu/batch_matmul.py +43 -0
- mindspore/ops/_op_impl/aicpu/batch_norm_grad_grad.py +49 -0
- mindspore/ops/_op_impl/aicpu/bernoulli.py +48 -0
- mindspore/ops/_op_impl/aicpu/bessel_i0.py +31 -0
- mindspore/ops/_op_impl/aicpu/bias_add.py +44 -0
- mindspore/ops/_op_impl/aicpu/bias_add_grad.py +43 -0
- mindspore/ops/_op_impl/aicpu/bincount.py +33 -0
- mindspore/{nn/probability/infer/variational/__init__.py → ops/_op_impl/aicpu/cauchy.py} +17 -10
- mindspore/ops/_op_impl/aicpu/channel_shuffle.py +40 -0
- mindspore/ops/_op_impl/aicpu/cholesky.py +1 -1
- mindspore/ops/_op_impl/{cpu/bias_add.py → aicpu/choleskygrad.py} +9 -7
- mindspore/ops/_op_impl/aicpu/combined_non_max_suppression.py +42 -0
- mindspore/ops/_op_impl/aicpu/concat_offset.py +42 -0
- mindspore/ops/_op_impl/aicpu/concat_offset_v1.py +31 -0
- mindspore/ops/_op_impl/aicpu/conj.py +11 -0
- mindspore/ops/_op_impl/aicpu/crop_and_resize_grad_image.py +38 -0
- mindspore/ops/_op_impl/aicpu/cumulative_logsumexp.py +36 -0
- mindspore/ops/_op_impl/aicpu/deformable_offsets.py +38 -0
- mindspore/ops/_op_impl/aicpu/deformable_offsets_grad.py +2 -2
- mindspore/ops/_op_impl/aicpu/dense_to_sparse_set_operation.py +48 -0
- mindspore/ops/_op_impl/aicpu/diag.py +36 -0
- mindspore/ops/_op_impl/aicpu/diag_part.py +36 -0
- mindspore/ops/_op_impl/aicpu/diagonal.py +35 -0
- mindspore/ops/_op_impl/{cpu/bias_add_grad.py → aicpu/digamma.py} +9 -7
- mindspore/ops/_op_impl/aicpu/eig.py +35 -0
- mindspore/ops/_op_impl/aicpu/fft_with_size.py +41 -0
- mindspore/ops/_op_impl/aicpu/flatten.py +1 -0
- mindspore/ops/_op_impl/aicpu/fmax.py +36 -0
- mindspore/ops/_op_impl/aicpu/fmin.py +37 -0
- mindspore/ops/_op_impl/aicpu/fractional_max_pool3d_with_fixed_ksize.py +1 -1
- mindspore/ops/_op_impl/aicpu/fse_decode.py +43 -0
- mindspore/ops/_op_impl/aicpu/glu.py +33 -0
- mindspore/ops/_op_impl/aicpu/glu_grad.py +34 -0
- mindspore/ops/_op_impl/aicpu/greater.py +41 -0
- mindspore/ops/_op_impl/aicpu/greater_equal.py +41 -0
- mindspore/ops/_op_impl/aicpu/index_put.py +50 -0
- mindspore/ops/_op_impl/{tbe/scatter_add_ds.py → aicpu/inplace_index_add.py} +17 -21
- mindspore/ops/_op_impl/aicpu/instance_norm_v2.py +41 -0
- mindspore/ops/_op_impl/aicpu/instance_norm_v2_grad.py +44 -0
- mindspore/ops/_op_impl/aicpu/layer_norm_grad_grad.py +47 -0
- mindspore/ops/_op_impl/aicpu/less.py +41 -0
- mindspore/ops/_op_impl/aicpu/less_equal.py +41 -0
- mindspore/ops/_op_impl/aicpu/lgamma.py +32 -0
- mindspore/ops/_op_impl/aicpu/log_normal_reverse.py +33 -0
- mindspore/ops/_op_impl/aicpu/logit.py +33 -0
- mindspore/ops/_op_impl/aicpu/logit_grad.py +34 -0
- mindspore/ops/_op_impl/aicpu/masked_fill.py +42 -0
- mindspore/ops/_op_impl/aicpu/masked_scatter.py +39 -0
- mindspore/ops/_op_impl/aicpu/matmul.py +39 -0
- mindspore/ops/_op_impl/aicpu/matrix_logarithm.py +31 -0
- mindspore/ops/_op_impl/aicpu/matrix_power.py +32 -0
- mindspore/ops/_op_impl/aicpu/matrix_solve_ls.py +36 -0
- mindspore/ops/_op_impl/aicpu/matrix_triangular_solve.py +36 -0
- mindspore/ops/_op_impl/aicpu/mirror_pad.py +2 -0
- mindspore/ops/_op_impl/aicpu/mirror_pad_grad.py +0 -4
- mindspore/ops/_op_impl/aicpu/mul.py +3 -1
- mindspore/ops/_op_impl/aicpu/multinomial.py +14 -6
- mindspore/ops/_op_impl/aicpu/multinomial_with_replacement.py +35 -0
- mindspore/ops/_op_impl/aicpu/nan_to_num.py +34 -0
- mindspore/ops/_op_impl/aicpu/nllloss.py +38 -0
- mindspore/ops/_op_impl/aicpu/nllloss_grad.py +39 -0
- mindspore/ops/_op_impl/aicpu/ones_like.py +0 -2
- mindspore/ops/_op_impl/aicpu/polar.py +32 -0
- mindspore/ops/_op_impl/aicpu/polygamma.py +34 -0
- mindspore/ops/_op_impl/aicpu/qr.py +36 -0
- mindspore/ops/_op_impl/aicpu/quant_dtype_cast.py +40 -0
- mindspore/ops/_op_impl/aicpu/quantile.py +35 -0
- mindspore/ops/_op_impl/aicpu/ragged_tensor_to_sparse.py +73 -0
- mindspore/ops/_op_impl/aicpu/ragged_tensor_to_tensor.py +74 -0
- mindspore/ops/_op_impl/aicpu/random_shuffle.py +3 -0
- mindspore/ops/_op_impl/aicpu/randperm_v2.py +41 -0
- mindspore/ops/_op_impl/aicpu/range.py +36 -0
- mindspore/ops/_op_impl/aicpu/reciprocal.py +34 -0
- mindspore/ops/_op_impl/aicpu/reciprocal_grad.py +35 -0
- mindspore/ops/_op_impl/aicpu/reduce_sum.py +57 -0
- mindspore/ops/_op_impl/aicpu/resize_bicubic.py +2 -8
- mindspore/ops/_op_impl/aicpu/resize_bicubic_grad.py +1 -1
- mindspore/ops/_op_impl/aicpu/resize_v2.py +68 -0
- mindspore/ops/_op_impl/aicpu/resize_v2_grad.py +68 -0
- mindspore/ops/_op_impl/aicpu/scatter_elements.py +4 -0
- mindspore/ops/_op_impl/aicpu/scatter_nd_update.py +2 -0
- mindspore/ops/_op_impl/aicpu/search_sorted.py +12 -6
- mindspore/ops/_op_impl/aicpu/self_adjoint_eig.py +34 -0
- mindspore/ops/_op_impl/aicpu/sequence_add.py +34 -0
- mindspore/ops/_op_impl/aicpu/sequence_add_offset.py +34 -0
- mindspore/ops/_op_impl/aicpu/sequence_addn.py +38 -0
- mindspore/ops/_op_impl/aicpu/slice_grad.py +76 -0
- mindspore/ops/_op_impl/aicpu/smooth_l1_loss.py +35 -0
- mindspore/ops/_op_impl/aicpu/smooth_l1_loss_grad.py +37 -0
- mindspore/ops/_op_impl/aicpu/sort.py +39 -0
- mindspore/ops/_op_impl/aicpu/sparse_apply_adagrad_da.py +0 -24
- mindspore/ops/_op_impl/aicpu/sparse_cross.py +42 -0
- mindspore/ops/_op_impl/aicpu/sparse_fill_empty_rows.py +63 -0
- mindspore/ops/_op_impl/aicpu/sparse_fill_empty_rows_grad.py +45 -0
- mindspore/ops/_op_impl/aicpu/sparse_matrix_mat_mul.py +56 -0
- mindspore/ops/_op_impl/{tbe/slice_ds.py → aicpu/sparse_segment_sum.py} +16 -24
- mindspore/ops/_op_impl/aicpu/sparse_segment_sum_with_num_segments.py +68 -0
- mindspore/ops/_op_impl/aicpu/sparse_slice.py +63 -0
- mindspore/ops/_op_impl/aicpu/sparse_slice_grad.py +61 -0
- mindspore/ops/_op_impl/aicpu/squared_difference.py +2 -0
- mindspore/ops/_op_impl/aicpu/strided_slice_v2.py +93 -0
- mindspore/ops/_op_impl/aicpu/strided_slice_v2_grad.py +66 -0
- mindspore/ops/_op_impl/aicpu/tensor_scatter_update.py +59 -0
- mindspore/ops/_op_impl/{tbe/gather_v2.py → aicpu/tile.py} +24 -24
- mindspore/ops/_op_impl/aicpu/tridiagonal_solve.py +35 -0
- mindspore/ops/_op_impl/aicpu/tril_indices.py +34 -0
- mindspore/ops/_op_impl/aicpu/triu_indices.py +34 -0
- mindspore/ops/_op_impl/aicpu/uniform.py +34 -0
- mindspore/ops/_op_impl/aicpu/uniform_candidate_sampler.py +1 -0
- mindspore/ops/_op_impl/aicpu/unique_consecutive.py +10 -2
- mindspore/ops/_op_impl/cpu/__init__.py +1 -2
- mindspore/ops/_op_impl/cpu/dynamic_shape.py +5 -1
- mindspore/ops/_op_impl/cpu/maximum_grad.py +2 -0
- mindspore/{compression/common/__init__.py → ops/_op_impl/cpu/pyexecute.py} +13 -8
- mindspore/ops/_op_impl/cpu/reduce_sum.py +8 -0
- mindspore/ops/_op_impl/cpu/sparse_slice.py +62 -0
- mindspore/ops/_op_impl/cpu/sparse_slice_grad.py +60 -0
- mindspore/ops/_op_impl/cpu/tensor_shape.py +5 -1
- mindspore/ops/_op_impl/tbe/__init__.py +27 -608
- mindspore/ops/_op_impl/tbe/addcdiv_ds.py +42 -0
- mindspore/ops/_op_impl/tbe/addcmul_ds.py +44 -0
- mindspore/ops/_op_impl/tbe/assign_add_ds.py +1 -0
- mindspore/ops/_op_impl/tbe/atomic_addr_clean.py +1 -1
- mindspore/ops/_op_impl/tbe/avg_pool_3d_grad.py +1 -1
- mindspore/ops/_op_impl/tbe/basic_lstm_cell_c_state_grad_v2.py +0 -1
- mindspore/ops/_op_impl/tbe/batch_to_space.py +1 -1
- mindspore/ops/_op_impl/tbe/batch_to_space_nd.py +1 -1
- mindspore/ops/_op_impl/tbe/batch_to_space_nd_v2.py +41 -0
- mindspore/ops/_op_impl/tbe/bce_with_logits_loss.py +1 -0
- mindspore/ops/_op_impl/tbe/bias_add_grad.py +2 -0
- mindspore/ops/_op_impl/tbe/bn_infer_grad.py +4 -2
- mindspore/ops/_op_impl/tbe/bn_infer_grad_ds.py +40 -0
- mindspore/ops/_op_impl/tbe/bn_training_update.py +0 -1
- mindspore/ops/_op_impl/tbe/bn_training_update_ds.py +0 -1
- mindspore/ops/_op_impl/tbe/broadcast_to_ds.py +6 -4
- mindspore/ops/_op_impl/tbe/cast.py +0 -2
- mindspore/ops/_op_impl/tbe/cast_ds.py +3 -3
- mindspore/ops/_op_impl/tbe/ctc_loss_v2.py +0 -2
- mindspore/ops/_op_impl/tbe/ctc_loss_v2_grad.py +0 -2
- mindspore/ops/_op_impl/tbe/data_format_dim_map_ds.py +1 -0
- mindspore/ops/_op_impl/tbe/deformable_offsets.py +1 -0
- mindspore/ops/_op_impl/tbe/depthwise_conv2d.py +1 -1
- mindspore/ops/_op_impl/tbe/dynamic_atomic_addr_clean.py +1 -1
- mindspore/ops/_op_impl/tbe/gather_nd.py +1 -0
- mindspore/ops/_op_impl/tbe/greater.py +2 -0
- mindspore/ops/_op_impl/tbe/{index_add.py → inplace_index_add.py} +3 -6
- mindspore/ops/_op_impl/tbe/layer_norm_beta_gamma_backprop_v2.py +0 -1
- mindspore/ops/_op_impl/tbe/npu_clear_float_status_v2.py +35 -0
- mindspore/ops/_op_impl/tbe/npu_get_float_status_v2.py +35 -0
- mindspore/ops/_op_impl/tbe/one_hot_ds.py +0 -6
- mindspore/ops/_op_impl/tbe/{greater_ds.py → reduce_all_ds.py} +13 -16
- mindspore/ops/_op_impl/tbe/reduce_any_ds.py +39 -0
- mindspore/ops/_op_impl/tbe/roi_align_ds.py +44 -0
- mindspore/ops/_op_impl/tbe/roi_align_grad_ds.py +44 -0
- mindspore/ops/_op_impl/tbe/scatter_add.py +2 -0
- mindspore/ops/_op_impl/tbe/scatter_nd_add.py +2 -2
- mindspore/ops/_op_impl/tbe/slice.py +26 -15
- mindspore/ops/_op_impl/tbe/space_to_batch.py +1 -1
- mindspore/ops/_op_impl/tbe/space_to_batch_nd.py +1 -1
- mindspore/ops/_op_impl/tbe/strided_slice_grad_d.py +1 -0
- mindspore/ops/_op_impl/tbe/trans_data_ds.py +15 -5
- mindspore/ops/_op_impl/tbe/unsorted_segment_sum.py +1 -1
- mindspore/ops/_op_impl/tbe/unsorted_segment_sum_ds.py +2 -0
- mindspore/ops/_primitive_cache.py +3 -2
- mindspore/ops/_register_for_op.py +11 -0
- mindspore/ops/_utils/__init__.py +1 -1
- mindspore/ops/_utils/utils.py +20 -41
- mindspore/ops/_vmap/__init__.py +2 -2
- mindspore/ops/_vmap/vmap_array_ops.py +170 -78
- mindspore/ops/_vmap/vmap_base.py +24 -10
- mindspore/ops/_vmap/vmap_convolution_ops.py +7 -10
- mindspore/ops/_vmap/vmap_grad_math_ops.py +4 -4
- mindspore/ops/_vmap/vmap_grad_nn_ops.py +41 -9
- mindspore/ops/_vmap/vmap_image_ops.py +52 -0
- mindspore/ops/_vmap/vmap_math_ops.py +77 -6
- mindspore/ops/_vmap/vmap_nn_ops.py +78 -29
- mindspore/ops/_vmap/vmap_other_ops.py +3 -1
- mindspore/ops/_vmap/vmap_random_ops.py +55 -3
- mindspore/ops/_vmap/vmap_sparse_ops.py +1 -0
- mindspore/ops/bprop_mindir/AdaptiveAvgPool2D_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/AdaptiveMaxPool2D_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/ApproximateEqual_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/Argmax_bprop.mindir +13 -12
- mindspore/ops/bprop_mindir/Argmin_bprop.mindir +14 -13
- mindspore/ops/bprop_mindir/AssignSub_bprop.mindir +17 -18
- mindspore/ops/bprop_mindir/Assign_bprop.mindir +16 -16
- mindspore/ops/bprop_mindir/AvgPool3D_bprop.mindir +150 -0
- mindspore/ops/bprop_mindir/AvgPool_bprop.mindir +66 -0
- mindspore/ops/bprop_mindir/BCEWithLogitsLoss_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/BNTrainingReduce_bprop.mindir +13 -12
- mindspore/ops/bprop_mindir/BatchNormGrad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/BatchToSpaceND_bprop.mindir +28 -0
- mindspore/ops/bprop_mindir/BiasAddGrad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/BinaryCrossEntropy_bprop.mindir +33 -0
- mindspore/ops/bprop_mindir/BroadcastTo_bprop.mindir +306 -0
- mindspore/ops/bprop_mindir/Broadcast_bprop.mindir +12 -8
- mindspore/ops/bprop_mindir/CTCLoss_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Concat_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Conv2DBackpropFilter_bprop.mindir +240 -0
- mindspore/ops/bprop_mindir/Conv2DBackpropInput_bprop.mindir +247 -0
- mindspore/ops/bprop_mindir/Conv2DTranspose_bprop.mindir +247 -0
- mindspore/ops/bprop_mindir/Conv3DTranspose_bprop.mindir +315 -0
- mindspore/ops/bprop_mindir/Conv3D_bprop.mindir +278 -0
- mindspore/ops/bprop_mindir/DType_bprop.mindir +12 -12
- mindspore/ops/bprop_mindir/DeformableOffsets_bprop.mindir +58 -0
- mindspore/ops/bprop_mindir/Depend_bprop.mindir +12 -13
- mindspore/ops/bprop_mindir/DepthToSpace_bprop.mindir +23 -0
- mindspore/ops/bprop_mindir/DepthwiseConv2dNative_bprop.mindir +138 -0
- mindspore/ops/bprop_mindir/DiagPart_bprop.mindir +15 -0
- mindspore/ops/bprop_mindir/Dropout2D_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Dropout3D_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/DropoutDoMask_bprop.mindir +22 -24
- mindspore/ops/bprop_mindir/DropoutGenMask_bprop.mindir +16 -14
- mindspore/ops/bprop_mindir/DropoutGrad_bprop.mindir +27 -0
- mindspore/ops/bprop_mindir/Dropout_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/DynamicGRUV2_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/DynamicRNN_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/DynamicShape_bprop.mindir +12 -12
- mindspore/ops/bprop_mindir/Elu_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/EmbeddingLookup_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Equal_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/ExpandDims_bprop.mindir +58 -0
- mindspore/ops/bprop_mindir/FastGeLU_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/Flatten_bprop.mindir +54 -0
- mindspore/ops/bprop_mindir/FloorDiv_bprop.mindir +18 -15
- mindspore/ops/bprop_mindir/GatherD_bprop.mindir +26 -0
- mindspore/ops/bprop_mindir/GatherNd_bprop.mindir +57 -0
- mindspore/ops/bprop_mindir/Gather_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/GreaterEqual_bprop.mindir +17 -18
- mindspore/ops/bprop_mindir/Greater_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/HSigmoid_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/HSwish_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/IOU_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/InstanceNorm_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/IsFinite_bprop.mindir +13 -12
- mindspore/ops/bprop_mindir/IsInf_bprop.mindir +13 -10
- mindspore/ops/bprop_mindir/IsNan_bprop.mindir +14 -11
- mindspore/ops/bprop_mindir/KLDivLoss_bprop.mindir +126 -0
- mindspore/ops/bprop_mindir/L2Loss_bprop.mindir +15 -0
- mindspore/ops/bprop_mindir/L2Normalize_bprop.mindir +30 -0
- mindspore/ops/bprop_mindir/LRN_bprop.mindir +43 -0
- mindspore/ops/bprop_mindir/LayerNormGrad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/LessEqual_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/Less_bprop.mindir +17 -18
- mindspore/ops/bprop_mindir/LinSpace_bprop.mindir +22 -19
- mindspore/ops/bprop_mindir/Load_bprop.mindir +12 -13
- mindspore/ops/bprop_mindir/LogSoftmax_bprop.mindir +23 -0
- mindspore/ops/bprop_mindir/LogicalAnd_bprop.mindir +17 -18
- mindspore/ops/bprop_mindir/LogicalNot_bprop.mindir +14 -13
- mindspore/ops/bprop_mindir/MaskedSelect_bprop.mindir +21 -0
- mindspore/ops/bprop_mindir/MaxPool3DGradGrad_bprop.mindir +74 -0
- mindspore/ops/bprop_mindir/MaxPool3DGrad_bprop.mindir +74 -0
- mindspore/ops/bprop_mindir/MaxPool3D_bprop.mindir +75 -0
- mindspore/ops/bprop_mindir/MaxPoolGradGrad_bprop.mindir +65 -0
- mindspore/ops/bprop_mindir/MaxPoolWithArgmax_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Maximum_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Minimum_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/MirrorPad_bprop.mindir +27 -0
- mindspore/ops/bprop_mindir/Mish_bprop.mindir +35 -0
- mindspore/ops/bprop_mindir/MulNoNan_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/NLLLoss_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/NonZero_bprop.mindir +14 -0
- mindspore/ops/bprop_mindir/NotEqual_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/OneHot_bprop.mindir +25 -23
- mindspore/ops/bprop_mindir/OnesLike_bprop.mindir +13 -13
- mindspore/ops/bprop_mindir/PReLU_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Pad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Padding_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/RNNTLoss_bprop.mindir +29 -0
- mindspore/ops/bprop_mindir/ROIAlign_bprop.mindir +82 -0
- mindspore/ops/bprop_mindir/Range_bprop.mindir +21 -19
- mindspore/ops/bprop_mindir/Rank_bprop.mindir +11 -11
- mindspore/ops/bprop_mindir/ReLU6_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/ReLUV2_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/ReduceAll_bprop.mindir +18 -17
- mindspore/ops/bprop_mindir/ReduceAny_bprop.mindir +18 -17
- mindspore/ops/bprop_mindir/ReluGrad_bprop.mindir +19 -23
- mindspore/ops/bprop_mindir/Reshape_bprop.mindir +60 -0
- mindspore/ops/bprop_mindir/ResizeBilinear_bprop.mindir +29 -0
- mindspore/ops/bprop_mindir/ResizeNearestNeighbor_bprop.mindir +89 -0
- mindspore/ops/bprop_mindir/ReverseSequence_bprop.mindir +52 -0
- mindspore/ops/bprop_mindir/ReverseV2_bprop.mindir +22 -0
- mindspore/ops/bprop_mindir/Round_bprop.mindir +14 -13
- mindspore/ops/bprop_mindir/ScatterMax_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/ScatterMin_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/ScatterNdUpdate_bprop.mindir +22 -0
- mindspore/ops/bprop_mindir/ScatterNd_bprop.mindir +24 -0
- mindspore/ops/bprop_mindir/ScatterNonAliasingAdd_bprop.mindir +22 -0
- mindspore/ops/bprop_mindir/ScatterUpdate_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/SeLU_bprop.mindir +21 -0
- mindspore/ops/bprop_mindir/Select_bprop.mindir +30 -34
- mindspore/ops/bprop_mindir/Shape_bprop.mindir +12 -12
- mindspore/ops/bprop_mindir/SigmoidCrossEntropyWithLogits_bprop.mindir +21 -0
- mindspore/ops/bprop_mindir/SigmoidGrad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Sigmoid_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/Sign_bprop.mindir +13 -12
- mindspore/ops/bprop_mindir/Slice_bprop.mindir +26 -0
- mindspore/ops/bprop_mindir/SmoothL1Loss_bprop.mindir +36 -0
- mindspore/ops/bprop_mindir/SoftmaxCrossEntropyWithLogits_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Softplus_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/Softsign_bprop.mindir +33 -0
- mindspore/ops/bprop_mindir/Sort_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/SpaceToBatchND_bprop.mindir +28 -0
- mindspore/ops/bprop_mindir/SpaceToDepth_bprop.mindir +23 -0
- mindspore/ops/bprop_mindir/SparseGatherV2_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/SparseSoftmaxCrossEntropyWithLogits_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Split_bprop.mindir +22 -0
- mindspore/ops/bprop_mindir/Squeeze_bprop.mindir +54 -0
- mindspore/ops/bprop_mindir/StridedSliceGrad_bprop.mindir +95 -0
- mindspore/ops/bprop_mindir/StridedSlice_bprop.mindir +98 -0
- mindspore/ops/bprop_mindir/Switch_bprop.mindir +28 -32
- mindspore/ops/bprop_mindir/TanhGrad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Tanh_bprop.mindir +66 -0
- mindspore/ops/bprop_mindir/TensorScatterAdd_bprop.mindir +22 -0
- mindspore/ops/bprop_mindir/TensorScatterUpdate_bprop.mindir +29 -0
- mindspore/ops/bprop_mindir/TensorShape_bprop.mindir +14 -0
- mindspore/ops/bprop_mindir/Tile_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/TopK_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/TransShape_bprop.mindir +23 -0
- mindspore/ops/bprop_mindir/TruncateDiv_bprop.mindir +18 -15
- mindspore/ops/bprop_mindir/TupleGetItem_bprop.mindir +11 -13
- mindspore/ops/bprop_mindir/Unique_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/Unstack_bprop.mindir +22 -0
- mindspore/ops/bprop_mindir/UpsampleNearest3D_bprop.mindir +32 -0
- mindspore/ops/bprop_mindir/UpsampleTrilinear3D_bprop.mindir +38 -0
- mindspore/ops/bprop_mindir/ZerosLike_bprop.mindir +13 -12
- mindspore/ops/bprop_mindir/__init__.py +1 -4
- mindspore/ops/bprop_mindir/generate_mindir.py +32 -20
- mindspore/ops/composite/__init__.py +12 -13
- mindspore/ops/composite/base.py +261 -254
- mindspore/ops/composite/env_ops.py +41 -0
- mindspore/ops/composite/math_ops.py +197 -156
- mindspore/ops/composite/multitype_ops/_compile_utils.py +428 -176
- mindspore/ops/composite/multitype_ops/_constexpr_utils.py +188 -87
- mindspore/ops/composite/multitype_ops/add_impl.py +23 -1
- mindspore/ops/composite/multitype_ops/div_impl.py +3 -3
- mindspore/ops/composite/multitype_ops/equal_impl.py +1 -0
- mindspore/ops/composite/multitype_ops/floordiv_impl.py +1 -1
- mindspore/ops/composite/multitype_ops/getitem_impl.py +52 -5
- mindspore/ops/composite/multitype_ops/greater_equal_impl.py +31 -0
- mindspore/ops/composite/multitype_ops/greater_impl.py +31 -0
- mindspore/ops/composite/multitype_ops/in_impl.py +15 -3
- mindspore/ops/composite/multitype_ops/less_equal_impl.py +33 -2
- mindspore/ops/composite/multitype_ops/less_impl.py +33 -0
- mindspore/ops/composite/multitype_ops/logical_and_impl.py +2 -2
- mindspore/ops/composite/multitype_ops/logical_or_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/mod_impl.py +1 -1
- mindspore/ops/composite/multitype_ops/mul_impl.py +21 -7
- mindspore/ops/composite/multitype_ops/not_in_impl.py +15 -3
- mindspore/ops/composite/multitype_ops/ones_like_impl.py +2 -4
- mindspore/ops/composite/multitype_ops/pow_impl.py +1 -0
- mindspore/ops/composite/multitype_ops/setitem_impl.py +62 -70
- mindspore/ops/composite/multitype_ops/sub_impl.py +3 -3
- mindspore/ops/composite/multitype_ops/zeros_like_impl.py +41 -4
- mindspore/ops/function/__init__.py +323 -8
- mindspore/ops/function/array_func.py +3511 -780
- mindspore/ops/function/clip_func.py +329 -0
- mindspore/ops/function/debug_func.py +6 -6
- mindspore/ops/function/grad/__init__.py +5 -1
- mindspore/ops/function/grad/grad_func.py +736 -65
- mindspore/ops/function/image_func.py +270 -0
- mindspore/ops/function/linalg_func.py +268 -8
- mindspore/ops/function/math_func.py +8032 -3164
- mindspore/ops/function/nn_func.py +5619 -1855
- mindspore/ops/function/other_func.py +115 -0
- mindspore/ops/function/parameter_func.py +11 -10
- mindspore/ops/function/random_func.py +939 -77
- mindspore/ops/function/sparse_func.py +249 -84
- mindspore/ops/function/sparse_unary_func.py +2303 -0
- mindspore/ops/function/spectral_func.py +146 -0
- mindspore/ops/function/vmap_func.py +114 -0
- mindspore/ops/functional.py +182 -254
- mindspore/ops/op_info_register.py +79 -34
- mindspore/ops/operations/__init__.py +210 -118
- mindspore/ops/operations/_csr_ops.py +7 -7
- mindspore/ops/operations/_embedding_cache_ops.py +25 -15
- mindspore/ops/operations/_grad_ops.py +447 -322
- mindspore/ops/operations/_inner_ops.py +547 -176
- mindspore/ops/operations/_map_tensor_ops.py +112 -0
- mindspore/ops/operations/_ms_kernel.py +29 -27
- mindspore/ops/operations/_ocr_ops.py +11 -11
- mindspore/ops/operations/_opaque_predicate_registry.py +41 -0
- mindspore/ops/operations/_quant_ops.py +186 -101
- mindspore/ops/operations/_rl_inner_ops.py +122 -61
- mindspore/ops/operations/_scalar_ops.py +466 -0
- mindspore/ops/operations/_sequence_ops.py +1047 -0
- mindspore/ops/operations/_tensor_array.py +10 -11
- mindspore/ops/operations/_thor_ops.py +4 -4
- mindspore/ops/operations/array_ops.py +1428 -1226
- mindspore/ops/operations/comm_ops.py +180 -117
- mindspore/ops/operations/control_ops.py +4 -2
- mindspore/ops/operations/custom_ops.py +185 -98
- mindspore/ops/operations/debug_ops.py +92 -54
- mindspore/ops/operations/image_ops.py +406 -211
- mindspore/ops/operations/inner_ops.py +42 -53
- mindspore/ops/operations/linalg_ops.py +32 -29
- mindspore/ops/operations/math_ops.py +2076 -897
- mindspore/ops/operations/nn_ops.py +1282 -1252
- mindspore/ops/operations/other_ops.py +124 -278
- mindspore/ops/operations/random_ops.py +345 -178
- mindspore/ops/operations/rl_ops.py +8 -9
- mindspore/ops/operations/sparse_ops.py +502 -157
- mindspore/ops/operations/spectral_ops.py +107 -0
- mindspore/ops/primitive.py +192 -15
- mindspore/ops/vm_impl_registry.py +23 -2
- mindspore/parallel/__init__.py +6 -1
- mindspore/parallel/_auto_parallel_context.py +199 -92
- mindspore/parallel/_cell_wrapper.py +4 -2
- mindspore/parallel/_cost_model_context.py +3 -0
- mindspore/parallel/_dp_allreduce_fusion.py +2 -1
- mindspore/parallel/_offload_context.py +185 -0
- mindspore/parallel/_parallel_serialization.py +167 -28
- mindspore/parallel/_ps_context.py +9 -5
- mindspore/parallel/_recovery_context.py +1 -1
- mindspore/parallel/_tensor.py +9 -1
- mindspore/{nn/transformer → parallel/_transformer}/__init__.py +6 -6
- mindspore/{nn/transformer → parallel/_transformer}/layers.py +59 -37
- mindspore/{nn/transformer → parallel/_transformer}/loss.py +4 -7
- mindspore/{nn/transformer → parallel/_transformer}/moe.py +160 -35
- mindspore/{nn/transformer → parallel/_transformer}/op_parallel_config.py +3 -3
- mindspore/{nn/transformer → parallel/_transformer}/transformer.py +235 -196
- mindspore/parallel/_utils.py +47 -7
- mindspore/parallel/algo_parameter_config.py +5 -1
- mindspore/parallel/checkpoint_transform.py +329 -0
- mindspore/parallel/shard.py +229 -0
- mindspore/perf_msvcbuildinsights.dll +0 -0
- mindspore/pgodb140.dll +0 -0
- mindspore/pgort140.dll +0 -0
- mindspore/profiler/__init__.py +2 -1
- mindspore/profiler/common/util.py +4 -3
- mindspore/profiler/common/validator/validate_path.py +2 -2
- mindspore/profiler/envprofiling.py +249 -0
- mindspore/profiler/parser/aicpu_data_parser.py +38 -39
- mindspore/profiler/parser/ascend_timeline_generator.py +497 -0
- mindspore/profiler/parser/base_timeline_generator.py +471 -0
- mindspore/profiler/parser/cpu_gpu_timeline_generator.py +684 -0
- mindspore/profiler/parser/framework_parser.py +42 -16
- mindspore/profiler/parser/hccl_parser.py +158 -158
- mindspore/profiler/parser/hwts_log_parser.py +7 -6
- mindspore/profiler/parser/integrator.py +18 -1579
- mindspore/profiler/parser/minddata_analyzer.py +8 -8
- mindspore/profiler/parser/msadvisor_analyzer.py +14 -27
- mindspore/profiler/parser/msadvisor_parser.py +2 -4
- mindspore/profiler/parser/optime_parser.py +17 -18
- mindspore/profiler/parser/profiler_info.py +108 -0
- mindspore/profiler/parser/step_trace_parser.py +1 -1
- mindspore/profiler/profiling.py +396 -194
- mindspore/rewrite/__init__.py +6 -2
- mindspore/rewrite/api/node.py +51 -110
- mindspore/rewrite/api/node_type.py +10 -6
- mindspore/rewrite/api/pattern_engine.py +51 -7
- mindspore/rewrite/api/scoped_value.py +64 -53
- mindspore/rewrite/api/symbol_tree.py +108 -61
- mindspore/rewrite/api/tree_node_helper.py +2 -3
- mindspore/{compression/quant/__init__.py → rewrite/ast_creator_register.py} +20 -11
- mindspore/rewrite/ast_helpers/__init__.py +6 -3
- mindspore/rewrite/ast_helpers/ast_creator.py +115 -0
- mindspore/rewrite/ast_helpers/ast_finder.py +99 -1
- mindspore/rewrite/ast_helpers/ast_modifier.py +17 -4
- mindspore/rewrite/ast_helpers/ast_replacer.py +1 -1
- mindspore/rewrite/ast_transformers/__init__.py +0 -1
- mindspore/rewrite/ast_transformers/flatten_recursive_stmt.py +46 -5
- mindspore/rewrite/ast_transformers/remove_return_out_of_if.py +6 -3
- mindspore/rewrite/common/__init__.py +2 -0
- mindspore/rewrite/common/event.py +1 -1
- mindspore/rewrite/common/observable.py +1 -1
- mindspore/rewrite/common/observer.py +1 -1
- mindspore/rewrite/common/rewrite_elog.py +35 -0
- mindspore/rewrite/namer.py +2 -2
- mindspore/rewrite/namespace.py +14 -4
- mindspore/rewrite/node.py +161 -13
- mindspore/rewrite/parser.py +0 -1
- mindspore/rewrite/parser_register.py +0 -1
- mindspore/rewrite/parsers/arguments_parser.py +3 -2
- mindspore/rewrite/parsers/assign_parser.py +267 -67
- mindspore/rewrite/parsers/attribute_parser.py +56 -0
- mindspore/rewrite/parsers/class_def_parser.py +191 -108
- mindspore/rewrite/parsers/constant_parser.py +101 -0
- mindspore/rewrite/parsers/container_parser.py +88 -0
- mindspore/rewrite/parsers/for_parser.py +28 -15
- mindspore/rewrite/parsers/function_def_parser.py +21 -5
- mindspore/rewrite/parsers/if_parser.py +11 -28
- mindspore/rewrite/parsers/module_parser.py +9 -6
- mindspore/rewrite/parsers/return_parser.py +3 -2
- mindspore/rewrite/sparsify/__init__.py +0 -0
- mindspore/rewrite/sparsify/sparse_transformer.py +448 -0
- mindspore/rewrite/sparsify/sparsify.py +109 -0
- mindspore/rewrite/sparsify/utils.py +173 -0
- mindspore/rewrite/symbol_tree.py +322 -109
- mindspore/rewrite/symbol_tree_builder.py +45 -8
- mindspore/rewrite/symbol_tree_dumper.py +0 -1
- mindspore/rewrite/topological_manager.py +1 -2
- mindspore/run_check/_check_version.py +209 -112
- mindspore/run_check/run_check.py +2 -1
- mindspore/tbbmalloc.dll +0 -0
- mindspore/tinyxml2.dll +0 -0
- mindspore/train/__init__.py +6 -4
- mindspore/train/_utils.py +28 -5
- mindspore/train/amp.py +321 -50
- mindspore/train/callback/__init__.py +3 -1
- mindspore/train/callback/_backup_and_restore.py +120 -0
- mindspore/train/callback/_callback.py +8 -8
- mindspore/train/callback/_checkpoint.py +12 -9
- mindspore/train/callback/_early_stop.py +13 -7
- mindspore/train/callback/_history.py +8 -8
- mindspore/train/callback/_lambda_callback.py +6 -6
- mindspore/train/callback/_landscape.py +36 -38
- mindspore/train/callback/_loss_monitor.py +12 -6
- mindspore/train/callback/_lr_scheduler_callback.py +2 -4
- mindspore/train/callback/_on_request_exit.py +212 -0
- mindspore/train/callback/_reduce_lr_on_plateau.py +13 -7
- mindspore/train/callback/_summary_collector.py +27 -19
- mindspore/train/callback/_time_monitor.py +13 -7
- mindspore/train/checkpoint_pb2.py +68 -8
- mindspore/train/data_sink.py +122 -33
- mindspore/train/dataset_helper.py +28 -87
- mindspore/train/loss_scale_manager.py +4 -7
- mindspore/{nn → train}/metrics/__init__.py +20 -20
- mindspore/{nn → train}/metrics/accuracy.py +12 -10
- mindspore/{nn → train}/metrics/auc.py +4 -4
- mindspore/{nn → train}/metrics/bleu_score.py +4 -4
- mindspore/{nn → train}/metrics/confusion_matrix.py +10 -8
- mindspore/{nn → train}/metrics/cosine_similarity.py +4 -4
- mindspore/{nn → train}/metrics/dice.py +6 -5
- mindspore/{nn → train}/metrics/error.py +7 -5
- mindspore/{nn → train}/metrics/fbeta.py +9 -7
- mindspore/{nn → train}/metrics/hausdorff_distance.py +8 -6
- mindspore/{nn → train}/metrics/loss.py +4 -3
- mindspore/{nn → train}/metrics/mean_surface_distance.py +6 -5
- mindspore/{nn → train}/metrics/metric.py +6 -5
- mindspore/{nn → train}/metrics/occlusion_sensitivity.py +4 -3
- mindspore/{nn → train}/metrics/perplexity.py +5 -4
- mindspore/{nn → train}/metrics/precision.py +5 -4
- mindspore/{nn → train}/metrics/recall.py +5 -4
- mindspore/{nn → train}/metrics/roc.py +7 -6
- mindspore/{nn → train}/metrics/root_mean_square_surface_distance.py +6 -5
- mindspore/{nn → train}/metrics/topk.py +7 -5
- mindspore/train/mind_ir_pb2.py +339 -32
- mindspore/train/model.py +113 -84
- mindspore/train/serialization.py +547 -167
- mindspore/train/summary/_summary_adapter.py +1 -1
- mindspore/train/summary/summary_record.py +43 -12
- mindspore/train/train_thor/convert_utils.py +7 -1
- mindspore/train/train_thor/dataset_helper.py +3 -3
- mindspore/train/train_thor/model_thor.py +0 -4
- mindspore/turbojpeg.dll +0 -0
- mindspore/vcmeta.dll +0 -0
- mindspore/vcruntime140.dll +0 -0
- mindspore/vcruntime140_1.dll +0 -0
- mindspore/version.py +1 -1
- {mindspore-1.10.0.dist-info → mindspore-2.0.0rc1.dist-info}/METADATA +4 -3
- {mindspore-1.10.0.dist-info → mindspore-2.0.0rc1.dist-info}/RECORD +901 -660
- mindspore/compression/common/constant.py +0 -124
- mindspore/compression/export/__init__.py +0 -19
- mindspore/compression/export/quant_export.py +0 -514
- mindspore/compression/quant/qat.py +0 -636
- mindspore/compression/quant/quant_utils.py +0 -462
- mindspore/compression/quant/quantizer.py +0 -68
- mindspore/libatomic-1.dll +0 -0
- mindspore/libgcc_s_seh-1.dll +0 -0
- mindspore/libgfortran-4.dll +0 -0
- mindspore/libgomp-1.dll +0 -0
- mindspore/libjpeg-62.dll +0 -0
- mindspore/libmindspore.dll +0 -0
- mindspore/libmindspore_common.dll +0 -0
- mindspore/libmindspore_core.dll +0 -0
- mindspore/libmindspore_glog.dll +0 -0
- mindspore/libnnacl.dll +0 -0
- mindspore/libopencv_core452.dll +0 -0
- mindspore/libopencv_imgcodecs452.dll +0 -0
- mindspore/libopencv_imgproc452.dll +0 -0
- mindspore/libquadmath-0.dll +0 -0
- mindspore/libsqlite3.dll +0 -0
- mindspore/libssp-0.dll +0 -0
- mindspore/libstdc++-6.dll +0 -0
- mindspore/libtinyxml2.dll +0 -0
- mindspore/libturbojpeg.dll +0 -0
- mindspore/libwinpthread-1.dll +0 -0
- mindspore/nn/layer/quant.py +0 -1868
- mindspore/nn/layer/rnn_utils.py +0 -90
- mindspore/nn/probability/dpn/__init__.py +0 -22
- mindspore/nn/probability/dpn/vae/__init__.py +0 -25
- mindspore/nn/probability/dpn/vae/cvae.py +0 -138
- mindspore/nn/probability/dpn/vae/vae.py +0 -122
- mindspore/nn/probability/infer/__init__.py +0 -22
- mindspore/nn/probability/infer/variational/elbo.py +0 -70
- mindspore/nn/probability/infer/variational/svi.py +0 -84
- mindspore/nn/probability/toolbox/__init__.py +0 -22
- mindspore/nn/probability/toolbox/anomaly_detection.py +0 -99
- mindspore/nn/probability/toolbox/uncertainty_evaluation.py +0 -363
- mindspore/nn/probability/transforms/__init__.py +0 -22
- mindspore/nn/probability/transforms/transform_bnn.py +0 -262
- mindspore/nn/probability/zhusuan/__init__.py +0 -18
- mindspore/nn/probability/zhusuan/framework/__init__.py +0 -18
- mindspore/nn/probability/zhusuan/framework/bn.py +0 -95
- mindspore/nn/probability/zhusuan/variational/__init__.py +0 -18
- mindspore/nn/probability/zhusuan/variational/elbo.py +0 -46
- mindspore/ops/_op_impl/tbe/bias_add_grad_ds.py +0 -52
- mindspore/ops/_op_impl/tbe/scatter_nd_add_ds.py +0 -43
- mindspore/ops/bprop_mindir/AssignAdd_bprop.mindir +0 -20
- mindspore/ops/bprop_mindir/Identity_bprop.mindir +0 -9
- mindspore/ops/bprop_mindir/LogicalOr_bprop.mindir +0 -20
- mindspore/ops/bprop_mindir/ReLU_bprop.mindir +0 -16
- mindspore/ops/bprop_mindir/UpdateState_bprop.mindir +0 -17
- mindspore/ops/bprop_mindir/stop_gradient_bprop.mindir +0 -12
- mindspore/ops/composite/array_ops.py +0 -210
- mindspore/ops/composite/clip_ops.py +0 -238
- mindspore/ops/composite/random_ops.py +0 -426
- mindspore/ops/composite/vmap_ops.py +0 -38
- mindspore/ops/operations/sponge_ops.py +0 -3531
- mindspore/ops/operations/sponge_update_ops.py +0 -2546
- mindspore/parallel/nn/__init__.py +0 -42
- mindspore/parallel/nn/loss.py +0 -22
- mindspore/parallel/nn/moe.py +0 -21
- mindspore/parallel/nn/op_parallel_config.py +0 -22
- mindspore/parallel/nn/transformer.py +0 -31
- mindspore/run_check/_check_deps_version.py +0 -84
- {mindspore-1.10.0.dist-info → mindspore-2.0.0rc1.dist-info}/WHEEL +0 -0
- {mindspore-1.10.0.dist-info → mindspore-2.0.0rc1.dist-info}/entry_points.txt +0 -0
- {mindspore-1.10.0.dist-info → mindspore-2.0.0rc1.dist-info}/top_level.txt +0 -0
mindspore/context.py
CHANGED
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
|
|
1
|
+
# Copyright 2020-2022 Huawei Technologies Co., Ltd
|
|
2
2
|
#
|
|
3
3
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
4
4
|
# you may not use this file except in compliance with the License.
|
|
@@ -27,22 +27,23 @@ from types import FunctionType
|
|
|
27
27
|
|
|
28
28
|
from mindspore import log as logger
|
|
29
29
|
from mindspore._c_expression import MSContext, ms_ctx_param
|
|
30
|
-
from mindspore
|
|
30
|
+
from mindspore import _checkparam as Validator
|
|
31
|
+
from mindspore._checkparam import args_type_check
|
|
31
32
|
from mindspore.parallel._auto_parallel_context import _set_auto_parallel_context, _get_auto_parallel_context, \
|
|
32
33
|
_reset_auto_parallel_context
|
|
33
34
|
from mindspore.parallel._ps_context import _set_ps_context, _get_ps_context, _reset_ps_context, \
|
|
34
35
|
_need_reset_device_target_for_ps
|
|
35
|
-
from mindspore.
|
|
36
|
+
from mindspore.parallel._offload_context import _set_offload_context, _get_offload_context
|
|
36
37
|
|
|
37
38
|
__all__ = ['GRAPH_MODE', 'PYNATIVE_MODE', 'set_context', 'get_context', 'set_auto_parallel_context',
|
|
38
39
|
'get_auto_parallel_context', 'reset_auto_parallel_context', 'ParallelMode', 'set_ps_context',
|
|
39
|
-
'get_ps_context', 'reset_ps_context']
|
|
40
|
+
'get_ps_context', 'reset_ps_context', 'set_offload_context', 'get_offload_context']
|
|
40
41
|
|
|
41
42
|
GRAPH_MODE = 0
|
|
42
43
|
PYNATIVE_MODE = 1
|
|
43
44
|
_DEVICE_APP_MEMORY_SIZE = 31 # The max memory size of graph plus variable.
|
|
44
45
|
_re_pattern = r'[1-9][0-9]*(\.)?[0-9]*GB|0\.[0-9]*GB'
|
|
45
|
-
|
|
46
|
+
K_CONTEXT = None
|
|
46
47
|
|
|
47
48
|
|
|
48
49
|
def _make_directory(path):
|
|
@@ -184,11 +185,12 @@ class _Context:
|
|
|
184
185
|
if self.enable_debug_runtime:
|
|
185
186
|
self.set_backend_policy("vm")
|
|
186
187
|
parallel_mode = _get_auto_parallel_context("parallel_mode")
|
|
187
|
-
if parallel_mode not in (ParallelMode.DATA_PARALLEL, ParallelMode.STAND_ALONE):
|
|
188
|
-
raise ValueError(f"Got {parallel_mode}, when the user enabled SEMI_AUTO_PARALELL
|
|
189
|
-
f"pynative mode dose not support, you should set "
|
|
190
|
-
f"context.set_auto_parallel_context(parallel_mode='data_parallel') "
|
|
191
|
-
f"
|
|
188
|
+
if parallel_mode not in (ParallelMode.DATA_PARALLEL, ParallelMode.STAND_ALONE, ParallelMode.AUTO_PARALLEL):
|
|
189
|
+
raise ValueError(f"Got {parallel_mode}, when the user enabled SEMI_AUTO_PARALELL, "
|
|
190
|
+
f"pynative mode dose not support, you should set either "
|
|
191
|
+
f"context.set_auto_parallel_context(parallel_mode='data_parallel'), "
|
|
192
|
+
f"context.set_auto_parallel_context(parallel_mode='stand_alone') "
|
|
193
|
+
f"or context.set_auto_parallel_context(parallel_mode='auto_parallel').")
|
|
192
194
|
self._context_switches.push(True, None)
|
|
193
195
|
elif mode == GRAPH_MODE:
|
|
194
196
|
if self.enable_debug_runtime:
|
|
@@ -199,6 +201,83 @@ class _Context:
|
|
|
199
201
|
f"or context.PYNATIVE_MODE (1), but got {mode}.")
|
|
200
202
|
self.set_param(ms_ctx_param.mode, mode)
|
|
201
203
|
|
|
204
|
+
def set_memory_optimize_level(self, memory_optimize_level):
|
|
205
|
+
"""
|
|
206
|
+
The memory optimize level, support "O0", "O1".
|
|
207
|
+
|
|
208
|
+
Args:
|
|
209
|
+
target (str): "O0", "O1"
|
|
210
|
+
"""
|
|
211
|
+
memory_optimize_levels = ["O0", "O1"]
|
|
212
|
+
if memory_optimize_level not in memory_optimize_levels:
|
|
213
|
+
raise ValueError(f"For 'context.set_context', the argument 'memory_optimize_level' must be one of "
|
|
214
|
+
f"{memory_optimize_levels}, but got {memory_optimize_level}.")
|
|
215
|
+
if memory_optimize_level == "O0":
|
|
216
|
+
self.set_param(ms_ctx_param.memory_optimize_level, 0)
|
|
217
|
+
else:
|
|
218
|
+
self.set_param(ms_ctx_param.memory_optimize_level, 1)
|
|
219
|
+
|
|
220
|
+
def set_memory_offload(self, memory_offload):
|
|
221
|
+
"""
|
|
222
|
+
Enable memory offload or not, support "ON", "OFF".
|
|
223
|
+
|
|
224
|
+
Args:
|
|
225
|
+
memory_offload (str): "ON", "OFF"
|
|
226
|
+
"""
|
|
227
|
+
memory_offload_options = ["ON", "OFF"]
|
|
228
|
+
if memory_offload not in memory_offload_options:
|
|
229
|
+
raise ValueError(f"For 'context.set_context', the argument 'memory_offload' must be one of "
|
|
230
|
+
f"{memory_offload_options}, but got {memory_offload}.")
|
|
231
|
+
if memory_offload == "ON":
|
|
232
|
+
self.set_param(ms_ctx_param.memory_offload, True)
|
|
233
|
+
else:
|
|
234
|
+
self.set_param(ms_ctx_param.memory_offload, False)
|
|
235
|
+
|
|
236
|
+
def set_deterministic(self, deterministic):
|
|
237
|
+
"""
|
|
238
|
+
Enable model run in deterministic, and support the values "ON" and "OFF".
|
|
239
|
+
|
|
240
|
+
Args:
|
|
241
|
+
deterministic (str): "ON", "OFF"
|
|
242
|
+
"""
|
|
243
|
+
deterministic_options = ["ON", "OFF"]
|
|
244
|
+
if deterministic not in deterministic_options:
|
|
245
|
+
raise ValueError(f"For 'context.set_context', the argument 'deterministic' must be one of "
|
|
246
|
+
f"{deterministic_options}, but got {deterministic}.")
|
|
247
|
+
self.set_param(ms_ctx_param.deterministic, deterministic)
|
|
248
|
+
|
|
249
|
+
def set_ascend_config(self, ascend_config):
|
|
250
|
+
"""
|
|
251
|
+
Enable ascend config.
|
|
252
|
+
|
|
253
|
+
Args:
|
|
254
|
+
ascend_config (dict): 'precision_mode'
|
|
255
|
+
- precision_mode (str): "force_fp16", "allow_fp32_to_fp16", "allow_mix_precision",
|
|
256
|
+
"must_keep_origin_dtype", "force_fp32", "force_lowerprecision", "allow_fp32_to_bf16",
|
|
257
|
+
"allow_fp32_to_lowprecision", "allow_mix_precision_fp16" and "allow_mix_precision_bf16".
|
|
258
|
+
"""
|
|
259
|
+
|
|
260
|
+
ascend_cfgs = {'precision_mode': ["force_fp16", "allow_fp32_to_fp16", "allow_mix_precision",
|
|
261
|
+
"must_keep_origin_dtype", "force_fp32", "force_lowerprecision",
|
|
262
|
+
"allow_fp32_to_bf16", "allow_fp32_to_lowprecision",
|
|
263
|
+
"allow_mix_precision_fp16", "allow_mix_precision_bf16"],
|
|
264
|
+
'jit_compile': [True, False]}
|
|
265
|
+
for ascend_key in ascend_config:
|
|
266
|
+
if ascend_key not in ascend_cfgs:
|
|
267
|
+
raise ValueError(f"For 'context.set_context', the key of argument 'ascend_config' must be one of "
|
|
268
|
+
f"{ascend_cfgs}, but got {ascend_key}.")
|
|
269
|
+
supported_modes = ascend_cfgs.get(ascend_key)
|
|
270
|
+
if ascend_config[ascend_key] not in supported_modes:
|
|
271
|
+
raise ValueError(f"For 'ascend_config', the value of argument {ascend_key} must be one of "
|
|
272
|
+
f"{supported_modes}, but got {ascend_config[ascend_key]}.")
|
|
273
|
+
if ascend_key == 'precision_mode':
|
|
274
|
+
self.set_param(ms_ctx_param.precision_mode, ascend_config[ascend_key])
|
|
275
|
+
if ascend_key == 'jit_compile':
|
|
276
|
+
if ascend_config[ascend_key] is True:
|
|
277
|
+
self.set_param(ms_ctx_param.jit_compile, "1")
|
|
278
|
+
else:
|
|
279
|
+
self.set_param(ms_ctx_param.jit_compile, "0")
|
|
280
|
+
|
|
202
281
|
def set_backend_policy(self, policy):
|
|
203
282
|
success = self._context_handle.set_backend_policy(policy)
|
|
204
283
|
if not success:
|
|
@@ -236,7 +315,7 @@ class _Context:
|
|
|
236
315
|
def set_auto_tune_mode(self, tune_mode):
|
|
237
316
|
candidate = ["NO_TUNE", "RL", "GA", "RL,GA", "GA,RL"]
|
|
238
317
|
if tune_mode in candidate:
|
|
239
|
-
self.set_param(ms_ctx_param.
|
|
318
|
+
self.set_param(ms_ctx_param.auto_tune_mode, tune_mode)
|
|
240
319
|
else:
|
|
241
320
|
raise ValueError(f"For 'context.set_context', the argument 'auto_tune_mode' must be in "
|
|
242
321
|
f"['NO_TUNE', 'RL', 'GA', 'RL,GA', 'GA,RL'], but got {tune_mode}.")
|
|
@@ -336,14 +415,14 @@ class _Context:
|
|
|
336
415
|
json.load(f)
|
|
337
416
|
except (TypeError, ValueError) as exo:
|
|
338
417
|
raise ValueError(str(exo) + "\nFor 'context.set_context', open or load the 'env_config_path' file {} "
|
|
339
|
-
|
|
340
|
-
|
|
418
|
+
"failed, please check whether 'env_config_path' is json file and correct, "
|
|
419
|
+
"or may not have permission to read it.".format(env_config_path))
|
|
341
420
|
self.set_param(ms_ctx_param.env_config_path, env_config_path)
|
|
342
421
|
|
|
343
422
|
def set_runtime_num_threads(self, runtime_num_threads):
|
|
344
423
|
"""Check and set runtime_num_threads."""
|
|
345
|
-
if runtime_num_threads
|
|
346
|
-
raise ValueError("The num of thread must bigger than 0.")
|
|
424
|
+
if runtime_num_threads < 0:
|
|
425
|
+
raise ValueError("The num of thread must bigger than or equal to 0.")
|
|
347
426
|
self.set_param(ms_ctx_param.runtime_num_threads, runtime_num_threads)
|
|
348
427
|
|
|
349
428
|
def set_op_timeout(self, op_timeout):
|
|
@@ -352,6 +431,12 @@ class _Context:
|
|
|
352
431
|
raise ValueError("The num of op exe timeout must bigger than or equal to 0.")
|
|
353
432
|
self.set_param(ms_ctx_param.op_timeout, op_timeout)
|
|
354
433
|
|
|
434
|
+
def set_inter_op_parallel_num(self, inter_op_parallel_num):
|
|
435
|
+
"""Check and set inter_op_parallel_num."""
|
|
436
|
+
if inter_op_parallel_num < 0:
|
|
437
|
+
raise ValueError("The num of parallel thread must bigger than or equal to 0.")
|
|
438
|
+
self.set_param(ms_ctx_param.inter_op_parallel_num, inter_op_parallel_num)
|
|
439
|
+
|
|
355
440
|
setters = {
|
|
356
441
|
'mode': set_mode,
|
|
357
442
|
'save_graphs_path': set_save_graphs_path,
|
|
@@ -365,8 +450,13 @@ class _Context:
|
|
|
365
450
|
'mempool_block_size': set_mempool_block_size,
|
|
366
451
|
'print_file_path': set_print_file_path,
|
|
367
452
|
'env_config_path': set_env_config_path,
|
|
453
|
+
'inter_op_parallel_num': set_inter_op_parallel_num,
|
|
368
454
|
'runtime_num_threads': set_runtime_num_threads,
|
|
369
|
-
'
|
|
455
|
+
'memory_optimize_level': set_memory_optimize_level,
|
|
456
|
+
'op_timeout': set_op_timeout,
|
|
457
|
+
'memory_offload': set_memory_offload,
|
|
458
|
+
'deterministic': set_deterministic,
|
|
459
|
+
'ascend_config': set_ascend_config
|
|
370
460
|
}
|
|
371
461
|
|
|
372
462
|
@property
|
|
@@ -407,7 +497,6 @@ class _Context:
|
|
|
407
497
|
self._support_binary = support
|
|
408
498
|
|
|
409
499
|
|
|
410
|
-
|
|
411
500
|
def _context():
|
|
412
501
|
"""
|
|
413
502
|
Get the global _context, if context is not created, create a new one.
|
|
@@ -415,28 +504,28 @@ def _context():
|
|
|
415
504
|
Returns:
|
|
416
505
|
_Context, the global context in PyNative mode.
|
|
417
506
|
"""
|
|
418
|
-
global
|
|
419
|
-
if
|
|
507
|
+
global K_CONTEXT
|
|
508
|
+
if K_CONTEXT is None:
|
|
420
509
|
default_backend = 'debug'
|
|
421
510
|
try:
|
|
422
511
|
from mindspore import default_config
|
|
423
512
|
default_backend = default_config.__backend__
|
|
424
513
|
except ImportError:
|
|
425
514
|
logger.error("import default config fail")
|
|
426
|
-
|
|
427
|
-
|
|
515
|
+
K_CONTEXT = _Context()
|
|
516
|
+
K_CONTEXT.enable_debug_runtime = False
|
|
428
517
|
if default_backend == 'debug':
|
|
429
|
-
|
|
518
|
+
K_CONTEXT.enable_debug_runtime = True
|
|
430
519
|
default_backend = 'vm'
|
|
431
|
-
|
|
432
|
-
return
|
|
520
|
+
K_CONTEXT.set_backend_policy(default_backend)
|
|
521
|
+
return K_CONTEXT
|
|
433
522
|
|
|
434
523
|
|
|
435
524
|
@args_type_check(device_num=int, global_rank=int, gradients_mean=bool, gradient_fp32_sync=bool, parallel_mode=str,
|
|
436
525
|
auto_parallel_search_mode=str, search_mode=str, parameter_broadcast=bool, strategy_ckpt_load_file=str,
|
|
437
526
|
strategy_ckpt_save_file=str, full_batch=bool, enable_parallel_optimizer=bool, enable_alltoall=bool,
|
|
438
527
|
all_reduce_fusion_config=list, pipeline_stages=int, grad_accumulation_step=int,
|
|
439
|
-
parallel_optimizer_config=dict, comm_fusion=dict)
|
|
528
|
+
parallel_optimizer_config=dict, comm_fusion=dict, strategy_ckpt_config=dict)
|
|
440
529
|
def set_auto_parallel_context(**kwargs):
|
|
441
530
|
r"""
|
|
442
531
|
Set auto parallel context, only data parallel supported on CPU.
|
|
@@ -464,6 +553,7 @@ def set_auto_parallel_context(**kwargs):
|
|
|
464
553
|
enable_alltoall grad_accumulation_step
|
|
465
554
|
\ auto_parallel_search_mode
|
|
466
555
|
\ comm_fusion
|
|
556
|
+
\ strategy_ckpt_config
|
|
467
557
|
=========================== ===========================
|
|
468
558
|
|
|
469
559
|
Args:
|
|
@@ -502,15 +592,18 @@ def set_auto_parallel_context(**kwargs):
|
|
|
502
592
|
data_parallel mode, all parameters are broadcast except for the parameter whose attribute
|
|
503
593
|
layerwise_parallel is True. Hybrid_parallel, semi_auto_parallel and auto_parallel mode, the
|
|
504
594
|
segmented parameters do not participate in broadcasting. Default: False.
|
|
505
|
-
strategy_ckpt_load_file (str): The path to load parallel strategy checkpoint.
|
|
506
|
-
|
|
595
|
+
strategy_ckpt_load_file (str): The path to load parallel strategy checkpoint. The parameter is not to be
|
|
596
|
+
recommended currently, it is better using 'strategy_ckpt_config' to replace it. Default: ''
|
|
597
|
+
strategy_ckpt_save_file (str): The path to save parallel strategy checkpoint. The parameter is not to be
|
|
598
|
+
recommended currently, it is better using 'strategy_ckpt_config' to replace it. Default: ''
|
|
507
599
|
full_batch (bool): If you load whole batch datasets in auto_parallel mode, this parameter
|
|
508
600
|
should be set as True. Default: False. The interface is not to be recommended currently,
|
|
509
601
|
it is better using 'dataset_strategy' to replace it.
|
|
510
602
|
dataset_strategy (Union[str, tuple]): Dataset sharding strategy. Default: "data_parallel".
|
|
511
603
|
dataset_strategy="data_parallel" is equal to full_batch=False, dataset_strategy="full_batch" is
|
|
512
|
-
equal to full_batch=True. For dataset load into net by model
|
|
513
|
-
ds_stra ((1, 8), (1, 8)), it requires using
|
|
604
|
+
equal to full_batch=True. For execution mode is 'GRAPH_MODE' and dataset load into net by model
|
|
605
|
+
parallel strategy likes ds_stra ((1, 8), (1, 8)), it requires using
|
|
606
|
+
set_auto_parallel_context(dataset_strategy=ds_stra).
|
|
514
607
|
enable_parallel_optimizer (bool): This is a developing feature, which shards the weight update computation for
|
|
515
608
|
data parallel training in the benefit of time and memory saving. Currently, auto and semi auto
|
|
516
609
|
parallel mode support all optimizers in both Ascend and GPU. Data parallel mode only supports
|
|
@@ -551,6 +644,9 @@ def set_auto_parallel_context(**kwargs):
|
|
|
551
644
|
communication fusion config has two keys: "mode" and "config".
|
|
552
645
|
It supports following communication fusion types and configurations:
|
|
553
646
|
|
|
647
|
+
- openstate: Whether turn on the communication fusion or not. If `openstate` is `True`, turn on
|
|
648
|
+
the communication fusion, otherwise, turn off the communication fusion. Default: `True`.
|
|
649
|
+
|
|
554
650
|
- allreduce: If communication fusion type is `allreduce`. The `mode` contains: `auto`, `size`
|
|
555
651
|
and `index`. In `auto` mode, AllReduce fusion is configured by gradients size and the default
|
|
556
652
|
fusion threshold is `64` MB. In 'size' mode, AllReduce fusion is configured by gradients size
|
|
@@ -565,6 +661,24 @@ def set_auto_parallel_context(**kwargs):
|
|
|
565
661
|
- reducescatter: If communication fusion type is `reducescatter`. The `mode` contains: `auto`
|
|
566
662
|
and `size`. Config is same as `allgather`.
|
|
567
663
|
|
|
664
|
+
strategy_ckpt_config (dict): A dict contains the configurations for setting the parallel strategy file. This
|
|
665
|
+
interface contains the functions of parameter `strategy_ckpt_load_file` and
|
|
666
|
+
`strategy_ckpt_save_file`, it is recommonded to use this parameter to replace those two
|
|
667
|
+
parameters.
|
|
668
|
+
It contains following configurations:
|
|
669
|
+
|
|
670
|
+
- load_file (str): The path to load parallel strategy checkpoint. If the file name extension is
|
|
671
|
+
`.json`, the file is loaded in JSON format. Otherwise, the file is loaded in ProtoBuf
|
|
672
|
+
format.
|
|
673
|
+
Default: ''
|
|
674
|
+
|
|
675
|
+
- save_file (str): The path to save parallel strategy checkpoint. If the file name extension is
|
|
676
|
+
`.json`, the file is saved in JSON format. Otherwise, the file is saved in ProtoBuf format.
|
|
677
|
+
Default: ''
|
|
678
|
+
|
|
679
|
+
- only_trainable_params (bool): Only save/load the strategy information for trainable parameter.
|
|
680
|
+
Default: True.
|
|
681
|
+
|
|
568
682
|
Raises:
|
|
569
683
|
ValueError: If input key is not attribute in auto parallel context.
|
|
570
684
|
|
|
@@ -589,6 +703,8 @@ def set_auto_parallel_context(**kwargs):
|
|
|
589
703
|
>>> ms.set_auto_parallel_context(parallel_optimizer_config=parallel_config, enable_parallel_optimizer=True)
|
|
590
704
|
>>> config = {"allreduce": {"mode": "size", "config": 32}, "allgather": {"mode": "size", "config": 32}}
|
|
591
705
|
>>> ms.set_auto_parallel_context(comm_fusion=config)
|
|
706
|
+
>>> stra_ckpt_dict = {"load_file": "./stra0.ckpt", "save_file": "./stra1.ckpt", "only_trainable_params": False}
|
|
707
|
+
>>> ms.set_auto_parallel_context(strategy_ckpt_config=stra_ckpt_dict)
|
|
592
708
|
"""
|
|
593
709
|
_set_auto_parallel_context(**kwargs)
|
|
594
710
|
|
|
@@ -616,7 +732,7 @@ def get_auto_parallel_context(attr_key):
|
|
|
616
732
|
|
|
617
733
|
def reset_auto_parallel_context():
|
|
618
734
|
"""
|
|
619
|
-
Reset auto parallel context attributes to the default values
|
|
735
|
+
Reset auto parallel context attributes to the default values.
|
|
620
736
|
|
|
621
737
|
- device_num: 1.
|
|
622
738
|
- global_rank: 0.
|
|
@@ -637,6 +753,46 @@ def reset_auto_parallel_context():
|
|
|
637
753
|
_reset_auto_parallel_context()
|
|
638
754
|
|
|
639
755
|
|
|
756
|
+
@args_type_check(offload_config=dict)
|
|
757
|
+
def set_offload_context(offload_config):
|
|
758
|
+
r"""
|
|
759
|
+
Set offload context.
|
|
760
|
+
Some configurations are offload specific, see the below table for details:
|
|
761
|
+
|
|
762
|
+
Args:
|
|
763
|
+
offload_config (dict): A dict contains the keys and values for setting the offload context
|
|
764
|
+
configure.It supports the following keys.
|
|
765
|
+
enable_offload (bool): The flag of whether enabling offload. Default: False.
|
|
766
|
+
offload_param (str): The param for offload destination, cpu or disk.
|
|
767
|
+
offload_path (str): The path of offload.
|
|
768
|
+
offload_checkpoint (str): The checkpoint for offload destination, cpu or disk.
|
|
769
|
+
offload_ddr_size (int): The ddr size for offload.
|
|
770
|
+
offload_disk_size (int): The disk size for offload.
|
|
771
|
+
enable_aio (bool): The flag of whether enabling aio. Default: True.
|
|
772
|
+
aio_block_size (int): The size of aio block.
|
|
773
|
+
aio_queue_depth (int): The depth of aio queue.
|
|
774
|
+
enable_pinned_mem (bool): The flag of whether enabling pinned memory.
|
|
775
|
+
|
|
776
|
+
Raises:
|
|
777
|
+
ValueError: If input key is not attribute in auto parallel context.
|
|
778
|
+
|
|
779
|
+
Examples:
|
|
780
|
+
>>> from mindspore import context
|
|
781
|
+
>>> context.set_offload_context(offload_config={"offload_param"="cpu"})
|
|
782
|
+
"""
|
|
783
|
+
_set_offload_context(offload_config)
|
|
784
|
+
|
|
785
|
+
|
|
786
|
+
def get_offload_context():
|
|
787
|
+
"""
|
|
788
|
+
Get offload context.
|
|
789
|
+
Examples:
|
|
790
|
+
>>> from mindspore import context
|
|
791
|
+
>>> offload_config = context.get_offload_context()
|
|
792
|
+
"""
|
|
793
|
+
return _get_offload_context()
|
|
794
|
+
|
|
795
|
+
|
|
640
796
|
def _check_target_specific_cfgs(device, arg_key):
|
|
641
797
|
"""Checking whether a config is suitable for a specified device"""
|
|
642
798
|
device_cfgs = {
|
|
@@ -648,7 +804,8 @@ def _check_target_specific_cfgs(device, arg_key):
|
|
|
648
804
|
'auto_tune_mode': ['Ascend'],
|
|
649
805
|
'max_device_memory': ['Ascend', 'GPU'],
|
|
650
806
|
'mempool_block_size': ['GPU', 'Ascend'],
|
|
651
|
-
'disable_format_transform': ['GPU']
|
|
807
|
+
'disable_format_transform': ['GPU'],
|
|
808
|
+
'ascend_config': ['Ascend']
|
|
652
809
|
}
|
|
653
810
|
# configs not in map device_cfgs are supposed to be suitable for all devices
|
|
654
811
|
if arg_key not in device_cfgs:
|
|
@@ -662,16 +819,15 @@ def _check_target_specific_cfgs(device, arg_key):
|
|
|
662
819
|
return False
|
|
663
820
|
|
|
664
821
|
|
|
665
|
-
@
|
|
666
|
-
@args_type_check(mode=int, precompile_only=bool, device_target=str, device_id=int, save_graphs=bool,
|
|
822
|
+
@args_type_check(mode=int, precompile_only=bool, device_target=str, device_id=int, save_graphs=(bool, int),
|
|
667
823
|
save_graphs_path=str, enable_dump=bool, auto_tune_mode=str,
|
|
668
824
|
save_dump_path=str, enable_reduce_precision=bool, variable_memory_max_size=str,
|
|
669
|
-
enable_auto_mixed_precision=bool,
|
|
825
|
+
enable_auto_mixed_precision=bool, inter_op_parallel_num=int,
|
|
670
826
|
enable_graph_kernel=bool, reserve_class_name_in_scope=bool, check_bprop=bool,
|
|
671
827
|
max_device_memory=str, print_file_path=str, max_call_depth=int, env_config_path=str,
|
|
672
828
|
graph_kernel_flags=str, save_compile_cache=bool, runtime_num_threads=int, load_compile_cache=bool,
|
|
673
829
|
grad_for_scalar=bool, pynative_synchronize=bool, mempool_block_size=str, disable_format_transform=bool,
|
|
674
|
-
op_timeout=int)
|
|
830
|
+
op_timeout=int, deterministic=str, ascend_config=dict)
|
|
675
831
|
def set_context(**kwargs):
|
|
676
832
|
"""
|
|
677
833
|
Set context for running environment.
|
|
@@ -682,7 +838,7 @@ def set_context(**kwargs):
|
|
|
682
838
|
Note:
|
|
683
839
|
Attribute name is required for setting attributes.
|
|
684
840
|
The mode is not recommended to be changed after net was initialized because the implementations of some
|
|
685
|
-
operations are different in graph mode and pynative mode. Default:
|
|
841
|
+
operations are different in graph mode and pynative mode. Default: PYNATIVE_MODE.
|
|
686
842
|
|
|
687
843
|
Some configurations are device specific, see the below table for details:
|
|
688
844
|
|
|
@@ -709,6 +865,8 @@ def set_context(**kwargs):
|
|
|
709
865
|
| +------------------------------+----------------------------+
|
|
710
866
|
| | save_dump_path | Ascend |
|
|
711
867
|
| +------------------------------+----------------------------+
|
|
868
|
+
| | deterministic | Ascend |
|
|
869
|
+
| +------------------------------+----------------------------+
|
|
712
870
|
| | print_file_path | Ascend |
|
|
713
871
|
| +------------------------------+----------------------------+
|
|
714
872
|
| | env_config_path | CPU/GPU/Ascend |
|
|
@@ -737,6 +895,8 @@ def set_context(**kwargs):
|
|
|
737
895
|
| +------------------------------+----------------------------+
|
|
738
896
|
| | enable_compile_cache | CPU/GPU/Ascend |
|
|
739
897
|
| +------------------------------+----------------------------+
|
|
898
|
+
| | inter_op_parallel_num | CPU/GPU/Ascend |
|
|
899
|
+
| +------------------------------+----------------------------+
|
|
740
900
|
| | runtime_num_threads | CPU/GPU/Ascend |
|
|
741
901
|
| +------------------------------+----------------------------+
|
|
742
902
|
| | compile_cache_path | CPU/GPU/Ascend |
|
|
@@ -744,6 +904,12 @@ def set_context(**kwargs):
|
|
|
744
904
|
| | disable_format_transform | GPU |
|
|
745
905
|
| +------------------------------+----------------------------+
|
|
746
906
|
| | support_binary | CPU/GPU/Ascend |
|
|
907
|
+
| +------------------------------+----------------------------+
|
|
908
|
+
| | memory_optimize_level | CPU/GPU/Ascend |
|
|
909
|
+
| +------------------------------+----------------------------+
|
|
910
|
+
| | memory_offload | GPU/Ascend |
|
|
911
|
+
| +------------------------------+----------------------------+
|
|
912
|
+
| | ascend_config | Ascend |
|
|
747
913
|
+-------------------------+------------------------------+----------------------------+
|
|
748
914
|
|
|
749
915
|
Args:
|
|
@@ -760,14 +926,31 @@ def set_context(**kwargs):
|
|
|
760
926
|
of the available memory of the device and mempool_block_size.
|
|
761
927
|
op_timeout (int): Set the maximum duration of executing an operator in seconds.
|
|
762
928
|
If the execution time exceeds this value, system will terminate the task. 0 means endless wait.
|
|
763
|
-
Default:
|
|
764
|
-
save_graphs (bool): Whether to save graphs. Default:
|
|
765
|
-
|
|
766
|
-
|
|
929
|
+
Default: 1900.
|
|
930
|
+
save_graphs (bool or int): Whether to save intermediate compilation graphs. Default: 0.
|
|
931
|
+
Available values are:
|
|
932
|
+
|
|
933
|
+
- False or 0: disable saving of intermediate compilation graphs.
|
|
934
|
+
- 1: some intermediate files will be generated during graph compliation.
|
|
935
|
+
- True or 2: Generate more ir files related to backend process.
|
|
936
|
+
- 3: Generate visualization computing graphs and detailed frontend ir graphs.
|
|
937
|
+
|
|
938
|
+
When the `save_graphs` attribute is set as True, 1, 2 or 3, attribute of `save_graphs_path` is used
|
|
939
|
+
to set the intermediate compilation graph storage path. By default, the graphs are saved in the current
|
|
940
|
+
directory.
|
|
767
941
|
save_graphs_path (str): Path to save graphs. Default: ".".
|
|
768
942
|
If the specified directory does not exist, the system will automatically create the directory.
|
|
769
943
|
During distributed training, graphs will be saved to the directory of
|
|
770
944
|
`save_graphs_path/rank_${rank_id}/`. `rank_id` is the ID of the current device in the cluster.
|
|
945
|
+
deterministic (str): Whether to enable op run in deterministic mode. The value must be in the
|
|
946
|
+
range of ['ON', 'OFF'], and the default value is 'OFF'.
|
|
947
|
+
|
|
948
|
+
- "ON": Enable operator deterministic running mode.
|
|
949
|
+
- "OFF": Disable operator deterministic running mode.
|
|
950
|
+
|
|
951
|
+
When deterministic mode is on, model ops will be deterministic in Ascend. This means that if op run multiple
|
|
952
|
+
times with the same inputs on the same hardware, it will have the exact same outputs each time. This is
|
|
953
|
+
useful for debugging models.
|
|
771
954
|
enable_dump (bool): This parameters is deprecated, and will be deleted in the next version.
|
|
772
955
|
save_dump_path (str): This parameters is deprecated, and will be deleted in the next version.
|
|
773
956
|
print_file_path (str): The path of saving print data. If this parameter is set, print data is saved to
|
|
@@ -785,7 +968,7 @@ def set_context(**kwargs):
|
|
|
785
968
|
When set to false, the RDR will be turned off.
|
|
786
969
|
- mode: sets the mode of RDR on exporting data. When set to 1, the RDR only exports data
|
|
787
970
|
in the fault scenario. When set to 2, the RDR exports data in the fault scenario and the
|
|
788
|
-
normal end scenario. Default
|
|
971
|
+
normal end scenario. Default: 1.
|
|
789
972
|
- path: sets the path where RDR saves data. The current path must be absolute.
|
|
790
973
|
|
|
791
974
|
Memory reuse:
|
|
@@ -810,16 +993,15 @@ def set_context(**kwargs):
|
|
|
810
993
|
be located, when the value is set to True, the operator is executed synchronously on the device. It will
|
|
811
994
|
reduce the execution performance of the program. At this time, when an error occurs in the execution of
|
|
812
995
|
the operator, the location of the error script code can be located according to the call stack of the error.
|
|
813
|
-
mode (int): Running in GRAPH_MODE(0) or PYNATIVE_MODE(1).
|
|
814
|
-
|
|
815
|
-
mode is GRAPH_MODE.
|
|
996
|
+
mode (int): Running in GRAPH_MODE(0) or PYNATIVE_MODE(1).
|
|
997
|
+
Both modes support all backends. Default: PYNATIVE_MODE.
|
|
816
998
|
enable_graph_kernel (bool): Whether to enable graph kernel fusion to optimize network execution performance.
|
|
817
999
|
Default: False.
|
|
818
1000
|
Indicates whether to enable image-computing convergence to optimize network execution performance.
|
|
819
1001
|
If enable_graph_kernel is set to True, acceleration can be enabled.
|
|
820
1002
|
For details of graph kernel fusion, please check
|
|
821
1003
|
`Enabling Graph Kernel Fusion
|
|
822
|
-
<https://www.mindspore.cn/tutorials/experts/en/
|
|
1004
|
+
<https://www.mindspore.cn/tutorials/experts/en/r2.0/debug/graph_fusion_engine.html>`_.
|
|
823
1005
|
graph_kernel_flags (str):
|
|
824
1006
|
Optimization options of graph kernel fusion, and the priority is higher when it conflicts
|
|
825
1007
|
with enable_graph_kernel. Only for experienced users.
|
|
@@ -854,7 +1036,7 @@ def set_context(**kwargs):
|
|
|
854
1036
|
|
|
855
1037
|
For more information about the enable operator tuning tool settings, please check
|
|
856
1038
|
`Enable the operator optimization tool
|
|
857
|
-
<https://www.mindspore.cn/tutorials/experts/en/
|
|
1039
|
+
<https://www.mindspore.cn/tutorials/experts/en/r2.0/debug/auto_tune.html>`_.
|
|
858
1040
|
check_bprop (bool): Whether to check back propagation nodes. The checking ensures that the shape and dtype
|
|
859
1041
|
of back propagation node outputs is the same as input parameters. Default: False.
|
|
860
1042
|
max_call_depth (int): Specify the maximum depth of function call. Must be positive integer. Default: 1000.
|
|
@@ -872,12 +1054,14 @@ def set_context(**kwargs):
|
|
|
872
1054
|
the compile cache is loaded. Note that only limited automatic detection for the changes of
|
|
873
1055
|
python scripts is supported by now, which means that there is a correctness risk. Default: False.
|
|
874
1056
|
This is an experimental prototype that is subject to change and/or deletion.
|
|
875
|
-
compile_cache_path (str): Path to save the cache
|
|
1057
|
+
compile_cache_path (str): Path to save the compile cache. Default: ".".
|
|
876
1058
|
If the specified directory does not exist, the system will automatically create the directory.
|
|
877
1059
|
The cache will be saved to the directory of `compile_cache_path/rank_${rank_id}/`. The `rank_id` is
|
|
878
1060
|
the ID of the current device in the cluster.
|
|
879
|
-
|
|
880
|
-
which
|
|
1061
|
+
inter_op_parallel_num(int): The thread number of op parallel at the same time. Default value is 0,
|
|
1062
|
+
which means use the default num.
|
|
1063
|
+
runtime_num_threads(int): The thread pool number of cpu kernel used in runtime,
|
|
1064
|
+
which must bigger than or equal to 0. Default value is 30, if you run many processes at
|
|
881
1065
|
the same time, you should set the value smaller to avoid thread contention.
|
|
882
1066
|
disable_format_transform (bool): Whether to disable the automatic format transform function from NCHW to NHWC.
|
|
883
1067
|
When the network training performance of fp16 is worse than fp32,
|
|
@@ -886,6 +1070,49 @@ def set_context(**kwargs):
|
|
|
886
1070
|
in graph mode, coulde set 'support_binary' to be True, and run once .py file. It would save the source
|
|
887
1071
|
of the interfaces would be compiled by MindSpore to the interfaces definition .py file that should be
|
|
888
1072
|
guaranteed to be writable. Then compile the .py file to the .pyc or .so file, and could run in Graph mode.
|
|
1073
|
+
memory_optimize_level (str): The memory optimize level.
|
|
1074
|
+
Default: O0. The value must be in ['O0', 'O1'].
|
|
1075
|
+
|
|
1076
|
+
- O0: priority performance option, disable SOMAS (Safe Optimized Memory Allocation Solver).
|
|
1077
|
+
- O1: priority memory option, enable SOMAS.
|
|
1078
|
+
memory_offload (str): Whether to enable the memory offload function. When it is enabled, the idle data will be
|
|
1079
|
+
temporarily copied to the host side in the case of insufficient device memory. The value must be in the
|
|
1080
|
+
range of ['ON', 'OFF'], and the default value is 'OFF'.
|
|
1081
|
+
|
|
1082
|
+
- ON: Enable the memory Offload function. On Ascend hardware platform, this parameter does not take effect
|
|
1083
|
+
when the environment variable "GRAPH_OP_RUN=1" is not set; This parameter does not take effect when
|
|
1084
|
+
memory_optimize_level is set 'O1'.
|
|
1085
|
+
- OFF: Turn off the memory Offload function.
|
|
1086
|
+
ascend_config (dict): Set the parameters specific to Ascend hardware platform. It is not set by default.
|
|
1087
|
+
Currently, only setting `precision_mode` and `jit_compile` are supported on Ascend910B hardware platform.
|
|
1088
|
+
The default value of `precision_mode` and `jit_compile` are experimental parameters, may change
|
|
1089
|
+
in the future.
|
|
1090
|
+
|
|
1091
|
+
- precision_mode (str): Mixed precision mode setting, on Ascend910B hardware platform, the default
|
|
1092
|
+
value of training network is based on the value of CANN, and the default value of inference network
|
|
1093
|
+
is force_fp16. The value range is as follows:
|
|
1094
|
+
|
|
1095
|
+
- force_fp16: When the operator supports both float16 and float32, select float16 directly.
|
|
1096
|
+
- allow_fp32_to_fp16: When the operator does not support the float32 data type, directly reduce
|
|
1097
|
+
the precision of float16.
|
|
1098
|
+
- allow_mix_precision: Automatic mixing precision, facing the whole network operator, according
|
|
1099
|
+
to the built-in optimization strategy, automatically reduces the precision of some operators
|
|
1100
|
+
to float16 or bfloat16.
|
|
1101
|
+
- must_keep_origin_dtype: Keep the accuracy of the original drawing.
|
|
1102
|
+
- force_fp32: When the input of the matrix calculation operator is float16 and the output supports
|
|
1103
|
+
float16 and float32, output is forced to float32.
|
|
1104
|
+
- force_lowerprecision: When the operator supports both float16 or bfloat16 and float32, select
|
|
1105
|
+
float16 or bfloat16 directly.
|
|
1106
|
+
- allow_fp32_to_bf16: When the operator does not support the float32 data type, directly reduce
|
|
1107
|
+
the precision of bfloat16.
|
|
1108
|
+
- allow_fp32_to_lowprecision: When the operator does not support the float32 data type, directly
|
|
1109
|
+
reduce the precision of float16 or bfloat16.
|
|
1110
|
+
- allow_mix_precision_fp16: Automatic mixing precision, facing the whole network operator, automatically
|
|
1111
|
+
reduces the precision of some operators to float16 according to the built-in optimization strategy.
|
|
1112
|
+
- allow_mix_precision_bf16: Automatic mixing precision, facing the whole network operator, according to
|
|
1113
|
+
the built-in optimization strategy, automatically reduces the precision of some operators to bfloat16.
|
|
1114
|
+
|
|
1115
|
+
- jit_compile (bool): Whether to select online compilation. the default value is based on CANN.
|
|
889
1116
|
|
|
890
1117
|
Raises:
|
|
891
1118
|
ValueError: If input key is not an attribute in context.
|
|
@@ -913,16 +1140,17 @@ def set_context(**kwargs):
|
|
|
913
1140
|
>>> ms.set_context(enable_compile_cache=True, compile_cache_path="./cache.ms")
|
|
914
1141
|
>>> ms.set_context(pynative_synchronize=True)
|
|
915
1142
|
>>> ms.set_context(runtime_num_threads=10)
|
|
1143
|
+
>>> ms.set_context(inter_op_parallel_num=4)
|
|
916
1144
|
>>> ms.set_context(disable_format_transform=True)
|
|
1145
|
+
>>> ms.set_context(memory_optimize_level='O0')
|
|
1146
|
+
>>> ms.set_context(memory_offload='ON')
|
|
1147
|
+
>>> ms.set_context(deterministic='ON')
|
|
1148
|
+
>>> ms.set_context(ascend_config={"precision_mode": "force_fp16", "jit_compile": True})
|
|
917
1149
|
"""
|
|
918
1150
|
ctx = _context()
|
|
919
1151
|
# set device target first
|
|
920
1152
|
if 'device_target' in kwargs:
|
|
921
1153
|
ctx.set_device_target(kwargs['device_target'])
|
|
922
|
-
device = ctx.get_param(ms_ctx_param.device_target)
|
|
923
|
-
if not device.lower() in __device_target__:
|
|
924
|
-
raise ValueError(f"For 'context.set_context', package type {__package_name__} support 'device_target' "
|
|
925
|
-
f"type {__device_target__}, but got {device}.")
|
|
926
1154
|
device = ctx.get_param(ms_ctx_param.device_target)
|
|
927
1155
|
for key, value in kwargs.items():
|
|
928
1156
|
if key == 'enable_sparse':
|
|
@@ -933,6 +1161,15 @@ def set_context(**kwargs):
|
|
|
933
1161
|
logger.warning(f"For 'context.set_context', '{key}' parameter is deprecated. "
|
|
934
1162
|
"For details, please see the interface parameter API comments")
|
|
935
1163
|
continue
|
|
1164
|
+
if key in ('precision_mode', 'jit_compile'):
|
|
1165
|
+
raise ValueError(f"Please set '{key}' through parameter ascend_config")
|
|
1166
|
+
if key == 'save_graphs':
|
|
1167
|
+
if value is True:
|
|
1168
|
+
value = 2
|
|
1169
|
+
if value is False:
|
|
1170
|
+
value = 0
|
|
1171
|
+
if value > 3:
|
|
1172
|
+
raise ValueError(f"value for save_graphs should be 0-3 but got '{value}'")
|
|
936
1173
|
if not _check_target_specific_cfgs(device, key):
|
|
937
1174
|
continue
|
|
938
1175
|
if hasattr(ctx, key):
|
|
@@ -1020,6 +1257,7 @@ def set_ps_context(**kwargs):
|
|
|
1020
1257
|
Set parameter server training mode context.
|
|
1021
1258
|
|
|
1022
1259
|
Note:
|
|
1260
|
+
Parameter server mode is only supported in graph mode.
|
|
1023
1261
|
Some other environment variables should also be set for parameter server training mode.
|
|
1024
1262
|
These environment variables are listed below:
|
|
1025
1263
|
|
|
@@ -1104,9 +1342,6 @@ _hccl_connect_timeout = '600'
|
|
|
1104
1342
|
|
|
1105
1343
|
def _init_parallel_env():
|
|
1106
1344
|
"""Set hccl connect timeout."""
|
|
1107
|
-
if 'ascend' not in __device_target__:
|
|
1108
|
-
return
|
|
1109
|
-
|
|
1110
1345
|
if 'HCCL_CONNECT_TIMEOUT' not in os.environ:
|
|
1111
1346
|
os.environ['HCCL_CONNECT_TIMEOUT'] = _hccl_connect_timeout
|
|
1112
1347
|
|
|
Binary file
|