mindspore 1.10.0__cp37-cp37m-win_amd64.whl → 2.0.0rc1__cp37-cp37m-win_amd64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of mindspore might be problematic. Click here for more details.
- mindspore/.commit_id +1 -1
- mindspore/ConcurrencyCheck.dll +0 -0
- mindspore/CppBuildInsights.dll +0 -0
- mindspore/CppCoreCheck.dll +0 -0
- mindspore/EnumIndex.dll +0 -0
- mindspore/EspXEngine.dll +0 -0
- mindspore/HResultCheck.dll +0 -0
- mindspore/KernelTraceControl.dll +0 -0
- mindspore/LocalESPC.dll +0 -0
- mindspore/Microsoft.Diagnostics.Tracing.EventSource.dll +0 -0
- mindspore/Microsoft.VisualStudio.RemoteControl.dll +0 -0
- mindspore/Microsoft.VisualStudio.Telemetry.dll +0 -0
- mindspore/Microsoft.VisualStudio.Utilities.Internal.dll +0 -0
- mindspore/Newtonsoft.Json.dll +0 -0
- mindspore/System.Runtime.CompilerServices.Unsafe.dll +0 -0
- mindspore/VariantClear.dll +0 -0
- mindspore/__init__.py +9 -4
- mindspore/_c_dataengine.cp37-win_amd64.pyd +0 -0
- mindspore/_c_expression.cp37-win_amd64.pyd +0 -0
- mindspore/_c_mindrecord.cp37-win_amd64.pyd +0 -0
- mindspore/_check_jit_forbidden_api.py +102 -0
- mindspore/_checkparam.py +1066 -1001
- mindspore/_extends/builtin_operations.py +32 -4
- mindspore/_extends/graph_kernel/model/graph_split.py +66 -222
- mindspore/_extends/parallel_compile/akg_compiler/akg_process.py +12 -9
- mindspore/_extends/parallel_compile/akg_compiler/build_tbe_kernel.py +119 -26
- mindspore/_extends/parallel_compile/akg_compiler/tbe_topi.py +50 -50
- mindspore/_extends/parallel_compile/akg_compiler/util.py +9 -6
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_adapter.py +4 -25
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_helper.py +9 -4
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_job_manager.py +1 -27
- mindspore/_extends/parse/__init__.py +5 -3
- mindspore/_extends/parse/namespace.py +17 -2
- mindspore/_extends/parse/parser.py +193 -34
- mindspore/_extends/parse/resources.py +7 -8
- mindspore/_extends/parse/standard_method.py +1780 -435
- mindspore/_extends/parse/trope.py +3 -1
- mindspore/amp.py +53 -58
- mindspore/atlprov.dll +0 -0
- mindspore/boost/adasum.py +3 -2
- mindspore/boost/boost.py +2 -2
- mindspore/boost/boost_cell_wrapper.py +46 -26
- mindspore/boost/dim_reduce.py +6 -5
- mindspore/boost/grad_accumulation.py +2 -1
- mindspore/boost/group_loss_scale_manager.py +1 -1
- mindspore/c1.dll +0 -0
- mindspore/c1xx.dll +0 -0
- mindspore/c2.dll +0 -0
- mindspore/cfgpersist.dll +0 -0
- mindspore/clang_rt.asan_dbg_dynamic-x86_64.dll +0 -0
- mindspore/clang_rt.asan_dynamic-x86_64.dll +0 -0
- mindspore/common/__init__.py +11 -10
- mindspore/common/_decorator.py +2 -0
- mindspore/common/_register_for_adapter.py +55 -0
- mindspore/common/_stub_tensor.py +201 -0
- mindspore/common/_utils.py +57 -0
- mindspore/common/api.py +582 -297
- mindspore/common/dtype.py +66 -18
- mindspore/common/dump.py +2 -2
- mindspore/common/initializer.py +38 -1
- mindspore/common/jit_config.py +25 -13
- mindspore/common/mutable.py +53 -24
- mindspore/common/parameter.py +60 -37
- mindspore/common/seed.py +8 -24
- mindspore/common/sparse_tensor.py +927 -0
- mindspore/common/tensor.py +1627 -3900
- mindspore/communication/__init__.py +10 -5
- mindspore/communication/_comm_helper.py +78 -214
- mindspore/communication/_hccl_management.py +2 -1
- mindspore/communication/management.py +136 -47
- mindspore/config/op_info.config +501 -1008
- mindspore/context.py +291 -56
- mindspore/d3dcompiler_47.dll +0 -0
- mindspore/dataset/__init__.py +12 -8
- mindspore/dataset/audio/__init__.py +9 -9
- mindspore/dataset/audio/transforms.py +1090 -228
- mindspore/dataset/audio/utils.py +87 -39
- mindspore/dataset/audio/validators.py +223 -1
- mindspore/dataset/callback/ds_callback.py +17 -15
- mindspore/dataset/core/config.py +246 -17
- mindspore/dataset/core/py_util_helpers.py +4 -3
- mindspore/dataset/core/validator_helpers.py +10 -10
- mindspore/{parallel/nn/layers.py → dataset/debug/__init__.py} +7 -8
- mindspore/dataset/debug/debug_hook.py +65 -0
- mindspore/dataset/debug/pre_defined_hook.py +67 -0
- mindspore/dataset/engine/__init__.py +7 -3
- mindspore/dataset/engine/cache_client.py +9 -9
- mindspore/dataset/engine/datasets.py +648 -477
- mindspore/dataset/engine/datasets_audio.py +165 -167
- mindspore/dataset/engine/datasets_standard_format.py +93 -67
- mindspore/dataset/engine/datasets_text.py +492 -342
- mindspore/dataset/engine/datasets_user_defined.py +85 -50
- mindspore/dataset/engine/datasets_vision.py +1224 -699
- mindspore/dataset/engine/graphdata.py +134 -69
- mindspore/dataset/engine/iterators.py +50 -9
- mindspore/dataset/engine/offload.py +52 -31
- mindspore/dataset/engine/samplers.py +27 -24
- mindspore/dataset/engine/serializer_deserializer.py +14 -15
- mindspore/dataset/engine/validators.py +213 -52
- mindspore/dataset/text/__init__.py +10 -8
- mindspore/dataset/text/transforms.py +152 -57
- mindspore/dataset/text/utils.py +98 -49
- mindspore/dataset/text/validators.py +25 -0
- mindspore/dataset/transforms/__init__.py +4 -2
- mindspore/dataset/transforms/c_transforms.py +11 -13
- mindspore/dataset/transforms/py_transforms.py +2 -2
- mindspore/dataset/transforms/py_transforms_util.py +10 -0
- mindspore/dataset/transforms/transforms.py +13 -15
- mindspore/dataset/transforms/validators.py +7 -7
- mindspore/dataset/utils/__init__.py +2 -1
- mindspore/dataset/utils/browse_dataset.py +13 -13
- mindspore/dataset/utils/line_reader.py +121 -0
- mindspore/dataset/vision/__init__.py +8 -7
- mindspore/dataset/vision/c_transforms.py +125 -126
- mindspore/dataset/vision/py_transforms.py +37 -37
- mindspore/dataset/vision/py_transforms_util.py +23 -20
- mindspore/dataset/vision/transforms.py +316 -315
- mindspore/dataset/vision/utils.py +313 -17
- mindspore/dataset/vision/validators.py +6 -6
- mindspore/default_config.py +0 -1
- mindspore/dpcmi.dll +0 -0
- mindspore/{compression → experimental}/__init__.py +6 -5
- mindspore/experimental/map_parameter.py +275 -0
- mindspore/include/OWNERS +0 -1
- mindspore/include/api/callback/callback.h +9 -13
- mindspore/include/api/callback/ckpt_saver.h +2 -2
- mindspore/include/api/callback/loss_monitor.h +2 -2
- mindspore/include/api/callback/lr_scheduler.h +5 -5
- mindspore/include/api/callback/time_monitor.h +2 -2
- mindspore/include/api/callback/train_accuracy.h +4 -6
- mindspore/include/api/cfg.h +19 -6
- mindspore/include/api/context.h +70 -9
- mindspore/include/api/delegate.h +8 -1
- mindspore/include/api/dual_abi_helper.h +8 -24
- mindspore/include/api/metrics/accuracy.h +2 -2
- mindspore/include/api/metrics/metrics.h +4 -3
- mindspore/include/api/model.h +9 -4
- mindspore/include/api/model_group.h +68 -0
- mindspore/include/api/model_parallel_runner.h +17 -17
- mindspore/include/api/net.h +12 -11
- mindspore/include/api/serialization.h +20 -4
- mindspore/include/api/status.h +7 -1
- mindspore/include/api/types.h +25 -21
- mindspore/include/api/visible.h +4 -0
- mindspore/include/c_api/model_c.h +5 -0
- mindspore/include/c_api/status_c.h +1 -1
- mindspore/include/dataset/config.h +1 -1
- mindspore/include/dataset/constants.h +14 -0
- mindspore/include/dataset/text.h +59 -0
- mindspore/include/dataset/vision.h +56 -117
- mindspore/include/dataset/vision_lite.h +102 -0
- mindspore/jpeg62.dll +0 -0
- mindspore/log.py +28 -28
- mindspore/mindrecord/common/exceptions.py +2 -4
- mindspore/mindrecord/filereader.py +19 -1
- mindspore/mindrecord/filewriter.py +250 -88
- mindspore/mindrecord/mindpage.py +13 -13
- mindspore/mindrecord/shardheader.py +15 -15
- mindspore/mindrecord/shardreader.py +9 -0
- mindspore/mindrecord/shardwriter.py +29 -29
- mindspore/mindrecord/tools/cifar100_to_mr.py +9 -9
- mindspore/mindrecord/tools/cifar10_to_mr.py +9 -9
- mindspore/mindrecord/tools/csv_to_mr.py +4 -4
- mindspore/mindrecord/tools/imagenet_to_mr.py +70 -65
- mindspore/mindrecord/tools/mnist_to_mr.py +41 -41
- mindspore/mindrecord/tools/tfrecord_to_mr.py +6 -6
- mindspore/{libmindspore_backend.dll → mindspore_backend.dll} +0 -0
- mindspore/mindspore_common.dll +0 -0
- mindspore/mindspore_core.dll +0 -0
- mindspore/mindspore_glog.dll +0 -0
- mindspore/mindspore_shared_lib.dll +0 -0
- mindspore/msobj140.dll +0 -0
- mindspore/mspdb140.dll +0 -0
- mindspore/mspdbcore.dll +0 -0
- mindspore/mspdbst.dll +0 -0
- mindspore/mspft140.dll +0 -0
- mindspore/msvcdis140.dll +0 -0
- mindspore/msvcp140_1.dll +0 -0
- mindspore/msvcp140_2.dll +0 -0
- mindspore/msvcp140_atomic_wait.dll +0 -0
- mindspore/msvcp140_codecvt_ids.dll +0 -0
- mindspore/nn/__init__.py +1 -5
- mindspore/nn/cell.py +297 -234
- mindspore/nn/dynamic_lr.py +1 -1
- mindspore/nn/grad/cell_grad.py +17 -42
- mindspore/nn/layer/__init__.py +7 -4
- mindspore/nn/layer/activation.py +131 -88
- mindspore/nn/layer/basic.py +313 -613
- mindspore/nn/layer/channel_shuffle.py +103 -0
- mindspore/nn/layer/combined.py +1 -1
- mindspore/nn/layer/container.py +52 -6
- mindspore/nn/layer/conv.py +112 -43
- mindspore/nn/layer/dense.py +10 -9
- mindspore/nn/layer/embedding.py +36 -34
- mindspore/nn/layer/image.py +123 -27
- mindspore/nn/layer/math.py +108 -107
- mindspore/nn/layer/normalization.py +212 -366
- mindspore/nn/layer/padding.py +370 -42
- mindspore/nn/layer/pooling.py +1443 -219
- mindspore/nn/layer/rnn_cells.py +11 -16
- mindspore/nn/layer/rnns.py +38 -39
- mindspore/nn/layer/thor_layer.py +24 -25
- mindspore/nn/layer/timedistributed.py +5 -5
- mindspore/nn/layer/transformer.py +701 -0
- mindspore/nn/learning_rate_schedule.py +8 -8
- mindspore/nn/loss/__init__.py +9 -6
- mindspore/nn/loss/loss.py +678 -142
- mindspore/nn/metrics.py +53 -0
- mindspore/nn/optim/_dist_optimizer_registry.py +2 -2
- mindspore/nn/optim/ada_grad.py +8 -8
- mindspore/nn/optim/adadelta.py +2 -3
- mindspore/nn/optim/adafactor.py +18 -14
- mindspore/nn/optim/adam.py +429 -87
- mindspore/nn/optim/adamax.py +5 -6
- mindspore/nn/optim/adasum.py +10 -8
- mindspore/nn/optim/asgd.py +7 -7
- mindspore/nn/optim/ftrl.py +81 -11
- mindspore/nn/optim/lamb.py +7 -8
- mindspore/nn/optim/lars.py +4 -4
- mindspore/nn/optim/lazyadam.py +82 -7
- mindspore/nn/optim/momentum.py +8 -7
- mindspore/nn/optim/optimizer.py +19 -10
- mindspore/nn/optim/proximal_ada_grad.py +6 -5
- mindspore/nn/optim/rmsprop.py +3 -3
- mindspore/nn/optim/rprop.py +20 -16
- mindspore/nn/optim/sgd.py +21 -15
- mindspore/nn/optim/thor.py +23 -21
- mindspore/nn/probability/__init__.py +0 -2
- mindspore/nn/probability/bijector/bijector.py +7 -6
- mindspore/nn/probability/bijector/invert.py +4 -2
- mindspore/nn/probability/bijector/softplus.py +2 -2
- mindspore/nn/probability/bnn_layers/dense_variational.py +1 -1
- mindspore/nn/probability/bnn_layers/layer_distribution.py +2 -2
- mindspore/nn/probability/distribution/__init__.py +6 -0
- mindspore/nn/probability/distribution/_utils/custom_ops.py +3 -2
- mindspore/nn/probability/distribution/_utils/utils.py +11 -17
- mindspore/nn/probability/distribution/bernoulli.py +6 -6
- mindspore/nn/probability/distribution/beta.py +1 -1
- mindspore/nn/probability/distribution/categorical.py +9 -9
- mindspore/nn/probability/distribution/cauchy.py +8 -8
- mindspore/nn/probability/distribution/distribution.py +12 -6
- mindspore/nn/probability/distribution/exponential.py +5 -5
- mindspore/nn/probability/distribution/gamma.py +3 -3
- mindspore/nn/probability/distribution/geometric.py +6 -5
- mindspore/nn/probability/distribution/gumbel.py +5 -5
- mindspore/nn/probability/distribution/half_normal.py +133 -0
- mindspore/nn/probability/distribution/laplace.py +128 -0
- mindspore/nn/probability/distribution/log_normal.py +0 -1
- mindspore/nn/probability/distribution/logistic.py +4 -5
- mindspore/nn/probability/distribution/normal.py +11 -15
- mindspore/nn/probability/distribution/poisson.py +6 -2
- mindspore/nn/probability/distribution/student_t.py +150 -0
- mindspore/nn/probability/distribution/transformed_distribution.py +4 -4
- mindspore/nn/probability/distribution/uniform.py +5 -5
- mindspore/nn/reinforcement/_tensors_queue.py +3 -3
- mindspore/nn/reinforcement/tensor_array.py +2 -2
- mindspore/nn/sparse/sparse.py +8 -1
- mindspore/nn/wrap/cell_wrapper.py +55 -27
- mindspore/nn/wrap/grad_reducer.py +20 -11
- mindspore/nn/wrap/loss_scale.py +47 -30
- mindspore/numpy/array_creations.py +33 -22
- mindspore/numpy/array_ops.py +46 -42
- mindspore/numpy/logic_ops.py +6 -27
- mindspore/numpy/math_ops.py +26 -19
- mindspore/numpy/utils.py +1 -8
- mindspore/numpy/utils_const.py +112 -62
- mindspore/opencv_core452.dll +0 -0
- mindspore/opencv_imgcodecs452.dll +0 -0
- mindspore/opencv_imgproc452.dll +0 -0
- mindspore/ops/__init__.py +6 -3
- mindspore/ops/_constants.py +0 -6
- mindspore/ops/_grad/__init__.py +2 -1
- mindspore/ops/_grad/grad_array_ops.py +209 -152
- mindspore/ops/_grad/grad_base.py +55 -17
- mindspore/ops/_grad/grad_clip_ops.py +11 -3
- mindspore/ops/_grad/grad_comm_ops.py +58 -47
- mindspore/ops/_grad/grad_implementations.py +21 -61
- mindspore/ops/_grad/grad_inner_ops.py +48 -6
- mindspore/ops/_grad/grad_math_ops.py +306 -161
- mindspore/ops/_grad/grad_nn_ops.py +192 -181
- mindspore/ops/_grad/grad_other_ops.py +1 -1
- mindspore/ops/_grad/grad_quant_ops.py +5 -5
- mindspore/ops/_grad/grad_sequence_ops.py +296 -0
- mindspore/ops/_grad/grad_sparse.py +15 -9
- mindspore/ops/_grad_experimental/__init__.py +1 -0
- mindspore/ops/_grad_experimental/grad_array_ops.py +441 -55
- mindspore/ops/_grad_experimental/grad_image_ops.py +25 -7
- mindspore/ops/_grad_experimental/grad_inner_ops.py +3 -44
- mindspore/ops/_grad_experimental/grad_linalg_ops.py +16 -21
- mindspore/ops/_grad_experimental/grad_math_ops.py +979 -49
- mindspore/ops/_grad_experimental/grad_nn_ops.py +78 -8
- mindspore/ops/_grad_experimental/grad_scalar_ops.py +112 -0
- mindspore/ops/_grad_experimental/grad_sparse_ops.py +197 -13
- mindspore/ops/_op_impl/__init__.py +3 -3
- mindspore/ops/_op_impl/_custom_op/__init__.py +0 -1
- mindspore/ops/_op_impl/_custom_op/_basic.py +0 -1
- mindspore/ops/_op_impl/_custom_op/batch_matmul_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/batchnorm_fold.py +4 -2
- mindspore/ops/_op_impl/_custom_op/batchnorm_fold2.py +2 -2
- mindspore/ops/_op_impl/_custom_op/batchnorm_fold2_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/batchnorm_fold2_grad_reduce.py +5 -5
- mindspore/ops/_op_impl/_custom_op/batchnorm_fold_grad.py +3 -3
- mindspore/ops/_op_impl/_custom_op/cholesky_trsm_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/correction_mul.py +3 -3
- mindspore/ops/_op_impl/_custom_op/correction_mul_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/dsd_back_impl.py +4 -8
- mindspore/ops/_op_impl/_custom_op/dsd_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perchannel.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perchannel_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perchannel_grad_reduce.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perlayer.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perlayer_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perlayer_grad_reduce.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_quant_perchannel.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_quant_perchannel_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_quant_perlayer.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_quant_perlayer_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fused_abs_max1_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/img2col_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/matmul_cube_dense_left_impl.py +2 -2
- mindspore/ops/_op_impl/_custom_op/matmul_cube_dense_right_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/matmul_cube_fracz_left_cast_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/matmul_cube_fracz_right_mul_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/matmul_cube_impl.py +2 -2
- mindspore/ops/_op_impl/_custom_op/matmul_dds_grad_impl.py +0 -1
- mindspore/ops/_op_impl/_custom_op/matmul_dds_impl.py +0 -1
- mindspore/ops/_op_impl/_custom_op/matrix_combine_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/minmax_update_perchannel.py +2 -2
- mindspore/ops/_op_impl/_custom_op/minmax_update_perlayer.py +2 -2
- mindspore/ops/_op_impl/_custom_op/transpose02314_impl.py +1 -1
- mindspore/ops/_op_impl/aicpu/__init__.py +238 -3
- mindspore/ops/_op_impl/aicpu/abs.py +36 -0
- mindspore/ops/_op_impl/aicpu/adaptive_avg_pool_2d.py +34 -0
- mindspore/ops/_op_impl/aicpu/adaptive_avg_pool_2d_grad.py +34 -0
- mindspore/ops/_op_impl/aicpu/adaptive_avg_pool_3d.py +39 -0
- mindspore/ops/_op_impl/aicpu/adaptive_avg_pool_3d_grad.py +39 -0
- mindspore/ops/_op_impl/aicpu/adaptive_max_pool_2d_grad.py +37 -0
- mindspore/ops/_op_impl/aicpu/adaptive_max_pool_3d.py +42 -0
- mindspore/ops/_op_impl/aicpu/adaptive_max_pool_3d_grad.py +152 -0
- mindspore/ops/_op_impl/aicpu/add.py +43 -0
- mindspore/ops/_op_impl/aicpu/addcdiv.py +0 -32
- mindspore/ops/_op_impl/aicpu/addcmul.py +0 -84
- mindspore/ops/_op_impl/aicpu/affine_grid_grad.py +35 -0
- mindspore/ops/_op_impl/aicpu/arg_max.py +75 -0
- mindspore/ops/_op_impl/aicpu/arg_min.py +75 -0
- mindspore/ops/_op_impl/aicpu/argmin_with_value.py +43 -0
- mindspore/ops/_op_impl/aicpu/batch_matmul.py +43 -0
- mindspore/ops/_op_impl/aicpu/batch_norm_grad_grad.py +49 -0
- mindspore/ops/_op_impl/aicpu/bernoulli.py +48 -0
- mindspore/ops/_op_impl/aicpu/bessel_i0.py +31 -0
- mindspore/ops/_op_impl/aicpu/bias_add.py +44 -0
- mindspore/ops/_op_impl/aicpu/bias_add_grad.py +43 -0
- mindspore/ops/_op_impl/aicpu/bincount.py +33 -0
- mindspore/{nn/probability/infer/variational/__init__.py → ops/_op_impl/aicpu/cauchy.py} +17 -10
- mindspore/ops/_op_impl/aicpu/channel_shuffle.py +40 -0
- mindspore/ops/_op_impl/aicpu/cholesky.py +1 -1
- mindspore/ops/_op_impl/{cpu/bias_add.py → aicpu/choleskygrad.py} +9 -7
- mindspore/ops/_op_impl/aicpu/combined_non_max_suppression.py +42 -0
- mindspore/ops/_op_impl/aicpu/concat_offset.py +42 -0
- mindspore/ops/_op_impl/aicpu/concat_offset_v1.py +31 -0
- mindspore/ops/_op_impl/aicpu/conj.py +11 -0
- mindspore/ops/_op_impl/aicpu/crop_and_resize_grad_image.py +38 -0
- mindspore/ops/_op_impl/aicpu/cumulative_logsumexp.py +36 -0
- mindspore/ops/_op_impl/aicpu/deformable_offsets.py +38 -0
- mindspore/ops/_op_impl/aicpu/deformable_offsets_grad.py +2 -2
- mindspore/ops/_op_impl/aicpu/dense_to_sparse_set_operation.py +48 -0
- mindspore/ops/_op_impl/aicpu/diag.py +36 -0
- mindspore/ops/_op_impl/aicpu/diag_part.py +36 -0
- mindspore/ops/_op_impl/aicpu/diagonal.py +35 -0
- mindspore/ops/_op_impl/{cpu/bias_add_grad.py → aicpu/digamma.py} +9 -7
- mindspore/ops/_op_impl/aicpu/eig.py +35 -0
- mindspore/ops/_op_impl/aicpu/fft_with_size.py +41 -0
- mindspore/ops/_op_impl/aicpu/flatten.py +1 -0
- mindspore/ops/_op_impl/aicpu/fmax.py +36 -0
- mindspore/ops/_op_impl/aicpu/fmin.py +37 -0
- mindspore/ops/_op_impl/aicpu/fractional_max_pool3d_with_fixed_ksize.py +1 -1
- mindspore/ops/_op_impl/aicpu/fse_decode.py +43 -0
- mindspore/ops/_op_impl/aicpu/glu.py +33 -0
- mindspore/ops/_op_impl/aicpu/glu_grad.py +34 -0
- mindspore/ops/_op_impl/aicpu/greater.py +41 -0
- mindspore/ops/_op_impl/aicpu/greater_equal.py +41 -0
- mindspore/ops/_op_impl/aicpu/index_put.py +50 -0
- mindspore/ops/_op_impl/{tbe/scatter_add_ds.py → aicpu/inplace_index_add.py} +17 -21
- mindspore/ops/_op_impl/aicpu/instance_norm_v2.py +41 -0
- mindspore/ops/_op_impl/aicpu/instance_norm_v2_grad.py +44 -0
- mindspore/ops/_op_impl/aicpu/layer_norm_grad_grad.py +47 -0
- mindspore/ops/_op_impl/aicpu/less.py +41 -0
- mindspore/ops/_op_impl/aicpu/less_equal.py +41 -0
- mindspore/ops/_op_impl/aicpu/lgamma.py +32 -0
- mindspore/ops/_op_impl/aicpu/log_normal_reverse.py +33 -0
- mindspore/ops/_op_impl/aicpu/logit.py +33 -0
- mindspore/ops/_op_impl/aicpu/logit_grad.py +34 -0
- mindspore/ops/_op_impl/aicpu/masked_fill.py +42 -0
- mindspore/ops/_op_impl/aicpu/masked_scatter.py +39 -0
- mindspore/ops/_op_impl/aicpu/matmul.py +39 -0
- mindspore/ops/_op_impl/aicpu/matrix_logarithm.py +31 -0
- mindspore/ops/_op_impl/aicpu/matrix_power.py +32 -0
- mindspore/ops/_op_impl/aicpu/matrix_solve_ls.py +36 -0
- mindspore/ops/_op_impl/aicpu/matrix_triangular_solve.py +36 -0
- mindspore/ops/_op_impl/aicpu/mirror_pad.py +2 -0
- mindspore/ops/_op_impl/aicpu/mirror_pad_grad.py +0 -4
- mindspore/ops/_op_impl/aicpu/mul.py +3 -1
- mindspore/ops/_op_impl/aicpu/multinomial.py +14 -6
- mindspore/ops/_op_impl/aicpu/multinomial_with_replacement.py +35 -0
- mindspore/ops/_op_impl/aicpu/nan_to_num.py +34 -0
- mindspore/ops/_op_impl/aicpu/nllloss.py +38 -0
- mindspore/ops/_op_impl/aicpu/nllloss_grad.py +39 -0
- mindspore/ops/_op_impl/aicpu/ones_like.py +0 -2
- mindspore/ops/_op_impl/aicpu/polar.py +32 -0
- mindspore/ops/_op_impl/aicpu/polygamma.py +34 -0
- mindspore/ops/_op_impl/aicpu/qr.py +36 -0
- mindspore/ops/_op_impl/aicpu/quant_dtype_cast.py +40 -0
- mindspore/ops/_op_impl/aicpu/quantile.py +35 -0
- mindspore/ops/_op_impl/aicpu/ragged_tensor_to_sparse.py +73 -0
- mindspore/ops/_op_impl/aicpu/ragged_tensor_to_tensor.py +74 -0
- mindspore/ops/_op_impl/aicpu/random_shuffle.py +3 -0
- mindspore/ops/_op_impl/aicpu/randperm_v2.py +41 -0
- mindspore/ops/_op_impl/aicpu/range.py +36 -0
- mindspore/ops/_op_impl/aicpu/reciprocal.py +34 -0
- mindspore/ops/_op_impl/aicpu/reciprocal_grad.py +35 -0
- mindspore/ops/_op_impl/aicpu/reduce_sum.py +57 -0
- mindspore/ops/_op_impl/aicpu/resize_bicubic.py +2 -8
- mindspore/ops/_op_impl/aicpu/resize_bicubic_grad.py +1 -1
- mindspore/ops/_op_impl/aicpu/resize_v2.py +68 -0
- mindspore/ops/_op_impl/aicpu/resize_v2_grad.py +68 -0
- mindspore/ops/_op_impl/aicpu/scatter_elements.py +4 -0
- mindspore/ops/_op_impl/aicpu/scatter_nd_update.py +2 -0
- mindspore/ops/_op_impl/aicpu/search_sorted.py +12 -6
- mindspore/ops/_op_impl/aicpu/self_adjoint_eig.py +34 -0
- mindspore/ops/_op_impl/aicpu/sequence_add.py +34 -0
- mindspore/ops/_op_impl/aicpu/sequence_add_offset.py +34 -0
- mindspore/ops/_op_impl/aicpu/sequence_addn.py +38 -0
- mindspore/ops/_op_impl/aicpu/slice_grad.py +76 -0
- mindspore/ops/_op_impl/aicpu/smooth_l1_loss.py +35 -0
- mindspore/ops/_op_impl/aicpu/smooth_l1_loss_grad.py +37 -0
- mindspore/ops/_op_impl/aicpu/sort.py +39 -0
- mindspore/ops/_op_impl/aicpu/sparse_apply_adagrad_da.py +0 -24
- mindspore/ops/_op_impl/aicpu/sparse_cross.py +42 -0
- mindspore/ops/_op_impl/aicpu/sparse_fill_empty_rows.py +63 -0
- mindspore/ops/_op_impl/aicpu/sparse_fill_empty_rows_grad.py +45 -0
- mindspore/ops/_op_impl/aicpu/sparse_matrix_mat_mul.py +56 -0
- mindspore/ops/_op_impl/{tbe/slice_ds.py → aicpu/sparse_segment_sum.py} +16 -24
- mindspore/ops/_op_impl/aicpu/sparse_segment_sum_with_num_segments.py +68 -0
- mindspore/ops/_op_impl/aicpu/sparse_slice.py +63 -0
- mindspore/ops/_op_impl/aicpu/sparse_slice_grad.py +61 -0
- mindspore/ops/_op_impl/aicpu/squared_difference.py +2 -0
- mindspore/ops/_op_impl/aicpu/strided_slice_v2.py +93 -0
- mindspore/ops/_op_impl/aicpu/strided_slice_v2_grad.py +66 -0
- mindspore/ops/_op_impl/aicpu/tensor_scatter_update.py +59 -0
- mindspore/ops/_op_impl/{tbe/gather_v2.py → aicpu/tile.py} +24 -24
- mindspore/ops/_op_impl/aicpu/tridiagonal_solve.py +35 -0
- mindspore/ops/_op_impl/aicpu/tril_indices.py +34 -0
- mindspore/ops/_op_impl/aicpu/triu_indices.py +34 -0
- mindspore/ops/_op_impl/aicpu/uniform.py +34 -0
- mindspore/ops/_op_impl/aicpu/uniform_candidate_sampler.py +1 -0
- mindspore/ops/_op_impl/aicpu/unique_consecutive.py +10 -2
- mindspore/ops/_op_impl/cpu/__init__.py +1 -2
- mindspore/ops/_op_impl/cpu/dynamic_shape.py +5 -1
- mindspore/ops/_op_impl/cpu/maximum_grad.py +2 -0
- mindspore/{compression/common/__init__.py → ops/_op_impl/cpu/pyexecute.py} +13 -8
- mindspore/ops/_op_impl/cpu/reduce_sum.py +8 -0
- mindspore/ops/_op_impl/cpu/sparse_slice.py +62 -0
- mindspore/ops/_op_impl/cpu/sparse_slice_grad.py +60 -0
- mindspore/ops/_op_impl/cpu/tensor_shape.py +5 -1
- mindspore/ops/_op_impl/tbe/__init__.py +27 -608
- mindspore/ops/_op_impl/tbe/addcdiv_ds.py +42 -0
- mindspore/ops/_op_impl/tbe/addcmul_ds.py +44 -0
- mindspore/ops/_op_impl/tbe/assign_add_ds.py +1 -0
- mindspore/ops/_op_impl/tbe/atomic_addr_clean.py +1 -1
- mindspore/ops/_op_impl/tbe/avg_pool_3d_grad.py +1 -1
- mindspore/ops/_op_impl/tbe/basic_lstm_cell_c_state_grad_v2.py +0 -1
- mindspore/ops/_op_impl/tbe/batch_to_space.py +1 -1
- mindspore/ops/_op_impl/tbe/batch_to_space_nd.py +1 -1
- mindspore/ops/_op_impl/tbe/batch_to_space_nd_v2.py +41 -0
- mindspore/ops/_op_impl/tbe/bce_with_logits_loss.py +1 -0
- mindspore/ops/_op_impl/tbe/bias_add_grad.py +2 -0
- mindspore/ops/_op_impl/tbe/bn_infer_grad.py +4 -2
- mindspore/ops/_op_impl/tbe/bn_infer_grad_ds.py +40 -0
- mindspore/ops/_op_impl/tbe/bn_training_update.py +0 -1
- mindspore/ops/_op_impl/tbe/bn_training_update_ds.py +0 -1
- mindspore/ops/_op_impl/tbe/broadcast_to_ds.py +6 -4
- mindspore/ops/_op_impl/tbe/cast.py +0 -2
- mindspore/ops/_op_impl/tbe/cast_ds.py +3 -3
- mindspore/ops/_op_impl/tbe/ctc_loss_v2.py +0 -2
- mindspore/ops/_op_impl/tbe/ctc_loss_v2_grad.py +0 -2
- mindspore/ops/_op_impl/tbe/data_format_dim_map_ds.py +1 -0
- mindspore/ops/_op_impl/tbe/deformable_offsets.py +1 -0
- mindspore/ops/_op_impl/tbe/depthwise_conv2d.py +1 -1
- mindspore/ops/_op_impl/tbe/dynamic_atomic_addr_clean.py +1 -1
- mindspore/ops/_op_impl/tbe/gather_nd.py +1 -0
- mindspore/ops/_op_impl/tbe/greater.py +2 -0
- mindspore/ops/_op_impl/tbe/{index_add.py → inplace_index_add.py} +3 -6
- mindspore/ops/_op_impl/tbe/layer_norm_beta_gamma_backprop_v2.py +0 -1
- mindspore/ops/_op_impl/tbe/npu_clear_float_status_v2.py +35 -0
- mindspore/ops/_op_impl/tbe/npu_get_float_status_v2.py +35 -0
- mindspore/ops/_op_impl/tbe/one_hot_ds.py +0 -6
- mindspore/ops/_op_impl/tbe/{greater_ds.py → reduce_all_ds.py} +13 -16
- mindspore/ops/_op_impl/tbe/reduce_any_ds.py +39 -0
- mindspore/ops/_op_impl/tbe/roi_align_ds.py +44 -0
- mindspore/ops/_op_impl/tbe/roi_align_grad_ds.py +44 -0
- mindspore/ops/_op_impl/tbe/scatter_add.py +2 -0
- mindspore/ops/_op_impl/tbe/scatter_nd_add.py +2 -2
- mindspore/ops/_op_impl/tbe/slice.py +26 -15
- mindspore/ops/_op_impl/tbe/space_to_batch.py +1 -1
- mindspore/ops/_op_impl/tbe/space_to_batch_nd.py +1 -1
- mindspore/ops/_op_impl/tbe/strided_slice_grad_d.py +1 -0
- mindspore/ops/_op_impl/tbe/trans_data_ds.py +15 -5
- mindspore/ops/_op_impl/tbe/unsorted_segment_sum.py +1 -1
- mindspore/ops/_op_impl/tbe/unsorted_segment_sum_ds.py +2 -0
- mindspore/ops/_primitive_cache.py +3 -2
- mindspore/ops/_register_for_op.py +11 -0
- mindspore/ops/_utils/__init__.py +1 -1
- mindspore/ops/_utils/utils.py +20 -41
- mindspore/ops/_vmap/__init__.py +2 -2
- mindspore/ops/_vmap/vmap_array_ops.py +170 -78
- mindspore/ops/_vmap/vmap_base.py +24 -10
- mindspore/ops/_vmap/vmap_convolution_ops.py +7 -10
- mindspore/ops/_vmap/vmap_grad_math_ops.py +4 -4
- mindspore/ops/_vmap/vmap_grad_nn_ops.py +41 -9
- mindspore/ops/_vmap/vmap_image_ops.py +52 -0
- mindspore/ops/_vmap/vmap_math_ops.py +77 -6
- mindspore/ops/_vmap/vmap_nn_ops.py +78 -29
- mindspore/ops/_vmap/vmap_other_ops.py +3 -1
- mindspore/ops/_vmap/vmap_random_ops.py +55 -3
- mindspore/ops/_vmap/vmap_sparse_ops.py +1 -0
- mindspore/ops/bprop_mindir/AdaptiveAvgPool2D_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/AdaptiveMaxPool2D_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/ApproximateEqual_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/Argmax_bprop.mindir +13 -12
- mindspore/ops/bprop_mindir/Argmin_bprop.mindir +14 -13
- mindspore/ops/bprop_mindir/AssignSub_bprop.mindir +17 -18
- mindspore/ops/bprop_mindir/Assign_bprop.mindir +16 -16
- mindspore/ops/bprop_mindir/AvgPool3D_bprop.mindir +150 -0
- mindspore/ops/bprop_mindir/AvgPool_bprop.mindir +66 -0
- mindspore/ops/bprop_mindir/BCEWithLogitsLoss_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/BNTrainingReduce_bprop.mindir +13 -12
- mindspore/ops/bprop_mindir/BatchNormGrad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/BatchToSpaceND_bprop.mindir +28 -0
- mindspore/ops/bprop_mindir/BiasAddGrad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/BinaryCrossEntropy_bprop.mindir +33 -0
- mindspore/ops/bprop_mindir/BroadcastTo_bprop.mindir +306 -0
- mindspore/ops/bprop_mindir/Broadcast_bprop.mindir +12 -8
- mindspore/ops/bprop_mindir/CTCLoss_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Concat_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Conv2DBackpropFilter_bprop.mindir +240 -0
- mindspore/ops/bprop_mindir/Conv2DBackpropInput_bprop.mindir +247 -0
- mindspore/ops/bprop_mindir/Conv2DTranspose_bprop.mindir +247 -0
- mindspore/ops/bprop_mindir/Conv3DTranspose_bprop.mindir +315 -0
- mindspore/ops/bprop_mindir/Conv3D_bprop.mindir +278 -0
- mindspore/ops/bprop_mindir/DType_bprop.mindir +12 -12
- mindspore/ops/bprop_mindir/DeformableOffsets_bprop.mindir +58 -0
- mindspore/ops/bprop_mindir/Depend_bprop.mindir +12 -13
- mindspore/ops/bprop_mindir/DepthToSpace_bprop.mindir +23 -0
- mindspore/ops/bprop_mindir/DepthwiseConv2dNative_bprop.mindir +138 -0
- mindspore/ops/bprop_mindir/DiagPart_bprop.mindir +15 -0
- mindspore/ops/bprop_mindir/Dropout2D_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Dropout3D_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/DropoutDoMask_bprop.mindir +22 -24
- mindspore/ops/bprop_mindir/DropoutGenMask_bprop.mindir +16 -14
- mindspore/ops/bprop_mindir/DropoutGrad_bprop.mindir +27 -0
- mindspore/ops/bprop_mindir/Dropout_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/DynamicGRUV2_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/DynamicRNN_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/DynamicShape_bprop.mindir +12 -12
- mindspore/ops/bprop_mindir/Elu_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/EmbeddingLookup_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Equal_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/ExpandDims_bprop.mindir +58 -0
- mindspore/ops/bprop_mindir/FastGeLU_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/Flatten_bprop.mindir +54 -0
- mindspore/ops/bprop_mindir/FloorDiv_bprop.mindir +18 -15
- mindspore/ops/bprop_mindir/GatherD_bprop.mindir +26 -0
- mindspore/ops/bprop_mindir/GatherNd_bprop.mindir +57 -0
- mindspore/ops/bprop_mindir/Gather_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/GreaterEqual_bprop.mindir +17 -18
- mindspore/ops/bprop_mindir/Greater_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/HSigmoid_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/HSwish_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/IOU_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/InstanceNorm_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/IsFinite_bprop.mindir +13 -12
- mindspore/ops/bprop_mindir/IsInf_bprop.mindir +13 -10
- mindspore/ops/bprop_mindir/IsNan_bprop.mindir +14 -11
- mindspore/ops/bprop_mindir/KLDivLoss_bprop.mindir +126 -0
- mindspore/ops/bprop_mindir/L2Loss_bprop.mindir +15 -0
- mindspore/ops/bprop_mindir/L2Normalize_bprop.mindir +30 -0
- mindspore/ops/bprop_mindir/LRN_bprop.mindir +43 -0
- mindspore/ops/bprop_mindir/LayerNormGrad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/LessEqual_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/Less_bprop.mindir +17 -18
- mindspore/ops/bprop_mindir/LinSpace_bprop.mindir +22 -19
- mindspore/ops/bprop_mindir/Load_bprop.mindir +12 -13
- mindspore/ops/bprop_mindir/LogSoftmax_bprop.mindir +23 -0
- mindspore/ops/bprop_mindir/LogicalAnd_bprop.mindir +17 -18
- mindspore/ops/bprop_mindir/LogicalNot_bprop.mindir +14 -13
- mindspore/ops/bprop_mindir/MaskedSelect_bprop.mindir +21 -0
- mindspore/ops/bprop_mindir/MaxPool3DGradGrad_bprop.mindir +74 -0
- mindspore/ops/bprop_mindir/MaxPool3DGrad_bprop.mindir +74 -0
- mindspore/ops/bprop_mindir/MaxPool3D_bprop.mindir +75 -0
- mindspore/ops/bprop_mindir/MaxPoolGradGrad_bprop.mindir +65 -0
- mindspore/ops/bprop_mindir/MaxPoolWithArgmax_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Maximum_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Minimum_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/MirrorPad_bprop.mindir +27 -0
- mindspore/ops/bprop_mindir/Mish_bprop.mindir +35 -0
- mindspore/ops/bprop_mindir/MulNoNan_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/NLLLoss_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/NonZero_bprop.mindir +14 -0
- mindspore/ops/bprop_mindir/NotEqual_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/OneHot_bprop.mindir +25 -23
- mindspore/ops/bprop_mindir/OnesLike_bprop.mindir +13 -13
- mindspore/ops/bprop_mindir/PReLU_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Pad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Padding_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/RNNTLoss_bprop.mindir +29 -0
- mindspore/ops/bprop_mindir/ROIAlign_bprop.mindir +82 -0
- mindspore/ops/bprop_mindir/Range_bprop.mindir +21 -19
- mindspore/ops/bprop_mindir/Rank_bprop.mindir +11 -11
- mindspore/ops/bprop_mindir/ReLU6_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/ReLUV2_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/ReduceAll_bprop.mindir +18 -17
- mindspore/ops/bprop_mindir/ReduceAny_bprop.mindir +18 -17
- mindspore/ops/bprop_mindir/ReluGrad_bprop.mindir +19 -23
- mindspore/ops/bprop_mindir/Reshape_bprop.mindir +60 -0
- mindspore/ops/bprop_mindir/ResizeBilinear_bprop.mindir +29 -0
- mindspore/ops/bprop_mindir/ResizeNearestNeighbor_bprop.mindir +89 -0
- mindspore/ops/bprop_mindir/ReverseSequence_bprop.mindir +52 -0
- mindspore/ops/bprop_mindir/ReverseV2_bprop.mindir +22 -0
- mindspore/ops/bprop_mindir/Round_bprop.mindir +14 -13
- mindspore/ops/bprop_mindir/ScatterMax_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/ScatterMin_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/ScatterNdUpdate_bprop.mindir +22 -0
- mindspore/ops/bprop_mindir/ScatterNd_bprop.mindir +24 -0
- mindspore/ops/bprop_mindir/ScatterNonAliasingAdd_bprop.mindir +22 -0
- mindspore/ops/bprop_mindir/ScatterUpdate_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/SeLU_bprop.mindir +21 -0
- mindspore/ops/bprop_mindir/Select_bprop.mindir +30 -34
- mindspore/ops/bprop_mindir/Shape_bprop.mindir +12 -12
- mindspore/ops/bprop_mindir/SigmoidCrossEntropyWithLogits_bprop.mindir +21 -0
- mindspore/ops/bprop_mindir/SigmoidGrad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Sigmoid_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/Sign_bprop.mindir +13 -12
- mindspore/ops/bprop_mindir/Slice_bprop.mindir +26 -0
- mindspore/ops/bprop_mindir/SmoothL1Loss_bprop.mindir +36 -0
- mindspore/ops/bprop_mindir/SoftmaxCrossEntropyWithLogits_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Softplus_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/Softsign_bprop.mindir +33 -0
- mindspore/ops/bprop_mindir/Sort_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/SpaceToBatchND_bprop.mindir +28 -0
- mindspore/ops/bprop_mindir/SpaceToDepth_bprop.mindir +23 -0
- mindspore/ops/bprop_mindir/SparseGatherV2_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/SparseSoftmaxCrossEntropyWithLogits_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Split_bprop.mindir +22 -0
- mindspore/ops/bprop_mindir/Squeeze_bprop.mindir +54 -0
- mindspore/ops/bprop_mindir/StridedSliceGrad_bprop.mindir +95 -0
- mindspore/ops/bprop_mindir/StridedSlice_bprop.mindir +98 -0
- mindspore/ops/bprop_mindir/Switch_bprop.mindir +28 -32
- mindspore/ops/bprop_mindir/TanhGrad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Tanh_bprop.mindir +66 -0
- mindspore/ops/bprop_mindir/TensorScatterAdd_bprop.mindir +22 -0
- mindspore/ops/bprop_mindir/TensorScatterUpdate_bprop.mindir +29 -0
- mindspore/ops/bprop_mindir/TensorShape_bprop.mindir +14 -0
- mindspore/ops/bprop_mindir/Tile_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/TopK_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/TransShape_bprop.mindir +23 -0
- mindspore/ops/bprop_mindir/TruncateDiv_bprop.mindir +18 -15
- mindspore/ops/bprop_mindir/TupleGetItem_bprop.mindir +11 -13
- mindspore/ops/bprop_mindir/Unique_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/Unstack_bprop.mindir +22 -0
- mindspore/ops/bprop_mindir/UpsampleNearest3D_bprop.mindir +32 -0
- mindspore/ops/bprop_mindir/UpsampleTrilinear3D_bprop.mindir +38 -0
- mindspore/ops/bprop_mindir/ZerosLike_bprop.mindir +13 -12
- mindspore/ops/bprop_mindir/__init__.py +1 -4
- mindspore/ops/bprop_mindir/generate_mindir.py +32 -20
- mindspore/ops/composite/__init__.py +12 -13
- mindspore/ops/composite/base.py +261 -254
- mindspore/ops/composite/env_ops.py +41 -0
- mindspore/ops/composite/math_ops.py +197 -156
- mindspore/ops/composite/multitype_ops/_compile_utils.py +428 -176
- mindspore/ops/composite/multitype_ops/_constexpr_utils.py +188 -87
- mindspore/ops/composite/multitype_ops/add_impl.py +23 -1
- mindspore/ops/composite/multitype_ops/div_impl.py +3 -3
- mindspore/ops/composite/multitype_ops/equal_impl.py +1 -0
- mindspore/ops/composite/multitype_ops/floordiv_impl.py +1 -1
- mindspore/ops/composite/multitype_ops/getitem_impl.py +52 -5
- mindspore/ops/composite/multitype_ops/greater_equal_impl.py +31 -0
- mindspore/ops/composite/multitype_ops/greater_impl.py +31 -0
- mindspore/ops/composite/multitype_ops/in_impl.py +15 -3
- mindspore/ops/composite/multitype_ops/less_equal_impl.py +33 -2
- mindspore/ops/composite/multitype_ops/less_impl.py +33 -0
- mindspore/ops/composite/multitype_ops/logical_and_impl.py +2 -2
- mindspore/ops/composite/multitype_ops/logical_or_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/mod_impl.py +1 -1
- mindspore/ops/composite/multitype_ops/mul_impl.py +21 -7
- mindspore/ops/composite/multitype_ops/not_in_impl.py +15 -3
- mindspore/ops/composite/multitype_ops/ones_like_impl.py +2 -4
- mindspore/ops/composite/multitype_ops/pow_impl.py +1 -0
- mindspore/ops/composite/multitype_ops/setitem_impl.py +62 -70
- mindspore/ops/composite/multitype_ops/sub_impl.py +3 -3
- mindspore/ops/composite/multitype_ops/zeros_like_impl.py +41 -4
- mindspore/ops/function/__init__.py +323 -8
- mindspore/ops/function/array_func.py +3511 -780
- mindspore/ops/function/clip_func.py +329 -0
- mindspore/ops/function/debug_func.py +6 -6
- mindspore/ops/function/grad/__init__.py +5 -1
- mindspore/ops/function/grad/grad_func.py +736 -65
- mindspore/ops/function/image_func.py +270 -0
- mindspore/ops/function/linalg_func.py +268 -8
- mindspore/ops/function/math_func.py +8032 -3164
- mindspore/ops/function/nn_func.py +5619 -1855
- mindspore/ops/function/other_func.py +115 -0
- mindspore/ops/function/parameter_func.py +11 -10
- mindspore/ops/function/random_func.py +939 -77
- mindspore/ops/function/sparse_func.py +249 -84
- mindspore/ops/function/sparse_unary_func.py +2303 -0
- mindspore/ops/function/spectral_func.py +146 -0
- mindspore/ops/function/vmap_func.py +114 -0
- mindspore/ops/functional.py +182 -254
- mindspore/ops/op_info_register.py +79 -34
- mindspore/ops/operations/__init__.py +210 -118
- mindspore/ops/operations/_csr_ops.py +7 -7
- mindspore/ops/operations/_embedding_cache_ops.py +25 -15
- mindspore/ops/operations/_grad_ops.py +447 -322
- mindspore/ops/operations/_inner_ops.py +547 -176
- mindspore/ops/operations/_map_tensor_ops.py +112 -0
- mindspore/ops/operations/_ms_kernel.py +29 -27
- mindspore/ops/operations/_ocr_ops.py +11 -11
- mindspore/ops/operations/_opaque_predicate_registry.py +41 -0
- mindspore/ops/operations/_quant_ops.py +186 -101
- mindspore/ops/operations/_rl_inner_ops.py +122 -61
- mindspore/ops/operations/_scalar_ops.py +466 -0
- mindspore/ops/operations/_sequence_ops.py +1047 -0
- mindspore/ops/operations/_tensor_array.py +10 -11
- mindspore/ops/operations/_thor_ops.py +4 -4
- mindspore/ops/operations/array_ops.py +1428 -1226
- mindspore/ops/operations/comm_ops.py +180 -117
- mindspore/ops/operations/control_ops.py +4 -2
- mindspore/ops/operations/custom_ops.py +185 -98
- mindspore/ops/operations/debug_ops.py +92 -54
- mindspore/ops/operations/image_ops.py +406 -211
- mindspore/ops/operations/inner_ops.py +42 -53
- mindspore/ops/operations/linalg_ops.py +32 -29
- mindspore/ops/operations/math_ops.py +2076 -897
- mindspore/ops/operations/nn_ops.py +1282 -1252
- mindspore/ops/operations/other_ops.py +124 -278
- mindspore/ops/operations/random_ops.py +345 -178
- mindspore/ops/operations/rl_ops.py +8 -9
- mindspore/ops/operations/sparse_ops.py +502 -157
- mindspore/ops/operations/spectral_ops.py +107 -0
- mindspore/ops/primitive.py +192 -15
- mindspore/ops/vm_impl_registry.py +23 -2
- mindspore/parallel/__init__.py +6 -1
- mindspore/parallel/_auto_parallel_context.py +199 -92
- mindspore/parallel/_cell_wrapper.py +4 -2
- mindspore/parallel/_cost_model_context.py +3 -0
- mindspore/parallel/_dp_allreduce_fusion.py +2 -1
- mindspore/parallel/_offload_context.py +185 -0
- mindspore/parallel/_parallel_serialization.py +167 -28
- mindspore/parallel/_ps_context.py +9 -5
- mindspore/parallel/_recovery_context.py +1 -1
- mindspore/parallel/_tensor.py +9 -1
- mindspore/{nn/transformer → parallel/_transformer}/__init__.py +6 -6
- mindspore/{nn/transformer → parallel/_transformer}/layers.py +59 -37
- mindspore/{nn/transformer → parallel/_transformer}/loss.py +4 -7
- mindspore/{nn/transformer → parallel/_transformer}/moe.py +160 -35
- mindspore/{nn/transformer → parallel/_transformer}/op_parallel_config.py +3 -3
- mindspore/{nn/transformer → parallel/_transformer}/transformer.py +235 -196
- mindspore/parallel/_utils.py +47 -7
- mindspore/parallel/algo_parameter_config.py +5 -1
- mindspore/parallel/checkpoint_transform.py +329 -0
- mindspore/parallel/shard.py +229 -0
- mindspore/perf_msvcbuildinsights.dll +0 -0
- mindspore/pgodb140.dll +0 -0
- mindspore/pgort140.dll +0 -0
- mindspore/profiler/__init__.py +2 -1
- mindspore/profiler/common/util.py +4 -3
- mindspore/profiler/common/validator/validate_path.py +2 -2
- mindspore/profiler/envprofiling.py +249 -0
- mindspore/profiler/parser/aicpu_data_parser.py +38 -39
- mindspore/profiler/parser/ascend_timeline_generator.py +497 -0
- mindspore/profiler/parser/base_timeline_generator.py +471 -0
- mindspore/profiler/parser/cpu_gpu_timeline_generator.py +684 -0
- mindspore/profiler/parser/framework_parser.py +42 -16
- mindspore/profiler/parser/hccl_parser.py +158 -158
- mindspore/profiler/parser/hwts_log_parser.py +7 -6
- mindspore/profiler/parser/integrator.py +18 -1579
- mindspore/profiler/parser/minddata_analyzer.py +8 -8
- mindspore/profiler/parser/msadvisor_analyzer.py +14 -27
- mindspore/profiler/parser/msadvisor_parser.py +2 -4
- mindspore/profiler/parser/optime_parser.py +17 -18
- mindspore/profiler/parser/profiler_info.py +108 -0
- mindspore/profiler/parser/step_trace_parser.py +1 -1
- mindspore/profiler/profiling.py +396 -194
- mindspore/rewrite/__init__.py +6 -2
- mindspore/rewrite/api/node.py +51 -110
- mindspore/rewrite/api/node_type.py +10 -6
- mindspore/rewrite/api/pattern_engine.py +51 -7
- mindspore/rewrite/api/scoped_value.py +64 -53
- mindspore/rewrite/api/symbol_tree.py +108 -61
- mindspore/rewrite/api/tree_node_helper.py +2 -3
- mindspore/{compression/quant/__init__.py → rewrite/ast_creator_register.py} +20 -11
- mindspore/rewrite/ast_helpers/__init__.py +6 -3
- mindspore/rewrite/ast_helpers/ast_creator.py +115 -0
- mindspore/rewrite/ast_helpers/ast_finder.py +99 -1
- mindspore/rewrite/ast_helpers/ast_modifier.py +17 -4
- mindspore/rewrite/ast_helpers/ast_replacer.py +1 -1
- mindspore/rewrite/ast_transformers/__init__.py +0 -1
- mindspore/rewrite/ast_transformers/flatten_recursive_stmt.py +46 -5
- mindspore/rewrite/ast_transformers/remove_return_out_of_if.py +6 -3
- mindspore/rewrite/common/__init__.py +2 -0
- mindspore/rewrite/common/event.py +1 -1
- mindspore/rewrite/common/observable.py +1 -1
- mindspore/rewrite/common/observer.py +1 -1
- mindspore/rewrite/common/rewrite_elog.py +35 -0
- mindspore/rewrite/namer.py +2 -2
- mindspore/rewrite/namespace.py +14 -4
- mindspore/rewrite/node.py +161 -13
- mindspore/rewrite/parser.py +0 -1
- mindspore/rewrite/parser_register.py +0 -1
- mindspore/rewrite/parsers/arguments_parser.py +3 -2
- mindspore/rewrite/parsers/assign_parser.py +267 -67
- mindspore/rewrite/parsers/attribute_parser.py +56 -0
- mindspore/rewrite/parsers/class_def_parser.py +191 -108
- mindspore/rewrite/parsers/constant_parser.py +101 -0
- mindspore/rewrite/parsers/container_parser.py +88 -0
- mindspore/rewrite/parsers/for_parser.py +28 -15
- mindspore/rewrite/parsers/function_def_parser.py +21 -5
- mindspore/rewrite/parsers/if_parser.py +11 -28
- mindspore/rewrite/parsers/module_parser.py +9 -6
- mindspore/rewrite/parsers/return_parser.py +3 -2
- mindspore/rewrite/sparsify/__init__.py +0 -0
- mindspore/rewrite/sparsify/sparse_transformer.py +448 -0
- mindspore/rewrite/sparsify/sparsify.py +109 -0
- mindspore/rewrite/sparsify/utils.py +173 -0
- mindspore/rewrite/symbol_tree.py +322 -109
- mindspore/rewrite/symbol_tree_builder.py +45 -8
- mindspore/rewrite/symbol_tree_dumper.py +0 -1
- mindspore/rewrite/topological_manager.py +1 -2
- mindspore/run_check/_check_version.py +209 -112
- mindspore/run_check/run_check.py +2 -1
- mindspore/tbbmalloc.dll +0 -0
- mindspore/tinyxml2.dll +0 -0
- mindspore/train/__init__.py +6 -4
- mindspore/train/_utils.py +28 -5
- mindspore/train/amp.py +321 -50
- mindspore/train/callback/__init__.py +3 -1
- mindspore/train/callback/_backup_and_restore.py +120 -0
- mindspore/train/callback/_callback.py +8 -8
- mindspore/train/callback/_checkpoint.py +12 -9
- mindspore/train/callback/_early_stop.py +13 -7
- mindspore/train/callback/_history.py +8 -8
- mindspore/train/callback/_lambda_callback.py +6 -6
- mindspore/train/callback/_landscape.py +36 -38
- mindspore/train/callback/_loss_monitor.py +12 -6
- mindspore/train/callback/_lr_scheduler_callback.py +2 -4
- mindspore/train/callback/_on_request_exit.py +212 -0
- mindspore/train/callback/_reduce_lr_on_plateau.py +13 -7
- mindspore/train/callback/_summary_collector.py +27 -19
- mindspore/train/callback/_time_monitor.py +13 -7
- mindspore/train/checkpoint_pb2.py +68 -8
- mindspore/train/data_sink.py +122 -33
- mindspore/train/dataset_helper.py +28 -87
- mindspore/train/loss_scale_manager.py +4 -7
- mindspore/{nn → train}/metrics/__init__.py +20 -20
- mindspore/{nn → train}/metrics/accuracy.py +12 -10
- mindspore/{nn → train}/metrics/auc.py +4 -4
- mindspore/{nn → train}/metrics/bleu_score.py +4 -4
- mindspore/{nn → train}/metrics/confusion_matrix.py +10 -8
- mindspore/{nn → train}/metrics/cosine_similarity.py +4 -4
- mindspore/{nn → train}/metrics/dice.py +6 -5
- mindspore/{nn → train}/metrics/error.py +7 -5
- mindspore/{nn → train}/metrics/fbeta.py +9 -7
- mindspore/{nn → train}/metrics/hausdorff_distance.py +8 -6
- mindspore/{nn → train}/metrics/loss.py +4 -3
- mindspore/{nn → train}/metrics/mean_surface_distance.py +6 -5
- mindspore/{nn → train}/metrics/metric.py +6 -5
- mindspore/{nn → train}/metrics/occlusion_sensitivity.py +4 -3
- mindspore/{nn → train}/metrics/perplexity.py +5 -4
- mindspore/{nn → train}/metrics/precision.py +5 -4
- mindspore/{nn → train}/metrics/recall.py +5 -4
- mindspore/{nn → train}/metrics/roc.py +7 -6
- mindspore/{nn → train}/metrics/root_mean_square_surface_distance.py +6 -5
- mindspore/{nn → train}/metrics/topk.py +7 -5
- mindspore/train/mind_ir_pb2.py +339 -32
- mindspore/train/model.py +113 -84
- mindspore/train/serialization.py +547 -167
- mindspore/train/summary/_summary_adapter.py +1 -1
- mindspore/train/summary/summary_record.py +43 -12
- mindspore/train/train_thor/convert_utils.py +7 -1
- mindspore/train/train_thor/dataset_helper.py +3 -3
- mindspore/train/train_thor/model_thor.py +0 -4
- mindspore/turbojpeg.dll +0 -0
- mindspore/vcmeta.dll +0 -0
- mindspore/vcruntime140.dll +0 -0
- mindspore/vcruntime140_1.dll +0 -0
- mindspore/version.py +1 -1
- {mindspore-1.10.0.dist-info → mindspore-2.0.0rc1.dist-info}/METADATA +4 -3
- {mindspore-1.10.0.dist-info → mindspore-2.0.0rc1.dist-info}/RECORD +901 -660
- mindspore/compression/common/constant.py +0 -124
- mindspore/compression/export/__init__.py +0 -19
- mindspore/compression/export/quant_export.py +0 -514
- mindspore/compression/quant/qat.py +0 -636
- mindspore/compression/quant/quant_utils.py +0 -462
- mindspore/compression/quant/quantizer.py +0 -68
- mindspore/libatomic-1.dll +0 -0
- mindspore/libgcc_s_seh-1.dll +0 -0
- mindspore/libgfortran-4.dll +0 -0
- mindspore/libgomp-1.dll +0 -0
- mindspore/libjpeg-62.dll +0 -0
- mindspore/libmindspore.dll +0 -0
- mindspore/libmindspore_common.dll +0 -0
- mindspore/libmindspore_core.dll +0 -0
- mindspore/libmindspore_glog.dll +0 -0
- mindspore/libnnacl.dll +0 -0
- mindspore/libopencv_core452.dll +0 -0
- mindspore/libopencv_imgcodecs452.dll +0 -0
- mindspore/libopencv_imgproc452.dll +0 -0
- mindspore/libquadmath-0.dll +0 -0
- mindspore/libsqlite3.dll +0 -0
- mindspore/libssp-0.dll +0 -0
- mindspore/libstdc++-6.dll +0 -0
- mindspore/libtinyxml2.dll +0 -0
- mindspore/libturbojpeg.dll +0 -0
- mindspore/libwinpthread-1.dll +0 -0
- mindspore/nn/layer/quant.py +0 -1868
- mindspore/nn/layer/rnn_utils.py +0 -90
- mindspore/nn/probability/dpn/__init__.py +0 -22
- mindspore/nn/probability/dpn/vae/__init__.py +0 -25
- mindspore/nn/probability/dpn/vae/cvae.py +0 -138
- mindspore/nn/probability/dpn/vae/vae.py +0 -122
- mindspore/nn/probability/infer/__init__.py +0 -22
- mindspore/nn/probability/infer/variational/elbo.py +0 -70
- mindspore/nn/probability/infer/variational/svi.py +0 -84
- mindspore/nn/probability/toolbox/__init__.py +0 -22
- mindspore/nn/probability/toolbox/anomaly_detection.py +0 -99
- mindspore/nn/probability/toolbox/uncertainty_evaluation.py +0 -363
- mindspore/nn/probability/transforms/__init__.py +0 -22
- mindspore/nn/probability/transforms/transform_bnn.py +0 -262
- mindspore/nn/probability/zhusuan/__init__.py +0 -18
- mindspore/nn/probability/zhusuan/framework/__init__.py +0 -18
- mindspore/nn/probability/zhusuan/framework/bn.py +0 -95
- mindspore/nn/probability/zhusuan/variational/__init__.py +0 -18
- mindspore/nn/probability/zhusuan/variational/elbo.py +0 -46
- mindspore/ops/_op_impl/tbe/bias_add_grad_ds.py +0 -52
- mindspore/ops/_op_impl/tbe/scatter_nd_add_ds.py +0 -43
- mindspore/ops/bprop_mindir/AssignAdd_bprop.mindir +0 -20
- mindspore/ops/bprop_mindir/Identity_bprop.mindir +0 -9
- mindspore/ops/bprop_mindir/LogicalOr_bprop.mindir +0 -20
- mindspore/ops/bprop_mindir/ReLU_bprop.mindir +0 -16
- mindspore/ops/bprop_mindir/UpdateState_bprop.mindir +0 -17
- mindspore/ops/bprop_mindir/stop_gradient_bprop.mindir +0 -12
- mindspore/ops/composite/array_ops.py +0 -210
- mindspore/ops/composite/clip_ops.py +0 -238
- mindspore/ops/composite/random_ops.py +0 -426
- mindspore/ops/composite/vmap_ops.py +0 -38
- mindspore/ops/operations/sponge_ops.py +0 -3531
- mindspore/ops/operations/sponge_update_ops.py +0 -2546
- mindspore/parallel/nn/__init__.py +0 -42
- mindspore/parallel/nn/loss.py +0 -22
- mindspore/parallel/nn/moe.py +0 -21
- mindspore/parallel/nn/op_parallel_config.py +0 -22
- mindspore/parallel/nn/transformer.py +0 -31
- mindspore/run_check/_check_deps_version.py +0 -84
- {mindspore-1.10.0.dist-info → mindspore-2.0.0rc1.dist-info}/WHEEL +0 -0
- {mindspore-1.10.0.dist-info → mindspore-2.0.0rc1.dist-info}/entry_points.txt +0 -0
- {mindspore-1.10.0.dist-info → mindspore-2.0.0rc1.dist-info}/top_level.txt +0 -0
mindspore/nn/optim/optimizer.py
CHANGED
|
@@ -26,12 +26,13 @@ from mindspore.nn.cell import Cell
|
|
|
26
26
|
from mindspore.nn.layer.container import CellList
|
|
27
27
|
from mindspore.common.parameter import Parameter, ParameterTuple
|
|
28
28
|
from mindspore.common.initializer import initializer
|
|
29
|
-
from mindspore.common
|
|
29
|
+
from mindspore.common import Tensor
|
|
30
|
+
from mindspore.common.sparse_tensor import RowTensorInner
|
|
30
31
|
import mindspore.common.dtype as mstype
|
|
31
|
-
from mindspore
|
|
32
|
+
from mindspore import _checkparam as validator
|
|
32
33
|
from mindspore import log as logger
|
|
33
34
|
from mindspore.parallel._utils import _get_global_rank, _get_device_num, _get_parallel_mode
|
|
34
|
-
from mindspore.parallel._ps_context import _is_ps_mode
|
|
35
|
+
from mindspore.parallel._ps_context import _is_ps_mode
|
|
35
36
|
from mindspore.context import ParallelMode
|
|
36
37
|
from mindspore import context
|
|
37
38
|
from mindspore.nn.learning_rate_schedule import LearningRateSchedule
|
|
@@ -180,6 +181,7 @@ class Optimizer(Cell):
|
|
|
180
181
|
self._init_group_params(parameters, learning_rate, weight_decay, self.grad_centralization)
|
|
181
182
|
|
|
182
183
|
self._init_opt_attrs(learning_rate, parameters, weight_decay)
|
|
184
|
+
self.add_flags(skip_auto_parallel_compile=True)
|
|
183
185
|
|
|
184
186
|
def _init_opt_attrs(self, learning_rate, parameters, weight_decay):
|
|
185
187
|
"""initialize optimizer attributions"""
|
|
@@ -252,7 +254,7 @@ class Optimizer(Cell):
|
|
|
252
254
|
chunk_tensors = Tensor._get_flattened_tensors(parameters) # pylint: disable=W0212
|
|
253
255
|
if not chunk_tensors:
|
|
254
256
|
# Failed to get chunk tensors.
|
|
255
|
-
logger.warning("Parameters are not properly
|
|
257
|
+
logger.warning("Parameters are not properly flattened, fallback to not flattened parameters.")
|
|
256
258
|
return parameters
|
|
257
259
|
# Convert chunk tensors to parameters.
|
|
258
260
|
self._use_flattened_params = True
|
|
@@ -350,7 +352,7 @@ class Optimizer(Cell):
|
|
|
350
352
|
"""
|
|
351
353
|
Whether use distributed optimizers.
|
|
352
354
|
"""
|
|
353
|
-
return _is_ps_mode()
|
|
355
|
+
return _is_ps_mode()
|
|
354
356
|
|
|
355
357
|
def flatten_gradients(self, gradients):
|
|
356
358
|
"""
|
|
@@ -545,7 +547,7 @@ class Optimizer(Cell):
|
|
|
545
547
|
parameters = self._parameters_base_check(group_param['params'], "group `params`")
|
|
546
548
|
for index, param in enumerate(parameters):
|
|
547
549
|
if not isinstance(param, Parameter):
|
|
548
|
-
raise TypeError(f"For 'Optimizer', the
|
|
550
|
+
raise TypeError(f"For 'Optimizer', the element in group parameters must be Parameter type, "
|
|
549
551
|
f"but got {type(param)} at index {index}.")
|
|
550
552
|
|
|
551
553
|
def _parse_group_params(self, parameters, learning_rate):
|
|
@@ -717,6 +719,7 @@ class Optimizer(Cell):
|
|
|
717
719
|
|
|
718
720
|
Examples:
|
|
719
721
|
>>> from mindspore import nn
|
|
722
|
+
>>> # net = LeNet5()
|
|
720
723
|
>>> net = Net()
|
|
721
724
|
>>> conv_params = list(filter(lambda x: 'conv' in x.name, net.trainable_params()))
|
|
722
725
|
>>> no_conv_params = list(filter(lambda x: 'conv' not in x.name, net.trainable_params()))
|
|
@@ -837,7 +840,7 @@ def _tensor_apply_decay_with_sparse(weight_decay, if_apply, weight, gradient):
|
|
|
837
840
|
indices = gradient.indices
|
|
838
841
|
values = op_add((op_gather(weight, indices, 0) * F.cast(weight_decay, F.dtype(weight)), gradient.values))
|
|
839
842
|
shape = gradient.dense_shape
|
|
840
|
-
return
|
|
843
|
+
return RowTensorInner(indices, values, shape)
|
|
841
844
|
return gradient
|
|
842
845
|
|
|
843
846
|
|
|
@@ -863,7 +866,7 @@ def _tensor_apply_grad_centralization_with_sparse(if_apply, gradient):
|
|
|
863
866
|
if grad_shape[1] % 16 != 0:
|
|
864
867
|
return gradient
|
|
865
868
|
values = op_gc(gradient.values, axis)
|
|
866
|
-
return
|
|
869
|
+
return RowTensorInner(indices, values, shape)
|
|
867
870
|
return gradient
|
|
868
871
|
|
|
869
872
|
|
|
@@ -903,7 +906,13 @@ def tensor_grad_scale_with_tensor(scale, grad):
|
|
|
903
906
|
@_grad_scale.register("Tensor", "RowTensor")
|
|
904
907
|
def tensor_grad_scale_with_sparse(scale, grad):
|
|
905
908
|
"""Get grad with scale."""
|
|
906
|
-
return
|
|
909
|
+
return RowTensorInner(grad.indices, grad.values * F.cast(scale, F.dtype(grad.values)), grad.dense_shape)
|
|
910
|
+
|
|
911
|
+
|
|
912
|
+
@_grad_scale.register("Tensor", "MapTensor")
|
|
913
|
+
def tensor_grad_scale_with_map_tensor(scale, grad):
|
|
914
|
+
"""Get grad with scale."""
|
|
915
|
+
return grad
|
|
907
916
|
|
|
908
917
|
|
|
909
918
|
@_indices_deduplicate.register("RowTensor")
|
|
@@ -915,7 +924,7 @@ def rowtensor_deduplicate_indices_slices(grad):
|
|
|
915
924
|
unique_indices, index_position = P.Unique()(indices)
|
|
916
925
|
summed_values = P.UnsortedSegmentSum()(values, index_position, P.TensorShape()(unique_indices)[0])
|
|
917
926
|
|
|
918
|
-
return
|
|
927
|
+
return RowTensorInner(unique_indices, summed_values, grad.dense_shape)
|
|
919
928
|
|
|
920
929
|
|
|
921
930
|
@_indices_deduplicate.register("Tensor")
|
|
@@ -18,8 +18,8 @@ from __future__ import absolute_import
|
|
|
18
18
|
from mindspore.ops import functional as F, composite as C, operations as P
|
|
19
19
|
from mindspore.common import Tensor
|
|
20
20
|
import mindspore.common.dtype as mstype
|
|
21
|
-
from mindspore.common.api import
|
|
22
|
-
from mindspore
|
|
21
|
+
from mindspore.common.api import jit
|
|
22
|
+
from mindspore import _checkparam as validator
|
|
23
23
|
from mindspore.nn.optim.optimizer import Optimizer
|
|
24
24
|
from mindspore.nn.optim.optimizer import opt_init_args_register
|
|
25
25
|
|
|
@@ -158,7 +158,7 @@ class ProximalAdagrad(Optimizer):
|
|
|
158
158
|
ValueError: If `accum`, `l1`, `l2` or `weight_decay` is less than 0.
|
|
159
159
|
|
|
160
160
|
Supported Platforms:
|
|
161
|
-
``Ascend``
|
|
161
|
+
``Ascend`` ``GPU``
|
|
162
162
|
|
|
163
163
|
Examples:
|
|
164
164
|
>>> import mindspore as ms
|
|
@@ -197,7 +197,7 @@ class ProximalAdagrad(Optimizer):
|
|
|
197
197
|
self.opt = P.ApplyProximalAdagrad(use_locking=use_locking)
|
|
198
198
|
self.sparse_opt = P.SparseApplyProximalAdagrad(use_locking=use_locking)
|
|
199
199
|
|
|
200
|
-
@
|
|
200
|
+
@jit
|
|
201
201
|
def construct(self, grads):
|
|
202
202
|
params = self._parameters
|
|
203
203
|
accum = self.accum
|
|
@@ -231,7 +231,8 @@ class ProximalAdagrad(Optimizer):
|
|
|
231
231
|
"but got {}.".format(value))
|
|
232
232
|
|
|
233
233
|
if value == 'CPU':
|
|
234
|
-
self.sparse_opt = P.FusedSparseProximalAdagrad(self.use_locking)
|
|
234
|
+
self.sparse_opt = P.FusedSparseProximalAdagrad(self.use_locking)
|
|
235
|
+
self.sparse_opt.set_device("CPU")
|
|
235
236
|
else:
|
|
236
237
|
self.sparse_opt = P.SparseApplyProximalAdagrad(self.use_locking)
|
|
237
238
|
|
mindspore/nn/optim/rmsprop.py
CHANGED
|
@@ -16,8 +16,8 @@
|
|
|
16
16
|
from __future__ import absolute_import
|
|
17
17
|
|
|
18
18
|
from mindspore.ops import functional as F, composite as C, operations as P
|
|
19
|
-
from mindspore
|
|
20
|
-
from mindspore.common.api import
|
|
19
|
+
from mindspore import _checkparam as validator
|
|
20
|
+
from mindspore.common.api import jit
|
|
21
21
|
from mindspore.nn.optim.optimizer import Optimizer
|
|
22
22
|
from mindspore.nn.optim.optimizer import opt_init_args_register
|
|
23
23
|
|
|
@@ -225,7 +225,7 @@ class RMSProp(Optimizer):
|
|
|
225
225
|
self.epsilon = epsilon
|
|
226
226
|
self.decay = decay
|
|
227
227
|
|
|
228
|
-
@
|
|
228
|
+
@jit
|
|
229
229
|
def construct(self, gradients):
|
|
230
230
|
params = self._parameters
|
|
231
231
|
gradients = self.flatten_gradients(gradients)
|
mindspore/nn/optim/rprop.py
CHANGED
|
@@ -18,9 +18,8 @@ from __future__ import absolute_import
|
|
|
18
18
|
from mindspore import ops
|
|
19
19
|
from mindspore.ops import operations as P
|
|
20
20
|
import mindspore.common.dtype as mstype
|
|
21
|
-
from mindspore.common.api import
|
|
22
|
-
from mindspore
|
|
23
|
-
from mindspore._checkparam import Rel
|
|
21
|
+
from mindspore.common.api import jit
|
|
22
|
+
from mindspore import _checkparam as validator
|
|
24
23
|
from mindspore.nn.optim.optimizer import Optimizer
|
|
25
24
|
from mindspore.nn.optim.optimizer import opt_init_args_register
|
|
26
25
|
|
|
@@ -30,19 +29,19 @@ class Rprop(Optimizer):
|
|
|
30
29
|
Implements Resilient backpropagation.
|
|
31
30
|
|
|
32
31
|
Further information about this implementation can be found at `A Direct Adaptive Method for Faster Backpropagation
|
|
33
|
-
Learning: The RPROP Algorithm <https://ieeexplore.ieee.org/document/298623>`_
|
|
32
|
+
Learning: The RPROP Algorithm <https://ieeexplore.ieee.org/document/298623>`_.
|
|
34
33
|
|
|
35
34
|
The updating formulas are as follows:
|
|
36
35
|
|
|
37
36
|
.. math::
|
|
38
37
|
\begin{gather*}
|
|
39
38
|
&\hspace{-10mm} \textbf{if} \: g_{t-1} g_t > 0 \\
|
|
40
|
-
&\hspace{25mm} \Delta_t \leftarrow \mathrm{min}(\Delta_{t-1} \eta_{+}, \Delta_{max})
|
|
41
|
-
&\hspace{0mm} \textbf{else if} \: g_{t-1} g_t < 0
|
|
42
|
-
&\hspace{25mm} \Delta_t \leftarrow \mathrm{max}(\Delta_{t-1} \eta_{-}, \Delta_{min})
|
|
43
|
-
&\hspace{-25mm} \textbf{else} \:
|
|
44
|
-
&\hspace{-5mm} \Delta_t \leftarrow \Delta_{t-1}
|
|
45
|
-
&\hspace{15mm} w_{t} \leftarrow w_{t-1}- \Delta_{t} \mathrm{sign}(g_t)
|
|
39
|
+
&\hspace{25mm} \Delta_t \leftarrow \mathrm{min}(\Delta_{t-1} \eta_{+}, \Delta_{max}) \\
|
|
40
|
+
&\hspace{0mm} \textbf{else if} \: g_{t-1} g_t < 0 \\
|
|
41
|
+
&\hspace{25mm} \Delta_t \leftarrow \mathrm{max}(\Delta_{t-1} \eta_{-}, \Delta_{min}) \\
|
|
42
|
+
&\hspace{-25mm} \textbf{else} \: \\
|
|
43
|
+
&\hspace{-5mm} \Delta_t \leftarrow \Delta_{t-1} \\
|
|
44
|
+
&\hspace{15mm} w_{t} \leftarrow w_{t-1}- \Delta_{t} \mathrm{sign}(g_t) \\
|
|
46
45
|
\end{gather*}
|
|
47
46
|
|
|
48
47
|
:math:`\Delta_{min/max}` represents the min/max step size, :math:`\eta_{+/-}` represents the factors of
|
|
@@ -50,8 +49,12 @@ class Rprop(Optimizer):
|
|
|
50
49
|
|
|
51
50
|
Note:
|
|
52
51
|
If parameters are not grouped, the `weight_decay` in optimizer will be applied on the parameters without 'beta'
|
|
53
|
-
or 'gamma' in their names.
|
|
54
|
-
|
|
52
|
+
or 'gamma' in their names.
|
|
53
|
+
|
|
54
|
+
Users can group parameters to change the strategy of decaying weight.
|
|
55
|
+
|
|
56
|
+
When parameters are grouped, each group can set `weight_decay`. If not, the `weight_decay` in optimizer will be
|
|
57
|
+
applied.
|
|
55
58
|
|
|
56
59
|
Args:
|
|
57
60
|
params (Union[list[Parameter], list[dict]]): Must be list of `Parameter` or list of `dict`. When the
|
|
@@ -80,7 +83,7 @@ class Rprop(Optimizer):
|
|
|
80
83
|
If `order_params` in the keys, other keys will be ignored and the element of 'order_params' must be in
|
|
81
84
|
one group of `params`.
|
|
82
85
|
|
|
83
|
-
learning_rate (Union[float, int, Tensor, Iterable, LearningRateSchedule]):
|
|
86
|
+
learning_rate (Union[float, int, Tensor, Iterable, LearningRateSchedule]): Learning_rate. Default: 0.1.
|
|
84
87
|
|
|
85
88
|
- float: The fixed learning rate value. Must be equal to or greater than 0.
|
|
86
89
|
|
|
@@ -95,8 +98,9 @@ class Rprop(Optimizer):
|
|
|
95
98
|
LearningRateSchedule with step as the input to get the learning rate of current step.
|
|
96
99
|
|
|
97
100
|
etas (tuple[float, float]): The factor of multiplicative increasing or
|
|
98
|
-
descreasing(etaminus, etaplus).
|
|
101
|
+
descreasing(etaminus, etaplus). Default: (0.5, 1.2).
|
|
99
102
|
step_sizes(tuple[float, float]): The allowed minimal and maximal step size(min_step_sizes, max_step_size).
|
|
103
|
+
Default: (1e-6, 50.).
|
|
100
104
|
weight_decay (Union[float, int, Cell]): Weight decay (L2 penalty). Default: 0.0.
|
|
101
105
|
|
|
102
106
|
- float: The fixed weight decay value. Must be equal to or greater than 0.
|
|
@@ -170,7 +174,7 @@ class Rprop(Optimizer):
|
|
|
170
174
|
raise ValueError("For Rprop, maximal step size should not be less than minimal step size, "
|
|
171
175
|
"but got {} > {}.".format(step_sizes[0], step_sizes[1]))
|
|
172
176
|
|
|
173
|
-
validator.check_float_range(etas[0], 0.0, 1.0,
|
|
177
|
+
validator.check_float_range(etas[0], 0.0, 1.0, validator.INC_NEITHER, "etaminus", self.cls_name)
|
|
174
178
|
validator.check_value_type("etaplus", etas[1], [float], self.cls_name)
|
|
175
179
|
if etas[1] <= 1.0:
|
|
176
180
|
raise ValueError("For Rprop, etaplus must be greater than 1.0, but got etaplus {}.".format(etas[1]))
|
|
@@ -191,7 +195,7 @@ class Rprop(Optimizer):
|
|
|
191
195
|
self.select = P.Select()
|
|
192
196
|
self.ones_like = P.OnesLike()
|
|
193
197
|
|
|
194
|
-
@
|
|
198
|
+
@jit
|
|
195
199
|
def construct(self, gradients):
|
|
196
200
|
gradients = self.flatten_gradients(gradients)
|
|
197
201
|
gradients = self.decay_weight(gradients)
|
mindspore/nn/optim/sgd.py
CHANGED
|
@@ -18,17 +18,17 @@ from __future__ import absolute_import
|
|
|
18
18
|
from mindspore.ops import functional as F, composite as C, operations as P
|
|
19
19
|
from mindspore.common.parameter import Parameter
|
|
20
20
|
from mindspore.common.tensor import Tensor
|
|
21
|
-
from mindspore.common.api import
|
|
21
|
+
from mindspore.common.api import jit
|
|
22
22
|
import mindspore.common.dtype as mstype
|
|
23
|
-
from mindspore
|
|
23
|
+
from mindspore import _checkparam as validator
|
|
24
24
|
from mindspore.nn.optim.optimizer import Optimizer
|
|
25
25
|
from mindspore.nn.optim.optimizer import opt_init_args_register
|
|
26
26
|
|
|
27
27
|
_sgd_opt = C.MultitypeFuncGraph("sgd_opt")
|
|
28
28
|
|
|
29
29
|
|
|
30
|
-
@_sgd_opt.register("
|
|
31
|
-
def _tensor_run_opt_ext(
|
|
30
|
+
@_sgd_opt.register("Tensor", "Tensor", "Tensor", "Tensor", "Tensor", "Tensor", "Function")
|
|
31
|
+
def _tensor_run_opt_ext(momentum, learning_rate, gradient, weight, accum, stat, opt):
|
|
32
32
|
"""Apply sgd optimizer to the weight parameter using Tensor."""
|
|
33
33
|
success = True
|
|
34
34
|
success = F.depend(success, opt(weight, gradient, learning_rate, accum, momentum, stat))
|
|
@@ -76,7 +76,9 @@ class SGD(Optimizer):
|
|
|
76
76
|
- lr: Optional. If "lr" in the keys, the value of corresponding learning rate will be used.
|
|
77
77
|
If not, the `learning_rate` in optimizer will be used. Fixed and dynamic learning rate are supported.
|
|
78
78
|
|
|
79
|
-
- weight_decay:
|
|
79
|
+
- weight_decay: Optional. If "weight_decay" in the keys, the value of corresponding weight decay
|
|
80
|
+
will be used. If not, the `weight_decay` in the optimizer will be used. It should be noted that weight
|
|
81
|
+
decay must be float, dynamic weight decay is currently not supported.
|
|
80
82
|
|
|
81
83
|
- grad_centralization: Optional. Must be Boolean. If "grad_centralization" is in the keys, the set value
|
|
82
84
|
will be used. If not, the `grad_centralization` is False by default. This configuration only works on the
|
|
@@ -164,7 +166,7 @@ class SGD(Optimizer):
|
|
|
164
166
|
|
|
165
167
|
if isinstance(momentum, float) and momentum < 0.0:
|
|
166
168
|
raise ValueError("For 'SGD', the argument 'momentum' must be at least 0.0, "
|
|
167
|
-
"but got {}".format(momentum))
|
|
169
|
+
"but got {}.".format(momentum))
|
|
168
170
|
|
|
169
171
|
if isinstance(dampening, int):
|
|
170
172
|
dampening = float(dampening)
|
|
@@ -177,9 +179,6 @@ class SGD(Optimizer):
|
|
|
177
179
|
"but got 'dampening' {}".format(dampening))
|
|
178
180
|
self.dampening = dampening
|
|
179
181
|
|
|
180
|
-
if isinstance(weight_decay, int):
|
|
181
|
-
weight_decay = float(weight_decay)
|
|
182
|
-
|
|
183
182
|
validator.check_value_type("nesterov", nesterov, [bool], self.cls_name)
|
|
184
183
|
|
|
185
184
|
if nesterov and (momentum <= 0.0 or dampening != 0.0):
|
|
@@ -187,13 +186,20 @@ class SGD(Optimizer):
|
|
|
187
186
|
"equal to 0.0, but got 'momentum' {}, 'dampening' {}".format(momentum, dampening))
|
|
188
187
|
self.nesterov = nesterov
|
|
189
188
|
|
|
190
|
-
self.
|
|
189
|
+
if self.dynamic_weight_decay:
|
|
190
|
+
raise TypeError("For 'SGD', dynamic weight decay is currently not supported, the argument 'weight_decay' "
|
|
191
|
+
"or 'weight_decay' set in grouped 'params' must be float or int type.")
|
|
192
|
+
|
|
193
|
+
if hasattr(self, "group_weight_decay") and self.group_weight_decay:
|
|
194
|
+
self.opt = tuple(P.SGD(dampening, wd, nesterov) for wd in self.group_weight_decay)
|
|
195
|
+
else:
|
|
196
|
+
self.opt = tuple([P.SGD(dampening, float(weight_decay), nesterov)] * len(self._parameters))
|
|
191
197
|
|
|
192
198
|
self.momentum = Parameter(Tensor(momentum, mstype.float32), name="momentum")
|
|
193
199
|
self.accum = self._parameters.clone(prefix="accum", init='zeros')
|
|
194
200
|
self.stat = self._parameters.clone(prefix="stat", init='ones')
|
|
195
201
|
|
|
196
|
-
@
|
|
202
|
+
@jit
|
|
197
203
|
def construct(self, gradients):
|
|
198
204
|
params = self._parameters
|
|
199
205
|
accum = self.accum
|
|
@@ -203,9 +209,9 @@ class SGD(Optimizer):
|
|
|
203
209
|
gradients = self.scale_grad(gradients)
|
|
204
210
|
lr = self.get_lr()
|
|
205
211
|
if self.is_group_lr:
|
|
206
|
-
success = self.hyper_map_reverse(F.partial(_sgd_opt, self.
|
|
207
|
-
lr, gradients, params, accum, stat)
|
|
212
|
+
success = self.hyper_map_reverse(F.partial(_sgd_opt, self.momentum),
|
|
213
|
+
lr, gradients, params, accum, stat, self.opt)
|
|
208
214
|
else:
|
|
209
|
-
success = self.hyper_map_reverse(F.partial(_sgd_opt, self.
|
|
210
|
-
gradients, params, accum, stat)
|
|
215
|
+
success = self.hyper_map_reverse(F.partial(_sgd_opt, self.momentum, lr),
|
|
216
|
+
gradients, params, accum, stat, self.opt)
|
|
211
217
|
return success
|
mindspore/nn/optim/thor.py
CHANGED
|
@@ -21,10 +21,11 @@ from mindspore.ops import functional as F, composite as C, operations as P
|
|
|
21
21
|
from mindspore.common.initializer import initializer
|
|
22
22
|
from mindspore.common.parameter import Parameter, ParameterTuple
|
|
23
23
|
from mindspore.common.tensor import Tensor
|
|
24
|
+
import mindspore.ops as ops
|
|
24
25
|
import mindspore.nn as nn
|
|
25
26
|
import mindspore.common.dtype as mstype
|
|
26
27
|
import mindspore.log as logger
|
|
27
|
-
from mindspore
|
|
28
|
+
from mindspore import _checkparam as Validator
|
|
28
29
|
from mindspore.nn.optim.optimizer import Optimizer
|
|
29
30
|
from mindspore.parallel._utils import _get_device_num, _get_gradients_mean
|
|
30
31
|
from mindspore import context
|
|
@@ -87,8 +88,8 @@ def _clip_grad(clip_type, clip_value, grad):
|
|
|
87
88
|
return grad
|
|
88
89
|
dt = F.dtype(grad)
|
|
89
90
|
if clip_type == 0:
|
|
90
|
-
new_grad =
|
|
91
|
-
|
|
91
|
+
new_grad = ops.clip_by_value(grad, F.cast(F.tuple_to_array((-clip_value,)), dt),
|
|
92
|
+
F.cast(F.tuple_to_array((clip_value,)), dt))
|
|
92
93
|
else:
|
|
93
94
|
new_grad = nn.ClipByNorm()(grad, F.cast(F.tuple_to_array((clip_value,)), dt))
|
|
94
95
|
return new_grad
|
|
@@ -210,7 +211,7 @@ def find_net_layertype_recur(net, layertype_map):
|
|
|
210
211
|
if subcell.gamma.requires_grad:
|
|
211
212
|
layertype_map.append(BatchNorm)
|
|
212
213
|
elif isinstance(subcell, (nn.Conv2d, nn.Dense, nn.Embedding, nn.Conv2dTranspose, nn.Conv1d, nn.Conv1dTranspose,
|
|
213
|
-
nn.BatchNorm1d, nn.GroupNorm
|
|
214
|
+
nn.BatchNorm1d, nn.GroupNorm)):
|
|
214
215
|
if isinstance(subcell, (nn.Dense, nn.Conv2d)):
|
|
215
216
|
get_layer_type_for_dense_and_conv(subcell, prefix, layertype_map)
|
|
216
217
|
else:
|
|
@@ -253,11 +254,6 @@ def thor(net, learning_rate, damping, momentum, weight_decay=0.0, loss_scale=1.0
|
|
|
253
254
|
r"""
|
|
254
255
|
Updates gradients by second-order algorithm--THOR.
|
|
255
256
|
|
|
256
|
-
Trace-based Hardware-driven layer-ORiented Natural Gradient Descent Computation (THOR) algorithm is proposed in:
|
|
257
|
-
|
|
258
|
-
`THOR: Trace-based Hardware-driven layer-ORiented Natural Gradient Descent Computation
|
|
259
|
-
<https://www.aaai.org/AAAI21Papers/AAAI-6611.ChenM.pdf>`_
|
|
260
|
-
|
|
261
257
|
The updating formulas are as follows,
|
|
262
258
|
|
|
263
259
|
.. math::
|
|
@@ -270,11 +266,11 @@ def thor(net, learning_rate, damping, momentum, weight_decay=0.0, loss_scale=1.0
|
|
|
270
266
|
\otimes\left(G_{i}^{(k)}+\lambda I\right)^{-1}\right) \nabla_{w_{i}} J^{(k)}
|
|
271
267
|
\end{array}
|
|
272
268
|
|
|
273
|
-
:math:`a_{i-1}` represents the input of i-th layer,and which is the activations of previous layer
|
|
274
|
-
:math:`D_{s_i}` represents the derivative of the loss function of the output of the i-th layer
|
|
275
|
-
:math:`I` represents the identity matrix
|
|
276
|
-
:math:`\lambda` represents :math:`damping`, :math:`g_i` represents gradients of the i-th layer
|
|
277
|
-
:math:`\otimes` represents Kronecker product, :math:`\gamma` represents 'learning rate'
|
|
269
|
+
:math:`a_{i-1}` represents the input of i-th layer,and which is the activations of previous layer.
|
|
270
|
+
:math:`D_{s_i}` represents the derivative of the loss function of the output of the i-th layer.
|
|
271
|
+
:math:`I` represents the identity matrix.
|
|
272
|
+
:math:`\lambda` represents :math:`damping`, :math:`g_i` represents gradients of the i-th layer.
|
|
273
|
+
:math:`\otimes` represents Kronecker product, :math:`\gamma` represents 'learning rate'.
|
|
278
274
|
|
|
279
275
|
Note:
|
|
280
276
|
When a parameter group is separated, 'weight_decay' of each group is applied to the corresponding parameter.
|
|
@@ -313,9 +309,9 @@ def thor(net, learning_rate, damping, momentum, weight_decay=0.0, loss_scale=1.0
|
|
|
313
309
|
|
|
314
310
|
enable_clip_grad (bool): Whether to clip the gradients. Default: False
|
|
315
311
|
|
|
316
|
-
frequency(int): The update interval of A/G and
|
|
317
|
-
A/G and
|
|
318
|
-
|
|
312
|
+
frequency(int): The update interval of A/G and :math:`A^{-1}/G^{-1}`. When frequency equals N
|
|
313
|
+
(N is greater than 1), A/G and :math:`A^{-1}/G^{-1}` will be updated every N steps,
|
|
314
|
+
and other steps will use the stale A/G and :math:`A^{-1}/G^{-1}` to update weights. Default: 100.
|
|
319
315
|
|
|
320
316
|
Inputs:
|
|
321
317
|
- **gradients** (tuple[Tensor]) - The gradients of `params`, the shape is the same as `params`.
|
|
@@ -337,6 +333,12 @@ def thor(net, learning_rate, damping, momentum, weight_decay=0.0, loss_scale=1.0
|
|
|
337
333
|
``Ascend`` ``GPU``
|
|
338
334
|
|
|
339
335
|
Examples:
|
|
336
|
+
.. note::
|
|
337
|
+
Before running the following example, you need to customize the network Net and
|
|
338
|
+
dataset preparation function create_dataset. Refer to
|
|
339
|
+
`Building a Network <https://www.mindspore.cn/tutorials/en/r2.0/beginner/model.html>`_
|
|
340
|
+
and `Dataset <https://www.mindspore.cn/tutorials/en/r2.0/beginner/dataset.html>`_ .
|
|
341
|
+
|
|
340
342
|
>>> import mindspore as ms
|
|
341
343
|
>>> from mindspore.nn import thor
|
|
342
344
|
>>> from mindspore import nn
|
|
@@ -966,15 +968,15 @@ class ThorAscend(Optimizer):
|
|
|
966
968
|
matrix_g_combine_shape = self.shape(matrix_g_inv)
|
|
967
969
|
if matrix_a_inv_shape[0] == 2048 and matrix_g_combine_shape[0] == 1001:
|
|
968
970
|
matrix_a_inv = self.reshape(matrix_a_inv,
|
|
969
|
-
(matrix_a_inv_shape[0]
|
|
970
|
-
matrix_a_inv_shape[0]
|
|
971
|
+
(matrix_a_inv_shape[0] // 16, 16,
|
|
972
|
+
matrix_a_inv_shape[0] // 16, 16))
|
|
971
973
|
matrix_a_inv = self.transpose(matrix_a_inv, (2, 0, 1, 3))
|
|
972
974
|
matrix_g_inv = P.Pad(((0, 7), (0, 7)))(matrix_g_inv)
|
|
973
975
|
|
|
974
976
|
matrix_g_inv_shape = self.shape(matrix_g_inv)
|
|
975
977
|
matrix_g_inv = self.reshape(matrix_g_inv,
|
|
976
|
-
(matrix_g_inv_shape[0]
|
|
977
|
-
matrix_g_inv_shape[0]
|
|
978
|
+
(matrix_g_inv_shape[0] // 16, 16,
|
|
979
|
+
matrix_g_inv_shape[0] // 16, 16))
|
|
978
980
|
matrix_g_inv = self.transpose(matrix_g_inv, (2, 0, 1, 3))
|
|
979
981
|
|
|
980
982
|
matrix_a_allreduce = matrix_a_allreduce + (matrix_a_inv,)
|
|
@@ -16,9 +16,10 @@
|
|
|
16
16
|
from mindspore import context
|
|
17
17
|
from mindspore.nn.cell import Cell
|
|
18
18
|
from mindspore.ops import operations as P
|
|
19
|
+
from mindspore.ops.operations import _inner_ops as inner
|
|
19
20
|
from mindspore.common import dtype as mstype
|
|
20
21
|
from mindspore.common.tensor import Tensor
|
|
21
|
-
from mindspore
|
|
22
|
+
from mindspore import _checkparam as validator
|
|
22
23
|
from ..distribution._utils.utils import CheckTensor, cast_to_tensor, raise_type_error
|
|
23
24
|
from ..distribution import Distribution
|
|
24
25
|
from ..distribution import TransformedDistribution
|
|
@@ -27,9 +28,9 @@ from ..distribution import TransformedDistribution
|
|
|
27
28
|
class Bijector(Cell):
|
|
28
29
|
"""
|
|
29
30
|
Bijecotr class. A bijector perform a mapping from one distribution to the other via some function.
|
|
30
|
-
If X is a random variable following the original distribution,
|
|
31
|
-
and g(x) is the mapping function,
|
|
32
|
-
then Y = g(X) is the random variable following the transformed distribution.
|
|
31
|
+
If :math:`X` is a random variable following the original distribution,
|
|
32
|
+
and :math:`g(x)` is the mapping function,
|
|
33
|
+
then :math:`Y = g(X)` is the random variable following the transformed distribution.
|
|
33
34
|
|
|
34
35
|
Args:
|
|
35
36
|
is_constant_jacobian (bool): Whether the Bijector has constant derivative. Default: False.
|
|
@@ -96,8 +97,8 @@ class Bijector(Cell):
|
|
|
96
97
|
self.dtype_base = P.DType()
|
|
97
98
|
self.shape_base = P.Shape()
|
|
98
99
|
self.fill_base = P.Fill()
|
|
99
|
-
self.sametypeshape_base =
|
|
100
|
-
self.issubclass_base =
|
|
100
|
+
self.sametypeshape_base = inner.SameTypeShape()
|
|
101
|
+
self.issubclass_base = inner.IsSubClass()
|
|
101
102
|
|
|
102
103
|
@property
|
|
103
104
|
def name(self):
|
|
@@ -13,13 +13,15 @@
|
|
|
13
13
|
# limitations under the License.
|
|
14
14
|
# ============================================================================
|
|
15
15
|
"""Invert Bijector"""
|
|
16
|
-
from mindspore
|
|
16
|
+
from mindspore import _checkparam as validator
|
|
17
17
|
from .bijector import Bijector
|
|
18
18
|
|
|
19
19
|
|
|
20
20
|
class Invert(Bijector):
|
|
21
21
|
r"""
|
|
22
|
-
Invert Bijector. Compute the inverse function of the input bijector.
|
|
22
|
+
Invert Bijector. Compute the inverse function of the input bijector. If the function of the forward mapping,
|
|
23
|
+
namely the input of `bijector` below, is :math:`Y = g(X)`,
|
|
24
|
+
then the function of corresponding inverse mapping Bijector is :math:`Y = h(X) = g^{-1}(X)`.
|
|
23
25
|
|
|
24
26
|
Args:
|
|
25
27
|
bijector (Bijector): Base Bijector.
|
|
@@ -41,7 +41,7 @@ class Softplus(Bijector):
|
|
|
41
41
|
TypeError: When the dtype of the sharpness is not float.
|
|
42
42
|
|
|
43
43
|
Supported Platforms:
|
|
44
|
-
``Ascend`` ``GPU``
|
|
44
|
+
``Ascend`` ``GPU`` ``CPU``
|
|
45
45
|
|
|
46
46
|
Examples:
|
|
47
47
|
>>> import mindspore
|
|
@@ -51,7 +51,7 @@ class Softplus(Bijector):
|
|
|
51
51
|
>>>
|
|
52
52
|
>>> # To initialize a Softplus bijector of sharpness 2.0.
|
|
53
53
|
>>> softplus = msb.Softplus(2.0)
|
|
54
|
-
>>> # To use a
|
|
54
|
+
>>> # To use a Softplus bijector in a network.
|
|
55
55
|
>>> value = Tensor([1, 2, 3], dtype=mindspore.float32)
|
|
56
56
|
>>> ans1 = softplus.forward(value)
|
|
57
57
|
>>> print(ans1.shape)
|
|
@@ -14,7 +14,7 @@
|
|
|
14
14
|
# ============================================================================
|
|
15
15
|
"""dense_variational"""
|
|
16
16
|
from mindspore.ops import operations as P
|
|
17
|
-
from mindspore
|
|
17
|
+
from mindspore import _checkparam as Validator
|
|
18
18
|
from ...cell import Cell
|
|
19
19
|
from ...layer.activation import get_activation
|
|
20
20
|
from ..distribution.normal import Normal
|
|
@@ -29,7 +29,7 @@ class NormalPrior(Cell):
|
|
|
29
29
|
To initialize a normal distribution of mean 0 and standard deviation 0.1.
|
|
30
30
|
|
|
31
31
|
Args:
|
|
32
|
-
dtype (
|
|
32
|
+
dtype (mindspore.dtype): The argument is used to define the data type of the output tensor.
|
|
33
33
|
Default: mindspore.float32.
|
|
34
34
|
mean (int, float): Mean of normal distribution. Default: 0.
|
|
35
35
|
std (int, float): Standard deviation of normal distribution. Default: 0.1.
|
|
@@ -55,7 +55,7 @@ class NormalPosterior(Cell):
|
|
|
55
55
|
Args:
|
|
56
56
|
name (str): Name prepended to trainable parameter.
|
|
57
57
|
shape (list, tuple): Shape of the mean and standard deviation.
|
|
58
|
-
dtype (
|
|
58
|
+
dtype (mindspore.dtype): The argument is used to define the data type of the output tensor.
|
|
59
59
|
Default: mindspore.float32.
|
|
60
60
|
loc_mean (int, float): Mean of distribution to initialize trainable parameters. Default: 0.
|
|
61
61
|
loc_std (int, float): Standard deviation of distribution to initialize trainable parameters. Default: 0.1.
|
|
@@ -31,6 +31,9 @@ from .log_normal import LogNormal
|
|
|
31
31
|
from .normal import Normal
|
|
32
32
|
from .poisson import Poisson
|
|
33
33
|
from .uniform import Uniform
|
|
34
|
+
from .half_normal import HalfNormal
|
|
35
|
+
from .laplace import Laplace
|
|
36
|
+
from .student_t import StudentT
|
|
34
37
|
|
|
35
38
|
__all__ = ['Distribution',
|
|
36
39
|
'TransformedDistribution',
|
|
@@ -47,4 +50,7 @@ __all__ = ['Distribution',
|
|
|
47
50
|
'Normal',
|
|
48
51
|
'Poisson',
|
|
49
52
|
'Uniform',
|
|
53
|
+
'HalfNormal',
|
|
54
|
+
'Laplace',
|
|
55
|
+
'StudentT',
|
|
50
56
|
]
|
|
@@ -15,6 +15,7 @@
|
|
|
15
15
|
"""Utility functions to help distribution class."""
|
|
16
16
|
import numpy as np
|
|
17
17
|
from mindspore.ops import operations as P
|
|
18
|
+
from mindspore.ops.operations import _inner_ops as inner
|
|
18
19
|
from mindspore.common import dtype as mstype
|
|
19
20
|
|
|
20
21
|
|
|
@@ -26,7 +27,7 @@ def exp_generic(input_x):
|
|
|
26
27
|
exp = P.Exp()
|
|
27
28
|
cast = P.Cast()
|
|
28
29
|
dtype = P.DType()
|
|
29
|
-
checktype =
|
|
30
|
+
checktype = inner.IsSubClass()
|
|
30
31
|
|
|
31
32
|
if not checktype(dtype(input_x), mstype.float_):
|
|
32
33
|
input_x = cast(input_x, mstype.float32)
|
|
@@ -48,7 +49,7 @@ def log_generic(input_x):
|
|
|
48
49
|
dtype = P.DType()
|
|
49
50
|
shape = P.Shape()
|
|
50
51
|
select = P.Select()
|
|
51
|
-
checktype =
|
|
52
|
+
checktype = inner.IsSubClass()
|
|
52
53
|
|
|
53
54
|
if not checktype(dtype(input_x), mstype.float_):
|
|
54
55
|
input_x = cast(input_x, mstype.float32)
|