mindspore 1.10.0__cp37-none-any.whl → 2.0.0rc1__cp37-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of mindspore might be problematic. Click here for more details.
- mindspore/.commit_id +1 -1
- mindspore/Third_Party_Open_Source_Software_Notice +9064 -0
- mindspore/__init__.py +9 -4
- mindspore/_akg/akg/composite/build_module.py +11 -0
- mindspore/_akg/akg/config/repository_cuda.json +11 -0
- mindspore/_akg/akg/tvm/contrib/nvcc.py +4 -3
- mindspore/_c_dataengine.cpython-37m-aarch64-linux-gnu.so +0 -0
- mindspore/_c_expression.cpython-37m-aarch64-linux-gnu.so +0 -0
- mindspore/_c_mindrecord.cpython-37m-aarch64-linux-gnu.so +0 -0
- mindspore/_check_jit_forbidden_api.py +102 -0
- mindspore/_checkparam.py +1066 -1001
- mindspore/_extends/builtin_operations.py +32 -4
- mindspore/_extends/graph_kernel/model/graph_split.py +66 -222
- mindspore/_extends/parallel_compile/akg_compiler/akg_process.py +12 -9
- mindspore/_extends/parallel_compile/akg_compiler/build_tbe_kernel.py +119 -26
- mindspore/_extends/parallel_compile/akg_compiler/tbe_topi.py +50 -50
- mindspore/_extends/parallel_compile/akg_compiler/util.py +9 -6
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_adapter.py +4 -25
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_helper.py +9 -4
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_job_manager.py +1 -27
- mindspore/_extends/parse/__init__.py +5 -3
- mindspore/_extends/parse/namespace.py +17 -2
- mindspore/_extends/parse/parser.py +193 -34
- mindspore/_extends/parse/resources.py +7 -8
- mindspore/_extends/parse/standard_method.py +1780 -435
- mindspore/_extends/parse/trope.py +3 -1
- mindspore/_mindspore_offline_debug.cpython-37m-aarch64-linux-gnu.so +0 -0
- mindspore/amp.py +53 -58
- mindspore/bin/cache_admin +0 -0
- mindspore/bin/cache_server +0 -0
- mindspore/boost/adasum.py +3 -2
- mindspore/boost/boost.py +2 -2
- mindspore/boost/boost_cell_wrapper.py +46 -26
- mindspore/boost/dim_reduce.py +6 -5
- mindspore/boost/grad_accumulation.py +2 -1
- mindspore/boost/group_loss_scale_manager.py +1 -1
- mindspore/common/__init__.py +11 -10
- mindspore/common/_decorator.py +2 -0
- mindspore/common/_register_for_adapter.py +55 -0
- mindspore/common/_stub_tensor.py +201 -0
- mindspore/common/_utils.py +57 -0
- mindspore/common/api.py +582 -297
- mindspore/common/dtype.py +66 -18
- mindspore/common/dump.py +2 -2
- mindspore/common/initializer.py +38 -1
- mindspore/common/jit_config.py +25 -13
- mindspore/common/mutable.py +53 -24
- mindspore/common/parameter.py +60 -37
- mindspore/common/seed.py +8 -24
- mindspore/common/sparse_tensor.py +927 -0
- mindspore/common/tensor.py +1627 -3900
- mindspore/communication/__init__.py +10 -5
- mindspore/communication/_comm_helper.py +78 -214
- mindspore/communication/_hccl_management.py +2 -1
- mindspore/communication/management.py +136 -47
- mindspore/config/op_info.config +501 -1008
- mindspore/config/super_bar_config.json +512 -0
- mindspore/context.py +291 -56
- mindspore/dataset/__init__.py +12 -8
- mindspore/dataset/audio/__init__.py +9 -9
- mindspore/dataset/audio/transforms.py +1090 -228
- mindspore/dataset/audio/utils.py +87 -39
- mindspore/dataset/audio/validators.py +223 -1
- mindspore/dataset/callback/ds_callback.py +17 -15
- mindspore/dataset/core/config.py +246 -17
- mindspore/dataset/core/py_util_helpers.py +4 -3
- mindspore/dataset/core/validator_helpers.py +10 -10
- mindspore/{parallel/nn/layers.py → dataset/debug/__init__.py} +7 -8
- mindspore/dataset/debug/debug_hook.py +65 -0
- mindspore/dataset/debug/pre_defined_hook.py +67 -0
- mindspore/dataset/engine/__init__.py +7 -3
- mindspore/dataset/engine/cache_client.py +9 -9
- mindspore/dataset/engine/datasets.py +648 -477
- mindspore/dataset/engine/datasets_audio.py +165 -167
- mindspore/dataset/engine/datasets_standard_format.py +93 -67
- mindspore/dataset/engine/datasets_text.py +492 -342
- mindspore/dataset/engine/datasets_user_defined.py +85 -50
- mindspore/dataset/engine/datasets_vision.py +1224 -699
- mindspore/dataset/engine/graphdata.py +134 -69
- mindspore/dataset/engine/iterators.py +50 -9
- mindspore/dataset/engine/offload.py +52 -31
- mindspore/dataset/engine/samplers.py +27 -24
- mindspore/dataset/engine/serializer_deserializer.py +14 -15
- mindspore/dataset/engine/validators.py +213 -52
- mindspore/dataset/text/__init__.py +10 -8
- mindspore/dataset/text/transforms.py +152 -57
- mindspore/dataset/text/utils.py +98 -49
- mindspore/dataset/text/validators.py +25 -0
- mindspore/dataset/transforms/__init__.py +4 -2
- mindspore/dataset/transforms/c_transforms.py +11 -13
- mindspore/dataset/transforms/py_transforms.py +2 -2
- mindspore/dataset/transforms/py_transforms_util.py +10 -0
- mindspore/dataset/transforms/transforms.py +13 -15
- mindspore/dataset/transforms/validators.py +7 -7
- mindspore/dataset/utils/__init__.py +2 -1
- mindspore/dataset/utils/browse_dataset.py +13 -13
- mindspore/dataset/utils/line_reader.py +121 -0
- mindspore/dataset/vision/__init__.py +8 -7
- mindspore/dataset/vision/c_transforms.py +125 -126
- mindspore/dataset/vision/py_transforms.py +37 -37
- mindspore/dataset/vision/py_transforms_util.py +23 -20
- mindspore/dataset/vision/transforms.py +316 -315
- mindspore/dataset/vision/utils.py +313 -17
- mindspore/dataset/vision/validators.py +6 -6
- mindspore/default_config.py +0 -1
- mindspore/{compression → experimental}/__init__.py +6 -5
- mindspore/experimental/map_parameter.py +275 -0
- mindspore/include/OWNERS +0 -1
- mindspore/include/api/callback/callback.h +9 -13
- mindspore/include/api/callback/ckpt_saver.h +2 -2
- mindspore/include/api/callback/loss_monitor.h +2 -2
- mindspore/include/api/callback/lr_scheduler.h +5 -5
- mindspore/include/api/callback/time_monitor.h +2 -2
- mindspore/include/api/callback/train_accuracy.h +4 -6
- mindspore/include/api/cfg.h +19 -6
- mindspore/include/api/context.h +70 -9
- mindspore/include/api/delegate.h +8 -1
- mindspore/include/api/dual_abi_helper.h +8 -24
- mindspore/include/api/metrics/accuracy.h +2 -2
- mindspore/include/api/metrics/metrics.h +4 -3
- mindspore/include/api/model.h +9 -4
- mindspore/include/api/model_group.h +68 -0
- mindspore/include/api/model_parallel_runner.h +17 -17
- mindspore/include/api/net.h +12 -11
- mindspore/include/api/serialization.h +20 -4
- mindspore/include/api/status.h +7 -1
- mindspore/include/api/types.h +25 -21
- mindspore/include/api/visible.h +4 -0
- mindspore/include/c_api/model_c.h +5 -0
- mindspore/include/c_api/status_c.h +1 -1
- mindspore/include/dataset/config.h +1 -1
- mindspore/include/dataset/constants.h +14 -0
- mindspore/include/dataset/text.h +59 -0
- mindspore/include/dataset/vision.h +56 -117
- mindspore/include/dataset/vision_lite.h +102 -0
- mindspore/include/mindapi/base/type_id.h +42 -3
- mindspore/lib/libdnnl.so.2 +0 -0
- mindspore/lib/libicudata.so.69 +0 -0
- mindspore/lib/libicui18n.so.69 +0 -0
- mindspore/lib/libicuuc.so.69 +0 -0
- mindspore/lib/libmindspore.so +0 -0
- mindspore/lib/libmindspore_backend.so +0 -0
- mindspore/lib/libmindspore_common.so +0 -0
- mindspore/lib/libmindspore_core.so +0 -0
- mindspore/lib/libmindspore_glog.so.0 +0 -0
- mindspore/lib/libmindspore_gpr.so.15 +0 -0
- mindspore/lib/libmindspore_grpc++.so.1 +0 -0
- mindspore/lib/libmindspore_grpc.so.15 +0 -0
- mindspore/lib/libmindspore_shared_lib.so +0 -0
- mindspore/lib/libmpi_adapter.so +0 -0
- mindspore/lib/libmpi_collective.so +0 -0
- mindspore/lib/libnnacl.so +0 -0
- mindspore/lib/libopencv_core.so.4.5 +0 -0
- mindspore/lib/libopencv_imgcodecs.so.4.5 +0 -0
- mindspore/lib/libopencv_imgproc.so.4.5 +0 -0
- mindspore/lib/libps_cache.so +0 -0
- mindspore/lib/plugin/ascend/libakg.so +0 -0
- mindspore/lib/plugin/ascend/libascend_collective.so +0 -0
- mindspore/lib/plugin/ascend/libdvpp_utils.so +0 -0
- mindspore/lib/plugin/ascend/libhccl_plugin.so +0 -0
- mindspore/lib/plugin/ascend/libmindspore_aicpu_kernels.so +0 -0
- mindspore/lib/plugin/ascend/libmindspore_cpu_kernels.so +0 -0
- mindspore/lib/{libakg.so → plugin/cpu/libakg.so} +0 -0
- mindspore/lib/plugin/libmindspore_ascend.so.1 +0 -0
- mindspore/lib/plugin/libmindspore_ascend.so.2 +0 -0
- mindspore/log.py +28 -28
- mindspore/mindrecord/common/exceptions.py +2 -4
- mindspore/mindrecord/filereader.py +19 -1
- mindspore/mindrecord/filewriter.py +250 -88
- mindspore/mindrecord/mindpage.py +13 -13
- mindspore/mindrecord/shardheader.py +15 -15
- mindspore/mindrecord/shardreader.py +9 -0
- mindspore/mindrecord/shardwriter.py +29 -29
- mindspore/mindrecord/tools/cifar100_to_mr.py +9 -9
- mindspore/mindrecord/tools/cifar10_to_mr.py +9 -9
- mindspore/mindrecord/tools/csv_to_mr.py +4 -4
- mindspore/mindrecord/tools/imagenet_to_mr.py +70 -65
- mindspore/mindrecord/tools/mnist_to_mr.py +41 -41
- mindspore/mindrecord/tools/tfrecord_to_mr.py +6 -6
- mindspore/nn/__init__.py +1 -5
- mindspore/nn/cell.py +297 -234
- mindspore/nn/dynamic_lr.py +1 -1
- mindspore/nn/grad/cell_grad.py +17 -42
- mindspore/nn/layer/__init__.py +7 -4
- mindspore/nn/layer/activation.py +131 -88
- mindspore/nn/layer/basic.py +313 -613
- mindspore/nn/layer/channel_shuffle.py +103 -0
- mindspore/nn/layer/combined.py +1 -1
- mindspore/nn/layer/container.py +52 -6
- mindspore/nn/layer/conv.py +112 -43
- mindspore/nn/layer/dense.py +10 -9
- mindspore/nn/layer/embedding.py +36 -34
- mindspore/nn/layer/image.py +123 -27
- mindspore/nn/layer/math.py +108 -107
- mindspore/nn/layer/normalization.py +212 -366
- mindspore/nn/layer/padding.py +370 -42
- mindspore/nn/layer/pooling.py +1443 -219
- mindspore/nn/layer/rnn_cells.py +11 -16
- mindspore/nn/layer/rnns.py +38 -39
- mindspore/nn/layer/thor_layer.py +24 -25
- mindspore/nn/layer/timedistributed.py +5 -5
- mindspore/nn/layer/transformer.py +701 -0
- mindspore/nn/learning_rate_schedule.py +8 -8
- mindspore/nn/loss/__init__.py +9 -6
- mindspore/nn/loss/loss.py +678 -142
- mindspore/nn/metrics.py +53 -0
- mindspore/nn/optim/_dist_optimizer_registry.py +2 -2
- mindspore/nn/optim/ada_grad.py +8 -8
- mindspore/nn/optim/adadelta.py +2 -3
- mindspore/nn/optim/adafactor.py +18 -14
- mindspore/nn/optim/adam.py +429 -87
- mindspore/nn/optim/adamax.py +5 -6
- mindspore/nn/optim/adasum.py +10 -8
- mindspore/nn/optim/asgd.py +7 -7
- mindspore/nn/optim/ftrl.py +81 -11
- mindspore/nn/optim/lamb.py +7 -8
- mindspore/nn/optim/lars.py +4 -4
- mindspore/nn/optim/lazyadam.py +82 -7
- mindspore/nn/optim/momentum.py +8 -7
- mindspore/nn/optim/optimizer.py +19 -10
- mindspore/nn/optim/proximal_ada_grad.py +6 -5
- mindspore/nn/optim/rmsprop.py +3 -3
- mindspore/nn/optim/rprop.py +20 -16
- mindspore/nn/optim/sgd.py +21 -15
- mindspore/nn/optim/thor.py +23 -21
- mindspore/nn/probability/__init__.py +0 -2
- mindspore/nn/probability/bijector/bijector.py +7 -6
- mindspore/nn/probability/bijector/invert.py +4 -2
- mindspore/nn/probability/bijector/softplus.py +2 -2
- mindspore/nn/probability/bnn_layers/dense_variational.py +1 -1
- mindspore/nn/probability/bnn_layers/layer_distribution.py +2 -2
- mindspore/nn/probability/distribution/__init__.py +6 -0
- mindspore/nn/probability/distribution/_utils/custom_ops.py +3 -2
- mindspore/nn/probability/distribution/_utils/utils.py +11 -17
- mindspore/nn/probability/distribution/bernoulli.py +6 -6
- mindspore/nn/probability/distribution/beta.py +1 -1
- mindspore/nn/probability/distribution/categorical.py +9 -9
- mindspore/nn/probability/distribution/cauchy.py +8 -8
- mindspore/nn/probability/distribution/distribution.py +12 -6
- mindspore/nn/probability/distribution/exponential.py +5 -5
- mindspore/nn/probability/distribution/gamma.py +3 -3
- mindspore/nn/probability/distribution/geometric.py +6 -5
- mindspore/nn/probability/distribution/gumbel.py +5 -5
- mindspore/nn/probability/distribution/half_normal.py +133 -0
- mindspore/nn/probability/distribution/laplace.py +128 -0
- mindspore/nn/probability/distribution/log_normal.py +0 -1
- mindspore/nn/probability/distribution/logistic.py +4 -5
- mindspore/nn/probability/distribution/normal.py +11 -15
- mindspore/nn/probability/distribution/poisson.py +6 -2
- mindspore/nn/probability/distribution/student_t.py +150 -0
- mindspore/nn/probability/distribution/transformed_distribution.py +4 -4
- mindspore/nn/probability/distribution/uniform.py +5 -5
- mindspore/nn/reinforcement/_tensors_queue.py +3 -3
- mindspore/nn/reinforcement/tensor_array.py +2 -2
- mindspore/nn/sparse/sparse.py +8 -1
- mindspore/nn/wrap/cell_wrapper.py +55 -27
- mindspore/nn/wrap/grad_reducer.py +20 -11
- mindspore/nn/wrap/loss_scale.py +47 -30
- mindspore/numpy/array_creations.py +33 -22
- mindspore/numpy/array_ops.py +46 -42
- mindspore/numpy/logic_ops.py +6 -27
- mindspore/numpy/math_ops.py +26 -19
- mindspore/numpy/utils.py +1 -8
- mindspore/numpy/utils_const.py +112 -62
- mindspore/ops/__init__.py +6 -3
- mindspore/ops/_constants.py +0 -6
- mindspore/ops/_grad/__init__.py +2 -1
- mindspore/ops/_grad/grad_array_ops.py +209 -152
- mindspore/ops/_grad/grad_base.py +55 -17
- mindspore/ops/_grad/grad_clip_ops.py +11 -3
- mindspore/ops/_grad/grad_comm_ops.py +58 -47
- mindspore/ops/_grad/grad_implementations.py +21 -61
- mindspore/ops/_grad/grad_inner_ops.py +48 -6
- mindspore/ops/_grad/grad_math_ops.py +306 -161
- mindspore/ops/_grad/grad_nn_ops.py +192 -181
- mindspore/ops/_grad/grad_other_ops.py +1 -1
- mindspore/ops/_grad/grad_quant_ops.py +5 -5
- mindspore/ops/_grad/grad_sequence_ops.py +296 -0
- mindspore/ops/_grad/grad_sparse.py +15 -9
- mindspore/ops/_grad_experimental/__init__.py +1 -0
- mindspore/ops/_grad_experimental/grad_array_ops.py +441 -55
- mindspore/ops/_grad_experimental/grad_image_ops.py +25 -7
- mindspore/ops/_grad_experimental/grad_inner_ops.py +3 -44
- mindspore/ops/_grad_experimental/grad_linalg_ops.py +16 -21
- mindspore/ops/_grad_experimental/grad_math_ops.py +979 -49
- mindspore/ops/_grad_experimental/grad_nn_ops.py +78 -8
- mindspore/ops/_grad_experimental/grad_scalar_ops.py +112 -0
- mindspore/ops/_grad_experimental/grad_sparse_ops.py +197 -13
- mindspore/ops/_op_impl/__init__.py +3 -3
- mindspore/ops/_op_impl/_custom_op/__init__.py +0 -1
- mindspore/ops/_op_impl/_custom_op/_basic.py +0 -1
- mindspore/ops/_op_impl/_custom_op/batch_matmul_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/batchnorm_fold.py +4 -2
- mindspore/ops/_op_impl/_custom_op/batchnorm_fold2.py +2 -2
- mindspore/ops/_op_impl/_custom_op/batchnorm_fold2_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/batchnorm_fold2_grad_reduce.py +5 -5
- mindspore/ops/_op_impl/_custom_op/batchnorm_fold_grad.py +3 -3
- mindspore/ops/_op_impl/_custom_op/cholesky_trsm_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/correction_mul.py +3 -3
- mindspore/ops/_op_impl/_custom_op/correction_mul_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/dsd_back_impl.py +4 -8
- mindspore/ops/_op_impl/_custom_op/dsd_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perchannel.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perchannel_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perchannel_grad_reduce.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perlayer.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perlayer_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perlayer_grad_reduce.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_quant_perchannel.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_quant_perchannel_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_quant_perlayer.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_quant_perlayer_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fused_abs_max1_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/img2col_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/matmul_cube_dense_left_impl.py +2 -2
- mindspore/ops/_op_impl/_custom_op/matmul_cube_dense_right_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/matmul_cube_fracz_left_cast_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/matmul_cube_fracz_right_mul_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/matmul_cube_impl.py +2 -2
- mindspore/ops/_op_impl/_custom_op/matmul_dds_grad_impl.py +0 -1
- mindspore/ops/_op_impl/_custom_op/matmul_dds_impl.py +0 -1
- mindspore/ops/_op_impl/_custom_op/matrix_combine_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/minmax_update_perchannel.py +2 -2
- mindspore/ops/_op_impl/_custom_op/minmax_update_perlayer.py +2 -2
- mindspore/ops/_op_impl/_custom_op/transpose02314_impl.py +1 -1
- mindspore/ops/_op_impl/aicpu/__init__.py +238 -3
- mindspore/ops/_op_impl/aicpu/abs.py +36 -0
- mindspore/ops/_op_impl/aicpu/adaptive_avg_pool_2d.py +34 -0
- mindspore/ops/_op_impl/aicpu/adaptive_avg_pool_2d_grad.py +34 -0
- mindspore/ops/_op_impl/aicpu/adaptive_avg_pool_3d.py +39 -0
- mindspore/ops/_op_impl/aicpu/adaptive_avg_pool_3d_grad.py +39 -0
- mindspore/ops/_op_impl/aicpu/adaptive_max_pool_2d_grad.py +37 -0
- mindspore/ops/_op_impl/aicpu/adaptive_max_pool_3d.py +42 -0
- mindspore/ops/_op_impl/aicpu/adaptive_max_pool_3d_grad.py +152 -0
- mindspore/ops/_op_impl/aicpu/add.py +43 -0
- mindspore/ops/_op_impl/aicpu/addcdiv.py +0 -32
- mindspore/ops/_op_impl/aicpu/addcmul.py +0 -84
- mindspore/ops/_op_impl/aicpu/affine_grid_grad.py +35 -0
- mindspore/ops/_op_impl/aicpu/arg_max.py +75 -0
- mindspore/ops/_op_impl/aicpu/arg_min.py +75 -0
- mindspore/ops/_op_impl/aicpu/argmin_with_value.py +43 -0
- mindspore/ops/_op_impl/aicpu/batch_matmul.py +43 -0
- mindspore/ops/_op_impl/aicpu/batch_norm_grad_grad.py +49 -0
- mindspore/ops/_op_impl/aicpu/bernoulli.py +48 -0
- mindspore/ops/_op_impl/aicpu/bessel_i0.py +31 -0
- mindspore/ops/_op_impl/aicpu/bias_add.py +44 -0
- mindspore/ops/_op_impl/aicpu/bias_add_grad.py +43 -0
- mindspore/ops/_op_impl/aicpu/bincount.py +33 -0
- mindspore/{nn/probability/infer/variational/__init__.py → ops/_op_impl/aicpu/cauchy.py} +17 -10
- mindspore/ops/_op_impl/aicpu/channel_shuffle.py +40 -0
- mindspore/ops/_op_impl/aicpu/cholesky.py +1 -1
- mindspore/ops/_op_impl/{cpu/bias_add.py → aicpu/choleskygrad.py} +9 -7
- mindspore/ops/_op_impl/aicpu/combined_non_max_suppression.py +42 -0
- mindspore/ops/_op_impl/aicpu/concat_offset.py +42 -0
- mindspore/ops/_op_impl/aicpu/concat_offset_v1.py +31 -0
- mindspore/ops/_op_impl/aicpu/conj.py +11 -0
- mindspore/ops/_op_impl/aicpu/crop_and_resize_grad_image.py +38 -0
- mindspore/ops/_op_impl/aicpu/cumulative_logsumexp.py +36 -0
- mindspore/ops/_op_impl/aicpu/deformable_offsets.py +38 -0
- mindspore/ops/_op_impl/aicpu/deformable_offsets_grad.py +2 -2
- mindspore/ops/_op_impl/aicpu/dense_to_sparse_set_operation.py +48 -0
- mindspore/ops/_op_impl/aicpu/diag.py +36 -0
- mindspore/ops/_op_impl/aicpu/diag_part.py +36 -0
- mindspore/ops/_op_impl/aicpu/diagonal.py +35 -0
- mindspore/ops/_op_impl/{cpu/bias_add_grad.py → aicpu/digamma.py} +9 -7
- mindspore/ops/_op_impl/aicpu/eig.py +35 -0
- mindspore/ops/_op_impl/aicpu/fft_with_size.py +41 -0
- mindspore/ops/_op_impl/aicpu/flatten.py +1 -0
- mindspore/ops/_op_impl/aicpu/fmax.py +36 -0
- mindspore/ops/_op_impl/aicpu/fmin.py +37 -0
- mindspore/ops/_op_impl/aicpu/fractional_max_pool3d_with_fixed_ksize.py +1 -1
- mindspore/ops/_op_impl/aicpu/fse_decode.py +43 -0
- mindspore/ops/_op_impl/aicpu/glu.py +33 -0
- mindspore/ops/_op_impl/aicpu/glu_grad.py +34 -0
- mindspore/ops/_op_impl/aicpu/greater.py +41 -0
- mindspore/ops/_op_impl/aicpu/greater_equal.py +41 -0
- mindspore/ops/_op_impl/aicpu/index_put.py +50 -0
- mindspore/ops/_op_impl/{tbe/scatter_add_ds.py → aicpu/inplace_index_add.py} +17 -21
- mindspore/ops/_op_impl/aicpu/instance_norm_v2.py +41 -0
- mindspore/ops/_op_impl/aicpu/instance_norm_v2_grad.py +44 -0
- mindspore/ops/_op_impl/aicpu/layer_norm_grad_grad.py +47 -0
- mindspore/ops/_op_impl/aicpu/less.py +41 -0
- mindspore/ops/_op_impl/aicpu/less_equal.py +41 -0
- mindspore/ops/_op_impl/aicpu/lgamma.py +32 -0
- mindspore/ops/_op_impl/aicpu/log_normal_reverse.py +33 -0
- mindspore/ops/_op_impl/aicpu/logit.py +33 -0
- mindspore/ops/_op_impl/aicpu/logit_grad.py +34 -0
- mindspore/ops/_op_impl/aicpu/masked_fill.py +42 -0
- mindspore/ops/_op_impl/aicpu/masked_scatter.py +39 -0
- mindspore/ops/_op_impl/aicpu/matmul.py +39 -0
- mindspore/ops/_op_impl/aicpu/matrix_logarithm.py +31 -0
- mindspore/ops/_op_impl/aicpu/matrix_power.py +32 -0
- mindspore/ops/_op_impl/aicpu/matrix_solve_ls.py +36 -0
- mindspore/ops/_op_impl/aicpu/matrix_triangular_solve.py +36 -0
- mindspore/ops/_op_impl/aicpu/mirror_pad.py +2 -0
- mindspore/ops/_op_impl/aicpu/mirror_pad_grad.py +0 -4
- mindspore/ops/_op_impl/aicpu/mul.py +3 -1
- mindspore/ops/_op_impl/aicpu/multinomial.py +14 -6
- mindspore/ops/_op_impl/aicpu/multinomial_with_replacement.py +35 -0
- mindspore/ops/_op_impl/aicpu/nan_to_num.py +34 -0
- mindspore/ops/_op_impl/aicpu/nllloss.py +38 -0
- mindspore/ops/_op_impl/aicpu/nllloss_grad.py +39 -0
- mindspore/ops/_op_impl/aicpu/ones_like.py +0 -2
- mindspore/ops/_op_impl/aicpu/polar.py +32 -0
- mindspore/ops/_op_impl/aicpu/polygamma.py +34 -0
- mindspore/ops/_op_impl/aicpu/qr.py +36 -0
- mindspore/ops/_op_impl/aicpu/quant_dtype_cast.py +40 -0
- mindspore/ops/_op_impl/aicpu/quantile.py +35 -0
- mindspore/ops/_op_impl/aicpu/ragged_tensor_to_sparse.py +73 -0
- mindspore/ops/_op_impl/aicpu/ragged_tensor_to_tensor.py +74 -0
- mindspore/ops/_op_impl/aicpu/random_shuffle.py +3 -0
- mindspore/ops/_op_impl/aicpu/randperm_v2.py +41 -0
- mindspore/ops/_op_impl/aicpu/range.py +36 -0
- mindspore/ops/_op_impl/aicpu/reciprocal.py +34 -0
- mindspore/ops/_op_impl/aicpu/reciprocal_grad.py +35 -0
- mindspore/ops/_op_impl/aicpu/reduce_sum.py +57 -0
- mindspore/ops/_op_impl/aicpu/resize_bicubic.py +2 -8
- mindspore/ops/_op_impl/aicpu/resize_bicubic_grad.py +1 -1
- mindspore/ops/_op_impl/aicpu/resize_v2.py +68 -0
- mindspore/ops/_op_impl/aicpu/resize_v2_grad.py +68 -0
- mindspore/ops/_op_impl/aicpu/scatter_elements.py +4 -0
- mindspore/ops/_op_impl/aicpu/scatter_nd_update.py +2 -0
- mindspore/ops/_op_impl/aicpu/search_sorted.py +12 -6
- mindspore/ops/_op_impl/aicpu/self_adjoint_eig.py +34 -0
- mindspore/ops/_op_impl/aicpu/sequence_add.py +34 -0
- mindspore/ops/_op_impl/aicpu/sequence_add_offset.py +34 -0
- mindspore/ops/_op_impl/aicpu/sequence_addn.py +38 -0
- mindspore/ops/_op_impl/aicpu/slice_grad.py +76 -0
- mindspore/ops/_op_impl/aicpu/smooth_l1_loss.py +35 -0
- mindspore/ops/_op_impl/aicpu/smooth_l1_loss_grad.py +37 -0
- mindspore/ops/_op_impl/aicpu/sort.py +39 -0
- mindspore/ops/_op_impl/aicpu/sparse_apply_adagrad_da.py +0 -24
- mindspore/ops/_op_impl/aicpu/sparse_cross.py +42 -0
- mindspore/ops/_op_impl/aicpu/sparse_fill_empty_rows.py +63 -0
- mindspore/ops/_op_impl/aicpu/sparse_fill_empty_rows_grad.py +45 -0
- mindspore/ops/_op_impl/aicpu/sparse_matrix_mat_mul.py +56 -0
- mindspore/ops/_op_impl/{tbe/slice_ds.py → aicpu/sparse_segment_sum.py} +16 -24
- mindspore/ops/_op_impl/aicpu/sparse_segment_sum_with_num_segments.py +68 -0
- mindspore/ops/_op_impl/aicpu/sparse_slice.py +63 -0
- mindspore/ops/_op_impl/aicpu/sparse_slice_grad.py +61 -0
- mindspore/ops/_op_impl/aicpu/squared_difference.py +2 -0
- mindspore/ops/_op_impl/aicpu/strided_slice_v2.py +93 -0
- mindspore/ops/_op_impl/aicpu/strided_slice_v2_grad.py +66 -0
- mindspore/ops/_op_impl/aicpu/tensor_scatter_update.py +59 -0
- mindspore/ops/_op_impl/{tbe/gather_v2.py → aicpu/tile.py} +24 -24
- mindspore/ops/_op_impl/aicpu/tridiagonal_solve.py +35 -0
- mindspore/ops/_op_impl/aicpu/tril_indices.py +34 -0
- mindspore/ops/_op_impl/aicpu/triu_indices.py +34 -0
- mindspore/ops/_op_impl/aicpu/uniform.py +34 -0
- mindspore/ops/_op_impl/aicpu/uniform_candidate_sampler.py +1 -0
- mindspore/ops/_op_impl/aicpu/unique_consecutive.py +10 -2
- mindspore/ops/_op_impl/cpu/__init__.py +1 -2
- mindspore/ops/_op_impl/cpu/dynamic_shape.py +5 -1
- mindspore/ops/_op_impl/cpu/maximum_grad.py +2 -0
- mindspore/{compression/common/__init__.py → ops/_op_impl/cpu/pyexecute.py} +13 -8
- mindspore/ops/_op_impl/cpu/reduce_sum.py +8 -0
- mindspore/ops/_op_impl/cpu/sparse_slice.py +62 -0
- mindspore/ops/_op_impl/cpu/sparse_slice_grad.py +60 -0
- mindspore/ops/_op_impl/cpu/tensor_shape.py +5 -1
- mindspore/ops/_op_impl/tbe/__init__.py +27 -608
- mindspore/ops/_op_impl/tbe/addcdiv_ds.py +42 -0
- mindspore/ops/_op_impl/tbe/addcmul_ds.py +44 -0
- mindspore/ops/_op_impl/tbe/assign_add_ds.py +1 -0
- mindspore/ops/_op_impl/tbe/atomic_addr_clean.py +1 -1
- mindspore/ops/_op_impl/tbe/avg_pool_3d_grad.py +1 -1
- mindspore/ops/_op_impl/tbe/basic_lstm_cell_c_state_grad_v2.py +0 -1
- mindspore/ops/_op_impl/tbe/batch_to_space.py +1 -1
- mindspore/ops/_op_impl/tbe/batch_to_space_nd.py +1 -1
- mindspore/ops/_op_impl/tbe/batch_to_space_nd_v2.py +41 -0
- mindspore/ops/_op_impl/tbe/bce_with_logits_loss.py +1 -0
- mindspore/ops/_op_impl/tbe/bias_add_grad.py +2 -0
- mindspore/ops/_op_impl/tbe/bn_infer_grad.py +4 -2
- mindspore/ops/_op_impl/tbe/bn_infer_grad_ds.py +40 -0
- mindspore/ops/_op_impl/tbe/bn_training_update.py +0 -1
- mindspore/ops/_op_impl/tbe/bn_training_update_ds.py +0 -1
- mindspore/ops/_op_impl/tbe/broadcast_to_ds.py +6 -4
- mindspore/ops/_op_impl/tbe/cast.py +0 -2
- mindspore/ops/_op_impl/tbe/cast_ds.py +3 -3
- mindspore/ops/_op_impl/tbe/ctc_loss_v2.py +0 -2
- mindspore/ops/_op_impl/tbe/ctc_loss_v2_grad.py +0 -2
- mindspore/ops/_op_impl/tbe/data_format_dim_map_ds.py +1 -0
- mindspore/ops/_op_impl/tbe/deformable_offsets.py +1 -0
- mindspore/ops/_op_impl/tbe/depthwise_conv2d.py +1 -1
- mindspore/ops/_op_impl/tbe/dynamic_atomic_addr_clean.py +1 -1
- mindspore/ops/_op_impl/tbe/gather_nd.py +1 -0
- mindspore/ops/_op_impl/tbe/greater.py +2 -0
- mindspore/ops/_op_impl/tbe/{index_add.py → inplace_index_add.py} +3 -6
- mindspore/ops/_op_impl/tbe/layer_norm_beta_gamma_backprop_v2.py +0 -1
- mindspore/ops/_op_impl/tbe/npu_clear_float_status_v2.py +35 -0
- mindspore/ops/_op_impl/tbe/npu_get_float_status_v2.py +35 -0
- mindspore/ops/_op_impl/tbe/one_hot_ds.py +0 -6
- mindspore/ops/_op_impl/tbe/{greater_ds.py → reduce_all_ds.py} +13 -16
- mindspore/ops/_op_impl/tbe/reduce_any_ds.py +39 -0
- mindspore/ops/_op_impl/tbe/roi_align_ds.py +44 -0
- mindspore/ops/_op_impl/tbe/roi_align_grad_ds.py +44 -0
- mindspore/ops/_op_impl/tbe/scatter_add.py +2 -0
- mindspore/ops/_op_impl/tbe/scatter_nd_add.py +2 -2
- mindspore/ops/_op_impl/tbe/slice.py +26 -15
- mindspore/ops/_op_impl/tbe/space_to_batch.py +1 -1
- mindspore/ops/_op_impl/tbe/space_to_batch_nd.py +1 -1
- mindspore/ops/_op_impl/tbe/strided_slice_grad_d.py +1 -0
- mindspore/ops/_op_impl/tbe/trans_data_ds.py +15 -5
- mindspore/ops/_op_impl/tbe/unsorted_segment_sum.py +1 -1
- mindspore/ops/_op_impl/tbe/unsorted_segment_sum_ds.py +2 -0
- mindspore/ops/_primitive_cache.py +3 -2
- mindspore/ops/_register_for_op.py +11 -0
- mindspore/ops/_utils/__init__.py +1 -1
- mindspore/ops/_utils/utils.py +20 -41
- mindspore/ops/_vmap/__init__.py +2 -2
- mindspore/ops/_vmap/vmap_array_ops.py +170 -78
- mindspore/ops/_vmap/vmap_base.py +24 -10
- mindspore/ops/_vmap/vmap_convolution_ops.py +7 -10
- mindspore/ops/_vmap/vmap_grad_math_ops.py +4 -4
- mindspore/ops/_vmap/vmap_grad_nn_ops.py +41 -9
- mindspore/ops/_vmap/vmap_image_ops.py +52 -0
- mindspore/ops/_vmap/vmap_math_ops.py +77 -6
- mindspore/ops/_vmap/vmap_nn_ops.py +78 -29
- mindspore/ops/_vmap/vmap_other_ops.py +3 -1
- mindspore/ops/_vmap/vmap_random_ops.py +55 -3
- mindspore/ops/_vmap/vmap_sparse_ops.py +1 -0
- mindspore/ops/bprop_mindir/AdaptiveAvgPool2D_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/AdaptiveMaxPool2D_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/ApproximateEqual_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/Argmax_bprop.mindir +13 -12
- mindspore/ops/bprop_mindir/Argmin_bprop.mindir +14 -13
- mindspore/ops/bprop_mindir/AssignSub_bprop.mindir +17 -18
- mindspore/ops/bprop_mindir/Assign_bprop.mindir +16 -16
- mindspore/ops/bprop_mindir/AvgPool3D_bprop.mindir +150 -0
- mindspore/ops/bprop_mindir/AvgPool_bprop.mindir +66 -0
- mindspore/ops/bprop_mindir/BCEWithLogitsLoss_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/BNTrainingReduce_bprop.mindir +13 -12
- mindspore/ops/bprop_mindir/BatchNormGrad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/BatchToSpaceND_bprop.mindir +28 -0
- mindspore/ops/bprop_mindir/BiasAddGrad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/BinaryCrossEntropy_bprop.mindir +33 -0
- mindspore/ops/bprop_mindir/BroadcastTo_bprop.mindir +306 -0
- mindspore/ops/bprop_mindir/Broadcast_bprop.mindir +12 -8
- mindspore/ops/bprop_mindir/CTCLoss_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Concat_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Conv2DBackpropFilter_bprop.mindir +240 -0
- mindspore/ops/bprop_mindir/Conv2DBackpropInput_bprop.mindir +247 -0
- mindspore/ops/bprop_mindir/Conv2DTranspose_bprop.mindir +247 -0
- mindspore/ops/bprop_mindir/Conv3DTranspose_bprop.mindir +315 -0
- mindspore/ops/bprop_mindir/Conv3D_bprop.mindir +278 -0
- mindspore/ops/bprop_mindir/DType_bprop.mindir +12 -12
- mindspore/ops/bprop_mindir/DeformableOffsets_bprop.mindir +58 -0
- mindspore/ops/bprop_mindir/Depend_bprop.mindir +12 -13
- mindspore/ops/bprop_mindir/DepthToSpace_bprop.mindir +23 -0
- mindspore/ops/bprop_mindir/DepthwiseConv2dNative_bprop.mindir +138 -0
- mindspore/ops/bprop_mindir/DiagPart_bprop.mindir +15 -0
- mindspore/ops/bprop_mindir/Dropout2D_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Dropout3D_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/DropoutDoMask_bprop.mindir +22 -24
- mindspore/ops/bprop_mindir/DropoutGenMask_bprop.mindir +16 -14
- mindspore/ops/bprop_mindir/DropoutGrad_bprop.mindir +27 -0
- mindspore/ops/bprop_mindir/Dropout_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/DynamicGRUV2_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/DynamicRNN_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/DynamicShape_bprop.mindir +12 -12
- mindspore/ops/bprop_mindir/Elu_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/EmbeddingLookup_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Equal_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/ExpandDims_bprop.mindir +58 -0
- mindspore/ops/bprop_mindir/FastGeLU_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/Flatten_bprop.mindir +54 -0
- mindspore/ops/bprop_mindir/FloorDiv_bprop.mindir +18 -15
- mindspore/ops/bprop_mindir/GatherD_bprop.mindir +26 -0
- mindspore/ops/bprop_mindir/GatherNd_bprop.mindir +57 -0
- mindspore/ops/bprop_mindir/Gather_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/GreaterEqual_bprop.mindir +17 -18
- mindspore/ops/bprop_mindir/Greater_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/HSigmoid_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/HSwish_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/IOU_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/InstanceNorm_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/IsFinite_bprop.mindir +13 -12
- mindspore/ops/bprop_mindir/IsInf_bprop.mindir +13 -10
- mindspore/ops/bprop_mindir/IsNan_bprop.mindir +14 -11
- mindspore/ops/bprop_mindir/KLDivLoss_bprop.mindir +126 -0
- mindspore/ops/bprop_mindir/L2Loss_bprop.mindir +15 -0
- mindspore/ops/bprop_mindir/L2Normalize_bprop.mindir +30 -0
- mindspore/ops/bprop_mindir/LRN_bprop.mindir +43 -0
- mindspore/ops/bprop_mindir/LayerNormGrad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/LessEqual_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/Less_bprop.mindir +17 -18
- mindspore/ops/bprop_mindir/LinSpace_bprop.mindir +22 -19
- mindspore/ops/bprop_mindir/Load_bprop.mindir +12 -13
- mindspore/ops/bprop_mindir/LogSoftmax_bprop.mindir +23 -0
- mindspore/ops/bprop_mindir/LogicalAnd_bprop.mindir +17 -18
- mindspore/ops/bprop_mindir/LogicalNot_bprop.mindir +14 -13
- mindspore/ops/bprop_mindir/MaskedSelect_bprop.mindir +21 -0
- mindspore/ops/bprop_mindir/MaxPool3DGradGrad_bprop.mindir +74 -0
- mindspore/ops/bprop_mindir/MaxPool3DGrad_bprop.mindir +74 -0
- mindspore/ops/bprop_mindir/MaxPool3D_bprop.mindir +75 -0
- mindspore/ops/bprop_mindir/MaxPoolGradGrad_bprop.mindir +65 -0
- mindspore/ops/bprop_mindir/MaxPoolWithArgmax_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Maximum_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Minimum_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/MirrorPad_bprop.mindir +27 -0
- mindspore/ops/bprop_mindir/Mish_bprop.mindir +35 -0
- mindspore/ops/bprop_mindir/MulNoNan_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/NLLLoss_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/NonZero_bprop.mindir +14 -0
- mindspore/ops/bprop_mindir/NotEqual_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/OneHot_bprop.mindir +25 -23
- mindspore/ops/bprop_mindir/OnesLike_bprop.mindir +13 -13
- mindspore/ops/bprop_mindir/PReLU_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Pad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Padding_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/RNNTLoss_bprop.mindir +29 -0
- mindspore/ops/bprop_mindir/ROIAlign_bprop.mindir +82 -0
- mindspore/ops/bprop_mindir/Range_bprop.mindir +21 -19
- mindspore/ops/bprop_mindir/Rank_bprop.mindir +11 -11
- mindspore/ops/bprop_mindir/ReLU6_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/ReLUV2_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/ReduceAll_bprop.mindir +18 -17
- mindspore/ops/bprop_mindir/ReduceAny_bprop.mindir +18 -17
- mindspore/ops/bprop_mindir/ReluGrad_bprop.mindir +19 -23
- mindspore/ops/bprop_mindir/Reshape_bprop.mindir +60 -0
- mindspore/ops/bprop_mindir/ResizeBilinear_bprop.mindir +29 -0
- mindspore/ops/bprop_mindir/ResizeNearestNeighbor_bprop.mindir +89 -0
- mindspore/ops/bprop_mindir/ReverseSequence_bprop.mindir +52 -0
- mindspore/ops/bprop_mindir/ReverseV2_bprop.mindir +22 -0
- mindspore/ops/bprop_mindir/Round_bprop.mindir +14 -13
- mindspore/ops/bprop_mindir/ScatterMax_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/ScatterMin_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/ScatterNdUpdate_bprop.mindir +22 -0
- mindspore/ops/bprop_mindir/ScatterNd_bprop.mindir +24 -0
- mindspore/ops/bprop_mindir/ScatterNonAliasingAdd_bprop.mindir +22 -0
- mindspore/ops/bprop_mindir/ScatterUpdate_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/SeLU_bprop.mindir +21 -0
- mindspore/ops/bprop_mindir/Select_bprop.mindir +30 -34
- mindspore/ops/bprop_mindir/Shape_bprop.mindir +12 -12
- mindspore/ops/bprop_mindir/SigmoidCrossEntropyWithLogits_bprop.mindir +21 -0
- mindspore/ops/bprop_mindir/SigmoidGrad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Sigmoid_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/Sign_bprop.mindir +13 -12
- mindspore/ops/bprop_mindir/Slice_bprop.mindir +26 -0
- mindspore/ops/bprop_mindir/SmoothL1Loss_bprop.mindir +36 -0
- mindspore/ops/bprop_mindir/SoftmaxCrossEntropyWithLogits_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Softplus_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/Softsign_bprop.mindir +33 -0
- mindspore/ops/bprop_mindir/Sort_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/SpaceToBatchND_bprop.mindir +28 -0
- mindspore/ops/bprop_mindir/SpaceToDepth_bprop.mindir +23 -0
- mindspore/ops/bprop_mindir/SparseGatherV2_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/SparseSoftmaxCrossEntropyWithLogits_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Split_bprop.mindir +22 -0
- mindspore/ops/bprop_mindir/Squeeze_bprop.mindir +54 -0
- mindspore/ops/bprop_mindir/StridedSliceGrad_bprop.mindir +95 -0
- mindspore/ops/bprop_mindir/StridedSlice_bprop.mindir +98 -0
- mindspore/ops/bprop_mindir/Switch_bprop.mindir +28 -32
- mindspore/ops/bprop_mindir/TanhGrad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Tanh_bprop.mindir +66 -0
- mindspore/ops/bprop_mindir/TensorScatterAdd_bprop.mindir +22 -0
- mindspore/ops/bprop_mindir/TensorScatterUpdate_bprop.mindir +29 -0
- mindspore/ops/bprop_mindir/TensorShape_bprop.mindir +14 -0
- mindspore/ops/bprop_mindir/Tile_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/TopK_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/TransShape_bprop.mindir +23 -0
- mindspore/ops/bprop_mindir/TruncateDiv_bprop.mindir +18 -15
- mindspore/ops/bprop_mindir/TupleGetItem_bprop.mindir +11 -13
- mindspore/ops/bprop_mindir/Unique_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/Unstack_bprop.mindir +22 -0
- mindspore/ops/bprop_mindir/UpsampleNearest3D_bprop.mindir +32 -0
- mindspore/ops/bprop_mindir/UpsampleTrilinear3D_bprop.mindir +38 -0
- mindspore/ops/bprop_mindir/ZerosLike_bprop.mindir +13 -12
- mindspore/ops/bprop_mindir/__init__.py +1 -4
- mindspore/ops/bprop_mindir/generate_mindir.py +32 -20
- mindspore/ops/composite/__init__.py +12 -13
- mindspore/ops/composite/base.py +261 -254
- mindspore/ops/composite/env_ops.py +41 -0
- mindspore/ops/composite/math_ops.py +197 -156
- mindspore/ops/composite/multitype_ops/_compile_utils.py +428 -176
- mindspore/ops/composite/multitype_ops/_constexpr_utils.py +188 -87
- mindspore/ops/composite/multitype_ops/add_impl.py +23 -1
- mindspore/ops/composite/multitype_ops/div_impl.py +3 -3
- mindspore/ops/composite/multitype_ops/equal_impl.py +1 -0
- mindspore/ops/composite/multitype_ops/floordiv_impl.py +1 -1
- mindspore/ops/composite/multitype_ops/getitem_impl.py +52 -5
- mindspore/ops/composite/multitype_ops/greater_equal_impl.py +31 -0
- mindspore/ops/composite/multitype_ops/greater_impl.py +31 -0
- mindspore/ops/composite/multitype_ops/in_impl.py +15 -3
- mindspore/ops/composite/multitype_ops/less_equal_impl.py +33 -2
- mindspore/ops/composite/multitype_ops/less_impl.py +33 -0
- mindspore/ops/composite/multitype_ops/logical_and_impl.py +2 -2
- mindspore/ops/composite/multitype_ops/logical_or_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/mod_impl.py +1 -1
- mindspore/ops/composite/multitype_ops/mul_impl.py +21 -7
- mindspore/ops/composite/multitype_ops/not_in_impl.py +15 -3
- mindspore/ops/composite/multitype_ops/ones_like_impl.py +2 -4
- mindspore/ops/composite/multitype_ops/pow_impl.py +1 -0
- mindspore/ops/composite/multitype_ops/setitem_impl.py +62 -70
- mindspore/ops/composite/multitype_ops/sub_impl.py +3 -3
- mindspore/ops/composite/multitype_ops/zeros_like_impl.py +41 -4
- mindspore/ops/function/__init__.py +323 -8
- mindspore/ops/function/array_func.py +3511 -780
- mindspore/ops/function/clip_func.py +329 -0
- mindspore/ops/function/debug_func.py +6 -6
- mindspore/ops/function/grad/__init__.py +5 -1
- mindspore/ops/function/grad/grad_func.py +736 -65
- mindspore/ops/function/image_func.py +270 -0
- mindspore/ops/function/linalg_func.py +268 -8
- mindspore/ops/function/math_func.py +8032 -3164
- mindspore/ops/function/nn_func.py +5619 -1855
- mindspore/ops/function/other_func.py +115 -0
- mindspore/ops/function/parameter_func.py +11 -10
- mindspore/ops/function/random_func.py +939 -77
- mindspore/ops/function/sparse_func.py +249 -84
- mindspore/ops/function/sparse_unary_func.py +2303 -0
- mindspore/ops/function/spectral_func.py +146 -0
- mindspore/ops/function/vmap_func.py +114 -0
- mindspore/ops/functional.py +182 -254
- mindspore/ops/op_info_register.py +79 -34
- mindspore/ops/operations/__init__.py +210 -118
- mindspore/ops/operations/_csr_ops.py +7 -7
- mindspore/ops/operations/_embedding_cache_ops.py +25 -15
- mindspore/ops/operations/_grad_ops.py +447 -322
- mindspore/ops/operations/_inner_ops.py +547 -176
- mindspore/ops/operations/_map_tensor_ops.py +112 -0
- mindspore/ops/operations/_ms_kernel.py +29 -27
- mindspore/ops/operations/_ocr_ops.py +11 -11
- mindspore/ops/operations/_opaque_predicate_registry.py +41 -0
- mindspore/ops/operations/_quant_ops.py +186 -101
- mindspore/ops/operations/_rl_inner_ops.py +122 -61
- mindspore/ops/operations/_scalar_ops.py +466 -0
- mindspore/ops/operations/_sequence_ops.py +1047 -0
- mindspore/ops/operations/_tensor_array.py +10 -11
- mindspore/ops/operations/_thor_ops.py +4 -4
- mindspore/ops/operations/array_ops.py +1428 -1226
- mindspore/ops/operations/comm_ops.py +180 -117
- mindspore/ops/operations/control_ops.py +4 -2
- mindspore/ops/operations/custom_ops.py +185 -98
- mindspore/ops/operations/debug_ops.py +92 -54
- mindspore/ops/operations/image_ops.py +406 -211
- mindspore/ops/operations/inner_ops.py +42 -53
- mindspore/ops/operations/linalg_ops.py +32 -29
- mindspore/ops/operations/math_ops.py +2076 -897
- mindspore/ops/operations/nn_ops.py +1282 -1252
- mindspore/ops/operations/other_ops.py +124 -278
- mindspore/ops/operations/random_ops.py +345 -178
- mindspore/ops/operations/rl_ops.py +8 -9
- mindspore/ops/operations/sparse_ops.py +502 -157
- mindspore/ops/operations/spectral_ops.py +107 -0
- mindspore/ops/primitive.py +192 -15
- mindspore/ops/vm_impl_registry.py +23 -2
- mindspore/parallel/__init__.py +6 -1
- mindspore/parallel/_auto_parallel_context.py +199 -92
- mindspore/parallel/_cell_wrapper.py +4 -2
- mindspore/parallel/_cost_model_context.py +3 -0
- mindspore/parallel/_dp_allreduce_fusion.py +2 -1
- mindspore/parallel/_offload_context.py +185 -0
- mindspore/parallel/_parallel_serialization.py +167 -28
- mindspore/parallel/_ps_context.py +9 -5
- mindspore/parallel/_recovery_context.py +1 -1
- mindspore/parallel/_tensor.py +9 -1
- mindspore/{nn/transformer → parallel/_transformer}/__init__.py +6 -6
- mindspore/{nn/transformer → parallel/_transformer}/layers.py +59 -37
- mindspore/{nn/transformer → parallel/_transformer}/loss.py +4 -7
- mindspore/{nn/transformer → parallel/_transformer}/moe.py +160 -35
- mindspore/{nn/transformer → parallel/_transformer}/op_parallel_config.py +3 -3
- mindspore/{nn/transformer → parallel/_transformer}/transformer.py +235 -196
- mindspore/parallel/_utils.py +47 -7
- mindspore/parallel/algo_parameter_config.py +5 -1
- mindspore/parallel/checkpoint_transform.py +329 -0
- mindspore/parallel/shard.py +229 -0
- mindspore/profiler/__init__.py +2 -1
- mindspore/profiler/common/util.py +4 -3
- mindspore/profiler/common/validator/validate_path.py +2 -2
- mindspore/profiler/envprofiling.py +249 -0
- mindspore/profiler/parser/aicpu_data_parser.py +38 -39
- mindspore/profiler/parser/ascend_timeline_generator.py +497 -0
- mindspore/profiler/parser/base_timeline_generator.py +471 -0
- mindspore/profiler/parser/cpu_gpu_timeline_generator.py +684 -0
- mindspore/profiler/parser/framework_parser.py +42 -16
- mindspore/profiler/parser/hccl_parser.py +158 -158
- mindspore/profiler/parser/hwts_log_parser.py +7 -6
- mindspore/profiler/parser/integrator.py +18 -1579
- mindspore/profiler/parser/minddata_analyzer.py +8 -8
- mindspore/profiler/parser/msadvisor_analyzer.py +14 -27
- mindspore/profiler/parser/msadvisor_parser.py +2 -4
- mindspore/profiler/parser/optime_parser.py +17 -18
- mindspore/profiler/parser/profiler_info.py +108 -0
- mindspore/profiler/parser/step_trace_parser.py +1 -1
- mindspore/profiler/profiling.py +396 -194
- mindspore/rewrite/__init__.py +6 -2
- mindspore/rewrite/api/node.py +51 -110
- mindspore/rewrite/api/node_type.py +10 -6
- mindspore/rewrite/api/pattern_engine.py +51 -7
- mindspore/rewrite/api/scoped_value.py +64 -53
- mindspore/rewrite/api/symbol_tree.py +108 -61
- mindspore/rewrite/api/tree_node_helper.py +2 -3
- mindspore/{compression/quant/__init__.py → rewrite/ast_creator_register.py} +20 -11
- mindspore/rewrite/ast_helpers/__init__.py +6 -3
- mindspore/rewrite/ast_helpers/ast_creator.py +115 -0
- mindspore/rewrite/ast_helpers/ast_finder.py +99 -1
- mindspore/rewrite/ast_helpers/ast_modifier.py +17 -4
- mindspore/rewrite/ast_helpers/ast_replacer.py +1 -1
- mindspore/rewrite/ast_transformers/__init__.py +0 -1
- mindspore/rewrite/ast_transformers/flatten_recursive_stmt.py +46 -5
- mindspore/rewrite/ast_transformers/remove_return_out_of_if.py +6 -3
- mindspore/rewrite/common/__init__.py +2 -0
- mindspore/rewrite/common/event.py +1 -1
- mindspore/rewrite/common/observable.py +1 -1
- mindspore/rewrite/common/observer.py +1 -1
- mindspore/rewrite/common/rewrite_elog.py +35 -0
- mindspore/rewrite/namer.py +2 -2
- mindspore/rewrite/namespace.py +14 -4
- mindspore/rewrite/node.py +161 -13
- mindspore/rewrite/parser.py +0 -1
- mindspore/rewrite/parser_register.py +0 -1
- mindspore/rewrite/parsers/arguments_parser.py +3 -2
- mindspore/rewrite/parsers/assign_parser.py +267 -67
- mindspore/rewrite/parsers/attribute_parser.py +56 -0
- mindspore/rewrite/parsers/class_def_parser.py +191 -108
- mindspore/rewrite/parsers/constant_parser.py +101 -0
- mindspore/rewrite/parsers/container_parser.py +88 -0
- mindspore/rewrite/parsers/for_parser.py +28 -15
- mindspore/rewrite/parsers/function_def_parser.py +21 -5
- mindspore/rewrite/parsers/if_parser.py +11 -28
- mindspore/rewrite/parsers/module_parser.py +9 -6
- mindspore/rewrite/parsers/return_parser.py +3 -2
- mindspore/rewrite/sparsify/__init__.py +0 -0
- mindspore/rewrite/sparsify/sparse_transformer.py +448 -0
- mindspore/rewrite/sparsify/sparsify.py +109 -0
- mindspore/rewrite/sparsify/utils.py +173 -0
- mindspore/rewrite/symbol_tree.py +322 -109
- mindspore/rewrite/symbol_tree_builder.py +45 -8
- mindspore/rewrite/symbol_tree_dumper.py +0 -1
- mindspore/rewrite/topological_manager.py +1 -2
- mindspore/run_check/_check_version.py +209 -112
- mindspore/run_check/run_check.py +2 -1
- mindspore/scipy/linalg.py +13 -117
- mindspore/scipy/ops.py +5 -71
- mindspore/scipy/ops_grad.py +1 -25
- mindspore/scipy/ops_wrapper.py +1 -1
- mindspore/scipy/optimize/_bfgs.py +1 -1
- mindspore/scipy/optimize/_lagrange.py +200 -0
- mindspore/scipy/optimize/line_search.py +3 -2
- mindspore/scipy/optimize/minimize.py +43 -6
- mindspore/scipy/sparse/__init__.py +2 -2
- mindspore/scipy/sparse/linalg.py +5 -465
- mindspore/scipy/utils.py +2 -1
- mindspore/scipy/utils_const.py +7 -1
- mindspore/train/__init__.py +6 -4
- mindspore/train/_utils.py +28 -5
- mindspore/train/amp.py +321 -50
- mindspore/train/callback/__init__.py +3 -1
- mindspore/train/callback/_backup_and_restore.py +120 -0
- mindspore/train/callback/_callback.py +8 -8
- mindspore/train/callback/_checkpoint.py +12 -9
- mindspore/train/callback/_early_stop.py +13 -7
- mindspore/train/callback/_history.py +8 -8
- mindspore/train/callback/_lambda_callback.py +6 -6
- mindspore/train/callback/_landscape.py +36 -38
- mindspore/train/callback/_loss_monitor.py +12 -6
- mindspore/train/callback/_lr_scheduler_callback.py +2 -4
- mindspore/train/callback/_on_request_exit.py +212 -0
- mindspore/train/callback/_reduce_lr_on_plateau.py +13 -7
- mindspore/train/callback/_summary_collector.py +27 -19
- mindspore/train/callback/_time_monitor.py +13 -7
- mindspore/train/checkpoint_pb2.py +68 -8
- mindspore/train/data_sink.py +122 -33
- mindspore/train/dataset_helper.py +28 -87
- mindspore/train/loss_scale_manager.py +4 -7
- mindspore/{nn → train}/metrics/__init__.py +20 -20
- mindspore/{nn → train}/metrics/accuracy.py +12 -10
- mindspore/{nn → train}/metrics/auc.py +4 -4
- mindspore/{nn → train}/metrics/bleu_score.py +4 -4
- mindspore/{nn → train}/metrics/confusion_matrix.py +10 -8
- mindspore/{nn → train}/metrics/cosine_similarity.py +4 -4
- mindspore/{nn → train}/metrics/dice.py +6 -5
- mindspore/{nn → train}/metrics/error.py +7 -5
- mindspore/{nn → train}/metrics/fbeta.py +9 -7
- mindspore/{nn → train}/metrics/hausdorff_distance.py +8 -6
- mindspore/{nn → train}/metrics/loss.py +4 -3
- mindspore/{nn → train}/metrics/mean_surface_distance.py +6 -5
- mindspore/{nn → train}/metrics/metric.py +6 -5
- mindspore/{nn → train}/metrics/occlusion_sensitivity.py +4 -3
- mindspore/{nn → train}/metrics/perplexity.py +5 -4
- mindspore/{nn → train}/metrics/precision.py +5 -4
- mindspore/{nn → train}/metrics/recall.py +5 -4
- mindspore/{nn → train}/metrics/roc.py +7 -6
- mindspore/{nn → train}/metrics/root_mean_square_surface_distance.py +6 -5
- mindspore/{nn → train}/metrics/topk.py +7 -5
- mindspore/train/mind_ir_pb2.py +339 -32
- mindspore/train/model.py +113 -84
- mindspore/train/serialization.py +547 -167
- mindspore/train/summary/_summary_adapter.py +1 -1
- mindspore/train/summary/summary_record.py +43 -12
- mindspore/train/train_thor/convert_utils.py +7 -1
- mindspore/train/train_thor/dataset_helper.py +3 -3
- mindspore/train/train_thor/model_thor.py +0 -4
- mindspore/version.py +1 -1
- {mindspore-1.10.0.dist-info → mindspore-2.0.0rc1.dist-info}/METADATA +4 -3
- {mindspore-1.10.0.dist-info → mindspore-2.0.0rc1.dist-info}/RECORD +899 -675
- mindspore/compression/common/constant.py +0 -124
- mindspore/compression/export/__init__.py +0 -19
- mindspore/compression/export/quant_export.py +0 -514
- mindspore/compression/quant/qat.py +0 -636
- mindspore/compression/quant/quant_utils.py +0 -462
- mindspore/compression/quant/quantizer.py +0 -68
- mindspore/nn/layer/quant.py +0 -1868
- mindspore/nn/layer/rnn_utils.py +0 -90
- mindspore/nn/probability/dpn/__init__.py +0 -22
- mindspore/nn/probability/dpn/vae/__init__.py +0 -25
- mindspore/nn/probability/dpn/vae/cvae.py +0 -138
- mindspore/nn/probability/dpn/vae/vae.py +0 -122
- mindspore/nn/probability/infer/__init__.py +0 -22
- mindspore/nn/probability/infer/variational/elbo.py +0 -70
- mindspore/nn/probability/infer/variational/svi.py +0 -84
- mindspore/nn/probability/toolbox/__init__.py +0 -22
- mindspore/nn/probability/toolbox/anomaly_detection.py +0 -99
- mindspore/nn/probability/toolbox/uncertainty_evaluation.py +0 -363
- mindspore/nn/probability/transforms/__init__.py +0 -22
- mindspore/nn/probability/transforms/transform_bnn.py +0 -262
- mindspore/nn/probability/zhusuan/__init__.py +0 -18
- mindspore/nn/probability/zhusuan/framework/__init__.py +0 -18
- mindspore/nn/probability/zhusuan/framework/bn.py +0 -95
- mindspore/nn/probability/zhusuan/variational/__init__.py +0 -18
- mindspore/nn/probability/zhusuan/variational/elbo.py +0 -46
- mindspore/ops/_op_impl/tbe/bias_add_grad_ds.py +0 -52
- mindspore/ops/_op_impl/tbe/scatter_nd_add_ds.py +0 -43
- mindspore/ops/bprop_mindir/AssignAdd_bprop.mindir +0 -20
- mindspore/ops/bprop_mindir/Identity_bprop.mindir +0 -9
- mindspore/ops/bprop_mindir/LogicalOr_bprop.mindir +0 -20
- mindspore/ops/bprop_mindir/ReLU_bprop.mindir +0 -16
- mindspore/ops/bprop_mindir/UpdateState_bprop.mindir +0 -17
- mindspore/ops/bprop_mindir/stop_gradient_bprop.mindir +0 -12
- mindspore/ops/composite/array_ops.py +0 -210
- mindspore/ops/composite/clip_ops.py +0 -238
- mindspore/ops/composite/random_ops.py +0 -426
- mindspore/ops/composite/vmap_ops.py +0 -38
- mindspore/ops/operations/sponge_ops.py +0 -3531
- mindspore/ops/operations/sponge_update_ops.py +0 -2546
- mindspore/parallel/nn/__init__.py +0 -42
- mindspore/parallel/nn/loss.py +0 -22
- mindspore/parallel/nn/moe.py +0 -21
- mindspore/parallel/nn/op_parallel_config.py +0 -22
- mindspore/parallel/nn/transformer.py +0 -31
- mindspore/run_check/_check_deps_version.py +0 -84
- {mindspore-1.10.0.dist-info → mindspore-2.0.0rc1.dist-info}/WHEEL +0 -0
- {mindspore-1.10.0.dist-info → mindspore-2.0.0rc1.dist-info}/entry_points.txt +0 -0
- {mindspore-1.10.0.dist-info → mindspore-2.0.0rc1.dist-info}/top_level.txt +0 -0
|
@@ -18,12 +18,21 @@ from __future__ import absolute_import
|
|
|
18
18
|
from functools import partial
|
|
19
19
|
|
|
20
20
|
import mindspore.context as context
|
|
21
|
-
from mindspore
|
|
22
|
-
from mindspore._checkparam import Rel
|
|
21
|
+
from mindspore import _checkparam as validator
|
|
23
22
|
from mindspore.ops.primitive import Primitive, PrimitiveWithInfer, prim_attr_register
|
|
24
23
|
from mindspore.common import dtype as mstype
|
|
24
|
+
from mindspore.common.dtype import QuantDtype
|
|
25
25
|
|
|
26
|
-
|
|
26
|
+
|
|
27
|
+
def _support_te():
|
|
28
|
+
try:
|
|
29
|
+
import te # pylint: disable=unused-import
|
|
30
|
+
return True
|
|
31
|
+
# pylint: disable=broad-except
|
|
32
|
+
except Exception:
|
|
33
|
+
return False
|
|
34
|
+
|
|
35
|
+
if context.get_context('device_target') == "Ascend" and _support_te():
|
|
27
36
|
import mindspore.ops._op_impl._custom_op
|
|
28
37
|
|
|
29
38
|
__all__ = ["MinMaxUpdatePerLayer",
|
|
@@ -61,10 +70,71 @@ __all__ = ["MinMaxUpdatePerLayer",
|
|
|
61
70
|
"ActsULQInputGrad",
|
|
62
71
|
"ActULQClampMinGrad",
|
|
63
72
|
"ActULQClampMaxGrad",
|
|
64
|
-
"WtsARQ"
|
|
73
|
+
"WtsARQ",
|
|
74
|
+
"FakeQuantParam",
|
|
65
75
|
]
|
|
66
76
|
|
|
67
77
|
|
|
78
|
+
class FakeQuantParam(Primitive):
|
|
79
|
+
r"""
|
|
80
|
+
Define the operation for storing quant parameter. This operation passes through input tensor to output tensor
|
|
81
|
+
without any calculation.
|
|
82
|
+
|
|
83
|
+
Args:
|
|
84
|
+
quant_dtype (QuantDtype) - The valid data type of the input tensor.
|
|
85
|
+
quant_algo_name (str) - Define the name of quant algorithm. Use
|
|
86
|
+
`FakeQuantParam.attr_value_linear_quant_algo_name` for linear quantization specially.
|
|
87
|
+
is_per_channel (bool) - Define whether quant parameter is per-channel or per-layer.
|
|
88
|
+
kwargs (dict): Other quant parameter in key-value form. Please use classmethod `linear_quant_param` to create a
|
|
89
|
+
linear quantization specially because key of scale and zero-point is pre-defined by MindSpore.
|
|
90
|
+
|
|
91
|
+
Inputs:
|
|
92
|
+
- *input_x* (Tensor) : Input tensor.
|
|
93
|
+
|
|
94
|
+
Outputs:
|
|
95
|
+
- Tensor: Output tensor same with `input_x`.
|
|
96
|
+
|
|
97
|
+
Examples:
|
|
98
|
+
>>> input_tensor = mindspore.Tensor(numpy.random.rand(1, 16, 5, 5), mindspore.dtype.float32)
|
|
99
|
+
>>> fake_quant_param_op = FakeQuantParam.linear_quant_param(mindspore.common.dtype.QuantDtype.INT8,
|
|
100
|
+
>>> 0.5, 1)
|
|
101
|
+
>>> output_tensor = fake_quant_param_op(input_tensor)
|
|
102
|
+
"""
|
|
103
|
+
|
|
104
|
+
attr_key_linear_quant_scale = "linear_quant_scale"
|
|
105
|
+
attr_key_linear_quant_zero_point = "linear_quant_zero_point"
|
|
106
|
+
|
|
107
|
+
attr_value_linear_quant_algo_name = "linear_quant_algo"
|
|
108
|
+
|
|
109
|
+
@prim_attr_register
|
|
110
|
+
def __init__(self, quant_dtype: QuantDtype, quant_algo_name: str, is_per_channel: bool, **kwargs):
|
|
111
|
+
self.add_prim_attr("quant_algo_name", quant_algo_name)
|
|
112
|
+
self.add_prim_attr("is_per_channel", is_per_channel)
|
|
113
|
+
self.add_prim_attr("quant_dtype", quant_dtype.value())
|
|
114
|
+
for key, value in kwargs.items():
|
|
115
|
+
self.add_prim_attr(key, value)
|
|
116
|
+
|
|
117
|
+
@classmethod
|
|
118
|
+
def linear_quant_param(cls, quant_dtype, scale, zp, is_per_channel=False, **kwargs):
|
|
119
|
+
"""
|
|
120
|
+
Create a linear quantization operator based on scale and zero-point parameter.
|
|
121
|
+
"""
|
|
122
|
+
validator.check_value_type("scale", scale, [float, tuple, list], "FakeQuantParam")
|
|
123
|
+
if isinstance(scale, float):
|
|
124
|
+
scale_list = [scale]
|
|
125
|
+
else:
|
|
126
|
+
scale_list = scale
|
|
127
|
+
validator.check_value_type("zero_point", zp, [int, tuple, list], "FakeQuantParam")
|
|
128
|
+
if isinstance(zp, int):
|
|
129
|
+
zp_list = [zp]
|
|
130
|
+
else:
|
|
131
|
+
zp_list = zp
|
|
132
|
+
validator.check_value_type("is_per_channel", is_per_channel, [bool], "FakeQuantParam")
|
|
133
|
+
kwargs[FakeQuantParam.attr_key_linear_quant_scale] = scale_list
|
|
134
|
+
kwargs[FakeQuantParam.attr_key_linear_quant_zero_point] = zp_list
|
|
135
|
+
return cls(quant_dtype, FakeQuantParam.attr_value_linear_quant_algo_name, is_per_channel, **kwargs)
|
|
136
|
+
|
|
137
|
+
|
|
68
138
|
class MinMaxUpdatePerLayer(PrimitiveWithInfer):
|
|
69
139
|
r"""
|
|
70
140
|
Updates min and max per layer.
|
|
@@ -99,14 +169,14 @@ class MinMaxUpdatePerLayer(PrimitiveWithInfer):
|
|
|
99
169
|
f"For '{self.name}' attr \'ema\' and \'ema_decay\' should set together.")
|
|
100
170
|
|
|
101
171
|
self.ema = validator.check_value_type('ema', ema, (bool,), self.name)
|
|
102
|
-
self.ema_decay = validator.check_float_range(ema_decay, 0, 1,
|
|
172
|
+
self.ema_decay = validator.check_float_range(ema_decay, 0, 1, validator.INC_BOTH, 'ema_decay', self.name)
|
|
103
173
|
self.init_prim_io_names(inputs=['x', 'min', 'max'],
|
|
104
174
|
outputs=['min_up', 'max_up'])
|
|
105
175
|
|
|
106
176
|
def infer_shape(self, x_shape, min_shape, max_shape):
|
|
107
|
-
validator.check_int(len(x_shape), 1,
|
|
177
|
+
validator.check_int(len(x_shape), 1, validator.GE, "x rank", self.name)
|
|
108
178
|
validator.check("min shape", min_shape, "max shape",
|
|
109
|
-
max_shape,
|
|
179
|
+
max_shape, validator.EQ, self.name)
|
|
110
180
|
validator.check_equal_int(len(min_shape), 1, "min shape", self.name)
|
|
111
181
|
return min_shape, max_shape
|
|
112
182
|
|
|
@@ -155,9 +225,10 @@ class MinMaxUpdatePerChannel(PrimitiveWithInfer):
|
|
|
155
225
|
f"For '{self.name}' attr \'ema\' and \'ema_decay\' should set together.")
|
|
156
226
|
|
|
157
227
|
self.ema = validator.check_value_type('ema', ema, (bool,), self.name)
|
|
158
|
-
self.ema_decay = validator.check_float_range(ema_decay, 0, 1,
|
|
228
|
+
self.ema_decay = validator.check_float_range(ema_decay, 0, 1, validator.INC_BOTH, 'ema_decay', self.name)
|
|
159
229
|
if self.is_ascend:
|
|
160
|
-
self.channel_axis = validator.check_int_range(channel_axis, 0, 1,
|
|
230
|
+
self.channel_axis = validator.check_int_range(channel_axis, 0, 1, validator.INC_BOTH,
|
|
231
|
+
'channel_axis', self.name)
|
|
161
232
|
else:
|
|
162
233
|
self.channel_axis = validator.check_non_negative_int(channel_axis, 'channel_axis', self.name)
|
|
163
234
|
self.init_prim_io_names(
|
|
@@ -167,9 +238,9 @@ class MinMaxUpdatePerChannel(PrimitiveWithInfer):
|
|
|
167
238
|
if self.is_ascend and len(x_shape) not in self.ascend_support_x_rank:
|
|
168
239
|
raise ValueError(f"For '{self.name}' x rank must be in '{self.ascend_support_x_rank}'")
|
|
169
240
|
if not self.is_ascend:
|
|
170
|
-
validator.check_int(len(x_shape), 1,
|
|
241
|
+
validator.check_int(len(x_shape), 1, validator.GE, "x rank", self.name)
|
|
171
242
|
validator.check("min shape", min_shape, "max shape",
|
|
172
|
-
max_shape,
|
|
243
|
+
max_shape, validator.EQ, self.name)
|
|
173
244
|
validator.check_equal_int(len(min_shape), 1, "min shape", self.name)
|
|
174
245
|
return min_shape, max_shape
|
|
175
246
|
|
|
@@ -225,9 +296,9 @@ class FakeLearnedScaleQuantPerLayer(PrimitiveWithInfer):
|
|
|
225
296
|
outputs=['out'])
|
|
226
297
|
|
|
227
298
|
def infer_shape(self, input_x_shape, alpha_shape, quant_max_shape):
|
|
228
|
-
validator.check_int(len(input_x_shape), 1,
|
|
229
|
-
validator.check_int(len(alpha_shape), 1,
|
|
230
|
-
validator.check_int(len(quant_max_shape), 1,
|
|
299
|
+
validator.check_int(len(input_x_shape), 1, validator.GE, "input_x rank", self.name)
|
|
300
|
+
validator.check_int(len(alpha_shape), 1, validator.GE, "alpha rank", self.name)
|
|
301
|
+
validator.check_int(len(quant_max_shape), 1, validator.GE, "quant max rank", self.name)
|
|
231
302
|
return input_x_shape
|
|
232
303
|
|
|
233
304
|
def infer_dtype(self, input_x_type, alpha_type, quant_max_type):
|
|
@@ -266,9 +337,9 @@ class FakeLearnedScaleQuantPerLayerGrad(PrimitiveWithInfer):
|
|
|
266
337
|
inputs=['dout', 'x', 'alpha', 'quant_max'], outputs=['dx', 'dalpha'])
|
|
267
338
|
|
|
268
339
|
def infer_shape(self, dout_shape, x_shape, alpha_shape, quant_max_shape):
|
|
269
|
-
validator.check("dout shape", dout_shape, "x_shape", x_shape,
|
|
270
|
-
validator.check_int(len(alpha_shape), 1,
|
|
271
|
-
validator.check_int(len(quant_max_shape), 1,
|
|
340
|
+
validator.check("dout shape", dout_shape, "x_shape", x_shape, validator.EQ, self.name)
|
|
341
|
+
validator.check_int(len(alpha_shape), 1, validator.GE, "alpha rank", self.name)
|
|
342
|
+
validator.check_int(len(quant_max_shape), 1, validator.GE, "quant max rank", self.name)
|
|
272
343
|
return dout_shape, alpha_shape
|
|
273
344
|
|
|
274
345
|
def infer_dtype(self, dout_type, x_type, alpha_type, quant_max_type):
|
|
@@ -297,9 +368,9 @@ class FakeLearnedScaleQuantPerLayerGradD(PrimitiveWithInfer):
|
|
|
297
368
|
inputs=['dout', 'x', 'alpha', 'quant_max'], outputs=['dx', 'dalpha'])
|
|
298
369
|
|
|
299
370
|
def infer_shape(self, dout_shape, x_shape, alpha_shape, quant_max_shape):
|
|
300
|
-
validator.check("dout shape", dout_shape, "x_shape", x_shape,
|
|
301
|
-
validator.check_int(len(alpha_shape), 1,
|
|
302
|
-
validator.check_int(len(quant_max_shape), 1,
|
|
371
|
+
validator.check("dout shape", dout_shape, "x_shape", x_shape, validator.EQ, self.name)
|
|
372
|
+
validator.check_int(len(alpha_shape), 1, validator.GE, "alpha rank", self.name)
|
|
373
|
+
validator.check_int(len(quant_max_shape), 1, validator.GE, "quant max rank", self.name)
|
|
303
374
|
return dout_shape, dout_shape
|
|
304
375
|
|
|
305
376
|
def infer_dtype(self, dout_type, x_type, alpha_type, quant_max_type):
|
|
@@ -375,7 +446,8 @@ class FakeLearnedScaleQuantPerChannel(PrimitiveWithInfer):
|
|
|
375
446
|
self.training = validator.check_value_type(
|
|
376
447
|
'training', training, (bool,), self.name)
|
|
377
448
|
if self.is_ascend:
|
|
378
|
-
self.channel_axis = validator.check_int_range(channel_axis, 0, 1,
|
|
449
|
+
self.channel_axis = validator.check_int_range(channel_axis, 0, 1, validator.INC_BOTH,
|
|
450
|
+
'channel_axis', self.name)
|
|
379
451
|
else:
|
|
380
452
|
self.channel_axis = validator.check_non_negative_int(channel_axis, 'channel_axis', self.name)
|
|
381
453
|
self.init_prim_io_names(inputs=['input_x', 'alpha', 'quant_max'],
|
|
@@ -385,12 +457,12 @@ class FakeLearnedScaleQuantPerChannel(PrimitiveWithInfer):
|
|
|
385
457
|
if self.is_ascend and len(input_x_shape) not in self.ascend_support_x_rank:
|
|
386
458
|
raise ValueError(f"For '{self.name}' x rank must be in '{self.ascend_support_x_rank}'")
|
|
387
459
|
if not self.is_ascend:
|
|
388
|
-
validator.check_int(len(input_x_shape), 1,
|
|
460
|
+
validator.check_int(len(input_x_shape), 1, validator.GE, "input_x rank", self.name)
|
|
389
461
|
if len(input_x_shape) == 1:
|
|
390
462
|
self.channel_axis = 0
|
|
391
463
|
|
|
392
464
|
validator.check_equal_int(alpha_shape[0], input_x_shape[self.channel_axis], "alpha rank", self.name)
|
|
393
|
-
validator.check_int(len(quant_max_shape), 1,
|
|
465
|
+
validator.check_int(len(quant_max_shape), 1, validator.GE, "quant max rank", self.name)
|
|
394
466
|
return input_x_shape
|
|
395
467
|
|
|
396
468
|
def infer_dtype(self, input_x_type, alpha_type, quant_max_type):
|
|
@@ -431,7 +503,7 @@ class FakeLearnedScaleQuantPerChannelGrad(PrimitiveWithInfer):
|
|
|
431
503
|
inputs=['dout', 'x', 'alpha', 'quant_max'], outputs=['dx', 'dalpha'])
|
|
432
504
|
|
|
433
505
|
def infer_shape(self, dout_shape, x_shape, alpha_shape, quant_max_shape):
|
|
434
|
-
validator.check("dout shape", dout_shape, "x_shape", x_shape,
|
|
506
|
+
validator.check("dout shape", dout_shape, "x_shape", x_shape, validator.EQ, self.name)
|
|
435
507
|
return dout_shape, alpha_shape
|
|
436
508
|
|
|
437
509
|
def infer_dtype(self, dout_type, x_type, alpha_type, quant_max_type):
|
|
@@ -462,9 +534,9 @@ class FakeLearnedScaleQuantPerChannelGradD(PrimitiveWithInfer):
|
|
|
462
534
|
inputs=['dout', 'x', 'alpha', 'quant_max'], outputs=['dx', 'dalpha'])
|
|
463
535
|
|
|
464
536
|
def infer_shape(self, dout_shape, x_shape, alpha_shape, quant_max_shape):
|
|
465
|
-
validator.check("dout shape", dout_shape, "x_shape", x_shape,
|
|
466
|
-
validator.check_int(len(alpha_shape), 1,
|
|
467
|
-
validator.check_int(len(quant_max_shape), 1,
|
|
537
|
+
validator.check("dout shape", dout_shape, "x_shape", x_shape, validator.EQ, self.name)
|
|
538
|
+
validator.check_int(len(alpha_shape), 1, validator.GE, "alpha rank", self.name)
|
|
539
|
+
validator.check_int(len(quant_max_shape), 1, validator.GE, "quant max rank", self.name)
|
|
468
540
|
return dout_shape, dout_shape
|
|
469
541
|
|
|
470
542
|
def infer_dtype(self, dout_type, x_type, alpha_type, quant_max_type):
|
|
@@ -528,7 +600,7 @@ class FakeQuantWithMinMaxVars(PrimitiveWithInfer):
|
|
|
528
600
|
num_bits=8,
|
|
529
601
|
narrow_range=False):
|
|
530
602
|
self.num_bits = validator.check_positive_int(num_bits, 'num_bits', self.name)
|
|
531
|
-
self.num_bits = validator.check_int_range(self.num_bits, 2, 16,
|
|
603
|
+
self.num_bits = validator.check_int_range(self.num_bits, 2, 16, validator.INC_BOTH, 'num_bits', self.name)
|
|
532
604
|
self.narrow_range = validator.check_value_type(
|
|
533
605
|
'narrow_range', narrow_range, (bool,), self.name)
|
|
534
606
|
|
|
@@ -540,9 +612,9 @@ class FakeQuantWithMinMaxVars(PrimitiveWithInfer):
|
|
|
540
612
|
raise ValueError(f"For '{self.name}', the shape of \'min\' cannot broadcast to the shape of \'x\'.")
|
|
541
613
|
|
|
542
614
|
def infer_shape(self, x_shape, min_shape, max_shape):
|
|
543
|
-
validator.check_int(len(x_shape), 1,
|
|
544
|
-
validator.check("min shape", min_shape, "max shape", max_shape,
|
|
545
|
-
validator.check_int(len(min_shape), 1,
|
|
615
|
+
validator.check_int(len(x_shape), 1, validator.GE, "x rank", self.name)
|
|
616
|
+
validator.check("min shape", min_shape, "max shape", max_shape, validator.EQ, self.name)
|
|
617
|
+
validator.check_int(len(min_shape), 1, validator.EQ, "min shape", self.name)
|
|
546
618
|
self.check_broadcast(min_shape, x_shape)
|
|
547
619
|
return x_shape
|
|
548
620
|
|
|
@@ -592,7 +664,7 @@ class FakeQuantWithMinMaxVarsGradient(PrimitiveWithInfer):
|
|
|
592
664
|
num_bits=8,
|
|
593
665
|
narrow_range=False):
|
|
594
666
|
self.num_bits = validator.check_positive_int(num_bits, 'num_bits', self.name)
|
|
595
|
-
self.num_bits = validator.check_int_range(self.num_bits, 2, 16,
|
|
667
|
+
self.num_bits = validator.check_int_range(self.num_bits, 2, 16, validator.INC_BOTH, 'num_bits', self.name)
|
|
596
668
|
self.narrow_range = validator.check_value_type(
|
|
597
669
|
'narrow_range', narrow_range, (bool,), self.name)
|
|
598
670
|
|
|
@@ -604,10 +676,10 @@ class FakeQuantWithMinMaxVarsGradient(PrimitiveWithInfer):
|
|
|
604
676
|
raise ValueError(f"For '{self.name}', the shape of \'min\' cannot broadcast to the shape of \'x\'.")
|
|
605
677
|
|
|
606
678
|
def infer_shape(self, dout_shape, x_shape, min_shape, max_shape):
|
|
607
|
-
validator.check_int(len(x_shape), 1,
|
|
608
|
-
validator.check("dout shape", dout_shape, "x shape", x_shape,
|
|
609
|
-
validator.check("min shape", min_shape, "max shape", max_shape,
|
|
610
|
-
validator.check_int(len(min_shape), 1,
|
|
679
|
+
validator.check_int(len(x_shape), 1, validator.GE, "x rank", self.name)
|
|
680
|
+
validator.check("dout shape", dout_shape, "x shape", x_shape, validator.EQ, self.name)
|
|
681
|
+
validator.check("min shape", min_shape, "max shape", max_shape, validator.EQ, self.name)
|
|
682
|
+
validator.check_int(len(min_shape), 1, validator.EQ, "min shape", self.name)
|
|
611
683
|
self.check_broadcast(min_shape, x_shape)
|
|
612
684
|
return x_shape, min_shape, max_shape
|
|
613
685
|
|
|
@@ -651,15 +723,15 @@ class FakeQuantWithMinMaxVarsPerChannel(PrimitiveWithInfer):
|
|
|
651
723
|
num_bits=8,
|
|
652
724
|
narrow_range=False):
|
|
653
725
|
self.num_bits = validator.check_positive_int(num_bits, 'num_bits', self.name)
|
|
654
|
-
self.num_bits = validator.check_int_range(self.num_bits, 2, 16,
|
|
726
|
+
self.num_bits = validator.check_int_range(self.num_bits, 2, 16, validator.INC_BOTH, 'num_bits', self.name)
|
|
655
727
|
self.narrow_range = validator.check_value_type(
|
|
656
728
|
'narrow_range', narrow_range, (bool,), self.name)
|
|
657
729
|
|
|
658
730
|
def infer_shape(self, x_shape, min_shape, max_shape):
|
|
659
|
-
validator.check_int(len(x_shape), 1,
|
|
660
|
-
validator.check("min shape", min_shape, "max shape", max_shape,
|
|
661
|
-
validator.check_int(len(min_shape), 1,
|
|
662
|
-
validator.check("min shape", min_shape[0], "x shape", x_shape[-1],
|
|
731
|
+
validator.check_int(len(x_shape), 1, validator.GE, "x rank", self.name)
|
|
732
|
+
validator.check("min shape", min_shape, "max shape", max_shape, validator.EQ, self.name)
|
|
733
|
+
validator.check_int(len(min_shape), 1, validator.EQ, "min shape", self.name)
|
|
734
|
+
validator.check("min shape", min_shape[0], "x shape", x_shape[-1], validator.EQ, self.name)
|
|
663
735
|
return x_shape
|
|
664
736
|
|
|
665
737
|
def infer_dtype(self, x_type, min_type, max_type):
|
|
@@ -709,16 +781,16 @@ class FakeQuantWithMinMaxVarsPerChannelGradient(PrimitiveWithInfer):
|
|
|
709
781
|
num_bits=8,
|
|
710
782
|
narrow_range=False):
|
|
711
783
|
self.num_bits = validator.check_positive_int(num_bits, 'num_bits', self.name)
|
|
712
|
-
self.num_bits = validator.check_int_range(self.num_bits, 2, 16,
|
|
784
|
+
self.num_bits = validator.check_int_range(self.num_bits, 2, 16, validator.INC_BOTH, 'num_bits', self.name)
|
|
713
785
|
self.narrow_range = validator.check_value_type(
|
|
714
786
|
'narrow_range', narrow_range, (bool,), self.name)
|
|
715
787
|
|
|
716
788
|
def infer_shape(self, dout_shape, x_shape, min_shape, max_shape):
|
|
717
|
-
validator.check_int(len(x_shape), 1,
|
|
718
|
-
validator.check("dout shape", dout_shape, "x shape", x_shape,
|
|
719
|
-
validator.check("min shape", min_shape, "max shape", max_shape,
|
|
720
|
-
validator.check_int(len(min_shape), 1,
|
|
721
|
-
validator.check("min shape", min_shape[0], "x shape", x_shape[-1],
|
|
789
|
+
validator.check_int(len(x_shape), 1, validator.GE, "x rank", self.name)
|
|
790
|
+
validator.check("dout shape", dout_shape, "x shape", x_shape, validator.EQ, self.name)
|
|
791
|
+
validator.check("min shape", min_shape, "max shape", max_shape, validator.EQ, self.name)
|
|
792
|
+
validator.check_int(len(min_shape), 1, validator.EQ, "min shape", self.name)
|
|
793
|
+
validator.check("min shape", min_shape[0], "x shape", x_shape[-1], validator.EQ, self.name)
|
|
722
794
|
return x_shape, min_shape, max_shape
|
|
723
795
|
|
|
724
796
|
def infer_dtype(self, dout_type, x_type, min_type, max_type):
|
|
@@ -807,15 +879,15 @@ class FakeQuantPerLayer(PrimitiveWithInfer):
|
|
|
807
879
|
self.narrow_range = validator.check_value_type(
|
|
808
880
|
'narrow_range', narrow_range, (bool,), self.name)
|
|
809
881
|
self.training = validator.check_value_type('training', training, (bool,), self.name)
|
|
810
|
-
self.ema_decay = validator.check_float_range(ema_decay, 0, 1,
|
|
882
|
+
self.ema_decay = validator.check_float_range(ema_decay, 0, 1, validator.INC_BOTH, 'ema_decay', self.name)
|
|
811
883
|
self.num_bits = validator.check_positive_int(num_bits, 'num_bits', self.name)
|
|
812
884
|
self.quant_delay = validator.check_non_negative_int(quant_delay, 'quant_delay', self.name)
|
|
813
885
|
self.init_prim_io_names(inputs=['x', 'min', 'max'],
|
|
814
886
|
outputs=['out'])
|
|
815
887
|
|
|
816
888
|
def infer_shape(self, x_shape, min_shape, max_shape):
|
|
817
|
-
validator.check_int(len(x_shape), 1,
|
|
818
|
-
validator.check("min shape", min_shape, "max shape", max_shape,
|
|
889
|
+
validator.check_int(len(x_shape), 1, validator.GE, "x rank", self.name)
|
|
890
|
+
validator.check("min shape", min_shape, "max shape", max_shape, validator.EQ, self.name)
|
|
819
891
|
validator.check_equal_int(len(min_shape), 1, "min shape", self.name)
|
|
820
892
|
return x_shape
|
|
821
893
|
|
|
@@ -861,9 +933,9 @@ class FakeQuantPerLayerGrad(PrimitiveWithInfer):
|
|
|
861
933
|
|
|
862
934
|
def infer_shape(self, dout_shape, x_shape, min_shape, max_shape):
|
|
863
935
|
validator.check("dout shape", dout_shape, "x shape",
|
|
864
|
-
x_shape,
|
|
936
|
+
x_shape, validator.EQ, self.name)
|
|
865
937
|
validator.check("min shape", min_shape, "max shape",
|
|
866
|
-
max_shape,
|
|
938
|
+
max_shape, validator.EQ, self.name)
|
|
867
939
|
validator.check_equal_int(len(min_shape), 1, "min shape", self.name)
|
|
868
940
|
return dout_shape
|
|
869
941
|
|
|
@@ -933,11 +1005,12 @@ class FakeQuantPerChannel(PrimitiveWithInfer):
|
|
|
933
1005
|
'narrow_range', narrow_range, (bool,), self.name)
|
|
934
1006
|
self.training = validator.check_value_type(
|
|
935
1007
|
'training', training, (bool,), self.name)
|
|
936
|
-
self.ema_decay = validator.check_float_range(ema_decay, 0, 1,
|
|
1008
|
+
self.ema_decay = validator.check_float_range(ema_decay, 0, 1, validator.INC_BOTH, 'ema_decay', self.name)
|
|
937
1009
|
self.num_bits = validator.check_positive_int(num_bits, 'num_bits', self.name)
|
|
938
1010
|
self.quant_delay = validator.check_non_negative_int(quant_delay, 'quant_delay', self.name)
|
|
939
1011
|
if self.is_ascend:
|
|
940
|
-
self.channel_axis = validator.check_int_range(channel_axis, 0, 1,
|
|
1012
|
+
self.channel_axis = validator.check_int_range(channel_axis, 0, 1, validator.INC_BOTH,
|
|
1013
|
+
'channel_axis', self.name)
|
|
941
1014
|
else:
|
|
942
1015
|
self.channel_axis = validator.check_non_negative_int(channel_axis, 'channel_axis', self.name)
|
|
943
1016
|
self.init_prim_io_names(inputs=['x', 'min', 'max'], outputs=['out'])
|
|
@@ -946,10 +1019,10 @@ class FakeQuantPerChannel(PrimitiveWithInfer):
|
|
|
946
1019
|
if self.is_ascend and len(x_shape) not in self.ascend_support_x_rank:
|
|
947
1020
|
raise ValueError(f"For '{self.name}' x rank must be in '{self.ascend_support_x_rank}'")
|
|
948
1021
|
if not self.is_ascend:
|
|
949
|
-
validator.check_int(len(x_shape), 1,
|
|
1022
|
+
validator.check_int(len(x_shape), 1, validator.GE, "x rank", self.name)
|
|
950
1023
|
if len(x_shape) == 1:
|
|
951
1024
|
self.channel_axis = 0
|
|
952
|
-
validator.check("min shape", min_shape, "max shape", max_shape,
|
|
1025
|
+
validator.check("min shape", min_shape, "max shape", max_shape, validator.EQ, self.name)
|
|
953
1026
|
validator.check_equal_int(min_shape[0], x_shape[self.channel_axis], "min shape", self.name)
|
|
954
1027
|
validator.check_equal_int(max_shape[0], x_shape[self.channel_axis], "max shape", self.name)
|
|
955
1028
|
return x_shape
|
|
@@ -1045,7 +1118,7 @@ class BatchNormFold(PrimitiveWithInfer):
|
|
|
1045
1118
|
@prim_attr_register
|
|
1046
1119
|
def __init__(self, momentum=0.9, epsilon=1e-5, is_training=True, freeze_bn=0):
|
|
1047
1120
|
"""Initialize batch norm fold layer"""
|
|
1048
|
-
self.momentum = validator.check_float_range(momentum, 0, 1,
|
|
1121
|
+
self.momentum = validator.check_float_range(momentum, 0, 1, validator.INC_BOTH, 'momentum', self.name)
|
|
1049
1122
|
self.epsilon = validator.check_positive_float(epsilon, 'epsilon', self.name)
|
|
1050
1123
|
self.is_training = validator.check_value_type('is_training', is_training, (bool,), self.name)
|
|
1051
1124
|
self.freeze_bn = validator.check_value_type('freeze_bn', freeze_bn, (int,), self.name)
|
|
@@ -1054,8 +1127,9 @@ class BatchNormFold(PrimitiveWithInfer):
|
|
|
1054
1127
|
outputs=['batch_mean', 'batch_std', 'running_mean', 'running_std'])
|
|
1055
1128
|
|
|
1056
1129
|
def infer_shape(self, x_shape, mean_shape, variance_shape, global_step_shape):
|
|
1057
|
-
validator.check("mean shape", mean_shape, "gamma_shape", variance_shape,
|
|
1058
|
-
validator.check("mean_shape[0]", mean_shape[0], "input channel",
|
|
1130
|
+
validator.check("mean shape", mean_shape, "gamma_shape", variance_shape, validator.EQ, self.name)
|
|
1131
|
+
validator.check("mean_shape[0]", mean_shape[0], "input channel",
|
|
1132
|
+
x_shape[self.channel_axis], validator.EQ, self.name)
|
|
1059
1133
|
validator.check_equal_int(len(global_step_shape), 1, "global step shape len", self.name)
|
|
1060
1134
|
return mean_shape, mean_shape, mean_shape, mean_shape
|
|
1061
1135
|
|
|
@@ -1096,13 +1170,13 @@ class BatchNormFoldGrad(PrimitiveWithInfer):
|
|
|
1096
1170
|
def infer_shape(self, d_batch_mean_shape, d_batch_std_shape, x_shape, batch_mean_shape, batch_std_shape,
|
|
1097
1171
|
global_step_shape):
|
|
1098
1172
|
validator.check("d_batch_mean shape", d_batch_mean_shape,
|
|
1099
|
-
"d_batch_std shape", d_batch_std_shape,
|
|
1173
|
+
"d_batch_std shape", d_batch_std_shape, validator.EQ, self.name)
|
|
1100
1174
|
validator.check("d_batch_mean shape", d_batch_mean_shape,
|
|
1101
|
-
"batch_mean shape", batch_mean_shape,
|
|
1175
|
+
"batch_mean shape", batch_mean_shape, validator.EQ, self.name)
|
|
1102
1176
|
validator.check("d_batch_mean shape", d_batch_mean_shape,
|
|
1103
|
-
"batch_std shape", batch_std_shape,
|
|
1177
|
+
"batch_std shape", batch_std_shape, validator.EQ, self.name)
|
|
1104
1178
|
validator.check("d_batch_mean_shape[0]", d_batch_mean_shape[0],
|
|
1105
|
-
"input channel", x_shape[self.channel_axis],
|
|
1179
|
+
"input channel", x_shape[self.channel_axis], validator.EQ, self.name)
|
|
1106
1180
|
validator.check_equal_int(len(global_step_shape), 1, "global step shape len", self.name)
|
|
1107
1181
|
return x_shape
|
|
1108
1182
|
|
|
@@ -1147,9 +1221,10 @@ class CorrectionMul(PrimitiveWithInfer):
|
|
|
1147
1221
|
outputs=['out'])
|
|
1148
1222
|
|
|
1149
1223
|
def infer_shape(self, x_shape, batch_std_shape, running_std_shape):
|
|
1150
|
-
validator.check("batch_std shape", batch_std_shape, "running_std shape",
|
|
1224
|
+
validator.check("batch_std shape", batch_std_shape, "running_std shape",
|
|
1225
|
+
running_std_shape, validator.EQ, self.name)
|
|
1151
1226
|
validator.check("batch_std_shape[0]", batch_std_shape[0], "x_shape channel size", x_shape[self.channel_axis],
|
|
1152
|
-
|
|
1227
|
+
validator.EQ, self.name)
|
|
1153
1228
|
return x_shape
|
|
1154
1229
|
|
|
1155
1230
|
def infer_dtype(self, x_type, batch_std_type, running_std_type):
|
|
@@ -1181,11 +1256,11 @@ class CorrectionMulGrad(PrimitiveWithInfer):
|
|
|
1181
1256
|
outputs=['dx', 'mul_dx'])
|
|
1182
1257
|
|
|
1183
1258
|
def infer_shape(self, dout_shape, x_shape, gamma_shape, running_std_shape):
|
|
1184
|
-
validator.check("dout shape", dout_shape, "x_shape x", x_shape,
|
|
1259
|
+
validator.check("dout shape", dout_shape, "x_shape x", x_shape, validator.EQ, self.name)
|
|
1185
1260
|
validator.check("gamma_shape[0]", gamma_shape[0], "dout channel size", dout_shape[self.channel_axis],
|
|
1186
|
-
|
|
1261
|
+
validator.EQ, self.name)
|
|
1187
1262
|
validator.check("running_std_shape[0]", running_std_shape[0],
|
|
1188
|
-
"dout channel size", dout_shape[self.channel_axis],
|
|
1263
|
+
"dout channel size", dout_shape[self.channel_axis], validator.EQ, self.name)
|
|
1189
1264
|
if context.get_context('device_target') == "Ascend":
|
|
1190
1265
|
return x_shape, x_shape
|
|
1191
1266
|
return x_shape, gamma_shape
|
|
@@ -1271,14 +1346,16 @@ class BatchNormFold2(PrimitiveWithInfer):
|
|
|
1271
1346
|
|
|
1272
1347
|
def infer_shape(self, x_shape, beta_shape, gamma_shape, batch_std_shape, running_std_shape, batch_mean_shape,
|
|
1273
1348
|
running_mean_shape, global_step_shape):
|
|
1274
|
-
validator.check("batch_std shape", batch_std_shape, "running_std shape",
|
|
1275
|
-
|
|
1276
|
-
validator.check("batch_std shape", batch_std_shape, "
|
|
1349
|
+
validator.check("batch_std shape", batch_std_shape, "running_std shape",
|
|
1350
|
+
running_std_shape, validator.EQ, self.name)
|
|
1351
|
+
validator.check("batch_std shape", batch_std_shape, "batch_mean shape",
|
|
1352
|
+
batch_mean_shape, validator.EQ, self.name)
|
|
1353
|
+
validator.check("batch_std shape", batch_std_shape, "beta shape", beta_shape, validator.EQ, self.name)
|
|
1277
1354
|
validator.check("batch_std shape", batch_std_shape, "running_mean shape", running_mean_shape,
|
|
1278
|
-
|
|
1279
|
-
validator.check("batch_std shape", batch_std_shape, "batch_mean shape", gamma_shape,
|
|
1355
|
+
validator.EQ, self.name)
|
|
1356
|
+
validator.check("batch_std shape", batch_std_shape, "batch_mean shape", gamma_shape, validator.EQ, self.name)
|
|
1280
1357
|
validator.check("batch_std_shape[0]", batch_std_shape[0], "x_shape channel size", x_shape[self.channel_axis],
|
|
1281
|
-
|
|
1358
|
+
validator.EQ, self.name)
|
|
1282
1359
|
validator.check_equal_int(len(global_step_shape), 1, "global step shape len", self.name)
|
|
1283
1360
|
return x_shape
|
|
1284
1361
|
|
|
@@ -1321,13 +1398,15 @@ class BatchNormFold2Grad(PrimitiveWithInfer):
|
|
|
1321
1398
|
def infer_shape(self, dout_shape, x_shape, gamma_shape,
|
|
1322
1399
|
batch_std_shape, batch_mean_shape,
|
|
1323
1400
|
running_std_shape, running_mean_shape, global_step_shape):
|
|
1324
|
-
validator.check("batch_std shape", batch_std_shape, "batch_mean shape",
|
|
1325
|
-
|
|
1401
|
+
validator.check("batch_std shape", batch_std_shape, "batch_mean shape",
|
|
1402
|
+
batch_mean_shape, validator.EQ, self.name)
|
|
1403
|
+
validator.check("batch_std shape", batch_std_shape, "running_std shape",
|
|
1404
|
+
running_std_shape, validator.EQ, self.name)
|
|
1326
1405
|
validator.check("batch_std shape", batch_std_shape, "running_mean shape", running_mean_shape,
|
|
1327
|
-
|
|
1328
|
-
validator.check("batch_std shape", batch_std_shape, "gamma shape", gamma_shape,
|
|
1406
|
+
validator.EQ, self.name)
|
|
1407
|
+
validator.check("batch_std shape", batch_std_shape, "gamma shape", gamma_shape, validator.EQ, self.name)
|
|
1329
1408
|
validator.check("batch_std size", batch_std_shape[0], "dout channel size", dout_shape[self.channel_axis],
|
|
1330
|
-
|
|
1409
|
+
validator.EQ, self.name)
|
|
1331
1410
|
validator.check_equal_int(len(global_step_shape), 1, "global step shape len", self.name)
|
|
1332
1411
|
return gamma_shape, gamma_shape, gamma_shape, gamma_shape, x_shape
|
|
1333
1412
|
|
|
@@ -1358,7 +1437,7 @@ class BatchNormFoldD(PrimitiveWithInfer):
|
|
|
1358
1437
|
def __init__(self, momentum=0.9, epsilon=1e-5, is_training=True, freeze_bn=0):
|
|
1359
1438
|
"""Initialize _BatchNormFold layer"""
|
|
1360
1439
|
from mindspore.ops._op_impl._custom_op import batchnorm_fold
|
|
1361
|
-
self.momentum = validator.check_float_range(momentum, 0, 1,
|
|
1440
|
+
self.momentum = validator.check_float_range(momentum, 0, 1, validator.INC_BOTH, 'momentum', self.name)
|
|
1362
1441
|
self.epsilon = validator.check_positive_float(epsilon, 'epsilon', self.name)
|
|
1363
1442
|
self.is_training = validator.check_value_type('is_training', is_training, (bool,), self.name)
|
|
1364
1443
|
self.freeze_bn = validator.check_value_type('freeze_bn', freeze_bn, (int,), self.name)
|
|
@@ -1368,8 +1447,8 @@ class BatchNormFoldD(PrimitiveWithInfer):
|
|
|
1368
1447
|
'mean_updated', 'variance_updated'])
|
|
1369
1448
|
|
|
1370
1449
|
def infer_shape(self, x_shape, x_sum_shape, x_square_sum_shape, mean_shape, variance_shape):
|
|
1371
|
-
validator.check("mean shape", mean_shape, "gamma_shape", variance_shape,
|
|
1372
|
-
validator.check("mean_shape[0]", mean_shape[0], "input channel", x_shape[1],
|
|
1450
|
+
validator.check("mean shape", mean_shape, "gamma_shape", variance_shape, validator.EQ, self.name)
|
|
1451
|
+
validator.check("mean_shape[0]", mean_shape[0], "input channel", x_shape[1], validator.EQ, self.name)
|
|
1373
1452
|
return x_shape, mean_shape, mean_shape, mean_shape, mean_shape, mean_shape, mean_shape
|
|
1374
1453
|
|
|
1375
1454
|
def infer_dtype(self, x_type, x_sum_type, x_square_sum_type, mean_type, variance_type):
|
|
@@ -1439,12 +1518,14 @@ class BatchNormFold2D(PrimitiveWithInfer):
|
|
|
1439
1518
|
outputs=['y'])
|
|
1440
1519
|
|
|
1441
1520
|
def infer_shape(self, x_shape, beta_shape, gamma_shape, batch_std_shape, running_std_shape, batch_mean_shape):
|
|
1442
|
-
validator.check("batch_std shape", batch_std_shape, "running_std shape",
|
|
1443
|
-
|
|
1444
|
-
validator.check("batch_std shape", batch_std_shape, "
|
|
1445
|
-
|
|
1521
|
+
validator.check("batch_std shape", batch_std_shape, "running_std shape",
|
|
1522
|
+
running_std_shape, validator.EQ, self.name)
|
|
1523
|
+
validator.check("batch_std shape", batch_std_shape, "batch_mean shape",
|
|
1524
|
+
batch_mean_shape, validator.EQ, self.name)
|
|
1525
|
+
validator.check("batch_std shape", batch_std_shape, "beta shape", beta_shape, validator.EQ, self.name)
|
|
1526
|
+
validator.check("batch_std shape", batch_std_shape, "batch_mean shape", gamma_shape, validator.EQ, self.name)
|
|
1446
1527
|
validator.check("batch_std_shape[0]", batch_std_shape[0], "x_shape channel size", x_shape[self.channel_axis],
|
|
1447
|
-
|
|
1528
|
+
validator.EQ, self.name)
|
|
1448
1529
|
return x_shape
|
|
1449
1530
|
|
|
1450
1531
|
def infer_dtype(self, x_type, beta_type, gamma_type, batch_std_type, running_std_type, batch_mean_type):
|
|
@@ -1469,11 +1550,13 @@ class BatchNormFold2GradD(PrimitiveWithInfer):
|
|
|
1469
1550
|
|
|
1470
1551
|
def infer_shape(self, dout_shape, dout_reduce_shape, dout_x_reduce_shape, gamma_shape, batch_std_shape,
|
|
1471
1552
|
batch_mean_shape, running_std_shape):
|
|
1472
|
-
validator.check("batch_std shape", batch_std_shape, "batch_mean shape",
|
|
1473
|
-
|
|
1474
|
-
validator.check("batch_std shape", batch_std_shape, "
|
|
1553
|
+
validator.check("batch_std shape", batch_std_shape, "batch_mean shape",
|
|
1554
|
+
batch_mean_shape, validator.EQ, self.name)
|
|
1555
|
+
validator.check("batch_std shape", batch_std_shape, "running_std shape",
|
|
1556
|
+
running_std_shape, validator.EQ, self.name)
|
|
1557
|
+
validator.check("batch_std shape", batch_std_shape, "gamma shape", gamma_shape, validator.EQ, self.name)
|
|
1475
1558
|
validator.check("batch_std size", batch_std_shape[0], "dout channel size", dout_shape[self.channel_axis],
|
|
1476
|
-
|
|
1559
|
+
validator.EQ, self.name)
|
|
1477
1560
|
return gamma_shape, gamma_shape, gamma_shape, dout_shape
|
|
1478
1561
|
|
|
1479
1562
|
def infer_dtype(self, dout_type, dout_reduce_type, dout_x_reduce_type, gamma_type, batch_std_type,
|
|
@@ -1505,7 +1588,7 @@ class BatchNormFold2GradReduce(PrimitiveWithInfer):
|
|
|
1505
1588
|
outputs=['dout_reduce', 'dout_x_reduce'])
|
|
1506
1589
|
|
|
1507
1590
|
def infer_shape(self, dout_shape, x_shape):
|
|
1508
|
-
validator.check("dout shape", dout_shape, "x shape", x_shape,
|
|
1591
|
+
validator.check("dout shape", dout_shape, "x shape", x_shape, validator.EQ, self.name)
|
|
1509
1592
|
return (dout_shape[self.channel_axis],), (dout_shape[self.channel_axis],)
|
|
1510
1593
|
|
|
1511
1594
|
def infer_dtype(self, dout_type, x_type):
|
|
@@ -1547,17 +1630,17 @@ class ActsULQ(PrimitiveWithInfer):
|
|
|
1547
1630
|
def __init__(self, fixed_min=False, num_bits=8):
|
|
1548
1631
|
validator.check_value_type("fixed_min", fixed_min, [bool], self.name)
|
|
1549
1632
|
validator.check_value_type("num_bits", num_bits, [int], self.name)
|
|
1550
|
-
validator.check_int(num_bits, 8,
|
|
1633
|
+
validator.check_int(num_bits, 8, validator.EQ, "value of num_bits", self.name)
|
|
1551
1634
|
|
|
1552
1635
|
def infer_shape(self, x_shape, clamp_min_shape, clamp_max_shape):
|
|
1553
1636
|
"""infer shape of primitive"""
|
|
1554
|
-
validator.check_int(len(clamp_min_shape), len(x_shape),
|
|
1555
|
-
validator.check_int(len(clamp_max_shape), len(x_shape),
|
|
1637
|
+
validator.check_int(len(clamp_min_shape), len(x_shape), validator.EQ, "dims of clamp_min", self.name)
|
|
1638
|
+
validator.check_int(len(clamp_max_shape), len(x_shape), validator.EQ, "dims of clamp_max", self.name)
|
|
1556
1639
|
|
|
1557
1640
|
x_shape_len = len(x_shape)
|
|
1558
1641
|
for i in range(x_shape_len):
|
|
1559
|
-
validator.check_int(clamp_min_shape[i], 1,
|
|
1560
|
-
validator.check_int(clamp_max_shape[i], 1,
|
|
1642
|
+
validator.check_int(clamp_min_shape[i], 1, validator.EQ, "dims of clamp_min", self.name)
|
|
1643
|
+
validator.check_int(clamp_max_shape[i], 1, validator.EQ, "dims of clamp_max", self.name)
|
|
1561
1644
|
|
|
1562
1645
|
return x_shape, x_shape, x_shape, x_shape
|
|
1563
1646
|
|
|
@@ -1698,12 +1781,12 @@ class WtsARQ(PrimitiveWithInfer):
|
|
|
1698
1781
|
@prim_attr_register
|
|
1699
1782
|
def __init__(self, num_bits, offset_flag):
|
|
1700
1783
|
validator.check_value_type("num_bits", num_bits, [int], self.name)
|
|
1701
|
-
validator.check_int(num_bits, 8,
|
|
1784
|
+
validator.check_int(num_bits, 8, validator.EQ, "value of num_bits", self.name)
|
|
1702
1785
|
validator.check_value_type("offset_flag", offset_flag, [bool], self.name)
|
|
1703
1786
|
|
|
1704
1787
|
def infer_shape(self, w_shape, w_min_shape, w_max_shape):
|
|
1705
|
-
validator.check_int(len(w_min_shape), len(w_shape),
|
|
1706
|
-
validator.check_int(len(w_max_shape), len(w_shape),
|
|
1788
|
+
validator.check_int(len(w_min_shape), len(w_shape), validator.EQ, "dims of w_min", self.name)
|
|
1789
|
+
validator.check_int(len(w_max_shape), len(w_shape), validator.EQ, "dims of w_max", self.name)
|
|
1707
1790
|
return w_shape
|
|
1708
1791
|
|
|
1709
1792
|
def infer_dtype(self, w_dtype, w_min_dtype, w_max_dtype):
|
|
@@ -1753,11 +1836,13 @@ class IFMR(Primitive):
|
|
|
1753
1836
|
@prim_attr_register
|
|
1754
1837
|
def __init__(self, min_percentile=0.999999, max_percentile=0.999999, search_range=(0.7, 1.3), search_step=0.01,
|
|
1755
1838
|
with_offset=True):
|
|
1839
|
+
self.init_prim_io_names(
|
|
1840
|
+
inputs=['data', 'data_min', 'data_max', 'cumsum'], outputs=['scale', 'offset'])
|
|
1756
1841
|
validator.check_value_type("min_percentile", min_percentile, [float], self.name)
|
|
1757
1842
|
validator.check_value_type("max_percentile", max_percentile, [float], self.name)
|
|
1758
1843
|
validator.check_value_type("search_range", search_range, [list, tuple], self.name)
|
|
1759
1844
|
for item in search_range:
|
|
1760
1845
|
validator.check_positive_float(item, "item of search_range", self.name)
|
|
1761
|
-
validator.check('search_range[1]', search_range[1], 'search_range[0]', search_range[0],
|
|
1846
|
+
validator.check('search_range[1]', search_range[1], 'search_range[0]', search_range[0], validator.GE, self.name)
|
|
1762
1847
|
validator.check_value_type("search_step", search_step, [float], self.name)
|
|
1763
1848
|
validator.check_value_type("offset_flag", with_offset, [bool], self.name)
|