mindspore 1.10.0__cp37-cp37m-win_amd64.whl → 2.0.0rc1__cp37-cp37m-win_amd64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of mindspore might be problematic. Click here for more details.
- mindspore/.commit_id +1 -1
- mindspore/ConcurrencyCheck.dll +0 -0
- mindspore/CppBuildInsights.dll +0 -0
- mindspore/CppCoreCheck.dll +0 -0
- mindspore/EnumIndex.dll +0 -0
- mindspore/EspXEngine.dll +0 -0
- mindspore/HResultCheck.dll +0 -0
- mindspore/KernelTraceControl.dll +0 -0
- mindspore/LocalESPC.dll +0 -0
- mindspore/Microsoft.Diagnostics.Tracing.EventSource.dll +0 -0
- mindspore/Microsoft.VisualStudio.RemoteControl.dll +0 -0
- mindspore/Microsoft.VisualStudio.Telemetry.dll +0 -0
- mindspore/Microsoft.VisualStudio.Utilities.Internal.dll +0 -0
- mindspore/Newtonsoft.Json.dll +0 -0
- mindspore/System.Runtime.CompilerServices.Unsafe.dll +0 -0
- mindspore/VariantClear.dll +0 -0
- mindspore/__init__.py +9 -4
- mindspore/_c_dataengine.cp37-win_amd64.pyd +0 -0
- mindspore/_c_expression.cp37-win_amd64.pyd +0 -0
- mindspore/_c_mindrecord.cp37-win_amd64.pyd +0 -0
- mindspore/_check_jit_forbidden_api.py +102 -0
- mindspore/_checkparam.py +1066 -1001
- mindspore/_extends/builtin_operations.py +32 -4
- mindspore/_extends/graph_kernel/model/graph_split.py +66 -222
- mindspore/_extends/parallel_compile/akg_compiler/akg_process.py +12 -9
- mindspore/_extends/parallel_compile/akg_compiler/build_tbe_kernel.py +119 -26
- mindspore/_extends/parallel_compile/akg_compiler/tbe_topi.py +50 -50
- mindspore/_extends/parallel_compile/akg_compiler/util.py +9 -6
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_adapter.py +4 -25
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_helper.py +9 -4
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_job_manager.py +1 -27
- mindspore/_extends/parse/__init__.py +5 -3
- mindspore/_extends/parse/namespace.py +17 -2
- mindspore/_extends/parse/parser.py +193 -34
- mindspore/_extends/parse/resources.py +7 -8
- mindspore/_extends/parse/standard_method.py +1780 -435
- mindspore/_extends/parse/trope.py +3 -1
- mindspore/amp.py +53 -58
- mindspore/atlprov.dll +0 -0
- mindspore/boost/adasum.py +3 -2
- mindspore/boost/boost.py +2 -2
- mindspore/boost/boost_cell_wrapper.py +46 -26
- mindspore/boost/dim_reduce.py +6 -5
- mindspore/boost/grad_accumulation.py +2 -1
- mindspore/boost/group_loss_scale_manager.py +1 -1
- mindspore/c1.dll +0 -0
- mindspore/c1xx.dll +0 -0
- mindspore/c2.dll +0 -0
- mindspore/cfgpersist.dll +0 -0
- mindspore/clang_rt.asan_dbg_dynamic-x86_64.dll +0 -0
- mindspore/clang_rt.asan_dynamic-x86_64.dll +0 -0
- mindspore/common/__init__.py +11 -10
- mindspore/common/_decorator.py +2 -0
- mindspore/common/_register_for_adapter.py +55 -0
- mindspore/common/_stub_tensor.py +201 -0
- mindspore/common/_utils.py +57 -0
- mindspore/common/api.py +582 -297
- mindspore/common/dtype.py +66 -18
- mindspore/common/dump.py +2 -2
- mindspore/common/initializer.py +38 -1
- mindspore/common/jit_config.py +25 -13
- mindspore/common/mutable.py +53 -24
- mindspore/common/parameter.py +60 -37
- mindspore/common/seed.py +8 -24
- mindspore/common/sparse_tensor.py +927 -0
- mindspore/common/tensor.py +1627 -3900
- mindspore/communication/__init__.py +10 -5
- mindspore/communication/_comm_helper.py +78 -214
- mindspore/communication/_hccl_management.py +2 -1
- mindspore/communication/management.py +136 -47
- mindspore/config/op_info.config +501 -1008
- mindspore/context.py +291 -56
- mindspore/d3dcompiler_47.dll +0 -0
- mindspore/dataset/__init__.py +12 -8
- mindspore/dataset/audio/__init__.py +9 -9
- mindspore/dataset/audio/transforms.py +1090 -228
- mindspore/dataset/audio/utils.py +87 -39
- mindspore/dataset/audio/validators.py +223 -1
- mindspore/dataset/callback/ds_callback.py +17 -15
- mindspore/dataset/core/config.py +246 -17
- mindspore/dataset/core/py_util_helpers.py +4 -3
- mindspore/dataset/core/validator_helpers.py +10 -10
- mindspore/{parallel/nn/layers.py → dataset/debug/__init__.py} +7 -8
- mindspore/dataset/debug/debug_hook.py +65 -0
- mindspore/dataset/debug/pre_defined_hook.py +67 -0
- mindspore/dataset/engine/__init__.py +7 -3
- mindspore/dataset/engine/cache_client.py +9 -9
- mindspore/dataset/engine/datasets.py +648 -477
- mindspore/dataset/engine/datasets_audio.py +165 -167
- mindspore/dataset/engine/datasets_standard_format.py +93 -67
- mindspore/dataset/engine/datasets_text.py +492 -342
- mindspore/dataset/engine/datasets_user_defined.py +85 -50
- mindspore/dataset/engine/datasets_vision.py +1224 -699
- mindspore/dataset/engine/graphdata.py +134 -69
- mindspore/dataset/engine/iterators.py +50 -9
- mindspore/dataset/engine/offload.py +52 -31
- mindspore/dataset/engine/samplers.py +27 -24
- mindspore/dataset/engine/serializer_deserializer.py +14 -15
- mindspore/dataset/engine/validators.py +213 -52
- mindspore/dataset/text/__init__.py +10 -8
- mindspore/dataset/text/transforms.py +152 -57
- mindspore/dataset/text/utils.py +98 -49
- mindspore/dataset/text/validators.py +25 -0
- mindspore/dataset/transforms/__init__.py +4 -2
- mindspore/dataset/transforms/c_transforms.py +11 -13
- mindspore/dataset/transforms/py_transforms.py +2 -2
- mindspore/dataset/transforms/py_transforms_util.py +10 -0
- mindspore/dataset/transforms/transforms.py +13 -15
- mindspore/dataset/transforms/validators.py +7 -7
- mindspore/dataset/utils/__init__.py +2 -1
- mindspore/dataset/utils/browse_dataset.py +13 -13
- mindspore/dataset/utils/line_reader.py +121 -0
- mindspore/dataset/vision/__init__.py +8 -7
- mindspore/dataset/vision/c_transforms.py +125 -126
- mindspore/dataset/vision/py_transforms.py +37 -37
- mindspore/dataset/vision/py_transforms_util.py +23 -20
- mindspore/dataset/vision/transforms.py +316 -315
- mindspore/dataset/vision/utils.py +313 -17
- mindspore/dataset/vision/validators.py +6 -6
- mindspore/default_config.py +0 -1
- mindspore/dpcmi.dll +0 -0
- mindspore/{compression → experimental}/__init__.py +6 -5
- mindspore/experimental/map_parameter.py +275 -0
- mindspore/include/OWNERS +0 -1
- mindspore/include/api/callback/callback.h +9 -13
- mindspore/include/api/callback/ckpt_saver.h +2 -2
- mindspore/include/api/callback/loss_monitor.h +2 -2
- mindspore/include/api/callback/lr_scheduler.h +5 -5
- mindspore/include/api/callback/time_monitor.h +2 -2
- mindspore/include/api/callback/train_accuracy.h +4 -6
- mindspore/include/api/cfg.h +19 -6
- mindspore/include/api/context.h +70 -9
- mindspore/include/api/delegate.h +8 -1
- mindspore/include/api/dual_abi_helper.h +8 -24
- mindspore/include/api/metrics/accuracy.h +2 -2
- mindspore/include/api/metrics/metrics.h +4 -3
- mindspore/include/api/model.h +9 -4
- mindspore/include/api/model_group.h +68 -0
- mindspore/include/api/model_parallel_runner.h +17 -17
- mindspore/include/api/net.h +12 -11
- mindspore/include/api/serialization.h +20 -4
- mindspore/include/api/status.h +7 -1
- mindspore/include/api/types.h +25 -21
- mindspore/include/api/visible.h +4 -0
- mindspore/include/c_api/model_c.h +5 -0
- mindspore/include/c_api/status_c.h +1 -1
- mindspore/include/dataset/config.h +1 -1
- mindspore/include/dataset/constants.h +14 -0
- mindspore/include/dataset/text.h +59 -0
- mindspore/include/dataset/vision.h +56 -117
- mindspore/include/dataset/vision_lite.h +102 -0
- mindspore/jpeg62.dll +0 -0
- mindspore/log.py +28 -28
- mindspore/mindrecord/common/exceptions.py +2 -4
- mindspore/mindrecord/filereader.py +19 -1
- mindspore/mindrecord/filewriter.py +250 -88
- mindspore/mindrecord/mindpage.py +13 -13
- mindspore/mindrecord/shardheader.py +15 -15
- mindspore/mindrecord/shardreader.py +9 -0
- mindspore/mindrecord/shardwriter.py +29 -29
- mindspore/mindrecord/tools/cifar100_to_mr.py +9 -9
- mindspore/mindrecord/tools/cifar10_to_mr.py +9 -9
- mindspore/mindrecord/tools/csv_to_mr.py +4 -4
- mindspore/mindrecord/tools/imagenet_to_mr.py +70 -65
- mindspore/mindrecord/tools/mnist_to_mr.py +41 -41
- mindspore/mindrecord/tools/tfrecord_to_mr.py +6 -6
- mindspore/{libmindspore_backend.dll → mindspore_backend.dll} +0 -0
- mindspore/mindspore_common.dll +0 -0
- mindspore/mindspore_core.dll +0 -0
- mindspore/mindspore_glog.dll +0 -0
- mindspore/mindspore_shared_lib.dll +0 -0
- mindspore/msobj140.dll +0 -0
- mindspore/mspdb140.dll +0 -0
- mindspore/mspdbcore.dll +0 -0
- mindspore/mspdbst.dll +0 -0
- mindspore/mspft140.dll +0 -0
- mindspore/msvcdis140.dll +0 -0
- mindspore/msvcp140_1.dll +0 -0
- mindspore/msvcp140_2.dll +0 -0
- mindspore/msvcp140_atomic_wait.dll +0 -0
- mindspore/msvcp140_codecvt_ids.dll +0 -0
- mindspore/nn/__init__.py +1 -5
- mindspore/nn/cell.py +297 -234
- mindspore/nn/dynamic_lr.py +1 -1
- mindspore/nn/grad/cell_grad.py +17 -42
- mindspore/nn/layer/__init__.py +7 -4
- mindspore/nn/layer/activation.py +131 -88
- mindspore/nn/layer/basic.py +313 -613
- mindspore/nn/layer/channel_shuffle.py +103 -0
- mindspore/nn/layer/combined.py +1 -1
- mindspore/nn/layer/container.py +52 -6
- mindspore/nn/layer/conv.py +112 -43
- mindspore/nn/layer/dense.py +10 -9
- mindspore/nn/layer/embedding.py +36 -34
- mindspore/nn/layer/image.py +123 -27
- mindspore/nn/layer/math.py +108 -107
- mindspore/nn/layer/normalization.py +212 -366
- mindspore/nn/layer/padding.py +370 -42
- mindspore/nn/layer/pooling.py +1443 -219
- mindspore/nn/layer/rnn_cells.py +11 -16
- mindspore/nn/layer/rnns.py +38 -39
- mindspore/nn/layer/thor_layer.py +24 -25
- mindspore/nn/layer/timedistributed.py +5 -5
- mindspore/nn/layer/transformer.py +701 -0
- mindspore/nn/learning_rate_schedule.py +8 -8
- mindspore/nn/loss/__init__.py +9 -6
- mindspore/nn/loss/loss.py +678 -142
- mindspore/nn/metrics.py +53 -0
- mindspore/nn/optim/_dist_optimizer_registry.py +2 -2
- mindspore/nn/optim/ada_grad.py +8 -8
- mindspore/nn/optim/adadelta.py +2 -3
- mindspore/nn/optim/adafactor.py +18 -14
- mindspore/nn/optim/adam.py +429 -87
- mindspore/nn/optim/adamax.py +5 -6
- mindspore/nn/optim/adasum.py +10 -8
- mindspore/nn/optim/asgd.py +7 -7
- mindspore/nn/optim/ftrl.py +81 -11
- mindspore/nn/optim/lamb.py +7 -8
- mindspore/nn/optim/lars.py +4 -4
- mindspore/nn/optim/lazyadam.py +82 -7
- mindspore/nn/optim/momentum.py +8 -7
- mindspore/nn/optim/optimizer.py +19 -10
- mindspore/nn/optim/proximal_ada_grad.py +6 -5
- mindspore/nn/optim/rmsprop.py +3 -3
- mindspore/nn/optim/rprop.py +20 -16
- mindspore/nn/optim/sgd.py +21 -15
- mindspore/nn/optim/thor.py +23 -21
- mindspore/nn/probability/__init__.py +0 -2
- mindspore/nn/probability/bijector/bijector.py +7 -6
- mindspore/nn/probability/bijector/invert.py +4 -2
- mindspore/nn/probability/bijector/softplus.py +2 -2
- mindspore/nn/probability/bnn_layers/dense_variational.py +1 -1
- mindspore/nn/probability/bnn_layers/layer_distribution.py +2 -2
- mindspore/nn/probability/distribution/__init__.py +6 -0
- mindspore/nn/probability/distribution/_utils/custom_ops.py +3 -2
- mindspore/nn/probability/distribution/_utils/utils.py +11 -17
- mindspore/nn/probability/distribution/bernoulli.py +6 -6
- mindspore/nn/probability/distribution/beta.py +1 -1
- mindspore/nn/probability/distribution/categorical.py +9 -9
- mindspore/nn/probability/distribution/cauchy.py +8 -8
- mindspore/nn/probability/distribution/distribution.py +12 -6
- mindspore/nn/probability/distribution/exponential.py +5 -5
- mindspore/nn/probability/distribution/gamma.py +3 -3
- mindspore/nn/probability/distribution/geometric.py +6 -5
- mindspore/nn/probability/distribution/gumbel.py +5 -5
- mindspore/nn/probability/distribution/half_normal.py +133 -0
- mindspore/nn/probability/distribution/laplace.py +128 -0
- mindspore/nn/probability/distribution/log_normal.py +0 -1
- mindspore/nn/probability/distribution/logistic.py +4 -5
- mindspore/nn/probability/distribution/normal.py +11 -15
- mindspore/nn/probability/distribution/poisson.py +6 -2
- mindspore/nn/probability/distribution/student_t.py +150 -0
- mindspore/nn/probability/distribution/transformed_distribution.py +4 -4
- mindspore/nn/probability/distribution/uniform.py +5 -5
- mindspore/nn/reinforcement/_tensors_queue.py +3 -3
- mindspore/nn/reinforcement/tensor_array.py +2 -2
- mindspore/nn/sparse/sparse.py +8 -1
- mindspore/nn/wrap/cell_wrapper.py +55 -27
- mindspore/nn/wrap/grad_reducer.py +20 -11
- mindspore/nn/wrap/loss_scale.py +47 -30
- mindspore/numpy/array_creations.py +33 -22
- mindspore/numpy/array_ops.py +46 -42
- mindspore/numpy/logic_ops.py +6 -27
- mindspore/numpy/math_ops.py +26 -19
- mindspore/numpy/utils.py +1 -8
- mindspore/numpy/utils_const.py +112 -62
- mindspore/opencv_core452.dll +0 -0
- mindspore/opencv_imgcodecs452.dll +0 -0
- mindspore/opencv_imgproc452.dll +0 -0
- mindspore/ops/__init__.py +6 -3
- mindspore/ops/_constants.py +0 -6
- mindspore/ops/_grad/__init__.py +2 -1
- mindspore/ops/_grad/grad_array_ops.py +209 -152
- mindspore/ops/_grad/grad_base.py +55 -17
- mindspore/ops/_grad/grad_clip_ops.py +11 -3
- mindspore/ops/_grad/grad_comm_ops.py +58 -47
- mindspore/ops/_grad/grad_implementations.py +21 -61
- mindspore/ops/_grad/grad_inner_ops.py +48 -6
- mindspore/ops/_grad/grad_math_ops.py +306 -161
- mindspore/ops/_grad/grad_nn_ops.py +192 -181
- mindspore/ops/_grad/grad_other_ops.py +1 -1
- mindspore/ops/_grad/grad_quant_ops.py +5 -5
- mindspore/ops/_grad/grad_sequence_ops.py +296 -0
- mindspore/ops/_grad/grad_sparse.py +15 -9
- mindspore/ops/_grad_experimental/__init__.py +1 -0
- mindspore/ops/_grad_experimental/grad_array_ops.py +441 -55
- mindspore/ops/_grad_experimental/grad_image_ops.py +25 -7
- mindspore/ops/_grad_experimental/grad_inner_ops.py +3 -44
- mindspore/ops/_grad_experimental/grad_linalg_ops.py +16 -21
- mindspore/ops/_grad_experimental/grad_math_ops.py +979 -49
- mindspore/ops/_grad_experimental/grad_nn_ops.py +78 -8
- mindspore/ops/_grad_experimental/grad_scalar_ops.py +112 -0
- mindspore/ops/_grad_experimental/grad_sparse_ops.py +197 -13
- mindspore/ops/_op_impl/__init__.py +3 -3
- mindspore/ops/_op_impl/_custom_op/__init__.py +0 -1
- mindspore/ops/_op_impl/_custom_op/_basic.py +0 -1
- mindspore/ops/_op_impl/_custom_op/batch_matmul_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/batchnorm_fold.py +4 -2
- mindspore/ops/_op_impl/_custom_op/batchnorm_fold2.py +2 -2
- mindspore/ops/_op_impl/_custom_op/batchnorm_fold2_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/batchnorm_fold2_grad_reduce.py +5 -5
- mindspore/ops/_op_impl/_custom_op/batchnorm_fold_grad.py +3 -3
- mindspore/ops/_op_impl/_custom_op/cholesky_trsm_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/correction_mul.py +3 -3
- mindspore/ops/_op_impl/_custom_op/correction_mul_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/dsd_back_impl.py +4 -8
- mindspore/ops/_op_impl/_custom_op/dsd_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perchannel.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perchannel_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perchannel_grad_reduce.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perlayer.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perlayer_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perlayer_grad_reduce.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_quant_perchannel.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_quant_perchannel_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_quant_perlayer.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_quant_perlayer_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fused_abs_max1_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/img2col_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/matmul_cube_dense_left_impl.py +2 -2
- mindspore/ops/_op_impl/_custom_op/matmul_cube_dense_right_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/matmul_cube_fracz_left_cast_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/matmul_cube_fracz_right_mul_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/matmul_cube_impl.py +2 -2
- mindspore/ops/_op_impl/_custom_op/matmul_dds_grad_impl.py +0 -1
- mindspore/ops/_op_impl/_custom_op/matmul_dds_impl.py +0 -1
- mindspore/ops/_op_impl/_custom_op/matrix_combine_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/minmax_update_perchannel.py +2 -2
- mindspore/ops/_op_impl/_custom_op/minmax_update_perlayer.py +2 -2
- mindspore/ops/_op_impl/_custom_op/transpose02314_impl.py +1 -1
- mindspore/ops/_op_impl/aicpu/__init__.py +238 -3
- mindspore/ops/_op_impl/aicpu/abs.py +36 -0
- mindspore/ops/_op_impl/aicpu/adaptive_avg_pool_2d.py +34 -0
- mindspore/ops/_op_impl/aicpu/adaptive_avg_pool_2d_grad.py +34 -0
- mindspore/ops/_op_impl/aicpu/adaptive_avg_pool_3d.py +39 -0
- mindspore/ops/_op_impl/aicpu/adaptive_avg_pool_3d_grad.py +39 -0
- mindspore/ops/_op_impl/aicpu/adaptive_max_pool_2d_grad.py +37 -0
- mindspore/ops/_op_impl/aicpu/adaptive_max_pool_3d.py +42 -0
- mindspore/ops/_op_impl/aicpu/adaptive_max_pool_3d_grad.py +152 -0
- mindspore/ops/_op_impl/aicpu/add.py +43 -0
- mindspore/ops/_op_impl/aicpu/addcdiv.py +0 -32
- mindspore/ops/_op_impl/aicpu/addcmul.py +0 -84
- mindspore/ops/_op_impl/aicpu/affine_grid_grad.py +35 -0
- mindspore/ops/_op_impl/aicpu/arg_max.py +75 -0
- mindspore/ops/_op_impl/aicpu/arg_min.py +75 -0
- mindspore/ops/_op_impl/aicpu/argmin_with_value.py +43 -0
- mindspore/ops/_op_impl/aicpu/batch_matmul.py +43 -0
- mindspore/ops/_op_impl/aicpu/batch_norm_grad_grad.py +49 -0
- mindspore/ops/_op_impl/aicpu/bernoulli.py +48 -0
- mindspore/ops/_op_impl/aicpu/bessel_i0.py +31 -0
- mindspore/ops/_op_impl/aicpu/bias_add.py +44 -0
- mindspore/ops/_op_impl/aicpu/bias_add_grad.py +43 -0
- mindspore/ops/_op_impl/aicpu/bincount.py +33 -0
- mindspore/{nn/probability/infer/variational/__init__.py → ops/_op_impl/aicpu/cauchy.py} +17 -10
- mindspore/ops/_op_impl/aicpu/channel_shuffle.py +40 -0
- mindspore/ops/_op_impl/aicpu/cholesky.py +1 -1
- mindspore/ops/_op_impl/{cpu/bias_add.py → aicpu/choleskygrad.py} +9 -7
- mindspore/ops/_op_impl/aicpu/combined_non_max_suppression.py +42 -0
- mindspore/ops/_op_impl/aicpu/concat_offset.py +42 -0
- mindspore/ops/_op_impl/aicpu/concat_offset_v1.py +31 -0
- mindspore/ops/_op_impl/aicpu/conj.py +11 -0
- mindspore/ops/_op_impl/aicpu/crop_and_resize_grad_image.py +38 -0
- mindspore/ops/_op_impl/aicpu/cumulative_logsumexp.py +36 -0
- mindspore/ops/_op_impl/aicpu/deformable_offsets.py +38 -0
- mindspore/ops/_op_impl/aicpu/deformable_offsets_grad.py +2 -2
- mindspore/ops/_op_impl/aicpu/dense_to_sparse_set_operation.py +48 -0
- mindspore/ops/_op_impl/aicpu/diag.py +36 -0
- mindspore/ops/_op_impl/aicpu/diag_part.py +36 -0
- mindspore/ops/_op_impl/aicpu/diagonal.py +35 -0
- mindspore/ops/_op_impl/{cpu/bias_add_grad.py → aicpu/digamma.py} +9 -7
- mindspore/ops/_op_impl/aicpu/eig.py +35 -0
- mindspore/ops/_op_impl/aicpu/fft_with_size.py +41 -0
- mindspore/ops/_op_impl/aicpu/flatten.py +1 -0
- mindspore/ops/_op_impl/aicpu/fmax.py +36 -0
- mindspore/ops/_op_impl/aicpu/fmin.py +37 -0
- mindspore/ops/_op_impl/aicpu/fractional_max_pool3d_with_fixed_ksize.py +1 -1
- mindspore/ops/_op_impl/aicpu/fse_decode.py +43 -0
- mindspore/ops/_op_impl/aicpu/glu.py +33 -0
- mindspore/ops/_op_impl/aicpu/glu_grad.py +34 -0
- mindspore/ops/_op_impl/aicpu/greater.py +41 -0
- mindspore/ops/_op_impl/aicpu/greater_equal.py +41 -0
- mindspore/ops/_op_impl/aicpu/index_put.py +50 -0
- mindspore/ops/_op_impl/{tbe/scatter_add_ds.py → aicpu/inplace_index_add.py} +17 -21
- mindspore/ops/_op_impl/aicpu/instance_norm_v2.py +41 -0
- mindspore/ops/_op_impl/aicpu/instance_norm_v2_grad.py +44 -0
- mindspore/ops/_op_impl/aicpu/layer_norm_grad_grad.py +47 -0
- mindspore/ops/_op_impl/aicpu/less.py +41 -0
- mindspore/ops/_op_impl/aicpu/less_equal.py +41 -0
- mindspore/ops/_op_impl/aicpu/lgamma.py +32 -0
- mindspore/ops/_op_impl/aicpu/log_normal_reverse.py +33 -0
- mindspore/ops/_op_impl/aicpu/logit.py +33 -0
- mindspore/ops/_op_impl/aicpu/logit_grad.py +34 -0
- mindspore/ops/_op_impl/aicpu/masked_fill.py +42 -0
- mindspore/ops/_op_impl/aicpu/masked_scatter.py +39 -0
- mindspore/ops/_op_impl/aicpu/matmul.py +39 -0
- mindspore/ops/_op_impl/aicpu/matrix_logarithm.py +31 -0
- mindspore/ops/_op_impl/aicpu/matrix_power.py +32 -0
- mindspore/ops/_op_impl/aicpu/matrix_solve_ls.py +36 -0
- mindspore/ops/_op_impl/aicpu/matrix_triangular_solve.py +36 -0
- mindspore/ops/_op_impl/aicpu/mirror_pad.py +2 -0
- mindspore/ops/_op_impl/aicpu/mirror_pad_grad.py +0 -4
- mindspore/ops/_op_impl/aicpu/mul.py +3 -1
- mindspore/ops/_op_impl/aicpu/multinomial.py +14 -6
- mindspore/ops/_op_impl/aicpu/multinomial_with_replacement.py +35 -0
- mindspore/ops/_op_impl/aicpu/nan_to_num.py +34 -0
- mindspore/ops/_op_impl/aicpu/nllloss.py +38 -0
- mindspore/ops/_op_impl/aicpu/nllloss_grad.py +39 -0
- mindspore/ops/_op_impl/aicpu/ones_like.py +0 -2
- mindspore/ops/_op_impl/aicpu/polar.py +32 -0
- mindspore/ops/_op_impl/aicpu/polygamma.py +34 -0
- mindspore/ops/_op_impl/aicpu/qr.py +36 -0
- mindspore/ops/_op_impl/aicpu/quant_dtype_cast.py +40 -0
- mindspore/ops/_op_impl/aicpu/quantile.py +35 -0
- mindspore/ops/_op_impl/aicpu/ragged_tensor_to_sparse.py +73 -0
- mindspore/ops/_op_impl/aicpu/ragged_tensor_to_tensor.py +74 -0
- mindspore/ops/_op_impl/aicpu/random_shuffle.py +3 -0
- mindspore/ops/_op_impl/aicpu/randperm_v2.py +41 -0
- mindspore/ops/_op_impl/aicpu/range.py +36 -0
- mindspore/ops/_op_impl/aicpu/reciprocal.py +34 -0
- mindspore/ops/_op_impl/aicpu/reciprocal_grad.py +35 -0
- mindspore/ops/_op_impl/aicpu/reduce_sum.py +57 -0
- mindspore/ops/_op_impl/aicpu/resize_bicubic.py +2 -8
- mindspore/ops/_op_impl/aicpu/resize_bicubic_grad.py +1 -1
- mindspore/ops/_op_impl/aicpu/resize_v2.py +68 -0
- mindspore/ops/_op_impl/aicpu/resize_v2_grad.py +68 -0
- mindspore/ops/_op_impl/aicpu/scatter_elements.py +4 -0
- mindspore/ops/_op_impl/aicpu/scatter_nd_update.py +2 -0
- mindspore/ops/_op_impl/aicpu/search_sorted.py +12 -6
- mindspore/ops/_op_impl/aicpu/self_adjoint_eig.py +34 -0
- mindspore/ops/_op_impl/aicpu/sequence_add.py +34 -0
- mindspore/ops/_op_impl/aicpu/sequence_add_offset.py +34 -0
- mindspore/ops/_op_impl/aicpu/sequence_addn.py +38 -0
- mindspore/ops/_op_impl/aicpu/slice_grad.py +76 -0
- mindspore/ops/_op_impl/aicpu/smooth_l1_loss.py +35 -0
- mindspore/ops/_op_impl/aicpu/smooth_l1_loss_grad.py +37 -0
- mindspore/ops/_op_impl/aicpu/sort.py +39 -0
- mindspore/ops/_op_impl/aicpu/sparse_apply_adagrad_da.py +0 -24
- mindspore/ops/_op_impl/aicpu/sparse_cross.py +42 -0
- mindspore/ops/_op_impl/aicpu/sparse_fill_empty_rows.py +63 -0
- mindspore/ops/_op_impl/aicpu/sparse_fill_empty_rows_grad.py +45 -0
- mindspore/ops/_op_impl/aicpu/sparse_matrix_mat_mul.py +56 -0
- mindspore/ops/_op_impl/{tbe/slice_ds.py → aicpu/sparse_segment_sum.py} +16 -24
- mindspore/ops/_op_impl/aicpu/sparse_segment_sum_with_num_segments.py +68 -0
- mindspore/ops/_op_impl/aicpu/sparse_slice.py +63 -0
- mindspore/ops/_op_impl/aicpu/sparse_slice_grad.py +61 -0
- mindspore/ops/_op_impl/aicpu/squared_difference.py +2 -0
- mindspore/ops/_op_impl/aicpu/strided_slice_v2.py +93 -0
- mindspore/ops/_op_impl/aicpu/strided_slice_v2_grad.py +66 -0
- mindspore/ops/_op_impl/aicpu/tensor_scatter_update.py +59 -0
- mindspore/ops/_op_impl/{tbe/gather_v2.py → aicpu/tile.py} +24 -24
- mindspore/ops/_op_impl/aicpu/tridiagonal_solve.py +35 -0
- mindspore/ops/_op_impl/aicpu/tril_indices.py +34 -0
- mindspore/ops/_op_impl/aicpu/triu_indices.py +34 -0
- mindspore/ops/_op_impl/aicpu/uniform.py +34 -0
- mindspore/ops/_op_impl/aicpu/uniform_candidate_sampler.py +1 -0
- mindspore/ops/_op_impl/aicpu/unique_consecutive.py +10 -2
- mindspore/ops/_op_impl/cpu/__init__.py +1 -2
- mindspore/ops/_op_impl/cpu/dynamic_shape.py +5 -1
- mindspore/ops/_op_impl/cpu/maximum_grad.py +2 -0
- mindspore/{compression/common/__init__.py → ops/_op_impl/cpu/pyexecute.py} +13 -8
- mindspore/ops/_op_impl/cpu/reduce_sum.py +8 -0
- mindspore/ops/_op_impl/cpu/sparse_slice.py +62 -0
- mindspore/ops/_op_impl/cpu/sparse_slice_grad.py +60 -0
- mindspore/ops/_op_impl/cpu/tensor_shape.py +5 -1
- mindspore/ops/_op_impl/tbe/__init__.py +27 -608
- mindspore/ops/_op_impl/tbe/addcdiv_ds.py +42 -0
- mindspore/ops/_op_impl/tbe/addcmul_ds.py +44 -0
- mindspore/ops/_op_impl/tbe/assign_add_ds.py +1 -0
- mindspore/ops/_op_impl/tbe/atomic_addr_clean.py +1 -1
- mindspore/ops/_op_impl/tbe/avg_pool_3d_grad.py +1 -1
- mindspore/ops/_op_impl/tbe/basic_lstm_cell_c_state_grad_v2.py +0 -1
- mindspore/ops/_op_impl/tbe/batch_to_space.py +1 -1
- mindspore/ops/_op_impl/tbe/batch_to_space_nd.py +1 -1
- mindspore/ops/_op_impl/tbe/batch_to_space_nd_v2.py +41 -0
- mindspore/ops/_op_impl/tbe/bce_with_logits_loss.py +1 -0
- mindspore/ops/_op_impl/tbe/bias_add_grad.py +2 -0
- mindspore/ops/_op_impl/tbe/bn_infer_grad.py +4 -2
- mindspore/ops/_op_impl/tbe/bn_infer_grad_ds.py +40 -0
- mindspore/ops/_op_impl/tbe/bn_training_update.py +0 -1
- mindspore/ops/_op_impl/tbe/bn_training_update_ds.py +0 -1
- mindspore/ops/_op_impl/tbe/broadcast_to_ds.py +6 -4
- mindspore/ops/_op_impl/tbe/cast.py +0 -2
- mindspore/ops/_op_impl/tbe/cast_ds.py +3 -3
- mindspore/ops/_op_impl/tbe/ctc_loss_v2.py +0 -2
- mindspore/ops/_op_impl/tbe/ctc_loss_v2_grad.py +0 -2
- mindspore/ops/_op_impl/tbe/data_format_dim_map_ds.py +1 -0
- mindspore/ops/_op_impl/tbe/deformable_offsets.py +1 -0
- mindspore/ops/_op_impl/tbe/depthwise_conv2d.py +1 -1
- mindspore/ops/_op_impl/tbe/dynamic_atomic_addr_clean.py +1 -1
- mindspore/ops/_op_impl/tbe/gather_nd.py +1 -0
- mindspore/ops/_op_impl/tbe/greater.py +2 -0
- mindspore/ops/_op_impl/tbe/{index_add.py → inplace_index_add.py} +3 -6
- mindspore/ops/_op_impl/tbe/layer_norm_beta_gamma_backprop_v2.py +0 -1
- mindspore/ops/_op_impl/tbe/npu_clear_float_status_v2.py +35 -0
- mindspore/ops/_op_impl/tbe/npu_get_float_status_v2.py +35 -0
- mindspore/ops/_op_impl/tbe/one_hot_ds.py +0 -6
- mindspore/ops/_op_impl/tbe/{greater_ds.py → reduce_all_ds.py} +13 -16
- mindspore/ops/_op_impl/tbe/reduce_any_ds.py +39 -0
- mindspore/ops/_op_impl/tbe/roi_align_ds.py +44 -0
- mindspore/ops/_op_impl/tbe/roi_align_grad_ds.py +44 -0
- mindspore/ops/_op_impl/tbe/scatter_add.py +2 -0
- mindspore/ops/_op_impl/tbe/scatter_nd_add.py +2 -2
- mindspore/ops/_op_impl/tbe/slice.py +26 -15
- mindspore/ops/_op_impl/tbe/space_to_batch.py +1 -1
- mindspore/ops/_op_impl/tbe/space_to_batch_nd.py +1 -1
- mindspore/ops/_op_impl/tbe/strided_slice_grad_d.py +1 -0
- mindspore/ops/_op_impl/tbe/trans_data_ds.py +15 -5
- mindspore/ops/_op_impl/tbe/unsorted_segment_sum.py +1 -1
- mindspore/ops/_op_impl/tbe/unsorted_segment_sum_ds.py +2 -0
- mindspore/ops/_primitive_cache.py +3 -2
- mindspore/ops/_register_for_op.py +11 -0
- mindspore/ops/_utils/__init__.py +1 -1
- mindspore/ops/_utils/utils.py +20 -41
- mindspore/ops/_vmap/__init__.py +2 -2
- mindspore/ops/_vmap/vmap_array_ops.py +170 -78
- mindspore/ops/_vmap/vmap_base.py +24 -10
- mindspore/ops/_vmap/vmap_convolution_ops.py +7 -10
- mindspore/ops/_vmap/vmap_grad_math_ops.py +4 -4
- mindspore/ops/_vmap/vmap_grad_nn_ops.py +41 -9
- mindspore/ops/_vmap/vmap_image_ops.py +52 -0
- mindspore/ops/_vmap/vmap_math_ops.py +77 -6
- mindspore/ops/_vmap/vmap_nn_ops.py +78 -29
- mindspore/ops/_vmap/vmap_other_ops.py +3 -1
- mindspore/ops/_vmap/vmap_random_ops.py +55 -3
- mindspore/ops/_vmap/vmap_sparse_ops.py +1 -0
- mindspore/ops/bprop_mindir/AdaptiveAvgPool2D_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/AdaptiveMaxPool2D_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/ApproximateEqual_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/Argmax_bprop.mindir +13 -12
- mindspore/ops/bprop_mindir/Argmin_bprop.mindir +14 -13
- mindspore/ops/bprop_mindir/AssignSub_bprop.mindir +17 -18
- mindspore/ops/bprop_mindir/Assign_bprop.mindir +16 -16
- mindspore/ops/bprop_mindir/AvgPool3D_bprop.mindir +150 -0
- mindspore/ops/bprop_mindir/AvgPool_bprop.mindir +66 -0
- mindspore/ops/bprop_mindir/BCEWithLogitsLoss_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/BNTrainingReduce_bprop.mindir +13 -12
- mindspore/ops/bprop_mindir/BatchNormGrad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/BatchToSpaceND_bprop.mindir +28 -0
- mindspore/ops/bprop_mindir/BiasAddGrad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/BinaryCrossEntropy_bprop.mindir +33 -0
- mindspore/ops/bprop_mindir/BroadcastTo_bprop.mindir +306 -0
- mindspore/ops/bprop_mindir/Broadcast_bprop.mindir +12 -8
- mindspore/ops/bprop_mindir/CTCLoss_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Concat_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Conv2DBackpropFilter_bprop.mindir +240 -0
- mindspore/ops/bprop_mindir/Conv2DBackpropInput_bprop.mindir +247 -0
- mindspore/ops/bprop_mindir/Conv2DTranspose_bprop.mindir +247 -0
- mindspore/ops/bprop_mindir/Conv3DTranspose_bprop.mindir +315 -0
- mindspore/ops/bprop_mindir/Conv3D_bprop.mindir +278 -0
- mindspore/ops/bprop_mindir/DType_bprop.mindir +12 -12
- mindspore/ops/bprop_mindir/DeformableOffsets_bprop.mindir +58 -0
- mindspore/ops/bprop_mindir/Depend_bprop.mindir +12 -13
- mindspore/ops/bprop_mindir/DepthToSpace_bprop.mindir +23 -0
- mindspore/ops/bprop_mindir/DepthwiseConv2dNative_bprop.mindir +138 -0
- mindspore/ops/bprop_mindir/DiagPart_bprop.mindir +15 -0
- mindspore/ops/bprop_mindir/Dropout2D_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Dropout3D_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/DropoutDoMask_bprop.mindir +22 -24
- mindspore/ops/bprop_mindir/DropoutGenMask_bprop.mindir +16 -14
- mindspore/ops/bprop_mindir/DropoutGrad_bprop.mindir +27 -0
- mindspore/ops/bprop_mindir/Dropout_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/DynamicGRUV2_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/DynamicRNN_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/DynamicShape_bprop.mindir +12 -12
- mindspore/ops/bprop_mindir/Elu_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/EmbeddingLookup_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Equal_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/ExpandDims_bprop.mindir +58 -0
- mindspore/ops/bprop_mindir/FastGeLU_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/Flatten_bprop.mindir +54 -0
- mindspore/ops/bprop_mindir/FloorDiv_bprop.mindir +18 -15
- mindspore/ops/bprop_mindir/GatherD_bprop.mindir +26 -0
- mindspore/ops/bprop_mindir/GatherNd_bprop.mindir +57 -0
- mindspore/ops/bprop_mindir/Gather_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/GreaterEqual_bprop.mindir +17 -18
- mindspore/ops/bprop_mindir/Greater_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/HSigmoid_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/HSwish_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/IOU_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/InstanceNorm_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/IsFinite_bprop.mindir +13 -12
- mindspore/ops/bprop_mindir/IsInf_bprop.mindir +13 -10
- mindspore/ops/bprop_mindir/IsNan_bprop.mindir +14 -11
- mindspore/ops/bprop_mindir/KLDivLoss_bprop.mindir +126 -0
- mindspore/ops/bprop_mindir/L2Loss_bprop.mindir +15 -0
- mindspore/ops/bprop_mindir/L2Normalize_bprop.mindir +30 -0
- mindspore/ops/bprop_mindir/LRN_bprop.mindir +43 -0
- mindspore/ops/bprop_mindir/LayerNormGrad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/LessEqual_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/Less_bprop.mindir +17 -18
- mindspore/ops/bprop_mindir/LinSpace_bprop.mindir +22 -19
- mindspore/ops/bprop_mindir/Load_bprop.mindir +12 -13
- mindspore/ops/bprop_mindir/LogSoftmax_bprop.mindir +23 -0
- mindspore/ops/bprop_mindir/LogicalAnd_bprop.mindir +17 -18
- mindspore/ops/bprop_mindir/LogicalNot_bprop.mindir +14 -13
- mindspore/ops/bprop_mindir/MaskedSelect_bprop.mindir +21 -0
- mindspore/ops/bprop_mindir/MaxPool3DGradGrad_bprop.mindir +74 -0
- mindspore/ops/bprop_mindir/MaxPool3DGrad_bprop.mindir +74 -0
- mindspore/ops/bprop_mindir/MaxPool3D_bprop.mindir +75 -0
- mindspore/ops/bprop_mindir/MaxPoolGradGrad_bprop.mindir +65 -0
- mindspore/ops/bprop_mindir/MaxPoolWithArgmax_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Maximum_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Minimum_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/MirrorPad_bprop.mindir +27 -0
- mindspore/ops/bprop_mindir/Mish_bprop.mindir +35 -0
- mindspore/ops/bprop_mindir/MulNoNan_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/NLLLoss_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/NonZero_bprop.mindir +14 -0
- mindspore/ops/bprop_mindir/NotEqual_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/OneHot_bprop.mindir +25 -23
- mindspore/ops/bprop_mindir/OnesLike_bprop.mindir +13 -13
- mindspore/ops/bprop_mindir/PReLU_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Pad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Padding_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/RNNTLoss_bprop.mindir +29 -0
- mindspore/ops/bprop_mindir/ROIAlign_bprop.mindir +82 -0
- mindspore/ops/bprop_mindir/Range_bprop.mindir +21 -19
- mindspore/ops/bprop_mindir/Rank_bprop.mindir +11 -11
- mindspore/ops/bprop_mindir/ReLU6_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/ReLUV2_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/ReduceAll_bprop.mindir +18 -17
- mindspore/ops/bprop_mindir/ReduceAny_bprop.mindir +18 -17
- mindspore/ops/bprop_mindir/ReluGrad_bprop.mindir +19 -23
- mindspore/ops/bprop_mindir/Reshape_bprop.mindir +60 -0
- mindspore/ops/bprop_mindir/ResizeBilinear_bprop.mindir +29 -0
- mindspore/ops/bprop_mindir/ResizeNearestNeighbor_bprop.mindir +89 -0
- mindspore/ops/bprop_mindir/ReverseSequence_bprop.mindir +52 -0
- mindspore/ops/bprop_mindir/ReverseV2_bprop.mindir +22 -0
- mindspore/ops/bprop_mindir/Round_bprop.mindir +14 -13
- mindspore/ops/bprop_mindir/ScatterMax_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/ScatterMin_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/ScatterNdUpdate_bprop.mindir +22 -0
- mindspore/ops/bprop_mindir/ScatterNd_bprop.mindir +24 -0
- mindspore/ops/bprop_mindir/ScatterNonAliasingAdd_bprop.mindir +22 -0
- mindspore/ops/bprop_mindir/ScatterUpdate_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/SeLU_bprop.mindir +21 -0
- mindspore/ops/bprop_mindir/Select_bprop.mindir +30 -34
- mindspore/ops/bprop_mindir/Shape_bprop.mindir +12 -12
- mindspore/ops/bprop_mindir/SigmoidCrossEntropyWithLogits_bprop.mindir +21 -0
- mindspore/ops/bprop_mindir/SigmoidGrad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Sigmoid_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/Sign_bprop.mindir +13 -12
- mindspore/ops/bprop_mindir/Slice_bprop.mindir +26 -0
- mindspore/ops/bprop_mindir/SmoothL1Loss_bprop.mindir +36 -0
- mindspore/ops/bprop_mindir/SoftmaxCrossEntropyWithLogits_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Softplus_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/Softsign_bprop.mindir +33 -0
- mindspore/ops/bprop_mindir/Sort_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/SpaceToBatchND_bprop.mindir +28 -0
- mindspore/ops/bprop_mindir/SpaceToDepth_bprop.mindir +23 -0
- mindspore/ops/bprop_mindir/SparseGatherV2_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/SparseSoftmaxCrossEntropyWithLogits_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Split_bprop.mindir +22 -0
- mindspore/ops/bprop_mindir/Squeeze_bprop.mindir +54 -0
- mindspore/ops/bprop_mindir/StridedSliceGrad_bprop.mindir +95 -0
- mindspore/ops/bprop_mindir/StridedSlice_bprop.mindir +98 -0
- mindspore/ops/bprop_mindir/Switch_bprop.mindir +28 -32
- mindspore/ops/bprop_mindir/TanhGrad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Tanh_bprop.mindir +66 -0
- mindspore/ops/bprop_mindir/TensorScatterAdd_bprop.mindir +22 -0
- mindspore/ops/bprop_mindir/TensorScatterUpdate_bprop.mindir +29 -0
- mindspore/ops/bprop_mindir/TensorShape_bprop.mindir +14 -0
- mindspore/ops/bprop_mindir/Tile_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/TopK_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/TransShape_bprop.mindir +23 -0
- mindspore/ops/bprop_mindir/TruncateDiv_bprop.mindir +18 -15
- mindspore/ops/bprop_mindir/TupleGetItem_bprop.mindir +11 -13
- mindspore/ops/bprop_mindir/Unique_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/Unstack_bprop.mindir +22 -0
- mindspore/ops/bprop_mindir/UpsampleNearest3D_bprop.mindir +32 -0
- mindspore/ops/bprop_mindir/UpsampleTrilinear3D_bprop.mindir +38 -0
- mindspore/ops/bprop_mindir/ZerosLike_bprop.mindir +13 -12
- mindspore/ops/bprop_mindir/__init__.py +1 -4
- mindspore/ops/bprop_mindir/generate_mindir.py +32 -20
- mindspore/ops/composite/__init__.py +12 -13
- mindspore/ops/composite/base.py +261 -254
- mindspore/ops/composite/env_ops.py +41 -0
- mindspore/ops/composite/math_ops.py +197 -156
- mindspore/ops/composite/multitype_ops/_compile_utils.py +428 -176
- mindspore/ops/composite/multitype_ops/_constexpr_utils.py +188 -87
- mindspore/ops/composite/multitype_ops/add_impl.py +23 -1
- mindspore/ops/composite/multitype_ops/div_impl.py +3 -3
- mindspore/ops/composite/multitype_ops/equal_impl.py +1 -0
- mindspore/ops/composite/multitype_ops/floordiv_impl.py +1 -1
- mindspore/ops/composite/multitype_ops/getitem_impl.py +52 -5
- mindspore/ops/composite/multitype_ops/greater_equal_impl.py +31 -0
- mindspore/ops/composite/multitype_ops/greater_impl.py +31 -0
- mindspore/ops/composite/multitype_ops/in_impl.py +15 -3
- mindspore/ops/composite/multitype_ops/less_equal_impl.py +33 -2
- mindspore/ops/composite/multitype_ops/less_impl.py +33 -0
- mindspore/ops/composite/multitype_ops/logical_and_impl.py +2 -2
- mindspore/ops/composite/multitype_ops/logical_or_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/mod_impl.py +1 -1
- mindspore/ops/composite/multitype_ops/mul_impl.py +21 -7
- mindspore/ops/composite/multitype_ops/not_in_impl.py +15 -3
- mindspore/ops/composite/multitype_ops/ones_like_impl.py +2 -4
- mindspore/ops/composite/multitype_ops/pow_impl.py +1 -0
- mindspore/ops/composite/multitype_ops/setitem_impl.py +62 -70
- mindspore/ops/composite/multitype_ops/sub_impl.py +3 -3
- mindspore/ops/composite/multitype_ops/zeros_like_impl.py +41 -4
- mindspore/ops/function/__init__.py +323 -8
- mindspore/ops/function/array_func.py +3511 -780
- mindspore/ops/function/clip_func.py +329 -0
- mindspore/ops/function/debug_func.py +6 -6
- mindspore/ops/function/grad/__init__.py +5 -1
- mindspore/ops/function/grad/grad_func.py +736 -65
- mindspore/ops/function/image_func.py +270 -0
- mindspore/ops/function/linalg_func.py +268 -8
- mindspore/ops/function/math_func.py +8032 -3164
- mindspore/ops/function/nn_func.py +5619 -1855
- mindspore/ops/function/other_func.py +115 -0
- mindspore/ops/function/parameter_func.py +11 -10
- mindspore/ops/function/random_func.py +939 -77
- mindspore/ops/function/sparse_func.py +249 -84
- mindspore/ops/function/sparse_unary_func.py +2303 -0
- mindspore/ops/function/spectral_func.py +146 -0
- mindspore/ops/function/vmap_func.py +114 -0
- mindspore/ops/functional.py +182 -254
- mindspore/ops/op_info_register.py +79 -34
- mindspore/ops/operations/__init__.py +210 -118
- mindspore/ops/operations/_csr_ops.py +7 -7
- mindspore/ops/operations/_embedding_cache_ops.py +25 -15
- mindspore/ops/operations/_grad_ops.py +447 -322
- mindspore/ops/operations/_inner_ops.py +547 -176
- mindspore/ops/operations/_map_tensor_ops.py +112 -0
- mindspore/ops/operations/_ms_kernel.py +29 -27
- mindspore/ops/operations/_ocr_ops.py +11 -11
- mindspore/ops/operations/_opaque_predicate_registry.py +41 -0
- mindspore/ops/operations/_quant_ops.py +186 -101
- mindspore/ops/operations/_rl_inner_ops.py +122 -61
- mindspore/ops/operations/_scalar_ops.py +466 -0
- mindspore/ops/operations/_sequence_ops.py +1047 -0
- mindspore/ops/operations/_tensor_array.py +10 -11
- mindspore/ops/operations/_thor_ops.py +4 -4
- mindspore/ops/operations/array_ops.py +1428 -1226
- mindspore/ops/operations/comm_ops.py +180 -117
- mindspore/ops/operations/control_ops.py +4 -2
- mindspore/ops/operations/custom_ops.py +185 -98
- mindspore/ops/operations/debug_ops.py +92 -54
- mindspore/ops/operations/image_ops.py +406 -211
- mindspore/ops/operations/inner_ops.py +42 -53
- mindspore/ops/operations/linalg_ops.py +32 -29
- mindspore/ops/operations/math_ops.py +2076 -897
- mindspore/ops/operations/nn_ops.py +1282 -1252
- mindspore/ops/operations/other_ops.py +124 -278
- mindspore/ops/operations/random_ops.py +345 -178
- mindspore/ops/operations/rl_ops.py +8 -9
- mindspore/ops/operations/sparse_ops.py +502 -157
- mindspore/ops/operations/spectral_ops.py +107 -0
- mindspore/ops/primitive.py +192 -15
- mindspore/ops/vm_impl_registry.py +23 -2
- mindspore/parallel/__init__.py +6 -1
- mindspore/parallel/_auto_parallel_context.py +199 -92
- mindspore/parallel/_cell_wrapper.py +4 -2
- mindspore/parallel/_cost_model_context.py +3 -0
- mindspore/parallel/_dp_allreduce_fusion.py +2 -1
- mindspore/parallel/_offload_context.py +185 -0
- mindspore/parallel/_parallel_serialization.py +167 -28
- mindspore/parallel/_ps_context.py +9 -5
- mindspore/parallel/_recovery_context.py +1 -1
- mindspore/parallel/_tensor.py +9 -1
- mindspore/{nn/transformer → parallel/_transformer}/__init__.py +6 -6
- mindspore/{nn/transformer → parallel/_transformer}/layers.py +59 -37
- mindspore/{nn/transformer → parallel/_transformer}/loss.py +4 -7
- mindspore/{nn/transformer → parallel/_transformer}/moe.py +160 -35
- mindspore/{nn/transformer → parallel/_transformer}/op_parallel_config.py +3 -3
- mindspore/{nn/transformer → parallel/_transformer}/transformer.py +235 -196
- mindspore/parallel/_utils.py +47 -7
- mindspore/parallel/algo_parameter_config.py +5 -1
- mindspore/parallel/checkpoint_transform.py +329 -0
- mindspore/parallel/shard.py +229 -0
- mindspore/perf_msvcbuildinsights.dll +0 -0
- mindspore/pgodb140.dll +0 -0
- mindspore/pgort140.dll +0 -0
- mindspore/profiler/__init__.py +2 -1
- mindspore/profiler/common/util.py +4 -3
- mindspore/profiler/common/validator/validate_path.py +2 -2
- mindspore/profiler/envprofiling.py +249 -0
- mindspore/profiler/parser/aicpu_data_parser.py +38 -39
- mindspore/profiler/parser/ascend_timeline_generator.py +497 -0
- mindspore/profiler/parser/base_timeline_generator.py +471 -0
- mindspore/profiler/parser/cpu_gpu_timeline_generator.py +684 -0
- mindspore/profiler/parser/framework_parser.py +42 -16
- mindspore/profiler/parser/hccl_parser.py +158 -158
- mindspore/profiler/parser/hwts_log_parser.py +7 -6
- mindspore/profiler/parser/integrator.py +18 -1579
- mindspore/profiler/parser/minddata_analyzer.py +8 -8
- mindspore/profiler/parser/msadvisor_analyzer.py +14 -27
- mindspore/profiler/parser/msadvisor_parser.py +2 -4
- mindspore/profiler/parser/optime_parser.py +17 -18
- mindspore/profiler/parser/profiler_info.py +108 -0
- mindspore/profiler/parser/step_trace_parser.py +1 -1
- mindspore/profiler/profiling.py +396 -194
- mindspore/rewrite/__init__.py +6 -2
- mindspore/rewrite/api/node.py +51 -110
- mindspore/rewrite/api/node_type.py +10 -6
- mindspore/rewrite/api/pattern_engine.py +51 -7
- mindspore/rewrite/api/scoped_value.py +64 -53
- mindspore/rewrite/api/symbol_tree.py +108 -61
- mindspore/rewrite/api/tree_node_helper.py +2 -3
- mindspore/{compression/quant/__init__.py → rewrite/ast_creator_register.py} +20 -11
- mindspore/rewrite/ast_helpers/__init__.py +6 -3
- mindspore/rewrite/ast_helpers/ast_creator.py +115 -0
- mindspore/rewrite/ast_helpers/ast_finder.py +99 -1
- mindspore/rewrite/ast_helpers/ast_modifier.py +17 -4
- mindspore/rewrite/ast_helpers/ast_replacer.py +1 -1
- mindspore/rewrite/ast_transformers/__init__.py +0 -1
- mindspore/rewrite/ast_transformers/flatten_recursive_stmt.py +46 -5
- mindspore/rewrite/ast_transformers/remove_return_out_of_if.py +6 -3
- mindspore/rewrite/common/__init__.py +2 -0
- mindspore/rewrite/common/event.py +1 -1
- mindspore/rewrite/common/observable.py +1 -1
- mindspore/rewrite/common/observer.py +1 -1
- mindspore/rewrite/common/rewrite_elog.py +35 -0
- mindspore/rewrite/namer.py +2 -2
- mindspore/rewrite/namespace.py +14 -4
- mindspore/rewrite/node.py +161 -13
- mindspore/rewrite/parser.py +0 -1
- mindspore/rewrite/parser_register.py +0 -1
- mindspore/rewrite/parsers/arguments_parser.py +3 -2
- mindspore/rewrite/parsers/assign_parser.py +267 -67
- mindspore/rewrite/parsers/attribute_parser.py +56 -0
- mindspore/rewrite/parsers/class_def_parser.py +191 -108
- mindspore/rewrite/parsers/constant_parser.py +101 -0
- mindspore/rewrite/parsers/container_parser.py +88 -0
- mindspore/rewrite/parsers/for_parser.py +28 -15
- mindspore/rewrite/parsers/function_def_parser.py +21 -5
- mindspore/rewrite/parsers/if_parser.py +11 -28
- mindspore/rewrite/parsers/module_parser.py +9 -6
- mindspore/rewrite/parsers/return_parser.py +3 -2
- mindspore/rewrite/sparsify/__init__.py +0 -0
- mindspore/rewrite/sparsify/sparse_transformer.py +448 -0
- mindspore/rewrite/sparsify/sparsify.py +109 -0
- mindspore/rewrite/sparsify/utils.py +173 -0
- mindspore/rewrite/symbol_tree.py +322 -109
- mindspore/rewrite/symbol_tree_builder.py +45 -8
- mindspore/rewrite/symbol_tree_dumper.py +0 -1
- mindspore/rewrite/topological_manager.py +1 -2
- mindspore/run_check/_check_version.py +209 -112
- mindspore/run_check/run_check.py +2 -1
- mindspore/tbbmalloc.dll +0 -0
- mindspore/tinyxml2.dll +0 -0
- mindspore/train/__init__.py +6 -4
- mindspore/train/_utils.py +28 -5
- mindspore/train/amp.py +321 -50
- mindspore/train/callback/__init__.py +3 -1
- mindspore/train/callback/_backup_and_restore.py +120 -0
- mindspore/train/callback/_callback.py +8 -8
- mindspore/train/callback/_checkpoint.py +12 -9
- mindspore/train/callback/_early_stop.py +13 -7
- mindspore/train/callback/_history.py +8 -8
- mindspore/train/callback/_lambda_callback.py +6 -6
- mindspore/train/callback/_landscape.py +36 -38
- mindspore/train/callback/_loss_monitor.py +12 -6
- mindspore/train/callback/_lr_scheduler_callback.py +2 -4
- mindspore/train/callback/_on_request_exit.py +212 -0
- mindspore/train/callback/_reduce_lr_on_plateau.py +13 -7
- mindspore/train/callback/_summary_collector.py +27 -19
- mindspore/train/callback/_time_monitor.py +13 -7
- mindspore/train/checkpoint_pb2.py +68 -8
- mindspore/train/data_sink.py +122 -33
- mindspore/train/dataset_helper.py +28 -87
- mindspore/train/loss_scale_manager.py +4 -7
- mindspore/{nn → train}/metrics/__init__.py +20 -20
- mindspore/{nn → train}/metrics/accuracy.py +12 -10
- mindspore/{nn → train}/metrics/auc.py +4 -4
- mindspore/{nn → train}/metrics/bleu_score.py +4 -4
- mindspore/{nn → train}/metrics/confusion_matrix.py +10 -8
- mindspore/{nn → train}/metrics/cosine_similarity.py +4 -4
- mindspore/{nn → train}/metrics/dice.py +6 -5
- mindspore/{nn → train}/metrics/error.py +7 -5
- mindspore/{nn → train}/metrics/fbeta.py +9 -7
- mindspore/{nn → train}/metrics/hausdorff_distance.py +8 -6
- mindspore/{nn → train}/metrics/loss.py +4 -3
- mindspore/{nn → train}/metrics/mean_surface_distance.py +6 -5
- mindspore/{nn → train}/metrics/metric.py +6 -5
- mindspore/{nn → train}/metrics/occlusion_sensitivity.py +4 -3
- mindspore/{nn → train}/metrics/perplexity.py +5 -4
- mindspore/{nn → train}/metrics/precision.py +5 -4
- mindspore/{nn → train}/metrics/recall.py +5 -4
- mindspore/{nn → train}/metrics/roc.py +7 -6
- mindspore/{nn → train}/metrics/root_mean_square_surface_distance.py +6 -5
- mindspore/{nn → train}/metrics/topk.py +7 -5
- mindspore/train/mind_ir_pb2.py +339 -32
- mindspore/train/model.py +113 -84
- mindspore/train/serialization.py +547 -167
- mindspore/train/summary/_summary_adapter.py +1 -1
- mindspore/train/summary/summary_record.py +43 -12
- mindspore/train/train_thor/convert_utils.py +7 -1
- mindspore/train/train_thor/dataset_helper.py +3 -3
- mindspore/train/train_thor/model_thor.py +0 -4
- mindspore/turbojpeg.dll +0 -0
- mindspore/vcmeta.dll +0 -0
- mindspore/vcruntime140.dll +0 -0
- mindspore/vcruntime140_1.dll +0 -0
- mindspore/version.py +1 -1
- {mindspore-1.10.0.dist-info → mindspore-2.0.0rc1.dist-info}/METADATA +4 -3
- {mindspore-1.10.0.dist-info → mindspore-2.0.0rc1.dist-info}/RECORD +901 -660
- mindspore/compression/common/constant.py +0 -124
- mindspore/compression/export/__init__.py +0 -19
- mindspore/compression/export/quant_export.py +0 -514
- mindspore/compression/quant/qat.py +0 -636
- mindspore/compression/quant/quant_utils.py +0 -462
- mindspore/compression/quant/quantizer.py +0 -68
- mindspore/libatomic-1.dll +0 -0
- mindspore/libgcc_s_seh-1.dll +0 -0
- mindspore/libgfortran-4.dll +0 -0
- mindspore/libgomp-1.dll +0 -0
- mindspore/libjpeg-62.dll +0 -0
- mindspore/libmindspore.dll +0 -0
- mindspore/libmindspore_common.dll +0 -0
- mindspore/libmindspore_core.dll +0 -0
- mindspore/libmindspore_glog.dll +0 -0
- mindspore/libnnacl.dll +0 -0
- mindspore/libopencv_core452.dll +0 -0
- mindspore/libopencv_imgcodecs452.dll +0 -0
- mindspore/libopencv_imgproc452.dll +0 -0
- mindspore/libquadmath-0.dll +0 -0
- mindspore/libsqlite3.dll +0 -0
- mindspore/libssp-0.dll +0 -0
- mindspore/libstdc++-6.dll +0 -0
- mindspore/libtinyxml2.dll +0 -0
- mindspore/libturbojpeg.dll +0 -0
- mindspore/libwinpthread-1.dll +0 -0
- mindspore/nn/layer/quant.py +0 -1868
- mindspore/nn/layer/rnn_utils.py +0 -90
- mindspore/nn/probability/dpn/__init__.py +0 -22
- mindspore/nn/probability/dpn/vae/__init__.py +0 -25
- mindspore/nn/probability/dpn/vae/cvae.py +0 -138
- mindspore/nn/probability/dpn/vae/vae.py +0 -122
- mindspore/nn/probability/infer/__init__.py +0 -22
- mindspore/nn/probability/infer/variational/elbo.py +0 -70
- mindspore/nn/probability/infer/variational/svi.py +0 -84
- mindspore/nn/probability/toolbox/__init__.py +0 -22
- mindspore/nn/probability/toolbox/anomaly_detection.py +0 -99
- mindspore/nn/probability/toolbox/uncertainty_evaluation.py +0 -363
- mindspore/nn/probability/transforms/__init__.py +0 -22
- mindspore/nn/probability/transforms/transform_bnn.py +0 -262
- mindspore/nn/probability/zhusuan/__init__.py +0 -18
- mindspore/nn/probability/zhusuan/framework/__init__.py +0 -18
- mindspore/nn/probability/zhusuan/framework/bn.py +0 -95
- mindspore/nn/probability/zhusuan/variational/__init__.py +0 -18
- mindspore/nn/probability/zhusuan/variational/elbo.py +0 -46
- mindspore/ops/_op_impl/tbe/bias_add_grad_ds.py +0 -52
- mindspore/ops/_op_impl/tbe/scatter_nd_add_ds.py +0 -43
- mindspore/ops/bprop_mindir/AssignAdd_bprop.mindir +0 -20
- mindspore/ops/bprop_mindir/Identity_bprop.mindir +0 -9
- mindspore/ops/bprop_mindir/LogicalOr_bprop.mindir +0 -20
- mindspore/ops/bprop_mindir/ReLU_bprop.mindir +0 -16
- mindspore/ops/bprop_mindir/UpdateState_bprop.mindir +0 -17
- mindspore/ops/bprop_mindir/stop_gradient_bprop.mindir +0 -12
- mindspore/ops/composite/array_ops.py +0 -210
- mindspore/ops/composite/clip_ops.py +0 -238
- mindspore/ops/composite/random_ops.py +0 -426
- mindspore/ops/composite/vmap_ops.py +0 -38
- mindspore/ops/operations/sponge_ops.py +0 -3531
- mindspore/ops/operations/sponge_update_ops.py +0 -2546
- mindspore/parallel/nn/__init__.py +0 -42
- mindspore/parallel/nn/loss.py +0 -22
- mindspore/parallel/nn/moe.py +0 -21
- mindspore/parallel/nn/op_parallel_config.py +0 -22
- mindspore/parallel/nn/transformer.py +0 -31
- mindspore/run_check/_check_deps_version.py +0 -84
- {mindspore-1.10.0.dist-info → mindspore-2.0.0rc1.dist-info}/WHEEL +0 -0
- {mindspore-1.10.0.dist-info → mindspore-2.0.0rc1.dist-info}/entry_points.txt +0 -0
- {mindspore-1.10.0.dist-info → mindspore-2.0.0rc1.dist-info}/top_level.txt +0 -0
|
@@ -1,124 +0,0 @@
|
|
|
1
|
-
# Copyright 2020 Huawei Technologies Co., Ltd
|
|
2
|
-
#
|
|
3
|
-
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
4
|
-
# you may not use this file except in compliance with the License.
|
|
5
|
-
# You may obtain a copy of the License at
|
|
6
|
-
#
|
|
7
|
-
# http://www.apache.org/licenses/LICENSE-2.0
|
|
8
|
-
#
|
|
9
|
-
# Unless required by applicable law or agreed to in writing, software
|
|
10
|
-
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
11
|
-
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
12
|
-
# See the License for the specific language governing permissions and
|
|
13
|
-
# limitations under the License.
|
|
14
|
-
# ============================================================================
|
|
15
|
-
"""
|
|
16
|
-
Note:
|
|
17
|
-
Constant module for compression. This is interface that is subject to change or deletion.
|
|
18
|
-
"""
|
|
19
|
-
from __future__ import absolute_import
|
|
20
|
-
|
|
21
|
-
import enum
|
|
22
|
-
import re
|
|
23
|
-
from types import DynamicClassAttribute
|
|
24
|
-
|
|
25
|
-
|
|
26
|
-
__all__ = ["QuantDtype"]
|
|
27
|
-
|
|
28
|
-
|
|
29
|
-
@enum.unique
|
|
30
|
-
class QuantDtype(enum.Enum):
|
|
31
|
-
"""
|
|
32
|
-
An enum for quant datatype, contains `INT2` ~ `INT8`, `UINT2` ~ `UINT8`.
|
|
33
|
-
"""
|
|
34
|
-
INT2 = "INT2"
|
|
35
|
-
INT3 = "INT3"
|
|
36
|
-
INT4 = "INT4"
|
|
37
|
-
INT5 = "INT5"
|
|
38
|
-
INT6 = "INT6"
|
|
39
|
-
INT7 = "INT7"
|
|
40
|
-
INT8 = "INT8"
|
|
41
|
-
|
|
42
|
-
UINT2 = "UINT2"
|
|
43
|
-
UINT3 = "UINT3"
|
|
44
|
-
UINT4 = "UINT4"
|
|
45
|
-
UINT5 = "UINT5"
|
|
46
|
-
UINT6 = "UINT6"
|
|
47
|
-
UINT7 = "UINT7"
|
|
48
|
-
UINT8 = "UINT8"
|
|
49
|
-
|
|
50
|
-
def __str__(self):
|
|
51
|
-
return f"{self.name}"
|
|
52
|
-
|
|
53
|
-
@staticmethod
|
|
54
|
-
def is_signed(dtype):
|
|
55
|
-
"""
|
|
56
|
-
Get whether the quant datatype is signed.
|
|
57
|
-
|
|
58
|
-
Args:
|
|
59
|
-
dtype (QuantDtype): quant datatype.
|
|
60
|
-
|
|
61
|
-
Returns:
|
|
62
|
-
bool, whether the input quant datatype is signed.
|
|
63
|
-
|
|
64
|
-
Examples:
|
|
65
|
-
>>> quant_dtype = QuantDtype.INT8
|
|
66
|
-
>>> is_signed = QuantDtype.is_signed(quant_dtype)
|
|
67
|
-
"""
|
|
68
|
-
return dtype in [QuantDtype.INT2, QuantDtype.INT3, QuantDtype.INT4, QuantDtype.INT5,
|
|
69
|
-
QuantDtype.INT6, QuantDtype.INT7, QuantDtype.INT8]
|
|
70
|
-
|
|
71
|
-
@staticmethod
|
|
72
|
-
def switch_signed(dtype):
|
|
73
|
-
"""
|
|
74
|
-
Switch the signed state of the input quant datatype.
|
|
75
|
-
|
|
76
|
-
Args:
|
|
77
|
-
dtype (QuantDtype): quant datatype.
|
|
78
|
-
|
|
79
|
-
Returns:
|
|
80
|
-
QuantDtype, quant datatype with opposite signed state as the input.
|
|
81
|
-
|
|
82
|
-
Examples:
|
|
83
|
-
>>> quant_dtype = QuantDtype.INT8
|
|
84
|
-
>>> quant_dtype = QuantDtype.switch_signed(quant_dtype)
|
|
85
|
-
"""
|
|
86
|
-
type_map = {
|
|
87
|
-
QuantDtype.INT2: QuantDtype.UINT2,
|
|
88
|
-
QuantDtype.INT3: QuantDtype.UINT3,
|
|
89
|
-
QuantDtype.INT4: QuantDtype.UINT4,
|
|
90
|
-
QuantDtype.INT5: QuantDtype.UINT5,
|
|
91
|
-
QuantDtype.INT6: QuantDtype.UINT6,
|
|
92
|
-
QuantDtype.INT7: QuantDtype.UINT7,
|
|
93
|
-
QuantDtype.INT8: QuantDtype.UINT8,
|
|
94
|
-
QuantDtype.UINT2: QuantDtype.INT2,
|
|
95
|
-
QuantDtype.UINT3: QuantDtype.INT3,
|
|
96
|
-
QuantDtype.UINT4: QuantDtype.INT4,
|
|
97
|
-
QuantDtype.UINT5: QuantDtype.INT5,
|
|
98
|
-
QuantDtype.UINT6: QuantDtype.INT6,
|
|
99
|
-
QuantDtype.UINT7: QuantDtype.INT7,
|
|
100
|
-
QuantDtype.UINT8: QuantDtype.INT8
|
|
101
|
-
}
|
|
102
|
-
return type_map.get(dtype)
|
|
103
|
-
|
|
104
|
-
@DynamicClassAttribute
|
|
105
|
-
def _value(self):
|
|
106
|
-
"""The value of the Enum member."""
|
|
107
|
-
return int(re.search(r"(\d+)", self._value_).group(1))
|
|
108
|
-
|
|
109
|
-
@DynamicClassAttribute
|
|
110
|
-
def num_bits(self):
|
|
111
|
-
"""
|
|
112
|
-
Get the num bits of the QuantDtype member.
|
|
113
|
-
|
|
114
|
-
Returns:
|
|
115
|
-
int, the num bits of the QuantDtype member.
|
|
116
|
-
|
|
117
|
-
Examples:
|
|
118
|
-
>>> from mindspore.compression.common import QuantDtype
|
|
119
|
-
>>> quant_dtype = QuantDtype.INT8
|
|
120
|
-
>>> num_bits = quant_dtype.num_bits
|
|
121
|
-
>>> print(num_bits)
|
|
122
|
-
8
|
|
123
|
-
"""
|
|
124
|
-
return self._value
|
|
@@ -1,19 +0,0 @@
|
|
|
1
|
-
# Copyright 2020 Huawei Technologies Co., Ltd
|
|
2
|
-
#
|
|
3
|
-
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
4
|
-
# you may not use this file except in compliance with the License.
|
|
5
|
-
# You may obtain a copy of the License at
|
|
6
|
-
#
|
|
7
|
-
# http://www.apache.org/licenses/LICENSE-2.0
|
|
8
|
-
#
|
|
9
|
-
# Unless required by applicable law or agreed to in writing, software
|
|
10
|
-
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
11
|
-
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
12
|
-
# See the License for the specific language governing permissions and
|
|
13
|
-
# limitations under the License.
|
|
14
|
-
# ============================================================================
|
|
15
|
-
"""
|
|
16
|
-
Compression export module.
|
|
17
|
-
|
|
18
|
-
Note: This is an experimental interface that is subject to change and/or deletion.
|
|
19
|
-
"""
|
|
@@ -1,514 +0,0 @@
|
|
|
1
|
-
# Copyright 2020 Huawei Technologies Co., Ltd
|
|
2
|
-
#
|
|
3
|
-
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
4
|
-
# you may not use this file except in compliance with the License.
|
|
5
|
-
# You may obtain a copy of the License at
|
|
6
|
-
#
|
|
7
|
-
# http://www.apache.org/licenses/LICENSE-2.0
|
|
8
|
-
#
|
|
9
|
-
# Unless required by applicable law or agreed to in writing, software
|
|
10
|
-
# distributed under the License is distributed on an "AS IS" BASIS,
|
|
11
|
-
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
12
|
-
# See the License for the specific language governing permissions and
|
|
13
|
-
# limitations under the License.
|
|
14
|
-
# ============================================================================
|
|
15
|
-
"""
|
|
16
|
-
Note:
|
|
17
|
-
Export for quantization. This is interface that is subject to change or deletion.
|
|
18
|
-
"""
|
|
19
|
-
|
|
20
|
-
from __future__ import absolute_import
|
|
21
|
-
|
|
22
|
-
import copy
|
|
23
|
-
|
|
24
|
-
import numpy as np
|
|
25
|
-
|
|
26
|
-
from mindspore import log as logger
|
|
27
|
-
from mindspore import nn, ops
|
|
28
|
-
from mindspore._checkparam import Validator
|
|
29
|
-
from mindspore.common import Tensor
|
|
30
|
-
from mindspore.common import dtype as mstype
|
|
31
|
-
from mindspore.common.api import _cell_graph_executor as _executor
|
|
32
|
-
from mindspore.common.parameter import Parameter
|
|
33
|
-
from mindspore.nn import Cell
|
|
34
|
-
from mindspore.nn.layer import quant
|
|
35
|
-
from mindspore.ops import operations as P
|
|
36
|
-
from mindspore.ops import functional as F
|
|
37
|
-
from mindspore.ops.operations import _inner_ops as inner
|
|
38
|
-
from mindspore.compression.quant import quant_utils
|
|
39
|
-
from mindspore.compression.quant.qat import _AddFakeQuantInput, _AddFakeQuantAfterSubCell
|
|
40
|
-
|
|
41
|
-
|
|
42
|
-
__all__ = ["ExportToQuantInferNetwork"]
|
|
43
|
-
|
|
44
|
-
|
|
45
|
-
class QuantBlock(Cell):
|
|
46
|
-
r"""
|
|
47
|
-
A quant block of Conv/Dense, activation layer for Ascend deploy.
|
|
48
|
-
|
|
49
|
-
Calculate Conv or Dense in Int8, with Quant and DeQuant.
|
|
50
|
-
|
|
51
|
-
Notes:
|
|
52
|
-
This block is only for deploy, and not trainable.
|
|
53
|
-
|
|
54
|
-
Args:
|
|
55
|
-
in_channels (int): The number of channels in the input space.
|
|
56
|
-
out_channels (int): The number of channels in the output space.
|
|
57
|
-
weight_init (Union[Tensor, str, Initializer, numbers.Number]): The trainable weight_init parameter. The dtype
|
|
58
|
-
is same as input x. The values of str refer to the function `initializer`. Default: 'normal'.
|
|
59
|
-
bias_init (Union[Tensor, str, Initializer, numbers.Number]): The trainable bias_init parameter. The dtype is
|
|
60
|
-
same as input x. The values of str refer to the function `initializer`. Default: 'zeros'.
|
|
61
|
-
has_bias (bool): Specifies whether the layer uses a bias vector. Default: True.
|
|
62
|
-
activation (str): The regularization function applied to the output of the layer, eg. 'relu'. Default: None.
|
|
63
|
-
batchnorm (bool): Specifies to used batchnorm or not. Default: None.
|
|
64
|
-
activation (string): Specifies activation type. The optional values are as following:
|
|
65
|
-
'softmax', 'logsoftmax', 'relu', 'relu6', 'tanh', 'gelu', 'sigmoid',
|
|
66
|
-
'prelu', 'leakyrelu', 'hswish', 'hsigmoid'. Default: None.
|
|
67
|
-
|
|
68
|
-
Inputs:
|
|
69
|
-
- **input** (Tensor) - Tensor of shape :math:`(N, in\_channels)`.
|
|
70
|
-
|
|
71
|
-
Outputs:
|
|
72
|
-
Tensor of shape :math:`(N, out\_channels)`.
|
|
73
|
-
"""
|
|
74
|
-
|
|
75
|
-
def __init__(self,
|
|
76
|
-
core_op,
|
|
77
|
-
weight,
|
|
78
|
-
quant_op,
|
|
79
|
-
dequant_op,
|
|
80
|
-
dequant_scale,
|
|
81
|
-
bias=None,
|
|
82
|
-
activation=None):
|
|
83
|
-
super(QuantBlock, self).__init__()
|
|
84
|
-
self.core_op = core_op
|
|
85
|
-
self.weight = weight
|
|
86
|
-
self.quant = quant_op
|
|
87
|
-
self.dequant = dequant_op
|
|
88
|
-
self.dequant_scale = dequant_scale
|
|
89
|
-
self.bias = bias
|
|
90
|
-
self.has_bias = bias is not None
|
|
91
|
-
self.activation = activation
|
|
92
|
-
self.has_act = activation is not None
|
|
93
|
-
self.bias_add = P.BiasAdd()
|
|
94
|
-
self.sub = P.Sub()
|
|
95
|
-
self.weight_offset = Parameter(np.zeros(1, dtype=np.int8), name='weight_offset')
|
|
96
|
-
|
|
97
|
-
def construct(self, x):
|
|
98
|
-
x = self.quant(x)
|
|
99
|
-
if self.has_bias:
|
|
100
|
-
weight = self.sub(self.weight, self.weight_offset)
|
|
101
|
-
x = self.core_op(x, weight)
|
|
102
|
-
x = self.bias_add(x, self.bias)
|
|
103
|
-
else:
|
|
104
|
-
x = self.core_op(x, self.weight)
|
|
105
|
-
x = self.dequant(x, self.dequant_scale)
|
|
106
|
-
x = F.cast(x, mstype.float32)
|
|
107
|
-
if self.has_act:
|
|
108
|
-
x = self.activation(x)
|
|
109
|
-
return x
|
|
110
|
-
|
|
111
|
-
def extend_repr(self):
|
|
112
|
-
s = f'quant={self.quant}, core_op={type(self.core_op)}, weight=shape[{self.weight.shape}]'
|
|
113
|
-
if self.has_bias:
|
|
114
|
-
s += f', bias=shape[{self.bias.shape}]'
|
|
115
|
-
if self.has_act:
|
|
116
|
-
s += f', activation={self.activation}'
|
|
117
|
-
s += f', dequant={self.dequant}'
|
|
118
|
-
return s
|
|
119
|
-
|
|
120
|
-
|
|
121
|
-
class QuantMindirBlock(Cell):
|
|
122
|
-
"""A quant binary block of Conv/Dense, activation layer for export MINDIR model.
|
|
123
|
-
|
|
124
|
-
Args:
|
|
125
|
-
core_op (Cell): The operation cell.
|
|
126
|
-
weight (Tensor): The weight of the cell.
|
|
127
|
-
bias (Tensor): The bias of the cell. Default: None.
|
|
128
|
-
activation (str): The regularization function applied to the output of the layer, eg. 'relu'. Default: None.
|
|
129
|
-
param_dict (dict): The information of the cell.
|
|
130
|
-
"""
|
|
131
|
-
|
|
132
|
-
def __init__(self,
|
|
133
|
-
core_op,
|
|
134
|
-
weight,
|
|
135
|
-
bias=None,
|
|
136
|
-
activation=None,
|
|
137
|
-
param_dict=None):
|
|
138
|
-
|
|
139
|
-
super(QuantMindirBlock, self).__init__()
|
|
140
|
-
self.core_op = core_op
|
|
141
|
-
if activation is not None:
|
|
142
|
-
self.core_op.add_prim_attr("activation_name", activation.__class__.__name__)
|
|
143
|
-
self.core_op.add_prim_attr("filter_maxq", Tensor(param_dict["filter_maxq"]))
|
|
144
|
-
self.core_op.add_prim_attr("filter_minq", Tensor(param_dict["filter_minq"]))
|
|
145
|
-
if param_dict["output_maxq"] is not None:
|
|
146
|
-
self.core_op.add_prim_attr("output_maxq", Tensor(param_dict["output_maxq"]))
|
|
147
|
-
self.core_op.add_prim_attr("output_minq", Tensor(param_dict["output_minq"]))
|
|
148
|
-
self.core_op.add_prim_attr("symmetric", Tensor(param_dict["symmetric"]))
|
|
149
|
-
if hasattr(core_op, 'pad_mode'):
|
|
150
|
-
self.core_op.add_prim_attr("pad_mode", core_op.pad_mode)
|
|
151
|
-
self.core_op.add_prim_attr("act_num_bits", Tensor(8))
|
|
152
|
-
self.core_op.add_prim_attr("weight_num_bits", Tensor(param_dict["weight_num_bits"]))
|
|
153
|
-
self.core_op.add_prim_attr("weight_narrow_range", Tensor(param_dict["weight_narrow_range"]))
|
|
154
|
-
if param_dict["input_narrow_range"] is not None:
|
|
155
|
-
self.core_op.add_prim_attr("input_narrow_range", Tensor(param_dict["input_narrow_range"]))
|
|
156
|
-
if param_dict["output_narrow_range"] is not None:
|
|
157
|
-
self.core_op.add_prim_attr("output_narrow_range", Tensor(param_dict["output_narrow_range"]))
|
|
158
|
-
if param_dict["input_maxq"] == 'None':
|
|
159
|
-
self.core_op.add_prim_attr("mean", Tensor(param_dict["mean"]))
|
|
160
|
-
self.core_op.add_prim_attr("std_dev", Tensor(param_dict["std_dev"]))
|
|
161
|
-
elif param_dict["input_maxq"] is not None:
|
|
162
|
-
self.core_op.add_prim_attr("input_maxq", Tensor(param_dict["input_maxq"]))
|
|
163
|
-
self.core_op.add_prim_attr("input_minq", Tensor(param_dict["input_minq"]))
|
|
164
|
-
|
|
165
|
-
self.weight = weight
|
|
166
|
-
self.bias = bias
|
|
167
|
-
self.has_bias = bias is not None
|
|
168
|
-
self.activation = activation
|
|
169
|
-
self.has_act = activation is not None
|
|
170
|
-
self.bias_add = P.BiasAdd()
|
|
171
|
-
|
|
172
|
-
def construct(self, x):
|
|
173
|
-
if self.has_bias:
|
|
174
|
-
x = self.core_op(x, self.weight)
|
|
175
|
-
x = self.bias_add(x, self.bias)
|
|
176
|
-
else:
|
|
177
|
-
x = self.core_op(x, self.weight)
|
|
178
|
-
if self.has_act:
|
|
179
|
-
x = self.activation(x)
|
|
180
|
-
return x
|
|
181
|
-
|
|
182
|
-
def extend_repr(self):
|
|
183
|
-
s = f'core_op={type(self.core_op)}, weight=shape[{self.weight.shape}]'
|
|
184
|
-
if self.has_bias:
|
|
185
|
-
s += f', bias=shape[{self.bias.shape}]'
|
|
186
|
-
if self.has_act:
|
|
187
|
-
s += f', activation={self.activation}'
|
|
188
|
-
return s
|
|
189
|
-
|
|
190
|
-
|
|
191
|
-
class ExportToQuantInferNetwork:
|
|
192
|
-
"""
|
|
193
|
-
Convert quantization aware network to infer network.
|
|
194
|
-
|
|
195
|
-
Args:
|
|
196
|
-
network (Cell): MindSpore quantization aware training network.
|
|
197
|
-
inputs (Tensor): Input tensors of the `quantization aware training network`.
|
|
198
|
-
mean (int, float): The mean of input data after preprocessing, used for quantizing the first layer of network.
|
|
199
|
-
Default: 127.5.
|
|
200
|
-
std_dev (int, float): The variance of input data after preprocessing, used for quantizing the first layer
|
|
201
|
-
of network. Default: 127.5.
|
|
202
|
-
is_mindir (bool): Whether export MINDIR format. Default: False.
|
|
203
|
-
|
|
204
|
-
Returns:
|
|
205
|
-
Cell, Infer network.
|
|
206
|
-
"""
|
|
207
|
-
|
|
208
|
-
def __init__(self, network, mean, std_dev, *inputs, is_mindir=False):
|
|
209
|
-
network = Validator.check_isinstance('network', network, (nn.Cell,))
|
|
210
|
-
self.data_type = mstype.int8
|
|
211
|
-
self.network = copy.deepcopy(network)
|
|
212
|
-
self.network_bk = copy.deepcopy(network)
|
|
213
|
-
self.get_inputs_table(inputs)
|
|
214
|
-
self.mean = mean
|
|
215
|
-
self.std_dev = std_dev
|
|
216
|
-
self.is_mindir = is_mindir
|
|
217
|
-
self.upcell = None
|
|
218
|
-
|
|
219
|
-
def get_inputs_table(self, inputs):
|
|
220
|
-
"""Get the input quantization parameters of quantization cell for quant export."""
|
|
221
|
-
phase_name = 'export_quant'
|
|
222
|
-
graph_id, _ = _executor.compile(self.network, *inputs, phase=phase_name, do_convert=False)
|
|
223
|
-
self.quant_info_table = _executor.fetch_info_for_quant_export(graph_id)
|
|
224
|
-
|
|
225
|
-
def run(self):
|
|
226
|
-
"""Start to convert."""
|
|
227
|
-
logger.warning("The compression module is deprecated and may not be supported in later version, please use "
|
|
228
|
-
"MindSpore Golden Stick(https://gitee.com/mindspore/golden-stick) instead.")
|
|
229
|
-
self.network.update_cell_prefix()
|
|
230
|
-
network = self.network
|
|
231
|
-
if isinstance(network, _AddFakeQuantInput):
|
|
232
|
-
network = network.network
|
|
233
|
-
network = self._convert_quant2deploy(network)
|
|
234
|
-
return network
|
|
235
|
-
|
|
236
|
-
def _get_quant_block(self, cell_core, activation, fake_quant_a_out):
|
|
237
|
-
"""convert network's quant subcell to deploy subcell"""
|
|
238
|
-
scale_a_in, zp_a_in, scale_w, zp_w, param_dict = self.__get_quant_param(cell_core, fake_quant_a_out)
|
|
239
|
-
|
|
240
|
-
# Build the `Quant` `Dequant` op.
|
|
241
|
-
# Quant only support perlayer version. Need check here.
|
|
242
|
-
quant_op = inner.Quant(1 / float(scale_a_in), float(zp_a_in))
|
|
243
|
-
scale_deq = self.__get_dequant_scale(scale_a_in, scale_w)
|
|
244
|
-
dequant_op = inner.Dequant()
|
|
245
|
-
|
|
246
|
-
if isinstance(activation, _AddFakeQuantAfterSubCell):
|
|
247
|
-
activation = activation.subcell
|
|
248
|
-
elif hasattr(activation, "get_origin"):
|
|
249
|
-
activation = activation.get_origin()
|
|
250
|
-
|
|
251
|
-
# get op
|
|
252
|
-
if isinstance(cell_core, quant.DenseQuant):
|
|
253
|
-
op_core = P.MatMul()
|
|
254
|
-
else:
|
|
255
|
-
op_core = cell_core.conv
|
|
256
|
-
|
|
257
|
-
# get the `weight` and `bias`
|
|
258
|
-
weight, bias, weight_b, bias_b = self.__get_weight_bias(cell_core, scale_a_in, scale_w, zp_w)
|
|
259
|
-
|
|
260
|
-
if self.is_mindir:
|
|
261
|
-
block = QuantMindirBlock(op_core, weight_b, bias_b, activation, param_dict)
|
|
262
|
-
else:
|
|
263
|
-
block = QuantBlock(op_core, weight, quant_op, dequant_op, scale_deq, bias, activation)
|
|
264
|
-
return block
|
|
265
|
-
|
|
266
|
-
def _get_input_quant_param(self, minq_name, np_type, param_dict):
|
|
267
|
-
"""get input quant parameter for quant block"""
|
|
268
|
-
fake_quant_a_in_prefix = minq_name[:-5]
|
|
269
|
-
cells = self.network_bk.cells_and_names()
|
|
270
|
-
for cell in cells:
|
|
271
|
-
if cell[0].endswith(fake_quant_a_in_prefix):
|
|
272
|
-
fake_quant_a_in = cell[1]
|
|
273
|
-
break
|
|
274
|
-
scale_a_in, zp_a_in, param_dict["input_maxq"], param_dict["input_minq"] = \
|
|
275
|
-
quant_utils.scale_zp_max_min_from_fake_quant_cell(fake_quant_a_in, np_type)
|
|
276
|
-
param_dict["input_narrow_range"] = fake_quant_a_in.narrow_range
|
|
277
|
-
return scale_a_in, zp_a_in
|
|
278
|
-
|
|
279
|
-
def __get_quant_param(self, cell_core, fake_quant_a_out):
|
|
280
|
-
"""get parameter for quant block"""
|
|
281
|
-
w_minq_name = cell_core.fake_quant_weight.minq.name
|
|
282
|
-
w_maxq_name = cell_core.fake_quant_weight.maxq.name
|
|
283
|
-
np_type = mstype.dtype_to_nptype(self.data_type)
|
|
284
|
-
param_dict = dict()
|
|
285
|
-
param_dict["filter_maxq"] = None
|
|
286
|
-
param_dict["filter_minq"] = None
|
|
287
|
-
param_dict["output_maxq"] = None
|
|
288
|
-
param_dict["output_minq"] = None
|
|
289
|
-
param_dict["input_maxq"] = None
|
|
290
|
-
param_dict["input_minq"] = None
|
|
291
|
-
param_dict["input_narrow_range"] = None
|
|
292
|
-
param_dict["output_narrow_range"] = None
|
|
293
|
-
param_dict["weight_narrow_range"] = cell_core.fake_quant_weight.narrow_range
|
|
294
|
-
param_dict["mean"] = self.mean
|
|
295
|
-
param_dict["std_dev"] = self.std_dev
|
|
296
|
-
param_dict["symmetric"] = cell_core.fake_quant_weight.symmetric
|
|
297
|
-
param_dict["weight_num_bits"] = cell_core.fake_quant_weight.num_bits
|
|
298
|
-
|
|
299
|
-
scale_w, zp_w, param_dict["filter_maxq"], param_dict["filter_minq"] = \
|
|
300
|
-
quant_utils.scale_zp_max_min_from_fake_quant_cell(cell_core.fake_quant_weight, np_type)
|
|
301
|
-
if fake_quant_a_out is not None:
|
|
302
|
-
_, _, param_dict["output_maxq"], param_dict["output_minq"] = \
|
|
303
|
-
quant_utils.scale_zp_max_min_from_fake_quant_cell(fake_quant_a_out, np_type)
|
|
304
|
-
param_dict["output_narrow_range"] = fake_quant_a_out.narrow_range
|
|
305
|
-
|
|
306
|
-
info = self.quant_info_table.get(w_minq_name, None)
|
|
307
|
-
if not info:
|
|
308
|
-
info = self.quant_info_table.get(w_maxq_name, None)
|
|
309
|
-
if info:
|
|
310
|
-
_, minq_name = info
|
|
311
|
-
if minq_name == 'input':
|
|
312
|
-
scale_a_in, zp_a_in, param_dict["input_maxq"], param_dict["input_minq"] = \
|
|
313
|
-
(1 / self.std_dev), round(self.mean), 'None', 'None'
|
|
314
|
-
else:
|
|
315
|
-
scale_a_in, zp_a_in = self._get_input_quant_param(minq_name, np_type, param_dict)
|
|
316
|
-
else:
|
|
317
|
-
# skip quant layer
|
|
318
|
-
scale_a_in, zp_a_in = 1.0, 0.0
|
|
319
|
-
return scale_a_in, zp_a_in, scale_w, zp_w, param_dict
|
|
320
|
-
|
|
321
|
-
@staticmethod
|
|
322
|
-
def __get_dequant_scale(scale_a_in, scale_w):
|
|
323
|
-
"""Get dequant scale"""
|
|
324
|
-
scale_deq = scale_a_in * scale_w
|
|
325
|
-
|
|
326
|
-
# fuse parameter
|
|
327
|
-
# |--------|47:40|--------|39:32|--------|31:0|
|
|
328
|
-
# offset_w [8] shift_N [8] deq_scale [32]
|
|
329
|
-
float32_deq_scale = scale_deq.astype(np.float32)
|
|
330
|
-
uint32_deq_scale = np.frombuffer(float32_deq_scale, np.uint32)
|
|
331
|
-
scale_length = scale_deq.size # channel
|
|
332
|
-
dequant_param = np.zeros(scale_length, dtype=np.uint64)
|
|
333
|
-
for index in range(scale_length):
|
|
334
|
-
dequant_param[index] += uint32_deq_scale[index]
|
|
335
|
-
scale_deq = Tensor(dequant_param, mstype.uint64)
|
|
336
|
-
return scale_deq
|
|
337
|
-
|
|
338
|
-
def __get_weight_bias(self, cell_core, scale_a_in, scale_w, zp_w):
|
|
339
|
-
"""Get weight and bias for quantizaiton"""
|
|
340
|
-
np_type = mstype.dtype_to_nptype(self.data_type)
|
|
341
|
-
weight = cell_core.weight.data.asnumpy()
|
|
342
|
-
bias = None
|
|
343
|
-
if isinstance(cell_core, (quant.DenseQuant, quant.Conv2dQuant)):
|
|
344
|
-
if cell_core.has_bias:
|
|
345
|
-
bias = cell_core.bias.data.asnumpy()
|
|
346
|
-
elif isinstance(cell_core, (quant.Conv2dBnFoldQuant, quant.Conv2dBnFoldQuantOneConv)):
|
|
347
|
-
weight, bias = quant_utils.fold_batchnorm(weight, cell_core)
|
|
348
|
-
elif isinstance(cell_core, quant.Conv2dBnWithoutFoldQuant):
|
|
349
|
-
weight, bias = quant_utils.without_fold_batchnorm(weight, cell_core)
|
|
350
|
-
weight_b = weight
|
|
351
|
-
bias_b = bias
|
|
352
|
-
# apply the quant
|
|
353
|
-
quant_min, quant_max = quant_utils.get_quant_min_max(np_type,
|
|
354
|
-
cell_core.fake_quant_weight.num_bits,
|
|
355
|
-
cell_core.fake_quant_weight.narrow_range)
|
|
356
|
-
weight = quant_utils.weight2int(weight, scale_w, zp_w, quant_min, quant_max)
|
|
357
|
-
if bias is not None:
|
|
358
|
-
if 0 in scale_a_in:
|
|
359
|
-
raise ValueError("Zero exist in `scale_a_in` which will lead to divide zero error.")
|
|
360
|
-
if 0 in scale_w:
|
|
361
|
-
raise ValueError("Zero exist in `scale_w` which will lead to divide zero error.")
|
|
362
|
-
bias = Tensor(bias / scale_a_in / scale_w, mstype.int32)
|
|
363
|
-
|
|
364
|
-
if isinstance(cell_core, quant.DenseQuant):
|
|
365
|
-
weight = np.transpose(weight)
|
|
366
|
-
weight_b = np.transpose(weight_b)
|
|
367
|
-
|
|
368
|
-
weight_tensor = Tensor(weight, self.data_type)
|
|
369
|
-
weight_b_tensor = Tensor(weight_b)
|
|
370
|
-
if bias_b is not None:
|
|
371
|
-
bias_b_tensor = Tensor(bias_b, mstype.float32)
|
|
372
|
-
return weight_tensor, bias, weight_b_tensor, bias_b_tensor
|
|
373
|
-
return weight_tensor, bias, weight_b_tensor, None
|
|
374
|
-
|
|
375
|
-
def _add_output_min_max_for_op(self, origin_op, fake_quant_cell):
|
|
376
|
-
"""add output quant info for quant op for export mindir."""
|
|
377
|
-
if self.is_mindir:
|
|
378
|
-
if isinstance(origin_op, ops.Primitive) and not hasattr(origin_op, 'output_minq'):
|
|
379
|
-
np_type = mstype.dtype_to_nptype(self.data_type)
|
|
380
|
-
_, _, maxq, minq = quant_utils.scale_zp_max_min_from_fake_quant_cell(fake_quant_cell, np_type)
|
|
381
|
-
origin_op.add_prim_attr('output_maxq', Tensor(maxq))
|
|
382
|
-
origin_op.add_prim_attr('output_minq', Tensor(minq))
|
|
383
|
-
|
|
384
|
-
def _convert_subcell(self, network, change, name, subcell):
|
|
385
|
-
"""Convert subcell to ant subcell."""
|
|
386
|
-
if subcell is not None and hasattr(subcell, "fake_quant_weight"):
|
|
387
|
-
new_subcell = self._get_quant_block(subcell, None, None)
|
|
388
|
-
prefix = subcell.param_prefix
|
|
389
|
-
new_subcell.update_parameters_name(prefix + '.')
|
|
390
|
-
self.upcell = new_subcell
|
|
391
|
-
network.insert_child_to_cell(name, new_subcell)
|
|
392
|
-
change = True
|
|
393
|
-
return network, change
|
|
394
|
-
|
|
395
|
-
def _convert_conv(self, network, change, name, subcell):
|
|
396
|
-
"""Convert subcell to ant subcell for conv."""
|
|
397
|
-
cell_core = subcell.conv
|
|
398
|
-
activation = subcell.activation
|
|
399
|
-
fake_quant_act = None
|
|
400
|
-
if hasattr(activation, 'fake_quant_act_before'):
|
|
401
|
-
fake_quant_act = activation.fake_quant_act_before
|
|
402
|
-
elif hasattr(activation, 'fake_quant_act'):
|
|
403
|
-
fake_quant_act = activation.fake_quant_act
|
|
404
|
-
if cell_core is not None and hasattr(cell_core, "fake_quant_weight"):
|
|
405
|
-
new_subcell = self._get_quant_block(cell_core, activation, fake_quant_act)
|
|
406
|
-
self.upcell = None
|
|
407
|
-
prefix = subcell.param_prefix
|
|
408
|
-
new_subcell.update_parameters_name(prefix + '.')
|
|
409
|
-
network.insert_child_to_cell(name, new_subcell)
|
|
410
|
-
change = True
|
|
411
|
-
return network, change
|
|
412
|
-
|
|
413
|
-
def _convert_dense(self, network, change, name, subcell):
|
|
414
|
-
"""Convert subcell to ant subcell for dense."""
|
|
415
|
-
cell_core = subcell.dense
|
|
416
|
-
activation = subcell.activation
|
|
417
|
-
fake_quant_act = None
|
|
418
|
-
if hasattr(activation, 'fake_quant_act_before'):
|
|
419
|
-
fake_quant_act = activation.fake_quant_act_before
|
|
420
|
-
elif hasattr(activation, 'fake_quant_act'):
|
|
421
|
-
fake_quant_act = activation.fake_quant_act
|
|
422
|
-
if cell_core is not None and hasattr(cell_core, "fake_quant_weight"):
|
|
423
|
-
new_subcell = self._get_quant_block(cell_core, activation, fake_quant_act)
|
|
424
|
-
prefix = subcell.param_prefix
|
|
425
|
-
new_subcell.update_parameters_name(prefix + '.')
|
|
426
|
-
network.insert_child_to_cell(name, new_subcell)
|
|
427
|
-
self.upcell = None
|
|
428
|
-
change = True
|
|
429
|
-
return network, change
|
|
430
|
-
|
|
431
|
-
def _convert_act(self, subcell):
|
|
432
|
-
"""Convert subcell to ant subcell for activation."""
|
|
433
|
-
activation = subcell.get_origin()
|
|
434
|
-
if isinstance(activation, nn.ReLU):
|
|
435
|
-
self._add_output_min_max_for_op(activation.relu, subcell.fake_quant_act)
|
|
436
|
-
elif isinstance(activation, nn.ReLU6):
|
|
437
|
-
self._add_output_min_max_for_op(activation.relu6, subcell.fake_quant_act)
|
|
438
|
-
if self.upcell:
|
|
439
|
-
self._add_output_min_max_for_op(self.upcell.core_op, subcell.fake_quant_act)
|
|
440
|
-
return activation
|
|
441
|
-
|
|
442
|
-
def _convert_add(self, subcell):
|
|
443
|
-
"""Convert subcell to ant subcell for add."""
|
|
444
|
-
if isinstance(subcell.add, _AddFakeQuantAfterSubCell):
|
|
445
|
-
add_op = subcell.add.subcell
|
|
446
|
-
subcell.__delattr__("add")
|
|
447
|
-
subcell.__setattr__("add", add_op)
|
|
448
|
-
add_op = subcell.add
|
|
449
|
-
self._add_output_min_max_for_op(add_op, subcell.fake_quant_act)
|
|
450
|
-
subcell.__delattr__("fake_quant_act")
|
|
451
|
-
subcell.__setattr__("fake_quant_act", P.identity())
|
|
452
|
-
|
|
453
|
-
def _convert_observer(self, network, name, subcell):
|
|
454
|
-
"""Convert subcell to ant subcell for FakeQuantWithMinMaxObserver."""
|
|
455
|
-
if self.upcell:
|
|
456
|
-
self._add_output_min_max_for_op(self.upcell.core_op, subcell)
|
|
457
|
-
network.__delattr__(name)
|
|
458
|
-
network.__setattr__(name, P.identity())
|
|
459
|
-
|
|
460
|
-
def _convert_fake_quant_after_cell(self, network, name, subcell):
|
|
461
|
-
"""Convert subcell to ant subcell for _AddFakeQuantAfterSubCell."""
|
|
462
|
-
op = subcell.subcell
|
|
463
|
-
self._add_output_min_max_for_op(op, subcell.fake_quant_act)
|
|
464
|
-
network.__delattr__(name)
|
|
465
|
-
network.__setattr__(name, op)
|
|
466
|
-
|
|
467
|
-
def _convert_core_quant_subcell(self, network, change, name, subcell):
|
|
468
|
-
"""Convert subcell to ant subcell for conv and dense."""
|
|
469
|
-
is_core_subcell = True
|
|
470
|
-
if isinstance(subcell, nn.Conv2dBnAct):
|
|
471
|
-
network, change = self._convert_conv(network, change, name, subcell)
|
|
472
|
-
elif isinstance(subcell, nn.DenseBnAct):
|
|
473
|
-
network, change = self._convert_dense(network, change, name, subcell)
|
|
474
|
-
elif isinstance(subcell, (quant.Conv2dBnFoldQuant, quant.Conv2dBnFoldQuantOneConv,
|
|
475
|
-
quant.Conv2dBnWithoutFoldQuant, quant.Conv2dQuant, quant.DenseQuant)):
|
|
476
|
-
network, change = self._convert_subcell(network, change, name, subcell)
|
|
477
|
-
else:
|
|
478
|
-
is_core_subcell = False
|
|
479
|
-
return is_core_subcell, network, change
|
|
480
|
-
|
|
481
|
-
def _convert_other_quant_subcell(self, network, change, name, subcell):
|
|
482
|
-
"""Convert subcell to ant subcell for cell except conv and dense."""
|
|
483
|
-
is_other_subcell = True
|
|
484
|
-
if isinstance(subcell, nn.ActQuant) and hasattr(subcell, "get_origin"):
|
|
485
|
-
activation = self._convert_act(subcell)
|
|
486
|
-
network.insert_child_to_cell(name, activation)
|
|
487
|
-
change = True
|
|
488
|
-
elif isinstance(subcell, nn.TensorAddQuant):
|
|
489
|
-
self._convert_add(subcell)
|
|
490
|
-
elif isinstance(subcell, quant.FakeQuantWithMinMaxObserver):
|
|
491
|
-
self._convert_observer(network, name, subcell)
|
|
492
|
-
elif isinstance(subcell, _AddFakeQuantAfterSubCell):
|
|
493
|
-
self._convert_fake_quant_after_cell(network, name, subcell)
|
|
494
|
-
change = True
|
|
495
|
-
else:
|
|
496
|
-
is_other_subcell = False
|
|
497
|
-
return is_other_subcell, network, change
|
|
498
|
-
|
|
499
|
-
def _convert_quant2deploy(self, network):
|
|
500
|
-
"""Convert network's all quant subcell to deploy subcell."""
|
|
501
|
-
cells = network.name_cells()
|
|
502
|
-
change = False
|
|
503
|
-
for name in cells:
|
|
504
|
-
subcell = cells[name]
|
|
505
|
-
if subcell == network:
|
|
506
|
-
continue
|
|
507
|
-
is_core_quant_subcell, network, change = self._convert_core_quant_subcell(network, change, name, subcell)
|
|
508
|
-
is_other_quant_subcell, network, change = self._convert_other_quant_subcell(network, change, name, subcell)
|
|
509
|
-
if not is_core_quant_subcell and not is_other_quant_subcell:
|
|
510
|
-
self.upcell = None
|
|
511
|
-
self._convert_quant2deploy(subcell)
|
|
512
|
-
if isinstance(network, nn.SequentialCell) and change:
|
|
513
|
-
network.cell_list = list(network.cells())
|
|
514
|
-
return network
|