mindspore 1.10.0__cp37-cp37m-win_amd64.whl → 2.0.0rc1__cp37-cp37m-win_amd64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of mindspore might be problematic. Click here for more details.
- mindspore/.commit_id +1 -1
- mindspore/ConcurrencyCheck.dll +0 -0
- mindspore/CppBuildInsights.dll +0 -0
- mindspore/CppCoreCheck.dll +0 -0
- mindspore/EnumIndex.dll +0 -0
- mindspore/EspXEngine.dll +0 -0
- mindspore/HResultCheck.dll +0 -0
- mindspore/KernelTraceControl.dll +0 -0
- mindspore/LocalESPC.dll +0 -0
- mindspore/Microsoft.Diagnostics.Tracing.EventSource.dll +0 -0
- mindspore/Microsoft.VisualStudio.RemoteControl.dll +0 -0
- mindspore/Microsoft.VisualStudio.Telemetry.dll +0 -0
- mindspore/Microsoft.VisualStudio.Utilities.Internal.dll +0 -0
- mindspore/Newtonsoft.Json.dll +0 -0
- mindspore/System.Runtime.CompilerServices.Unsafe.dll +0 -0
- mindspore/VariantClear.dll +0 -0
- mindspore/__init__.py +9 -4
- mindspore/_c_dataengine.cp37-win_amd64.pyd +0 -0
- mindspore/_c_expression.cp37-win_amd64.pyd +0 -0
- mindspore/_c_mindrecord.cp37-win_amd64.pyd +0 -0
- mindspore/_check_jit_forbidden_api.py +102 -0
- mindspore/_checkparam.py +1066 -1001
- mindspore/_extends/builtin_operations.py +32 -4
- mindspore/_extends/graph_kernel/model/graph_split.py +66 -222
- mindspore/_extends/parallel_compile/akg_compiler/akg_process.py +12 -9
- mindspore/_extends/parallel_compile/akg_compiler/build_tbe_kernel.py +119 -26
- mindspore/_extends/parallel_compile/akg_compiler/tbe_topi.py +50 -50
- mindspore/_extends/parallel_compile/akg_compiler/util.py +9 -6
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_adapter.py +4 -25
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_helper.py +9 -4
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_job_manager.py +1 -27
- mindspore/_extends/parse/__init__.py +5 -3
- mindspore/_extends/parse/namespace.py +17 -2
- mindspore/_extends/parse/parser.py +193 -34
- mindspore/_extends/parse/resources.py +7 -8
- mindspore/_extends/parse/standard_method.py +1780 -435
- mindspore/_extends/parse/trope.py +3 -1
- mindspore/amp.py +53 -58
- mindspore/atlprov.dll +0 -0
- mindspore/boost/adasum.py +3 -2
- mindspore/boost/boost.py +2 -2
- mindspore/boost/boost_cell_wrapper.py +46 -26
- mindspore/boost/dim_reduce.py +6 -5
- mindspore/boost/grad_accumulation.py +2 -1
- mindspore/boost/group_loss_scale_manager.py +1 -1
- mindspore/c1.dll +0 -0
- mindspore/c1xx.dll +0 -0
- mindspore/c2.dll +0 -0
- mindspore/cfgpersist.dll +0 -0
- mindspore/clang_rt.asan_dbg_dynamic-x86_64.dll +0 -0
- mindspore/clang_rt.asan_dynamic-x86_64.dll +0 -0
- mindspore/common/__init__.py +11 -10
- mindspore/common/_decorator.py +2 -0
- mindspore/common/_register_for_adapter.py +55 -0
- mindspore/common/_stub_tensor.py +201 -0
- mindspore/common/_utils.py +57 -0
- mindspore/common/api.py +582 -297
- mindspore/common/dtype.py +66 -18
- mindspore/common/dump.py +2 -2
- mindspore/common/initializer.py +38 -1
- mindspore/common/jit_config.py +25 -13
- mindspore/common/mutable.py +53 -24
- mindspore/common/parameter.py +60 -37
- mindspore/common/seed.py +8 -24
- mindspore/common/sparse_tensor.py +927 -0
- mindspore/common/tensor.py +1627 -3900
- mindspore/communication/__init__.py +10 -5
- mindspore/communication/_comm_helper.py +78 -214
- mindspore/communication/_hccl_management.py +2 -1
- mindspore/communication/management.py +136 -47
- mindspore/config/op_info.config +501 -1008
- mindspore/context.py +291 -56
- mindspore/d3dcompiler_47.dll +0 -0
- mindspore/dataset/__init__.py +12 -8
- mindspore/dataset/audio/__init__.py +9 -9
- mindspore/dataset/audio/transforms.py +1090 -228
- mindspore/dataset/audio/utils.py +87 -39
- mindspore/dataset/audio/validators.py +223 -1
- mindspore/dataset/callback/ds_callback.py +17 -15
- mindspore/dataset/core/config.py +246 -17
- mindspore/dataset/core/py_util_helpers.py +4 -3
- mindspore/dataset/core/validator_helpers.py +10 -10
- mindspore/{parallel/nn/layers.py → dataset/debug/__init__.py} +7 -8
- mindspore/dataset/debug/debug_hook.py +65 -0
- mindspore/dataset/debug/pre_defined_hook.py +67 -0
- mindspore/dataset/engine/__init__.py +7 -3
- mindspore/dataset/engine/cache_client.py +9 -9
- mindspore/dataset/engine/datasets.py +648 -477
- mindspore/dataset/engine/datasets_audio.py +165 -167
- mindspore/dataset/engine/datasets_standard_format.py +93 -67
- mindspore/dataset/engine/datasets_text.py +492 -342
- mindspore/dataset/engine/datasets_user_defined.py +85 -50
- mindspore/dataset/engine/datasets_vision.py +1224 -699
- mindspore/dataset/engine/graphdata.py +134 -69
- mindspore/dataset/engine/iterators.py +50 -9
- mindspore/dataset/engine/offload.py +52 -31
- mindspore/dataset/engine/samplers.py +27 -24
- mindspore/dataset/engine/serializer_deserializer.py +14 -15
- mindspore/dataset/engine/validators.py +213 -52
- mindspore/dataset/text/__init__.py +10 -8
- mindspore/dataset/text/transforms.py +152 -57
- mindspore/dataset/text/utils.py +98 -49
- mindspore/dataset/text/validators.py +25 -0
- mindspore/dataset/transforms/__init__.py +4 -2
- mindspore/dataset/transforms/c_transforms.py +11 -13
- mindspore/dataset/transforms/py_transforms.py +2 -2
- mindspore/dataset/transforms/py_transforms_util.py +10 -0
- mindspore/dataset/transforms/transforms.py +13 -15
- mindspore/dataset/transforms/validators.py +7 -7
- mindspore/dataset/utils/__init__.py +2 -1
- mindspore/dataset/utils/browse_dataset.py +13 -13
- mindspore/dataset/utils/line_reader.py +121 -0
- mindspore/dataset/vision/__init__.py +8 -7
- mindspore/dataset/vision/c_transforms.py +125 -126
- mindspore/dataset/vision/py_transforms.py +37 -37
- mindspore/dataset/vision/py_transforms_util.py +23 -20
- mindspore/dataset/vision/transforms.py +316 -315
- mindspore/dataset/vision/utils.py +313 -17
- mindspore/dataset/vision/validators.py +6 -6
- mindspore/default_config.py +0 -1
- mindspore/dpcmi.dll +0 -0
- mindspore/{compression → experimental}/__init__.py +6 -5
- mindspore/experimental/map_parameter.py +275 -0
- mindspore/include/OWNERS +0 -1
- mindspore/include/api/callback/callback.h +9 -13
- mindspore/include/api/callback/ckpt_saver.h +2 -2
- mindspore/include/api/callback/loss_monitor.h +2 -2
- mindspore/include/api/callback/lr_scheduler.h +5 -5
- mindspore/include/api/callback/time_monitor.h +2 -2
- mindspore/include/api/callback/train_accuracy.h +4 -6
- mindspore/include/api/cfg.h +19 -6
- mindspore/include/api/context.h +70 -9
- mindspore/include/api/delegate.h +8 -1
- mindspore/include/api/dual_abi_helper.h +8 -24
- mindspore/include/api/metrics/accuracy.h +2 -2
- mindspore/include/api/metrics/metrics.h +4 -3
- mindspore/include/api/model.h +9 -4
- mindspore/include/api/model_group.h +68 -0
- mindspore/include/api/model_parallel_runner.h +17 -17
- mindspore/include/api/net.h +12 -11
- mindspore/include/api/serialization.h +20 -4
- mindspore/include/api/status.h +7 -1
- mindspore/include/api/types.h +25 -21
- mindspore/include/api/visible.h +4 -0
- mindspore/include/c_api/model_c.h +5 -0
- mindspore/include/c_api/status_c.h +1 -1
- mindspore/include/dataset/config.h +1 -1
- mindspore/include/dataset/constants.h +14 -0
- mindspore/include/dataset/text.h +59 -0
- mindspore/include/dataset/vision.h +56 -117
- mindspore/include/dataset/vision_lite.h +102 -0
- mindspore/jpeg62.dll +0 -0
- mindspore/log.py +28 -28
- mindspore/mindrecord/common/exceptions.py +2 -4
- mindspore/mindrecord/filereader.py +19 -1
- mindspore/mindrecord/filewriter.py +250 -88
- mindspore/mindrecord/mindpage.py +13 -13
- mindspore/mindrecord/shardheader.py +15 -15
- mindspore/mindrecord/shardreader.py +9 -0
- mindspore/mindrecord/shardwriter.py +29 -29
- mindspore/mindrecord/tools/cifar100_to_mr.py +9 -9
- mindspore/mindrecord/tools/cifar10_to_mr.py +9 -9
- mindspore/mindrecord/tools/csv_to_mr.py +4 -4
- mindspore/mindrecord/tools/imagenet_to_mr.py +70 -65
- mindspore/mindrecord/tools/mnist_to_mr.py +41 -41
- mindspore/mindrecord/tools/tfrecord_to_mr.py +6 -6
- mindspore/{libmindspore_backend.dll → mindspore_backend.dll} +0 -0
- mindspore/mindspore_common.dll +0 -0
- mindspore/mindspore_core.dll +0 -0
- mindspore/mindspore_glog.dll +0 -0
- mindspore/mindspore_shared_lib.dll +0 -0
- mindspore/msobj140.dll +0 -0
- mindspore/mspdb140.dll +0 -0
- mindspore/mspdbcore.dll +0 -0
- mindspore/mspdbst.dll +0 -0
- mindspore/mspft140.dll +0 -0
- mindspore/msvcdis140.dll +0 -0
- mindspore/msvcp140_1.dll +0 -0
- mindspore/msvcp140_2.dll +0 -0
- mindspore/msvcp140_atomic_wait.dll +0 -0
- mindspore/msvcp140_codecvt_ids.dll +0 -0
- mindspore/nn/__init__.py +1 -5
- mindspore/nn/cell.py +297 -234
- mindspore/nn/dynamic_lr.py +1 -1
- mindspore/nn/grad/cell_grad.py +17 -42
- mindspore/nn/layer/__init__.py +7 -4
- mindspore/nn/layer/activation.py +131 -88
- mindspore/nn/layer/basic.py +313 -613
- mindspore/nn/layer/channel_shuffle.py +103 -0
- mindspore/nn/layer/combined.py +1 -1
- mindspore/nn/layer/container.py +52 -6
- mindspore/nn/layer/conv.py +112 -43
- mindspore/nn/layer/dense.py +10 -9
- mindspore/nn/layer/embedding.py +36 -34
- mindspore/nn/layer/image.py +123 -27
- mindspore/nn/layer/math.py +108 -107
- mindspore/nn/layer/normalization.py +212 -366
- mindspore/nn/layer/padding.py +370 -42
- mindspore/nn/layer/pooling.py +1443 -219
- mindspore/nn/layer/rnn_cells.py +11 -16
- mindspore/nn/layer/rnns.py +38 -39
- mindspore/nn/layer/thor_layer.py +24 -25
- mindspore/nn/layer/timedistributed.py +5 -5
- mindspore/nn/layer/transformer.py +701 -0
- mindspore/nn/learning_rate_schedule.py +8 -8
- mindspore/nn/loss/__init__.py +9 -6
- mindspore/nn/loss/loss.py +678 -142
- mindspore/nn/metrics.py +53 -0
- mindspore/nn/optim/_dist_optimizer_registry.py +2 -2
- mindspore/nn/optim/ada_grad.py +8 -8
- mindspore/nn/optim/adadelta.py +2 -3
- mindspore/nn/optim/adafactor.py +18 -14
- mindspore/nn/optim/adam.py +429 -87
- mindspore/nn/optim/adamax.py +5 -6
- mindspore/nn/optim/adasum.py +10 -8
- mindspore/nn/optim/asgd.py +7 -7
- mindspore/nn/optim/ftrl.py +81 -11
- mindspore/nn/optim/lamb.py +7 -8
- mindspore/nn/optim/lars.py +4 -4
- mindspore/nn/optim/lazyadam.py +82 -7
- mindspore/nn/optim/momentum.py +8 -7
- mindspore/nn/optim/optimizer.py +19 -10
- mindspore/nn/optim/proximal_ada_grad.py +6 -5
- mindspore/nn/optim/rmsprop.py +3 -3
- mindspore/nn/optim/rprop.py +20 -16
- mindspore/nn/optim/sgd.py +21 -15
- mindspore/nn/optim/thor.py +23 -21
- mindspore/nn/probability/__init__.py +0 -2
- mindspore/nn/probability/bijector/bijector.py +7 -6
- mindspore/nn/probability/bijector/invert.py +4 -2
- mindspore/nn/probability/bijector/softplus.py +2 -2
- mindspore/nn/probability/bnn_layers/dense_variational.py +1 -1
- mindspore/nn/probability/bnn_layers/layer_distribution.py +2 -2
- mindspore/nn/probability/distribution/__init__.py +6 -0
- mindspore/nn/probability/distribution/_utils/custom_ops.py +3 -2
- mindspore/nn/probability/distribution/_utils/utils.py +11 -17
- mindspore/nn/probability/distribution/bernoulli.py +6 -6
- mindspore/nn/probability/distribution/beta.py +1 -1
- mindspore/nn/probability/distribution/categorical.py +9 -9
- mindspore/nn/probability/distribution/cauchy.py +8 -8
- mindspore/nn/probability/distribution/distribution.py +12 -6
- mindspore/nn/probability/distribution/exponential.py +5 -5
- mindspore/nn/probability/distribution/gamma.py +3 -3
- mindspore/nn/probability/distribution/geometric.py +6 -5
- mindspore/nn/probability/distribution/gumbel.py +5 -5
- mindspore/nn/probability/distribution/half_normal.py +133 -0
- mindspore/nn/probability/distribution/laplace.py +128 -0
- mindspore/nn/probability/distribution/log_normal.py +0 -1
- mindspore/nn/probability/distribution/logistic.py +4 -5
- mindspore/nn/probability/distribution/normal.py +11 -15
- mindspore/nn/probability/distribution/poisson.py +6 -2
- mindspore/nn/probability/distribution/student_t.py +150 -0
- mindspore/nn/probability/distribution/transformed_distribution.py +4 -4
- mindspore/nn/probability/distribution/uniform.py +5 -5
- mindspore/nn/reinforcement/_tensors_queue.py +3 -3
- mindspore/nn/reinforcement/tensor_array.py +2 -2
- mindspore/nn/sparse/sparse.py +8 -1
- mindspore/nn/wrap/cell_wrapper.py +55 -27
- mindspore/nn/wrap/grad_reducer.py +20 -11
- mindspore/nn/wrap/loss_scale.py +47 -30
- mindspore/numpy/array_creations.py +33 -22
- mindspore/numpy/array_ops.py +46 -42
- mindspore/numpy/logic_ops.py +6 -27
- mindspore/numpy/math_ops.py +26 -19
- mindspore/numpy/utils.py +1 -8
- mindspore/numpy/utils_const.py +112 -62
- mindspore/opencv_core452.dll +0 -0
- mindspore/opencv_imgcodecs452.dll +0 -0
- mindspore/opencv_imgproc452.dll +0 -0
- mindspore/ops/__init__.py +6 -3
- mindspore/ops/_constants.py +0 -6
- mindspore/ops/_grad/__init__.py +2 -1
- mindspore/ops/_grad/grad_array_ops.py +209 -152
- mindspore/ops/_grad/grad_base.py +55 -17
- mindspore/ops/_grad/grad_clip_ops.py +11 -3
- mindspore/ops/_grad/grad_comm_ops.py +58 -47
- mindspore/ops/_grad/grad_implementations.py +21 -61
- mindspore/ops/_grad/grad_inner_ops.py +48 -6
- mindspore/ops/_grad/grad_math_ops.py +306 -161
- mindspore/ops/_grad/grad_nn_ops.py +192 -181
- mindspore/ops/_grad/grad_other_ops.py +1 -1
- mindspore/ops/_grad/grad_quant_ops.py +5 -5
- mindspore/ops/_grad/grad_sequence_ops.py +296 -0
- mindspore/ops/_grad/grad_sparse.py +15 -9
- mindspore/ops/_grad_experimental/__init__.py +1 -0
- mindspore/ops/_grad_experimental/grad_array_ops.py +441 -55
- mindspore/ops/_grad_experimental/grad_image_ops.py +25 -7
- mindspore/ops/_grad_experimental/grad_inner_ops.py +3 -44
- mindspore/ops/_grad_experimental/grad_linalg_ops.py +16 -21
- mindspore/ops/_grad_experimental/grad_math_ops.py +979 -49
- mindspore/ops/_grad_experimental/grad_nn_ops.py +78 -8
- mindspore/ops/_grad_experimental/grad_scalar_ops.py +112 -0
- mindspore/ops/_grad_experimental/grad_sparse_ops.py +197 -13
- mindspore/ops/_op_impl/__init__.py +3 -3
- mindspore/ops/_op_impl/_custom_op/__init__.py +0 -1
- mindspore/ops/_op_impl/_custom_op/_basic.py +0 -1
- mindspore/ops/_op_impl/_custom_op/batch_matmul_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/batchnorm_fold.py +4 -2
- mindspore/ops/_op_impl/_custom_op/batchnorm_fold2.py +2 -2
- mindspore/ops/_op_impl/_custom_op/batchnorm_fold2_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/batchnorm_fold2_grad_reduce.py +5 -5
- mindspore/ops/_op_impl/_custom_op/batchnorm_fold_grad.py +3 -3
- mindspore/ops/_op_impl/_custom_op/cholesky_trsm_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/correction_mul.py +3 -3
- mindspore/ops/_op_impl/_custom_op/correction_mul_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/dsd_back_impl.py +4 -8
- mindspore/ops/_op_impl/_custom_op/dsd_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perchannel.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perchannel_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perchannel_grad_reduce.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perlayer.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perlayer_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perlayer_grad_reduce.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_quant_perchannel.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_quant_perchannel_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_quant_perlayer.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_quant_perlayer_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fused_abs_max1_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/img2col_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/matmul_cube_dense_left_impl.py +2 -2
- mindspore/ops/_op_impl/_custom_op/matmul_cube_dense_right_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/matmul_cube_fracz_left_cast_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/matmul_cube_fracz_right_mul_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/matmul_cube_impl.py +2 -2
- mindspore/ops/_op_impl/_custom_op/matmul_dds_grad_impl.py +0 -1
- mindspore/ops/_op_impl/_custom_op/matmul_dds_impl.py +0 -1
- mindspore/ops/_op_impl/_custom_op/matrix_combine_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/minmax_update_perchannel.py +2 -2
- mindspore/ops/_op_impl/_custom_op/minmax_update_perlayer.py +2 -2
- mindspore/ops/_op_impl/_custom_op/transpose02314_impl.py +1 -1
- mindspore/ops/_op_impl/aicpu/__init__.py +238 -3
- mindspore/ops/_op_impl/aicpu/abs.py +36 -0
- mindspore/ops/_op_impl/aicpu/adaptive_avg_pool_2d.py +34 -0
- mindspore/ops/_op_impl/aicpu/adaptive_avg_pool_2d_grad.py +34 -0
- mindspore/ops/_op_impl/aicpu/adaptive_avg_pool_3d.py +39 -0
- mindspore/ops/_op_impl/aicpu/adaptive_avg_pool_3d_grad.py +39 -0
- mindspore/ops/_op_impl/aicpu/adaptive_max_pool_2d_grad.py +37 -0
- mindspore/ops/_op_impl/aicpu/adaptive_max_pool_3d.py +42 -0
- mindspore/ops/_op_impl/aicpu/adaptive_max_pool_3d_grad.py +152 -0
- mindspore/ops/_op_impl/aicpu/add.py +43 -0
- mindspore/ops/_op_impl/aicpu/addcdiv.py +0 -32
- mindspore/ops/_op_impl/aicpu/addcmul.py +0 -84
- mindspore/ops/_op_impl/aicpu/affine_grid_grad.py +35 -0
- mindspore/ops/_op_impl/aicpu/arg_max.py +75 -0
- mindspore/ops/_op_impl/aicpu/arg_min.py +75 -0
- mindspore/ops/_op_impl/aicpu/argmin_with_value.py +43 -0
- mindspore/ops/_op_impl/aicpu/batch_matmul.py +43 -0
- mindspore/ops/_op_impl/aicpu/batch_norm_grad_grad.py +49 -0
- mindspore/ops/_op_impl/aicpu/bernoulli.py +48 -0
- mindspore/ops/_op_impl/aicpu/bessel_i0.py +31 -0
- mindspore/ops/_op_impl/aicpu/bias_add.py +44 -0
- mindspore/ops/_op_impl/aicpu/bias_add_grad.py +43 -0
- mindspore/ops/_op_impl/aicpu/bincount.py +33 -0
- mindspore/{nn/probability/infer/variational/__init__.py → ops/_op_impl/aicpu/cauchy.py} +17 -10
- mindspore/ops/_op_impl/aicpu/channel_shuffle.py +40 -0
- mindspore/ops/_op_impl/aicpu/cholesky.py +1 -1
- mindspore/ops/_op_impl/{cpu/bias_add.py → aicpu/choleskygrad.py} +9 -7
- mindspore/ops/_op_impl/aicpu/combined_non_max_suppression.py +42 -0
- mindspore/ops/_op_impl/aicpu/concat_offset.py +42 -0
- mindspore/ops/_op_impl/aicpu/concat_offset_v1.py +31 -0
- mindspore/ops/_op_impl/aicpu/conj.py +11 -0
- mindspore/ops/_op_impl/aicpu/crop_and_resize_grad_image.py +38 -0
- mindspore/ops/_op_impl/aicpu/cumulative_logsumexp.py +36 -0
- mindspore/ops/_op_impl/aicpu/deformable_offsets.py +38 -0
- mindspore/ops/_op_impl/aicpu/deformable_offsets_grad.py +2 -2
- mindspore/ops/_op_impl/aicpu/dense_to_sparse_set_operation.py +48 -0
- mindspore/ops/_op_impl/aicpu/diag.py +36 -0
- mindspore/ops/_op_impl/aicpu/diag_part.py +36 -0
- mindspore/ops/_op_impl/aicpu/diagonal.py +35 -0
- mindspore/ops/_op_impl/{cpu/bias_add_grad.py → aicpu/digamma.py} +9 -7
- mindspore/ops/_op_impl/aicpu/eig.py +35 -0
- mindspore/ops/_op_impl/aicpu/fft_with_size.py +41 -0
- mindspore/ops/_op_impl/aicpu/flatten.py +1 -0
- mindspore/ops/_op_impl/aicpu/fmax.py +36 -0
- mindspore/ops/_op_impl/aicpu/fmin.py +37 -0
- mindspore/ops/_op_impl/aicpu/fractional_max_pool3d_with_fixed_ksize.py +1 -1
- mindspore/ops/_op_impl/aicpu/fse_decode.py +43 -0
- mindspore/ops/_op_impl/aicpu/glu.py +33 -0
- mindspore/ops/_op_impl/aicpu/glu_grad.py +34 -0
- mindspore/ops/_op_impl/aicpu/greater.py +41 -0
- mindspore/ops/_op_impl/aicpu/greater_equal.py +41 -0
- mindspore/ops/_op_impl/aicpu/index_put.py +50 -0
- mindspore/ops/_op_impl/{tbe/scatter_add_ds.py → aicpu/inplace_index_add.py} +17 -21
- mindspore/ops/_op_impl/aicpu/instance_norm_v2.py +41 -0
- mindspore/ops/_op_impl/aicpu/instance_norm_v2_grad.py +44 -0
- mindspore/ops/_op_impl/aicpu/layer_norm_grad_grad.py +47 -0
- mindspore/ops/_op_impl/aicpu/less.py +41 -0
- mindspore/ops/_op_impl/aicpu/less_equal.py +41 -0
- mindspore/ops/_op_impl/aicpu/lgamma.py +32 -0
- mindspore/ops/_op_impl/aicpu/log_normal_reverse.py +33 -0
- mindspore/ops/_op_impl/aicpu/logit.py +33 -0
- mindspore/ops/_op_impl/aicpu/logit_grad.py +34 -0
- mindspore/ops/_op_impl/aicpu/masked_fill.py +42 -0
- mindspore/ops/_op_impl/aicpu/masked_scatter.py +39 -0
- mindspore/ops/_op_impl/aicpu/matmul.py +39 -0
- mindspore/ops/_op_impl/aicpu/matrix_logarithm.py +31 -0
- mindspore/ops/_op_impl/aicpu/matrix_power.py +32 -0
- mindspore/ops/_op_impl/aicpu/matrix_solve_ls.py +36 -0
- mindspore/ops/_op_impl/aicpu/matrix_triangular_solve.py +36 -0
- mindspore/ops/_op_impl/aicpu/mirror_pad.py +2 -0
- mindspore/ops/_op_impl/aicpu/mirror_pad_grad.py +0 -4
- mindspore/ops/_op_impl/aicpu/mul.py +3 -1
- mindspore/ops/_op_impl/aicpu/multinomial.py +14 -6
- mindspore/ops/_op_impl/aicpu/multinomial_with_replacement.py +35 -0
- mindspore/ops/_op_impl/aicpu/nan_to_num.py +34 -0
- mindspore/ops/_op_impl/aicpu/nllloss.py +38 -0
- mindspore/ops/_op_impl/aicpu/nllloss_grad.py +39 -0
- mindspore/ops/_op_impl/aicpu/ones_like.py +0 -2
- mindspore/ops/_op_impl/aicpu/polar.py +32 -0
- mindspore/ops/_op_impl/aicpu/polygamma.py +34 -0
- mindspore/ops/_op_impl/aicpu/qr.py +36 -0
- mindspore/ops/_op_impl/aicpu/quant_dtype_cast.py +40 -0
- mindspore/ops/_op_impl/aicpu/quantile.py +35 -0
- mindspore/ops/_op_impl/aicpu/ragged_tensor_to_sparse.py +73 -0
- mindspore/ops/_op_impl/aicpu/ragged_tensor_to_tensor.py +74 -0
- mindspore/ops/_op_impl/aicpu/random_shuffle.py +3 -0
- mindspore/ops/_op_impl/aicpu/randperm_v2.py +41 -0
- mindspore/ops/_op_impl/aicpu/range.py +36 -0
- mindspore/ops/_op_impl/aicpu/reciprocal.py +34 -0
- mindspore/ops/_op_impl/aicpu/reciprocal_grad.py +35 -0
- mindspore/ops/_op_impl/aicpu/reduce_sum.py +57 -0
- mindspore/ops/_op_impl/aicpu/resize_bicubic.py +2 -8
- mindspore/ops/_op_impl/aicpu/resize_bicubic_grad.py +1 -1
- mindspore/ops/_op_impl/aicpu/resize_v2.py +68 -0
- mindspore/ops/_op_impl/aicpu/resize_v2_grad.py +68 -0
- mindspore/ops/_op_impl/aicpu/scatter_elements.py +4 -0
- mindspore/ops/_op_impl/aicpu/scatter_nd_update.py +2 -0
- mindspore/ops/_op_impl/aicpu/search_sorted.py +12 -6
- mindspore/ops/_op_impl/aicpu/self_adjoint_eig.py +34 -0
- mindspore/ops/_op_impl/aicpu/sequence_add.py +34 -0
- mindspore/ops/_op_impl/aicpu/sequence_add_offset.py +34 -0
- mindspore/ops/_op_impl/aicpu/sequence_addn.py +38 -0
- mindspore/ops/_op_impl/aicpu/slice_grad.py +76 -0
- mindspore/ops/_op_impl/aicpu/smooth_l1_loss.py +35 -0
- mindspore/ops/_op_impl/aicpu/smooth_l1_loss_grad.py +37 -0
- mindspore/ops/_op_impl/aicpu/sort.py +39 -0
- mindspore/ops/_op_impl/aicpu/sparse_apply_adagrad_da.py +0 -24
- mindspore/ops/_op_impl/aicpu/sparse_cross.py +42 -0
- mindspore/ops/_op_impl/aicpu/sparse_fill_empty_rows.py +63 -0
- mindspore/ops/_op_impl/aicpu/sparse_fill_empty_rows_grad.py +45 -0
- mindspore/ops/_op_impl/aicpu/sparse_matrix_mat_mul.py +56 -0
- mindspore/ops/_op_impl/{tbe/slice_ds.py → aicpu/sparse_segment_sum.py} +16 -24
- mindspore/ops/_op_impl/aicpu/sparse_segment_sum_with_num_segments.py +68 -0
- mindspore/ops/_op_impl/aicpu/sparse_slice.py +63 -0
- mindspore/ops/_op_impl/aicpu/sparse_slice_grad.py +61 -0
- mindspore/ops/_op_impl/aicpu/squared_difference.py +2 -0
- mindspore/ops/_op_impl/aicpu/strided_slice_v2.py +93 -0
- mindspore/ops/_op_impl/aicpu/strided_slice_v2_grad.py +66 -0
- mindspore/ops/_op_impl/aicpu/tensor_scatter_update.py +59 -0
- mindspore/ops/_op_impl/{tbe/gather_v2.py → aicpu/tile.py} +24 -24
- mindspore/ops/_op_impl/aicpu/tridiagonal_solve.py +35 -0
- mindspore/ops/_op_impl/aicpu/tril_indices.py +34 -0
- mindspore/ops/_op_impl/aicpu/triu_indices.py +34 -0
- mindspore/ops/_op_impl/aicpu/uniform.py +34 -0
- mindspore/ops/_op_impl/aicpu/uniform_candidate_sampler.py +1 -0
- mindspore/ops/_op_impl/aicpu/unique_consecutive.py +10 -2
- mindspore/ops/_op_impl/cpu/__init__.py +1 -2
- mindspore/ops/_op_impl/cpu/dynamic_shape.py +5 -1
- mindspore/ops/_op_impl/cpu/maximum_grad.py +2 -0
- mindspore/{compression/common/__init__.py → ops/_op_impl/cpu/pyexecute.py} +13 -8
- mindspore/ops/_op_impl/cpu/reduce_sum.py +8 -0
- mindspore/ops/_op_impl/cpu/sparse_slice.py +62 -0
- mindspore/ops/_op_impl/cpu/sparse_slice_grad.py +60 -0
- mindspore/ops/_op_impl/cpu/tensor_shape.py +5 -1
- mindspore/ops/_op_impl/tbe/__init__.py +27 -608
- mindspore/ops/_op_impl/tbe/addcdiv_ds.py +42 -0
- mindspore/ops/_op_impl/tbe/addcmul_ds.py +44 -0
- mindspore/ops/_op_impl/tbe/assign_add_ds.py +1 -0
- mindspore/ops/_op_impl/tbe/atomic_addr_clean.py +1 -1
- mindspore/ops/_op_impl/tbe/avg_pool_3d_grad.py +1 -1
- mindspore/ops/_op_impl/tbe/basic_lstm_cell_c_state_grad_v2.py +0 -1
- mindspore/ops/_op_impl/tbe/batch_to_space.py +1 -1
- mindspore/ops/_op_impl/tbe/batch_to_space_nd.py +1 -1
- mindspore/ops/_op_impl/tbe/batch_to_space_nd_v2.py +41 -0
- mindspore/ops/_op_impl/tbe/bce_with_logits_loss.py +1 -0
- mindspore/ops/_op_impl/tbe/bias_add_grad.py +2 -0
- mindspore/ops/_op_impl/tbe/bn_infer_grad.py +4 -2
- mindspore/ops/_op_impl/tbe/bn_infer_grad_ds.py +40 -0
- mindspore/ops/_op_impl/tbe/bn_training_update.py +0 -1
- mindspore/ops/_op_impl/tbe/bn_training_update_ds.py +0 -1
- mindspore/ops/_op_impl/tbe/broadcast_to_ds.py +6 -4
- mindspore/ops/_op_impl/tbe/cast.py +0 -2
- mindspore/ops/_op_impl/tbe/cast_ds.py +3 -3
- mindspore/ops/_op_impl/tbe/ctc_loss_v2.py +0 -2
- mindspore/ops/_op_impl/tbe/ctc_loss_v2_grad.py +0 -2
- mindspore/ops/_op_impl/tbe/data_format_dim_map_ds.py +1 -0
- mindspore/ops/_op_impl/tbe/deformable_offsets.py +1 -0
- mindspore/ops/_op_impl/tbe/depthwise_conv2d.py +1 -1
- mindspore/ops/_op_impl/tbe/dynamic_atomic_addr_clean.py +1 -1
- mindspore/ops/_op_impl/tbe/gather_nd.py +1 -0
- mindspore/ops/_op_impl/tbe/greater.py +2 -0
- mindspore/ops/_op_impl/tbe/{index_add.py → inplace_index_add.py} +3 -6
- mindspore/ops/_op_impl/tbe/layer_norm_beta_gamma_backprop_v2.py +0 -1
- mindspore/ops/_op_impl/tbe/npu_clear_float_status_v2.py +35 -0
- mindspore/ops/_op_impl/tbe/npu_get_float_status_v2.py +35 -0
- mindspore/ops/_op_impl/tbe/one_hot_ds.py +0 -6
- mindspore/ops/_op_impl/tbe/{greater_ds.py → reduce_all_ds.py} +13 -16
- mindspore/ops/_op_impl/tbe/reduce_any_ds.py +39 -0
- mindspore/ops/_op_impl/tbe/roi_align_ds.py +44 -0
- mindspore/ops/_op_impl/tbe/roi_align_grad_ds.py +44 -0
- mindspore/ops/_op_impl/tbe/scatter_add.py +2 -0
- mindspore/ops/_op_impl/tbe/scatter_nd_add.py +2 -2
- mindspore/ops/_op_impl/tbe/slice.py +26 -15
- mindspore/ops/_op_impl/tbe/space_to_batch.py +1 -1
- mindspore/ops/_op_impl/tbe/space_to_batch_nd.py +1 -1
- mindspore/ops/_op_impl/tbe/strided_slice_grad_d.py +1 -0
- mindspore/ops/_op_impl/tbe/trans_data_ds.py +15 -5
- mindspore/ops/_op_impl/tbe/unsorted_segment_sum.py +1 -1
- mindspore/ops/_op_impl/tbe/unsorted_segment_sum_ds.py +2 -0
- mindspore/ops/_primitive_cache.py +3 -2
- mindspore/ops/_register_for_op.py +11 -0
- mindspore/ops/_utils/__init__.py +1 -1
- mindspore/ops/_utils/utils.py +20 -41
- mindspore/ops/_vmap/__init__.py +2 -2
- mindspore/ops/_vmap/vmap_array_ops.py +170 -78
- mindspore/ops/_vmap/vmap_base.py +24 -10
- mindspore/ops/_vmap/vmap_convolution_ops.py +7 -10
- mindspore/ops/_vmap/vmap_grad_math_ops.py +4 -4
- mindspore/ops/_vmap/vmap_grad_nn_ops.py +41 -9
- mindspore/ops/_vmap/vmap_image_ops.py +52 -0
- mindspore/ops/_vmap/vmap_math_ops.py +77 -6
- mindspore/ops/_vmap/vmap_nn_ops.py +78 -29
- mindspore/ops/_vmap/vmap_other_ops.py +3 -1
- mindspore/ops/_vmap/vmap_random_ops.py +55 -3
- mindspore/ops/_vmap/vmap_sparse_ops.py +1 -0
- mindspore/ops/bprop_mindir/AdaptiveAvgPool2D_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/AdaptiveMaxPool2D_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/ApproximateEqual_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/Argmax_bprop.mindir +13 -12
- mindspore/ops/bprop_mindir/Argmin_bprop.mindir +14 -13
- mindspore/ops/bprop_mindir/AssignSub_bprop.mindir +17 -18
- mindspore/ops/bprop_mindir/Assign_bprop.mindir +16 -16
- mindspore/ops/bprop_mindir/AvgPool3D_bprop.mindir +150 -0
- mindspore/ops/bprop_mindir/AvgPool_bprop.mindir +66 -0
- mindspore/ops/bprop_mindir/BCEWithLogitsLoss_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/BNTrainingReduce_bprop.mindir +13 -12
- mindspore/ops/bprop_mindir/BatchNormGrad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/BatchToSpaceND_bprop.mindir +28 -0
- mindspore/ops/bprop_mindir/BiasAddGrad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/BinaryCrossEntropy_bprop.mindir +33 -0
- mindspore/ops/bprop_mindir/BroadcastTo_bprop.mindir +306 -0
- mindspore/ops/bprop_mindir/Broadcast_bprop.mindir +12 -8
- mindspore/ops/bprop_mindir/CTCLoss_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Concat_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Conv2DBackpropFilter_bprop.mindir +240 -0
- mindspore/ops/bprop_mindir/Conv2DBackpropInput_bprop.mindir +247 -0
- mindspore/ops/bprop_mindir/Conv2DTranspose_bprop.mindir +247 -0
- mindspore/ops/bprop_mindir/Conv3DTranspose_bprop.mindir +315 -0
- mindspore/ops/bprop_mindir/Conv3D_bprop.mindir +278 -0
- mindspore/ops/bprop_mindir/DType_bprop.mindir +12 -12
- mindspore/ops/bprop_mindir/DeformableOffsets_bprop.mindir +58 -0
- mindspore/ops/bprop_mindir/Depend_bprop.mindir +12 -13
- mindspore/ops/bprop_mindir/DepthToSpace_bprop.mindir +23 -0
- mindspore/ops/bprop_mindir/DepthwiseConv2dNative_bprop.mindir +138 -0
- mindspore/ops/bprop_mindir/DiagPart_bprop.mindir +15 -0
- mindspore/ops/bprop_mindir/Dropout2D_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Dropout3D_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/DropoutDoMask_bprop.mindir +22 -24
- mindspore/ops/bprop_mindir/DropoutGenMask_bprop.mindir +16 -14
- mindspore/ops/bprop_mindir/DropoutGrad_bprop.mindir +27 -0
- mindspore/ops/bprop_mindir/Dropout_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/DynamicGRUV2_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/DynamicRNN_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/DynamicShape_bprop.mindir +12 -12
- mindspore/ops/bprop_mindir/Elu_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/EmbeddingLookup_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Equal_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/ExpandDims_bprop.mindir +58 -0
- mindspore/ops/bprop_mindir/FastGeLU_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/Flatten_bprop.mindir +54 -0
- mindspore/ops/bprop_mindir/FloorDiv_bprop.mindir +18 -15
- mindspore/ops/bprop_mindir/GatherD_bprop.mindir +26 -0
- mindspore/ops/bprop_mindir/GatherNd_bprop.mindir +57 -0
- mindspore/ops/bprop_mindir/Gather_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/GreaterEqual_bprop.mindir +17 -18
- mindspore/ops/bprop_mindir/Greater_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/HSigmoid_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/HSwish_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/IOU_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/InstanceNorm_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/IsFinite_bprop.mindir +13 -12
- mindspore/ops/bprop_mindir/IsInf_bprop.mindir +13 -10
- mindspore/ops/bprop_mindir/IsNan_bprop.mindir +14 -11
- mindspore/ops/bprop_mindir/KLDivLoss_bprop.mindir +126 -0
- mindspore/ops/bprop_mindir/L2Loss_bprop.mindir +15 -0
- mindspore/ops/bprop_mindir/L2Normalize_bprop.mindir +30 -0
- mindspore/ops/bprop_mindir/LRN_bprop.mindir +43 -0
- mindspore/ops/bprop_mindir/LayerNormGrad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/LessEqual_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/Less_bprop.mindir +17 -18
- mindspore/ops/bprop_mindir/LinSpace_bprop.mindir +22 -19
- mindspore/ops/bprop_mindir/Load_bprop.mindir +12 -13
- mindspore/ops/bprop_mindir/LogSoftmax_bprop.mindir +23 -0
- mindspore/ops/bprop_mindir/LogicalAnd_bprop.mindir +17 -18
- mindspore/ops/bprop_mindir/LogicalNot_bprop.mindir +14 -13
- mindspore/ops/bprop_mindir/MaskedSelect_bprop.mindir +21 -0
- mindspore/ops/bprop_mindir/MaxPool3DGradGrad_bprop.mindir +74 -0
- mindspore/ops/bprop_mindir/MaxPool3DGrad_bprop.mindir +74 -0
- mindspore/ops/bprop_mindir/MaxPool3D_bprop.mindir +75 -0
- mindspore/ops/bprop_mindir/MaxPoolGradGrad_bprop.mindir +65 -0
- mindspore/ops/bprop_mindir/MaxPoolWithArgmax_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Maximum_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Minimum_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/MirrorPad_bprop.mindir +27 -0
- mindspore/ops/bprop_mindir/Mish_bprop.mindir +35 -0
- mindspore/ops/bprop_mindir/MulNoNan_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/NLLLoss_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/NonZero_bprop.mindir +14 -0
- mindspore/ops/bprop_mindir/NotEqual_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/OneHot_bprop.mindir +25 -23
- mindspore/ops/bprop_mindir/OnesLike_bprop.mindir +13 -13
- mindspore/ops/bprop_mindir/PReLU_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Pad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Padding_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/RNNTLoss_bprop.mindir +29 -0
- mindspore/ops/bprop_mindir/ROIAlign_bprop.mindir +82 -0
- mindspore/ops/bprop_mindir/Range_bprop.mindir +21 -19
- mindspore/ops/bprop_mindir/Rank_bprop.mindir +11 -11
- mindspore/ops/bprop_mindir/ReLU6_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/ReLUV2_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/ReduceAll_bprop.mindir +18 -17
- mindspore/ops/bprop_mindir/ReduceAny_bprop.mindir +18 -17
- mindspore/ops/bprop_mindir/ReluGrad_bprop.mindir +19 -23
- mindspore/ops/bprop_mindir/Reshape_bprop.mindir +60 -0
- mindspore/ops/bprop_mindir/ResizeBilinear_bprop.mindir +29 -0
- mindspore/ops/bprop_mindir/ResizeNearestNeighbor_bprop.mindir +89 -0
- mindspore/ops/bprop_mindir/ReverseSequence_bprop.mindir +52 -0
- mindspore/ops/bprop_mindir/ReverseV2_bprop.mindir +22 -0
- mindspore/ops/bprop_mindir/Round_bprop.mindir +14 -13
- mindspore/ops/bprop_mindir/ScatterMax_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/ScatterMin_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/ScatterNdUpdate_bprop.mindir +22 -0
- mindspore/ops/bprop_mindir/ScatterNd_bprop.mindir +24 -0
- mindspore/ops/bprop_mindir/ScatterNonAliasingAdd_bprop.mindir +22 -0
- mindspore/ops/bprop_mindir/ScatterUpdate_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/SeLU_bprop.mindir +21 -0
- mindspore/ops/bprop_mindir/Select_bprop.mindir +30 -34
- mindspore/ops/bprop_mindir/Shape_bprop.mindir +12 -12
- mindspore/ops/bprop_mindir/SigmoidCrossEntropyWithLogits_bprop.mindir +21 -0
- mindspore/ops/bprop_mindir/SigmoidGrad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Sigmoid_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/Sign_bprop.mindir +13 -12
- mindspore/ops/bprop_mindir/Slice_bprop.mindir +26 -0
- mindspore/ops/bprop_mindir/SmoothL1Loss_bprop.mindir +36 -0
- mindspore/ops/bprop_mindir/SoftmaxCrossEntropyWithLogits_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Softplus_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/Softsign_bprop.mindir +33 -0
- mindspore/ops/bprop_mindir/Sort_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/SpaceToBatchND_bprop.mindir +28 -0
- mindspore/ops/bprop_mindir/SpaceToDepth_bprop.mindir +23 -0
- mindspore/ops/bprop_mindir/SparseGatherV2_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/SparseSoftmaxCrossEntropyWithLogits_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Split_bprop.mindir +22 -0
- mindspore/ops/bprop_mindir/Squeeze_bprop.mindir +54 -0
- mindspore/ops/bprop_mindir/StridedSliceGrad_bprop.mindir +95 -0
- mindspore/ops/bprop_mindir/StridedSlice_bprop.mindir +98 -0
- mindspore/ops/bprop_mindir/Switch_bprop.mindir +28 -32
- mindspore/ops/bprop_mindir/TanhGrad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Tanh_bprop.mindir +66 -0
- mindspore/ops/bprop_mindir/TensorScatterAdd_bprop.mindir +22 -0
- mindspore/ops/bprop_mindir/TensorScatterUpdate_bprop.mindir +29 -0
- mindspore/ops/bprop_mindir/TensorShape_bprop.mindir +14 -0
- mindspore/ops/bprop_mindir/Tile_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/TopK_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/TransShape_bprop.mindir +23 -0
- mindspore/ops/bprop_mindir/TruncateDiv_bprop.mindir +18 -15
- mindspore/ops/bprop_mindir/TupleGetItem_bprop.mindir +11 -13
- mindspore/ops/bprop_mindir/Unique_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/Unstack_bprop.mindir +22 -0
- mindspore/ops/bprop_mindir/UpsampleNearest3D_bprop.mindir +32 -0
- mindspore/ops/bprop_mindir/UpsampleTrilinear3D_bprop.mindir +38 -0
- mindspore/ops/bprop_mindir/ZerosLike_bprop.mindir +13 -12
- mindspore/ops/bprop_mindir/__init__.py +1 -4
- mindspore/ops/bprop_mindir/generate_mindir.py +32 -20
- mindspore/ops/composite/__init__.py +12 -13
- mindspore/ops/composite/base.py +261 -254
- mindspore/ops/composite/env_ops.py +41 -0
- mindspore/ops/composite/math_ops.py +197 -156
- mindspore/ops/composite/multitype_ops/_compile_utils.py +428 -176
- mindspore/ops/composite/multitype_ops/_constexpr_utils.py +188 -87
- mindspore/ops/composite/multitype_ops/add_impl.py +23 -1
- mindspore/ops/composite/multitype_ops/div_impl.py +3 -3
- mindspore/ops/composite/multitype_ops/equal_impl.py +1 -0
- mindspore/ops/composite/multitype_ops/floordiv_impl.py +1 -1
- mindspore/ops/composite/multitype_ops/getitem_impl.py +52 -5
- mindspore/ops/composite/multitype_ops/greater_equal_impl.py +31 -0
- mindspore/ops/composite/multitype_ops/greater_impl.py +31 -0
- mindspore/ops/composite/multitype_ops/in_impl.py +15 -3
- mindspore/ops/composite/multitype_ops/less_equal_impl.py +33 -2
- mindspore/ops/composite/multitype_ops/less_impl.py +33 -0
- mindspore/ops/composite/multitype_ops/logical_and_impl.py +2 -2
- mindspore/ops/composite/multitype_ops/logical_or_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/mod_impl.py +1 -1
- mindspore/ops/composite/multitype_ops/mul_impl.py +21 -7
- mindspore/ops/composite/multitype_ops/not_in_impl.py +15 -3
- mindspore/ops/composite/multitype_ops/ones_like_impl.py +2 -4
- mindspore/ops/composite/multitype_ops/pow_impl.py +1 -0
- mindspore/ops/composite/multitype_ops/setitem_impl.py +62 -70
- mindspore/ops/composite/multitype_ops/sub_impl.py +3 -3
- mindspore/ops/composite/multitype_ops/zeros_like_impl.py +41 -4
- mindspore/ops/function/__init__.py +323 -8
- mindspore/ops/function/array_func.py +3511 -780
- mindspore/ops/function/clip_func.py +329 -0
- mindspore/ops/function/debug_func.py +6 -6
- mindspore/ops/function/grad/__init__.py +5 -1
- mindspore/ops/function/grad/grad_func.py +736 -65
- mindspore/ops/function/image_func.py +270 -0
- mindspore/ops/function/linalg_func.py +268 -8
- mindspore/ops/function/math_func.py +8032 -3164
- mindspore/ops/function/nn_func.py +5619 -1855
- mindspore/ops/function/other_func.py +115 -0
- mindspore/ops/function/parameter_func.py +11 -10
- mindspore/ops/function/random_func.py +939 -77
- mindspore/ops/function/sparse_func.py +249 -84
- mindspore/ops/function/sparse_unary_func.py +2303 -0
- mindspore/ops/function/spectral_func.py +146 -0
- mindspore/ops/function/vmap_func.py +114 -0
- mindspore/ops/functional.py +182 -254
- mindspore/ops/op_info_register.py +79 -34
- mindspore/ops/operations/__init__.py +210 -118
- mindspore/ops/operations/_csr_ops.py +7 -7
- mindspore/ops/operations/_embedding_cache_ops.py +25 -15
- mindspore/ops/operations/_grad_ops.py +447 -322
- mindspore/ops/operations/_inner_ops.py +547 -176
- mindspore/ops/operations/_map_tensor_ops.py +112 -0
- mindspore/ops/operations/_ms_kernel.py +29 -27
- mindspore/ops/operations/_ocr_ops.py +11 -11
- mindspore/ops/operations/_opaque_predicate_registry.py +41 -0
- mindspore/ops/operations/_quant_ops.py +186 -101
- mindspore/ops/operations/_rl_inner_ops.py +122 -61
- mindspore/ops/operations/_scalar_ops.py +466 -0
- mindspore/ops/operations/_sequence_ops.py +1047 -0
- mindspore/ops/operations/_tensor_array.py +10 -11
- mindspore/ops/operations/_thor_ops.py +4 -4
- mindspore/ops/operations/array_ops.py +1428 -1226
- mindspore/ops/operations/comm_ops.py +180 -117
- mindspore/ops/operations/control_ops.py +4 -2
- mindspore/ops/operations/custom_ops.py +185 -98
- mindspore/ops/operations/debug_ops.py +92 -54
- mindspore/ops/operations/image_ops.py +406 -211
- mindspore/ops/operations/inner_ops.py +42 -53
- mindspore/ops/operations/linalg_ops.py +32 -29
- mindspore/ops/operations/math_ops.py +2076 -897
- mindspore/ops/operations/nn_ops.py +1282 -1252
- mindspore/ops/operations/other_ops.py +124 -278
- mindspore/ops/operations/random_ops.py +345 -178
- mindspore/ops/operations/rl_ops.py +8 -9
- mindspore/ops/operations/sparse_ops.py +502 -157
- mindspore/ops/operations/spectral_ops.py +107 -0
- mindspore/ops/primitive.py +192 -15
- mindspore/ops/vm_impl_registry.py +23 -2
- mindspore/parallel/__init__.py +6 -1
- mindspore/parallel/_auto_parallel_context.py +199 -92
- mindspore/parallel/_cell_wrapper.py +4 -2
- mindspore/parallel/_cost_model_context.py +3 -0
- mindspore/parallel/_dp_allreduce_fusion.py +2 -1
- mindspore/parallel/_offload_context.py +185 -0
- mindspore/parallel/_parallel_serialization.py +167 -28
- mindspore/parallel/_ps_context.py +9 -5
- mindspore/parallel/_recovery_context.py +1 -1
- mindspore/parallel/_tensor.py +9 -1
- mindspore/{nn/transformer → parallel/_transformer}/__init__.py +6 -6
- mindspore/{nn/transformer → parallel/_transformer}/layers.py +59 -37
- mindspore/{nn/transformer → parallel/_transformer}/loss.py +4 -7
- mindspore/{nn/transformer → parallel/_transformer}/moe.py +160 -35
- mindspore/{nn/transformer → parallel/_transformer}/op_parallel_config.py +3 -3
- mindspore/{nn/transformer → parallel/_transformer}/transformer.py +235 -196
- mindspore/parallel/_utils.py +47 -7
- mindspore/parallel/algo_parameter_config.py +5 -1
- mindspore/parallel/checkpoint_transform.py +329 -0
- mindspore/parallel/shard.py +229 -0
- mindspore/perf_msvcbuildinsights.dll +0 -0
- mindspore/pgodb140.dll +0 -0
- mindspore/pgort140.dll +0 -0
- mindspore/profiler/__init__.py +2 -1
- mindspore/profiler/common/util.py +4 -3
- mindspore/profiler/common/validator/validate_path.py +2 -2
- mindspore/profiler/envprofiling.py +249 -0
- mindspore/profiler/parser/aicpu_data_parser.py +38 -39
- mindspore/profiler/parser/ascend_timeline_generator.py +497 -0
- mindspore/profiler/parser/base_timeline_generator.py +471 -0
- mindspore/profiler/parser/cpu_gpu_timeline_generator.py +684 -0
- mindspore/profiler/parser/framework_parser.py +42 -16
- mindspore/profiler/parser/hccl_parser.py +158 -158
- mindspore/profiler/parser/hwts_log_parser.py +7 -6
- mindspore/profiler/parser/integrator.py +18 -1579
- mindspore/profiler/parser/minddata_analyzer.py +8 -8
- mindspore/profiler/parser/msadvisor_analyzer.py +14 -27
- mindspore/profiler/parser/msadvisor_parser.py +2 -4
- mindspore/profiler/parser/optime_parser.py +17 -18
- mindspore/profiler/parser/profiler_info.py +108 -0
- mindspore/profiler/parser/step_trace_parser.py +1 -1
- mindspore/profiler/profiling.py +396 -194
- mindspore/rewrite/__init__.py +6 -2
- mindspore/rewrite/api/node.py +51 -110
- mindspore/rewrite/api/node_type.py +10 -6
- mindspore/rewrite/api/pattern_engine.py +51 -7
- mindspore/rewrite/api/scoped_value.py +64 -53
- mindspore/rewrite/api/symbol_tree.py +108 -61
- mindspore/rewrite/api/tree_node_helper.py +2 -3
- mindspore/{compression/quant/__init__.py → rewrite/ast_creator_register.py} +20 -11
- mindspore/rewrite/ast_helpers/__init__.py +6 -3
- mindspore/rewrite/ast_helpers/ast_creator.py +115 -0
- mindspore/rewrite/ast_helpers/ast_finder.py +99 -1
- mindspore/rewrite/ast_helpers/ast_modifier.py +17 -4
- mindspore/rewrite/ast_helpers/ast_replacer.py +1 -1
- mindspore/rewrite/ast_transformers/__init__.py +0 -1
- mindspore/rewrite/ast_transformers/flatten_recursive_stmt.py +46 -5
- mindspore/rewrite/ast_transformers/remove_return_out_of_if.py +6 -3
- mindspore/rewrite/common/__init__.py +2 -0
- mindspore/rewrite/common/event.py +1 -1
- mindspore/rewrite/common/observable.py +1 -1
- mindspore/rewrite/common/observer.py +1 -1
- mindspore/rewrite/common/rewrite_elog.py +35 -0
- mindspore/rewrite/namer.py +2 -2
- mindspore/rewrite/namespace.py +14 -4
- mindspore/rewrite/node.py +161 -13
- mindspore/rewrite/parser.py +0 -1
- mindspore/rewrite/parser_register.py +0 -1
- mindspore/rewrite/parsers/arguments_parser.py +3 -2
- mindspore/rewrite/parsers/assign_parser.py +267 -67
- mindspore/rewrite/parsers/attribute_parser.py +56 -0
- mindspore/rewrite/parsers/class_def_parser.py +191 -108
- mindspore/rewrite/parsers/constant_parser.py +101 -0
- mindspore/rewrite/parsers/container_parser.py +88 -0
- mindspore/rewrite/parsers/for_parser.py +28 -15
- mindspore/rewrite/parsers/function_def_parser.py +21 -5
- mindspore/rewrite/parsers/if_parser.py +11 -28
- mindspore/rewrite/parsers/module_parser.py +9 -6
- mindspore/rewrite/parsers/return_parser.py +3 -2
- mindspore/rewrite/sparsify/__init__.py +0 -0
- mindspore/rewrite/sparsify/sparse_transformer.py +448 -0
- mindspore/rewrite/sparsify/sparsify.py +109 -0
- mindspore/rewrite/sparsify/utils.py +173 -0
- mindspore/rewrite/symbol_tree.py +322 -109
- mindspore/rewrite/symbol_tree_builder.py +45 -8
- mindspore/rewrite/symbol_tree_dumper.py +0 -1
- mindspore/rewrite/topological_manager.py +1 -2
- mindspore/run_check/_check_version.py +209 -112
- mindspore/run_check/run_check.py +2 -1
- mindspore/tbbmalloc.dll +0 -0
- mindspore/tinyxml2.dll +0 -0
- mindspore/train/__init__.py +6 -4
- mindspore/train/_utils.py +28 -5
- mindspore/train/amp.py +321 -50
- mindspore/train/callback/__init__.py +3 -1
- mindspore/train/callback/_backup_and_restore.py +120 -0
- mindspore/train/callback/_callback.py +8 -8
- mindspore/train/callback/_checkpoint.py +12 -9
- mindspore/train/callback/_early_stop.py +13 -7
- mindspore/train/callback/_history.py +8 -8
- mindspore/train/callback/_lambda_callback.py +6 -6
- mindspore/train/callback/_landscape.py +36 -38
- mindspore/train/callback/_loss_monitor.py +12 -6
- mindspore/train/callback/_lr_scheduler_callback.py +2 -4
- mindspore/train/callback/_on_request_exit.py +212 -0
- mindspore/train/callback/_reduce_lr_on_plateau.py +13 -7
- mindspore/train/callback/_summary_collector.py +27 -19
- mindspore/train/callback/_time_monitor.py +13 -7
- mindspore/train/checkpoint_pb2.py +68 -8
- mindspore/train/data_sink.py +122 -33
- mindspore/train/dataset_helper.py +28 -87
- mindspore/train/loss_scale_manager.py +4 -7
- mindspore/{nn → train}/metrics/__init__.py +20 -20
- mindspore/{nn → train}/metrics/accuracy.py +12 -10
- mindspore/{nn → train}/metrics/auc.py +4 -4
- mindspore/{nn → train}/metrics/bleu_score.py +4 -4
- mindspore/{nn → train}/metrics/confusion_matrix.py +10 -8
- mindspore/{nn → train}/metrics/cosine_similarity.py +4 -4
- mindspore/{nn → train}/metrics/dice.py +6 -5
- mindspore/{nn → train}/metrics/error.py +7 -5
- mindspore/{nn → train}/metrics/fbeta.py +9 -7
- mindspore/{nn → train}/metrics/hausdorff_distance.py +8 -6
- mindspore/{nn → train}/metrics/loss.py +4 -3
- mindspore/{nn → train}/metrics/mean_surface_distance.py +6 -5
- mindspore/{nn → train}/metrics/metric.py +6 -5
- mindspore/{nn → train}/metrics/occlusion_sensitivity.py +4 -3
- mindspore/{nn → train}/metrics/perplexity.py +5 -4
- mindspore/{nn → train}/metrics/precision.py +5 -4
- mindspore/{nn → train}/metrics/recall.py +5 -4
- mindspore/{nn → train}/metrics/roc.py +7 -6
- mindspore/{nn → train}/metrics/root_mean_square_surface_distance.py +6 -5
- mindspore/{nn → train}/metrics/topk.py +7 -5
- mindspore/train/mind_ir_pb2.py +339 -32
- mindspore/train/model.py +113 -84
- mindspore/train/serialization.py +547 -167
- mindspore/train/summary/_summary_adapter.py +1 -1
- mindspore/train/summary/summary_record.py +43 -12
- mindspore/train/train_thor/convert_utils.py +7 -1
- mindspore/train/train_thor/dataset_helper.py +3 -3
- mindspore/train/train_thor/model_thor.py +0 -4
- mindspore/turbojpeg.dll +0 -0
- mindspore/vcmeta.dll +0 -0
- mindspore/vcruntime140.dll +0 -0
- mindspore/vcruntime140_1.dll +0 -0
- mindspore/version.py +1 -1
- {mindspore-1.10.0.dist-info → mindspore-2.0.0rc1.dist-info}/METADATA +4 -3
- {mindspore-1.10.0.dist-info → mindspore-2.0.0rc1.dist-info}/RECORD +901 -660
- mindspore/compression/common/constant.py +0 -124
- mindspore/compression/export/__init__.py +0 -19
- mindspore/compression/export/quant_export.py +0 -514
- mindspore/compression/quant/qat.py +0 -636
- mindspore/compression/quant/quant_utils.py +0 -462
- mindspore/compression/quant/quantizer.py +0 -68
- mindspore/libatomic-1.dll +0 -0
- mindspore/libgcc_s_seh-1.dll +0 -0
- mindspore/libgfortran-4.dll +0 -0
- mindspore/libgomp-1.dll +0 -0
- mindspore/libjpeg-62.dll +0 -0
- mindspore/libmindspore.dll +0 -0
- mindspore/libmindspore_common.dll +0 -0
- mindspore/libmindspore_core.dll +0 -0
- mindspore/libmindspore_glog.dll +0 -0
- mindspore/libnnacl.dll +0 -0
- mindspore/libopencv_core452.dll +0 -0
- mindspore/libopencv_imgcodecs452.dll +0 -0
- mindspore/libopencv_imgproc452.dll +0 -0
- mindspore/libquadmath-0.dll +0 -0
- mindspore/libsqlite3.dll +0 -0
- mindspore/libssp-0.dll +0 -0
- mindspore/libstdc++-6.dll +0 -0
- mindspore/libtinyxml2.dll +0 -0
- mindspore/libturbojpeg.dll +0 -0
- mindspore/libwinpthread-1.dll +0 -0
- mindspore/nn/layer/quant.py +0 -1868
- mindspore/nn/layer/rnn_utils.py +0 -90
- mindspore/nn/probability/dpn/__init__.py +0 -22
- mindspore/nn/probability/dpn/vae/__init__.py +0 -25
- mindspore/nn/probability/dpn/vae/cvae.py +0 -138
- mindspore/nn/probability/dpn/vae/vae.py +0 -122
- mindspore/nn/probability/infer/__init__.py +0 -22
- mindspore/nn/probability/infer/variational/elbo.py +0 -70
- mindspore/nn/probability/infer/variational/svi.py +0 -84
- mindspore/nn/probability/toolbox/__init__.py +0 -22
- mindspore/nn/probability/toolbox/anomaly_detection.py +0 -99
- mindspore/nn/probability/toolbox/uncertainty_evaluation.py +0 -363
- mindspore/nn/probability/transforms/__init__.py +0 -22
- mindspore/nn/probability/transforms/transform_bnn.py +0 -262
- mindspore/nn/probability/zhusuan/__init__.py +0 -18
- mindspore/nn/probability/zhusuan/framework/__init__.py +0 -18
- mindspore/nn/probability/zhusuan/framework/bn.py +0 -95
- mindspore/nn/probability/zhusuan/variational/__init__.py +0 -18
- mindspore/nn/probability/zhusuan/variational/elbo.py +0 -46
- mindspore/ops/_op_impl/tbe/bias_add_grad_ds.py +0 -52
- mindspore/ops/_op_impl/tbe/scatter_nd_add_ds.py +0 -43
- mindspore/ops/bprop_mindir/AssignAdd_bprop.mindir +0 -20
- mindspore/ops/bprop_mindir/Identity_bprop.mindir +0 -9
- mindspore/ops/bprop_mindir/LogicalOr_bprop.mindir +0 -20
- mindspore/ops/bprop_mindir/ReLU_bprop.mindir +0 -16
- mindspore/ops/bprop_mindir/UpdateState_bprop.mindir +0 -17
- mindspore/ops/bprop_mindir/stop_gradient_bprop.mindir +0 -12
- mindspore/ops/composite/array_ops.py +0 -210
- mindspore/ops/composite/clip_ops.py +0 -238
- mindspore/ops/composite/random_ops.py +0 -426
- mindspore/ops/composite/vmap_ops.py +0 -38
- mindspore/ops/operations/sponge_ops.py +0 -3531
- mindspore/ops/operations/sponge_update_ops.py +0 -2546
- mindspore/parallel/nn/__init__.py +0 -42
- mindspore/parallel/nn/loss.py +0 -22
- mindspore/parallel/nn/moe.py +0 -21
- mindspore/parallel/nn/op_parallel_config.py +0 -22
- mindspore/parallel/nn/transformer.py +0 -31
- mindspore/run_check/_check_deps_version.py +0 -84
- {mindspore-1.10.0.dist-info → mindspore-2.0.0rc1.dist-info}/WHEEL +0 -0
- {mindspore-1.10.0.dist-info → mindspore-2.0.0rc1.dist-info}/entry_points.txt +0 -0
- {mindspore-1.10.0.dist-info → mindspore-2.0.0rc1.dist-info}/top_level.txt +0 -0
mindspore/train/model.py
CHANGED
|
@@ -23,28 +23,30 @@ import math
|
|
|
23
23
|
import copy
|
|
24
24
|
import numpy as np
|
|
25
25
|
|
|
26
|
+
import mindspore
|
|
26
27
|
from mindspore import log as logger
|
|
27
28
|
from mindspore.train.serialization import save_checkpoint, load_checkpoint
|
|
28
29
|
from mindspore.train.callback._checkpoint import ModelCheckpoint, _chg_ckpt_file_name_if_same_exist
|
|
29
30
|
from mindspore.common.tensor import Tensor
|
|
30
|
-
from mindspore.
|
|
31
|
-
from mindspore._checkparam import check_input_data, check_output_data
|
|
31
|
+
from mindspore.train.metrics import get_metrics, get_metric_fn
|
|
32
|
+
from mindspore._checkparam import check_input_data, check_output_data
|
|
33
|
+
from mindspore import _checkparam as Validator
|
|
32
34
|
from mindspore.train.callback import _InternalCallbackParam, RunContext, _CallbackManager, Callback, TimeMonitor
|
|
33
35
|
from mindspore.train.callback import __all__ as internal_cb_names
|
|
34
36
|
from mindspore import context
|
|
35
|
-
from mindspore.parallel._utils import _get_parallel_mode, _get_device_num,
|
|
36
|
-
|
|
37
|
+
from mindspore.parallel._utils import _get_parallel_mode, _get_device_num, _get_parameter_broadcast, \
|
|
38
|
+
_device_number_check, _parameter_broadcast_check, _parallel_predict_check, \
|
|
37
39
|
_reset_op_id_with_offset
|
|
38
40
|
from mindspore.parallel._ps_context import _is_role_worker, _is_role_pserver, _is_role_sched, _is_ps_mode, \
|
|
39
41
|
_cache_enable, _enable_distributed_mindrt
|
|
40
|
-
from mindspore.
|
|
42
|
+
from mindspore.train.metrics import Loss
|
|
41
43
|
from mindspore import nn
|
|
42
44
|
from mindspore.boost import AutoBoost
|
|
43
45
|
from mindspore.context import ParallelMode
|
|
44
|
-
from mindspore.parallel._cost_model_context import _set_multi_subgraphs
|
|
45
46
|
from mindspore.parallel._recovery_context import _set_recovery_context, _get_recovery_context
|
|
46
47
|
from mindspore.train.dataset_helper import DatasetHelper, connect_network_with_dataset
|
|
47
48
|
from mindspore.common.api import _pynative_executor
|
|
49
|
+
from mindspore.dataset.core.config import get_debug_mode
|
|
48
50
|
from mindspore.dataset.engine.datasets import _set_training_dataset, _reset_training_dataset
|
|
49
51
|
from mindspore.train import amp
|
|
50
52
|
|
|
@@ -125,14 +127,15 @@ class Model:
|
|
|
125
127
|
three elements: the positions of loss value, predicted value and label in outputs of the
|
|
126
128
|
`eval_network`. In this case, the loss value will be passed to the `Loss` metric, the
|
|
127
129
|
predicted value and label will be passed to other metrics.
|
|
128
|
-
:func
|
|
130
|
+
:func:`mindspore.train.Metric.set_indexes` is recommended instead of `eval_indexes`.
|
|
129
131
|
Default: None.
|
|
130
|
-
amp_level (str): Option for argument `level` in :func:`mindspore.build_train_network`, level for mixed
|
|
132
|
+
amp_level (str): Option for argument `level` in :func:`mindspore.amp.build_train_network`, level for mixed
|
|
131
133
|
precision training. Supports ["O0", "O1", "O2", "O3", "auto"]. Default: "O0".
|
|
132
134
|
|
|
133
135
|
- "O0": Do not change.
|
|
134
136
|
- "O1": Cast the operators in white_list to float16, the remaining operators are kept in float32.
|
|
135
|
-
|
|
137
|
+
The operators in the whitelist: [Conv1d, Conv2d, Conv3d, Conv1dTranspose, Conv2dTranspose,
|
|
138
|
+
Conv3dTranspose, Dense, LSTMCell, RNNCell, GRUCell, MatMul, BatchMatMul, PReLU, ReLU, Ger].
|
|
136
139
|
- "O2": Cast network to float16, keep BatchNorm run in float32, using dynamic loss scale.
|
|
137
140
|
- "O3": Cast network to float16, the BatchNorm is also cast to float16, loss scale will not be used.
|
|
138
141
|
- auto: Set level to recommended level in different devices. Set level to "O2" on GPU, set
|
|
@@ -143,7 +146,7 @@ class Model:
|
|
|
143
146
|
The BatchNorm strategy can be changed by `keep_batchnorm_fp32` settings in `kwargs`. `keep_batchnorm_fp32`
|
|
144
147
|
must be a bool. The loss scale strategy can be changed by `loss_scale_manager` setting in `kwargs`.
|
|
145
148
|
`loss_scale_manager` should be a subclass of :class:`mindspore.amp.LossScaleManager`.
|
|
146
|
-
The more detailed explanation of `amp_level` setting can be found at `mindspore.build_train_network`.
|
|
149
|
+
The more detailed explanation of `amp_level` setting can be found at `mindspore.amp.build_train_network`.
|
|
147
150
|
|
|
148
151
|
boost_level (str): Option for argument `level` in `mindspore.boost`, level for boost mode
|
|
149
152
|
training. Supports ["O0", "O1", "O2"]. Default: "O0".
|
|
@@ -163,8 +166,8 @@ class Model:
|
|
|
163
166
|
the Graph mode + Ascend platform, and for better acceleration, refer to the documentation to configure
|
|
164
167
|
boost_config_dict.
|
|
165
168
|
Examples:
|
|
166
|
-
>>> import mindspore as ms
|
|
167
169
|
>>> from mindspore import nn
|
|
170
|
+
>>> from mindspore.train import Model
|
|
168
171
|
>>>
|
|
169
172
|
>>> class Net(nn.Cell):
|
|
170
173
|
... def __init__(self, num_class=10, num_channel=1):
|
|
@@ -190,10 +193,10 @@ class Model:
|
|
|
190
193
|
>>> net = Net()
|
|
191
194
|
>>> loss = nn.SoftmaxCrossEntropyWithLogits()
|
|
192
195
|
>>> optim = nn.Momentum(params=net.trainable_params(), learning_rate=0.1, momentum=0.9)
|
|
193
|
-
>>> model =
|
|
196
|
+
>>> model = Model(net, loss_fn=loss, optimizer=optim, metrics=None)
|
|
194
197
|
>>> # For details about how to build the dataset, please refer to the variable `dataset_train` in tutorial
|
|
195
198
|
>>> # document on the official website:
|
|
196
|
-
>>> # https://www.mindspore.cn/tutorials/zh-CN/
|
|
199
|
+
>>> # https://www.mindspore.cn/tutorials/zh-CN/r2.0/beginner/quick_start.html
|
|
197
200
|
>>> dataset = create_custom_dataset()
|
|
198
201
|
>>> model.train(2, dataset)
|
|
199
202
|
"""
|
|
@@ -213,7 +216,6 @@ class Model:
|
|
|
213
216
|
self._process_amp_args(kwargs)
|
|
214
217
|
self._parallel_mode = _get_parallel_mode()
|
|
215
218
|
self._device_number = _get_device_num()
|
|
216
|
-
self._global_rank = _get_global_rank()
|
|
217
219
|
self._parameter_broadcast = _get_parameter_broadcast()
|
|
218
220
|
self._metrics = metrics
|
|
219
221
|
|
|
@@ -323,11 +325,6 @@ class Model:
|
|
|
323
325
|
network = nn.WithLossCell(network, self._loss_fn)
|
|
324
326
|
# If need to check if loss_fn is not None, but optimizer is None
|
|
325
327
|
|
|
326
|
-
if self._parallel_mode in (ParallelMode.SEMI_AUTO_PARALLEL, ParallelMode.AUTO_PARALLEL):
|
|
327
|
-
network.set_auto_parallel()
|
|
328
|
-
if self._optimizer is None:
|
|
329
|
-
# In this case, multiple optimizer(s) is supposed to be included in 'self._network'
|
|
330
|
-
_set_multi_subgraphs()
|
|
331
328
|
if net_inputs is not None:
|
|
332
329
|
network.set_inputs(*net_inputs)
|
|
333
330
|
return network
|
|
@@ -368,18 +365,10 @@ class Model:
|
|
|
368
365
|
self._eval_network.set_inputs(*net_inputs)
|
|
369
366
|
self._eval_indexes = [0, 1, 2]
|
|
370
367
|
|
|
371
|
-
if self._parallel_mode in (ParallelMode.SEMI_AUTO_PARALLEL, ParallelMode.AUTO_PARALLEL):
|
|
372
|
-
if self._optimizer is None:
|
|
373
|
-
# In this case, multiple optimizer(s) is supposed to be included in 'self._network'
|
|
374
|
-
_set_multi_subgraphs()
|
|
375
|
-
self._eval_network.set_auto_parallel()
|
|
376
|
-
|
|
377
368
|
def _build_predict_network(self):
|
|
378
369
|
"""Build the network for prediction."""
|
|
379
370
|
self._predict_network = self._network
|
|
380
|
-
|
|
381
|
-
# Unlike the cases in build_train_network() and build_eval_network(), 'multi_subgraphs' is not set
|
|
382
|
-
self._predict_network.set_auto_parallel()
|
|
371
|
+
# Unlike the cases in build_train_network() and build_eval_network(), 'multi_subgraphs' is not set
|
|
383
372
|
|
|
384
373
|
def _clear_metrics(self):
|
|
385
374
|
"""Clear metrics local values."""
|
|
@@ -408,13 +397,12 @@ class Model:
|
|
|
408
397
|
def _get_metrics(self):
|
|
409
398
|
"""Get metrics local values."""
|
|
410
399
|
metrics = dict()
|
|
411
|
-
#
|
|
412
|
-
is_embedding_cache_server = _is_role_pserver() and _cache_enable()
|
|
400
|
+
# There's no need for server to execute eval, just give fake metrics.
|
|
413
401
|
for key, value in self._metric_fns.items():
|
|
414
|
-
if not
|
|
402
|
+
if not _is_role_pserver():
|
|
415
403
|
metrics[key] = value.eval()
|
|
416
404
|
else:
|
|
417
|
-
metrics[key] =
|
|
405
|
+
metrics[key] = 1
|
|
418
406
|
return metrics
|
|
419
407
|
|
|
420
408
|
def _get_scaling_sens(self):
|
|
@@ -444,7 +432,7 @@ class Model:
|
|
|
444
432
|
if dataset_sink_mode:
|
|
445
433
|
network = connect_network_with_dataset(network, dataset_helper)
|
|
446
434
|
|
|
447
|
-
if is_train:
|
|
435
|
+
if _get_recovery_context("enable_recovery") and is_train:
|
|
448
436
|
_set_training_dataset(dataset_helper)
|
|
449
437
|
|
|
450
438
|
|
|
@@ -452,9 +440,6 @@ class Model:
|
|
|
452
440
|
network.phase = phase
|
|
453
441
|
self._backbone_is_train = is_train
|
|
454
442
|
|
|
455
|
-
if self._parallel_mode in (ParallelMode.SEMI_AUTO_PARALLEL, ParallelMode.AUTO_PARALLEL):
|
|
456
|
-
network.set_auto_parallel()
|
|
457
|
-
|
|
458
443
|
return dataset_helper, network
|
|
459
444
|
|
|
460
445
|
def _check_network_mode(self, network, is_train):
|
|
@@ -515,6 +500,10 @@ class Model:
|
|
|
515
500
|
_device_number_check(self._parallel_mode, self._device_number)
|
|
516
501
|
|
|
517
502
|
if train_dataset:
|
|
503
|
+
if not isinstance(train_dataset, mindspore.dataset.Dataset):
|
|
504
|
+
raise TypeError("The type of 'train_dataset' must be `Dataset`, "
|
|
505
|
+
"but got {}.".format(type(train_dataset)))
|
|
506
|
+
|
|
518
507
|
_parameter_broadcast_check(self._parallel_mode, self._parameter_broadcast)
|
|
519
508
|
if self._parameter_broadcast:
|
|
520
509
|
self._train_network.set_broadcast_flag()
|
|
@@ -532,6 +521,9 @@ class Model:
|
|
|
532
521
|
break
|
|
533
522
|
|
|
534
523
|
if valid_dataset:
|
|
524
|
+
if not isinstance(valid_dataset, mindspore.dataset.Dataset):
|
|
525
|
+
raise TypeError("The type of 'valid_dataset' must be `Dataset`, "
|
|
526
|
+
"but got {}.".format(type(valid_dataset)))
|
|
535
527
|
if not self._metric_fns:
|
|
536
528
|
raise RuntimeError("If define `valid_dataset`, metric fn can not be None or empty, "
|
|
537
529
|
"you should set the argument 'metrics' for model.")
|
|
@@ -603,7 +595,7 @@ class Model:
|
|
|
603
595
|
callbacks = cb_params.list_callback
|
|
604
596
|
cb_params.train_dataset_element = None
|
|
605
597
|
cb_params.network = self._network
|
|
606
|
-
if
|
|
598
|
+
if _is_role_sched():
|
|
607
599
|
epoch = 1
|
|
608
600
|
# Embedding cache server only run one step.
|
|
609
601
|
if (_is_role_pserver() or _is_role_sched()) and _cache_enable():
|
|
@@ -686,7 +678,7 @@ class Model:
|
|
|
686
678
|
cb_params.train_network = train_network
|
|
687
679
|
|
|
688
680
|
# Perform recovery for process which is restarted.
|
|
689
|
-
self._reset_training_step_for_abnormal_process(cb_params)
|
|
681
|
+
self._reset_training_step_for_abnormal_process(cb_params, dataset_helper)
|
|
690
682
|
# Perform recovery for process which is not restarted.
|
|
691
683
|
self._reset_training_step_for_normal_process(cb_params, dataset_helper)
|
|
692
684
|
|
|
@@ -702,19 +694,27 @@ class Model:
|
|
|
702
694
|
train_network = self._check_network_mode(train_network, True)
|
|
703
695
|
outputs = train_network(*inputs)
|
|
704
696
|
cb_params.net_outputs = outputs
|
|
697
|
+
|
|
698
|
+
if _is_role_sched():
|
|
699
|
+
os._exit(0)
|
|
700
|
+
|
|
705
701
|
# In disaster recovery scenarios, need not to execute callbacks if this step executes failed.
|
|
706
702
|
need_exec_callback_step_end = not (self.enable_recovery and _get_recovery_context("need_reset"))
|
|
707
703
|
if need_exec_callback_step_end:
|
|
708
704
|
list_callback.on_train_step_end(run_context)
|
|
709
705
|
|
|
710
|
-
if (_is_role_pserver() and not _enable_distributed_mindrt()) or _is_role_sched():
|
|
711
|
-
os._exit(0)
|
|
712
706
|
# Embedding cache server only run one step.
|
|
713
707
|
if is_embedding_cache_server:
|
|
714
708
|
break
|
|
715
709
|
|
|
716
710
|
dataset_helper.continue_send()
|
|
717
711
|
|
|
712
|
+
# When it's distributed training and using MindRT,
|
|
713
|
+
# the node id should be reset to start from 0.
|
|
714
|
+
# This is to avoid the timeout when finding the actor route tables in 'train' and 'eval' case(or 'fit').
|
|
715
|
+
if _enable_distributed_mindrt():
|
|
716
|
+
_reset_op_id_with_offset()
|
|
717
|
+
|
|
718
718
|
self._eval_during_train(valid_infos, cb_params, list_callback)
|
|
719
719
|
|
|
720
720
|
# In disaster recovery scenarios, need not to execute callbacks if this epoch executes failed.
|
|
@@ -809,7 +809,7 @@ class Model:
|
|
|
809
809
|
else:
|
|
810
810
|
self.need_load_ckpt = False
|
|
811
811
|
|
|
812
|
-
def _reset_training_step_for_abnormal_process(self, cb_params):
|
|
812
|
+
def _reset_training_step_for_abnormal_process(self, cb_params, dataset_helper):
|
|
813
813
|
"""
|
|
814
814
|
Execute recovery for abnormal exit process when restart.
|
|
815
815
|
|
|
@@ -824,7 +824,7 @@ class Model:
|
|
|
824
824
|
os.remove(cb_params.latest_ckpt_file)
|
|
825
825
|
raise RuntimeError(e.__str__() + ", load ckpt failed and remove the ckpt: "\
|
|
826
826
|
+ cb_params.latest_ckpt_file) from e
|
|
827
|
-
_reset_training_dataset(cb_params.cur_step_num)
|
|
827
|
+
_reset_training_dataset(cb_params.cur_step_num, dataset_helper.sink_size())
|
|
828
828
|
self.need_load_ckpt = False
|
|
829
829
|
|
|
830
830
|
def _reset_training_step_for_normal_process(self, cb_params, dataset_helper):
|
|
@@ -853,9 +853,9 @@ class Model:
|
|
|
853
853
|
self.epoch_iter = recovery_epoch_num
|
|
854
854
|
cb_params.cur_epoch_num = self.epoch_iter + 1
|
|
855
855
|
cb_params.last_save_ckpt_step = cb_params.cur_step_num
|
|
856
|
-
_reset_training_dataset(cb_params.cur_step_num)
|
|
856
|
+
_reset_training_dataset(cb_params.cur_step_num, dataset_helper.sink_size())
|
|
857
857
|
else:
|
|
858
|
-
_reset_training_dataset(0)
|
|
858
|
+
_reset_training_dataset(0, dataset_helper.sink_size())
|
|
859
859
|
|
|
860
860
|
_set_recovery_context(need_reset=False)
|
|
861
861
|
|
|
@@ -909,12 +909,12 @@ class Model:
|
|
|
909
909
|
outputs = self._train_network(*next_element)
|
|
910
910
|
cb_params.net_outputs = outputs
|
|
911
911
|
if self._loss_scale_manager and self._loss_scale_manager.get_drop_overflow_update():
|
|
912
|
-
|
|
912
|
+
overflow = outputs[1]
|
|
913
913
|
overflow = np.all(overflow.asnumpy())
|
|
914
914
|
self._loss_scale_manager.update_loss_scale(overflow)
|
|
915
915
|
|
|
916
916
|
list_callback.on_train_step_end(run_context)
|
|
917
|
-
if
|
|
917
|
+
if _is_role_sched():
|
|
918
918
|
os._exit(0)
|
|
919
919
|
# Embedding cache server only run one step.
|
|
920
920
|
if is_embedding_cache_server:
|
|
@@ -923,6 +923,12 @@ class Model:
|
|
|
923
923
|
if should_stop:
|
|
924
924
|
break
|
|
925
925
|
|
|
926
|
+
# When it's distributed training and using MindRT,
|
|
927
|
+
# the node id should be reset to start from 0.
|
|
928
|
+
# This is to avoid the timeout when finding the actor route tables in 'train' and 'eval' case(or 'fit').
|
|
929
|
+
if _enable_distributed_mindrt():
|
|
930
|
+
_reset_op_id_with_offset()
|
|
931
|
+
|
|
926
932
|
self._eval_during_train(valid_infos, cb_params, list_callback)
|
|
927
933
|
|
|
928
934
|
train_dataset.reset()
|
|
@@ -942,7 +948,7 @@ class Model:
|
|
|
942
948
|
|
|
943
949
|
list_callback.on_train_end(run_context)
|
|
944
950
|
|
|
945
|
-
def train(self, epoch, train_dataset, callbacks=None, dataset_sink_mode=
|
|
951
|
+
def train(self, epoch, train_dataset, callbacks=None, dataset_sink_mode=False, sink_size=-1, initial_epoch=0):
|
|
946
952
|
"""
|
|
947
953
|
Training API.
|
|
948
954
|
|
|
@@ -980,7 +986,7 @@ class Model:
|
|
|
980
986
|
Default: None.
|
|
981
987
|
dataset_sink_mode (bool): Determines whether to pass the data through dataset channel.
|
|
982
988
|
Configure pynative mode or CPU, the training process will be performed with
|
|
983
|
-
dataset not sink. Default:
|
|
989
|
+
dataset not sink. Default: False.
|
|
984
990
|
sink_size (int): Control the amount of data in each sink. `sink_size` is invalid if `dataset_sink_mode`
|
|
985
991
|
is False.
|
|
986
992
|
If sink_size = -1, sink the complete dataset for each epoch.
|
|
@@ -990,20 +996,25 @@ class Model:
|
|
|
990
996
|
Default: 0.
|
|
991
997
|
|
|
992
998
|
Examples:
|
|
993
|
-
>>> import mindspore as ms
|
|
994
999
|
>>> from mindspore import nn
|
|
1000
|
+
>>> from mindspore.train import Model
|
|
995
1001
|
>>>
|
|
996
1002
|
>>> # For details about how to build the dataset, please refer to the tutorial
|
|
997
1003
|
>>> # document on the official website.
|
|
998
1004
|
>>> dataset = create_custom_dataset()
|
|
999
1005
|
>>> net = Net()
|
|
1000
1006
|
>>> loss = nn.SoftmaxCrossEntropyWithLogits()
|
|
1001
|
-
>>> loss_scale_manager = ms.FixedLossScaleManager()
|
|
1007
|
+
>>> loss_scale_manager = ms.FixedLossScaleManager(1024., False)
|
|
1002
1008
|
>>> optim = nn.Momentum(params=net.trainable_params(), learning_rate=0.1, momentum=0.9)
|
|
1003
|
-
>>> model =
|
|
1009
|
+
>>> model = Model(net, loss_fn=loss, optimizer=optim, metrics=None,
|
|
1004
1010
|
... loss_scale_manager=loss_scale_manager)
|
|
1005
1011
|
>>> model.train(2, dataset)
|
|
1006
1012
|
"""
|
|
1013
|
+
device_target = context.get_context("device_target")
|
|
1014
|
+
if _is_ps_mode() and not _cache_enable() and (device_target in ["Ascend", "CPU"]) and dataset_sink_mode:
|
|
1015
|
+
logger.info("For PS mode, reset datasink mode to False when using Ascend or CPU backend.")
|
|
1016
|
+
dataset_sink_mode = False
|
|
1017
|
+
|
|
1007
1018
|
Validator.check_bool(dataset_sink_mode)
|
|
1008
1019
|
if isinstance(self._train_network, nn.GraphCell) and dataset_sink_mode:
|
|
1009
1020
|
raise ValueError("Dataset sink mode is currently not supported when training with a GraphCell.")
|
|
@@ -1016,11 +1027,10 @@ class Model:
|
|
|
1016
1027
|
|
|
1017
1028
|
# Parameter server and embedding cache mode check.
|
|
1018
1029
|
if _is_ps_mode():
|
|
1019
|
-
if dataset_sink_mode and not _cache_enable():
|
|
1020
|
-
raise ValueError("Parameter server mode does not support 'data_sink_mode=True'.")
|
|
1021
1030
|
if not dataset_sink_mode and _cache_enable():
|
|
1022
|
-
raise ValueError("Embedding cache mode should run with '
|
|
1031
|
+
raise ValueError("Embedding cache mode should run with 'dataset_sink_mode=True'.")
|
|
1023
1032
|
|
|
1033
|
+
self._check_sink_mode_for_ds_debug_mode(dataset_sink_mode)
|
|
1024
1034
|
|
|
1025
1035
|
Validator.check_is_int(sink_size)
|
|
1026
1036
|
Validator.check_non_negative_int(epoch)
|
|
@@ -1050,11 +1060,18 @@ class Model:
|
|
|
1050
1060
|
sink_size=sink_size,
|
|
1051
1061
|
initial_epoch=initial_epoch)
|
|
1052
1062
|
|
|
1053
|
-
# When it's
|
|
1063
|
+
# When it's distributed training and using MindRT,
|
|
1054
1064
|
# the node id should be reset to start from 0.
|
|
1065
|
+
# This is to avoid the timeout when finding the actor route tables in 'train' and 'eval' case(or 'fit').
|
|
1055
1066
|
if _enable_distributed_mindrt():
|
|
1056
1067
|
_reset_op_id_with_offset()
|
|
1057
1068
|
|
|
1069
|
+
@staticmethod
|
|
1070
|
+
def _check_sink_mode_for_ds_debug_mode(dataset_sink_mode):
|
|
1071
|
+
if get_debug_mode() and dataset_sink_mode:
|
|
1072
|
+
raise ValueError("Dataset sink mode is not supported when dataset pipeline debug mode is on. "
|
|
1073
|
+
"Please manually turn off sink mode.")
|
|
1074
|
+
|
|
1058
1075
|
@staticmethod
|
|
1059
1076
|
def _check_methods_for_custom_callbacks(callbacks, current_mode):
|
|
1060
1077
|
"""
|
|
@@ -1083,13 +1100,14 @@ class Model:
|
|
|
1083
1100
|
"using customized callbacks." % (cb_name, invalid_methods_names))
|
|
1084
1101
|
|
|
1085
1102
|
def fit(self, epoch, train_dataset, valid_dataset=None, valid_frequency=1, callbacks=None,
|
|
1086
|
-
dataset_sink_mode=
|
|
1103
|
+
dataset_sink_mode=False, valid_dataset_sink_mode=False, sink_size=-1, initial_epoch=0):
|
|
1087
1104
|
"""
|
|
1088
1105
|
Fit API.
|
|
1089
1106
|
|
|
1090
1107
|
Evaluation process will be performed during training process if `valid_dataset` is provided.
|
|
1091
1108
|
|
|
1092
|
-
More details please refer to
|
|
1109
|
+
More details please refer to :func:`mindspore.train.Model.train` and
|
|
1110
|
+
:func:`mindspore.train.Model.eval`.
|
|
1093
1111
|
|
|
1094
1112
|
Args:
|
|
1095
1113
|
epoch (int): Total training epochs. Generally, train network will be trained on complete dataset per epoch.
|
|
@@ -1115,9 +1133,9 @@ class Model:
|
|
|
1115
1133
|
Default: None.
|
|
1116
1134
|
dataset_sink_mode (bool): Determines whether to pass the train data through dataset channel.
|
|
1117
1135
|
Configure pynative mode or CPU, the training process will be performed with
|
|
1118
|
-
dataset not sink. Default:
|
|
1136
|
+
dataset not sink. Default: False.
|
|
1119
1137
|
valid_dataset_sink_mode (bool): Determines whether to pass the validation data through dataset channel.
|
|
1120
|
-
Default:
|
|
1138
|
+
Default: False.
|
|
1121
1139
|
sink_size (int): Control the amount of data in each sink. `sink_size` is invalid if `dataset_sink_mode`
|
|
1122
1140
|
is False.
|
|
1123
1141
|
If sink_size = -1, sink the complete dataset for each epoch.
|
|
@@ -1127,8 +1145,8 @@ class Model:
|
|
|
1127
1145
|
Default: 0.
|
|
1128
1146
|
|
|
1129
1147
|
Examples:
|
|
1130
|
-
>>> import mindspore as ms
|
|
1131
1148
|
>>> from mindspore import nn
|
|
1149
|
+
>>> from mindspore.train import Model
|
|
1132
1150
|
>>>
|
|
1133
1151
|
>>> # For details about how to build the dataset, please refer to the tutorial
|
|
1134
1152
|
>>> # document on the official website.
|
|
@@ -1137,9 +1155,13 @@ class Model:
|
|
|
1137
1155
|
>>> net = Net()
|
|
1138
1156
|
>>> loss = nn.SoftmaxCrossEntropyWithLogits()
|
|
1139
1157
|
>>> optim = nn.Momentum(params=net.trainable_params(), learning_rate=0.1, momentum=0.9)
|
|
1140
|
-
>>> model =
|
|
1158
|
+
>>> model = Model(net, loss_fn=loss, optimizer=optim, metrics={"accuracy"})
|
|
1141
1159
|
>>> model.fit(2, train_dataset, valid_dataset)
|
|
1142
1160
|
"""
|
|
1161
|
+
device_target = context.get_context("device_target")
|
|
1162
|
+
if _is_ps_mode() and not _cache_enable() and (device_target in ["Ascend", "CPU"]) and dataset_sink_mode:
|
|
1163
|
+
logger.info("For PS mode, reset datasink mode to False when using Ascend or CPU backend.")
|
|
1164
|
+
dataset_sink_mode = False
|
|
1143
1165
|
|
|
1144
1166
|
dataset_sink_mode = Validator.check_bool(dataset_sink_mode)
|
|
1145
1167
|
valid_dataset_sink_mode = Validator.check_bool(valid_dataset_sink_mode)
|
|
@@ -1152,9 +1174,6 @@ class Model:
|
|
|
1152
1174
|
"should be equal to value in Model.fit, but got {} and {} separately."
|
|
1153
1175
|
.format(train_dataset._warmup_epoch, epoch))
|
|
1154
1176
|
|
|
1155
|
-
if dataset_sink_mode and _is_ps_mode() and not _cache_enable():
|
|
1156
|
-
raise ValueError("Parameter server mode does not support 'data_sink_mode=True'.")
|
|
1157
|
-
|
|
1158
1177
|
Validator.check_is_int(sink_size)
|
|
1159
1178
|
Validator.check_non_negative_int(epoch)
|
|
1160
1179
|
Validator.check_non_negative_int(initial_epoch)
|
|
@@ -1173,7 +1192,7 @@ class Model:
|
|
|
1173
1192
|
_device_number_check(self._parallel_mode, self._device_number)
|
|
1174
1193
|
|
|
1175
1194
|
if not isinstance(valid_frequency, (int, list)):
|
|
1176
|
-
raise TypeError(f"For 'Model.fit', the type of 'valid_frequency' must be a list or
|
|
1195
|
+
raise TypeError(f"For 'Model.fit', the type of 'valid_frequency' must be a list or an integer, but got "
|
|
1177
1196
|
f"type {type(valid_frequency)}.")
|
|
1178
1197
|
|
|
1179
1198
|
if valid_dataset and not self._metric_fns:
|
|
@@ -1196,7 +1215,7 @@ class Model:
|
|
|
1196
1215
|
Build computational graphs and data graphs with the sink mode.
|
|
1197
1216
|
|
|
1198
1217
|
.. warning::
|
|
1199
|
-
This is an experimental
|
|
1218
|
+
This is an experimental API that is subject to change or deletion.
|
|
1200
1219
|
|
|
1201
1220
|
Note:
|
|
1202
1221
|
The interface builds the computational graphs, when the interface is executed first, 'Model.train' only
|
|
@@ -1212,17 +1231,18 @@ class Model:
|
|
|
1212
1231
|
epoch (int): Control the training epochs. Default: 1.
|
|
1213
1232
|
|
|
1214
1233
|
Examples:
|
|
1215
|
-
>>> import mindspore as ms
|
|
1216
1234
|
>>> from mindspore import nn
|
|
1235
|
+
>>> from mindspore.train import Model
|
|
1236
|
+
>>> from mindspore.amp import FixedLossScaleManager
|
|
1217
1237
|
>>>
|
|
1218
1238
|
>>> # For details about how to build the dataset, please refer to the tutorial
|
|
1219
1239
|
>>> # document on the official website.
|
|
1220
1240
|
>>> dataset = create_custom_dataset()
|
|
1221
1241
|
>>> net = Net()
|
|
1222
1242
|
>>> loss = nn.SoftmaxCrossEntropyWithLogits()
|
|
1223
|
-
>>> loss_scale_manager =
|
|
1243
|
+
>>> loss_scale_manager = FixedLossScaleManager()
|
|
1224
1244
|
>>> optim = nn.Momentum(params=net.trainable_params(), learning_rate=0.1, momentum=0.9)
|
|
1225
|
-
>>> model =
|
|
1245
|
+
>>> model = Model(net, loss_fn=loss, optimizer=optim, metrics=None,
|
|
1226
1246
|
... loss_scale_manager=loss_scale_manager)
|
|
1227
1247
|
>>> model.build(dataset, epoch=2)
|
|
1228
1248
|
>>> model.train(2, dataset)
|
|
@@ -1231,7 +1251,7 @@ class Model:
|
|
|
1231
1251
|
|
|
1232
1252
|
def _eval_in_fit(self, valid_dataset, callbacks=None, dataset_sink_mode=True, cb_params=None):
|
|
1233
1253
|
"""
|
|
1234
|
-
Evaluation process in `mindspore.Model.fit`.
|
|
1254
|
+
Evaluation process in `mindspore.train.Model.fit`.
|
|
1235
1255
|
|
|
1236
1256
|
Args:
|
|
1237
1257
|
valid_dataset (Dataset): Dataset to evaluate the model. If `valid_dataset` is provided, evaluation process
|
|
@@ -1286,6 +1306,8 @@ class Model:
|
|
|
1286
1306
|
list_callback.on_eval_epoch_begin(run_context)
|
|
1287
1307
|
for inputs in dataset_helper:
|
|
1288
1308
|
cb_params.cur_step_num += 1
|
|
1309
|
+
inputs = _transfer_tensor_to_tuple(inputs)
|
|
1310
|
+
cb_params.eval_dataset_element = inputs
|
|
1289
1311
|
list_callback.on_eval_step_begin(run_context)
|
|
1290
1312
|
eval_network = self._check_network_mode(eval_network, False)
|
|
1291
1313
|
outputs = eval_network(*inputs)
|
|
@@ -1330,8 +1352,9 @@ class Model:
|
|
|
1330
1352
|
list_callback.on_eval_epoch_begin(run_context)
|
|
1331
1353
|
for next_element in dataset_helper:
|
|
1332
1354
|
cb_params.cur_step_num += 1
|
|
1333
|
-
list_callback.on_eval_step_begin(run_context)
|
|
1334
1355
|
next_element = _transfer_tensor_to_tuple(next_element)
|
|
1356
|
+
cb_params.eval_dataset_element = next_element
|
|
1357
|
+
list_callback.on_eval_step_begin(run_context)
|
|
1335
1358
|
self._check_network_mode(self._eval_network, False)
|
|
1336
1359
|
outputs = self._eval_network(*next_element)
|
|
1337
1360
|
cb_params.net_outputs = outputs
|
|
@@ -1342,6 +1365,8 @@ class Model:
|
|
|
1342
1365
|
if add_eval_loss:
|
|
1343
1366
|
eval_loss_fn = get_metric_fn("loss")
|
|
1344
1367
|
eval_loss_fn.update(outputs[self._eval_indexes[0]])
|
|
1368
|
+
if run_context.get_stop_requested():
|
|
1369
|
+
break
|
|
1345
1370
|
|
|
1346
1371
|
list_callback.on_eval_epoch_end(run_context)
|
|
1347
1372
|
valid_dataset.reset()
|
|
@@ -1354,7 +1379,7 @@ class Model:
|
|
|
1354
1379
|
list_callback.on_eval_end(run_context)
|
|
1355
1380
|
return metrics
|
|
1356
1381
|
|
|
1357
|
-
def eval(self, valid_dataset, callbacks=None, dataset_sink_mode=
|
|
1382
|
+
def eval(self, valid_dataset, callbacks=None, dataset_sink_mode=False):
|
|
1358
1383
|
"""
|
|
1359
1384
|
Evaluation API.
|
|
1360
1385
|
|
|
@@ -1374,22 +1399,22 @@ class Model:
|
|
|
1374
1399
|
which should be executed while evaluation.
|
|
1375
1400
|
Default: None.
|
|
1376
1401
|
dataset_sink_mode (bool): Determines whether to pass the data through dataset channel.
|
|
1377
|
-
Default:
|
|
1402
|
+
Default: False.
|
|
1378
1403
|
|
|
1379
1404
|
Returns:
|
|
1380
1405
|
Dict, the key is the metric name defined by users and the value is the metrics value for
|
|
1381
1406
|
the model in the test mode.
|
|
1382
1407
|
|
|
1383
1408
|
Examples:
|
|
1384
|
-
>>> import mindspore as ms
|
|
1385
1409
|
>>> from mindspore import nn
|
|
1410
|
+
>>> from mindspore.train import Model
|
|
1386
1411
|
>>>
|
|
1387
1412
|
>>> # For details about how to build the dataset, please refer to the tutorial
|
|
1388
1413
|
>>> # document on the official website.
|
|
1389
1414
|
>>> dataset = create_custom_dataset()
|
|
1390
1415
|
>>> net = Net()
|
|
1391
1416
|
>>> loss = nn.SoftmaxCrossEntropyWithLogits()
|
|
1392
|
-
>>> model =
|
|
1417
|
+
>>> model = Model(net, loss_fn=loss, optimizer=None, metrics={'acc'})
|
|
1393
1418
|
>>> acc = model.eval(dataset, dataset_sink_mode=False)
|
|
1394
1419
|
"""
|
|
1395
1420
|
dataset_sink_mode = Validator.check_bool(dataset_sink_mode)
|
|
@@ -1431,8 +1456,9 @@ class Model:
|
|
|
1431
1456
|
else:
|
|
1432
1457
|
eval_result = self._eval_process(valid_dataset, list_callback, cb_params)
|
|
1433
1458
|
|
|
1434
|
-
# When it's
|
|
1459
|
+
# When it's distributed training and using MindRT,
|
|
1435
1460
|
# the node id should be reset to start from 0.
|
|
1461
|
+
# This is to avoid the timeout when finding the actor route tables in 'train' and 'eval' case(or 'fit').
|
|
1436
1462
|
if _enable_distributed_mindrt():
|
|
1437
1463
|
_reset_op_id_with_offset()
|
|
1438
1464
|
|
|
@@ -1452,11 +1478,12 @@ class Model:
|
|
|
1452
1478
|
|
|
1453
1479
|
Examples:
|
|
1454
1480
|
>>> import numpy as np
|
|
1455
|
-
>>> import mindspore
|
|
1481
|
+
>>> import mindspore
|
|
1456
1482
|
>>> from mindspore import Tensor
|
|
1483
|
+
>>> from mindspore.train import Model
|
|
1457
1484
|
>>>
|
|
1458
|
-
>>> input_data = Tensor(np.random.randint(0, 255, [1, 1, 32, 32]),
|
|
1459
|
-
>>> model =
|
|
1485
|
+
>>> input_data = Tensor(np.random.randint(0, 255, [1, 1, 32, 32]), mindspore.float32)
|
|
1486
|
+
>>> model = Model(Net())
|
|
1460
1487
|
>>> result = model.predict(input_data)
|
|
1461
1488
|
"""
|
|
1462
1489
|
self._check_network_mode(self._predict_network, False)
|
|
@@ -1466,8 +1493,9 @@ class Model:
|
|
|
1466
1493
|
|
|
1467
1494
|
check_output_data(result)
|
|
1468
1495
|
|
|
1469
|
-
# When it's
|
|
1496
|
+
# When it's distributed training and using MindRT,
|
|
1470
1497
|
# the node id should be reset to start from 0.
|
|
1498
|
+
# This is to avoid the timeout when finding the actor route tables in 'train' and 'eval' case(or 'fit').
|
|
1471
1499
|
if _enable_distributed_mindrt():
|
|
1472
1500
|
_reset_op_id_with_offset()
|
|
1473
1501
|
|
|
@@ -1509,7 +1537,7 @@ class Model:
|
|
|
1509
1537
|
Only dataset sink mode is supported for now.
|
|
1510
1538
|
|
|
1511
1539
|
.. warning::
|
|
1512
|
-
This is an experimental
|
|
1540
|
+
This is an experimental API that is subject to change or deletion.
|
|
1513
1541
|
|
|
1514
1542
|
Note:
|
|
1515
1543
|
This is a pre-compile function. The arguments should be the same as model.train() function.
|
|
@@ -1538,6 +1566,7 @@ class Model:
|
|
|
1538
1566
|
>>> import numpy as np
|
|
1539
1567
|
>>> import mindspore as ms
|
|
1540
1568
|
>>> from mindspore import Tensor, nn
|
|
1569
|
+
>>> from mindspore.train import Model
|
|
1541
1570
|
>>> from mindspore.communication import init
|
|
1542
1571
|
>>>
|
|
1543
1572
|
>>> ms.set_context(mode=ms.GRAPH_MODE)
|
|
@@ -1551,7 +1580,7 @@ class Model:
|
|
|
1551
1580
|
>>> loss = nn.SoftmaxCrossEntropyWithLogits()
|
|
1552
1581
|
>>> loss_scale_manager = ms.FixedLossScaleManager()
|
|
1553
1582
|
>>> optim = nn.Momentum(params=net.trainable_params(), learning_rate=0.1, momentum=0.9)
|
|
1554
|
-
>>> model =
|
|
1583
|
+
>>> model = Model(net, loss_fn=loss, optimizer=optim, metrics=None,
|
|
1555
1584
|
... loss_scale_manager=loss_scale_manager)
|
|
1556
1585
|
>>> layout_dict = model.infer_train_layout(dataset)
|
|
1557
1586
|
"""
|
|
@@ -1596,13 +1625,14 @@ class Model:
|
|
|
1596
1625
|
>>> import numpy as np
|
|
1597
1626
|
>>> import mindspore as ms
|
|
1598
1627
|
>>> from mindspore import Tensor
|
|
1628
|
+
>>> from mindspore.train import Model
|
|
1599
1629
|
>>> from mindspore.communication import init
|
|
1600
1630
|
>>>
|
|
1601
1631
|
>>> ms.set_context(mode=ms.GRAPH_MODE)
|
|
1602
1632
|
>>> init()
|
|
1603
1633
|
>>> ms.set_auto_parallel_context(full_batch=True, parallel_mode=ms.ParallelMode.SEMI_AUTO_PARALLEL)
|
|
1604
1634
|
>>> input_data = Tensor(np.random.randint(0, 255, [1, 1, 32, 32]), ms.float32)
|
|
1605
|
-
>>> model =
|
|
1635
|
+
>>> model = Model(Net())
|
|
1606
1636
|
>>> predict_map = model.infer_predict_layout(input_data)
|
|
1607
1637
|
"""
|
|
1608
1638
|
if context.get_context("mode") != context.GRAPH_MODE:
|
|
@@ -1615,7 +1645,6 @@ class Model:
|
|
|
1615
1645
|
|
|
1616
1646
|
predict_net = self._predict_network
|
|
1617
1647
|
# Unlike the cases in build_train_network() and build_eval_network(), 'multi_subgraphs' is not set
|
|
1618
|
-
predict_net.set_auto_parallel()
|
|
1619
1648
|
predict_net = self._check_network_mode(predict_net, False)
|
|
1620
1649
|
predict_net.compile(*predict_data)
|
|
1621
1650
|
return predict_net.parameter_layout_dict
|