mindspore 2.6.0__cp311-cp311-win_amd64.whl → 2.7.0__cp311-cp311-win_amd64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of mindspore might be problematic. Click here for more details.
- mindspore/.commit_id +1 -1
- mindspore/Microsoft.VisualStudio.Telemetry.dll +0 -0
- mindspore/Newtonsoft.Json.dll +0 -0
- mindspore/__init__.py +2 -2
- mindspore/_c_dataengine.cp311-win_amd64.pyd +0 -0
- mindspore/_c_expression.cp311-win_amd64.pyd +0 -0
- mindspore/_c_mindrecord.cp311-win_amd64.pyd +0 -0
- mindspore/_checkparam.py +42 -11
- mindspore/_extends/builtin_operations.py +3 -3
- mindspore/{_deprecated → _extends/optimize}/__init__.py +9 -3
- mindspore/_extends/optimize/cell_utils.py +96 -0
- mindspore/_extends/parallel_compile/akg_compiler/custom.py +1109 -0
- mindspore/_extends/parallel_compile/akg_compiler/gen_custom_op_files.py +1 -1
- mindspore/_extends/parse/__init__.py +3 -3
- mindspore/_extends/parse/compile_config.py +44 -22
- mindspore/_extends/parse/deprecated/deprecated_tensor_method.py +1 -2
- mindspore/_extends/parse/parser.py +64 -83
- mindspore/_extends/parse/resources.py +39 -0
- mindspore/_extends/parse/standard_method.py +47 -14
- mindspore/_extends/parse/trope.py +8 -1
- mindspore/_extends/pijit/__init__.py +1 -2
- mindspore/_extends/pijit/pijit_func_white_list.py +2 -5
- mindspore/amp.py +4 -22
- mindspore/atlprov.dll +0 -0
- mindspore/avcodec-59.dll +0 -0
- mindspore/avdevice-59.dll +0 -0
- mindspore/avfilter-8.dll +0 -0
- mindspore/avformat-59.dll +0 -0
- mindspore/avutil-57.dll +0 -0
- mindspore/boost/adasum.py +1 -1
- mindspore/boost/boost_cell_wrapper.py +4 -4
- mindspore/c1.dll +0 -0
- mindspore/c1xx.dll +0 -0
- mindspore/c2.dll +0 -0
- mindspore/common/__init__.py +43 -12
- mindspore/common/_grad_function.py +2 -1
- mindspore/common/_pijit_context.py +28 -7
- mindspore/common/_stub_tensor.py +1 -209
- mindspore/common/_tensor_cpp_method.py +1 -1
- mindspore/common/_tensor_docs.py +177 -52
- mindspore/common/_utils.py +9 -1
- mindspore/common/api.py +338 -208
- mindspore/common/dtype.py +108 -57
- mindspore/common/dump.py +11 -16
- mindspore/common/dynamic_shape/__init__.py +0 -0
- mindspore/common/{auto_dynamic_shape.py → dynamic_shape/auto_dynamic_shape.py} +17 -23
- mindspore/common/dynamic_shape/enable_dynamic.py +197 -0
- mindspore/common/file_system.py +59 -9
- mindspore/common/generator.py +2 -3
- mindspore/common/hook_handle.py +33 -5
- mindspore/common/jit_config.py +1 -1
- mindspore/common/jit_trace.py +84 -105
- mindspore/common/np_dtype.py +3 -3
- mindspore/common/parameter.py +27 -29
- mindspore/common/recompute.py +5 -7
- mindspore/common/sparse_tensor.py +0 -3
- mindspore/common/symbol.py +0 -1
- mindspore/common/tensor.py +84 -133
- mindspore/communication/_comm_helper.py +46 -4
- mindspore/communication/management.py +79 -7
- mindspore/context.py +47 -38
- mindspore/dataset/__init__.py +1 -1
- mindspore/dataset/audio/transforms.py +1 -1
- mindspore/dataset/core/config.py +38 -4
- mindspore/dataset/engine/datasets.py +350 -322
- mindspore/dataset/engine/datasets_user_defined.py +69 -23
- mindspore/dataset/engine/iterators.py +2 -2
- mindspore/dataset/engine/obs/config_loader.py +2 -2
- mindspore/dataset/engine/obs/obs_mindrecord_dataset.py +8 -0
- mindspore/dataset/transforms/c_transforms.py +2 -2
- mindspore/dataset/transforms/py_transforms.py +7 -3
- mindspore/dataset/transforms/transforms.py +10 -6
- mindspore/dataset/vision/__init__.py +1 -1
- mindspore/dataset/vision/py_transforms.py +8 -8
- mindspore/dataset/vision/transforms.py +17 -5
- mindspore/dataset/vision/utils.py +632 -21
- mindspore/dataset/vision/validators.py +1 -0
- mindspore/device_context/ascend/device.py +1 -1
- mindspore/device_context/ascend/op_tuning.py +35 -1
- mindspore/device_context/gpu/__init__.py +2 -2
- mindspore/device_context/gpu/device.py +1 -1
- mindspore/device_context/gpu/op_precision.py +4 -2
- mindspore/device_context/gpu/op_tuning.py +6 -3
- mindspore/device_manager.py +16 -9
- mindspore/dnnl.dll +0 -0
- mindspore/dpcmi.dll +0 -0
- mindspore/experimental/llm_boost/ascend_native/llama_boost_ascend_native.py +5 -4
- mindspore/experimental/llm_boost/atb/boost_base.py +2 -3
- mindspore/experimental/optim/adadelta.py +13 -20
- mindspore/experimental/optim/adagrad.py +15 -22
- mindspore/experimental/optim/adam.py +17 -24
- mindspore/experimental/optim/adamax.py +14 -22
- mindspore/experimental/optim/adamw.py +28 -34
- mindspore/experimental/optim/asgd.py +15 -25
- mindspore/experimental/optim/lr_scheduler.py +27 -45
- mindspore/experimental/optim/nadam.py +14 -24
- mindspore/experimental/optim/optimizer.py +13 -23
- mindspore/experimental/optim/radam.py +18 -24
- mindspore/experimental/optim/rmsprop.py +14 -25
- mindspore/experimental/optim/rprop.py +15 -26
- mindspore/experimental/optim/sgd.py +9 -19
- mindspore/hal/__init__.py +4 -4
- mindspore/hal/contiguous_tensors_handle.py +2 -2
- mindspore/hal/memory.py +1 -0
- mindspore/include/api/cell.h +65 -5
- mindspore/include/api/cfg.h +24 -7
- mindspore/include/api/context.h +1 -0
- mindspore/include/api/delegate.h +10 -2
- mindspore/include/api/dual_abi_helper.h +100 -19
- mindspore/include/api/graph.h +14 -1
- mindspore/include/api/kernel.h +16 -3
- mindspore/include/api/kernel_api.h +9 -1
- mindspore/include/api/metrics/accuracy.h +9 -0
- mindspore/include/api/model.h +8 -1
- mindspore/include/api/model_group.h +4 -0
- mindspore/include/api/model_parallel_runner.h +2 -0
- mindspore/include/api/status.h +48 -10
- mindspore/include/api/types.h +8 -3
- mindspore/include/c_api/model_c.h +0 -58
- mindspore/include/c_api/tensor_c.h +0 -26
- mindspore/include/dataset/constants.h +9 -0
- mindspore/include/dataset/vision_ascend.h +1 -1
- mindspore/jpeg62.dll +0 -0
- mindspore/mindrecord/tools/cifar10.py +61 -11
- mindspore/mindrecord/tools/cifar10_to_mr.py +5 -0
- mindspore/mindspore_backend_common.dll +0 -0
- mindspore/mindspore_backend_manager.dll +0 -0
- mindspore/mindspore_common.dll +0 -0
- mindspore/mindspore_core.dll +0 -0
- mindspore/mindspore_cpu_res_manager.dll +0 -0
- mindspore/mindspore_dump.dll +0 -0
- mindspore/mindspore_frontend.dll +0 -0
- mindspore/mindspore_glog.dll +0 -0
- mindspore/mindspore_memory_pool.dll +0 -0
- mindspore/mindspore_ms_backend.dll +0 -0
- mindspore/mindspore_ops.dll +0 -0
- mindspore/mindspore_ops_host.dll +0 -0
- mindspore/mindspore_ops_kernel_common.dll +0 -0
- mindspore/mindspore_profiler.dll +0 -0
- mindspore/mindspore_pyboost.dll +0 -0
- mindspore/mindspore_pynative.dll +0 -0
- mindspore/mindspore_res_manager.dll +0 -0
- mindspore/mindspore_runtime_pipeline.dll +0 -0
- mindspore/mint/__init__.py +4 -44
- mindspore/mint/distributed/__init__.py +5 -0
- mindspore/mint/distributed/distributed.py +425 -19
- mindspore/mint/nn/__init__.py +1 -1
- mindspore/mint/nn/functional.py +53 -6
- mindspore/mint/nn/layer/_functions.py +163 -294
- mindspore/mint/nn/layer/activation.py +8 -6
- mindspore/mint/nn/layer/conv.py +125 -101
- mindspore/mint/nn/layer/normalization.py +11 -25
- mindspore/mint/optim/adam.py +19 -18
- mindspore/mint/optim/adamw.py +14 -8
- mindspore/mint/optim/sgd.py +5 -5
- mindspore/msobj140.dll +0 -0
- mindspore/mspdb140.dll +0 -0
- mindspore/mspdbcore.dll +0 -0
- mindspore/mspdbst.dll +0 -0
- mindspore/mspft140.dll +0 -0
- mindspore/msvcdis140.dll +0 -0
- mindspore/msvcp140_1.dll +0 -0
- mindspore/msvcp140_2.dll +0 -0
- mindspore/msvcp140_atomic_wait.dll +0 -0
- mindspore/msvcp140_codecvt_ids.dll +0 -0
- mindspore/nn/cell.py +488 -620
- mindspore/nn/grad/cell_grad.py +11 -12
- mindspore/nn/layer/activation.py +36 -36
- mindspore/nn/layer/basic.py +74 -77
- mindspore/nn/layer/channel_shuffle.py +4 -4
- mindspore/nn/layer/combined.py +4 -2
- mindspore/nn/layer/conv.py +86 -85
- mindspore/nn/layer/dense.py +9 -7
- mindspore/nn/layer/embedding.py +50 -52
- mindspore/nn/layer/image.py +38 -40
- mindspore/nn/layer/math.py +111 -112
- mindspore/nn/layer/normalization.py +56 -44
- mindspore/nn/layer/pooling.py +58 -63
- mindspore/nn/layer/rnn_cells.py +33 -33
- mindspore/nn/layer/rnns.py +56 -56
- mindspore/nn/layer/thor_layer.py +74 -73
- mindspore/nn/layer/transformer.py +11 -1
- mindspore/nn/learning_rate_schedule.py +20 -20
- mindspore/nn/loss/loss.py +79 -81
- mindspore/nn/optim/adam.py +2 -4
- mindspore/nn/optim/adasum.py +2 -2
- mindspore/nn/optim/lamb.py +1 -3
- mindspore/nn/optim/optimizer.py +1 -1
- mindspore/nn/optim/tft_wrapper.py +2 -3
- mindspore/nn/optim/thor.py +2 -2
- mindspore/nn/probability/distribution/_utils/utils.py +2 -2
- mindspore/nn/probability/distribution/exponential.py +2 -1
- mindspore/nn/probability/distribution/poisson.py +2 -1
- mindspore/nn/sparse/sparse.py +3 -3
- mindspore/nn/wrap/cell_wrapper.py +73 -42
- mindspore/nn/wrap/grad_reducer.py +37 -52
- mindspore/nn/wrap/loss_scale.py +72 -74
- mindspore/numpy/array_creations.py +7 -7
- mindspore/numpy/fft.py +1 -1
- mindspore/numpy/math_ops.py +1 -1
- mindspore/numpy/utils_const.py +1 -1
- mindspore/opencv_core452.dll +0 -0
- mindspore/opencv_imgcodecs452.dll +0 -0
- mindspore/opencv_imgproc452.dll +0 -0
- mindspore/ops/_grad_experimental/grad_comm_ops.py +51 -13
- mindspore/ops/_grad_experimental/grad_debug_ops.py +14 -0
- mindspore/ops/_grad_experimental/grad_inner_ops.py +0 -9
- mindspore/ops/_op_impl/cpu/__init__.py +1 -0
- mindspore/{experimental/es/__init__.py → ops/_op_impl/cpu/joinedstr_op.py} +12 -6
- mindspore/ops/_vmap/vmap_array_ops.py +6 -13
- mindspore/ops/_vmap/vmap_nn_ops.py +8 -16
- mindspore/ops/auto_generate/cpp_create_prim_instance_helper.py +29 -10
- mindspore/ops/auto_generate/gen_extend_func.py +5 -55
- mindspore/ops/auto_generate/gen_ops_def.py +753 -273
- mindspore/ops/auto_generate/gen_ops_prim.py +1687 -958
- mindspore/ops/auto_generate/pyboost_inner_prim.py +31 -1
- mindspore/ops/composite/__init__.py +10 -0
- mindspore/ops/composite/base.py +9 -5
- mindspore/ops/composite/multitype_ops/__init__.py +12 -1
- mindspore/ops/composite/multitype_ops/_compile_utils.py +132 -108
- mindspore/ops/composite/multitype_ops/_constexpr_utils.py +1 -1
- mindspore/ops/composite/multitype_ops/add_impl.py +70 -2
- mindspore/ops/composite/multitype_ops/div_impl.py +49 -0
- mindspore/ops/composite/multitype_ops/floordiv_impl.py +29 -0
- mindspore/ops/composite/multitype_ops/getitem_impl.py +11 -0
- mindspore/ops/composite/multitype_ops/mod_impl.py +5 -3
- mindspore/ops/composite/multitype_ops/mul_impl.py +49 -0
- mindspore/ops/composite/multitype_ops/setitem_impl.py +57 -0
- mindspore/ops/composite/multitype_ops/sub_impl.py +34 -0
- mindspore/ops/composite/multitype_ops/zeros_like_impl.py +14 -0
- mindspore/ops/function/__init__.py +4 -1
- mindspore/ops/function/_add_attr_func.py +11 -6
- mindspore/ops/function/array_func.py +17 -100
- mindspore/ops/function/debug_func.py +8 -5
- mindspore/ops/function/grad/grad_func.py +5 -13
- mindspore/ops/function/math_func.py +65 -399
- mindspore/ops/function/nn_func.py +44 -61
- mindspore/ops/function/other_func.py +4 -1
- mindspore/ops/function/random_func.py +31 -4
- mindspore/ops/functional.py +2 -3
- mindspore/ops/functional_overload.py +486 -18
- mindspore/ops/op_info_register.py +21 -0
- mindspore/ops/operations/__init__.py +5 -2
- mindspore/ops/operations/_custom_ops_utils.py +675 -8
- mindspore/ops/operations/_inner_ops.py +14 -18
- mindspore/ops/operations/_sequence_ops.py +1 -1
- mindspore/ops/operations/array_ops.py +4 -50
- mindspore/ops/operations/comm_ops.py +186 -41
- mindspore/ops/operations/custom_ops.py +244 -175
- mindspore/ops/operations/debug_ops.py +55 -4
- mindspore/ops/operations/image_ops.py +13 -13
- mindspore/ops/operations/manually_defined/ops_def.py +27 -28
- mindspore/ops/operations/math_ops.py +8 -9
- mindspore/ops/operations/nn_ops.py +6 -7
- mindspore/ops/primitive.py +9 -20
- mindspore/ops/tensor_method.py +52 -11
- mindspore/ops_generate/api/cpp_create_prim_instance_helper_generator.py +1 -1
- mindspore/ops_generate/api/functional_map_cpp_generator.py +10 -9
- mindspore/ops_generate/api/functions_cc_generator.py +58 -10
- mindspore/ops_generate/api/tensor_func_reg_cpp_generator.py +1 -1
- mindspore/ops_generate/common/base_generator.py +14 -0
- mindspore/ops_generate/common/gen_constants.py +7 -2
- mindspore/ops_generate/common/gen_utils.py +0 -19
- mindspore/ops_generate/common/op_proto.py +11 -4
- mindspore/ops_generate/common/template.py +88 -11
- mindspore/ops_generate/gen_ops.py +1 -1
- mindspore/ops_generate/op_def/lite_ops_cpp_generator.py +4 -4
- mindspore/ops_generate/op_def/ops_name_h_generator.py +0 -3
- mindspore/ops_generate/op_def/ops_primitive_h_generator.py +0 -4
- mindspore/ops_generate/op_def_py/op_prim_py_generator.py +5 -2
- mindspore/ops_generate/pyboost/auto_grad_impl_cc_generator.py +49 -8
- mindspore/ops_generate/pyboost/auto_grad_reg_cc_generator.py +2 -2
- mindspore/ops_generate/pyboost/gen_pyboost_func.py +31 -16
- mindspore/ops_generate/pyboost/op_template_parser.py +98 -72
- mindspore/ops_generate/pyboost/pyboost_functions_cpp_generator.py +70 -273
- mindspore/ops_generate/pyboost/pyboost_functions_h_generator.py +14 -6
- mindspore/ops_generate/pyboost/pyboost_functions_impl_cpp_generator.py +316 -0
- mindspore/ops_generate/pyboost/pyboost_functions_py_generator.py +1 -1
- mindspore/ops_generate/pyboost/pyboost_grad_function_cpp_generator.py +5 -3
- mindspore/ops_generate/pyboost/pyboost_inner_prim_generator.py +1 -1
- mindspore/ops_generate/pyboost/pyboost_internal_functions_cpp_generator.py +76 -0
- mindspore/ops_generate/pyboost/pyboost_internal_functions_h_generator.py +76 -0
- mindspore/ops_generate/pyboost/pyboost_internal_kernel_info_adapter_generator.py +125 -0
- mindspore/ops_generate/pyboost/pyboost_native_grad_functions_generator.py +4 -3
- mindspore/ops_generate/pyboost/pyboost_op_cpp_code_generator.py +348 -61
- mindspore/ops_generate/pyboost/pyboost_overload_functions_cpp_generator.py +1 -1
- mindspore/ops_generate/pyboost/pyboost_utils.py +118 -9
- mindspore/ops_generate/tensor_py_cc_generator.py +1 -24
- mindspore/parallel/_auto_parallel_context.py +9 -17
- mindspore/parallel/_cell_wrapper.py +106 -40
- mindspore/parallel/_parallel_serialization.py +4 -3
- mindspore/parallel/_ps_context.py +4 -6
- mindspore/parallel/_tensor.py +167 -12
- mindspore/parallel/_transformer/moe.py +1 -1
- mindspore/parallel/_transformer/transformer.py +17 -12
- mindspore/parallel/_utils.py +5 -11
- mindspore/parallel/auto_parallel.py +33 -12
- mindspore/parallel/checkpoint_convert.py +3 -3
- mindspore/parallel/checkpoint_transform.py +5 -1
- mindspore/parallel/cluster/process_entity/_api.py +88 -49
- mindspore/parallel/cluster/process_entity/_utils.py +95 -7
- mindspore/parallel/cluster/run.py +48 -7
- mindspore/parallel/function/__init__.py +8 -1
- mindspore/parallel/function/reshard_func.py +7 -6
- mindspore/parallel/nn/__init__.py +15 -2
- mindspore/parallel/nn/parallel_cell_wrapper.py +50 -14
- mindspore/parallel/nn/parallel_grad_reducer.py +7 -14
- mindspore/parallel/shard.py +9 -23
- mindspore/parallel/transform_safetensors.py +468 -174
- mindspore/pgodb140.dll +0 -0
- mindspore/pgort140.dll +0 -0
- mindspore/profiler/__init__.py +2 -1
- mindspore/profiler/analysis/parser/timeline_assembly_factory/ascend_timeline_assembler.py +7 -7
- mindspore/profiler/analysis/parser/timeline_assembly_factory/base_timeline_assembler.py +3 -0
- mindspore/profiler/analysis/parser/timeline_assembly_factory/trace_view_container.py +3 -0
- mindspore/profiler/analysis/parser/timeline_creator/cpu_op_timeline_creator.py +3 -3
- mindspore/profiler/analysis/parser/timeline_creator/fwk_timeline_creator.py +3 -3
- mindspore/profiler/analysis/parser/timeline_creator/msprof_timeline_creator.py +4 -4
- mindspore/profiler/analysis/parser/timeline_creator/scope_layer_timeline_creator.py +3 -3
- mindspore/profiler/analysis/parser/timeline_event/fwk_event.py +4 -1
- mindspore/profiler/analysis/parser/timeline_event/timeline_event_pool.py +2 -1
- mindspore/profiler/analysis/task_manager.py +1 -1
- mindspore/profiler/analysis/viewer/ascend_communication_viewer.py +5 -1
- mindspore/profiler/analysis/viewer/ascend_integrate_viewer.py +2 -1
- mindspore/profiler/analysis/viewer/ascend_kernel_details_viewer.py +10 -9
- mindspore/profiler/analysis/viewer/ascend_op_memory_viewer.py +43 -23
- mindspore/profiler/analysis/viewer/ascend_step_trace_time_viewer.py +3 -2
- mindspore/profiler/analysis/viewer/ms_minddata_viewer.py +9 -5
- mindspore/profiler/analysis/viewer/ms_operator_details_viewer.py +132 -0
- mindspore/profiler/common/constant.py +16 -0
- mindspore/profiler/common/msprof_cmd_tool.py +2 -2
- mindspore/profiler/common/path_manager.py +9 -0
- mindspore/profiler/common/profiler_context.py +50 -29
- mindspore/profiler/common/profiler_info.py +0 -16
- mindspore/profiler/common/profiler_meta_data.py +1 -0
- mindspore/profiler/common/profiler_op_analyse.py +239 -0
- mindspore/profiler/common/profiler_output_path.py +23 -8
- mindspore/profiler/common/profiler_parameters.py +128 -35
- mindspore/profiler/dynamic_profile/__init__.py +0 -0
- mindspore/profiler/dynamic_profile/dynamic_monitor_proxy.py +39 -0
- mindspore/profiler/dynamic_profile/dynamic_profiler_config_context.py +666 -0
- mindspore/profiler/dynamic_profile/dynamic_profiler_utils.py +62 -0
- mindspore/profiler/dynamic_profiler.py +374 -338
- mindspore/profiler/envprofiler.py +42 -12
- mindspore/profiler/experimental_config.py +112 -7
- mindspore/profiler/mstx.py +33 -12
- mindspore/profiler/platform/__init__.py +2 -3
- mindspore/profiler/platform/cpu_profiler.py +10 -4
- mindspore/profiler/platform/npu_profiler.py +30 -20
- mindspore/profiler/profiler.py +218 -154
- mindspore/profiler/profiler_action_controller.py +65 -77
- mindspore/profiler/profiler_interface.py +2 -2
- mindspore/profiler/schedule.py +10 -4
- mindspore/rewrite/common/config.py +1 -0
- mindspore/rewrite/common/namer.py +1 -0
- mindspore/rewrite/common/namespace.py +1 -0
- mindspore/rewrite/node/node.py +31 -11
- mindspore/rewrite/parsers/assign_parser.py +1 -1
- mindspore/rewrite/symbol_tree/symbol_tree.py +2 -2
- mindspore/run_check/_check_version.py +7 -10
- mindspore/runtime/__init__.py +8 -6
- mindspore/runtime/event.py +10 -4
- mindspore/runtime/executor.py +87 -45
- mindspore/runtime/memory.py +22 -30
- mindspore/runtime/thread_bind_core.py +299 -165
- mindspore/safeguard/rewrite_obfuscation.py +12 -13
- mindspore/swresample-4.dll +0 -0
- mindspore/swscale-6.dll +0 -0
- mindspore/tbbmalloc.dll +0 -0
- mindspore/tinyxml2.dll +0 -0
- mindspore/train/_utils.py +9 -5
- mindspore/train/amp.py +43 -23
- mindspore/train/callback/__init__.py +5 -5
- mindspore/train/callback/_callback.py +2 -1
- mindspore/train/callback/_checkpoint.py +4 -14
- mindspore/train/callback/_flops_collector.py +11 -7
- mindspore/train/callback/_landscape.py +0 -1
- mindspore/train/callback/_train_fault_tolerance.py +72 -18
- mindspore/train/data_sink.py +15 -6
- mindspore/train/dataset_helper.py +14 -5
- mindspore/train/model.py +49 -47
- mindspore/train/serialization.py +168 -126
- mindspore/train/summary/summary_record.py +13 -2
- mindspore/train/train_thor/model_thor.py +2 -2
- mindspore/turbojpeg.dll +0 -0
- mindspore/utils/__init__.py +3 -2
- mindspore/utils/dryrun.py +0 -6
- mindspore/utils/runtime_execution_order_check.py +162 -78
- mindspore/utils/sdc_detect.py +68 -0
- mindspore/utils/utils.py +14 -17
- mindspore/vcmeta.dll +0 -0
- mindspore/vcruntime140.dll +0 -0
- mindspore/vcruntime140_1.dll +0 -0
- mindspore/version.py +1 -1
- {mindspore-2.6.0.dist-info → mindspore-2.7.0.dist-info}/METADATA +5 -4
- {mindspore-2.6.0.dist-info → mindspore-2.7.0.dist-info}/RECORD +400 -439
- mindspore/_deprecated/jit.py +0 -198
- mindspore/_extends/remote/kernel_build_server_ascend.py +0 -75
- mindspore/communication/_hccl_management.py +0 -297
- mindspore/experimental/es/embedding_service.py +0 -891
- mindspore/experimental/es/embedding_service_layer.py +0 -581
- mindspore/profiler/common/validator/__init__.py +0 -14
- mindspore/profiler/common/validator/validate_path.py +0 -84
- mindspore/profiler/parser/__init__.py +0 -14
- mindspore/profiler/parser/aicpu_data_parser.py +0 -272
- mindspore/profiler/parser/ascend_analysis/__init__.py +0 -14
- mindspore/profiler/parser/ascend_analysis/constant.py +0 -71
- mindspore/profiler/parser/ascend_analysis/file_manager.py +0 -180
- mindspore/profiler/parser/ascend_analysis/function_event.py +0 -185
- mindspore/profiler/parser/ascend_analysis/fwk_cann_parser.py +0 -136
- mindspore/profiler/parser/ascend_analysis/fwk_file_parser.py +0 -131
- mindspore/profiler/parser/ascend_analysis/msprof_timeline_parser.py +0 -104
- mindspore/profiler/parser/ascend_analysis/path_manager.py +0 -313
- mindspore/profiler/parser/ascend_analysis/profiler_info_parser.py +0 -123
- mindspore/profiler/parser/ascend_analysis/tlv_decoder.py +0 -86
- mindspore/profiler/parser/ascend_analysis/trace_event_manager.py +0 -75
- mindspore/profiler/parser/ascend_cluster_generator.py +0 -116
- mindspore/profiler/parser/ascend_communicate_generator.py +0 -314
- mindspore/profiler/parser/ascend_flops_generator.py +0 -116
- mindspore/profiler/parser/ascend_fpbp_generator.py +0 -82
- mindspore/profiler/parser/ascend_hccl_generator.py +0 -271
- mindspore/profiler/parser/ascend_integrate_generator.py +0 -42
- mindspore/profiler/parser/ascend_memory_generator.py +0 -185
- mindspore/profiler/parser/ascend_msprof_exporter.py +0 -282
- mindspore/profiler/parser/ascend_msprof_generator.py +0 -187
- mindspore/profiler/parser/ascend_op_generator.py +0 -334
- mindspore/profiler/parser/ascend_steptrace_generator.py +0 -94
- mindspore/profiler/parser/ascend_timeline_generator.py +0 -545
- mindspore/profiler/parser/base_timeline_generator.py +0 -483
- mindspore/profiler/parser/container.py +0 -229
- mindspore/profiler/parser/cpu_gpu_timeline_generator.py +0 -697
- mindspore/profiler/parser/flops_parser.py +0 -531
- mindspore/profiler/parser/framework_enum.py +0 -111
- mindspore/profiler/parser/framework_parser.py +0 -464
- mindspore/profiler/parser/framework_struct.py +0 -61
- mindspore/profiler/parser/gpu_analysis/__init__.py +0 -14
- mindspore/profiler/parser/gpu_analysis/function_event.py +0 -44
- mindspore/profiler/parser/gpu_analysis/fwk_file_parser.py +0 -89
- mindspore/profiler/parser/gpu_analysis/profiler_info_parser.py +0 -72
- mindspore/profiler/parser/hccl_parser.py +0 -573
- mindspore/profiler/parser/hwts_log_parser.py +0 -122
- mindspore/profiler/parser/integrator.py +0 -526
- mindspore/profiler/parser/memory_usage_parser.py +0 -277
- mindspore/profiler/parser/minddata_analyzer.py +0 -800
- mindspore/profiler/parser/minddata_parser.py +0 -186
- mindspore/profiler/parser/minddata_pipeline_parser.py +0 -299
- mindspore/profiler/parser/op_intermediate_parser.py +0 -149
- mindspore/profiler/parser/optime_parser.py +0 -250
- mindspore/profiler/parser/profiler_info.py +0 -213
- mindspore/profiler/parser/step_trace_parser.py +0 -666
- mindspore/utils/hooks.py +0 -81
- /mindspore/common/{_auto_dynamic.py → dynamic_shape/_auto_dynamic.py} +0 -0
- {mindspore-2.6.0.dist-info → mindspore-2.7.0.dist-info}/WHEEL +0 -0
- {mindspore-2.6.0.dist-info → mindspore-2.7.0.dist-info}/entry_points.txt +0 -0
- {mindspore-2.6.0.dist-info → mindspore-2.7.0.dist-info}/top_level.txt +0 -0
mindspore/nn/layer/basic.py
CHANGED
|
@@ -26,8 +26,6 @@ from mindspore.common.seed import _get_graph_seed
|
|
|
26
26
|
from mindspore.common.tensor import Tensor
|
|
27
27
|
from mindspore.common.initializer import initializer, HeUniform, Uniform
|
|
28
28
|
from mindspore import ops
|
|
29
|
-
from mindspore.ops import operations as P
|
|
30
|
-
from mindspore.ops import functional as F
|
|
31
29
|
from mindspore.ops.operations import _inner_ops as inner
|
|
32
30
|
from mindspore.ops.primitive import constexpr, Primitive, _primexpr
|
|
33
31
|
from mindspore.common.parameter import Parameter
|
|
@@ -97,13 +95,13 @@ class L1Regularizer(Cell):
|
|
|
97
95
|
if math.isinf(scale) or math.isnan(scale):
|
|
98
96
|
raise ValueError(
|
|
99
97
|
f"For '{self.cls_name}', the 'scale' can not be INF or NAN, but got {scale}.")
|
|
100
|
-
self.abs =
|
|
101
|
-
self.reduce_sum =
|
|
98
|
+
self.abs = ops.Abs()
|
|
99
|
+
self.reduce_sum = ops.ReduceSum()
|
|
102
100
|
self.scale = Tensor(scale, dtype=mstype.float32)
|
|
103
101
|
|
|
104
102
|
def construct(self, weights):
|
|
105
103
|
const_utils.check_type_valid(
|
|
106
|
-
|
|
104
|
+
ops.dtype(weights), mstype.number_type, 'weights')
|
|
107
105
|
l1_regularization = self.scale * self.reduce_sum(self.abs(weights))
|
|
108
106
|
return l1_regularization
|
|
109
107
|
|
|
@@ -179,14 +177,14 @@ class Dropout(Cell):
|
|
|
179
177
|
raise ValueError(f"For '{self.cls_name}', the 'keep_prob' must be a number in range (0, 1], "
|
|
180
178
|
f"but got {keep_prob}.")
|
|
181
179
|
seed0, seed1 = _get_graph_seed(0, "dropout")
|
|
182
|
-
self.dropout =
|
|
180
|
+
self.dropout = ops.Dropout(keep_prob, seed0, seed1)
|
|
183
181
|
else:
|
|
184
182
|
Validator.check_value_type('p', p, [float, int], self.cls_name)
|
|
185
183
|
if p < 0 or p >= 1:
|
|
186
184
|
raise ValueError(f"For '{self.cls_name}', the 'p' must be a number in range [0, 1), "
|
|
187
185
|
f"but got {p}.")
|
|
188
186
|
seed0, seed1 = _get_graph_seed(0, "dropout")
|
|
189
|
-
self.dropout =
|
|
187
|
+
self.dropout = ops.Dropout(1.0 - p, seed0, seed1)
|
|
190
188
|
self.p = p
|
|
191
189
|
self.keep_prob = keep_prob
|
|
192
190
|
|
|
@@ -211,7 +209,7 @@ class DropoutExt(Cell):
|
|
|
211
209
|
Dropout is a means of regularization that reduces overfitting by preventing correlations between neuronal nodes.
|
|
212
210
|
The operator randomly sets some neurons output to 0 according to `p`, which means the probability of discarding
|
|
213
211
|
during training. And the return will be multiplied by :math:`\frac{1}{1-p}` during training.
|
|
214
|
-
During the reasoning, this layer returns the same Tensor as the `
|
|
212
|
+
During the reasoning, this layer returns the same Tensor as the `input`.
|
|
215
213
|
|
|
216
214
|
This technique is proposed in paper `Dropout: A Simple Way to Prevent Neural Networks from Overfitting
|
|
217
215
|
<http://www.cs.toronto.edu/~rsalakhu/papers/srivastava14a.pdf>`_ and proved to be effective to reduce
|
|
@@ -230,14 +228,13 @@ class DropoutExt(Cell):
|
|
|
230
228
|
If set to ``True`` , will do this operation in-place. Default: ``False`` .
|
|
231
229
|
|
|
232
230
|
Inputs:
|
|
233
|
-
- **
|
|
231
|
+
- **input** (Tensor) - The input of Dropout.
|
|
234
232
|
|
|
235
233
|
Outputs:
|
|
236
|
-
Tensor, output tensor with the same shape as the `
|
|
234
|
+
Tensor, output tensor with the same shape as the `input`.
|
|
237
235
|
|
|
238
236
|
Raises:
|
|
239
|
-
TypeError: If the dtype of `
|
|
240
|
-
ValueError: If length of shape of `x` is less than 1.
|
|
237
|
+
TypeError: If the dtype of `inplace` is not bool.
|
|
241
238
|
|
|
242
239
|
Supported Platforms:
|
|
243
240
|
``Ascend``
|
|
@@ -261,17 +258,12 @@ class DropoutExt(Cell):
|
|
|
261
258
|
self.inplace = inplace
|
|
262
259
|
self.generator_step = Tensor(12, mstype.int64)
|
|
263
260
|
|
|
264
|
-
def construct(self,
|
|
261
|
+
def construct(self, input):
|
|
265
262
|
if not self.training or self.p == 0:
|
|
266
|
-
return
|
|
263
|
+
return input
|
|
267
264
|
|
|
268
265
|
seed, offset = default_generator._step(self.generator_step) # pylint: disable=protected-access
|
|
269
|
-
|
|
270
|
-
|
|
271
|
-
if self.inplace:
|
|
272
|
-
x.copy_(out)
|
|
273
|
-
return x
|
|
274
|
-
return out
|
|
266
|
+
return ops.auto_generate.func_dropout_ext_op(input, self.p, self.training, self.inplace, seed, offset)
|
|
275
267
|
|
|
276
268
|
|
|
277
269
|
class Dropout1d(Cell):
|
|
@@ -336,7 +328,7 @@ class Dropout1d(Cell):
|
|
|
336
328
|
if not self.training or self.prob == 0:
|
|
337
329
|
return x
|
|
338
330
|
|
|
339
|
-
out =
|
|
331
|
+
out = ops.dropout1d(x, self.prob)
|
|
340
332
|
return out
|
|
341
333
|
|
|
342
334
|
|
|
@@ -377,7 +369,7 @@ class Dropout2d(Cell):
|
|
|
377
369
|
raise ValueError(f"For '{self.cls_name}', the 'p' must be a number in range [0, 1], "
|
|
378
370
|
f"but got {p}.")
|
|
379
371
|
self.keep_prob = 1.0 - p
|
|
380
|
-
self.dropout2d =
|
|
372
|
+
self.dropout2d = ops.Dropout2D(self.keep_prob)
|
|
381
373
|
|
|
382
374
|
def construct(self, x):
|
|
383
375
|
if not self.training or self.keep_prob == 1:
|
|
@@ -428,7 +420,7 @@ class Dropout3d(Cell):
|
|
|
428
420
|
raise ValueError(f"For '{self.cls_name}', the 'p' must be a number in range [0, 1], "
|
|
429
421
|
f"but got {p}.")
|
|
430
422
|
self.keep_prob = 1.0 - p
|
|
431
|
-
self.dropout3d =
|
|
423
|
+
self.dropout3d = ops.Dropout3D(self.keep_prob)
|
|
432
424
|
|
|
433
425
|
def construct(self, x):
|
|
434
426
|
if not self.training or self.keep_prob == 1:
|
|
@@ -476,8 +468,8 @@ class Upsample(Cell):
|
|
|
476
468
|
self.recompute_scale_factor = recompute_scale_factor
|
|
477
469
|
|
|
478
470
|
def construct(self, x):
|
|
479
|
-
out =
|
|
480
|
-
|
|
471
|
+
out = ops.interpolate(x, self.size, self.scale_factor, self.mode,
|
|
472
|
+
self.align_corners, self.recompute_scale_factor)
|
|
481
473
|
return out
|
|
482
474
|
|
|
483
475
|
|
|
@@ -577,11 +569,11 @@ class Flatten(Cell):
|
|
|
577
569
|
raise ValueError("'start_dim' or 'end_dim' out of range.")
|
|
578
570
|
|
|
579
571
|
def construct(self, x):
|
|
580
|
-
x_rank =
|
|
572
|
+
x_rank = ops.rank(x)
|
|
581
573
|
ndim = x_rank if x_rank != 0 else 1
|
|
582
574
|
self.check_axis_valid(self.start_dim, ndim)
|
|
583
575
|
self.check_axis_valid(self.end_dim, ndim)
|
|
584
|
-
return
|
|
576
|
+
return ops.flatten(x, start_dim=self.start_dim, end_dim=self.end_dim)
|
|
585
577
|
|
|
586
578
|
|
|
587
579
|
class Identity(Cell):
|
|
@@ -641,10 +633,12 @@ class Dense(Cell):
|
|
|
641
633
|
in_channels (int): The number of channels in the input space.
|
|
642
634
|
out_channels (int): The number of channels in the output space.
|
|
643
635
|
weight_init (Union[Tensor, str, Initializer, numbers.Number], optional): The trainable weight_init parameter.
|
|
644
|
-
The dtype is same as `x`. The values of str refer to the function
|
|
636
|
+
The dtype is same as `x`. The values of str refer to the function
|
|
637
|
+
:func:`mindspore.common.initializer.initializer`. Default: ``None`` ,
|
|
645
638
|
weight will be initialized using HeUniform.
|
|
646
639
|
bias_init (Union[Tensor, str, Initializer, numbers.Number], optional): The trainable bias_init parameter.
|
|
647
|
-
The dtype is same as `x`. The values of str refer to the function
|
|
640
|
+
The dtype is same as `x`. The values of str refer to the function
|
|
641
|
+
:func:`mindspore.common.initializer.initializer`. Default: ``None`` ,
|
|
648
642
|
bias will be initialized using Uniform.
|
|
649
643
|
has_bias (bool, optional): Specifies whether the layer uses a bias vector :math:`\text{bias}`.
|
|
650
644
|
Default: ``True``.
|
|
@@ -704,8 +698,8 @@ class Dense(Cell):
|
|
|
704
698
|
out_channels, "out_channels", self.cls_name)
|
|
705
699
|
self.has_bias = Validator.check_bool(
|
|
706
700
|
has_bias, "has_bias", self.cls_name)
|
|
707
|
-
self.reshape =
|
|
708
|
-
self.shape_op =
|
|
701
|
+
self.reshape = ops.Reshape()
|
|
702
|
+
self.shape_op = ops.Shape()
|
|
709
703
|
|
|
710
704
|
if isinstance(weight_init, Tensor):
|
|
711
705
|
if weight_init.ndim != 2 or weight_init.shape[0] != out_channels or \
|
|
@@ -731,9 +725,9 @@ class Dense(Cell):
|
|
|
731
725
|
bias_init = Uniform(scale=bound)
|
|
732
726
|
self.bias = Parameter(initializer(
|
|
733
727
|
bias_init, [out_channels], dtype=dtype), name="bias")
|
|
734
|
-
self.bias_add =
|
|
728
|
+
self.bias_add = ops.BiasAdd()
|
|
735
729
|
|
|
736
|
-
self.matmul =
|
|
730
|
+
self.matmul = ops.MatMul(transpose_b=True)
|
|
737
731
|
self.activation = get_activation(activation) if isinstance(
|
|
738
732
|
activation, str) else activation
|
|
739
733
|
if activation is not None and not isinstance(self.activation, (Cell, Primitive)):
|
|
@@ -751,7 +745,7 @@ class Dense(Cell):
|
|
|
751
745
|
if self.activation_flag:
|
|
752
746
|
x = self.activation(x)
|
|
753
747
|
if len(x_shape) != 2:
|
|
754
|
-
out_shape = x_shape[:-1] + (
|
|
748
|
+
out_shape = x_shape[:-1] + (ops.shape(x)[-1],)
|
|
755
749
|
x = self.reshape(x, out_shape)
|
|
756
750
|
return x
|
|
757
751
|
|
|
@@ -789,11 +783,13 @@ class Linear(Cell):
|
|
|
789
783
|
bias (bool, optional): Specifies whether the layer uses a bias vector :math:`\text{bias}`. Default: ``True``.
|
|
790
784
|
weight_init (Union[Tensor, str, Initializer, numbers.Number], optional):
|
|
791
785
|
The trainable weight_init parameter. The dtype
|
|
792
|
-
is same as `x`. The values of str refer to the function
|
|
786
|
+
is same as `x`. The values of str refer to the function :func:`mindspore.common.initializer.initializer`.
|
|
787
|
+
Default: ``None`` ,
|
|
793
788
|
weight will be initialized using HeUniform.
|
|
794
789
|
bias_init (Union[Tensor, str, Initializer, numbers.Number], optional):
|
|
795
790
|
The trainable bias_init parameter. The dtype is
|
|
796
|
-
same as `x`. The values of str refer to the function
|
|
791
|
+
same as `x`. The values of str refer to the function :func:`mindspore.common.initializer.initializer`.
|
|
792
|
+
Default: ``None`` ,
|
|
797
793
|
bias will be initialized using Uniform.
|
|
798
794
|
dtype (:class:`mindspore.dtype`, optional): Data type of Parameter. Default: ``None`` .
|
|
799
795
|
If `dtype` is ``None`` , `dtype` is set to ``mstype.float32`` when initializing the method.
|
|
@@ -847,7 +843,7 @@ class Linear(Cell):
|
|
|
847
843
|
out_features, "out_features", self.cls_name)
|
|
848
844
|
self.has_bias = Validator.check_bool(
|
|
849
845
|
bias, "has_bias", self.cls_name)
|
|
850
|
-
self.dense =
|
|
846
|
+
self.dense = ops.Dense()
|
|
851
847
|
if dtype is None:
|
|
852
848
|
dtype = mstype.float32
|
|
853
849
|
if isinstance(weight_init, Tensor):
|
|
@@ -890,7 +886,7 @@ class Linear(Cell):
|
|
|
890
886
|
def _is_equal_one(x):
|
|
891
887
|
if x is None:
|
|
892
888
|
return False
|
|
893
|
-
return
|
|
889
|
+
return ops.equal(ops.reduce_mean(x), 1.0)
|
|
894
890
|
|
|
895
891
|
|
|
896
892
|
@constexpr
|
|
@@ -983,12 +979,12 @@ class Norm(Cell):
|
|
|
983
979
|
"keep_dims", keep_dims, [bool], self.cls_name)
|
|
984
980
|
self.axis = axis
|
|
985
981
|
self.keep_dims = keep_dims
|
|
986
|
-
self.reduce_sum =
|
|
987
|
-
self.sqrt =
|
|
988
|
-
self.squeeze =
|
|
982
|
+
self.reduce_sum = ops.ReduceSum(True)
|
|
983
|
+
self.sqrt = ops.Sqrt()
|
|
984
|
+
self.squeeze = ops.Squeeze(self.axis)
|
|
989
985
|
|
|
990
986
|
def construct(self, x):
|
|
991
|
-
x = self.sqrt(self.reduce_sum(
|
|
987
|
+
x = self.sqrt(self.reduce_sum(ops.square(x), self.axis))
|
|
992
988
|
|
|
993
989
|
if not self.keep_dims:
|
|
994
990
|
x = self.squeeze(x)
|
|
@@ -1008,14 +1004,15 @@ class OneHot(Cell):
|
|
|
1008
1004
|
def __init__(self, axis=-1, depth=1, on_value=1.0, off_value=0.0, dtype=mstype.float32):
|
|
1009
1005
|
"""Initialize OneHot."""
|
|
1010
1006
|
super(OneHot, self).__init__()
|
|
1011
|
-
self.onehot =
|
|
1007
|
+
self.onehot = ops.OneHot(axis)
|
|
1012
1008
|
self.depth = depth
|
|
1013
1009
|
self.dtype = dtype
|
|
1014
1010
|
self.on_value = on_value
|
|
1015
1011
|
self.off_value = off_value
|
|
1016
1012
|
|
|
1017
1013
|
def construct(self, indices):
|
|
1018
|
-
return self.onehot(indices, self.depth,
|
|
1014
|
+
return self.onehot(indices, self.depth, ops.cast(self.on_value, self.dtype),
|
|
1015
|
+
ops.cast(self.off_value, self.dtype))
|
|
1019
1016
|
|
|
1020
1017
|
|
|
1021
1018
|
class Pad(Cell):
|
|
@@ -1171,10 +1168,10 @@ class Pad(Cell):
|
|
|
1171
1168
|
raise ValueError(f"For '{self.cls_name}', only 'paddings' up to 4 dims is supported, but got "
|
|
1172
1169
|
f"{len(paddings)}.")
|
|
1173
1170
|
if mode == "CONSTANT":
|
|
1174
|
-
self.pad =
|
|
1171
|
+
self.pad = ops.Pad(self.paddings)
|
|
1175
1172
|
else:
|
|
1176
1173
|
self.paddings = Tensor(np.array(self.paddings), dtype=mstype.int64)
|
|
1177
|
-
self.pad =
|
|
1174
|
+
self.pad = ops.MirrorPad(mode=mode)
|
|
1178
1175
|
|
|
1179
1176
|
def construct(self, x):
|
|
1180
1177
|
if self.mode == "CONSTANT":
|
|
@@ -1340,7 +1337,7 @@ class Fold(Cell):
|
|
|
1340
1337
|
def tril(x_shape, x_dtype, k):
|
|
1341
1338
|
Validator.check_int(len(x_shape), 1, Validator.GE, "x rank", "tril")
|
|
1342
1339
|
Validator.check_is_int(k, "k value", "tril")
|
|
1343
|
-
value =
|
|
1340
|
+
value = ops.cast(ops.Tril(diagonal=k)(ops.ones(x_shape, x_dtype)), x_dtype)
|
|
1344
1341
|
return value
|
|
1345
1342
|
|
|
1346
1343
|
|
|
@@ -1354,9 +1351,9 @@ class Tril(Cell):
|
|
|
1354
1351
|
def __init__(self):
|
|
1355
1352
|
"""Initialize Tril."""
|
|
1356
1353
|
super(Tril, self).__init__()
|
|
1357
|
-
self.dtype =
|
|
1358
|
-
self.mul =
|
|
1359
|
-
self.cast =
|
|
1354
|
+
self.dtype = ops.DType()
|
|
1355
|
+
self.mul = ops.Mul()
|
|
1356
|
+
self.cast = ops.Cast()
|
|
1360
1357
|
|
|
1361
1358
|
def construct(self, x, k=0):
|
|
1362
1359
|
assist = tril(x.shape, self.dtype(x), k)
|
|
@@ -1369,7 +1366,7 @@ class Tril(Cell):
|
|
|
1369
1366
|
def triu(x_shape, x_dtype, k):
|
|
1370
1367
|
Validator.check_int(len(x_shape), 1, Validator.GE, "x rank", "triu")
|
|
1371
1368
|
Validator.check_is_int(k, "k value", "triu")
|
|
1372
|
-
value =
|
|
1369
|
+
value = ops.cast(ops.Triu(k)(ops.ones(x_shape, x_dtype)), x_dtype)
|
|
1373
1370
|
return value
|
|
1374
1371
|
|
|
1375
1372
|
|
|
@@ -1383,9 +1380,9 @@ class Triu(Cell):
|
|
|
1383
1380
|
def __init__(self):
|
|
1384
1381
|
"""Initialize Triu."""
|
|
1385
1382
|
super(Triu, self).__init__()
|
|
1386
|
-
self.dtype =
|
|
1387
|
-
self.mul =
|
|
1388
|
-
self.cast =
|
|
1383
|
+
self.dtype = ops.DType()
|
|
1384
|
+
self.mul = ops.Mul()
|
|
1385
|
+
self.cast = ops.Cast()
|
|
1389
1386
|
|
|
1390
1387
|
def construct(self, x, k=0):
|
|
1391
1388
|
assist = triu(x.shape, self.dtype(x), k)
|
|
@@ -1398,14 +1395,14 @@ class Triu(Cell):
|
|
|
1398
1395
|
def _get_matrix_diag_assist(x_shape, x_dtype):
|
|
1399
1396
|
"""Get matrix diag assist"""
|
|
1400
1397
|
Validator.check_int(len(x_shape), 1, Validator.GE, "x rank", "_get_matrix_diag_assist")
|
|
1401
|
-
base_eye =
|
|
1402
|
-
|
|
1398
|
+
base_eye = ops.reshape(
|
|
1399
|
+
ops.eye(x_shape[-1], x_shape[-1], x_dtype), (x_shape[-1] * x_shape[-1],))
|
|
1403
1400
|
if len(x_shape) == 1:
|
|
1404
|
-
assist =
|
|
1401
|
+
assist = ops.reshape(base_eye, x_shape + (x_shape[-1],))
|
|
1405
1402
|
else:
|
|
1406
|
-
assist =
|
|
1407
|
-
|
|
1408
|
-
value =
|
|
1403
|
+
assist = ops.reshape(
|
|
1404
|
+
ops.tile(base_eye, x_shape[:-1]), x_shape + (x_shape[-1],))
|
|
1405
|
+
value = ops.cast(assist, x_dtype)
|
|
1409
1406
|
return value
|
|
1410
1407
|
|
|
1411
1408
|
|
|
@@ -1413,13 +1410,13 @@ def _get_matrix_diag_assist(x_shape, x_dtype):
|
|
|
1413
1410
|
def _get_matrix_diag_part_assist(x_shape, x_dtype):
|
|
1414
1411
|
"""Get matrix diag part assist"""
|
|
1415
1412
|
Validator.check_int(len(x_shape), 2, Validator.GE, "x rank", "_get_matrix_diag_part_assist")
|
|
1416
|
-
base_eye =
|
|
1417
|
-
|
|
1413
|
+
base_eye = ops.reshape(
|
|
1414
|
+
ops.eye(x_shape[-2], x_shape[-1], x_dtype), (x_shape[-2] * x_shape[-1],))
|
|
1418
1415
|
if len(x_shape) <= 2:
|
|
1419
|
-
assist =
|
|
1416
|
+
assist = ops.reshape(base_eye, x_shape)
|
|
1420
1417
|
else:
|
|
1421
|
-
assist =
|
|
1422
|
-
value =
|
|
1418
|
+
assist = ops.reshape(ops.tile(base_eye, x_shape[:-2]), x_shape)
|
|
1419
|
+
value = ops.cast(assist, x_dtype)
|
|
1423
1420
|
return value
|
|
1424
1421
|
|
|
1425
1422
|
|
|
@@ -1434,10 +1431,10 @@ class MatrixDiag(Cell):
|
|
|
1434
1431
|
"""Initialize MatrixDiag."""
|
|
1435
1432
|
super(MatrixDiag, self).__init__()
|
|
1436
1433
|
self.matrix_diag = inner.MatrixDiag()
|
|
1437
|
-
self.dtype =
|
|
1434
|
+
self.dtype = ops.DType()
|
|
1438
1435
|
|
|
1439
1436
|
def construct(self, input_x):
|
|
1440
|
-
x_shape =
|
|
1437
|
+
x_shape = ops.shape(input_x)
|
|
1441
1438
|
x_dtype = self.dtype(input_x)
|
|
1442
1439
|
assist = _get_matrix_diag_assist(x_shape, x_dtype)
|
|
1443
1440
|
out_matrix_diag = self.matrix_diag(input_x, assist)
|
|
@@ -1455,10 +1452,10 @@ class MatrixDiagPart(Cell):
|
|
|
1455
1452
|
"""Initialize MatrixDiagPart."""
|
|
1456
1453
|
super(MatrixDiagPart, self).__init__()
|
|
1457
1454
|
self.matrix_diag_part = inner.MatrixDiagPart()
|
|
1458
|
-
self.dtype =
|
|
1455
|
+
self.dtype = ops.DType()
|
|
1459
1456
|
|
|
1460
1457
|
def construct(self, input_x):
|
|
1461
|
-
x_shape =
|
|
1458
|
+
x_shape = ops.shape(input_x)
|
|
1462
1459
|
x_dtype = self.dtype(input_x)
|
|
1463
1460
|
assist = _get_matrix_diag_part_assist(x_shape, x_dtype)
|
|
1464
1461
|
out_matrix_diag_part = self.matrix_diag_part(input_x, assist)
|
|
@@ -1516,10 +1513,10 @@ class MatrixSetDiag(Cell):
|
|
|
1516
1513
|
"""Initialize MatrixSetDiag."""
|
|
1517
1514
|
super(MatrixSetDiag, self).__init__()
|
|
1518
1515
|
self.matrix_set_diag = inner.MatrixSetDiag()
|
|
1519
|
-
self.dtype =
|
|
1516
|
+
self.dtype = ops.DType()
|
|
1520
1517
|
|
|
1521
1518
|
def construct(self, input_x, diagonal):
|
|
1522
|
-
x_shape =
|
|
1519
|
+
x_shape = ops.shape(input_x)
|
|
1523
1520
|
x_dtype = self.dtype(input_x)
|
|
1524
1521
|
assist = _get_matrix_diag_part_assist(x_shape, x_dtype)
|
|
1525
1522
|
out_matrix_set_diag = self.matrix_set_diag(input_x, diagonal, assist)
|
|
@@ -1545,7 +1542,7 @@ class Roll(Cell):
|
|
|
1545
1542
|
"shift", shift, [int, tuple, list], self.cls_name)
|
|
1546
1543
|
Validator.check_value_type(
|
|
1547
1544
|
"axis", axis, [int, tuple, list], self.cls_name)
|
|
1548
|
-
self.shape_op =
|
|
1545
|
+
self.shape_op = ops.Shape()
|
|
1549
1546
|
self.shift = shift
|
|
1550
1547
|
self.axis = axis
|
|
1551
1548
|
self.op_list = []
|
|
@@ -1562,7 +1559,7 @@ class Roll(Cell):
|
|
|
1562
1559
|
Validator.check_is_int(s_axis, "axis", "Roll")
|
|
1563
1560
|
for s_shift in self.shift:
|
|
1564
1561
|
Validator.check_is_int(s_shift, "shift", "Roll")
|
|
1565
|
-
self.roll =
|
|
1562
|
+
self.roll = ops.Roll(self.shift, self.axis)
|
|
1566
1563
|
self.gpu = True
|
|
1567
1564
|
if len(self.shift) != len(self.axis):
|
|
1568
1565
|
raise ValueError(f"For '{self.cls_name}', the shape of 'shift' and the shape of 'axis' must be "
|
|
@@ -1571,7 +1568,7 @@ class Roll(Cell):
|
|
|
1571
1568
|
else:
|
|
1572
1569
|
if not isinstance(self.axis, (list, tuple)):
|
|
1573
1570
|
self.op_list.append(
|
|
1574
|
-
(
|
|
1571
|
+
(ops.Roll(shifts=self.shift, dims=0), self.axis))
|
|
1575
1572
|
else:
|
|
1576
1573
|
if len(self.shift) != len(self.axis):
|
|
1577
1574
|
raise ValueError(f"For '{self.cls_name}', the shape of 'shift' and the shape of 'axis' must be "
|
|
@@ -1579,7 +1576,7 @@ class Roll(Cell):
|
|
|
1579
1576
|
f"and the length of 'axis' {len(self.axis)}.")
|
|
1580
1577
|
for idx, _ in enumerate(self.axis):
|
|
1581
1578
|
self.op_list.append(
|
|
1582
|
-
(
|
|
1579
|
+
(ops.Roll(shifts=self.shift[idx], dims=0), self.axis[idx]))
|
|
1583
1580
|
|
|
1584
1581
|
def construct(self, input_x):
|
|
1585
1582
|
dim = len(self.shape_op(input_x))
|
|
@@ -1642,8 +1639,8 @@ class Unflatten(Cell):
|
|
|
1642
1639
|
def __init__(self, axis, unflattened_size):
|
|
1643
1640
|
"""Initialize Unflatten."""
|
|
1644
1641
|
super(Unflatten, self).__init__()
|
|
1645
|
-
self.shape =
|
|
1646
|
-
self.reshape =
|
|
1642
|
+
self.shape = ops.Shape()
|
|
1643
|
+
self.reshape = ops.Reshape()
|
|
1647
1644
|
Validator.check_is_int(axis, 'axis', 'Unflatten')
|
|
1648
1645
|
Validator.check_value_type(
|
|
1649
1646
|
'unflattended_size', unflattened_size, (list, tuple), 'Unflatten')
|
|
@@ -13,7 +13,7 @@
|
|
|
13
13
|
# limitations under the License.
|
|
14
14
|
# ============================================================================
|
|
15
15
|
"""channel shuffle"""
|
|
16
|
-
from mindspore
|
|
16
|
+
from mindspore import ops
|
|
17
17
|
from mindspore.nn.cell import Cell
|
|
18
18
|
|
|
19
19
|
__all__ = ['ChannelShuffle']
|
|
@@ -78,9 +78,9 @@ class ChannelShuffle(Cell):
|
|
|
78
78
|
raise ValueError(f"For ChannelShuffle, the param `groups` must be larger than 0, but got {groups}.")
|
|
79
79
|
|
|
80
80
|
self.groups = groups
|
|
81
|
-
self.shape =
|
|
82
|
-
self.reshape =
|
|
83
|
-
self.transpose =
|
|
81
|
+
self.shape = ops.Shape()
|
|
82
|
+
self.reshape = ops.Reshape()
|
|
83
|
+
self.transpose = ops.Transpose()
|
|
84
84
|
|
|
85
85
|
def construct(self, x):
|
|
86
86
|
x_shape = self.shape(x)
|
mindspore/nn/layer/combined.py
CHANGED
|
@@ -164,9 +164,11 @@ class DenseBnAct(Cell):
|
|
|
164
164
|
in_channels (int): The number of channels in the input space.
|
|
165
165
|
out_channels (int): The number of channels in the output space.
|
|
166
166
|
weight_init (Union[Tensor, str, Initializer, numbers.Number]): The trainable weight_init parameter. The dtype
|
|
167
|
-
is same as `x`. The values of str refer to the function
|
|
167
|
+
is same as `x`. The values of str refer to the function :func:`mindspore.common.initializer.initializer`.
|
|
168
|
+
Default: ``'normal'`` .
|
|
168
169
|
bias_init (Union[Tensor, str, Initializer, numbers.Number]): The trainable bias_init parameter. The dtype is
|
|
169
|
-
same as `x`. The values of str refer to the function
|
|
170
|
+
same as `x`. The values of str refer to the function :func:`mindspore.common.initializer.initializer`.
|
|
171
|
+
Default: ``'zeros'`` .
|
|
170
172
|
has_bias (bool): Specifies whether the layer uses a bias vector. Default: ``True`` .
|
|
171
173
|
has_bn (bool): Specifies to use batchnorm or not. Default: ``False`` .
|
|
172
174
|
momentum (float): Momentum for moving average for batchnorm, must be [0, 1]. Default: ``0.9`` .
|