mindspore 2.6.0rc1__cp310-cp310-win_amd64.whl → 2.7.0rc1__cp310-cp310-win_amd64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of mindspore might be problematic. Click here for more details.
- mindspore/.commit_id +1 -1
- mindspore/Microsoft.VisualStudio.Telemetry.dll +0 -0
- mindspore/Newtonsoft.Json.dll +0 -0
- mindspore/__init__.py +1 -1
- mindspore/_c_dataengine.cp310-win_amd64.pyd +0 -0
- mindspore/_c_expression.cp310-win_amd64.pyd +0 -0
- mindspore/_c_mindrecord.cp310-win_amd64.pyd +0 -0
- mindspore/_checkparam.py +40 -9
- mindspore/{_deprecated → _extends/optimize}/__init__.py +9 -3
- mindspore/_extends/optimize/cell_utils.py +96 -0
- mindspore/_extends/parse/__init__.py +2 -2
- mindspore/_extends/parse/compile_config.py +44 -22
- mindspore/_extends/parse/deprecated/deprecated_tensor_method.py +1 -1
- mindspore/_extends/parse/parser.py +37 -62
- mindspore/_extends/parse/resources.py +39 -0
- mindspore/_extends/parse/standard_method.py +43 -13
- mindspore/_extends/parse/trope.py +8 -1
- mindspore/_extends/pijit/__init__.py +1 -2
- mindspore/amp.py +4 -4
- mindspore/atlprov.dll +0 -0
- mindspore/avcodec-59.dll +0 -0
- mindspore/avdevice-59.dll +0 -0
- mindspore/avfilter-8.dll +0 -0
- mindspore/avformat-59.dll +0 -0
- mindspore/avutil-57.dll +0 -0
- mindspore/boost/adasum.py +1 -1
- mindspore/boost/boost_cell_wrapper.py +4 -4
- mindspore/c1.dll +0 -0
- mindspore/c1xx.dll +0 -0
- mindspore/c2.dll +0 -0
- mindspore/common/__init__.py +27 -2
- mindspore/common/_grad_function.py +2 -1
- mindspore/common/_pijit_context.py +28 -7
- mindspore/common/_stub_tensor.py +1 -209
- mindspore/common/_tensor_cpp_method.py +1 -1
- mindspore/common/_tensor_docs.py +77 -16
- mindspore/common/api.py +238 -113
- mindspore/common/dtype.py +21 -11
- mindspore/common/dump.py +10 -15
- mindspore/common/generator.py +5 -3
- mindspore/common/hook_handle.py +11 -2
- mindspore/common/jit_config.py +1 -1
- mindspore/common/jit_trace.py +84 -105
- mindspore/common/parameter.py +26 -12
- mindspore/common/recompute.py +3 -3
- mindspore/common/sparse_tensor.py +0 -3
- mindspore/common/symbol.py +0 -1
- mindspore/common/tensor.py +81 -81
- mindspore/communication/_comm_helper.py +46 -4
- mindspore/communication/management.py +79 -7
- mindspore/context.py +58 -40
- mindspore/dataset/core/config.py +3 -3
- mindspore/dataset/engine/datasets.py +20 -7
- mindspore/dataset/engine/datasets_user_defined.py +33 -3
- mindspore/dataset/engine/iterators.py +2 -2
- mindspore/dataset/engine/obs/config_loader.py +2 -2
- mindspore/dataset/engine/obs/obs_mindrecord_dataset.py +8 -0
- mindspore/dataset/transforms/py_transforms.py +7 -3
- mindspore/dataset/transforms/transforms.py +7 -3
- mindspore/dataset/vision/validators.py +1 -0
- mindspore/device_context/ascend/device.py +1 -1
- mindspore/device_context/gpu/__init__.py +2 -2
- mindspore/device_context/gpu/device.py +1 -1
- mindspore/device_context/gpu/op_precision.py +4 -2
- mindspore/device_context/gpu/op_tuning.py +6 -3
- mindspore/device_manager.py +16 -9
- mindspore/dnnl.dll +0 -0
- mindspore/dpcmi.dll +0 -0
- mindspore/experimental/llm_boost/ascend_native/llama_boost_ascend_native.py +3 -7
- mindspore/experimental/llm_boost/atb/boost_base.py +2 -3
- mindspore/experimental/optim/adadelta.py +13 -20
- mindspore/experimental/optim/adagrad.py +15 -22
- mindspore/experimental/optim/adam.py +17 -24
- mindspore/experimental/optim/adamax.py +14 -22
- mindspore/experimental/optim/adamw.py +28 -34
- mindspore/experimental/optim/asgd.py +15 -25
- mindspore/experimental/optim/lr_scheduler.py +27 -45
- mindspore/experimental/optim/nadam.py +14 -24
- mindspore/experimental/optim/optimizer.py +13 -23
- mindspore/experimental/optim/radam.py +18 -24
- mindspore/experimental/optim/rmsprop.py +14 -25
- mindspore/experimental/optim/rprop.py +15 -26
- mindspore/experimental/optim/sgd.py +9 -19
- mindspore/hal/__init__.py +4 -4
- mindspore/hal/contiguous_tensors_handle.py +2 -2
- mindspore/hal/memory.py +27 -7
- mindspore/include/api/cell.h +37 -1
- mindspore/include/api/delegate.h +10 -0
- mindspore/include/api/model.h +3 -0
- mindspore/include/api/types.h +2 -2
- mindspore/include/c_api/model_c.h +0 -58
- mindspore/include/c_api/tensor_c.h +0 -26
- mindspore/include/dataset/vision_ascend.h +1 -1
- mindspore/jpeg62.dll +0 -0
- mindspore/mindrecord/tools/cifar10.py +60 -11
- mindspore/mindrecord/tools/cifar10_to_mr.py +5 -0
- mindspore/mindspore_backend_common.dll +0 -0
- mindspore/mindspore_backend_manager.dll +0 -0
- mindspore/mindspore_common.dll +0 -0
- mindspore/mindspore_core.dll +0 -0
- mindspore/mindspore_cpu_res_manager.dll +0 -0
- mindspore/mindspore_dump.dll +0 -0
- mindspore/mindspore_frontend.dll +0 -0
- mindspore/mindspore_glog.dll +0 -0
- mindspore/mindspore_memory_pool.dll +0 -0
- mindspore/mindspore_ms_backend.dll +0 -0
- mindspore/mindspore_ops.dll +0 -0
- mindspore/mindspore_ops_host.dll +0 -0
- mindspore/mindspore_ops_kernel_common.dll +0 -0
- mindspore/mindspore_profiler.dll +0 -0
- mindspore/mindspore_pyboost.dll +0 -0
- mindspore/mindspore_pynative.dll +0 -0
- mindspore/mindspore_res_manager.dll +0 -0
- mindspore/mindspore_runtime_pipeline.dll +0 -0
- mindspore/mint/__init__.py +6 -46
- mindspore/mint/distributed/__init__.py +1 -0
- mindspore/mint/distributed/distributed.py +212 -9
- mindspore/mint/nn/__init__.py +1 -1
- mindspore/mint/nn/functional.py +53 -6
- mindspore/mint/nn/layer/_functions.py +164 -294
- mindspore/mint/nn/layer/activation.py +8 -6
- mindspore/mint/nn/layer/conv.py +137 -101
- mindspore/mint/nn/layer/normalization.py +8 -22
- mindspore/mint/optim/adam.py +19 -18
- mindspore/mint/optim/adamw.py +14 -8
- mindspore/mint/optim/sgd.py +5 -5
- mindspore/msobj140.dll +0 -0
- mindspore/mspdb140.dll +0 -0
- mindspore/mspdbcore.dll +0 -0
- mindspore/mspdbst.dll +0 -0
- mindspore/mspft140.dll +0 -0
- mindspore/msvcdis140.dll +0 -0
- mindspore/msvcp140_1.dll +0 -0
- mindspore/msvcp140_2.dll +0 -0
- mindspore/msvcp140_atomic_wait.dll +0 -0
- mindspore/msvcp140_codecvt_ids.dll +0 -0
- mindspore/nn/cell.py +328 -502
- mindspore/nn/grad/cell_grad.py +11 -12
- mindspore/nn/layer/activation.py +32 -34
- mindspore/nn/layer/basic.py +67 -64
- mindspore/nn/layer/channel_shuffle.py +4 -4
- mindspore/nn/layer/combined.py +4 -2
- mindspore/nn/layer/conv.py +117 -110
- mindspore/nn/layer/dense.py +9 -7
- mindspore/nn/layer/embedding.py +50 -52
- mindspore/nn/layer/image.py +37 -39
- mindspore/nn/layer/math.py +111 -112
- mindspore/nn/layer/normalization.py +56 -44
- mindspore/nn/layer/pooling.py +58 -63
- mindspore/nn/layer/rnn_cells.py +33 -33
- mindspore/nn/layer/rnns.py +56 -56
- mindspore/nn/layer/thor_layer.py +74 -73
- mindspore/nn/layer/transformer.py +11 -1
- mindspore/nn/learning_rate_schedule.py +20 -20
- mindspore/nn/loss/loss.py +79 -81
- mindspore/nn/optim/adam.py +3 -3
- mindspore/nn/optim/adasum.py +2 -2
- mindspore/nn/optim/asgd.py +2 -0
- mindspore/nn/optim/optimizer.py +1 -1
- mindspore/nn/optim/thor.py +2 -2
- mindspore/nn/probability/distribution/exponential.py +2 -1
- mindspore/nn/probability/distribution/poisson.py +2 -1
- mindspore/nn/sparse/sparse.py +3 -3
- mindspore/nn/wrap/cell_wrapper.py +34 -37
- mindspore/nn/wrap/grad_reducer.py +37 -37
- mindspore/nn/wrap/loss_scale.py +72 -74
- mindspore/numpy/array_creations.py +5 -5
- mindspore/numpy/fft.py +1 -1
- mindspore/numpy/math_ops.py +5 -5
- mindspore/opencv_core452.dll +0 -0
- mindspore/opencv_imgcodecs452.dll +0 -0
- mindspore/opencv_imgproc452.dll +0 -0
- mindspore/ops/_grad_experimental/grad_comm_ops.py +51 -13
- mindspore/ops/_grad_experimental/grad_debug_ops.py +14 -0
- mindspore/ops/_vmap/vmap_array_ops.py +31 -13
- mindspore/ops/_vmap/vmap_nn_ops.py +8 -16
- mindspore/ops/auto_generate/cpp_create_prim_instance_helper.py +42 -11
- mindspore/ops/auto_generate/gen_extend_func.py +23 -141
- mindspore/ops/auto_generate/gen_ops_def.py +727 -321
- mindspore/ops/auto_generate/gen_ops_prim.py +1721 -984
- mindspore/ops/auto_generate/pyboost_inner_prim.py +31 -1
- mindspore/ops/composite/__init__.py +10 -0
- mindspore/ops/composite/base.py +8 -4
- mindspore/ops/composite/multitype_ops/__init__.py +12 -1
- mindspore/ops/composite/multitype_ops/_compile_utils.py +133 -109
- mindspore/ops/composite/multitype_ops/add_impl.py +70 -2
- mindspore/ops/composite/multitype_ops/div_impl.py +49 -0
- mindspore/ops/composite/multitype_ops/floordiv_impl.py +29 -0
- mindspore/ops/composite/multitype_ops/getitem_impl.py +11 -0
- mindspore/ops/composite/multitype_ops/mod_impl.py +5 -3
- mindspore/ops/composite/multitype_ops/mul_impl.py +49 -0
- mindspore/ops/composite/multitype_ops/setitem_impl.py +57 -0
- mindspore/ops/composite/multitype_ops/sub_impl.py +34 -0
- mindspore/ops/composite/multitype_ops/zeros_like_impl.py +14 -0
- mindspore/ops/function/__init__.py +3 -1
- mindspore/ops/function/_add_attr_func.py +11 -6
- mindspore/ops/function/array_func.py +9 -96
- mindspore/ops/function/debug_func.py +4 -3
- mindspore/ops/function/grad/grad_func.py +1 -1
- mindspore/ops/function/math_func.py +33 -540
- mindspore/ops/function/nn_func.py +28 -74
- mindspore/ops/function/other_func.py +4 -1
- mindspore/ops/function/random_func.py +44 -5
- mindspore/ops/function/vmap_func.py +2 -1
- mindspore/ops/functional.py +2 -3
- mindspore/ops/functional_overload.py +571 -6
- mindspore/ops/op_info_register.py +21 -0
- mindspore/ops/operations/__init__.py +16 -11
- mindspore/ops/operations/_custom_ops_utils.py +689 -34
- mindspore/ops/operations/_inner_ops.py +3 -6
- mindspore/ops/operations/_sequence_ops.py +1 -1
- mindspore/ops/operations/array_ops.py +2 -2
- mindspore/ops/operations/comm_ops.py +185 -26
- mindspore/ops/operations/custom_ops.py +294 -174
- mindspore/ops/operations/debug_ops.py +59 -4
- mindspore/ops/operations/image_ops.py +13 -13
- mindspore/ops/operations/manually_defined/ops_def.py +15 -16
- mindspore/ops/operations/math_ops.py +3 -4
- mindspore/ops/operations/nn_ops.py +7 -39
- mindspore/ops/primitive.py +6 -10
- mindspore/ops/tensor_method.py +47 -8
- mindspore/ops_generate/api/cpp_create_prim_instance_helper_generator.py +1 -1
- mindspore/ops_generate/api/functional_map_cpp_generator.py +10 -9
- mindspore/ops_generate/api/functions_cc_generator.py +58 -10
- mindspore/ops_generate/api/tensor_func_reg_cpp_generator.py +1 -1
- mindspore/ops_generate/common/base_generator.py +14 -0
- mindspore/ops_generate/common/gen_constants.py +8 -3
- mindspore/ops_generate/common/gen_utils.py +0 -19
- mindspore/ops_generate/common/op_proto.py +11 -4
- mindspore/ops_generate/common/template.py +88 -11
- mindspore/ops_generate/gen_ops.py +1 -1
- mindspore/ops_generate/op_def/lite_ops_cpp_generator.py +4 -4
- mindspore/ops_generate/op_def/ops_def_cc_generator.py +0 -3
- mindspore/ops_generate/op_def/ops_name_h_generator.py +0 -3
- mindspore/ops_generate/op_def/ops_primitive_h_generator.py +0 -4
- mindspore/ops_generate/op_def_py/op_prim_py_generator.py +5 -2
- mindspore/ops_generate/pyboost/auto_grad_impl_cc_generator.py +49 -8
- mindspore/ops_generate/pyboost/auto_grad_reg_cc_generator.py +2 -2
- mindspore/ops_generate/pyboost/gen_pyboost_func.py +31 -0
- mindspore/ops_generate/pyboost/op_template_parser.py +98 -72
- mindspore/ops_generate/pyboost/pyboost_functions_cpp_generator.py +70 -273
- mindspore/ops_generate/pyboost/pyboost_functions_h_generator.py +14 -6
- mindspore/ops_generate/pyboost/pyboost_functions_impl_cpp_generator.py +316 -0
- mindspore/ops_generate/pyboost/pyboost_functions_py_generator.py +1 -1
- mindspore/ops_generate/pyboost/pyboost_grad_function_cpp_generator.py +5 -3
- mindspore/ops_generate/pyboost/pyboost_inner_prim_generator.py +1 -1
- mindspore/ops_generate/pyboost/pyboost_internal_functions_cpp_generator.py +76 -0
- mindspore/ops_generate/pyboost/pyboost_internal_functions_h_generator.py +76 -0
- mindspore/ops_generate/pyboost/pyboost_internal_kernel_info_adapter_generator.py +125 -0
- mindspore/ops_generate/pyboost/pyboost_native_grad_functions_generator.py +4 -3
- mindspore/ops_generate/pyboost/pyboost_op_cpp_code_generator.py +348 -61
- mindspore/ops_generate/pyboost/pyboost_overload_functions_cpp_generator.py +1 -1
- mindspore/ops_generate/pyboost/pyboost_utils.py +118 -9
- mindspore/ops_generate/tensor_py_cc_generator.py +1 -24
- mindspore/parallel/_auto_parallel_context.py +11 -8
- mindspore/parallel/_cell_wrapper.py +113 -45
- mindspore/parallel/_parallel_serialization.py +1 -1
- mindspore/parallel/_ps_context.py +4 -6
- mindspore/parallel/_tensor.py +167 -12
- mindspore/parallel/_transformer/moe.py +1 -1
- mindspore/parallel/_transformer/transformer.py +13 -8
- mindspore/parallel/auto_parallel.py +14 -7
- mindspore/parallel/checkpoint_convert.py +3 -3
- mindspore/parallel/checkpoint_transform.py +11 -7
- mindspore/parallel/cluster/process_entity/_api.py +84 -48
- mindspore/parallel/cluster/process_entity/_utils.py +95 -7
- mindspore/parallel/cluster/run.py +43 -4
- mindspore/parallel/function/__init__.py +8 -1
- mindspore/parallel/function/reshard_func.py +6 -7
- mindspore/parallel/nn/__init__.py +15 -2
- mindspore/parallel/nn/parallel_cell_wrapper.py +9 -10
- mindspore/parallel/nn/parallel_grad_reducer.py +7 -6
- mindspore/parallel/shard.py +3 -4
- mindspore/parallel/transform_safetensors.py +463 -174
- mindspore/pgodb140.dll +0 -0
- mindspore/pgort140.dll +0 -0
- mindspore/profiler/__init__.py +2 -1
- mindspore/profiler/analysis/parser/timeline_assembly_factory/ascend_timeline_assembler.py +7 -7
- mindspore/profiler/analysis/parser/timeline_assembly_factory/base_timeline_assembler.py +3 -0
- mindspore/profiler/analysis/parser/timeline_assembly_factory/trace_view_container.py +12 -6
- mindspore/profiler/analysis/parser/timeline_creator/cpu_op_timeline_creator.py +3 -3
- mindspore/profiler/analysis/parser/timeline_creator/fwk_timeline_creator.py +3 -3
- mindspore/profiler/analysis/parser/timeline_creator/msprof_timeline_creator.py +4 -4
- mindspore/profiler/analysis/parser/timeline_creator/scope_layer_timeline_creator.py +3 -3
- mindspore/profiler/analysis/parser/timeline_event/fwk_event.py +4 -1
- mindspore/profiler/analysis/parser/timeline_event/timeline_event_pool.py +2 -1
- mindspore/profiler/analysis/task_manager.py +1 -1
- mindspore/profiler/analysis/viewer/ascend_communication_viewer.py +5 -1
- mindspore/profiler/analysis/viewer/ascend_integrate_viewer.py +2 -1
- mindspore/profiler/analysis/viewer/ascend_op_memory_viewer.py +42 -22
- mindspore/profiler/analysis/viewer/ascend_step_trace_time_viewer.py +3 -2
- mindspore/profiler/analysis/viewer/ms_minddata_viewer.py +9 -5
- mindspore/profiler/analysis/viewer/ms_operator_details_viewer.py +132 -0
- mindspore/profiler/common/constant.py +16 -0
- mindspore/profiler/common/profiler_context.py +25 -27
- mindspore/profiler/common/profiler_info.py +0 -16
- mindspore/profiler/common/profiler_op_analyse.py +235 -0
- mindspore/profiler/common/profiler_output_path.py +23 -8
- mindspore/profiler/common/profiler_parameters.py +128 -35
- mindspore/profiler/dynamic_profile/__init__.py +0 -0
- mindspore/profiler/dynamic_profile/dynamic_monitor_proxy.py +39 -0
- mindspore/profiler/dynamic_profile/dynamic_profiler_config_context.py +666 -0
- mindspore/profiler/dynamic_profile/dynamic_profiler_utils.py +62 -0
- mindspore/profiler/dynamic_profiler.py +305 -314
- mindspore/profiler/envprofiler.py +12 -7
- mindspore/profiler/experimental_config.py +96 -6
- mindspore/profiler/mstx.py +33 -12
- mindspore/profiler/platform/__init__.py +2 -3
- mindspore/profiler/platform/npu_profiler.py +29 -19
- mindspore/profiler/profiler.py +35 -19
- mindspore/profiler/profiler_action_controller.py +64 -76
- mindspore/profiler/schedule.py +10 -4
- mindspore/rewrite/common/config.py +1 -0
- mindspore/rewrite/common/namer.py +1 -0
- mindspore/rewrite/common/namespace.py +1 -0
- mindspore/rewrite/node/node.py +31 -11
- mindspore/rewrite/parsers/assign_parser.py +1 -1
- mindspore/rewrite/symbol_tree/symbol_tree.py +1 -1
- mindspore/run_check/_check_version.py +7 -10
- mindspore/runtime/__init__.py +5 -5
- mindspore/runtime/event.py +10 -4
- mindspore/runtime/executor.py +60 -45
- mindspore/runtime/memory.py +30 -32
- mindspore/runtime/thread_bind_core.py +298 -164
- mindspore/safeguard/rewrite_obfuscation.py +12 -13
- mindspore/swresample-4.dll +0 -0
- mindspore/swscale-6.dll +0 -0
- mindspore/tbbmalloc.dll +0 -0
- mindspore/tinyxml2.dll +0 -0
- mindspore/train/_utils.py +14 -4
- mindspore/train/amp.py +43 -20
- mindspore/train/callback/__init__.py +5 -5
- mindspore/train/callback/_checkpoint.py +3 -6
- mindspore/train/callback/_flops_collector.py +1 -1
- mindspore/train/callback/_landscape.py +0 -1
- mindspore/train/callback/_train_fault_tolerance.py +97 -16
- mindspore/train/data_sink.py +11 -2
- mindspore/train/dataset_helper.py +9 -0
- mindspore/train/model.py +135 -55
- mindspore/train/serialization.py +133 -111
- mindspore/train/summary/summary_record.py +13 -2
- mindspore/turbojpeg.dll +0 -0
- mindspore/utils/__init__.py +3 -2
- mindspore/utils/dryrun.py +0 -6
- mindspore/utils/runtime_execution_order_check.py +163 -77
- mindspore/utils/sdc_detect.py +68 -0
- mindspore/utils/utils.py +6 -9
- mindspore/vcmeta.dll +0 -0
- mindspore/vcruntime140.dll +0 -0
- mindspore/vcruntime140_1.dll +0 -0
- mindspore/version.py +1 -1
- {mindspore-2.6.0rc1.dist-info → mindspore-2.7.0rc1.dist-info}/METADATA +5 -4
- {mindspore-2.6.0rc1.dist-info → mindspore-2.7.0rc1.dist-info}/RECORD +356 -394
- mindspore/_deprecated/jit.py +0 -198
- mindspore/experimental/es/__init__.py +0 -22
- mindspore/experimental/es/embedding_service.py +0 -891
- mindspore/experimental/es/embedding_service_layer.py +0 -581
- mindspore/profiler/parser/__init__.py +0 -14
- mindspore/profiler/parser/aicpu_data_parser.py +0 -272
- mindspore/profiler/parser/ascend_analysis/__init__.py +0 -14
- mindspore/profiler/parser/ascend_analysis/constant.py +0 -71
- mindspore/profiler/parser/ascend_analysis/file_manager.py +0 -180
- mindspore/profiler/parser/ascend_analysis/function_event.py +0 -185
- mindspore/profiler/parser/ascend_analysis/fwk_cann_parser.py +0 -136
- mindspore/profiler/parser/ascend_analysis/fwk_file_parser.py +0 -131
- mindspore/profiler/parser/ascend_analysis/msprof_timeline_parser.py +0 -104
- mindspore/profiler/parser/ascend_analysis/path_manager.py +0 -313
- mindspore/profiler/parser/ascend_analysis/profiler_info_parser.py +0 -123
- mindspore/profiler/parser/ascend_analysis/tlv_decoder.py +0 -86
- mindspore/profiler/parser/ascend_analysis/trace_event_manager.py +0 -75
- mindspore/profiler/parser/ascend_cluster_generator.py +0 -116
- mindspore/profiler/parser/ascend_communicate_generator.py +0 -314
- mindspore/profiler/parser/ascend_flops_generator.py +0 -116
- mindspore/profiler/parser/ascend_fpbp_generator.py +0 -82
- mindspore/profiler/parser/ascend_hccl_generator.py +0 -271
- mindspore/profiler/parser/ascend_integrate_generator.py +0 -42
- mindspore/profiler/parser/ascend_memory_generator.py +0 -185
- mindspore/profiler/parser/ascend_msprof_exporter.py +0 -282
- mindspore/profiler/parser/ascend_msprof_generator.py +0 -187
- mindspore/profiler/parser/ascend_op_generator.py +0 -334
- mindspore/profiler/parser/ascend_steptrace_generator.py +0 -94
- mindspore/profiler/parser/ascend_timeline_generator.py +0 -545
- mindspore/profiler/parser/base_timeline_generator.py +0 -483
- mindspore/profiler/parser/container.py +0 -229
- mindspore/profiler/parser/cpu_gpu_timeline_generator.py +0 -697
- mindspore/profiler/parser/flops_parser.py +0 -531
- mindspore/profiler/parser/framework_enum.py +0 -111
- mindspore/profiler/parser/framework_parser.py +0 -464
- mindspore/profiler/parser/framework_struct.py +0 -61
- mindspore/profiler/parser/gpu_analysis/__init__.py +0 -14
- mindspore/profiler/parser/gpu_analysis/function_event.py +0 -44
- mindspore/profiler/parser/gpu_analysis/fwk_file_parser.py +0 -89
- mindspore/profiler/parser/gpu_analysis/profiler_info_parser.py +0 -72
- mindspore/profiler/parser/hccl_parser.py +0 -573
- mindspore/profiler/parser/hwts_log_parser.py +0 -122
- mindspore/profiler/parser/integrator.py +0 -526
- mindspore/profiler/parser/memory_usage_parser.py +0 -277
- mindspore/profiler/parser/minddata_analyzer.py +0 -800
- mindspore/profiler/parser/minddata_parser.py +0 -186
- mindspore/profiler/parser/minddata_pipeline_parser.py +0 -299
- mindspore/profiler/parser/op_intermediate_parser.py +0 -149
- mindspore/profiler/parser/optime_parser.py +0 -250
- mindspore/profiler/parser/profiler_info.py +0 -213
- mindspore/profiler/parser/step_trace_parser.py +0 -666
- {mindspore-2.6.0rc1.dist-info → mindspore-2.7.0rc1.dist-info}/WHEEL +0 -0
- {mindspore-2.6.0rc1.dist-info → mindspore-2.7.0rc1.dist-info}/entry_points.txt +0 -0
- {mindspore-2.6.0rc1.dist-info → mindspore-2.7.0rc1.dist-info}/top_level.txt +0 -0
mindspore/nn/layer/basic.py
CHANGED
|
@@ -26,8 +26,6 @@ from mindspore.common.seed import _get_graph_seed
|
|
|
26
26
|
from mindspore.common.tensor import Tensor
|
|
27
27
|
from mindspore.common.initializer import initializer, HeUniform, Uniform
|
|
28
28
|
from mindspore import ops
|
|
29
|
-
from mindspore.ops import operations as P
|
|
30
|
-
from mindspore.ops import functional as F
|
|
31
29
|
from mindspore.ops.operations import _inner_ops as inner
|
|
32
30
|
from mindspore.ops.primitive import constexpr, Primitive, _primexpr
|
|
33
31
|
from mindspore.common.parameter import Parameter
|
|
@@ -97,13 +95,13 @@ class L1Regularizer(Cell):
|
|
|
97
95
|
if math.isinf(scale) or math.isnan(scale):
|
|
98
96
|
raise ValueError(
|
|
99
97
|
f"For '{self.cls_name}', the 'scale' can not be INF or NAN, but got {scale}.")
|
|
100
|
-
self.abs =
|
|
101
|
-
self.reduce_sum =
|
|
98
|
+
self.abs = ops.Abs()
|
|
99
|
+
self.reduce_sum = ops.ReduceSum()
|
|
102
100
|
self.scale = Tensor(scale, dtype=mstype.float32)
|
|
103
101
|
|
|
104
102
|
def construct(self, weights):
|
|
105
103
|
const_utils.check_type_valid(
|
|
106
|
-
|
|
104
|
+
ops.dtype(weights), mstype.number_type, 'weights')
|
|
107
105
|
l1_regularization = self.scale * self.reduce_sum(self.abs(weights))
|
|
108
106
|
return l1_regularization
|
|
109
107
|
|
|
@@ -179,14 +177,14 @@ class Dropout(Cell):
|
|
|
179
177
|
raise ValueError(f"For '{self.cls_name}', the 'keep_prob' must be a number in range (0, 1], "
|
|
180
178
|
f"but got {keep_prob}.")
|
|
181
179
|
seed0, seed1 = _get_graph_seed(0, "dropout")
|
|
182
|
-
self.dropout =
|
|
180
|
+
self.dropout = ops.Dropout(keep_prob, seed0, seed1)
|
|
183
181
|
else:
|
|
184
182
|
Validator.check_value_type('p', p, [float, int], self.cls_name)
|
|
185
183
|
if p < 0 or p >= 1:
|
|
186
184
|
raise ValueError(f"For '{self.cls_name}', the 'p' must be a number in range [0, 1), "
|
|
187
185
|
f"but got {p}.")
|
|
188
186
|
seed0, seed1 = _get_graph_seed(0, "dropout")
|
|
189
|
-
self.dropout =
|
|
187
|
+
self.dropout = ops.Dropout(1.0 - p, seed0, seed1)
|
|
190
188
|
self.p = p
|
|
191
189
|
self.keep_prob = keep_prob
|
|
192
190
|
|
|
@@ -336,7 +334,7 @@ class Dropout1d(Cell):
|
|
|
336
334
|
if not self.training or self.prob == 0:
|
|
337
335
|
return x
|
|
338
336
|
|
|
339
|
-
out =
|
|
337
|
+
out = ops.dropout1d(x, self.prob)
|
|
340
338
|
return out
|
|
341
339
|
|
|
342
340
|
|
|
@@ -377,7 +375,7 @@ class Dropout2d(Cell):
|
|
|
377
375
|
raise ValueError(f"For '{self.cls_name}', the 'p' must be a number in range [0, 1], "
|
|
378
376
|
f"but got {p}.")
|
|
379
377
|
self.keep_prob = 1.0 - p
|
|
380
|
-
self.dropout2d =
|
|
378
|
+
self.dropout2d = ops.Dropout2D(self.keep_prob)
|
|
381
379
|
|
|
382
380
|
def construct(self, x):
|
|
383
381
|
if not self.training or self.keep_prob == 1:
|
|
@@ -428,7 +426,7 @@ class Dropout3d(Cell):
|
|
|
428
426
|
raise ValueError(f"For '{self.cls_name}', the 'p' must be a number in range [0, 1], "
|
|
429
427
|
f"but got {p}.")
|
|
430
428
|
self.keep_prob = 1.0 - p
|
|
431
|
-
self.dropout3d =
|
|
429
|
+
self.dropout3d = ops.Dropout3D(self.keep_prob)
|
|
432
430
|
|
|
433
431
|
def construct(self, x):
|
|
434
432
|
if not self.training or self.keep_prob == 1:
|
|
@@ -476,8 +474,8 @@ class Upsample(Cell):
|
|
|
476
474
|
self.recompute_scale_factor = recompute_scale_factor
|
|
477
475
|
|
|
478
476
|
def construct(self, x):
|
|
479
|
-
out =
|
|
480
|
-
|
|
477
|
+
out = ops.interpolate(x, self.size, self.scale_factor, self.mode,
|
|
478
|
+
self.align_corners, self.recompute_scale_factor)
|
|
481
479
|
return out
|
|
482
480
|
|
|
483
481
|
|
|
@@ -577,11 +575,11 @@ class Flatten(Cell):
|
|
|
577
575
|
raise ValueError("'start_dim' or 'end_dim' out of range.")
|
|
578
576
|
|
|
579
577
|
def construct(self, x):
|
|
580
|
-
x_rank =
|
|
578
|
+
x_rank = ops.rank(x)
|
|
581
579
|
ndim = x_rank if x_rank != 0 else 1
|
|
582
580
|
self.check_axis_valid(self.start_dim, ndim)
|
|
583
581
|
self.check_axis_valid(self.end_dim, ndim)
|
|
584
|
-
return
|
|
582
|
+
return ops.flatten(x, start_dim=self.start_dim, end_dim=self.end_dim)
|
|
585
583
|
|
|
586
584
|
|
|
587
585
|
class Identity(Cell):
|
|
@@ -641,10 +639,12 @@ class Dense(Cell):
|
|
|
641
639
|
in_channels (int): The number of channels in the input space.
|
|
642
640
|
out_channels (int): The number of channels in the output space.
|
|
643
641
|
weight_init (Union[Tensor, str, Initializer, numbers.Number], optional): The trainable weight_init parameter.
|
|
644
|
-
The dtype is same as `x`. The values of str refer to the function
|
|
642
|
+
The dtype is same as `x`. The values of str refer to the function
|
|
643
|
+
:func:`mindspore.common.initializer.initializer`. Default: ``None`` ,
|
|
645
644
|
weight will be initialized using HeUniform.
|
|
646
645
|
bias_init (Union[Tensor, str, Initializer, numbers.Number], optional): The trainable bias_init parameter.
|
|
647
|
-
The dtype is same as `x`. The values of str refer to the function
|
|
646
|
+
The dtype is same as `x`. The values of str refer to the function
|
|
647
|
+
:func:`mindspore.common.initializer.initializer`. Default: ``None`` ,
|
|
648
648
|
bias will be initialized using Uniform.
|
|
649
649
|
has_bias (bool, optional): Specifies whether the layer uses a bias vector :math:`\text{bias}`.
|
|
650
650
|
Default: ``True``.
|
|
@@ -704,8 +704,8 @@ class Dense(Cell):
|
|
|
704
704
|
out_channels, "out_channels", self.cls_name)
|
|
705
705
|
self.has_bias = Validator.check_bool(
|
|
706
706
|
has_bias, "has_bias", self.cls_name)
|
|
707
|
-
self.reshape =
|
|
708
|
-
self.shape_op =
|
|
707
|
+
self.reshape = ops.Reshape()
|
|
708
|
+
self.shape_op = ops.Shape()
|
|
709
709
|
|
|
710
710
|
if isinstance(weight_init, Tensor):
|
|
711
711
|
if weight_init.ndim != 2 or weight_init.shape[0] != out_channels or \
|
|
@@ -731,9 +731,9 @@ class Dense(Cell):
|
|
|
731
731
|
bias_init = Uniform(scale=bound)
|
|
732
732
|
self.bias = Parameter(initializer(
|
|
733
733
|
bias_init, [out_channels], dtype=dtype), name="bias")
|
|
734
|
-
self.bias_add =
|
|
734
|
+
self.bias_add = ops.BiasAdd()
|
|
735
735
|
|
|
736
|
-
self.matmul =
|
|
736
|
+
self.matmul = ops.MatMul(transpose_b=True)
|
|
737
737
|
self.activation = get_activation(activation) if isinstance(
|
|
738
738
|
activation, str) else activation
|
|
739
739
|
if activation is not None and not isinstance(self.activation, (Cell, Primitive)):
|
|
@@ -751,7 +751,7 @@ class Dense(Cell):
|
|
|
751
751
|
if self.activation_flag:
|
|
752
752
|
x = self.activation(x)
|
|
753
753
|
if len(x_shape) != 2:
|
|
754
|
-
out_shape = x_shape[:-1] + (
|
|
754
|
+
out_shape = x_shape[:-1] + (ops.shape(x)[-1],)
|
|
755
755
|
x = self.reshape(x, out_shape)
|
|
756
756
|
return x
|
|
757
757
|
|
|
@@ -789,11 +789,13 @@ class Linear(Cell):
|
|
|
789
789
|
bias (bool, optional): Specifies whether the layer uses a bias vector :math:`\text{bias}`. Default: ``True``.
|
|
790
790
|
weight_init (Union[Tensor, str, Initializer, numbers.Number], optional):
|
|
791
791
|
The trainable weight_init parameter. The dtype
|
|
792
|
-
is same as `x`. The values of str refer to the function
|
|
792
|
+
is same as `x`. The values of str refer to the function :func:`mindspore.common.initializer.initializer`.
|
|
793
|
+
Default: ``None`` ,
|
|
793
794
|
weight will be initialized using HeUniform.
|
|
794
795
|
bias_init (Union[Tensor, str, Initializer, numbers.Number], optional):
|
|
795
796
|
The trainable bias_init parameter. The dtype is
|
|
796
|
-
same as `x`. The values of str refer to the function
|
|
797
|
+
same as `x`. The values of str refer to the function :func:`mindspore.common.initializer.initializer`.
|
|
798
|
+
Default: ``None`` ,
|
|
797
799
|
bias will be initialized using Uniform.
|
|
798
800
|
dtype (:class:`mindspore.dtype`, optional): Data type of Parameter. Default: ``None`` .
|
|
799
801
|
If `dtype` is ``None`` , `dtype` is set to ``mstype.float32`` when initializing the method.
|
|
@@ -847,7 +849,7 @@ class Linear(Cell):
|
|
|
847
849
|
out_features, "out_features", self.cls_name)
|
|
848
850
|
self.has_bias = Validator.check_bool(
|
|
849
851
|
bias, "has_bias", self.cls_name)
|
|
850
|
-
self.dense =
|
|
852
|
+
self.dense = ops.Dense()
|
|
851
853
|
if dtype is None:
|
|
852
854
|
dtype = mstype.float32
|
|
853
855
|
if isinstance(weight_init, Tensor):
|
|
@@ -890,7 +892,7 @@ class Linear(Cell):
|
|
|
890
892
|
def _is_equal_one(x):
|
|
891
893
|
if x is None:
|
|
892
894
|
return False
|
|
893
|
-
return
|
|
895
|
+
return ops.equal(ops.reduce_mean(x), 1.0)
|
|
894
896
|
|
|
895
897
|
|
|
896
898
|
@constexpr
|
|
@@ -983,12 +985,12 @@ class Norm(Cell):
|
|
|
983
985
|
"keep_dims", keep_dims, [bool], self.cls_name)
|
|
984
986
|
self.axis = axis
|
|
985
987
|
self.keep_dims = keep_dims
|
|
986
|
-
self.reduce_sum =
|
|
987
|
-
self.sqrt =
|
|
988
|
-
self.squeeze =
|
|
988
|
+
self.reduce_sum = ops.ReduceSum(True)
|
|
989
|
+
self.sqrt = ops.Sqrt()
|
|
990
|
+
self.squeeze = ops.Squeeze(self.axis)
|
|
989
991
|
|
|
990
992
|
def construct(self, x):
|
|
991
|
-
x = self.sqrt(self.reduce_sum(
|
|
993
|
+
x = self.sqrt(self.reduce_sum(ops.square(x), self.axis))
|
|
992
994
|
|
|
993
995
|
if not self.keep_dims:
|
|
994
996
|
x = self.squeeze(x)
|
|
@@ -1008,14 +1010,15 @@ class OneHot(Cell):
|
|
|
1008
1010
|
def __init__(self, axis=-1, depth=1, on_value=1.0, off_value=0.0, dtype=mstype.float32):
|
|
1009
1011
|
"""Initialize OneHot."""
|
|
1010
1012
|
super(OneHot, self).__init__()
|
|
1011
|
-
self.onehot =
|
|
1013
|
+
self.onehot = ops.OneHot(axis)
|
|
1012
1014
|
self.depth = depth
|
|
1013
1015
|
self.dtype = dtype
|
|
1014
1016
|
self.on_value = on_value
|
|
1015
1017
|
self.off_value = off_value
|
|
1016
1018
|
|
|
1017
1019
|
def construct(self, indices):
|
|
1018
|
-
return self.onehot(indices, self.depth,
|
|
1020
|
+
return self.onehot(indices, self.depth, ops.cast(self.on_value, self.dtype),
|
|
1021
|
+
ops.cast(self.off_value, self.dtype))
|
|
1019
1022
|
|
|
1020
1023
|
|
|
1021
1024
|
class Pad(Cell):
|
|
@@ -1171,10 +1174,10 @@ class Pad(Cell):
|
|
|
1171
1174
|
raise ValueError(f"For '{self.cls_name}', only 'paddings' up to 4 dims is supported, but got "
|
|
1172
1175
|
f"{len(paddings)}.")
|
|
1173
1176
|
if mode == "CONSTANT":
|
|
1174
|
-
self.pad =
|
|
1177
|
+
self.pad = ops.Pad(self.paddings)
|
|
1175
1178
|
else:
|
|
1176
1179
|
self.paddings = Tensor(np.array(self.paddings), dtype=mstype.int64)
|
|
1177
|
-
self.pad =
|
|
1180
|
+
self.pad = ops.MirrorPad(mode=mode)
|
|
1178
1181
|
|
|
1179
1182
|
def construct(self, x):
|
|
1180
1183
|
if self.mode == "CONSTANT":
|
|
@@ -1340,7 +1343,7 @@ class Fold(Cell):
|
|
|
1340
1343
|
def tril(x_shape, x_dtype, k):
|
|
1341
1344
|
Validator.check_int(len(x_shape), 1, Validator.GE, "x rank", "tril")
|
|
1342
1345
|
Validator.check_is_int(k, "k value", "tril")
|
|
1343
|
-
value =
|
|
1346
|
+
value = ops.cast(ops.Tril(diagonal=k)(ops.ones(x_shape, x_dtype)), x_dtype)
|
|
1344
1347
|
return value
|
|
1345
1348
|
|
|
1346
1349
|
|
|
@@ -1354,9 +1357,9 @@ class Tril(Cell):
|
|
|
1354
1357
|
def __init__(self):
|
|
1355
1358
|
"""Initialize Tril."""
|
|
1356
1359
|
super(Tril, self).__init__()
|
|
1357
|
-
self.dtype =
|
|
1358
|
-
self.mul =
|
|
1359
|
-
self.cast =
|
|
1360
|
+
self.dtype = ops.DType()
|
|
1361
|
+
self.mul = ops.Mul()
|
|
1362
|
+
self.cast = ops.Cast()
|
|
1360
1363
|
|
|
1361
1364
|
def construct(self, x, k=0):
|
|
1362
1365
|
assist = tril(x.shape, self.dtype(x), k)
|
|
@@ -1369,7 +1372,7 @@ class Tril(Cell):
|
|
|
1369
1372
|
def triu(x_shape, x_dtype, k):
|
|
1370
1373
|
Validator.check_int(len(x_shape), 1, Validator.GE, "x rank", "triu")
|
|
1371
1374
|
Validator.check_is_int(k, "k value", "triu")
|
|
1372
|
-
value =
|
|
1375
|
+
value = ops.cast(ops.Triu(k)(ops.ones(x_shape, x_dtype)), x_dtype)
|
|
1373
1376
|
return value
|
|
1374
1377
|
|
|
1375
1378
|
|
|
@@ -1383,9 +1386,9 @@ class Triu(Cell):
|
|
|
1383
1386
|
def __init__(self):
|
|
1384
1387
|
"""Initialize Triu."""
|
|
1385
1388
|
super(Triu, self).__init__()
|
|
1386
|
-
self.dtype =
|
|
1387
|
-
self.mul =
|
|
1388
|
-
self.cast =
|
|
1389
|
+
self.dtype = ops.DType()
|
|
1390
|
+
self.mul = ops.Mul()
|
|
1391
|
+
self.cast = ops.Cast()
|
|
1389
1392
|
|
|
1390
1393
|
def construct(self, x, k=0):
|
|
1391
1394
|
assist = triu(x.shape, self.dtype(x), k)
|
|
@@ -1398,14 +1401,14 @@ class Triu(Cell):
|
|
|
1398
1401
|
def _get_matrix_diag_assist(x_shape, x_dtype):
|
|
1399
1402
|
"""Get matrix diag assist"""
|
|
1400
1403
|
Validator.check_int(len(x_shape), 1, Validator.GE, "x rank", "_get_matrix_diag_assist")
|
|
1401
|
-
base_eye =
|
|
1402
|
-
|
|
1404
|
+
base_eye = ops.reshape(
|
|
1405
|
+
ops.eye(x_shape[-1], x_shape[-1], x_dtype), (x_shape[-1] * x_shape[-1],))
|
|
1403
1406
|
if len(x_shape) == 1:
|
|
1404
|
-
assist =
|
|
1407
|
+
assist = ops.reshape(base_eye, x_shape + (x_shape[-1],))
|
|
1405
1408
|
else:
|
|
1406
|
-
assist =
|
|
1407
|
-
|
|
1408
|
-
value =
|
|
1409
|
+
assist = ops.reshape(
|
|
1410
|
+
ops.tile(base_eye, x_shape[:-1]), x_shape + (x_shape[-1],))
|
|
1411
|
+
value = ops.cast(assist, x_dtype)
|
|
1409
1412
|
return value
|
|
1410
1413
|
|
|
1411
1414
|
|
|
@@ -1413,13 +1416,13 @@ def _get_matrix_diag_assist(x_shape, x_dtype):
|
|
|
1413
1416
|
def _get_matrix_diag_part_assist(x_shape, x_dtype):
|
|
1414
1417
|
"""Get matrix diag part assist"""
|
|
1415
1418
|
Validator.check_int(len(x_shape), 2, Validator.GE, "x rank", "_get_matrix_diag_part_assist")
|
|
1416
|
-
base_eye =
|
|
1417
|
-
|
|
1419
|
+
base_eye = ops.reshape(
|
|
1420
|
+
ops.eye(x_shape[-2], x_shape[-1], x_dtype), (x_shape[-2] * x_shape[-1],))
|
|
1418
1421
|
if len(x_shape) <= 2:
|
|
1419
|
-
assist =
|
|
1422
|
+
assist = ops.reshape(base_eye, x_shape)
|
|
1420
1423
|
else:
|
|
1421
|
-
assist =
|
|
1422
|
-
value =
|
|
1424
|
+
assist = ops.reshape(ops.tile(base_eye, x_shape[:-2]), x_shape)
|
|
1425
|
+
value = ops.cast(assist, x_dtype)
|
|
1423
1426
|
return value
|
|
1424
1427
|
|
|
1425
1428
|
|
|
@@ -1434,10 +1437,10 @@ class MatrixDiag(Cell):
|
|
|
1434
1437
|
"""Initialize MatrixDiag."""
|
|
1435
1438
|
super(MatrixDiag, self).__init__()
|
|
1436
1439
|
self.matrix_diag = inner.MatrixDiag()
|
|
1437
|
-
self.dtype =
|
|
1440
|
+
self.dtype = ops.DType()
|
|
1438
1441
|
|
|
1439
1442
|
def construct(self, input_x):
|
|
1440
|
-
x_shape =
|
|
1443
|
+
x_shape = ops.shape(input_x)
|
|
1441
1444
|
x_dtype = self.dtype(input_x)
|
|
1442
1445
|
assist = _get_matrix_diag_assist(x_shape, x_dtype)
|
|
1443
1446
|
out_matrix_diag = self.matrix_diag(input_x, assist)
|
|
@@ -1455,10 +1458,10 @@ class MatrixDiagPart(Cell):
|
|
|
1455
1458
|
"""Initialize MatrixDiagPart."""
|
|
1456
1459
|
super(MatrixDiagPart, self).__init__()
|
|
1457
1460
|
self.matrix_diag_part = inner.MatrixDiagPart()
|
|
1458
|
-
self.dtype =
|
|
1461
|
+
self.dtype = ops.DType()
|
|
1459
1462
|
|
|
1460
1463
|
def construct(self, input_x):
|
|
1461
|
-
x_shape =
|
|
1464
|
+
x_shape = ops.shape(input_x)
|
|
1462
1465
|
x_dtype = self.dtype(input_x)
|
|
1463
1466
|
assist = _get_matrix_diag_part_assist(x_shape, x_dtype)
|
|
1464
1467
|
out_matrix_diag_part = self.matrix_diag_part(input_x, assist)
|
|
@@ -1516,10 +1519,10 @@ class MatrixSetDiag(Cell):
|
|
|
1516
1519
|
"""Initialize MatrixSetDiag."""
|
|
1517
1520
|
super(MatrixSetDiag, self).__init__()
|
|
1518
1521
|
self.matrix_set_diag = inner.MatrixSetDiag()
|
|
1519
|
-
self.dtype =
|
|
1522
|
+
self.dtype = ops.DType()
|
|
1520
1523
|
|
|
1521
1524
|
def construct(self, input_x, diagonal):
|
|
1522
|
-
x_shape =
|
|
1525
|
+
x_shape = ops.shape(input_x)
|
|
1523
1526
|
x_dtype = self.dtype(input_x)
|
|
1524
1527
|
assist = _get_matrix_diag_part_assist(x_shape, x_dtype)
|
|
1525
1528
|
out_matrix_set_diag = self.matrix_set_diag(input_x, diagonal, assist)
|
|
@@ -1545,7 +1548,7 @@ class Roll(Cell):
|
|
|
1545
1548
|
"shift", shift, [int, tuple, list], self.cls_name)
|
|
1546
1549
|
Validator.check_value_type(
|
|
1547
1550
|
"axis", axis, [int, tuple, list], self.cls_name)
|
|
1548
|
-
self.shape_op =
|
|
1551
|
+
self.shape_op = ops.Shape()
|
|
1549
1552
|
self.shift = shift
|
|
1550
1553
|
self.axis = axis
|
|
1551
1554
|
self.op_list = []
|
|
@@ -1562,7 +1565,7 @@ class Roll(Cell):
|
|
|
1562
1565
|
Validator.check_is_int(s_axis, "axis", "Roll")
|
|
1563
1566
|
for s_shift in self.shift:
|
|
1564
1567
|
Validator.check_is_int(s_shift, "shift", "Roll")
|
|
1565
|
-
self.roll =
|
|
1568
|
+
self.roll = ops.Roll(self.shift, self.axis)
|
|
1566
1569
|
self.gpu = True
|
|
1567
1570
|
if len(self.shift) != len(self.axis):
|
|
1568
1571
|
raise ValueError(f"For '{self.cls_name}', the shape of 'shift' and the shape of 'axis' must be "
|
|
@@ -1571,7 +1574,7 @@ class Roll(Cell):
|
|
|
1571
1574
|
else:
|
|
1572
1575
|
if not isinstance(self.axis, (list, tuple)):
|
|
1573
1576
|
self.op_list.append(
|
|
1574
|
-
(
|
|
1577
|
+
(ops.Roll(shifts=self.shift, dims=0), self.axis))
|
|
1575
1578
|
else:
|
|
1576
1579
|
if len(self.shift) != len(self.axis):
|
|
1577
1580
|
raise ValueError(f"For '{self.cls_name}', the shape of 'shift' and the shape of 'axis' must be "
|
|
@@ -1579,7 +1582,7 @@ class Roll(Cell):
|
|
|
1579
1582
|
f"and the length of 'axis' {len(self.axis)}.")
|
|
1580
1583
|
for idx, _ in enumerate(self.axis):
|
|
1581
1584
|
self.op_list.append(
|
|
1582
|
-
(
|
|
1585
|
+
(ops.Roll(shifts=self.shift[idx], dims=0), self.axis[idx]))
|
|
1583
1586
|
|
|
1584
1587
|
def construct(self, input_x):
|
|
1585
1588
|
dim = len(self.shape_op(input_x))
|
|
@@ -1642,8 +1645,8 @@ class Unflatten(Cell):
|
|
|
1642
1645
|
def __init__(self, axis, unflattened_size):
|
|
1643
1646
|
"""Initialize Unflatten."""
|
|
1644
1647
|
super(Unflatten, self).__init__()
|
|
1645
|
-
self.shape =
|
|
1646
|
-
self.reshape =
|
|
1648
|
+
self.shape = ops.Shape()
|
|
1649
|
+
self.reshape = ops.Reshape()
|
|
1647
1650
|
Validator.check_is_int(axis, 'axis', 'Unflatten')
|
|
1648
1651
|
Validator.check_value_type(
|
|
1649
1652
|
'unflattended_size', unflattened_size, (list, tuple), 'Unflatten')
|
|
@@ -13,7 +13,7 @@
|
|
|
13
13
|
# limitations under the License.
|
|
14
14
|
# ============================================================================
|
|
15
15
|
"""channel shuffle"""
|
|
16
|
-
from mindspore
|
|
16
|
+
from mindspore import ops
|
|
17
17
|
from mindspore.nn.cell import Cell
|
|
18
18
|
|
|
19
19
|
__all__ = ['ChannelShuffle']
|
|
@@ -78,9 +78,9 @@ class ChannelShuffle(Cell):
|
|
|
78
78
|
raise ValueError(f"For ChannelShuffle, the param `groups` must be larger than 0, but got {groups}.")
|
|
79
79
|
|
|
80
80
|
self.groups = groups
|
|
81
|
-
self.shape =
|
|
82
|
-
self.reshape =
|
|
83
|
-
self.transpose =
|
|
81
|
+
self.shape = ops.Shape()
|
|
82
|
+
self.reshape = ops.Reshape()
|
|
83
|
+
self.transpose = ops.Transpose()
|
|
84
84
|
|
|
85
85
|
def construct(self, x):
|
|
86
86
|
x_shape = self.shape(x)
|
mindspore/nn/layer/combined.py
CHANGED
|
@@ -164,9 +164,11 @@ class DenseBnAct(Cell):
|
|
|
164
164
|
in_channels (int): The number of channels in the input space.
|
|
165
165
|
out_channels (int): The number of channels in the output space.
|
|
166
166
|
weight_init (Union[Tensor, str, Initializer, numbers.Number]): The trainable weight_init parameter. The dtype
|
|
167
|
-
is same as `x`. The values of str refer to the function
|
|
167
|
+
is same as `x`. The values of str refer to the function :func:`mindspore.common.initializer.initializer`.
|
|
168
|
+
Default: ``'normal'`` .
|
|
168
169
|
bias_init (Union[Tensor, str, Initializer, numbers.Number]): The trainable bias_init parameter. The dtype is
|
|
169
|
-
same as `x`. The values of str refer to the function
|
|
170
|
+
same as `x`. The values of str refer to the function :func:`mindspore.common.initializer.initializer`.
|
|
171
|
+
Default: ``'zeros'`` .
|
|
170
172
|
has_bias (bool): Specifies whether the layer uses a bias vector. Default: ``True`` .
|
|
171
173
|
has_bn (bool): Specifies to use batchnorm or not. Default: ``False`` .
|
|
172
174
|
momentum (float): Momentum for moving average for batchnorm, must be [0, 1]. Default: ``0.9`` .
|