mindspore 2.6.0__cp311-cp311-win_amd64.whl → 2.7.0__cp311-cp311-win_amd64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of mindspore might be problematic. Click here for more details.
- mindspore/.commit_id +1 -1
- mindspore/Microsoft.VisualStudio.Telemetry.dll +0 -0
- mindspore/Newtonsoft.Json.dll +0 -0
- mindspore/__init__.py +2 -2
- mindspore/_c_dataengine.cp311-win_amd64.pyd +0 -0
- mindspore/_c_expression.cp311-win_amd64.pyd +0 -0
- mindspore/_c_mindrecord.cp311-win_amd64.pyd +0 -0
- mindspore/_checkparam.py +42 -11
- mindspore/_extends/builtin_operations.py +3 -3
- mindspore/{_deprecated → _extends/optimize}/__init__.py +9 -3
- mindspore/_extends/optimize/cell_utils.py +96 -0
- mindspore/_extends/parallel_compile/akg_compiler/custom.py +1109 -0
- mindspore/_extends/parallel_compile/akg_compiler/gen_custom_op_files.py +1 -1
- mindspore/_extends/parse/__init__.py +3 -3
- mindspore/_extends/parse/compile_config.py +44 -22
- mindspore/_extends/parse/deprecated/deprecated_tensor_method.py +1 -2
- mindspore/_extends/parse/parser.py +64 -83
- mindspore/_extends/parse/resources.py +39 -0
- mindspore/_extends/parse/standard_method.py +47 -14
- mindspore/_extends/parse/trope.py +8 -1
- mindspore/_extends/pijit/__init__.py +1 -2
- mindspore/_extends/pijit/pijit_func_white_list.py +2 -5
- mindspore/amp.py +4 -22
- mindspore/atlprov.dll +0 -0
- mindspore/avcodec-59.dll +0 -0
- mindspore/avdevice-59.dll +0 -0
- mindspore/avfilter-8.dll +0 -0
- mindspore/avformat-59.dll +0 -0
- mindspore/avutil-57.dll +0 -0
- mindspore/boost/adasum.py +1 -1
- mindspore/boost/boost_cell_wrapper.py +4 -4
- mindspore/c1.dll +0 -0
- mindspore/c1xx.dll +0 -0
- mindspore/c2.dll +0 -0
- mindspore/common/__init__.py +43 -12
- mindspore/common/_grad_function.py +2 -1
- mindspore/common/_pijit_context.py +28 -7
- mindspore/common/_stub_tensor.py +1 -209
- mindspore/common/_tensor_cpp_method.py +1 -1
- mindspore/common/_tensor_docs.py +177 -52
- mindspore/common/_utils.py +9 -1
- mindspore/common/api.py +338 -208
- mindspore/common/dtype.py +108 -57
- mindspore/common/dump.py +11 -16
- mindspore/common/dynamic_shape/__init__.py +0 -0
- mindspore/common/{auto_dynamic_shape.py → dynamic_shape/auto_dynamic_shape.py} +17 -23
- mindspore/common/dynamic_shape/enable_dynamic.py +197 -0
- mindspore/common/file_system.py +59 -9
- mindspore/common/generator.py +2 -3
- mindspore/common/hook_handle.py +33 -5
- mindspore/common/jit_config.py +1 -1
- mindspore/common/jit_trace.py +84 -105
- mindspore/common/np_dtype.py +3 -3
- mindspore/common/parameter.py +27 -29
- mindspore/common/recompute.py +5 -7
- mindspore/common/sparse_tensor.py +0 -3
- mindspore/common/symbol.py +0 -1
- mindspore/common/tensor.py +84 -133
- mindspore/communication/_comm_helper.py +46 -4
- mindspore/communication/management.py +79 -7
- mindspore/context.py +47 -38
- mindspore/dataset/__init__.py +1 -1
- mindspore/dataset/audio/transforms.py +1 -1
- mindspore/dataset/core/config.py +38 -4
- mindspore/dataset/engine/datasets.py +350 -322
- mindspore/dataset/engine/datasets_user_defined.py +69 -23
- mindspore/dataset/engine/iterators.py +2 -2
- mindspore/dataset/engine/obs/config_loader.py +2 -2
- mindspore/dataset/engine/obs/obs_mindrecord_dataset.py +8 -0
- mindspore/dataset/transforms/c_transforms.py +2 -2
- mindspore/dataset/transforms/py_transforms.py +7 -3
- mindspore/dataset/transforms/transforms.py +10 -6
- mindspore/dataset/vision/__init__.py +1 -1
- mindspore/dataset/vision/py_transforms.py +8 -8
- mindspore/dataset/vision/transforms.py +17 -5
- mindspore/dataset/vision/utils.py +632 -21
- mindspore/dataset/vision/validators.py +1 -0
- mindspore/device_context/ascend/device.py +1 -1
- mindspore/device_context/ascend/op_tuning.py +35 -1
- mindspore/device_context/gpu/__init__.py +2 -2
- mindspore/device_context/gpu/device.py +1 -1
- mindspore/device_context/gpu/op_precision.py +4 -2
- mindspore/device_context/gpu/op_tuning.py +6 -3
- mindspore/device_manager.py +16 -9
- mindspore/dnnl.dll +0 -0
- mindspore/dpcmi.dll +0 -0
- mindspore/experimental/llm_boost/ascend_native/llama_boost_ascend_native.py +5 -4
- mindspore/experimental/llm_boost/atb/boost_base.py +2 -3
- mindspore/experimental/optim/adadelta.py +13 -20
- mindspore/experimental/optim/adagrad.py +15 -22
- mindspore/experimental/optim/adam.py +17 -24
- mindspore/experimental/optim/adamax.py +14 -22
- mindspore/experimental/optim/adamw.py +28 -34
- mindspore/experimental/optim/asgd.py +15 -25
- mindspore/experimental/optim/lr_scheduler.py +27 -45
- mindspore/experimental/optim/nadam.py +14 -24
- mindspore/experimental/optim/optimizer.py +13 -23
- mindspore/experimental/optim/radam.py +18 -24
- mindspore/experimental/optim/rmsprop.py +14 -25
- mindspore/experimental/optim/rprop.py +15 -26
- mindspore/experimental/optim/sgd.py +9 -19
- mindspore/hal/__init__.py +4 -4
- mindspore/hal/contiguous_tensors_handle.py +2 -2
- mindspore/hal/memory.py +1 -0
- mindspore/include/api/cell.h +65 -5
- mindspore/include/api/cfg.h +24 -7
- mindspore/include/api/context.h +1 -0
- mindspore/include/api/delegate.h +10 -2
- mindspore/include/api/dual_abi_helper.h +100 -19
- mindspore/include/api/graph.h +14 -1
- mindspore/include/api/kernel.h +16 -3
- mindspore/include/api/kernel_api.h +9 -1
- mindspore/include/api/metrics/accuracy.h +9 -0
- mindspore/include/api/model.h +8 -1
- mindspore/include/api/model_group.h +4 -0
- mindspore/include/api/model_parallel_runner.h +2 -0
- mindspore/include/api/status.h +48 -10
- mindspore/include/api/types.h +8 -3
- mindspore/include/c_api/model_c.h +0 -58
- mindspore/include/c_api/tensor_c.h +0 -26
- mindspore/include/dataset/constants.h +9 -0
- mindspore/include/dataset/vision_ascend.h +1 -1
- mindspore/jpeg62.dll +0 -0
- mindspore/mindrecord/tools/cifar10.py +61 -11
- mindspore/mindrecord/tools/cifar10_to_mr.py +5 -0
- mindspore/mindspore_backend_common.dll +0 -0
- mindspore/mindspore_backend_manager.dll +0 -0
- mindspore/mindspore_common.dll +0 -0
- mindspore/mindspore_core.dll +0 -0
- mindspore/mindspore_cpu_res_manager.dll +0 -0
- mindspore/mindspore_dump.dll +0 -0
- mindspore/mindspore_frontend.dll +0 -0
- mindspore/mindspore_glog.dll +0 -0
- mindspore/mindspore_memory_pool.dll +0 -0
- mindspore/mindspore_ms_backend.dll +0 -0
- mindspore/mindspore_ops.dll +0 -0
- mindspore/mindspore_ops_host.dll +0 -0
- mindspore/mindspore_ops_kernel_common.dll +0 -0
- mindspore/mindspore_profiler.dll +0 -0
- mindspore/mindspore_pyboost.dll +0 -0
- mindspore/mindspore_pynative.dll +0 -0
- mindspore/mindspore_res_manager.dll +0 -0
- mindspore/mindspore_runtime_pipeline.dll +0 -0
- mindspore/mint/__init__.py +4 -44
- mindspore/mint/distributed/__init__.py +5 -0
- mindspore/mint/distributed/distributed.py +425 -19
- mindspore/mint/nn/__init__.py +1 -1
- mindspore/mint/nn/functional.py +53 -6
- mindspore/mint/nn/layer/_functions.py +163 -294
- mindspore/mint/nn/layer/activation.py +8 -6
- mindspore/mint/nn/layer/conv.py +125 -101
- mindspore/mint/nn/layer/normalization.py +11 -25
- mindspore/mint/optim/adam.py +19 -18
- mindspore/mint/optim/adamw.py +14 -8
- mindspore/mint/optim/sgd.py +5 -5
- mindspore/msobj140.dll +0 -0
- mindspore/mspdb140.dll +0 -0
- mindspore/mspdbcore.dll +0 -0
- mindspore/mspdbst.dll +0 -0
- mindspore/mspft140.dll +0 -0
- mindspore/msvcdis140.dll +0 -0
- mindspore/msvcp140_1.dll +0 -0
- mindspore/msvcp140_2.dll +0 -0
- mindspore/msvcp140_atomic_wait.dll +0 -0
- mindspore/msvcp140_codecvt_ids.dll +0 -0
- mindspore/nn/cell.py +488 -620
- mindspore/nn/grad/cell_grad.py +11 -12
- mindspore/nn/layer/activation.py +36 -36
- mindspore/nn/layer/basic.py +74 -77
- mindspore/nn/layer/channel_shuffle.py +4 -4
- mindspore/nn/layer/combined.py +4 -2
- mindspore/nn/layer/conv.py +86 -85
- mindspore/nn/layer/dense.py +9 -7
- mindspore/nn/layer/embedding.py +50 -52
- mindspore/nn/layer/image.py +38 -40
- mindspore/nn/layer/math.py +111 -112
- mindspore/nn/layer/normalization.py +56 -44
- mindspore/nn/layer/pooling.py +58 -63
- mindspore/nn/layer/rnn_cells.py +33 -33
- mindspore/nn/layer/rnns.py +56 -56
- mindspore/nn/layer/thor_layer.py +74 -73
- mindspore/nn/layer/transformer.py +11 -1
- mindspore/nn/learning_rate_schedule.py +20 -20
- mindspore/nn/loss/loss.py +79 -81
- mindspore/nn/optim/adam.py +2 -4
- mindspore/nn/optim/adasum.py +2 -2
- mindspore/nn/optim/lamb.py +1 -3
- mindspore/nn/optim/optimizer.py +1 -1
- mindspore/nn/optim/tft_wrapper.py +2 -3
- mindspore/nn/optim/thor.py +2 -2
- mindspore/nn/probability/distribution/_utils/utils.py +2 -2
- mindspore/nn/probability/distribution/exponential.py +2 -1
- mindspore/nn/probability/distribution/poisson.py +2 -1
- mindspore/nn/sparse/sparse.py +3 -3
- mindspore/nn/wrap/cell_wrapper.py +73 -42
- mindspore/nn/wrap/grad_reducer.py +37 -52
- mindspore/nn/wrap/loss_scale.py +72 -74
- mindspore/numpy/array_creations.py +7 -7
- mindspore/numpy/fft.py +1 -1
- mindspore/numpy/math_ops.py +1 -1
- mindspore/numpy/utils_const.py +1 -1
- mindspore/opencv_core452.dll +0 -0
- mindspore/opencv_imgcodecs452.dll +0 -0
- mindspore/opencv_imgproc452.dll +0 -0
- mindspore/ops/_grad_experimental/grad_comm_ops.py +51 -13
- mindspore/ops/_grad_experimental/grad_debug_ops.py +14 -0
- mindspore/ops/_grad_experimental/grad_inner_ops.py +0 -9
- mindspore/ops/_op_impl/cpu/__init__.py +1 -0
- mindspore/{experimental/es/__init__.py → ops/_op_impl/cpu/joinedstr_op.py} +12 -6
- mindspore/ops/_vmap/vmap_array_ops.py +6 -13
- mindspore/ops/_vmap/vmap_nn_ops.py +8 -16
- mindspore/ops/auto_generate/cpp_create_prim_instance_helper.py +29 -10
- mindspore/ops/auto_generate/gen_extend_func.py +5 -55
- mindspore/ops/auto_generate/gen_ops_def.py +753 -273
- mindspore/ops/auto_generate/gen_ops_prim.py +1687 -958
- mindspore/ops/auto_generate/pyboost_inner_prim.py +31 -1
- mindspore/ops/composite/__init__.py +10 -0
- mindspore/ops/composite/base.py +9 -5
- mindspore/ops/composite/multitype_ops/__init__.py +12 -1
- mindspore/ops/composite/multitype_ops/_compile_utils.py +132 -108
- mindspore/ops/composite/multitype_ops/_constexpr_utils.py +1 -1
- mindspore/ops/composite/multitype_ops/add_impl.py +70 -2
- mindspore/ops/composite/multitype_ops/div_impl.py +49 -0
- mindspore/ops/composite/multitype_ops/floordiv_impl.py +29 -0
- mindspore/ops/composite/multitype_ops/getitem_impl.py +11 -0
- mindspore/ops/composite/multitype_ops/mod_impl.py +5 -3
- mindspore/ops/composite/multitype_ops/mul_impl.py +49 -0
- mindspore/ops/composite/multitype_ops/setitem_impl.py +57 -0
- mindspore/ops/composite/multitype_ops/sub_impl.py +34 -0
- mindspore/ops/composite/multitype_ops/zeros_like_impl.py +14 -0
- mindspore/ops/function/__init__.py +4 -1
- mindspore/ops/function/_add_attr_func.py +11 -6
- mindspore/ops/function/array_func.py +17 -100
- mindspore/ops/function/debug_func.py +8 -5
- mindspore/ops/function/grad/grad_func.py +5 -13
- mindspore/ops/function/math_func.py +65 -399
- mindspore/ops/function/nn_func.py +44 -61
- mindspore/ops/function/other_func.py +4 -1
- mindspore/ops/function/random_func.py +31 -4
- mindspore/ops/functional.py +2 -3
- mindspore/ops/functional_overload.py +486 -18
- mindspore/ops/op_info_register.py +21 -0
- mindspore/ops/operations/__init__.py +5 -2
- mindspore/ops/operations/_custom_ops_utils.py +675 -8
- mindspore/ops/operations/_inner_ops.py +14 -18
- mindspore/ops/operations/_sequence_ops.py +1 -1
- mindspore/ops/operations/array_ops.py +4 -50
- mindspore/ops/operations/comm_ops.py +186 -41
- mindspore/ops/operations/custom_ops.py +244 -175
- mindspore/ops/operations/debug_ops.py +55 -4
- mindspore/ops/operations/image_ops.py +13 -13
- mindspore/ops/operations/manually_defined/ops_def.py +27 -28
- mindspore/ops/operations/math_ops.py +8 -9
- mindspore/ops/operations/nn_ops.py +6 -7
- mindspore/ops/primitive.py +9 -20
- mindspore/ops/tensor_method.py +52 -11
- mindspore/ops_generate/api/cpp_create_prim_instance_helper_generator.py +1 -1
- mindspore/ops_generate/api/functional_map_cpp_generator.py +10 -9
- mindspore/ops_generate/api/functions_cc_generator.py +58 -10
- mindspore/ops_generate/api/tensor_func_reg_cpp_generator.py +1 -1
- mindspore/ops_generate/common/base_generator.py +14 -0
- mindspore/ops_generate/common/gen_constants.py +7 -2
- mindspore/ops_generate/common/gen_utils.py +0 -19
- mindspore/ops_generate/common/op_proto.py +11 -4
- mindspore/ops_generate/common/template.py +88 -11
- mindspore/ops_generate/gen_ops.py +1 -1
- mindspore/ops_generate/op_def/lite_ops_cpp_generator.py +4 -4
- mindspore/ops_generate/op_def/ops_name_h_generator.py +0 -3
- mindspore/ops_generate/op_def/ops_primitive_h_generator.py +0 -4
- mindspore/ops_generate/op_def_py/op_prim_py_generator.py +5 -2
- mindspore/ops_generate/pyboost/auto_grad_impl_cc_generator.py +49 -8
- mindspore/ops_generate/pyboost/auto_grad_reg_cc_generator.py +2 -2
- mindspore/ops_generate/pyboost/gen_pyboost_func.py +31 -16
- mindspore/ops_generate/pyboost/op_template_parser.py +98 -72
- mindspore/ops_generate/pyboost/pyboost_functions_cpp_generator.py +70 -273
- mindspore/ops_generate/pyboost/pyboost_functions_h_generator.py +14 -6
- mindspore/ops_generate/pyboost/pyboost_functions_impl_cpp_generator.py +316 -0
- mindspore/ops_generate/pyboost/pyboost_functions_py_generator.py +1 -1
- mindspore/ops_generate/pyboost/pyboost_grad_function_cpp_generator.py +5 -3
- mindspore/ops_generate/pyboost/pyboost_inner_prim_generator.py +1 -1
- mindspore/ops_generate/pyboost/pyboost_internal_functions_cpp_generator.py +76 -0
- mindspore/ops_generate/pyboost/pyboost_internal_functions_h_generator.py +76 -0
- mindspore/ops_generate/pyboost/pyboost_internal_kernel_info_adapter_generator.py +125 -0
- mindspore/ops_generate/pyboost/pyboost_native_grad_functions_generator.py +4 -3
- mindspore/ops_generate/pyboost/pyboost_op_cpp_code_generator.py +348 -61
- mindspore/ops_generate/pyboost/pyboost_overload_functions_cpp_generator.py +1 -1
- mindspore/ops_generate/pyboost/pyboost_utils.py +118 -9
- mindspore/ops_generate/tensor_py_cc_generator.py +1 -24
- mindspore/parallel/_auto_parallel_context.py +9 -17
- mindspore/parallel/_cell_wrapper.py +106 -40
- mindspore/parallel/_parallel_serialization.py +4 -3
- mindspore/parallel/_ps_context.py +4 -6
- mindspore/parallel/_tensor.py +167 -12
- mindspore/parallel/_transformer/moe.py +1 -1
- mindspore/parallel/_transformer/transformer.py +17 -12
- mindspore/parallel/_utils.py +5 -11
- mindspore/parallel/auto_parallel.py +33 -12
- mindspore/parallel/checkpoint_convert.py +3 -3
- mindspore/parallel/checkpoint_transform.py +5 -1
- mindspore/parallel/cluster/process_entity/_api.py +88 -49
- mindspore/parallel/cluster/process_entity/_utils.py +95 -7
- mindspore/parallel/cluster/run.py +48 -7
- mindspore/parallel/function/__init__.py +8 -1
- mindspore/parallel/function/reshard_func.py +7 -6
- mindspore/parallel/nn/__init__.py +15 -2
- mindspore/parallel/nn/parallel_cell_wrapper.py +50 -14
- mindspore/parallel/nn/parallel_grad_reducer.py +7 -14
- mindspore/parallel/shard.py +9 -23
- mindspore/parallel/transform_safetensors.py +468 -174
- mindspore/pgodb140.dll +0 -0
- mindspore/pgort140.dll +0 -0
- mindspore/profiler/__init__.py +2 -1
- mindspore/profiler/analysis/parser/timeline_assembly_factory/ascend_timeline_assembler.py +7 -7
- mindspore/profiler/analysis/parser/timeline_assembly_factory/base_timeline_assembler.py +3 -0
- mindspore/profiler/analysis/parser/timeline_assembly_factory/trace_view_container.py +3 -0
- mindspore/profiler/analysis/parser/timeline_creator/cpu_op_timeline_creator.py +3 -3
- mindspore/profiler/analysis/parser/timeline_creator/fwk_timeline_creator.py +3 -3
- mindspore/profiler/analysis/parser/timeline_creator/msprof_timeline_creator.py +4 -4
- mindspore/profiler/analysis/parser/timeline_creator/scope_layer_timeline_creator.py +3 -3
- mindspore/profiler/analysis/parser/timeline_event/fwk_event.py +4 -1
- mindspore/profiler/analysis/parser/timeline_event/timeline_event_pool.py +2 -1
- mindspore/profiler/analysis/task_manager.py +1 -1
- mindspore/profiler/analysis/viewer/ascend_communication_viewer.py +5 -1
- mindspore/profiler/analysis/viewer/ascend_integrate_viewer.py +2 -1
- mindspore/profiler/analysis/viewer/ascend_kernel_details_viewer.py +10 -9
- mindspore/profiler/analysis/viewer/ascend_op_memory_viewer.py +43 -23
- mindspore/profiler/analysis/viewer/ascend_step_trace_time_viewer.py +3 -2
- mindspore/profiler/analysis/viewer/ms_minddata_viewer.py +9 -5
- mindspore/profiler/analysis/viewer/ms_operator_details_viewer.py +132 -0
- mindspore/profiler/common/constant.py +16 -0
- mindspore/profiler/common/msprof_cmd_tool.py +2 -2
- mindspore/profiler/common/path_manager.py +9 -0
- mindspore/profiler/common/profiler_context.py +50 -29
- mindspore/profiler/common/profiler_info.py +0 -16
- mindspore/profiler/common/profiler_meta_data.py +1 -0
- mindspore/profiler/common/profiler_op_analyse.py +239 -0
- mindspore/profiler/common/profiler_output_path.py +23 -8
- mindspore/profiler/common/profiler_parameters.py +128 -35
- mindspore/profiler/dynamic_profile/__init__.py +0 -0
- mindspore/profiler/dynamic_profile/dynamic_monitor_proxy.py +39 -0
- mindspore/profiler/dynamic_profile/dynamic_profiler_config_context.py +666 -0
- mindspore/profiler/dynamic_profile/dynamic_profiler_utils.py +62 -0
- mindspore/profiler/dynamic_profiler.py +374 -338
- mindspore/profiler/envprofiler.py +42 -12
- mindspore/profiler/experimental_config.py +112 -7
- mindspore/profiler/mstx.py +33 -12
- mindspore/profiler/platform/__init__.py +2 -3
- mindspore/profiler/platform/cpu_profiler.py +10 -4
- mindspore/profiler/platform/npu_profiler.py +30 -20
- mindspore/profiler/profiler.py +218 -154
- mindspore/profiler/profiler_action_controller.py +65 -77
- mindspore/profiler/profiler_interface.py +2 -2
- mindspore/profiler/schedule.py +10 -4
- mindspore/rewrite/common/config.py +1 -0
- mindspore/rewrite/common/namer.py +1 -0
- mindspore/rewrite/common/namespace.py +1 -0
- mindspore/rewrite/node/node.py +31 -11
- mindspore/rewrite/parsers/assign_parser.py +1 -1
- mindspore/rewrite/symbol_tree/symbol_tree.py +2 -2
- mindspore/run_check/_check_version.py +7 -10
- mindspore/runtime/__init__.py +8 -6
- mindspore/runtime/event.py +10 -4
- mindspore/runtime/executor.py +87 -45
- mindspore/runtime/memory.py +22 -30
- mindspore/runtime/thread_bind_core.py +299 -165
- mindspore/safeguard/rewrite_obfuscation.py +12 -13
- mindspore/swresample-4.dll +0 -0
- mindspore/swscale-6.dll +0 -0
- mindspore/tbbmalloc.dll +0 -0
- mindspore/tinyxml2.dll +0 -0
- mindspore/train/_utils.py +9 -5
- mindspore/train/amp.py +43 -23
- mindspore/train/callback/__init__.py +5 -5
- mindspore/train/callback/_callback.py +2 -1
- mindspore/train/callback/_checkpoint.py +4 -14
- mindspore/train/callback/_flops_collector.py +11 -7
- mindspore/train/callback/_landscape.py +0 -1
- mindspore/train/callback/_train_fault_tolerance.py +72 -18
- mindspore/train/data_sink.py +15 -6
- mindspore/train/dataset_helper.py +14 -5
- mindspore/train/model.py +49 -47
- mindspore/train/serialization.py +168 -126
- mindspore/train/summary/summary_record.py +13 -2
- mindspore/train/train_thor/model_thor.py +2 -2
- mindspore/turbojpeg.dll +0 -0
- mindspore/utils/__init__.py +3 -2
- mindspore/utils/dryrun.py +0 -6
- mindspore/utils/runtime_execution_order_check.py +162 -78
- mindspore/utils/sdc_detect.py +68 -0
- mindspore/utils/utils.py +14 -17
- mindspore/vcmeta.dll +0 -0
- mindspore/vcruntime140.dll +0 -0
- mindspore/vcruntime140_1.dll +0 -0
- mindspore/version.py +1 -1
- {mindspore-2.6.0.dist-info → mindspore-2.7.0.dist-info}/METADATA +5 -4
- {mindspore-2.6.0.dist-info → mindspore-2.7.0.dist-info}/RECORD +400 -439
- mindspore/_deprecated/jit.py +0 -198
- mindspore/_extends/remote/kernel_build_server_ascend.py +0 -75
- mindspore/communication/_hccl_management.py +0 -297
- mindspore/experimental/es/embedding_service.py +0 -891
- mindspore/experimental/es/embedding_service_layer.py +0 -581
- mindspore/profiler/common/validator/__init__.py +0 -14
- mindspore/profiler/common/validator/validate_path.py +0 -84
- mindspore/profiler/parser/__init__.py +0 -14
- mindspore/profiler/parser/aicpu_data_parser.py +0 -272
- mindspore/profiler/parser/ascend_analysis/__init__.py +0 -14
- mindspore/profiler/parser/ascend_analysis/constant.py +0 -71
- mindspore/profiler/parser/ascend_analysis/file_manager.py +0 -180
- mindspore/profiler/parser/ascend_analysis/function_event.py +0 -185
- mindspore/profiler/parser/ascend_analysis/fwk_cann_parser.py +0 -136
- mindspore/profiler/parser/ascend_analysis/fwk_file_parser.py +0 -131
- mindspore/profiler/parser/ascend_analysis/msprof_timeline_parser.py +0 -104
- mindspore/profiler/parser/ascend_analysis/path_manager.py +0 -313
- mindspore/profiler/parser/ascend_analysis/profiler_info_parser.py +0 -123
- mindspore/profiler/parser/ascend_analysis/tlv_decoder.py +0 -86
- mindspore/profiler/parser/ascend_analysis/trace_event_manager.py +0 -75
- mindspore/profiler/parser/ascend_cluster_generator.py +0 -116
- mindspore/profiler/parser/ascend_communicate_generator.py +0 -314
- mindspore/profiler/parser/ascend_flops_generator.py +0 -116
- mindspore/profiler/parser/ascend_fpbp_generator.py +0 -82
- mindspore/profiler/parser/ascend_hccl_generator.py +0 -271
- mindspore/profiler/parser/ascend_integrate_generator.py +0 -42
- mindspore/profiler/parser/ascend_memory_generator.py +0 -185
- mindspore/profiler/parser/ascend_msprof_exporter.py +0 -282
- mindspore/profiler/parser/ascend_msprof_generator.py +0 -187
- mindspore/profiler/parser/ascend_op_generator.py +0 -334
- mindspore/profiler/parser/ascend_steptrace_generator.py +0 -94
- mindspore/profiler/parser/ascend_timeline_generator.py +0 -545
- mindspore/profiler/parser/base_timeline_generator.py +0 -483
- mindspore/profiler/parser/container.py +0 -229
- mindspore/profiler/parser/cpu_gpu_timeline_generator.py +0 -697
- mindspore/profiler/parser/flops_parser.py +0 -531
- mindspore/profiler/parser/framework_enum.py +0 -111
- mindspore/profiler/parser/framework_parser.py +0 -464
- mindspore/profiler/parser/framework_struct.py +0 -61
- mindspore/profiler/parser/gpu_analysis/__init__.py +0 -14
- mindspore/profiler/parser/gpu_analysis/function_event.py +0 -44
- mindspore/profiler/parser/gpu_analysis/fwk_file_parser.py +0 -89
- mindspore/profiler/parser/gpu_analysis/profiler_info_parser.py +0 -72
- mindspore/profiler/parser/hccl_parser.py +0 -573
- mindspore/profiler/parser/hwts_log_parser.py +0 -122
- mindspore/profiler/parser/integrator.py +0 -526
- mindspore/profiler/parser/memory_usage_parser.py +0 -277
- mindspore/profiler/parser/minddata_analyzer.py +0 -800
- mindspore/profiler/parser/minddata_parser.py +0 -186
- mindspore/profiler/parser/minddata_pipeline_parser.py +0 -299
- mindspore/profiler/parser/op_intermediate_parser.py +0 -149
- mindspore/profiler/parser/optime_parser.py +0 -250
- mindspore/profiler/parser/profiler_info.py +0 -213
- mindspore/profiler/parser/step_trace_parser.py +0 -666
- mindspore/utils/hooks.py +0 -81
- /mindspore/common/{_auto_dynamic.py → dynamic_shape/_auto_dynamic.py} +0 -0
- {mindspore-2.6.0.dist-info → mindspore-2.7.0.dist-info}/WHEEL +0 -0
- {mindspore-2.6.0.dist-info → mindspore-2.7.0.dist-info}/entry_points.txt +0 -0
- {mindspore-2.6.0.dist-info → mindspore-2.7.0.dist-info}/top_level.txt +0 -0
|
@@ -41,7 +41,7 @@ from mindspore.ops.operations.nn_ops import TripletMarginLoss
|
|
|
41
41
|
from mindspore.ops.operations._sequence_ops import TupleToTensor, TensorToTuple, ListToTensor
|
|
42
42
|
from mindspore.common.api import _function_forbid_reuse
|
|
43
43
|
from mindspore.ops.auto_generate import log_softmax, dense, prelu, celu, fast_gelu, silu, elu, sigmoid, relu6, \
|
|
44
|
-
softmax_impl, swiglu, logsigmoid_op, kl_div_op, divs_op
|
|
44
|
+
softmax_impl, swiglu, logsigmoid_op, kl_div_op, divs_op, l1_loss_ext
|
|
45
45
|
from mindspore.ops.auto_generate import relu_op, inplace_relu_op
|
|
46
46
|
from mindspore.ops.auto_generate import group_norm_op, rms_norm, add_rms_norm, layer_norm_ext_op, batch_norm_ext_op,\
|
|
47
47
|
mse_loss_ext
|
|
@@ -49,13 +49,14 @@ from mindspore.ops.auto_generate import group_norm_op, rms_norm, add_rms_norm, l
|
|
|
49
49
|
from mindspore.ops.auto_generate import (reflection_pad_1d_op, reflection_pad_2d_op, add_layernorm_v2_op,
|
|
50
50
|
reflection_pad_3d_op, # pylint: disable=W0611
|
|
51
51
|
replication_pad_1d_op, replication_pad_2d_op, replication_pad_3d_op,
|
|
52
|
-
constant_pad_nd_op,
|
|
52
|
+
constant_pad_nd_op, func_dropout_ext_op, reverse_v2_impl, avg_pool2d_op,
|
|
53
53
|
upsample_nearest1d_op, upsample_nearest2d_op, upsample_nearest3d_op,
|
|
54
54
|
upsample_linear1d_op, upsample_bilinear2d_op, upsample_bicubic2d_op,
|
|
55
55
|
upsample_trilinear3d_impl, fill_scalar_op, floor_op, nllloss_2d_op,
|
|
56
|
-
masked_fill_op, masked_select, ones, flatten_ext, conv_transpose2d
|
|
56
|
+
masked_fill_op, masked_select, ones, flatten_ext, conv_transpose2d,
|
|
57
|
+
func_max_pool2d_op)
|
|
57
58
|
# 2
|
|
58
|
-
|
|
59
|
+
from mindspore.ops.auto_generate.pyboost_inner_prim import grid_sampler_2d_impl, grid_sampler_3d_impl
|
|
59
60
|
# 3
|
|
60
61
|
|
|
61
62
|
# 4
|
|
@@ -91,16 +92,16 @@ from mindspore.ops.auto_generate import avg_pool3d_ext_op
|
|
|
91
92
|
# 19
|
|
92
93
|
|
|
93
94
|
# 20
|
|
94
|
-
|
|
95
|
+
from mindspore.ops.functional_overload import conv3d as conv3d_op
|
|
95
96
|
from mindspore.ops.auto_generate.gen_ops_prim import embedding_op, MaxPoolWithIndices, \
|
|
96
97
|
PromptFlashAttention, MaxPoolWithMask
|
|
97
|
-
from mindspore.ops.auto_generate.gen_ops_prim import
|
|
98
|
+
from mindspore.ops.auto_generate.gen_ops_prim import conv2d_ext_op, \
|
|
98
99
|
conv2d_padding_op, conv1d_ext_op, conv1d_padding_op, speed_fusion_attention_op
|
|
99
100
|
from mindspore.common.generator import default_generator
|
|
100
101
|
from mindspore.ops.auto_generate import hardshrink, hardsigmoid, hardswish
|
|
101
102
|
from mindspore.ops.auto_generate import softshrink
|
|
102
103
|
from mindspore.ops.auto_generate import soft_margin_loss
|
|
103
|
-
from mindspore.ops.auto_generate import moe_token_permute, moe_token_unpermute
|
|
104
|
+
from mindspore.ops.auto_generate import moe_token_permute, moe_token_unpermute, ring_attention_update
|
|
104
105
|
from mindspore.ops.auto_generate import adaptive_avg_pool2d_ext_op
|
|
105
106
|
from mindspore.ops.auto_generate.pyboost_inner_prim import nllloss_impl
|
|
106
107
|
from mindspore.ops.auto_generate.pyboost_inner_prim import adaptive_max_pool2d_impl
|
|
@@ -1622,7 +1623,6 @@ def dropout_ext(input, p=0.5, training=True, inplace=False):
|
|
|
1622
1623
|
- **output** (Tensor) - Zeroed tensor, with the same shape and data type as `input`.
|
|
1623
1624
|
|
|
1624
1625
|
Raises:
|
|
1625
|
-
TypeError: If `p` is not a float.
|
|
1626
1626
|
TypeError: If `input` is not a Tensor.
|
|
1627
1627
|
|
|
1628
1628
|
Supported Platforms:
|
|
@@ -1636,16 +1636,8 @@ def dropout_ext(input, p=0.5, training=True, inplace=False):
|
|
|
1636
1636
|
>>> print(output.shape)
|
|
1637
1637
|
(2, 2)
|
|
1638
1638
|
"""
|
|
1639
|
-
check_bool_const(training, "training", "dropout_ext")
|
|
1640
|
-
check_bool_const(inplace, "inplace", "dropout_ext")
|
|
1641
|
-
if not training:
|
|
1642
|
-
return input
|
|
1643
1639
|
seed, offset = default_generator._step(generator_step_) # pylint: disable=protected-access
|
|
1644
|
-
|
|
1645
|
-
if inplace:
|
|
1646
|
-
input.copy_(out)
|
|
1647
|
-
return input
|
|
1648
|
-
return out
|
|
1640
|
+
return func_dropout_ext_op(input, p, training, inplace, seed, offset)
|
|
1649
1641
|
|
|
1650
1642
|
|
|
1651
1643
|
def dropout1d(input, p=0.5, training=True):
|
|
@@ -2394,7 +2386,7 @@ def is_floating_point(input):
|
|
|
2394
2386
|
>>> mindspore.ops.is_floating_point(input)
|
|
2395
2387
|
False
|
|
2396
2388
|
"""
|
|
2397
|
-
return input.dtype in
|
|
2389
|
+
return input.dtype in mstype.float_type
|
|
2398
2390
|
|
|
2399
2391
|
|
|
2400
2392
|
def _is_dim_unknown(shape):
|
|
@@ -4302,6 +4294,15 @@ def nll_loss(inputs, target, weight=None, ignore_index=-100, reduction='mean', l
|
|
|
4302
4294
|
\sum_{n=1}^{N} l_{n}, & \text { if reduction }=\text { 'sum' }
|
|
4303
4295
|
\end{array}\right.
|
|
4304
4296
|
|
|
4297
|
+
.. warning::
|
|
4298
|
+
- In GE mode, the rank of `inputs` should be 1D or 2D, the rank of `target` and `weight` should be 1D,
|
|
4299
|
+
and the following restraints should be met:
|
|
4300
|
+
|
|
4301
|
+
- when `inputs` is 1D: target_shape[0] == 1 and weight_shape[0] == inputs_shape[0].
|
|
4302
|
+
- when `inputs` is 2D: target_shape[0] == inputs_shape[0] and weight_shape[0] == inputs_shape[1].
|
|
4303
|
+
|
|
4304
|
+
- On GPU or CPU, `inputs` should be 2D.
|
|
4305
|
+
|
|
4305
4306
|
Args:
|
|
4306
4307
|
inputs (Tensor): :math:`(N, C)` where `C = number of classes` or :math:`(N, C, H, W)`
|
|
4307
4308
|
in case of 2D Loss, or :math:`(N, C, d_1, d_2, ..., d_K)`.
|
|
@@ -4420,7 +4421,7 @@ def nll_loss_ext(input, target, weight=None, ignore_index=-100, reduction='mean'
|
|
|
4420
4421
|
:math:`N` is the batch size, :math:`c` belonging to :math:`[0, C-1]` is class index,
|
|
4421
4422
|
where :math:`C` is the number of classes.
|
|
4422
4423
|
|
|
4423
|
-
If `reduction` is not ``'
|
|
4424
|
+
If `reduction` is not ``'none'`` (default ``'mean'``), then
|
|
4424
4425
|
|
|
4425
4426
|
.. math::
|
|
4426
4427
|
|
|
@@ -4444,7 +4445,7 @@ def nll_loss_ext(input, target, weight=None, ignore_index=-100, reduction='mean'
|
|
|
4444
4445
|
weight (Tensor, optional): A rescaling weight applied to the loss of each batch element.
|
|
4445
4446
|
If not None, the shape is :math:`(C,)`.
|
|
4446
4447
|
The data type must be float16 or float32 or bfloat16(only supported by Atlas A2 training series products).
|
|
4447
|
-
It should have the same data type as `input` . Default: ``
|
|
4448
|
+
It should have the same data type as `input` . Default: ``None`` .
|
|
4448
4449
|
ignore_index (int, optional): Specifies a target value that is ignored
|
|
4449
4450
|
and does not contribute to the input gradient. Default: ``-100`` .
|
|
4450
4451
|
reduction (str, optional): Apply specific reduction method to the output: ``'none'`` , ``'mean'`` ,
|
|
@@ -4481,10 +4482,10 @@ def _nllloss_nd(input, target, weight=None, ignore_index=-100, reduction='mean')
|
|
|
4481
4482
|
weight = ones(n_classes, input.dtype)
|
|
4482
4483
|
if input_dim < 1:
|
|
4483
4484
|
raise ValueError(f"input dim should be less than 1, but got {input_dim}")
|
|
4484
|
-
if input_dim != 1 and input.shape[0] != target.shape[0]:
|
|
4485
|
+
if F.isconstant(input_dim) and F.isconstant(target.ndim) and input_dim != 1 and input.shape[0] != target.shape[0]:
|
|
4485
4486
|
raise ValueError(f"input bacth_size should be equal to target batch_size, but got {input.shape[0]} and "
|
|
4486
4487
|
f"{target.shape[0]}")
|
|
4487
|
-
if input_dim
|
|
4488
|
+
if input_dim in [1, 2]:
|
|
4488
4489
|
return nllloss_impl(input, target, weight, reduction, ignore_index)[0]
|
|
4489
4490
|
if input_dim == 4:
|
|
4490
4491
|
return nllloss_2d_op(input, target, weight, reduction, ignore_index)[0]
|
|
@@ -4845,8 +4846,9 @@ def smooth_l1_loss(input, target, beta=1.0, reduction='none'):
|
|
|
4845
4846
|
>>> print(output)
|
|
4846
4847
|
[0. 0. 0.5]
|
|
4847
4848
|
"""
|
|
4848
|
-
|
|
4849
|
-
|
|
4849
|
+
if beta == 0.0:
|
|
4850
|
+
return l1_loss_ext(input, target, reduction)
|
|
4851
|
+
return ops.auto_generate.smooth_l1_loss(input, target, beta=beta, reduction=reduction)
|
|
4850
4852
|
|
|
4851
4853
|
|
|
4852
4854
|
def threshold(input, thr, value):
|
|
@@ -5365,7 +5367,7 @@ def max_pool3d(x, kernel_size, stride=None, padding=0, dilation=1, ceil_mode=Fal
|
|
|
5365
5367
|
return out
|
|
5366
5368
|
|
|
5367
5369
|
|
|
5368
|
-
def grid_sample(input, grid, mode='bilinear', padding_mode='zeros', align_corners=
|
|
5370
|
+
def grid_sample(input, grid, mode='bilinear', padding_mode='zeros', align_corners=None):
|
|
5369
5371
|
"""
|
|
5370
5372
|
Given an `input` and a flow-field `grid`, computes the `output` using `input` values and pixel locations from
|
|
5371
5373
|
`grid`. Only spatial (4-D) and volumetric (5-D) `input` is supported.
|
|
@@ -5411,10 +5413,10 @@ def grid_sample(input, grid, mode='bilinear', padding_mode='zeros', align_corner
|
|
|
5411
5413
|
padding_mode (str, optional): An optional string specifying the pad method.
|
|
5412
5414
|
The optional values are "zeros", "border" or
|
|
5413
5415
|
"reflection". Default: ``'zeros'`` .
|
|
5414
|
-
align_corners (bool, optional): If set to
|
|
5416
|
+
align_corners (bool, optional): If set to ``True``, the extrema (-1 and 1) are considered as referring to
|
|
5415
5417
|
the center points of the input's corner pixels. If set to `False`, they are instead considered as referring
|
|
5416
5418
|
to the corner points of the input's corner pixels, making the sampling more resolution agnostic. Default:
|
|
5417
|
-
``False`` .
|
|
5419
|
+
``None``, which is the same as ``False`` .
|
|
5418
5420
|
|
|
5419
5421
|
Returns:
|
|
5420
5422
|
Tensor, dtype is the same as `input` and whose shape is :math:`(N, C, H_{out}, W_{out})` (4-D) and
|
|
@@ -5451,11 +5453,10 @@ def grid_sample(input, grid, mode='bilinear', padding_mode='zeros', align_corner
|
|
|
5451
5453
|
[[14.5 ]
|
|
5452
5454
|
[14.8 ]]]]
|
|
5453
5455
|
"""
|
|
5456
|
+
align_corners = False if align_corners is None else align_corners
|
|
5454
5457
|
if input.ndim == 4:
|
|
5455
|
-
|
|
5456
|
-
|
|
5457
|
-
_grid_sampler_3d = _get_cache_prim(NN_OPS.GridSampler3D)(mode, padding_mode, align_corners)
|
|
5458
|
-
return _grid_sampler_3d(input, grid)
|
|
5458
|
+
return grid_sampler_2d_impl(input, grid, mode, padding_mode, align_corners)
|
|
5459
|
+
return grid_sampler_3d_impl(input, grid, mode, padding_mode, align_corners)
|
|
5459
5460
|
|
|
5460
5461
|
|
|
5461
5462
|
@constexpr
|
|
@@ -6307,7 +6308,7 @@ def conv1d_ext(input, weight, bias=None, stride=1, padding=0, dilation=1, groups
|
|
|
6307
6308
|
this 1D convolution layer also can be called 1D depthwise convolution layer. Default: ``1`` .
|
|
6308
6309
|
|
|
6309
6310
|
- :math:`(C_{in} \text{ % } \text{groups} == 0)` , :math:`(C_{out} \text{ % } \text{groups} == 0)` ,
|
|
6310
|
-
:math:`(C_{out} >= \text{groups})` , :math:`(\text{
|
|
6311
|
+
:math:`(C_{out} >= \text{groups})` , :math:`(\text{weight[1]} = C_{in} / \text{groups})`。
|
|
6311
6312
|
|
|
6312
6313
|
Returns:
|
|
6313
6314
|
Tensor, the value that applied 1D convolution. The shape is :math:`(N, C_{out}, L_{out})`.
|
|
@@ -6372,9 +6373,9 @@ def _get_pad_info(dilation, weight):
|
|
|
6372
6373
|
for i in range(2):
|
|
6373
6374
|
d = dilation[i]
|
|
6374
6375
|
weight_size = weight.shape[i + 2]
|
|
6375
|
-
|
|
6376
|
-
pad_l += (int(
|
|
6377
|
-
pad_r += (int(
|
|
6376
|
+
pad_item = d * (weight_size - 1)
|
|
6377
|
+
pad_l += (int(pad_item / 2),)
|
|
6378
|
+
pad_r += (int(pad_item - pad_l[i]),)
|
|
6378
6379
|
if pad_l[i] != pad_r[i]:
|
|
6379
6380
|
need_pad_nd = True
|
|
6380
6381
|
return need_pad_nd, pad_l, pad_r
|
|
@@ -6438,9 +6439,6 @@ def conv2d_ext(input, weight, bias=None, stride=1, padding=0, dilation=1, groups
|
|
|
6438
6439
|
<http://vision.stanford.edu/cs598_spring07/papers/Lecun98.pdf>`_ and
|
|
6439
6440
|
`ConvNets <http://cs231n.github.io/convolutional-networks/>`_.
|
|
6440
6441
|
|
|
6441
|
-
.. warning::
|
|
6442
|
-
This is an experimental API that is subject to change or deletion.
|
|
6443
|
-
|
|
6444
6442
|
Args:
|
|
6445
6443
|
input (Tensor): Tensor of shape :math:`(N, C_{in}, H_{in}, W_{in})` or :math:`(C_{in}, H_{in}, W_{in})`.
|
|
6446
6444
|
weight (Tensor): Tensor of shape
|
|
@@ -6477,7 +6475,7 @@ def conv2d_ext(input, weight, bias=None, stride=1, padding=0, dilation=1, groups
|
|
|
6477
6475
|
groups (int, optional): Splits `input` into groups. Default: ``1`` .
|
|
6478
6476
|
|
|
6479
6477
|
- :math:`(C_{in} \text{ % } \text{groups} == 0)` , :math:`(C_{out} \text{ % } \text{groups} == 0)` ,
|
|
6480
|
-
:math:`(C_{out} >= \text{groups})` , :math:`(\text{
|
|
6478
|
+
:math:`(C_{out} >= \text{groups})` , :math:`(\text{weight[1]} = C_{in} / \text{groups})`
|
|
6481
6479
|
|
|
6482
6480
|
Returns:
|
|
6483
6481
|
Tensor, the value that applied 2D convolution. The shape is :math:`(N, C_{out}, H_{out}, W_{out})`.
|
|
@@ -6981,10 +6979,6 @@ def batch_norm_ext(input, running_mean, running_var, weight=None, bias=None, tra
|
|
|
6981
6979
|
[[ 2.1621194 1.2360122]
|
|
6982
6980
|
[14.810596 10.180061 ]]
|
|
6983
6981
|
"""
|
|
6984
|
-
if weight is None:
|
|
6985
|
-
weight = ops.ones([input.shape[1]], dtype=input.dtype)
|
|
6986
|
-
if bias is None:
|
|
6987
|
-
bias = ops.zeros([input.shape[1]], dtype=input.dtype)
|
|
6988
6982
|
output = batch_norm_ext_op(input, weight, bias, running_mean, running_var, training, momentum, eps)
|
|
6989
6983
|
return output[0]
|
|
6990
6984
|
|
|
@@ -7382,8 +7376,8 @@ def conv3d_ext(input, weight, bias=None, stride=1, padding=0, dilation=1, groups
|
|
|
7382
7376
|
- :math:`H_{out} = (H_{in} + PadUp + PadDown - ((kh - 1) * DilationH + 1)) / StrideH + 1` .
|
|
7383
7377
|
- :math:`W_{out} = (W_{in} + PadLeft + PadRight - ((kw - 1) * DilationW + 1)) / StrideW + 1` .
|
|
7384
7378
|
- :math:`D_{out} = (D_{in} + PadFront + PadBack - ((kd - 1) * DilationD + 1)) / StrideD + 1` .
|
|
7385
|
-
- :math:`(D_{in}+PadFront+PadBack - ((kd-1)*DilationD+1))
|
|
7386
|
-
- :math:`(H_{in}+PadUp+PadDown - ((kh-1)*Dilationh+1))
|
|
7379
|
+
- :math:`(D_{in}+PadFront+PadBack - ((kd-1)*DilationD+1)) \% StrideD <= PadBack` .
|
|
7380
|
+
- :math:`(H_{in}+PadUp+PadDown - ((kh-1)*Dilationh+1)) \% StrideH <= PadDown` .
|
|
7387
7381
|
- :math:`stride_d <= kernel_d` .
|
|
7388
7382
|
- :math:`PadUp < kh` and :math:`PadDown < kh` . When `padding` = ``'valid'``, both PadUp and PadDown are zeros.
|
|
7389
7383
|
When `padding` = ``'same'``, pad can be calculated by
|
|
@@ -7396,7 +7390,6 @@ def conv3d_ext(input, weight, bias=None, stride=1, padding=0, dilation=1, groups
|
|
|
7396
7390
|
|
|
7397
7391
|
.. warning::
|
|
7398
7392
|
This API does not support Atlas series products.
|
|
7399
|
-
This is an experimental API that is subject to change or deletion.
|
|
7400
7393
|
|
|
7401
7394
|
Args:
|
|
7402
7395
|
input (Tensor): Tensor of shape :math:`(N, C_{in}, D_{in}, H_{in}, W_{in})`.
|
|
@@ -7450,12 +7443,7 @@ def conv3d_ext(input, weight, bias=None, stride=1, padding=0, dilation=1, groups
|
|
|
7450
7443
|
(12, 26, 59, 47, 5)
|
|
7451
7444
|
"""
|
|
7452
7445
|
|
|
7453
|
-
|
|
7454
|
-
return conv3d_ext_op(input, weight, bias, stride, padding, dilation, groups)
|
|
7455
|
-
if isinstance(padding, str):
|
|
7456
|
-
return conv3d_padding_op(input, weight, bias, stride, padding, dilation, groups)
|
|
7457
|
-
raise TypeError(f"For conv3d, the parameter 'padding' must be a tuple/list " \
|
|
7458
|
-
f"or a string, but got {type(padding)}")
|
|
7446
|
+
return conv3d_op(input, weight, bias, stride, padding, dilation, groups)
|
|
7459
7447
|
|
|
7460
7448
|
|
|
7461
7449
|
@_primexpr
|
|
@@ -7523,7 +7511,7 @@ def pixel_shuffle(input, upscale_factor):
|
|
|
7523
7511
|
c, h, w = idx[-3:]
|
|
7524
7512
|
_check_pxiel_shuffle_valid(c, upscale_factor)
|
|
7525
7513
|
c = c // upscale_factor ** 2
|
|
7526
|
-
input_perm =
|
|
7514
|
+
input_perm = pre + (c, upscale_factor, upscale_factor, h, w)
|
|
7527
7515
|
input = reshape_(input, input_perm)
|
|
7528
7516
|
input_perm = [i for i in range(length - 2)]
|
|
7529
7517
|
input_perm = input_perm + [length, length - 2, length + 1, length - 1]
|
|
@@ -7587,7 +7575,7 @@ def pixel_unshuffle(input, downscale_factor):
|
|
|
7587
7575
|
_check_pxiel_unshuffle_valid(h, w, downscale_factor)
|
|
7588
7576
|
h = h // downscale_factor
|
|
7589
7577
|
w = w // downscale_factor
|
|
7590
|
-
input_perm =
|
|
7578
|
+
input_perm = pre + (c, h, downscale_factor, w, downscale_factor)
|
|
7591
7579
|
input = reshape_(input, input_perm)
|
|
7592
7580
|
input_perm = [i for i in range(length - 2)]
|
|
7593
7581
|
input_perm = input_perm + [length - 1, length + 1, length - 2, length]
|
|
@@ -8912,13 +8900,7 @@ def max_pool2d_ext(input, kernel_size, stride=None, padding=0, dilation=1, ceil_
|
|
|
8912
8900
|
>>> print(argmax.shape)
|
|
8913
8901
|
(20, 16, 24, 31)
|
|
8914
8902
|
"""
|
|
8915
|
-
|
|
8916
|
-
if return_indices:
|
|
8917
|
-
max_pool_func_ = _get_cache_prim(MaxPoolWithIndices)(kernel_size, strides, padding, dilation, ceil_mode)
|
|
8918
|
-
out, indices = max_pool_func_(input)
|
|
8919
|
-
else:
|
|
8920
|
-
max_pool_func_ = _get_cache_prim(MaxPoolWithMask)(kernel_size, strides, padding, dilation, ceil_mode)
|
|
8921
|
-
out, indices = max_pool_func_(input)
|
|
8903
|
+
out, indices = func_max_pool2d_op(input, kernel_size, stride, padding, dilation, ceil_mode, return_indices)
|
|
8922
8904
|
if return_indices:
|
|
8923
8905
|
return out, indices
|
|
8924
8906
|
return out
|
|
@@ -9668,6 +9650,7 @@ __all__ = [
|
|
|
9668
9650
|
'lp_pool2d',
|
|
9669
9651
|
'moe_token_permute',
|
|
9670
9652
|
'moe_token_unpermute',
|
|
9653
|
+
'ring_attention_update',
|
|
9671
9654
|
'max_unpool1d',
|
|
9672
9655
|
'max_unpool2d',
|
|
9673
9656
|
'max_unpool3d',
|
|
@@ -15,6 +15,7 @@
|
|
|
15
15
|
"""Defines other operators with functional form."""
|
|
16
16
|
from mindspore.ops import operations as P
|
|
17
17
|
from mindspore.ops.auto_generate import rotary_position_embedding
|
|
18
|
+
from mindspore.ops.auto_generate import moe_distribute_dispatch, moe_distribute_combine
|
|
18
19
|
from mindspore.ops.auto_generate.gen_ops_prim import moe_init_routing_v2_op
|
|
19
20
|
|
|
20
21
|
partial_ = P.Partial()
|
|
@@ -268,6 +269,8 @@ __all__ = [
|
|
|
268
269
|
'partial',
|
|
269
270
|
'rotary_position_embedding',
|
|
270
271
|
'move_to',
|
|
271
|
-
'moe_init_routing_v2'
|
|
272
|
+
'moe_init_routing_v2',
|
|
273
|
+
'moe_distribute_dispatch',
|
|
274
|
+
'moe_distribute_combine'
|
|
272
275
|
]
|
|
273
276
|
__all__.sort()
|
|
@@ -296,7 +296,7 @@ def uniform_(input, from_=0, to=1, *, generator=None):
|
|
|
296
296
|
Returns:
|
|
297
297
|
Tensor, with the same shape and dtype as `input` tensor.
|
|
298
298
|
|
|
299
|
-
|
|
299
|
+
Raises:
|
|
300
300
|
TypeError: If `input` is not a Tensor.
|
|
301
301
|
TypeError: If dtype of `input` is not one of: bool, int8, int16, int32, int64, uint8, float16, float32, float64,
|
|
302
302
|
bfloat16.
|
|
@@ -765,9 +765,9 @@ def normal_ext(mean=0.0, std=1.0, size=None, generator=None):
|
|
|
765
765
|
Generates random numbers according to the standard Normal (or Gaussian) random number distribution.
|
|
766
766
|
|
|
767
767
|
Args:
|
|
768
|
-
mean (Union[
|
|
768
|
+
mean (Union[Tensor]): Mean value of each element, the shape of the `mean` tensor
|
|
769
769
|
should be the same as that of the `std` tensor.
|
|
770
|
-
std (Union[
|
|
770
|
+
std (Union[Tensor]): Standard deviation for each element, the shape of the `std` tensor
|
|
771
771
|
should be the same as that of the `mean` tensor. The value of `std` should be greater than or equal to 0.
|
|
772
772
|
|
|
773
773
|
Keyword Args:
|
|
@@ -793,6 +793,33 @@ def normal_ext(mean=0.0, std=1.0, size=None, generator=None):
|
|
|
793
793
|
>>> print(output.shape)
|
|
794
794
|
(3,)
|
|
795
795
|
|
|
796
|
+
.. function:: normal(mean, std) -> Tensor
|
|
797
|
+
:noindex:
|
|
798
|
+
|
|
799
|
+
Similar to the function above, but the means are shared among all drawn elements.
|
|
800
|
+
|
|
801
|
+
Args:
|
|
802
|
+
mean (float): Mean value of each element.
|
|
803
|
+
std (Tensor): Standard deviation for each element. The value of `std` should be greater
|
|
804
|
+
than or equal to 0.
|
|
805
|
+
|
|
806
|
+
Returns:
|
|
807
|
+
Outputs a tensor with the same shape as `std`.
|
|
808
|
+
|
|
809
|
+
Supported Platforms:
|
|
810
|
+
``Ascend``
|
|
811
|
+
|
|
812
|
+
Examples:
|
|
813
|
+
>>> import mindspore
|
|
814
|
+
>>> import numpy as np
|
|
815
|
+
>>> from mindspore import ops
|
|
816
|
+
>>> from mindspore import Tensor
|
|
817
|
+
>>> mean = 1.
|
|
818
|
+
>>> std = Tensor(np.array([1.0, 2.0, 3.0]), mindspore.float32)
|
|
819
|
+
>>> output = ops.function.random_func.normal_ext(mean, std)
|
|
820
|
+
>>> print(output.shape)
|
|
821
|
+
(3,)
|
|
822
|
+
|
|
796
823
|
.. function:: normal(mean, std=1.0) -> Tensor
|
|
797
824
|
:noindex:
|
|
798
825
|
|
|
@@ -1007,7 +1034,7 @@ def gamma(shape, alpha, beta, seed=None):
|
|
|
1007
1034
|
(3, 2, 2)
|
|
1008
1035
|
>>> # case 2: alpha_shape is (2, 3), so shape is (3, 1, 3)
|
|
1009
1036
|
>>> shape = (3, 1, 3)
|
|
1010
|
-
>>> alpha = mindspore.tensor([[1, 3, 4], [2, 5, 6]]
|
|
1037
|
+
>>> alpha = mindspore.tensor([[1, 3, 4], [2, 5, 6]], mindspore.float32)
|
|
1011
1038
|
>>> beta = mindspore.tensor([1.0], mindspore.float32)
|
|
1012
1039
|
>>> output = mindspore.ops.gamma(shape, alpha, beta, seed=5)
|
|
1013
1040
|
>>> result = output.shape
|
mindspore/ops/functional.py
CHANGED
|
@@ -20,7 +20,7 @@ from mindspore.common._register_for_tensor import tensor_operator_registry
|
|
|
20
20
|
from mindspore.ops import _constants
|
|
21
21
|
from mindspore.ops.function import *
|
|
22
22
|
from mindspore.ops.function.array_func import chunk_ext, zero_
|
|
23
|
-
from mindspore.ops.function.math_func import all, argmax_ext, float_power_ext, erfinv_, tanh_, bernoulli_ext
|
|
23
|
+
from mindspore.ops.function.math_func import all, argmax_ext, float_power_ext, erfinv_, tanh_, bernoulli_ext, bernoulli_
|
|
24
24
|
from mindspore.ops.function.random_func import random_, uniform_ext, uniform_, normal_, exponential_
|
|
25
25
|
from mindspore.ops import operations as P
|
|
26
26
|
from mindspore.ops.operations import array_ops
|
|
@@ -33,7 +33,6 @@ from mindspore.ops.operations.nn_ops import AdaptiveMaxPool2D
|
|
|
33
33
|
from mindspore.ops.operations.math_ops import Roll
|
|
34
34
|
from mindspore.ops.composite.math_ops import mm
|
|
35
35
|
from mindspore.ops.function.math_func import dot
|
|
36
|
-
from mindspore.ops.function.array_func import new_empty
|
|
37
36
|
from mindspore.ops import auto_generate
|
|
38
37
|
from mindspore.ops.auto_generate import cast
|
|
39
38
|
from mindspore.ops._utils.arg_dtype_cast import DtypeToEnum
|
|
@@ -383,7 +382,6 @@ setattr(tensor_operator_registry, 'nanmedian', nanmedian)
|
|
|
383
382
|
setattr(tensor_operator_registry, 'csr_to_coo', csr_to_coo)
|
|
384
383
|
setattr(tensor_operator_registry, 'zeros', zeros)
|
|
385
384
|
setattr(tensor_operator_registry, 'ones', ones)
|
|
386
|
-
setattr(tensor_operator_registry, 'new_empty', new_empty)
|
|
387
385
|
setattr(tensor_operator_registry, 'unsorted_segment_min', unsorted_segment_min)
|
|
388
386
|
setattr(tensor_operator_registry, 'unsorted_segment_max', unsorted_segment_max)
|
|
389
387
|
setattr(tensor_operator_registry, 'unsorted_segment_prod', unsorted_segment_prod)
|
|
@@ -399,6 +397,7 @@ setattr(tensor_operator_registry, 'inplace_scatter_add', auto_generate.inplace_s
|
|
|
399
397
|
setattr(tensor_operator_registry, 'slice_scatter', slice_scatter)
|
|
400
398
|
setattr(tensor_operator_registry, 'select_scatter', select_scatter)
|
|
401
399
|
setattr(tensor_operator_registry, 'bernoulli', bernoulli_ext)
|
|
400
|
+
setattr(tensor_operator_registry, 'bernoulli_', bernoulli_)
|
|
402
401
|
setattr(tensor_operator_registry, 'poisson', P.Poisson)
|
|
403
402
|
setattr(tensor_operator_registry, 'randperm', P.Randperm)
|
|
404
403
|
setattr(tensor_operator_registry, 'multinomial', multinomial)
|