mindspore 2.4.10__cp39-none-any.whl → 2.5.0__cp39-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of mindspore might be problematic. Click here for more details.
- mindspore/.commit_id +1 -1
- mindspore/Third_Party_Open_Source_Software_Notice +39 -0
- mindspore/__init__.py +8 -3
- mindspore/_akg/akg/composite/build_module.py +6 -2
- mindspore/_akg/akg/utils/kernel_exec.py +2 -2
- mindspore/_c_dataengine.cpython-39-aarch64-linux-gnu.so +0 -0
- mindspore/_c_expression.cpython-39-aarch64-linux-gnu.so +0 -0
- mindspore/_c_mindrecord.cpython-39-aarch64-linux-gnu.so +0 -0
- mindspore/_checkparam.py +0 -5
- mindspore/_extends/parallel_compile/akg_compiler/gen_custom_op_files.py +1 -1
- mindspore/_extends/parse/compile_config.py +64 -0
- mindspore/_extends/parse/deprecated/__init__.py +0 -0
- mindspore/_extends/parse/deprecated/deprecated_tensor_method.py +375 -0
- mindspore/_extends/parse/parser.py +23 -5
- mindspore/_extends/parse/standard_method.py +123 -27
- mindspore/_extends/pijit/pijit_func_white_list.py +1 -1
- mindspore/amp.py +7 -1
- mindspore/boost/boost_cell_wrapper.py +136 -41
- mindspore/common/__init__.py +3 -1
- mindspore/common/_register_for_tensor.py +0 -1
- mindspore/common/_stub_tensor.py +25 -4
- mindspore/common/_tensor_cpp_method.py +17 -0
- mindspore/common/_tensor_docs.py +6132 -0
- mindspore/common/api.py +98 -21
- mindspore/common/dtype.py +34 -34
- mindspore/common/dump.py +2 -1
- mindspore/common/file_system.py +8 -3
- mindspore/common/generator.py +2 -0
- mindspore/common/hook_handle.py +3 -1
- mindspore/common/initializer.py +3 -4
- mindspore/common/lazy_inline.py +8 -2
- mindspore/common/mindir_util.py +10 -2
- mindspore/common/parameter.py +31 -15
- mindspore/common/tensor.py +713 -1337
- mindspore/communication/__init__.py +1 -1
- mindspore/communication/_comm_helper.py +5 -0
- mindspore/communication/comm_func.py +215 -173
- mindspore/communication/management.py +23 -20
- mindspore/context.py +285 -191
- mindspore/dataset/__init__.py +23 -19
- mindspore/dataset/callback/ds_callback.py +2 -1
- mindspore/dataset/core/config.py +84 -3
- mindspore/dataset/engine/cache_admin.py +3 -3
- mindspore/dataset/engine/cache_client.py +5 -4
- mindspore/dataset/engine/datasets.py +192 -149
- mindspore/dataset/engine/datasets_audio.py +14 -0
- mindspore/dataset/engine/datasets_standard_format.py +11 -11
- mindspore/dataset/engine/datasets_text.py +38 -1
- mindspore/dataset/engine/datasets_user_defined.py +100 -66
- mindspore/dataset/engine/datasets_vision.py +81 -8
- mindspore/dataset/engine/iterators.py +281 -63
- mindspore/dataset/engine/obs/util.py +8 -0
- mindspore/dataset/engine/queue.py +40 -0
- mindspore/dataset/engine/samplers.py +26 -2
- mindspore/dataset/engine/serializer_deserializer.py +1 -1
- mindspore/dataset/engine/validators.py +43 -11
- mindspore/dataset/transforms/py_transforms_util.py +17 -0
- mindspore/dataset/transforms/transforms.py +29 -12
- mindspore/dataset/vision/validators.py +1 -2
- mindspore/device_context/__init__.py +21 -0
- mindspore/device_context/ascend/__init__.py +25 -0
- mindspore/device_context/ascend/device.py +72 -0
- mindspore/device_context/ascend/op_debug.py +94 -0
- mindspore/device_context/ascend/op_precision.py +193 -0
- mindspore/device_context/ascend/op_tuning.py +127 -0
- mindspore/device_context/cpu/__init__.py +25 -0
- mindspore/device_context/cpu/device.py +62 -0
- mindspore/device_context/cpu/op_tuning.py +43 -0
- mindspore/device_context/gpu/__init__.py +21 -0
- mindspore/device_context/gpu/device.py +70 -0
- mindspore/device_context/gpu/op_precision.py +67 -0
- mindspore/device_context/gpu/op_tuning.py +175 -0
- mindspore/device_manager.py +134 -0
- mindspore/experimental/llm_boost/__init__.py +1 -0
- mindspore/experimental/llm_boost/ascend_native/__init__.py +22 -0
- mindspore/experimental/llm_boost/ascend_native/llama_boost_ascend_native.py +211 -0
- mindspore/experimental/llm_boost/ascend_native/llm_boost.py +52 -0
- mindspore/experimental/llm_boost/atb/boost_base.py +2 -3
- mindspore/experimental/llm_boost/atb/llama_boost.py +6 -1
- mindspore/experimental/llm_boost/register.py +1 -0
- mindspore/experimental/optim/adadelta.py +26 -22
- mindspore/experimental/optim/adam.py +3 -0
- mindspore/experimental/optim/lr_scheduler.py +33 -24
- mindspore/experimental/optim/radam.py +33 -30
- mindspore/hal/device.py +28 -0
- mindspore/hal/event.py +17 -0
- mindspore/hal/memory.py +94 -3
- mindspore/hal/stream.py +91 -6
- mindspore/include/api/context.h +0 -1
- mindspore/lib/libavcodec.so.59 +0 -0
- mindspore/lib/libavdevice.so.59 +0 -0
- mindspore/lib/libavfilter.so.8 +0 -0
- mindspore/lib/libavformat.so.59 +0 -0
- mindspore/lib/libavutil.so.57 +0 -0
- mindspore/lib/libdnnl.so.2 +0 -0
- mindspore/lib/libmindspore_backend.so +0 -0
- mindspore/lib/libmindspore_common.so +0 -0
- mindspore/lib/libmindspore_core.so +0 -0
- mindspore/lib/libmindspore_gpr.so.15 +0 -0
- mindspore/lib/libmindspore_grpc++.so.1 +0 -0
- mindspore/lib/libmindspore_grpc.so.15 +0 -0
- mindspore/lib/libmindspore_ops.so +0 -0
- mindspore/lib/libmpi_adapter.so +0 -0
- mindspore/lib/libmpi_collective.so +0 -0
- mindspore/lib/libnnacl.so +0 -0
- mindspore/lib/libopencv_core.so.4.5 +0 -0
- mindspore/lib/libopencv_imgcodecs.so.4.5 +0 -0
- mindspore/lib/libopencv_imgproc.so.4.5 +0 -0
- mindspore/lib/libps_cache.so +0 -0
- mindspore/lib/libswresample.so.4 +0 -0
- mindspore/lib/libswscale.so.6 +0 -0
- mindspore/lib/plugin/ascend/custom_aicore_ops/op_impl/ai_core/tbe/config/ascend910_93/aic-ascend910_93-ops-info.json +2048 -0
- mindspore/lib/plugin/ascend/custom_aicpu_ops/op_impl/cpu/aicpu_kernel/impl/libcust_cpu_kernels.so +0 -0
- mindspore/lib/plugin/ascend/custom_aicpu_ops/op_proto/libcust_op_proto.so +0 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910/op_api/lib/libcust_opapi.so +0 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910/op_impl/ai_core/tbe/custom_ascendc_910_impl/dynamic/decoder_kv_cache.py +1 -1
- mindspore/lib/plugin/ascend/custom_ascendc_910/op_impl/ai_core/tbe/custom_ascendc_910_impl/dynamic/prompt_kv_cache.py +1 -1
- mindspore/lib/plugin/ascend/custom_ascendc_910/op_impl/ai_core/tbe/op_tiling/lib/linux/aarch64/libcust_opmaster_rt2.0.so +0 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910/op_impl/ai_core/tbe/op_tiling/liboptiling.so +0 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910/op_proto/lib/linux/aarch64/libcust_opsproto_rt2.0.so +0 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910/version.info +1 -1
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_api/lib/libcust_opapi.so +0 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/config/ascend910_93/aic-ascend910_93-ops-info.json +224 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/custom_ascendc_910b_impl/dynamic/all_finite.py +1 -1
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/custom_ascendc_910b_impl/dynamic/decoder_kv_cache.py +1 -1
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/custom_ascendc_910b_impl/dynamic/prompt_kv_cache.py +1 -1
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/all_finite/AllFinite_52f59e2a65d9b1bb002de35c2819754a.json +78 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/all_finite/AllFinite_52f59e2a65d9b1bb002de35c2819754a.o +0 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/all_finite/AllFinite_6b5e50e30256d85838d6ce83514df20f.json +78 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/all_finite/AllFinite_6b5e50e30256d85838d6ce83514df20f.o +0 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/all_finite/AllFinite_74e4ac02880d452e3308c94af273562e.json +78 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/all_finite/AllFinite_74e4ac02880d452e3308c94af273562e.o +0 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/decoder_kv_cache/DecoderKvCache_0d5520cc587ad44ce634bf3fbcffc272.json +156 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/decoder_kv_cache/DecoderKvCache_0d5520cc587ad44ce634bf3fbcffc272.o +0 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/decoder_kv_cache/DecoderKvCache_20390d30b3c4c0d23167ccca6c030c2b.json +156 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/decoder_kv_cache/DecoderKvCache_20390d30b3c4c0d23167ccca6c030c2b.o +0 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/decoder_kv_cache/DecoderKvCache_2d151f0b1d2db51faa2968d5b67544e2.json +156 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/decoder_kv_cache/DecoderKvCache_2d151f0b1d2db51faa2968d5b67544e2.o +0 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/decoder_kv_cache/DecoderKvCache_561690ec17cc1def3d2fcf68c1b07b56.json +156 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/decoder_kv_cache/DecoderKvCache_561690ec17cc1def3d2fcf68c1b07b56.o +0 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/decoder_kv_cache/DecoderKvCache_570f9aaa99e5e773b3dd0a33784363f4.json +156 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/decoder_kv_cache/DecoderKvCache_570f9aaa99e5e773b3dd0a33784363f4.o +0 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/decoder_kv_cache/DecoderKvCache_59668a0f0764afb98fda8ab9e84126f1.json +156 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/decoder_kv_cache/DecoderKvCache_59668a0f0764afb98fda8ab9e84126f1.o +0 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/decoder_kv_cache/DecoderKvCache_91d9833e4792b70b670e4e2b916abd86.json +156 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/decoder_kv_cache/DecoderKvCache_91d9833e4792b70b670e4e2b916abd86.o +0 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/decoder_kv_cache/DecoderKvCache_c74cdc5fef094383401856f8519504af.json +156 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/decoder_kv_cache/DecoderKvCache_c74cdc5fef094383401856f8519504af.o +0 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/prompt_kv_cache/PromptKvCache_0515c7b1a4cd614449e38c5e9a7e3f8d.json +165 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/prompt_kv_cache/PromptKvCache_0515c7b1a4cd614449e38c5e9a7e3f8d.o +0 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/prompt_kv_cache/PromptKvCache_09f22d898d6358c91e7c4fc48bac48e7.json +165 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/prompt_kv_cache/PromptKvCache_09f22d898d6358c91e7c4fc48bac48e7.o +0 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/prompt_kv_cache/PromptKvCache_0cb9a6f894b925250227136e5aab7061.json +165 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/prompt_kv_cache/PromptKvCache_0cb9a6f894b925250227136e5aab7061.o +0 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/prompt_kv_cache/PromptKvCache_2fa8702ffd7ca85e9e194f62644415d5.json +165 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/prompt_kv_cache/PromptKvCache_2fa8702ffd7ca85e9e194f62644415d5.o +0 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/prompt_kv_cache/PromptKvCache_570b62f187dfd439b64613d881deedb7.json +165 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/prompt_kv_cache/PromptKvCache_570b62f187dfd439b64613d881deedb7.o +0 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/prompt_kv_cache/PromptKvCache_585218c11411ff84709b9e725b66c435.json +165 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/prompt_kv_cache/PromptKvCache_585218c11411ff84709b9e725b66c435.o +0 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/prompt_kv_cache/PromptKvCache_5c9365ccde170b358c5b126d69dae13e.json +165 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/prompt_kv_cache/PromptKvCache_5c9365ccde170b358c5b126d69dae13e.o +0 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/prompt_kv_cache/PromptKvCache_6d97c45b7c43bc16fcff8baa5dacac4e.json +165 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/ascend910_93/prompt_kv_cache/PromptKvCache_6d97c45b7c43bc16fcff8baa5dacac4e.o +0 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/config/ascend910_93/all_finite.json +139 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/config/ascend910_93/binary_info_config.json +361 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/config/ascend910_93/decoder_kv_cache.json +892 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/kernel/config/ascend910_93/prompt_kv_cache.json +892 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/op_tiling/lib/linux/aarch64/libcust_opmaster_rt2.0.so +0 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_impl/ai_core/tbe/op_tiling/liboptiling.so +0 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/op_proto/lib/linux/aarch64/libcust_opsproto_rt2.0.so +0 -0
- mindspore/lib/plugin/ascend/custom_ascendc_910b/version.info +1 -1
- mindspore/lib/plugin/ascend/custom_compiler/setup.py +1 -1
- mindspore/lib/plugin/ascend/libascend_collective.so +0 -0
- mindspore/lib/plugin/ascend/libdvpp_utils.so +0 -0
- mindspore/lib/plugin/ascend/liblowlatency_collective.so +0 -0
- mindspore/lib/plugin/ascend/libmindspore_cpu_kernels.so +0 -0
- mindspore/lib/plugin/ascend/libmindspore_internal_kernels.so +0 -0
- mindspore/lib/plugin/ascend/libms_ascend_native_boost.so +0 -0
- mindspore/lib/plugin/ascend/libms_atb_boost.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/asdops/device/ascend910b/bin/ascend910b.bin +957 -955
- mindspore/lib/plugin/ascend/ms_kernels_internal/asdops/lib/libasdops_static.a +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/asdops/lib/liblcal_static.a +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/{acme/include/base_type.h → base_type.h} +25 -20
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/{cast/cast_tiling.h → internal.h} +6 -4
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/internal_op.h +114 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/llm/boost_kernel.h +70 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/llm/llama_impl.h +85 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/llm/model_interface.h +52 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/llm/tensor.h +81 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/op_creator.h +123 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/op_param.h +155 -110
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/{acme/include/tiling_info.h → tiling_info.h} +12 -9
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/tiling_utils.h +178 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libadd_layer_norm_op.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libadd_rms_norm_op.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libadd_rms_norm_quant_op.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libapply_rotary_pos_emb_310p_op.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libapply_rotary_pos_emb_op.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libcast_op.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libcompare_op.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libgelu_op.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libllama_op.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libmatmul_op.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libms_kernels_internal.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libms_optiling.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libmulti_weight_matmul_kernel_op.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libreshape_and_cache_nz_op.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libreshape_and_cache_op.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/librms_norm_op.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend310p/object_kernels/internal_pp_matmul_f16_nz/internal_pp_matmul_f16_nz.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend310p/object_kernels/internal_pp_matmul_f16_nz/internal_pp_matmul_f16_nz_0.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend310p/object_kernels/internal_pp_matmul_i8_nz_compress/internal_pp_matmul_i8_nz_compress.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend310p/object_kernels/internal_pp_matmul_i8_nz_compress/internal_pp_matmul_i8_nz_compress_0.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend310p/object_kernels/internal_pp_matmul_int8_nz/internal_pp_matmul_int8_nz.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend310p/object_kernels/internal_pp_matmul_int8_nz/internal_pp_matmul_int8_nz_0.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend310p/so_kernels/libadd_rms_norm_quant_ascend310p.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/{lib/libapply_rotary_pos_emb_310p_impl.so → op_kernels/ascend310p/so_kernels/libapply_rotary_pos_emb_310p_ascend310p.so} +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend310p/so_kernels/libcast_ascend310p.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend310p/so_kernels/libcompare_ascend310p.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend310p/so_kernels/libgelu_ascend310p.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend310p/so_kernels/libmatmul_ascend310p.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend310p/so_kernels/libreshape_and_cache_nz_ascend310p.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/hphol_kernels/add_rms_norm_dynamic_quant/AddRmsNormDynamicQuant_4b60f88cdc28b25a36bad2d8b0a88092.json +163 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/hphol_kernels/add_rms_norm_dynamic_quant/AddRmsNormDynamicQuant_4b60f88cdc28b25a36bad2d8b0a88092.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/hphol_kernels/add_rms_norm_dynamic_quant/AddRmsNormDynamicQuant_cde61da2bd6fededcb1ba310a6ad16ee.json +163 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/hphol_kernels/add_rms_norm_dynamic_quant/AddRmsNormDynamicQuant_cde61da2bd6fededcb1ba310a6ad16ee.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/object_kernels/flash_attention_score/flash_attention_score_bf16_bnsd_full_mix.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/object_kernels/flash_attention_score/flash_attention_score_bf16_bnsd_tri_mix.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/object_kernels/flash_attention_score/flash_attention_score_bf16_bsh_full_mix.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/object_kernels/flash_attention_score/flash_attention_score_bf16_bsh_tri_mix.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/object_kernels/flash_attention_score/flash_attention_score_fp16_bnsd_full_mix.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/object_kernels/flash_attention_score/flash_attention_score_fp16_bnsd_tri_mix.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/object_kernels/flash_attention_score/flash_attention_score_fp16_bsh_full_mix.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/object_kernels/flash_attention_score/flash_attention_score_fp16_bsh_tri_mix.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/object_kernels/internal_matmul_postfusion_mix/internal_matmul_postfusion_mix.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/object_kernels/internal_matmul_postfusion_mix/internal_matmul_postfusion_mix_mix_aic_0.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/object_kernels/internal_matmul_postfusion_mix/internal_matmul_postfusion_mix_mix_aiv_0.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/object_kernels/internal_multi_weight_matmul_postfusion_mix/internal_multi_weight_matmul_postfusion_mix.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/object_kernels/internal_multi_weight_matmul_postfusion_mix/internal_multi_weight_matmul_postfusion_mix_mix_aic_0.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/object_kernels/internal_multi_weight_matmul_postfusion_mix/internal_multi_weight_matmul_postfusion_mix_mix_aiv_0.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/{matmul_add_rmsnorm → object_kernels/matmul_add_rmsnorm}/matmul_add_rmsnorm_bf16_bf16.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/{matmul_add_rmsnorm → object_kernels/matmul_add_rmsnorm}/matmul_add_rmsnorm_bf16_fp16.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/{matmul_add_rmsnorm → object_kernels/matmul_add_rmsnorm}/matmul_add_rmsnorm_bf16_fp32.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/{matmul_add_rmsnorm → object_kernels/matmul_add_rmsnorm}/matmul_add_rmsnorm_fp16_bf16.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/{matmul_add_rmsnorm → object_kernels/matmul_add_rmsnorm}/matmul_add_rmsnorm_fp16_fp16.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/{matmul_add_rmsnorm → object_kernels/matmul_add_rmsnorm}/matmul_add_rmsnorm_fp16_fp32.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/object_kernels/paged_attention_v2/paged_attention_v2.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/object_kernels/paged_attention_v2/paged_attention_v2_mix_aic_0.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/object_kernels/paged_attention_v2/paged_attention_v2_mix_aiv_0.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/{lib/libadd_layer_norm_impl.so → op_kernels/ascend910b/so_kernels/libadd_layer_norm_ascend910b.so} +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/{lib/libadd_rms_norm_impl.so → op_kernels/ascend910b/so_kernels/libadd_rms_norm_ascend910b.so} +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/so_kernels/libadd_rms_norm_quant_ascend910b.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/{lib/libapply_rotary_pos_emb_impl.so → op_kernels/ascend910b/so_kernels/libapply_rotary_pos_emb_ascend910b.so} +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/{lib/libcast_impl.so → op_kernels/ascend910b/so_kernels/libcast_ascend910b.so} +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/{lib/libnot_equal_impl.so → op_kernels/ascend910b/so_kernels/libcompare_ascend910b.so} +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/{lib/libgelu_impl.so → op_kernels/ascend910b/so_kernels/libgelu_ascend910b.so} +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/so_kernels/libllama_ascend910b.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/{lib/libmatmul_impl.so → op_kernels/ascend910b/so_kernels/libmatmul_ascend910b.so} +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/{lib/libmulti_weight_matmul_kernel_impl.so → op_kernels/ascend910b/so_kernels/libmulti_weight_matmul_kernel_ascend910b.so} +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/{lib/libreshape_and_cache_impl.so → op_kernels/ascend910b/so_kernels/libreshape_and_cache_ascend910b.so} +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/{lib/librms_norm_impl.so → op_kernels/ascend910b/so_kernels/librms_norm_ascend910b.so} +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/lccl/lib/liblccl_wrapper.so +0 -0
- mindspore/lib/plugin/libmindspore_ascend.so.2 +0 -0
- mindspore/log.py +12 -0
- mindspore/mindrecord/__init__.py +1 -1
- mindspore/mindrecord/config.py +17 -316
- mindspore/mindrecord/filereader.py +1 -9
- mindspore/mindrecord/filewriter.py +5 -15
- mindspore/mindrecord/mindpage.py +1 -9
- mindspore/mint/__init__.py +824 -218
- mindspore/mint/distributed/__init__.py +66 -4
- mindspore/mint/distributed/distributed.py +2594 -44
- mindspore/mint/linalg/__init__.py +6 -0
- mindspore/mint/nn/__init__.py +473 -14
- mindspore/mint/nn/functional.py +486 -11
- mindspore/mint/nn/layer/__init__.py +17 -4
- mindspore/mint/nn/layer/_functions.py +330 -0
- mindspore/mint/nn/layer/activation.py +169 -1
- mindspore/mint/nn/layer/basic.py +123 -0
- mindspore/mint/nn/layer/conv.py +727 -0
- mindspore/mint/nn/layer/normalization.py +215 -19
- mindspore/mint/nn/layer/padding.py +797 -0
- mindspore/mint/nn/layer/pooling.py +170 -0
- mindspore/mint/optim/__init__.py +2 -1
- mindspore/mint/optim/adam.py +223 -0
- mindspore/mint/optim/adamw.py +26 -19
- mindspore/mint/special/__init__.py +2 -1
- mindspore/multiprocessing/__init__.py +5 -0
- mindspore/nn/cell.py +126 -19
- mindspore/nn/dynamic_lr.py +2 -1
- mindspore/nn/layer/activation.py +6 -6
- mindspore/nn/layer/basic.py +35 -25
- mindspore/nn/layer/channel_shuffle.py +3 -3
- mindspore/nn/layer/embedding.py +3 -3
- mindspore/nn/layer/normalization.py +8 -7
- mindspore/nn/layer/padding.py +4 -3
- mindspore/nn/layer/pooling.py +47 -13
- mindspore/nn/layer/rnn_cells.py +1 -1
- mindspore/nn/layer/rnns.py +2 -1
- mindspore/nn/layer/timedistributed.py +5 -5
- mindspore/nn/layer/transformer.py +48 -26
- mindspore/nn/learning_rate_schedule.py +5 -3
- mindspore/nn/loss/loss.py +31 -36
- mindspore/nn/optim/ada_grad.py +1 -0
- mindspore/nn/optim/adadelta.py +2 -2
- mindspore/nn/optim/adam.py +1 -1
- mindspore/nn/optim/lars.py +1 -4
- mindspore/nn/optim/optimizer.py +1 -1
- mindspore/nn/optim/rprop.py +2 -2
- mindspore/nn/optim/thor.py +2 -1
- mindspore/nn/utils/init.py +13 -11
- mindspore/nn/wrap/cell_wrapper.py +4 -6
- mindspore/nn/wrap/loss_scale.py +3 -4
- mindspore/numpy/array_creations.py +60 -62
- mindspore/numpy/array_ops.py +148 -143
- mindspore/numpy/logic_ops.py +41 -42
- mindspore/numpy/math_ops.py +361 -359
- mindspore/numpy/utils.py +16 -16
- mindspore/numpy/utils_const.py +4 -4
- mindspore/ops/__init__.py +2 -1
- mindspore/ops/_grad_experimental/grad_comm_ops.py +94 -13
- mindspore/ops/_grad_experimental/grad_debug_ops.py +6 -1
- mindspore/ops/_grad_experimental/grad_inner_ops.py +9 -0
- mindspore/ops/_grad_experimental/grad_math_ops.py +2 -1
- mindspore/ops/_op_impl/cpu/__init__.py +1 -0
- mindspore/ops/_op_impl/cpu/raise_op.py +28 -0
- mindspore/ops/_vmap/vmap_array_ops.py +20 -19
- mindspore/ops/_vmap/vmap_base.py +0 -2
- mindspore/ops/_vmap/vmap_grad_nn_ops.py +19 -13
- mindspore/ops/_vmap/vmap_math_ops.py +11 -9
- mindspore/ops/_vmap/vmap_nn_ops.py +20 -34
- mindspore/ops/auto_generate/cpp_create_prim_instance_helper.py +149 -12
- mindspore/ops/auto_generate/gen_arg_handler.py +0 -61
- mindspore/ops/auto_generate/gen_extend_func.py +554 -60
- mindspore/ops/auto_generate/gen_ops_def.py +1621 -115
- mindspore/ops/auto_generate/gen_ops_prim.py +8024 -3409
- mindspore/ops/auto_generate/pyboost_inner_prim.py +183 -79
- mindspore/ops/composite/base.py +1 -1
- mindspore/ops/composite/multitype_ops/_compile_utils.py +229 -30
- mindspore/ops/composite/multitype_ops/pow_impl.py +0 -29
- mindspore/ops/function/__init__.py +12 -0
- mindspore/ops/function/array_func.py +561 -159
- mindspore/ops/function/clip_func.py +64 -0
- mindspore/ops/function/debug_func.py +28 -20
- mindspore/ops/function/image_func.py +1 -1
- mindspore/ops/function/linalg_func.py +5 -4
- mindspore/ops/function/math_func.py +1659 -290
- mindspore/ops/function/nn_func.py +988 -317
- mindspore/ops/function/parameter_func.py +3 -56
- mindspore/ops/function/random_func.py +243 -33
- mindspore/ops/function/sparse_unary_func.py +1 -1
- mindspore/ops/functional.py +18 -5
- mindspore/ops/functional_overload.py +897 -0
- mindspore/ops/operations/__init__.py +3 -2
- mindspore/ops/operations/_embedding_cache_ops.py +4 -4
- mindspore/ops/operations/_grad_ops.py +2 -34
- mindspore/ops/operations/_infer_ops.py +2 -1
- mindspore/ops/operations/_inner_ops.py +38 -8
- mindspore/ops/operations/array_ops.py +45 -303
- mindspore/ops/operations/comm_ops.py +19 -16
- mindspore/ops/operations/custom_ops.py +11 -55
- mindspore/ops/operations/debug_ops.py +42 -47
- mindspore/ops/operations/inner_ops.py +6 -4
- mindspore/ops/operations/linalg_ops.py +3 -2
- mindspore/ops/operations/manually_defined/ops_def.py +185 -104
- mindspore/ops/operations/math_ops.py +11 -216
- mindspore/ops/operations/nn_ops.py +146 -308
- mindspore/ops/primitive.py +23 -21
- mindspore/ops/tensor_method.py +1669 -0
- mindspore/ops_generate/aclnn_kernel_register_auto_cc_generator.py +110 -0
- mindspore/ops_generate/add_tensor_docs_generator.py +54 -0
- mindspore/ops_generate/arg_handler.py +0 -61
- mindspore/ops_generate/auto_grad_impl_cc_generator.py +135 -0
- mindspore/ops_generate/auto_grad_reg_cc_generator.py +93 -0
- mindspore/ops_generate/base_generator.py +11 -0
- mindspore/ops_generate/cpp_create_prim_instance_helper_generator.py +108 -0
- mindspore/ops_generate/functional_map_cpp_generator.py +491 -0
- mindspore/ops_generate/functional_overload_py_generator.py +110 -0
- mindspore/ops_generate/functions_cc_generator.py +233 -0
- mindspore/ops_generate/gen_aclnn_implement.py +110 -114
- mindspore/ops_generate/gen_constants.py +157 -3
- mindspore/ops_generate/gen_ops.py +245 -990
- mindspore/ops_generate/gen_pyboost_func.py +97 -998
- mindspore/ops_generate/gen_utils.py +119 -33
- mindspore/ops_generate/lite_ops_cpp_generator.py +155 -0
- mindspore/ops_generate/op_api_proto.py +206 -0
- mindspore/ops_generate/op_def_py_generator.py +131 -0
- mindspore/ops_generate/op_prim_py_generator.py +480 -0
- mindspore/ops_generate/op_proto.py +373 -108
- mindspore/ops_generate/op_template_parser.py +436 -0
- mindspore/ops_generate/ops_def_cc_generator.py +288 -0
- mindspore/ops_generate/ops_def_h_generator.py +74 -0
- mindspore/ops_generate/ops_name_h_generator.py +68 -0
- mindspore/ops_generate/ops_primitive_h_generator.py +81 -0
- mindspore/ops_generate/pyboost_functions_cpp_generator.py +370 -0
- mindspore/ops_generate/pyboost_functions_h_generator.py +68 -0
- mindspore/ops_generate/pyboost_functions_py_generator.py +148 -0
- mindspore/ops_generate/pyboost_grad_function_cpp_generator.py +154 -0
- mindspore/ops_generate/pyboost_inner_prim_generator.py +131 -0
- mindspore/ops_generate/pyboost_native_grad_functions_generator.py +268 -0
- mindspore/ops_generate/pyboost_op_cpp_code_generator.py +851 -0
- mindspore/ops_generate/pyboost_overload_functions_cpp_generator.py +344 -0
- mindspore/ops_generate/pyboost_utils.py +92 -33
- mindspore/ops_generate/template.py +294 -44
- mindspore/ops_generate/tensor_func_reg_cpp_generator.py +422 -0
- mindspore/parallel/__init__.py +3 -3
- mindspore/parallel/_auto_parallel_context.py +24 -33
- mindspore/parallel/_parallel_serialization.py +13 -2
- mindspore/parallel/_utils.py +4 -1
- mindspore/parallel/algo_parameter_config.py +1 -1
- mindspore/parallel/checkpoint_transform.py +44 -0
- mindspore/parallel/cluster/process_entity/_api.py +131 -37
- mindspore/parallel/cluster/process_entity/_utils.py +41 -6
- mindspore/parallel/cluster/run.py +20 -3
- mindspore/parallel/parameter_broadcast.py +1 -1
- mindspore/parallel/shard.py +3 -0
- mindspore/parallel/transform_safetensors.py +119 -253
- mindspore/profiler/__init__.py +17 -4
- mindspore/profiler/analysis/__init__.py +0 -0
- mindspore/profiler/analysis/parser/__init__.py +0 -0
- mindspore/profiler/analysis/parser/ascend_cann_parser.py +166 -0
- mindspore/profiler/analysis/parser/base_parser.py +158 -0
- mindspore/profiler/analysis/parser/framework_cann_relation_parser.py +45 -0
- mindspore/profiler/analysis/parser/ms_framework_parser.py +142 -0
- mindspore/profiler/analysis/parser/ms_minddata_parser.py +145 -0
- mindspore/profiler/analysis/parser/timeline_assembly_factory/__init__.py +0 -0
- mindspore/profiler/analysis/parser/timeline_assembly_factory/ascend_timeline_assembler.py +261 -0
- mindspore/profiler/analysis/parser/timeline_assembly_factory/base_timeline_assembler.py +40 -0
- mindspore/profiler/analysis/parser/timeline_assembly_factory/trace_view_container.py +84 -0
- mindspore/profiler/analysis/parser/timeline_creator/__init__.py +0 -0
- mindspore/profiler/analysis/parser/timeline_creator/base_timeline_creator.py +44 -0
- mindspore/profiler/analysis/parser/timeline_creator/cpu_op_timeline_creator.py +90 -0
- mindspore/profiler/analysis/parser/timeline_creator/fwk_timeline_creator.py +76 -0
- mindspore/profiler/analysis/parser/timeline_creator/msprof_timeline_creator.py +103 -0
- mindspore/profiler/analysis/parser/timeline_creator/scope_layer_timeline_creator.py +134 -0
- mindspore/profiler/analysis/parser/timeline_event/__init__.py +0 -0
- mindspore/profiler/analysis/parser/timeline_event/base_event.py +233 -0
- mindspore/profiler/analysis/parser/timeline_event/cpu_op_event.py +47 -0
- mindspore/profiler/analysis/parser/timeline_event/flow_event.py +36 -0
- mindspore/profiler/analysis/parser/timeline_event/fwk_event.py +260 -0
- mindspore/profiler/analysis/parser/timeline_event/msprof_event.py +73 -0
- mindspore/profiler/analysis/parser/timeline_event/scope_layer_event.py +53 -0
- mindspore/profiler/analysis/parser/timeline_event/timeline_event_pool.py +146 -0
- mindspore/profiler/analysis/task_manager.py +131 -0
- mindspore/profiler/analysis/time_converter.py +84 -0
- mindspore/profiler/analysis/viewer/__init__.py +0 -0
- mindspore/profiler/analysis/viewer/ascend_communication_viewer.py +333 -0
- mindspore/profiler/analysis/viewer/ascend_integrate_viewer.py +87 -0
- mindspore/profiler/analysis/viewer/ascend_kernel_details_viewer.py +252 -0
- mindspore/profiler/analysis/viewer/ascend_memory_viewer.py +313 -0
- mindspore/profiler/analysis/viewer/ascend_op_memory_viewer.py +322 -0
- mindspore/profiler/analysis/viewer/ascend_step_trace_time_viewer.py +265 -0
- mindspore/profiler/analysis/viewer/ascend_timeline_viewer.py +58 -0
- mindspore/profiler/analysis/viewer/base_viewer.py +26 -0
- mindspore/profiler/analysis/viewer/ms_dataset_viewer.py +97 -0
- mindspore/profiler/analysis/viewer/ms_minddata_viewer.py +581 -0
- mindspore/profiler/analysis/work_flow.py +73 -0
- mindspore/profiler/common/ascend_msprof_exporter.py +138 -0
- mindspore/profiler/common/command_executor.py +90 -0
- mindspore/profiler/common/constant.py +174 -3
- mindspore/profiler/common/file_manager.py +208 -0
- mindspore/profiler/common/log.py +130 -0
- mindspore/profiler/common/msprof_cmd_tool.py +202 -0
- mindspore/profiler/common/path_manager.py +371 -0
- mindspore/profiler/common/process_bar.py +168 -0
- mindspore/profiler/common/process_pool.py +9 -3
- mindspore/profiler/common/profiler_context.py +476 -0
- mindspore/profiler/common/profiler_info.py +304 -0
- mindspore/profiler/common/profiler_output_path.py +284 -0
- mindspore/profiler/common/profiler_parameters.py +210 -0
- mindspore/profiler/common/profiler_path_manager.py +120 -0
- mindspore/profiler/common/record_function.py +76 -0
- mindspore/profiler/common/tlv_decoder.py +76 -0
- mindspore/profiler/common/util.py +75 -2
- mindspore/profiler/dynamic_profiler.py +270 -37
- mindspore/profiler/envprofiler.py +138 -0
- mindspore/profiler/mstx.py +199 -0
- mindspore/profiler/platform/__init__.py +21 -0
- mindspore/profiler/platform/base_profiler.py +40 -0
- mindspore/profiler/platform/cpu_profiler.py +124 -0
- mindspore/profiler/platform/gpu_profiler.py +74 -0
- mindspore/profiler/platform/npu_profiler.py +309 -0
- mindspore/profiler/profiler.py +580 -93
- mindspore/profiler/profiler_action_controller.py +187 -0
- mindspore/profiler/profiler_interface.py +114 -0
- mindspore/profiler/schedule.py +208 -0
- mindspore/rewrite/api/symbol_tree.py +1 -2
- mindspore/run_check/_check_version.py +2 -6
- mindspore/runtime/__init__.py +37 -0
- mindspore/runtime/device.py +27 -0
- mindspore/runtime/event.py +209 -0
- mindspore/runtime/executor.py +148 -0
- mindspore/runtime/memory.py +392 -0
- mindspore/runtime/stream.py +460 -0
- mindspore/runtime/thread_bind_core.py +401 -0
- mindspore/train/__init__.py +2 -2
- mindspore/train/_utils.py +53 -18
- mindspore/train/amp.py +8 -4
- mindspore/train/callback/_checkpoint.py +32 -18
- mindspore/train/callback/_early_stop.py +1 -1
- mindspore/train/callback/_flops_collector.py +105 -69
- mindspore/train/callback/_history.py +1 -1
- mindspore/train/callback/_summary_collector.py +44 -6
- mindspore/train/callback/_tft_register.py +31 -10
- mindspore/train/dataset_helper.py +11 -11
- mindspore/train/metrics/precision.py +4 -5
- mindspore/train/mind_ir_pb2.py +167 -46
- mindspore/train/model.py +13 -15
- mindspore/train/serialization.py +462 -76
- mindspore/train/summary/summary_record.py +1 -2
- mindspore/train/train_thor/model_thor.py +1 -1
- mindspore/utils/__init__.py +4 -2
- mindspore/utils/bin/dataset-cache +0 -0
- mindspore/utils/bin/dataset-cache-server +0 -0
- mindspore/utils/dryrun.py +138 -0
- mindspore/utils/runtime_execution_order_check.py +550 -0
- mindspore/version.py +1 -1
- {mindspore-2.4.10.dist-info → mindspore-2.5.0.dist-info}/METADATA +2 -3
- {mindspore-2.4.10.dist-info → mindspore-2.5.0.dist-info}/RECORD +523 -457
- {mindspore-2.4.10.dist-info → mindspore-2.5.0.dist-info}/entry_points.txt +1 -1
- mindspore/_data_dump.cpython-39-aarch64-linux-gnu.so +0 -0
- mindspore/bin/cache_admin +0 -0
- mindspore/bin/cache_server +0 -0
- mindspore/common/_tensor_overload.py +0 -139
- mindspore/lib/libmindspore_np_dtype.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/acme/include/acme.h +0 -24
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/acme/include/acme_op.h +0 -82
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/acme/include/op_creator.h +0 -113
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/acme/include/op_param.h +0 -193
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/acme/src/core/dtype_registry.h +0 -90
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/acme/src/core/kernel_register.h +0 -46
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/acme/src/core/platform/platform_configs.h +0 -89
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/acme/src/core/platform/rt_funcs.h +0 -135
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/acme/src/ops/host_src/add_layer_norm_op.h +0 -60
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/acme/src/ops/host_src/add_rms_norm_op.h +0 -50
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/acme/src/ops/host_src/add_rms_norm_quant_op.h +0 -50
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/acme/src/ops/host_src/apply_rotary_pos_emb_nz_op.h +0 -42
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/acme/src/ops/host_src/apply_rotary_pos_emb_op.h +0 -55
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/acme/src/ops/host_src/asd_elewise_op.h +0 -34
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/acme/src/ops/host_src/asd_only_ops.h +0 -94
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/acme/src/ops/host_src/asd_op_base.h +0 -97
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/acme/src/ops/host_src/cast_op.h +0 -52
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/acme/src/ops/host_src/flash_attention_score_op.h +0 -97
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/acme/src/ops/host_src/gelu_op.h +0 -44
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/acme/src/ops/host_src/matmul_add_rmsnorm_op.h +0 -73
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/acme/src/ops/host_src/matmul_op.h +0 -108
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/acme/src/ops/host_src/multi_impls_op.h +0 -64
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/acme/src/ops/host_src/multi_weight_matmul_op.h +0 -91
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/acme/src/ops/host_src/paged_attention_op.h +0 -99
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/acme/src/ops/host_src/reshape_and_cache_nz_op.h +0 -44
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/acme/src/ops/host_src/reshape_and_cache_op.h +0 -44
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/acme/src/ops/host_src/rms_norm_op.h +0 -64
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/acme/src/utils/asd_utils.h +0 -179
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/acme/src/utils/comm_utils.h +0 -69
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/acme/src/utils/profiling_util.h +0 -366
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/add/add_impl.h +0 -56
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/add/kernel/add.h +0 -21
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/add/tiling/add_tiling.h +0 -43
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/apply_rotary_pos_emb/apply_rotary_pos_emb_impl.h +0 -46
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/apply_rotary_pos_emb/kernel/apply_rotary_pos_emb.h +0 -23
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/apply_rotary_pos_emb/kernel/apply_rotary_pos_emb_base.h +0 -456
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/apply_rotary_pos_emb/kernel/apply_rotary_pos_emb_bf16.h +0 -217
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/apply_rotary_pos_emb/kernel/apply_rotary_pos_emb_fp.h +0 -391
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/apply_rotary_pos_emb/kernel/apply_rotary_pos_emb_fp16.h +0 -126
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/apply_rotary_pos_emb/kernel/apply_rotary_pos_emb_fp32.h +0 -230
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/apply_rotary_pos_emb/kernel/apply_rotary_pos_emb_tiling.h +0 -43
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/apply_rotary_pos_emb/kernel/apply_rotary_pos_emb_value.h +0 -27
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/apply_rotary_pos_emb_nz/apply_rotary_pos_emb_nz_impl.h +0 -34
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/apply_rotary_pos_emb_nz/kernel/apply_rotary_pos_emb_nz.h +0 -23
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/apply_rotary_pos_emb_nz/kernel/apply_rotary_pos_emb_nz_base.h +0 -460
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/apply_rotary_pos_emb_nz/kernel/apply_rotary_pos_emb_nz_fp16.h +0 -116
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/apply_rotary_pos_emb_nz/kernel/apply_rotary_pos_emb_nz_fp32.h +0 -230
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/apply_rotary_pos_emb_nz/kernel/apply_rotary_pos_emb_nz_tiling.h +0 -43
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/apply_rotary_pos_emb_nz/kernel/apply_rotary_pos_emb_nz_value.h +0 -27
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/asdop/asd_op_impl.h +0 -74
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/backend_param.h +0 -74
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/cast/cast_impl.h +0 -48
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/cast/kernel/cast_kernel.h +0 -21
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/compare/compare_impl.h +0 -55
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/compare/compare_tiling.h +0 -27
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/compare/kernel/compare_kernel.h +0 -23
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_binary/and_impl.h +0 -29
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_binary/div_impl.h +0 -29
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_binary/elewise_binary_impl.h +0 -48
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_binary/elewise_binary_tiling.h +0 -25
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_binary/kernel/and_kernel.h +0 -46
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_binary/kernel/div_kernel.h +0 -46
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_binary/kernel/elewise_binary_base.h +0 -260
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_binary/kernel/elewise_binary_kernel.h +0 -35
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_binary/kernel/max_kernel.h +0 -66
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_binary/kernel/min_kernel.h +0 -66
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_binary/kernel/mul_kernel.h +0 -66
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_binary/kernel/or_kernel.h +0 -46
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_binary/max_impl.h +0 -29
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_binary/min_impl.h +0 -29
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_binary/mul_impl.h +0 -29
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_binary/or_impl.h +0 -29
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_unary/abs_impl.h +0 -29
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_unary/elewise_unary_impl.h +0 -47
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_unary/elewise_unary_tiling.h +0 -24
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_unary/exp_impl.h +0 -29
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_unary/kernel/abs_kernel.h +0 -45
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_unary/kernel/elewise_unary_base.h +0 -148
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_unary/kernel/elewise_unary_kernel.h +0 -31
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_unary/kernel/exp_kernel.h +0 -45
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_unary/kernel/ln_kernel.h +0 -45
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_unary/kernel/not_kernel.h +0 -45
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_unary/kernel/reciprocal_kernel.h +0 -45
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_unary/kernel/relu_kernel.h +0 -55
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_unary/kernel/rsqrt_kernel.h +0 -45
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_unary/kernel/sqrt_kernel.h +0 -45
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_unary/ln_impl.h +0 -29
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_unary/not_impl.h +0 -29
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_unary/reciprocal_impl.h +0 -29
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_unary/relu_impl.h +0 -29
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_unary/rsqrt_impl.h +0 -29
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/elewise_unary/sqrt_impl.h +0 -29
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/flash_attention_score/flash_attention_score_impl.h +0 -68
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/internal_kernel.h +0 -99
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/internal_rtbackend.h +0 -21
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/lccl/lccl_wrapper.h +0 -58
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/ms_int_types.h +0 -91
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/ms_int_utils.h +0 -108
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/paged_attention/paged_attention_impl.h +0 -64
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/param/add_param.h +0 -68
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/param/attention_param.h +0 -40
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/param/cast_param.h +0 -30
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/param/compare_param.h +0 -31
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/param/elewise_param.h +0 -41
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/param/grouped_matmul_param.h +0 -40
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/param/matmul_ext_param.h +0 -38
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/param/matmul_qkv_param.h +0 -42
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/param/sub_param.h +0 -33
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/profiling_util.h +0 -377
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/reshape_and_cache_nz/kernel/reshape_and_cache_nz.h +0 -24
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/reshape_and_cache_nz/reshape_and_cache_nz_impl.h +0 -42
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/reshape_and_cache_nz/reshape_and_cache_nz_tiling.h +0 -27
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/rms_norm/rms_norm_impl.h +0 -46
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/sub/kernel/sub_kernel.h +0 -20
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/sub/sub_impl.h +0 -48
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/sub/sub_tiling.h +0 -25
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/tune_repo/matmul_table.h +0 -399
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/tune_repo/utils.h +0 -41
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/utils/backend.h +0 -45
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/utils/elewise_tiling.h +0 -29
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/utils/elewise_utils.h +0 -30
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/utils/log/log.h +0 -69
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/utils/log/log_core.h +0 -43
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/utils/log/log_entity.h +0 -38
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/utils/log/log_sink.h +0 -69
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/utils/log/log_stream.h +0 -41
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/utils/log/log_tiling.h +0 -71
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/utils/log/log_utils.h +0 -165
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/utils/math.h +0 -20
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/utils/register/kernel_creator.h +0 -39
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/utils/register/kernel_registry.h +0 -121
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/include/utils/utils.h +0 -106
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libAdd_impl.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libSub_impl.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libadd_rms_norm_quant_acme_impl.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libapply_rotary_pos_emb_310p_old_impl.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libapply_rotary_pos_emb_old_impl.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libreshape_and_cache_nz_impl.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/lib/libreshape_and_cache_nz_old_impl.so +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/AcmeMatMulPostFusionMixTactic/acme_matmul_postfusion_mix.json +0 -19
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/AcmeMatMulPostFusionMixTactic/acme_matmul_postfusion_mix.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/AcmeMatMulPostFusionMixTactic/acme_matmul_postfusion_mix_mix_aic_0.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/AcmeMatMulPostFusionMixTactic/acme_matmul_postfusion_mix_mix_aiv_0.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/AcmeMultiWeightMatMulPostFusionMixTactic/acme_multi_weight_matmul_postfusion_mix.json +0 -19
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/AcmeMultiWeightMatMulPostFusionMixTactic/acme_multi_weight_matmul_postfusion_mix.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/AcmeMultiWeightMatMulPostFusionMixTactic/acme_multi_weight_matmul_postfusion_mix_mix_aic_0.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/AcmeMultiWeightMatMulPostFusionMixTactic/acme_multi_weight_matmul_postfusion_mix_mix_aiv_0.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/flash_attention_score/flash_attention_score_bf16_bnsd_full_mix.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/flash_attention_score/flash_attention_score_bf16_bnsd_tri_mix.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/flash_attention_score/flash_attention_score_bf16_bsh_full_mix.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/flash_attention_score/flash_attention_score_bf16_bsh_tri_mix.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/flash_attention_score/flash_attention_score_fp16_bnsd_full_mix.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/flash_attention_score/flash_attention_score_fp16_bnsd_tri_mix.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/flash_attention_score/flash_attention_score_fp16_bsh_full_mix.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/flash_attention_score/flash_attention_score_fp16_bsh_tri_mix.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/paged_attention/paged_attention_bf16_bnsd_mix.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/paged_attention/paged_attention_bf16_bsh_mix.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/paged_attention/paged_attention_fp16_bnsd_mix.o +0 -0
- mindspore/lib/plugin/ascend/ms_kernels_internal/internal_kernel/op_kernels/ascend910b/paged_attention/paged_attention_fp16_bsh_mix.o +0 -0
- mindspore/profiler/envprofiling.py +0 -254
- mindspore/profiler/profiling.py +0 -1926
- {mindspore-2.4.10.dist-info → mindspore-2.5.0.dist-info}/WHEEL +0 -0
- {mindspore-2.4.10.dist-info → mindspore-2.5.0.dist-info}/top_level.txt +0 -0
mindspore/nn/layer/pooling.py
CHANGED
|
@@ -18,23 +18,21 @@ from __future__ import absolute_import
|
|
|
18
18
|
from mindspore.ops import operations as P
|
|
19
19
|
from mindspore.ops import functional as F
|
|
20
20
|
import mindspore.ops as ops
|
|
21
|
-
from mindspore.ops.function.nn_func import avg_pool2d_ext
|
|
22
21
|
from mindspore._checkparam import _check_3d_int_or_tuple
|
|
23
22
|
from mindspore import _checkparam as validator
|
|
24
23
|
from mindspore.ops.primitive import constexpr, _primexpr
|
|
25
24
|
from mindspore.common.tensor import Tensor
|
|
26
25
|
import mindspore.context as context
|
|
27
26
|
from mindspore.common import dtype as mstype
|
|
28
|
-
from mindspore.ops.operations.nn_ops import AdaptiveMaxPool2D
|
|
29
|
-
from mindspore.ops.operations.nn_ops import AdaptiveMaxPool3D, AdaptiveAvgPool3D
|
|
30
|
-
from mindspore.ops.auto_generate.gen_ops_prim import MaxPoolWithIndices, MaxPoolWithMask
|
|
31
27
|
from mindspore.nn.cell import Cell
|
|
32
28
|
from mindspore._c_expression import MSContext
|
|
29
|
+
from mindspore.ops.auto_generate import avg_pool1d_ext
|
|
30
|
+
|
|
33
31
|
|
|
34
32
|
__all__ = ['AvgPool3d', 'MaxPool3d', 'AvgPool2d', 'MaxPool2d', 'AvgPool1d', 'MaxPool1d', 'FractionalMaxPool2d',
|
|
35
33
|
'FractionalMaxPool3d', 'AdaptiveAvgPool1d', 'AdaptiveMaxPool1d', 'AdaptiveMaxPool2d', 'AdaptiveMaxPool3d',
|
|
36
34
|
'AdaptiveAvgPool2d', 'AdaptiveAvgPool3d', 'MaxUnpool1d', 'MaxUnpool2d', 'MaxUnpool3d', 'LPPool1d',
|
|
37
|
-
'LPPool2d', 'AvgPool2dExt', 'MaxPool2dExt']
|
|
35
|
+
'LPPool2d', 'AvgPool2dExt', 'MaxPool2dExt', 'AvgPool1dExt']
|
|
38
36
|
|
|
39
37
|
|
|
40
38
|
class _PoolNd(Cell):
|
|
@@ -689,9 +687,11 @@ class MaxPool2dExt(Cell):
|
|
|
689
687
|
self.return_indices = return_indices
|
|
690
688
|
strides = stride if (stride is not None) else kernel_size
|
|
691
689
|
if return_indices:
|
|
692
|
-
self.max_pool_func_ = MaxPoolWithIndices(kernel_size, strides, padding,
|
|
690
|
+
self.max_pool_func_ = ops.auto_generate.gen_ops_prim.MaxPoolWithIndices(kernel_size, strides, padding,
|
|
691
|
+
dilation, ceil_mode)
|
|
693
692
|
else:
|
|
694
|
-
self.max_pool_func_ = MaxPoolWithMask(kernel_size, strides, padding,
|
|
693
|
+
self.max_pool_func_ = ops.auto_generate.gen_ops_prim.MaxPoolWithMask(kernel_size, strides, padding,
|
|
694
|
+
dilation, ceil_mode)
|
|
695
695
|
|
|
696
696
|
def construct(self, input):
|
|
697
697
|
out, indices = self.max_pool_func_(input)
|
|
@@ -1021,6 +1021,40 @@ class AvgPool3d(_PoolNd):
|
|
|
1021
1021
|
return out
|
|
1022
1022
|
|
|
1023
1023
|
|
|
1024
|
+
class AvgPool1dExt(Cell):
|
|
1025
|
+
r"""
|
|
1026
|
+
Applies a 1D average pooling over an input Tensor which can be regarded as
|
|
1027
|
+
a composition of 2D input planes.
|
|
1028
|
+
|
|
1029
|
+
For details, please refer to :func:`mindspore.mint.nn.functional.avg_pool1d`.
|
|
1030
|
+
|
|
1031
|
+
Supported Platforms:
|
|
1032
|
+
``Ascend``
|
|
1033
|
+
|
|
1034
|
+
Examples:
|
|
1035
|
+
>>> import numpy as np
|
|
1036
|
+
>>> from mindspore import Tensor, nn
|
|
1037
|
+
>>> from mindspore import dtype as mstype
|
|
1038
|
+
>>> input = Tensor(np.arange(1 * 3 * 4).reshape(1, 3, 4), mstype.float32)
|
|
1039
|
+
>>> net = nn.AvgPool1dExt(kernel_size=2, stride=1)
|
|
1040
|
+
>>> output = net(input)
|
|
1041
|
+
>>> print(output.shape)
|
|
1042
|
+
(1, 3, 3)
|
|
1043
|
+
"""
|
|
1044
|
+
def __init__(self, kernel_size, stride=None, padding=0, ceil_mode=False,
|
|
1045
|
+
count_include_pad=True):
|
|
1046
|
+
super().__init__()
|
|
1047
|
+
self.kernel_size = kernel_size
|
|
1048
|
+
self.stride = stride
|
|
1049
|
+
self.padding = padding
|
|
1050
|
+
self.ceil_mode = ceil_mode
|
|
1051
|
+
self.count_include_pad = count_include_pad
|
|
1052
|
+
|
|
1053
|
+
def construct(self, input):
|
|
1054
|
+
return avg_pool1d_ext(input, self.kernel_size, self.stride, self.padding,
|
|
1055
|
+
self.ceil_mode, self.count_include_pad)
|
|
1056
|
+
|
|
1057
|
+
|
|
1024
1058
|
class AvgPool2dExt(Cell):
|
|
1025
1059
|
r"""
|
|
1026
1060
|
Applies a 2D average pooling over an input Tensor which can be regarded as
|
|
@@ -1052,8 +1086,8 @@ class AvgPool2dExt(Cell):
|
|
|
1052
1086
|
self.divisor_override = divisor_override
|
|
1053
1087
|
|
|
1054
1088
|
def construct(self, input):
|
|
1055
|
-
return avg_pool2d_ext(input, self.kernel_size, self.stride, self.padding,
|
|
1056
|
-
|
|
1089
|
+
return ops.function.nn_func.avg_pool2d_ext(input, self.kernel_size, self.stride, self.padding,
|
|
1090
|
+
self.ceil_mode, self.count_include_pad, self.divisor_override)
|
|
1057
1091
|
|
|
1058
1092
|
|
|
1059
1093
|
class AvgPool2d(_PoolNd):
|
|
@@ -1127,7 +1161,7 @@ class AvgPool2d(_PoolNd):
|
|
|
1127
1161
|
TypeError: If `kernel_size` or `strides` is neither int nor tuple.
|
|
1128
1162
|
ValueError: If `pad_mode` is not ``"valid"`` , ``"same"`` or ``"pad"`` with not case sensitive.
|
|
1129
1163
|
ValueError: If `data_format` is neither ``'NCHW'`` nor ``'NHWC'``.
|
|
1130
|
-
ValueError: If `padding`, `ceil_mode`, `count_include_pad`, or `divisor_override` is used
|
|
1164
|
+
ValueError: If `padding`, `ceil_mode`, `count_include_pad`, or `divisor_override` is used,
|
|
1131
1165
|
or `pad_mode` is ``"pad"`` when `data_format` is 'NHWC'.
|
|
1132
1166
|
ValueError: If `kernel_size` or `strides` is less than 1.
|
|
1133
1167
|
ValueError: If length of `padding` tuple/list is not 1 or 2.
|
|
@@ -1592,7 +1626,7 @@ class AdaptiveAvgPool3d(Cell):
|
|
|
1592
1626
|
def __init__(self, output_size):
|
|
1593
1627
|
"""Initialize AdaptiveAvgPool3d."""
|
|
1594
1628
|
super(AdaptiveAvgPool3d, self).__init__()
|
|
1595
|
-
self.adaptive_avg_pool3d = AdaptiveAvgPool3D(output_size)
|
|
1629
|
+
self.adaptive_avg_pool3d = ops.AdaptiveAvgPool3D(output_size)
|
|
1596
1630
|
|
|
1597
1631
|
def construct(self, input):
|
|
1598
1632
|
return self.adaptive_avg_pool3d(input)
|
|
@@ -1764,7 +1798,7 @@ class AdaptiveMaxPool2d(Cell):
|
|
|
1764
1798
|
"""Initialize AdaptiveMaxPool2d."""
|
|
1765
1799
|
super(AdaptiveMaxPool2d, self).__init__()
|
|
1766
1800
|
validator.check_value_type('return_indices', return_indices, [bool], self.cls_name)
|
|
1767
|
-
self.adaptive_max_pool2d = AdaptiveMaxPool2D(output_size)
|
|
1801
|
+
self.adaptive_max_pool2d = ops.AdaptiveMaxPool2D(output_size)
|
|
1768
1802
|
self.return_indices = return_indices
|
|
1769
1803
|
|
|
1770
1804
|
def construct(self, input):
|
|
@@ -1823,7 +1857,7 @@ class AdaptiveMaxPool3d(Cell):
|
|
|
1823
1857
|
output_size = (output_size, output_size, output_size)
|
|
1824
1858
|
self.output_size = Tensor(output_size, dtype=mstype.int32)
|
|
1825
1859
|
self.return_indices = return_indices
|
|
1826
|
-
self.adaptive_max_pool3d = AdaptiveMaxPool3D()
|
|
1860
|
+
self.adaptive_max_pool3d = ops.AdaptiveMaxPool3D()
|
|
1827
1861
|
|
|
1828
1862
|
def construct(self, input):
|
|
1829
1863
|
output = self.adaptive_max_pool3d(input, self.output_size)
|
mindspore/nn/layer/rnn_cells.py
CHANGED
|
@@ -340,7 +340,7 @@ class GRUCell(RNNCellBase):
|
|
|
340
340
|
:math:`r` is reset gate. :math:`z` is update gate. :math:`n` is n-th layer. For instance,
|
|
341
341
|
:math:`W_{ir}, b_{ir}` are the weight and bias used to transform from input :math:`x` to :math:`r`.
|
|
342
342
|
Details can be found in paper
|
|
343
|
-
`Learning Phrase Representations using RNN Encoder
|
|
343
|
+
`Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation
|
|
344
344
|
<https://aclanthology.org/D14-1179.pdf>`_.
|
|
345
345
|
|
|
346
346
|
Args:
|
mindspore/nn/layer/rnns.py
CHANGED
|
@@ -237,6 +237,7 @@ class _DynamicGRUCPUGPU(Cell):
|
|
|
237
237
|
h_0.view(1, *h_0.shape),
|
|
238
238
|
weights.astype(x.dtype)
|
|
239
239
|
)
|
|
240
|
+
|
|
240
241
|
if seq_length is not None:
|
|
241
242
|
h_n = get_hidden(output, seq_length)
|
|
242
243
|
mask = sequence_mask(seq_length, x.shape[0])
|
|
@@ -687,7 +688,7 @@ class GRU(_RNNBase):
|
|
|
687
688
|
are learnable weights between the output and the input in the formula. For instance,
|
|
688
689
|
:math:`W_{ir}, b_{ir}` are the weight and bias used to transform from input :math:`x` to :math:`r`.
|
|
689
690
|
Details can be found in paper
|
|
690
|
-
`Learning Phrase Representations using RNN Encoder
|
|
691
|
+
`Learning Phrase Representations using RNN Encoder-Decoder for Statistical Machine Translation
|
|
691
692
|
<https://aclanthology.org/D14-1179.pdf>`_.
|
|
692
693
|
|
|
693
694
|
Note:
|
|
@@ -15,8 +15,8 @@
|
|
|
15
15
|
"""Time Distributed."""
|
|
16
16
|
from __future__ import absolute_import
|
|
17
17
|
|
|
18
|
+
from mindspore import ops
|
|
18
19
|
from mindspore.ops.primitive import constexpr, Primitive, _primexpr
|
|
19
|
-
from mindspore.ops import Reshape, Transpose, Stack, Unstack
|
|
20
20
|
from mindspore.common import Tensor
|
|
21
21
|
from mindspore import _checkparam as Validator
|
|
22
22
|
from mindspore.nn.cell import Cell
|
|
@@ -116,8 +116,8 @@ class TimeDistributed(Cell):
|
|
|
116
116
|
self.layer = layer
|
|
117
117
|
self.time_axis = time_axis
|
|
118
118
|
self.reshape_with_axis = reshape_with_axis
|
|
119
|
-
self.transpose = Transpose()
|
|
120
|
-
self.reshape = Reshape()
|
|
119
|
+
self.transpose = ops.Transpose()
|
|
120
|
+
self.reshape = ops.Reshape()
|
|
121
121
|
|
|
122
122
|
def construct(self, inputs):
|
|
123
123
|
_check_data(isinstance(inputs, Tensor), self.cls_name)
|
|
@@ -143,7 +143,7 @@ class TimeDistributed(Cell):
|
|
|
143
143
|
outputs_shape_new = (-1,) + outputs_shape_new[1:]
|
|
144
144
|
return self.reshape(outputs, outputs_shape_new)
|
|
145
145
|
|
|
146
|
-
unstack = Unstack(time_axis)
|
|
146
|
+
unstack = ops.Unstack(time_axis)
|
|
147
147
|
inputs = unstack(inputs)
|
|
148
148
|
y = ()
|
|
149
149
|
for item in inputs:
|
|
@@ -151,5 +151,5 @@ class TimeDistributed(Cell):
|
|
|
151
151
|
_check_data(isinstance(outputs, Tensor), self.cls_name)
|
|
152
152
|
_check_expand_dims_axis(time_axis, outputs.ndim, self.cls_name)
|
|
153
153
|
y += (outputs,)
|
|
154
|
-
y = Stack(time_axis)(y)
|
|
154
|
+
y = ops.Stack(time_axis)(y)
|
|
155
155
|
return y
|
|
@@ -26,12 +26,12 @@ from mindspore.common.tensor import Tensor
|
|
|
26
26
|
from mindspore.common.parameter import Parameter
|
|
27
27
|
from mindspore.common.initializer import initializer, XavierNormal, XavierUniform, \
|
|
28
28
|
HeUniform, Uniform, _calculate_fan_in_and_fan_out
|
|
29
|
-
from mindspore.ops.function.nn_func import multi_head_attention_forward
|
|
30
29
|
from mindspore.nn.cell import Cell
|
|
31
30
|
from .basic import Dense, Dropout
|
|
32
31
|
from .activation import ReLU, GELU
|
|
33
32
|
from .normalization import LayerNorm
|
|
34
33
|
from .container import CellList
|
|
34
|
+
|
|
35
35
|
__all__ = ['MultiheadAttention', 'TransformerEncoderLayer', 'TransformerDecoderLayer',
|
|
36
36
|
'TransformerEncoder', 'TransformerDecoder', 'Transformer']
|
|
37
37
|
|
|
@@ -212,7 +212,7 @@ class MultiheadAttention(Cell):
|
|
|
212
212
|
query, key, value = [x.swapaxes(1, 0) for x in (query, key, value)]
|
|
213
213
|
|
|
214
214
|
if not self._qkv_same_embed_dim:
|
|
215
|
-
attn_output, attn_output_weights = multi_head_attention_forward(
|
|
215
|
+
attn_output, attn_output_weights = ops.function.nn_func.multi_head_attention_forward(
|
|
216
216
|
query, key, value, self.embed_dim, self.num_heads,
|
|
217
217
|
self.in_proj_weight, self.in_proj_bias,
|
|
218
218
|
self.bias_k, self.bias_v, self.add_zero_attn,
|
|
@@ -224,7 +224,7 @@ class MultiheadAttention(Cell):
|
|
|
224
224
|
v_proj_weight=self.v_proj_weight, average_attn_weights=average_attn_weights,
|
|
225
225
|
k_is_v=self.k_is_v, q_is_k=self.q_is_k, dtype=self.dtype)
|
|
226
226
|
else:
|
|
227
|
-
attn_output, attn_output_weights = multi_head_attention_forward(
|
|
227
|
+
attn_output, attn_output_weights = ops.function.nn_func.multi_head_attention_forward(
|
|
228
228
|
query, key, value, self.embed_dim, self.num_heads,
|
|
229
229
|
self.in_proj_weight, self.in_proj_bias,
|
|
230
230
|
self.bias_k, self.bias_v, self.add_zero_attn,
|
|
@@ -328,7 +328,7 @@ class TransformerEncoderLayer(Cell):
|
|
|
328
328
|
self.activation1 = activation
|
|
329
329
|
|
|
330
330
|
if not isinstance(activation, str) and not isinstance(activation, Cell) \
|
|
331
|
-
|
|
331
|
+
and not callable(activation):
|
|
332
332
|
raise ValueError(f"The argument 'activation' must be str, callable or Cell instance,"
|
|
333
333
|
f" but get {activation}.")
|
|
334
334
|
if isinstance(activation, Cell) and (not isinstance(activation, ReLU) and \
|
|
@@ -360,15 +360,23 @@ class TransformerEncoderLayer(Cell):
|
|
|
360
360
|
raise AssertionError(
|
|
361
361
|
"only bool and floating types of key_padding_mask are supported")
|
|
362
362
|
|
|
363
|
-
|
|
363
|
+
input_data = src
|
|
364
|
+
|
|
364
365
|
if self.norm_first:
|
|
365
|
-
|
|
366
|
-
|
|
366
|
+
normed_input = self.norm1(input_data)
|
|
367
|
+
sa_block_result = self._sa_block(normed_input, src_mask, src_key_padding_mask)
|
|
368
|
+
input_data = input_data + sa_block_result
|
|
369
|
+
normed_updated_input = self.norm2(input_data)
|
|
370
|
+
ff_block_result = self._ff_block(normed_updated_input)
|
|
371
|
+
input_data = input_data + ff_block_result
|
|
367
372
|
else:
|
|
368
|
-
|
|
369
|
-
|
|
373
|
+
sa_block_result = self._sa_block(input_data, src_mask, src_key_padding_mask)
|
|
374
|
+
normed_sa_result = self.norm1(input_data + sa_block_result)
|
|
375
|
+
input_data = normed_sa_result
|
|
376
|
+
ff_block_result = self._ff_block(input_data)
|
|
377
|
+
input_data = self.norm2(input_data + ff_block_result)
|
|
370
378
|
|
|
371
|
-
return
|
|
379
|
+
return input_data
|
|
372
380
|
|
|
373
381
|
def _sa_block(self, x, attn_mask, key_padding_mask):
|
|
374
382
|
x = self.self_attn(x, x, x,
|
|
@@ -480,7 +488,7 @@ class TransformerDecoderLayer(Cell):
|
|
|
480
488
|
self.activation1 = activation
|
|
481
489
|
|
|
482
490
|
if not isinstance(activation, str) and not isinstance(activation, Cell) \
|
|
483
|
-
|
|
491
|
+
and not callable(activation):
|
|
484
492
|
raise ValueError(f"The argument 'activation' must be str, callable or Cell instance,"
|
|
485
493
|
f" but get {activation}.")
|
|
486
494
|
if isinstance(activation, Cell) and (not isinstance(activation, ReLU) and \
|
|
@@ -507,17 +515,29 @@ class TransformerDecoderLayer(Cell):
|
|
|
507
515
|
def construct(self, tgt: Tensor, memory: Tensor, tgt_mask: Optional[Tensor] = None,
|
|
508
516
|
memory_mask: Optional[Tensor] = None, tgt_key_padding_mask: Optional[Tensor] = None,
|
|
509
517
|
memory_key_padding_mask: Optional[Tensor] = None):
|
|
510
|
-
|
|
518
|
+
input_data = tgt
|
|
519
|
+
|
|
511
520
|
if self.norm_first:
|
|
512
|
-
|
|
513
|
-
|
|
514
|
-
|
|
521
|
+
normed_input = self.norm1(input_data)
|
|
522
|
+
sa_block_result = self._sa_block(normed_input, tgt_mask, tgt_key_padding_mask)
|
|
523
|
+
input_data = input_data + sa_block_result
|
|
524
|
+
normed_updated_input_1 = self.norm2(input_data)
|
|
525
|
+
mha_block_result = self._mha_block(normed_updated_input_1, memory, memory_mask, memory_key_padding_mask)
|
|
526
|
+
input_data = input_data + mha_block_result
|
|
527
|
+
normed_updated_input_2 = self.norm3(input_data)
|
|
528
|
+
ff_block_result = self._ff_block(normed_updated_input_2)
|
|
529
|
+
input_data = input_data + ff_block_result
|
|
515
530
|
else:
|
|
516
|
-
|
|
517
|
-
|
|
518
|
-
|
|
531
|
+
sa_block_result = self._sa_block(input_data, tgt_mask, tgt_key_padding_mask)
|
|
532
|
+
normed_sa_result = self.norm1(input_data + sa_block_result)
|
|
533
|
+
input_data = normed_sa_result
|
|
534
|
+
mha_block_result = self._mha_block(input_data, memory, memory_mask, memory_key_padding_mask)
|
|
535
|
+
normed_mha_result = self.norm2(input_data + mha_block_result)
|
|
536
|
+
input_data = normed_mha_result
|
|
537
|
+
ff_block_result = self._ff_block(input_data)
|
|
538
|
+
input_data = self.norm3(input_data + ff_block_result)
|
|
519
539
|
|
|
520
|
-
return
|
|
540
|
+
return input_data
|
|
521
541
|
|
|
522
542
|
def _sa_block(self, x, attn_mask, key_padding_mask):
|
|
523
543
|
x = self.self_attn(x, x, x,
|
|
@@ -670,17 +690,19 @@ class TransformerDecoder(Cell):
|
|
|
670
690
|
def construct(self, tgt: Tensor, memory: Tensor, tgt_mask: Optional[Tensor] = None,
|
|
671
691
|
memory_mask: Optional[Tensor] = None, tgt_key_padding_mask: Optional[Tensor] = None,
|
|
672
692
|
memory_key_padding_mask: Optional[Tensor] = None):
|
|
673
|
-
|
|
693
|
+
processed_output = tgt
|
|
674
694
|
for mod in self.layers:
|
|
675
|
-
|
|
676
|
-
|
|
677
|
-
|
|
678
|
-
|
|
695
|
+
layer_output = mod(processed_output, memory,
|
|
696
|
+
tgt_mask=tgt_mask,
|
|
697
|
+
memory_mask=memory_mask,
|
|
698
|
+
tgt_key_padding_mask=tgt_key_padding_mask,
|
|
699
|
+
memory_key_padding_mask=memory_key_padding_mask)
|
|
700
|
+
processed_output = layer_output
|
|
679
701
|
|
|
680
702
|
if self.norm is not None:
|
|
681
|
-
|
|
703
|
+
processed_output = self.norm(processed_output)
|
|
682
704
|
|
|
683
|
-
return
|
|
705
|
+
return processed_output
|
|
684
706
|
|
|
685
707
|
|
|
686
708
|
class Transformer(Cell):
|
|
@@ -223,7 +223,8 @@ class InverseDecayLR(LearningRateSchedule):
|
|
|
223
223
|
learning_rate (float): The initial value of learning rate.
|
|
224
224
|
decay_rate (float): The decay rate.
|
|
225
225
|
decay_steps (int): Number of steps to decay over.
|
|
226
|
-
is_stair (bool): If true, learning rate decay once every `decay_steps` times.
|
|
226
|
+
is_stair (bool): If true, learning rate decay once every `decay_steps` times. If False, the learning rate
|
|
227
|
+
decays for every step. Default: ``False`` .
|
|
227
228
|
|
|
228
229
|
Inputs:
|
|
229
230
|
- **global_step** (Tensor) - The current step number.
|
|
@@ -454,8 +455,9 @@ class WarmUpLR(LearningRateSchedule):
|
|
|
454
455
|
tmp\_step= \min(current\_step, warmup\_steps)
|
|
455
456
|
|
|
456
457
|
Args:
|
|
457
|
-
learning_rate (float): The initial value of learning rate.
|
|
458
|
-
warmup_steps (int): The warm up steps of learning rate.
|
|
458
|
+
learning_rate (float): The initial value of learning rate. The value of `learning_rate` must be greater than 0.
|
|
459
|
+
warmup_steps (int): The warm up steps of learning rate. The value of `warmup_steps` must be greater than
|
|
460
|
+
or equal to 1.
|
|
459
461
|
|
|
460
462
|
Inputs:
|
|
461
463
|
- **global_step** (Tensor) - The current step number. Shape is :math:`()`.
|
mindspore/nn/loss/loss.py
CHANGED
|
@@ -24,8 +24,6 @@ from mindspore.common.tensor import Tensor
|
|
|
24
24
|
from mindspore.common.parameter import Parameter
|
|
25
25
|
from mindspore.ops import operations as P
|
|
26
26
|
from mindspore.ops.operations import _inner_ops as inner
|
|
27
|
-
from mindspore.ops.operations.nn_ops import MultiMarginLoss as MultiMarginLossOp
|
|
28
|
-
from mindspore.ops.operations.nn_ops import MultilabelMarginLoss as MultilabelMarginLossOp
|
|
29
27
|
from mindspore.ops import functional as F
|
|
30
28
|
from mindspore import nn
|
|
31
29
|
from mindspore.ops.primitive import constexpr, _primexpr
|
|
@@ -33,7 +31,6 @@ from mindspore.nn.cell import Cell
|
|
|
33
31
|
from mindspore.nn.layer.activation import get_activation
|
|
34
32
|
from mindspore import _checkparam as validator
|
|
35
33
|
from mindspore import context
|
|
36
|
-
from mindspore.ops.auto_generate import l1_loss_ext_op
|
|
37
34
|
|
|
38
35
|
|
|
39
36
|
class LossBase(Cell):
|
|
@@ -319,7 +316,7 @@ class L1LossExt(LossBase):
|
|
|
319
316
|
self.reduction = reduction
|
|
320
317
|
|
|
321
318
|
def construct(self, logits, labels):
|
|
322
|
-
return l1_loss_ext_op(logits, labels, self.reduction)
|
|
319
|
+
return ops.auto_generate.l1_loss_ext_op(logits, labels, self.reduction)
|
|
323
320
|
|
|
324
321
|
|
|
325
322
|
class MSELoss(LossBase):
|
|
@@ -628,11 +625,11 @@ class SmoothL1Loss(LossBase):
|
|
|
628
625
|
.. math::
|
|
629
626
|
L_{i} =
|
|
630
627
|
\begin{cases}
|
|
631
|
-
\frac{0.5 (x_i - y_i)^{2}}{\beta}, & \text{if } |x_i - y_i| < {
|
|
632
|
-
|x_i - y_i| - 0.5 {\beta}, & \text{otherwise.}
|
|
628
|
+
\frac{0.5 (x_i - y_i)^{2}}{\text{beta}}, & \text{if } |x_i - y_i| < \text{beta} \\
|
|
629
|
+
|x_i - y_i| - 0.5 * {\text{beta}}, & \text{otherwise.}
|
|
633
630
|
\end{cases}
|
|
634
631
|
|
|
635
|
-
Where :math:`{\beta}` represents the threshold `beta`.
|
|
632
|
+
Where :math:`{\text{beta}}` represents the threshold `beta`.
|
|
636
633
|
|
|
637
634
|
If `reduction` is not `none`, then:
|
|
638
635
|
|
|
@@ -653,8 +650,11 @@ class SmoothL1Loss(LossBase):
|
|
|
653
650
|
robust to outliers, and the loss function has better robustness.
|
|
654
651
|
|
|
655
652
|
Args:
|
|
656
|
-
beta (
|
|
657
|
-
Default: ``1.0`` .
|
|
653
|
+
beta (number, optional): The loss function calculates the threshold of the transformation
|
|
654
|
+
between L1Loss and L2Loss. Default: ``1.0`` .
|
|
655
|
+
|
|
656
|
+
- Ascend: The value should be equal to or greater than zero.
|
|
657
|
+
- CPU/GPU: The value should be greater than zero.
|
|
658
658
|
reduction (str, optional): Apply specific reduction method to the output: ``'none'`` , ``'mean'`` ,
|
|
659
659
|
``'sum'`` . Default: ``'none'`` .
|
|
660
660
|
|
|
@@ -663,22 +663,26 @@ class SmoothL1Loss(LossBase):
|
|
|
663
663
|
- ``'sum'``: the output elements will be summed.
|
|
664
664
|
|
|
665
665
|
Inputs:
|
|
666
|
-
- **logits** (Tensor) - Predictive value. Tensor of any dimension.
|
|
667
|
-
float32.
|
|
668
|
-
- **labels** (Tensor) - Ground truth data, same shape and dtype as the `logits`.
|
|
666
|
+
- **logits** (Tensor) - Predictive value. Tensor of any dimension. Supported dtypes:
|
|
669
667
|
|
|
668
|
+
- Ascend: float16, float32, bfloat16.
|
|
669
|
+
- CPU/GPU: float16, float32, float64.
|
|
670
|
+
- **labels** (Tensor) - Ground truth data.
|
|
671
|
+
|
|
672
|
+
- CPU/Ascend: has the same shape as the `logits`,
|
|
673
|
+
`logits` and `labels` comply with the implicit type conversion rules to make the data types consistent.
|
|
674
|
+
- GPU: has the same shape and dtype as the `logits`.
|
|
670
675
|
Outputs:
|
|
671
676
|
Tensor, if `reduction` is ``'none'``, then output is a tensor with the same shape as `logits`.
|
|
672
677
|
Otherwise the shape of output tensor is :math:`()`.
|
|
673
678
|
|
|
674
679
|
Raises:
|
|
675
|
-
TypeError: If `
|
|
676
|
-
|
|
677
|
-
TypeError: If `logits` or `labels` are not Tensor.
|
|
678
|
-
TypeError: If dtype of `logits` or `labels` is neither float16 not float32.
|
|
679
|
-
TypeError: If dtype of `logits` is not the same as `labels`.
|
|
680
|
-
ValueError: If `beta` is less than or equal to 0.
|
|
680
|
+
TypeError: If input `logits` or `labels` are not Tensor.
|
|
681
|
+
RuntimeError: If dtype of `logits` or `labels` is not one of float16, float32, float64, bfloat16.
|
|
681
682
|
ValueError: If shape of `logits` is not the same as `labels`.
|
|
683
|
+
ValueError: If `reduction` is not one of ``'none'``, ``'mean'``, ``'sum'``.
|
|
684
|
+
TypeError: If `beta` is not a float, int or bool.
|
|
685
|
+
RuntimeError: If `beta` is less than or equal to 0.
|
|
682
686
|
|
|
683
687
|
Supported Platforms:
|
|
684
688
|
``Ascend`` ``GPU`` ``CPU``
|
|
@@ -1631,7 +1635,7 @@ class MultiMarginLoss(LossBase):
|
|
|
1631
1635
|
def __init__(self, p=1, margin=1.0, reduction='mean', weight=None):
|
|
1632
1636
|
"""Initialize MultiMarginLoss."""
|
|
1633
1637
|
super(MultiMarginLoss, self).__init__()
|
|
1634
|
-
self.multi_margin_loss =
|
|
1638
|
+
self.multi_margin_loss = ops.MultiMarginLoss(p=p, margin=margin, reduction=reduction)
|
|
1635
1639
|
self.weight = weight
|
|
1636
1640
|
|
|
1637
1641
|
def construct(self, x, target, weight=None):
|
|
@@ -1718,22 +1722,11 @@ class BCELoss(LossBase):
|
|
|
1718
1722
|
def __init__(self, weight=None, reduction='mean'):
|
|
1719
1723
|
"""Initialize BCELoss."""
|
|
1720
1724
|
super(BCELoss, self).__init__(reduction)
|
|
1721
|
-
self.
|
|
1722
|
-
self.
|
|
1723
|
-
if not self.weight_one:
|
|
1724
|
-
self.weight = weight
|
|
1725
|
-
else:
|
|
1726
|
-
self.ones = P.OnesLike()
|
|
1725
|
+
self.reduction = reduction
|
|
1726
|
+
self.weight = weight
|
|
1727
1727
|
|
|
1728
1728
|
def construct(self, logits, labels):
|
|
1729
|
-
|
|
1730
|
-
_check_is_tensor('labels', labels, self.cls_name)
|
|
1731
|
-
if self.weight_one:
|
|
1732
|
-
weight = self.ones(logits)
|
|
1733
|
-
else:
|
|
1734
|
-
weight = self.weight
|
|
1735
|
-
loss = self.binary_cross_entropy(logits, labels, weight)
|
|
1736
|
-
return loss
|
|
1729
|
+
return F.binary_cross_entropy(logits, labels, self.weight, self.reduction)
|
|
1737
1730
|
|
|
1738
1731
|
|
|
1739
1732
|
class CosineEmbeddingLoss(LossBase):
|
|
@@ -1887,7 +1880,7 @@ class MultilabelMarginLoss(LossBase):
|
|
|
1887
1880
|
|
|
1888
1881
|
def __init__(self, reduction='mean'):
|
|
1889
1882
|
super(MultilabelMarginLoss, self).__init__()
|
|
1890
|
-
self.multilabel_margin_loss =
|
|
1883
|
+
self.multilabel_margin_loss = ops.MultilabelMarginLoss(reduction=reduction)
|
|
1891
1884
|
|
|
1892
1885
|
def construct(self, x, target):
|
|
1893
1886
|
loss, _ = self.multilabel_margin_loss(x, target)
|
|
@@ -2265,7 +2258,8 @@ class TripletMarginLoss(LossBase):
|
|
|
2265
2258
|
- ``'mean'``: compute and return the mean of elements in the output.
|
|
2266
2259
|
- ``'sum'``: the output elements will be summed.
|
|
2267
2260
|
|
|
2268
|
-
margin (Union[Tensor, float]): Make a margin between the positive pair and the negative pair.
|
|
2261
|
+
margin (Union[Tensor, float]): Make a margin between the positive pair and the negative pair. The length of
|
|
2262
|
+
shape of `margin` must be 0.
|
|
2269
2263
|
Default: ``1.0`` .
|
|
2270
2264
|
|
|
2271
2265
|
Inputs:
|
|
@@ -2275,7 +2269,8 @@ class TripletMarginLoss(LossBase):
|
|
|
2275
2269
|
shape as `x`. :math:`p` in the above formula.
|
|
2276
2270
|
- **negative** (Tensor) - A sample belonging to the different class from `x`, with the same type and shape
|
|
2277
2271
|
as `x`. :math:`n` in the above formula.
|
|
2278
|
-
- **margin** (Union[Tensor, float]) - Make a margin between the positive pair and the negative pair.
|
|
2272
|
+
- **margin** (Union[Tensor, float]) - Make a margin between the positive pair and the negative pair. The length
|
|
2273
|
+
of shape of `margin` must be 0.
|
|
2279
2274
|
Default: ``1.0`` .
|
|
2280
2275
|
|
|
2281
2276
|
Outputs:
|
mindspore/nn/optim/ada_grad.py
CHANGED
|
@@ -78,6 +78,7 @@ class Adagrad(Optimizer):
|
|
|
78
78
|
:math:`state\_sum` stands for the accumulated squared sum of the gradients :math:`accum`.
|
|
79
79
|
:math:`g` stands for `grads`, :math:`\lambda` stands for `weight_decay`.
|
|
80
80
|
:math:`\gamma` stands for `learning_rate`, :math:`w` stands for `params`.
|
|
81
|
+
:math:`t` represents current `step`.
|
|
81
82
|
|
|
82
83
|
Note:
|
|
83
84
|
If parameters are not grouped, the `weight_decay` in optimizer will be applied on the network parameters without
|
mindspore/nn/optim/adadelta.py
CHANGED
|
@@ -134,9 +134,9 @@ class Adadelta(Optimizer):
|
|
|
134
134
|
|
|
135
135
|
Raises:
|
|
136
136
|
TypeError: If `learning_rate` is not one of int, float, Tensor, Iterable, LearningRateSchedule.
|
|
137
|
-
TypeError: If element of `
|
|
137
|
+
TypeError: If element of `params` is neither Parameter nor dict.
|
|
138
138
|
TypeError: If `rho`, `epsilon` or `loss_scale` is not a float.
|
|
139
|
-
TypeError: If `weight_decay` is
|
|
139
|
+
TypeError: If `weight_decay` is not float, int or cell.
|
|
140
140
|
ValueError: if `rho` is not in range [0.0, 1.0].
|
|
141
141
|
ValueError: If `loss_scale` is less than or equal to 0.
|
|
142
142
|
ValueError: If `learning_rate`, `epsilon` or `weight_decay` is less than 0.
|
mindspore/nn/optim/adam.py
CHANGED
|
@@ -633,7 +633,7 @@ class Adam(Optimizer):
|
|
|
633
633
|
Raises:
|
|
634
634
|
KeyError: If kwargs got keys other than 'use_lazy' or 'use_offload'.
|
|
635
635
|
TypeError: If `learning_rate` is not one of int, float, Tensor, Iterable, LearningRateSchedule.
|
|
636
|
-
TypeError: If element of `
|
|
636
|
+
TypeError: If element of `params` is neither Parameter nor dict.
|
|
637
637
|
TypeError: If `beta1`, `beta2`, `eps` or `loss_scale` is not a float.
|
|
638
638
|
TypeError: If `weight_decay` is neither float nor int.
|
|
639
639
|
TypeError: If `use_locking`, `use_nesterov`, `use_amsgrad`, `use_lazy` or `use_offload` is not a bool.
|
mindspore/nn/optim/lars.py
CHANGED
|
@@ -82,7 +82,7 @@ class LARS(Optimizer):
|
|
|
82
82
|
&\hline \\[-1.ex]
|
|
83
83
|
\end{array}
|
|
84
84
|
|
|
85
|
-
:math:`w` represents the network
|
|
85
|
+
:math:`w` represents the network's params, :math:`g` represents `gradients`,
|
|
86
86
|
:math:`t` represents the current step, :math:`\lambda` represents `weight_decay` in `optimizer`,
|
|
87
87
|
:math:`\gamma` represents `learning_rate` in `optimizer`, :math:`\eta` represents `coefficient`.
|
|
88
88
|
|
|
@@ -98,9 +98,6 @@ class LARS(Optimizer):
|
|
|
98
98
|
- **gradients** (tuple[Tensor]) - The gradients of `params` in the optimizer, the shape is the
|
|
99
99
|
as same as the `params` in the optimizer.
|
|
100
100
|
|
|
101
|
-
Outputs:
|
|
102
|
-
Union[Tensor[bool], tuple[Parameter]], it depends on the output of `optimizer`.
|
|
103
|
-
|
|
104
101
|
Supported Platforms:
|
|
105
102
|
``Ascend``
|
|
106
103
|
|
mindspore/nn/optim/optimizer.py
CHANGED
|
@@ -848,7 +848,7 @@ class Optimizer(Cell):
|
|
|
848
848
|
optim_result(bool): The results of updating parameters. This input is used to ensure that the parameters are
|
|
849
849
|
updated before they are broadcast.
|
|
850
850
|
Returns:
|
|
851
|
-
|
|
851
|
+
The broadcast parameters.
|
|
852
852
|
"""
|
|
853
853
|
# If rank_id is 0, 1, 2, 3, there are param0 ~ param7,
|
|
854
854
|
# then the value is[(param0, param4), (param1, param5), (param2, param6), (param3, param7)]
|
mindspore/nn/optim/rprop.py
CHANGED
|
@@ -44,8 +44,8 @@ class Rprop(Optimizer):
|
|
|
44
44
|
&\hspace{15mm} w_{t} \leftarrow w_{t-1}- \Delta_{t} \mathrm{sign}(g_t) \\
|
|
45
45
|
\end{gather*}
|
|
46
46
|
|
|
47
|
-
:math
|
|
48
|
-
|
|
47
|
+
:math:`g` represents `gradients`, :math:`w` represents `parameters`, :math:`\Delta_{min/max}` represents the
|
|
48
|
+
min/max step size, :math:`\eta_{+/-}` represents the factors of etaminus and etaplus.
|
|
49
49
|
|
|
50
50
|
Note:
|
|
51
51
|
If parameters are not grouped, the `weight_decay` in optimizer will be applied on the parameters without 'beta'
|
mindspore/nn/optim/thor.py
CHANGED
|
@@ -21,6 +21,7 @@ from mindspore.ops import functional as F, composite as C, operations as P
|
|
|
21
21
|
from mindspore.common.initializer import initializer
|
|
22
22
|
from mindspore.common.parameter import Parameter, ParameterTuple
|
|
23
23
|
from mindspore.common.tensor import Tensor
|
|
24
|
+
from mindspore.common import set_recursion_limit
|
|
24
25
|
import mindspore.ops as ops
|
|
25
26
|
import mindspore.nn as nn
|
|
26
27
|
import mindspore.common.dtype as mstype
|
|
@@ -355,7 +356,7 @@ def thor(net, learning_rate, damping, momentum, weight_decay=0.0, loss_scale=1.0
|
|
|
355
356
|
... amp_level="O2", keep_batchnorm_fp32=False)
|
|
356
357
|
|
|
357
358
|
"""
|
|
358
|
-
|
|
359
|
+
set_recursion_limit(10000)
|
|
359
360
|
ConvertNetUtils().convert_to_thor_net(net)
|
|
360
361
|
if context.get_context("device_target") == "Ascend":
|
|
361
362
|
return ThorAscend(net, learning_rate, damping, momentum, weight_decay, loss_scale, batch_size, decay_filter,
|
mindspore/nn/utils/init.py
CHANGED
|
@@ -23,19 +23,21 @@ from mindspore.common.parameter import Parameter
|
|
|
23
23
|
@contextmanager
|
|
24
24
|
def no_init_parameters():
|
|
25
25
|
r"""
|
|
26
|
-
|
|
27
|
-
instantiated and occupy physical memory. Loading a checkpoint will replace the parameter values.
|
|
28
|
-
Decorator can be applied during network instantiation to add an attribute `init_param` to all
|
|
29
|
-
parameters within the current Cell, setting it to `init_param=False` .
|
|
30
|
-
When `init_param=False` is detected, the initialization of the parameters is skipped,
|
|
31
|
-
and the parameters are assigned values directly from the checkpoint during loading,
|
|
32
|
-
which can optimize performance and reduce physical memory usage.
|
|
26
|
+
This interface is used to skip parameter initialization.
|
|
33
27
|
|
|
34
|
-
|
|
35
|
-
|
|
36
|
-
|
|
28
|
+
In scenarios where a checkpoint is loaded, parameters within the network instantiation will be
|
|
29
|
+
instantiated and occupy physical memory. Loading a checkpoint will replace the parameter values.
|
|
30
|
+
Decorator can be applied during network instantiation to add an attribute `init_param` to all
|
|
31
|
+
parameters within the current Cell, setting it to `init_param=False` .
|
|
32
|
+
When `init_param=False` is detected, the initialization of the parameters is skipped,
|
|
33
|
+
and the parameters are assigned values directly from the checkpoint during loading,
|
|
34
|
+
which can optimize performance and reduce physical memory usage.
|
|
37
35
|
|
|
38
|
-
|
|
36
|
+
Note:
|
|
37
|
+
Initialization of parameters created with `initializer` can only be skipped.
|
|
38
|
+
Parameters created by `Tensor` or `numpy` cannot be skipped.
|
|
39
|
+
|
|
40
|
+
Examples:
|
|
39
41
|
>>> import mindspore as ms
|
|
40
42
|
>>> from mindspore import nn, ops, load_checkpoint
|
|
41
43
|
>>> from mindspore.common.initializer import initializer
|