mindspore 2.6.0rc1__cp39-cp39-win_amd64.whl → 2.7.0rc1__cp39-cp39-win_amd64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of mindspore might be problematic. Click here for more details.
- mindspore/.commit_id +1 -1
- mindspore/__init__.py +1 -1
- mindspore/_c_dataengine.cp39-win_amd64.pyd +0 -0
- mindspore/_c_expression.cp39-win_amd64.pyd +0 -0
- mindspore/_c_mindrecord.cp39-win_amd64.pyd +0 -0
- mindspore/_checkparam.py +40 -9
- mindspore/{_deprecated → _extends/optimize}/__init__.py +9 -3
- mindspore/_extends/optimize/cell_utils.py +96 -0
- mindspore/_extends/parse/__init__.py +2 -2
- mindspore/_extends/parse/compile_config.py +44 -22
- mindspore/_extends/parse/deprecated/deprecated_tensor_method.py +1 -1
- mindspore/_extends/parse/parser.py +37 -62
- mindspore/_extends/parse/resources.py +39 -0
- mindspore/_extends/parse/standard_method.py +43 -13
- mindspore/_extends/parse/trope.py +8 -1
- mindspore/_extends/pijit/__init__.py +1 -2
- mindspore/amp.py +4 -4
- mindspore/avcodec-59.dll +0 -0
- mindspore/avdevice-59.dll +0 -0
- mindspore/avfilter-8.dll +0 -0
- mindspore/avformat-59.dll +0 -0
- mindspore/avutil-57.dll +0 -0
- mindspore/boost/adasum.py +1 -1
- mindspore/boost/boost_cell_wrapper.py +4 -4
- mindspore/common/__init__.py +27 -2
- mindspore/common/_grad_function.py +2 -1
- mindspore/common/_pijit_context.py +28 -7
- mindspore/common/_stub_tensor.py +1 -209
- mindspore/common/_tensor_cpp_method.py +1 -1
- mindspore/common/_tensor_docs.py +77 -16
- mindspore/common/api.py +238 -113
- mindspore/common/dtype.py +21 -11
- mindspore/common/dump.py +10 -15
- mindspore/common/generator.py +5 -3
- mindspore/common/hook_handle.py +11 -2
- mindspore/common/jit_config.py +1 -1
- mindspore/common/jit_trace.py +84 -105
- mindspore/common/parameter.py +26 -12
- mindspore/common/recompute.py +3 -3
- mindspore/common/sparse_tensor.py +0 -3
- mindspore/common/symbol.py +0 -1
- mindspore/common/tensor.py +81 -81
- mindspore/communication/_comm_helper.py +46 -4
- mindspore/communication/management.py +79 -7
- mindspore/context.py +58 -40
- mindspore/dataset/core/config.py +3 -3
- mindspore/dataset/engine/datasets.py +20 -7
- mindspore/dataset/engine/datasets_user_defined.py +33 -3
- mindspore/dataset/engine/iterators.py +2 -2
- mindspore/dataset/engine/obs/config_loader.py +2 -2
- mindspore/dataset/engine/obs/obs_mindrecord_dataset.py +8 -0
- mindspore/dataset/transforms/py_transforms.py +7 -3
- mindspore/dataset/transforms/transforms.py +7 -3
- mindspore/dataset/vision/validators.py +1 -0
- mindspore/device_context/ascend/device.py +1 -1
- mindspore/device_context/gpu/__init__.py +2 -2
- mindspore/device_context/gpu/device.py +1 -1
- mindspore/device_context/gpu/op_precision.py +4 -2
- mindspore/device_context/gpu/op_tuning.py +6 -3
- mindspore/device_manager.py +16 -9
- mindspore/dnnl.dll +0 -0
- mindspore/experimental/llm_boost/ascend_native/llama_boost_ascend_native.py +3 -7
- mindspore/experimental/llm_boost/atb/boost_base.py +2 -3
- mindspore/experimental/optim/adadelta.py +13 -20
- mindspore/experimental/optim/adagrad.py +15 -22
- mindspore/experimental/optim/adam.py +17 -24
- mindspore/experimental/optim/adamax.py +14 -22
- mindspore/experimental/optim/adamw.py +28 -34
- mindspore/experimental/optim/asgd.py +15 -25
- mindspore/experimental/optim/lr_scheduler.py +27 -45
- mindspore/experimental/optim/nadam.py +14 -24
- mindspore/experimental/optim/optimizer.py +13 -23
- mindspore/experimental/optim/radam.py +18 -24
- mindspore/experimental/optim/rmsprop.py +14 -25
- mindspore/experimental/optim/rprop.py +15 -26
- mindspore/experimental/optim/sgd.py +9 -19
- mindspore/hal/__init__.py +4 -4
- mindspore/hal/contiguous_tensors_handle.py +2 -2
- mindspore/hal/memory.py +27 -7
- mindspore/include/api/cell.h +37 -1
- mindspore/include/api/delegate.h +10 -0
- mindspore/include/api/model.h +3 -0
- mindspore/include/api/types.h +2 -2
- mindspore/include/c_api/model_c.h +0 -58
- mindspore/include/c_api/tensor_c.h +0 -26
- mindspore/include/dataset/vision_ascend.h +1 -1
- mindspore/jpeg62.dll +0 -0
- mindspore/mindrecord/tools/cifar10.py +60 -11
- mindspore/mindrecord/tools/cifar10_to_mr.py +5 -0
- mindspore/mindspore_backend_common.dll +0 -0
- mindspore/mindspore_backend_manager.dll +0 -0
- mindspore/mindspore_common.dll +0 -0
- mindspore/mindspore_core.dll +0 -0
- mindspore/mindspore_cpu_res_manager.dll +0 -0
- mindspore/mindspore_dump.dll +0 -0
- mindspore/mindspore_frontend.dll +0 -0
- mindspore/mindspore_glog.dll +0 -0
- mindspore/mindspore_memory_pool.dll +0 -0
- mindspore/mindspore_ms_backend.dll +0 -0
- mindspore/mindspore_ops.dll +0 -0
- mindspore/mindspore_ops_host.dll +0 -0
- mindspore/mindspore_ops_kernel_common.dll +0 -0
- mindspore/mindspore_profiler.dll +0 -0
- mindspore/mindspore_pyboost.dll +0 -0
- mindspore/mindspore_pynative.dll +0 -0
- mindspore/mindspore_res_manager.dll +0 -0
- mindspore/mindspore_runtime_pipeline.dll +0 -0
- mindspore/mint/__init__.py +6 -46
- mindspore/mint/distributed/__init__.py +1 -0
- mindspore/mint/distributed/distributed.py +212 -9
- mindspore/mint/nn/__init__.py +1 -1
- mindspore/mint/nn/functional.py +53 -6
- mindspore/mint/nn/layer/_functions.py +164 -294
- mindspore/mint/nn/layer/activation.py +8 -6
- mindspore/mint/nn/layer/conv.py +137 -101
- mindspore/mint/nn/layer/normalization.py +8 -22
- mindspore/mint/optim/adam.py +19 -18
- mindspore/mint/optim/adamw.py +14 -8
- mindspore/mint/optim/sgd.py +5 -5
- mindspore/nn/cell.py +328 -502
- mindspore/nn/grad/cell_grad.py +11 -12
- mindspore/nn/layer/activation.py +32 -34
- mindspore/nn/layer/basic.py +67 -64
- mindspore/nn/layer/channel_shuffle.py +4 -4
- mindspore/nn/layer/combined.py +4 -2
- mindspore/nn/layer/conv.py +117 -110
- mindspore/nn/layer/dense.py +9 -7
- mindspore/nn/layer/embedding.py +50 -52
- mindspore/nn/layer/image.py +37 -39
- mindspore/nn/layer/math.py +111 -112
- mindspore/nn/layer/normalization.py +56 -44
- mindspore/nn/layer/pooling.py +58 -63
- mindspore/nn/layer/rnn_cells.py +33 -33
- mindspore/nn/layer/rnns.py +56 -56
- mindspore/nn/layer/thor_layer.py +74 -73
- mindspore/nn/layer/transformer.py +11 -1
- mindspore/nn/learning_rate_schedule.py +20 -20
- mindspore/nn/loss/loss.py +79 -81
- mindspore/nn/optim/adam.py +3 -3
- mindspore/nn/optim/adasum.py +2 -2
- mindspore/nn/optim/asgd.py +2 -0
- mindspore/nn/optim/optimizer.py +1 -1
- mindspore/nn/optim/thor.py +2 -2
- mindspore/nn/probability/distribution/exponential.py +2 -1
- mindspore/nn/probability/distribution/poisson.py +2 -1
- mindspore/nn/sparse/sparse.py +3 -3
- mindspore/nn/wrap/cell_wrapper.py +34 -37
- mindspore/nn/wrap/grad_reducer.py +37 -37
- mindspore/nn/wrap/loss_scale.py +72 -74
- mindspore/numpy/array_creations.py +5 -5
- mindspore/numpy/fft.py +1 -1
- mindspore/numpy/math_ops.py +5 -5
- mindspore/opencv_core452.dll +0 -0
- mindspore/opencv_imgcodecs452.dll +0 -0
- mindspore/opencv_imgproc452.dll +0 -0
- mindspore/ops/_grad_experimental/grad_comm_ops.py +51 -13
- mindspore/ops/_grad_experimental/grad_debug_ops.py +14 -0
- mindspore/ops/_vmap/vmap_array_ops.py +31 -13
- mindspore/ops/_vmap/vmap_nn_ops.py +8 -16
- mindspore/ops/auto_generate/cpp_create_prim_instance_helper.py +42 -11
- mindspore/ops/auto_generate/gen_extend_func.py +23 -141
- mindspore/ops/auto_generate/gen_ops_def.py +727 -321
- mindspore/ops/auto_generate/gen_ops_prim.py +1721 -984
- mindspore/ops/auto_generate/pyboost_inner_prim.py +31 -1
- mindspore/ops/composite/__init__.py +10 -0
- mindspore/ops/composite/base.py +8 -4
- mindspore/ops/composite/multitype_ops/__init__.py +12 -1
- mindspore/ops/composite/multitype_ops/_compile_utils.py +133 -109
- mindspore/ops/composite/multitype_ops/add_impl.py +70 -2
- mindspore/ops/composite/multitype_ops/div_impl.py +49 -0
- mindspore/ops/composite/multitype_ops/floordiv_impl.py +29 -0
- mindspore/ops/composite/multitype_ops/getitem_impl.py +11 -0
- mindspore/ops/composite/multitype_ops/mod_impl.py +5 -3
- mindspore/ops/composite/multitype_ops/mul_impl.py +49 -0
- mindspore/ops/composite/multitype_ops/setitem_impl.py +57 -0
- mindspore/ops/composite/multitype_ops/sub_impl.py +34 -0
- mindspore/ops/composite/multitype_ops/zeros_like_impl.py +14 -0
- mindspore/ops/function/__init__.py +3 -1
- mindspore/ops/function/_add_attr_func.py +11 -6
- mindspore/ops/function/array_func.py +9 -96
- mindspore/ops/function/debug_func.py +4 -3
- mindspore/ops/function/grad/grad_func.py +1 -1
- mindspore/ops/function/math_func.py +33 -540
- mindspore/ops/function/nn_func.py +28 -74
- mindspore/ops/function/other_func.py +4 -1
- mindspore/ops/function/random_func.py +44 -5
- mindspore/ops/function/vmap_func.py +2 -1
- mindspore/ops/functional.py +2 -3
- mindspore/ops/functional_overload.py +571 -6
- mindspore/ops/op_info_register.py +21 -0
- mindspore/ops/operations/__init__.py +16 -11
- mindspore/ops/operations/_custom_ops_utils.py +689 -34
- mindspore/ops/operations/_inner_ops.py +3 -6
- mindspore/ops/operations/_sequence_ops.py +1 -1
- mindspore/ops/operations/array_ops.py +2 -2
- mindspore/ops/operations/comm_ops.py +185 -26
- mindspore/ops/operations/custom_ops.py +294 -174
- mindspore/ops/operations/debug_ops.py +59 -4
- mindspore/ops/operations/image_ops.py +13 -13
- mindspore/ops/operations/manually_defined/ops_def.py +15 -16
- mindspore/ops/operations/math_ops.py +3 -4
- mindspore/ops/operations/nn_ops.py +7 -39
- mindspore/ops/primitive.py +6 -10
- mindspore/ops/tensor_method.py +47 -8
- mindspore/ops_generate/api/cpp_create_prim_instance_helper_generator.py +1 -1
- mindspore/ops_generate/api/functional_map_cpp_generator.py +10 -9
- mindspore/ops_generate/api/functions_cc_generator.py +58 -10
- mindspore/ops_generate/api/tensor_func_reg_cpp_generator.py +1 -1
- mindspore/ops_generate/common/base_generator.py +14 -0
- mindspore/ops_generate/common/gen_constants.py +8 -3
- mindspore/ops_generate/common/gen_utils.py +0 -19
- mindspore/ops_generate/common/op_proto.py +11 -4
- mindspore/ops_generate/common/template.py +88 -11
- mindspore/ops_generate/gen_ops.py +1 -1
- mindspore/ops_generate/op_def/lite_ops_cpp_generator.py +4 -4
- mindspore/ops_generate/op_def/ops_def_cc_generator.py +0 -3
- mindspore/ops_generate/op_def/ops_name_h_generator.py +0 -3
- mindspore/ops_generate/op_def/ops_primitive_h_generator.py +0 -4
- mindspore/ops_generate/op_def_py/op_prim_py_generator.py +5 -2
- mindspore/ops_generate/pyboost/auto_grad_impl_cc_generator.py +49 -8
- mindspore/ops_generate/pyboost/auto_grad_reg_cc_generator.py +2 -2
- mindspore/ops_generate/pyboost/gen_pyboost_func.py +31 -0
- mindspore/ops_generate/pyboost/op_template_parser.py +98 -72
- mindspore/ops_generate/pyboost/pyboost_functions_cpp_generator.py +70 -273
- mindspore/ops_generate/pyboost/pyboost_functions_h_generator.py +14 -6
- mindspore/ops_generate/pyboost/pyboost_functions_impl_cpp_generator.py +316 -0
- mindspore/ops_generate/pyboost/pyboost_functions_py_generator.py +1 -1
- mindspore/ops_generate/pyboost/pyboost_grad_function_cpp_generator.py +5 -3
- mindspore/ops_generate/pyboost/pyboost_inner_prim_generator.py +1 -1
- mindspore/ops_generate/pyboost/pyboost_internal_functions_cpp_generator.py +76 -0
- mindspore/ops_generate/pyboost/pyboost_internal_functions_h_generator.py +76 -0
- mindspore/ops_generate/pyboost/pyboost_internal_kernel_info_adapter_generator.py +125 -0
- mindspore/ops_generate/pyboost/pyboost_native_grad_functions_generator.py +4 -3
- mindspore/ops_generate/pyboost/pyboost_op_cpp_code_generator.py +348 -61
- mindspore/ops_generate/pyboost/pyboost_overload_functions_cpp_generator.py +1 -1
- mindspore/ops_generate/pyboost/pyboost_utils.py +118 -9
- mindspore/ops_generate/tensor_py_cc_generator.py +1 -24
- mindspore/parallel/_auto_parallel_context.py +11 -8
- mindspore/parallel/_cell_wrapper.py +113 -45
- mindspore/parallel/_parallel_serialization.py +1 -1
- mindspore/parallel/_ps_context.py +4 -6
- mindspore/parallel/_tensor.py +167 -12
- mindspore/parallel/_transformer/moe.py +1 -1
- mindspore/parallel/_transformer/transformer.py +13 -8
- mindspore/parallel/auto_parallel.py +14 -7
- mindspore/parallel/checkpoint_convert.py +3 -3
- mindspore/parallel/checkpoint_transform.py +11 -7
- mindspore/parallel/cluster/process_entity/_api.py +84 -48
- mindspore/parallel/cluster/process_entity/_utils.py +95 -7
- mindspore/parallel/cluster/run.py +43 -4
- mindspore/parallel/function/__init__.py +8 -1
- mindspore/parallel/function/reshard_func.py +6 -7
- mindspore/parallel/nn/__init__.py +15 -2
- mindspore/parallel/nn/parallel_cell_wrapper.py +9 -10
- mindspore/parallel/nn/parallel_grad_reducer.py +7 -6
- mindspore/parallel/shard.py +3 -4
- mindspore/parallel/transform_safetensors.py +463 -174
- mindspore/profiler/__init__.py +2 -1
- mindspore/profiler/analysis/parser/timeline_assembly_factory/ascend_timeline_assembler.py +7 -7
- mindspore/profiler/analysis/parser/timeline_assembly_factory/base_timeline_assembler.py +3 -0
- mindspore/profiler/analysis/parser/timeline_assembly_factory/trace_view_container.py +12 -6
- mindspore/profiler/analysis/parser/timeline_creator/cpu_op_timeline_creator.py +3 -3
- mindspore/profiler/analysis/parser/timeline_creator/fwk_timeline_creator.py +3 -3
- mindspore/profiler/analysis/parser/timeline_creator/msprof_timeline_creator.py +4 -4
- mindspore/profiler/analysis/parser/timeline_creator/scope_layer_timeline_creator.py +3 -3
- mindspore/profiler/analysis/parser/timeline_event/fwk_event.py +4 -1
- mindspore/profiler/analysis/parser/timeline_event/timeline_event_pool.py +2 -1
- mindspore/profiler/analysis/task_manager.py +1 -1
- mindspore/profiler/analysis/viewer/ascend_communication_viewer.py +5 -1
- mindspore/profiler/analysis/viewer/ascend_integrate_viewer.py +2 -1
- mindspore/profiler/analysis/viewer/ascend_op_memory_viewer.py +42 -22
- mindspore/profiler/analysis/viewer/ascend_step_trace_time_viewer.py +3 -2
- mindspore/profiler/analysis/viewer/ms_minddata_viewer.py +9 -5
- mindspore/profiler/analysis/viewer/ms_operator_details_viewer.py +132 -0
- mindspore/profiler/common/constant.py +16 -0
- mindspore/profiler/common/profiler_context.py +25 -27
- mindspore/profiler/common/profiler_info.py +0 -16
- mindspore/profiler/common/profiler_op_analyse.py +235 -0
- mindspore/profiler/common/profiler_output_path.py +23 -8
- mindspore/profiler/common/profiler_parameters.py +128 -35
- mindspore/profiler/dynamic_profile/__init__.py +0 -0
- mindspore/profiler/dynamic_profile/dynamic_monitor_proxy.py +39 -0
- mindspore/profiler/dynamic_profile/dynamic_profiler_config_context.py +666 -0
- mindspore/profiler/dynamic_profile/dynamic_profiler_utils.py +62 -0
- mindspore/profiler/dynamic_profiler.py +305 -314
- mindspore/profiler/envprofiler.py +12 -7
- mindspore/profiler/experimental_config.py +96 -6
- mindspore/profiler/mstx.py +33 -12
- mindspore/profiler/platform/__init__.py +2 -3
- mindspore/profiler/platform/npu_profiler.py +29 -19
- mindspore/profiler/profiler.py +35 -19
- mindspore/profiler/profiler_action_controller.py +64 -76
- mindspore/profiler/schedule.py +10 -4
- mindspore/rewrite/common/config.py +1 -0
- mindspore/rewrite/common/namer.py +1 -0
- mindspore/rewrite/common/namespace.py +1 -0
- mindspore/rewrite/node/node.py +31 -11
- mindspore/rewrite/parsers/assign_parser.py +1 -1
- mindspore/rewrite/symbol_tree/symbol_tree.py +1 -1
- mindspore/run_check/_check_version.py +7 -10
- mindspore/runtime/__init__.py +5 -5
- mindspore/runtime/event.py +10 -4
- mindspore/runtime/executor.py +60 -45
- mindspore/runtime/memory.py +30 -32
- mindspore/runtime/thread_bind_core.py +298 -164
- mindspore/safeguard/rewrite_obfuscation.py +12 -13
- mindspore/swresample-4.dll +0 -0
- mindspore/swscale-6.dll +0 -0
- mindspore/tinyxml2.dll +0 -0
- mindspore/train/_utils.py +14 -4
- mindspore/train/amp.py +43 -20
- mindspore/train/callback/__init__.py +5 -5
- mindspore/train/callback/_checkpoint.py +3 -6
- mindspore/train/callback/_flops_collector.py +1 -1
- mindspore/train/callback/_landscape.py +0 -1
- mindspore/train/callback/_train_fault_tolerance.py +97 -16
- mindspore/train/data_sink.py +11 -2
- mindspore/train/dataset_helper.py +9 -0
- mindspore/train/model.py +135 -55
- mindspore/train/serialization.py +133 -111
- mindspore/train/summary/summary_record.py +13 -2
- mindspore/turbojpeg.dll +0 -0
- mindspore/utils/__init__.py +3 -2
- mindspore/utils/dryrun.py +0 -6
- mindspore/utils/runtime_execution_order_check.py +163 -77
- mindspore/utils/sdc_detect.py +68 -0
- mindspore/utils/utils.py +6 -9
- mindspore/version.py +1 -1
- {mindspore-2.6.0rc1.dist-info → mindspore-2.7.0rc1.dist-info}/METADATA +5 -4
- {mindspore-2.6.0rc1.dist-info → mindspore-2.7.0rc1.dist-info}/RECORD +333 -371
- mindspore/_deprecated/jit.py +0 -198
- mindspore/experimental/es/__init__.py +0 -22
- mindspore/experimental/es/embedding_service.py +0 -891
- mindspore/experimental/es/embedding_service_layer.py +0 -581
- mindspore/profiler/parser/__init__.py +0 -14
- mindspore/profiler/parser/aicpu_data_parser.py +0 -272
- mindspore/profiler/parser/ascend_analysis/__init__.py +0 -14
- mindspore/profiler/parser/ascend_analysis/constant.py +0 -71
- mindspore/profiler/parser/ascend_analysis/file_manager.py +0 -180
- mindspore/profiler/parser/ascend_analysis/function_event.py +0 -185
- mindspore/profiler/parser/ascend_analysis/fwk_cann_parser.py +0 -136
- mindspore/profiler/parser/ascend_analysis/fwk_file_parser.py +0 -131
- mindspore/profiler/parser/ascend_analysis/msprof_timeline_parser.py +0 -104
- mindspore/profiler/parser/ascend_analysis/path_manager.py +0 -313
- mindspore/profiler/parser/ascend_analysis/profiler_info_parser.py +0 -123
- mindspore/profiler/parser/ascend_analysis/tlv_decoder.py +0 -86
- mindspore/profiler/parser/ascend_analysis/trace_event_manager.py +0 -75
- mindspore/profiler/parser/ascend_cluster_generator.py +0 -116
- mindspore/profiler/parser/ascend_communicate_generator.py +0 -314
- mindspore/profiler/parser/ascend_flops_generator.py +0 -116
- mindspore/profiler/parser/ascend_fpbp_generator.py +0 -82
- mindspore/profiler/parser/ascend_hccl_generator.py +0 -271
- mindspore/profiler/parser/ascend_integrate_generator.py +0 -42
- mindspore/profiler/parser/ascend_memory_generator.py +0 -185
- mindspore/profiler/parser/ascend_msprof_exporter.py +0 -282
- mindspore/profiler/parser/ascend_msprof_generator.py +0 -187
- mindspore/profiler/parser/ascend_op_generator.py +0 -334
- mindspore/profiler/parser/ascend_steptrace_generator.py +0 -94
- mindspore/profiler/parser/ascend_timeline_generator.py +0 -545
- mindspore/profiler/parser/base_timeline_generator.py +0 -483
- mindspore/profiler/parser/container.py +0 -229
- mindspore/profiler/parser/cpu_gpu_timeline_generator.py +0 -697
- mindspore/profiler/parser/flops_parser.py +0 -531
- mindspore/profiler/parser/framework_enum.py +0 -111
- mindspore/profiler/parser/framework_parser.py +0 -464
- mindspore/profiler/parser/framework_struct.py +0 -61
- mindspore/profiler/parser/gpu_analysis/__init__.py +0 -14
- mindspore/profiler/parser/gpu_analysis/function_event.py +0 -44
- mindspore/profiler/parser/gpu_analysis/fwk_file_parser.py +0 -89
- mindspore/profiler/parser/gpu_analysis/profiler_info_parser.py +0 -72
- mindspore/profiler/parser/hccl_parser.py +0 -573
- mindspore/profiler/parser/hwts_log_parser.py +0 -122
- mindspore/profiler/parser/integrator.py +0 -526
- mindspore/profiler/parser/memory_usage_parser.py +0 -277
- mindspore/profiler/parser/minddata_analyzer.py +0 -800
- mindspore/profiler/parser/minddata_parser.py +0 -186
- mindspore/profiler/parser/minddata_pipeline_parser.py +0 -299
- mindspore/profiler/parser/op_intermediate_parser.py +0 -149
- mindspore/profiler/parser/optime_parser.py +0 -250
- mindspore/profiler/parser/profiler_info.py +0 -213
- mindspore/profiler/parser/step_trace_parser.py +0 -666
- {mindspore-2.6.0rc1.dist-info → mindspore-2.7.0rc1.dist-info}/WHEEL +0 -0
- {mindspore-2.6.0rc1.dist-info → mindspore-2.7.0rc1.dist-info}/entry_points.txt +0 -0
- {mindspore-2.6.0rc1.dist-info → mindspore-2.7.0rc1.dist-info}/top_level.txt +0 -0
|
@@ -14,9 +14,7 @@
|
|
|
14
14
|
# ============================================================================
|
|
15
15
|
|
|
16
16
|
"""Inner operators."""
|
|
17
|
-
from types import FunctionType, MethodType
|
|
18
17
|
from collections.abc import Iterable
|
|
19
|
-
import os
|
|
20
18
|
import weakref
|
|
21
19
|
import numpy as np
|
|
22
20
|
|
|
@@ -34,11 +32,10 @@ from mindspore._c_expression import pyboost_generator
|
|
|
34
32
|
from mindspore import _checkparam as validator
|
|
35
33
|
from mindspore.common import dtype as mstype
|
|
36
34
|
from mindspore.common.parameter import Parameter
|
|
37
|
-
from mindspore.common._stub_tensor import _convert_stub
|
|
38
35
|
from mindspore.communication.management import GlobalComm, get_rank, _get_group, get_group_size
|
|
39
36
|
from mindspore.common.api import _pynative_executor
|
|
40
|
-
from ..auto_generate import TensorCopySlices, SiLU, Cummin, TopKRouter,
|
|
41
|
-
|
|
37
|
+
from ..auto_generate import TensorCopySlices, SiLU, Cummin, TopKRouter, TopPRouter, ExtractImagePatches, \
|
|
38
|
+
ApplyCamePart1, ApplyCamePart2, ApplyCamePart3, ApplyCamePart4
|
|
42
39
|
|
|
43
40
|
# Bit operation
|
|
44
41
|
bit_and = bit_and()
|
|
@@ -78,7 +75,7 @@ class Generator(Primitive):
|
|
|
78
75
|
def __call__(self, cmd, inputs):
|
|
79
76
|
if cmd == 0: # step cmd
|
|
80
77
|
return inputs[0], inputs[1]
|
|
81
|
-
return
|
|
78
|
+
return pyboost_generator(self, [cmd, inputs])
|
|
82
79
|
|
|
83
80
|
|
|
84
81
|
class Quant(PrimitiveWithInfer):
|
|
@@ -36,13 +36,13 @@ from mindspore._c_expression import TensorPy as Tensor_
|
|
|
36
36
|
from mindspore._c_expression import CSRTensor as CSRTensor_
|
|
37
37
|
from mindspore._c_expression import COOTensor as COOTensor_
|
|
38
38
|
from ..auto_generate import (
|
|
39
|
-
ExpandDims, Reshape, TensorShape, Transpose, Gather, OnesLike, ZerosLike,
|
|
39
|
+
ExpandDims, Reshape, TensorShape, Transpose, TransposeView, Gather, OnesLike, ZerosLike,
|
|
40
40
|
Argmax, ArgMaxExt, ReverseV2, Diag, Eye, ScatterNd,
|
|
41
41
|
ResizeNearestNeighborV2, GatherNd, GatherD, Range, MaskedFill, RightShift,
|
|
42
42
|
NonZero, ResizeNearestNeighbor, Identity, Split, CumSum, CumProd,
|
|
43
43
|
MaskedSelect, Cummax, Cummin, Argmin, Concat, UnsortedSegmentSum, UniqueConsecutive,
|
|
44
44
|
ScalarToTensor, Triu, BroadcastTo, StridedSlice, Select, TopkExt,
|
|
45
|
-
SearchSorted, Meshgrid, Squeeze, Slice)
|
|
45
|
+
SearchSorted, Meshgrid, Squeeze, Slice, TransposeExtView)
|
|
46
46
|
from .manually_defined import Rank, Shape, Tile, Cast, Ones, Zeros, TypeAs
|
|
47
47
|
from ..auto_generate import ArgMaxWithValue, ArgMinWithValue
|
|
48
48
|
from ..auto_generate import TensorScatterElements as TensorScatterElementsExt
|
|
@@ -300,7 +300,7 @@ class AllGather(PrimitiveWithInfer):
|
|
|
300
300
|
|
|
301
301
|
Outputs:
|
|
302
302
|
Tensor. If the number of devices in the group is N,
|
|
303
|
-
then the shape of output is :math:`(N
|
|
303
|
+
then the shape of output is :math:`(N*x_1, x_2, ..., x_R)`.
|
|
304
304
|
|
|
305
305
|
Raises:
|
|
306
306
|
TypeError: If `group` is not a str.
|
|
@@ -957,25 +957,8 @@ class AlltoAll(PrimitiveWithInfer):
|
|
|
957
957
|
self.concat_dim = concat_dim
|
|
958
958
|
self.add_prim_attr('group', _get_group(group))
|
|
959
959
|
self.add_prim_attr('no_eliminate', True)
|
|
960
|
-
|
|
961
|
-
|
|
962
|
-
rank_size = get_group_size(_get_group(self.group))
|
|
963
|
-
if self.split_count != rank_size:
|
|
964
|
-
raise ValueError(f"For '{self.name}', the 'split_count' must be equal to 'rank_size', "
|
|
965
|
-
f"but got 'split_count': {self.split_count}, 'rank_size': {rank_size}.")
|
|
966
|
-
if x_shape[self.split_dim] >= 0 and x_shape[self.split_dim] % self.split_count != 0:
|
|
967
|
-
raise ValueError(f"For '{self.name}', the 'x_shape[self.split_dim]' must be divisible by 'split_count', "
|
|
968
|
-
f"but got 'x_shape[self.split_dim]' {x_shape[self.split_dim]}, "
|
|
969
|
-
f"'split_count' {self.split_count}.")
|
|
970
|
-
if x_shape[self.concat_dim] >= 0:
|
|
971
|
-
x_shape[self.concat_dim] = x_shape[self.concat_dim] * self.split_count
|
|
972
|
-
if x_shape[self.split_dim] >= 0:
|
|
973
|
-
x_shape[self.split_dim] = int(x_shape[self.split_dim] / self.split_count)
|
|
974
|
-
return x_shape
|
|
975
|
-
|
|
976
|
-
def infer_dtype(self, x_dtype):
|
|
977
|
-
check_collective_target_dtype('x', x_dtype, self.name)
|
|
978
|
-
return x_dtype
|
|
960
|
+
self.rank_size = get_group_size(_get_group(group))
|
|
961
|
+
self.add_prim_attr('rank_size', self.rank_size)
|
|
979
962
|
|
|
980
963
|
|
|
981
964
|
class NeighborExchangeV2(Primitive):
|
|
@@ -1749,10 +1732,10 @@ class _VirtualAssignKvCache(PrimitiveWithInfer):
|
|
|
1749
1732
|
self.add_prim_attr('order_enforce_skip', True)
|
|
1750
1733
|
self.add_prim_attr('side_effect_backprop_mem', True)
|
|
1751
1734
|
|
|
1752
|
-
def infer_shape(self, x_shape, y_shape, kv_equal_shape):
|
|
1735
|
+
def infer_shape(self, x_shape, y_shape, kv_equal_shape=None):
|
|
1753
1736
|
return x_shape
|
|
1754
1737
|
|
|
1755
|
-
def infer_dtype(self, x_dtype, y_dtype, kv_equal_dtype):
|
|
1738
|
+
def infer_dtype(self, x_dtype, y_dtype, kv_equal_dtype=None):
|
|
1756
1739
|
return x_dtype
|
|
1757
1740
|
virtual_assign_kv_cache = _VirtualAssignKvCache()
|
|
1758
1741
|
|
|
@@ -1975,8 +1958,8 @@ class AlltoAllV(PrimitiveWithInfer):
|
|
|
1975
1958
|
AllToAllV which support uneven scatter and gather compared with AllToAll.
|
|
1976
1959
|
|
|
1977
1960
|
Note:
|
|
1978
|
-
|
|
1979
|
-
|
|
1961
|
+
Only support flatten tensor as input. input tensor should be flattened and
|
|
1962
|
+
concatenated before call this primitive.
|
|
1980
1963
|
|
|
1981
1964
|
Args:
|
|
1982
1965
|
group (str, optional): The communication group to work on. Default: ``GlobalComm.WORLD_COMM_GROUP``, which
|
|
@@ -2012,9 +1995,9 @@ class AlltoAllV(PrimitiveWithInfer):
|
|
|
2012
1995
|
|
|
2013
1996
|
This example should be run with 2 devices.
|
|
2014
1997
|
|
|
2015
|
-
>>> from mindspore import ops
|
|
2016
1998
|
>>> import mindspore.nn as nn
|
|
2017
1999
|
>>> from mindspore.communication import init, get_rank
|
|
2000
|
+
>>> from mindspore.ops import AlltoAllV
|
|
2018
2001
|
>>> from mindspore import Tensor
|
|
2019
2002
|
>>>
|
|
2020
2003
|
>>> init()
|
|
@@ -2022,7 +2005,7 @@ class AlltoAllV(PrimitiveWithInfer):
|
|
|
2022
2005
|
>>> class Net(nn.Cell):
|
|
2023
2006
|
... def __init__(self):
|
|
2024
2007
|
... super(Net, self).__init__()
|
|
2025
|
-
... self.all_to_all =
|
|
2008
|
+
... self.all_to_all = AlltoAllV()
|
|
2026
2009
|
...
|
|
2027
2010
|
... def construct(self, x, send_numel_list, recv_numel_list):
|
|
2028
2011
|
... return self.all_to_all(x, send_numel_list, recv_numel_list)
|
|
@@ -2053,3 +2036,179 @@ class AlltoAllV(PrimitiveWithInfer):
|
|
|
2053
2036
|
self.add_prim_attr('group', self.group)
|
|
2054
2037
|
validator.check_value_type("block_size", block_size, [int], self.name)
|
|
2055
2038
|
self.add_prim_attr('block_size', self.block_size)
|
|
2039
|
+
|
|
2040
|
+
|
|
2041
|
+
class AllGatherV(PrimitiveWithInfer):
|
|
2042
|
+
"""
|
|
2043
|
+
Gathers uneven tensors from the specified communication group and returns the tensor which is all gathered.
|
|
2044
|
+
|
|
2045
|
+
Note:
|
|
2046
|
+
Only support flatten tensor as input. input tensor should be flattened and
|
|
2047
|
+
concatenated before call this primitive.
|
|
2048
|
+
|
|
2049
|
+
Args:
|
|
2050
|
+
group (str, optional): The communication group to work on. Default: ``GlobalComm.WORLD_COMM_GROUP`` , which
|
|
2051
|
+
means ``"hccl_world_group"`` in Ascend, and ``"nccl_world_group"`` in GPU.
|
|
2052
|
+
|
|
2053
|
+
Inputs:
|
|
2054
|
+
- **input_x** (Tensor) - One-dimensional tensor to be gathered, with the shape :math:`(x_1)`.
|
|
2055
|
+
- **output_split_sizes** (Union[tuple[int], list[int], Tensor]) - One-dimensional tensor, a list of the
|
|
2056
|
+
amount of data gathered by all ranks. The basic unit is the data type of Tensor.
|
|
2057
|
+
|
|
2058
|
+
Outputs:
|
|
2059
|
+
Tensor. flattened and concatenated tensor gather from remote ranks.
|
|
2060
|
+
If gather result is empty, it will return a Tensor with shape `()`, and value has no actual meaning.
|
|
2061
|
+
|
|
2062
|
+
Raises:
|
|
2063
|
+
RuntimeError: Device target is invalid, backend is invalid, or distributed initialization fails.
|
|
2064
|
+
|
|
2065
|
+
Supported Platforms:
|
|
2066
|
+
``Ascend`` ``GPU``
|
|
2067
|
+
|
|
2068
|
+
Examples:
|
|
2069
|
+
.. note::
|
|
2070
|
+
Before running the following examples, you need to configure the communication environment variables.
|
|
2071
|
+
|
|
2072
|
+
For Ascend/GPU/CPU devices, it is recommended to use the msrun startup method
|
|
2073
|
+
without any third-party or configuration file dependencies.
|
|
2074
|
+
Please see the `msrun start up
|
|
2075
|
+
<https://www.mindspore.cn/tutorials/en/master/parallel/msrun_launcher.html>`_
|
|
2076
|
+
for more details.
|
|
2077
|
+
|
|
2078
|
+
This example should be run with 2 devices.
|
|
2079
|
+
|
|
2080
|
+
>>> import mindspore as ms
|
|
2081
|
+
>>> from mindspore.ops import AllGatherV
|
|
2082
|
+
>>> import mindspore.nn as nn
|
|
2083
|
+
>>> from mindspore.communication import init, get_rank
|
|
2084
|
+
>>> from mindspore import Tensor
|
|
2085
|
+
>>>
|
|
2086
|
+
>>> ms.set_context(mode=ms.GRAPH_MODE)
|
|
2087
|
+
>>> init()
|
|
2088
|
+
>>> class Net(nn.Cell):
|
|
2089
|
+
... def __init__(self):
|
|
2090
|
+
... super(Net, self).__init__()
|
|
2091
|
+
... self.allgatherv = AllGatherV()
|
|
2092
|
+
...
|
|
2093
|
+
... def construct(self, x, output_split_sizes):
|
|
2094
|
+
... return self.allgatherv(x, output_split_sizes)
|
|
2095
|
+
...
|
|
2096
|
+
>>> rank = get_rank()
|
|
2097
|
+
>>> data = [i for i in range(rank + 3)]
|
|
2098
|
+
>>> input_x = Tensor(data)
|
|
2099
|
+
>>> output_split_sizes = [3, 4]
|
|
2100
|
+
>>> net = Net()
|
|
2101
|
+
>>> output = net(input_x, output_split_sizes)
|
|
2102
|
+
>>> print(output)
|
|
2103
|
+
[0 1 2 0 1 2 3]
|
|
2104
|
+
|
|
2105
|
+
Tutorial Examples:
|
|
2106
|
+
- `Distributed Set Communication Primitives - AllGatherV
|
|
2107
|
+
<https://www.mindspore.cn/docs/en/master/api_python/samples/ops/communicate_ops.html#allgatherv>`_
|
|
2108
|
+
|
|
2109
|
+
"""
|
|
2110
|
+
|
|
2111
|
+
@prim_attr_register
|
|
2112
|
+
def __init__(self, group=GlobalComm.WORLD_COMM_GROUP):
|
|
2113
|
+
"""Initialize AllGatherV."""
|
|
2114
|
+
self.group = _get_group(group)
|
|
2115
|
+
validator.check_value_type('group', self.group, (str,), self.name)
|
|
2116
|
+
self.rank = get_rank(self.group)
|
|
2117
|
+
self.rank_size = get_group_size(self.group)
|
|
2118
|
+
validator.check('rank', self.rank, 'rank_size', self.rank_size, validator.LT, self.name)
|
|
2119
|
+
self.add_prim_attr('rank_size', self.rank_size)
|
|
2120
|
+
self.add_prim_attr('group', self.group)
|
|
2121
|
+
self.add_prim_attr('mean_flag', False)
|
|
2122
|
+
self.rank_id = get_rank(_get_group(self.group))
|
|
2123
|
+
self.add_prim_attr('rank_id', self.rank_id)
|
|
2124
|
+
|
|
2125
|
+
|
|
2126
|
+
class ReduceScatterV(PrimitiveWithInfer):
|
|
2127
|
+
r"""
|
|
2128
|
+
Reduces and scatters uneven tensors from the specified communication group
|
|
2129
|
+
and returns the tensor which is reduced and scattered.
|
|
2130
|
+
|
|
2131
|
+
Note:
|
|
2132
|
+
Only support flatten tensor as input. The input tensor should be flattened and
|
|
2133
|
+
concatenated before call this primitive.
|
|
2134
|
+
|
|
2135
|
+
Args:
|
|
2136
|
+
op (str, optional): Specifies an operation used for element-wise reductions,
|
|
2137
|
+
like SUM, MIN and MAX, currently PROD is not supported. Default: ``ReduceOp.SUM`` .
|
|
2138
|
+
group (str, optional): The communication group to work on. Default: ``GlobalComm.WORLD_COMM_GROUP`` , which
|
|
2139
|
+
means ``"hccl_world_group"`` in Ascend, and ``"nccl_world_group"`` in GPU.
|
|
2140
|
+
|
|
2141
|
+
Inputs:
|
|
2142
|
+
- **input_x** (Tensor) - One-dimensional tensor to be distributed, with the shape :math:`(x_1)`.
|
|
2143
|
+
- **input_split_sizes** (Union[tuple[int], list[int], Tensor]) - One-dimensional tensor, a list of
|
|
2144
|
+
received data volumes for all ranks. The basic unit is the data type of Tensor. The value is not
|
|
2145
|
+
verified, and the user guarantees its correctness.
|
|
2146
|
+
|
|
2147
|
+
Outputs:
|
|
2148
|
+
Tensor. Reduces and scatters tensor from remote ranks.
|
|
2149
|
+
If the result is empty, it will return a Tensor with shape `()`, and value has no actual meaning.
|
|
2150
|
+
|
|
2151
|
+
Raises:
|
|
2152
|
+
RuntimeError: Device target is invalid, backend is invalid, or distributed initialization fails.
|
|
2153
|
+
|
|
2154
|
+
Supported Platforms:
|
|
2155
|
+
``Ascend`` ``GPU``
|
|
2156
|
+
|
|
2157
|
+
Examples:
|
|
2158
|
+
.. note::
|
|
2159
|
+
Before running the following examples, you need to configure the communication environment variables.
|
|
2160
|
+
|
|
2161
|
+
For Ascend/GPU/CPU devices, it is recommended to use the msrun startup method
|
|
2162
|
+
without any third-party or configuration file dependencies.
|
|
2163
|
+
Please see the `msrun start up
|
|
2164
|
+
<https://www.mindspore.cn/tutorials/en/master/parallel/msrun_launcher.html>`_
|
|
2165
|
+
for more details.
|
|
2166
|
+
|
|
2167
|
+
This example should be run with 2 devices.
|
|
2168
|
+
|
|
2169
|
+
>>> import mindspore as ms
|
|
2170
|
+
>>> from mindspore import Tensor
|
|
2171
|
+
>>> from mindspore.communication import init, get_rank
|
|
2172
|
+
>>> from mindspore.ops import ReduceOp
|
|
2173
|
+
>>> import mindspore.nn as nn
|
|
2174
|
+
>>> from mindspore.ops.operations.comm_ops import ReduceScatterV
|
|
2175
|
+
>>>
|
|
2176
|
+
>>> ms.set_context(mode=ms.GRAPH_MODE)
|
|
2177
|
+
>>> init()
|
|
2178
|
+
>>> class Net(nn.Cell):
|
|
2179
|
+
... def __init__(self):
|
|
2180
|
+
... super(Net, self).__init__()
|
|
2181
|
+
... self.reducescatterv = ReduceScatterV(ReduceOp.SUM)
|
|
2182
|
+
...
|
|
2183
|
+
... def construct(self, x, input_split_sizes):
|
|
2184
|
+
... return self.reducescatterv(x, input_split_sizes)
|
|
2185
|
+
...
|
|
2186
|
+
>>> rank = get_rank()
|
|
2187
|
+
>>> input_x = Tensor([0, 1, 2.0])
|
|
2188
|
+
>>> input_split_sizes = [2, 1]
|
|
2189
|
+
>>> net = Net()
|
|
2190
|
+
>>> output = net(input_x, input_split_sizes)
|
|
2191
|
+
>>> print(output)
|
|
2192
|
+
rank 0:
|
|
2193
|
+
[0. 2.]
|
|
2194
|
+
rank 1:
|
|
2195
|
+
[4.]
|
|
2196
|
+
|
|
2197
|
+
Tutorial Examples:
|
|
2198
|
+
- `Distributed Set Communication Primitives - ReduceScatterV
|
|
2199
|
+
<https://www.mindspore.cn/docs/en/master/api_python/samples/ops/communicate_ops.html#reducescatterv>`_
|
|
2200
|
+
|
|
2201
|
+
"""
|
|
2202
|
+
|
|
2203
|
+
@prim_attr_register
|
|
2204
|
+
def __init__(self, op=ReduceOp.SUM, group=GlobalComm.WORLD_COMM_GROUP):
|
|
2205
|
+
"""Initialize ReduceScatterV."""
|
|
2206
|
+
validator.check_value_type('op', op, (type(ReduceOp.SUM),), self.name)
|
|
2207
|
+
self.group = _get_group(group)
|
|
2208
|
+
validator.check_value_type('group', self.group, (str,), self.name)
|
|
2209
|
+
self.op = op
|
|
2210
|
+
self.rank_size = get_group_size(self.group)
|
|
2211
|
+
self.add_prim_attr('rank_size', self.rank_size)
|
|
2212
|
+
self.add_prim_attr('group', self.group)
|
|
2213
|
+
self.rank_id = get_rank(_get_group(self.group))
|
|
2214
|
+
self.add_prim_attr('rank_id', self.rank_id)
|