mindspore 2.6.0__cp39-cp39-win_amd64.whl → 2.7.0rc1__cp39-cp39-win_amd64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of mindspore might be problematic. Click here for more details.
- mindspore/.commit_id +1 -1
- mindspore/__init__.py +1 -1
- mindspore/_c_dataengine.cp39-win_amd64.pyd +0 -0
- mindspore/_c_expression.cp39-win_amd64.pyd +0 -0
- mindspore/_c_mindrecord.cp39-win_amd64.pyd +0 -0
- mindspore/_checkparam.py +40 -9
- mindspore/{_deprecated → _extends/optimize}/__init__.py +9 -3
- mindspore/_extends/optimize/cell_utils.py +96 -0
- mindspore/_extends/parse/__init__.py +2 -2
- mindspore/_extends/parse/compile_config.py +44 -22
- mindspore/_extends/parse/deprecated/deprecated_tensor_method.py +1 -1
- mindspore/_extends/parse/parser.py +36 -61
- mindspore/_extends/parse/resources.py +39 -0
- mindspore/_extends/parse/standard_method.py +32 -13
- mindspore/_extends/parse/trope.py +8 -1
- mindspore/_extends/pijit/__init__.py +1 -2
- mindspore/amp.py +4 -4
- mindspore/avcodec-59.dll +0 -0
- mindspore/avdevice-59.dll +0 -0
- mindspore/avfilter-8.dll +0 -0
- mindspore/avformat-59.dll +0 -0
- mindspore/avutil-57.dll +0 -0
- mindspore/boost/adasum.py +1 -1
- mindspore/boost/boost_cell_wrapper.py +4 -4
- mindspore/common/__init__.py +27 -2
- mindspore/common/_grad_function.py +2 -1
- mindspore/common/_pijit_context.py +28 -7
- mindspore/common/_stub_tensor.py +1 -209
- mindspore/common/_tensor_cpp_method.py +1 -1
- mindspore/common/_tensor_docs.py +76 -15
- mindspore/common/api.py +193 -112
- mindspore/common/dtype.py +21 -11
- mindspore/common/dump.py +10 -15
- mindspore/common/generator.py +2 -3
- mindspore/common/hook_handle.py +11 -2
- mindspore/common/jit_config.py +1 -1
- mindspore/common/jit_trace.py +84 -105
- mindspore/common/parameter.py +26 -12
- mindspore/common/recompute.py +3 -3
- mindspore/common/sparse_tensor.py +0 -3
- mindspore/common/symbol.py +0 -1
- mindspore/common/tensor.py +48 -83
- mindspore/communication/_comm_helper.py +46 -4
- mindspore/communication/management.py +79 -7
- mindspore/context.py +38 -23
- mindspore/dataset/core/config.py +3 -3
- mindspore/dataset/engine/datasets.py +20 -7
- mindspore/dataset/engine/datasets_user_defined.py +32 -2
- mindspore/dataset/engine/iterators.py +2 -2
- mindspore/dataset/engine/obs/config_loader.py +2 -2
- mindspore/dataset/engine/obs/obs_mindrecord_dataset.py +8 -0
- mindspore/dataset/transforms/py_transforms.py +7 -3
- mindspore/dataset/transforms/transforms.py +7 -3
- mindspore/dataset/vision/validators.py +1 -0
- mindspore/device_context/ascend/device.py +1 -1
- mindspore/device_context/gpu/__init__.py +2 -2
- mindspore/device_context/gpu/device.py +1 -1
- mindspore/device_context/gpu/op_precision.py +4 -2
- mindspore/device_context/gpu/op_tuning.py +6 -3
- mindspore/device_manager.py +16 -9
- mindspore/dnnl.dll +0 -0
- mindspore/experimental/llm_boost/ascend_native/llama_boost_ascend_native.py +3 -5
- mindspore/experimental/llm_boost/atb/boost_base.py +2 -3
- mindspore/experimental/optim/adadelta.py +13 -20
- mindspore/experimental/optim/adagrad.py +15 -22
- mindspore/experimental/optim/adam.py +17 -24
- mindspore/experimental/optim/adamax.py +14 -22
- mindspore/experimental/optim/adamw.py +28 -34
- mindspore/experimental/optim/asgd.py +15 -25
- mindspore/experimental/optim/lr_scheduler.py +27 -45
- mindspore/experimental/optim/nadam.py +14 -24
- mindspore/experimental/optim/optimizer.py +13 -23
- mindspore/experimental/optim/radam.py +18 -24
- mindspore/experimental/optim/rmsprop.py +14 -25
- mindspore/experimental/optim/rprop.py +15 -26
- mindspore/experimental/optim/sgd.py +9 -19
- mindspore/hal/__init__.py +4 -4
- mindspore/hal/contiguous_tensors_handle.py +2 -2
- mindspore/hal/memory.py +1 -0
- mindspore/include/api/cell.h +37 -1
- mindspore/include/api/delegate.h +10 -0
- mindspore/include/api/model.h +3 -0
- mindspore/include/api/types.h +2 -2
- mindspore/include/c_api/model_c.h +0 -58
- mindspore/include/c_api/tensor_c.h +0 -26
- mindspore/include/dataset/vision_ascend.h +1 -1
- mindspore/jpeg62.dll +0 -0
- mindspore/mindrecord/tools/cifar10.py +60 -11
- mindspore/mindrecord/tools/cifar10_to_mr.py +5 -0
- mindspore/mindspore_backend_common.dll +0 -0
- mindspore/mindspore_backend_manager.dll +0 -0
- mindspore/mindspore_common.dll +0 -0
- mindspore/mindspore_core.dll +0 -0
- mindspore/mindspore_cpu_res_manager.dll +0 -0
- mindspore/mindspore_dump.dll +0 -0
- mindspore/mindspore_frontend.dll +0 -0
- mindspore/mindspore_glog.dll +0 -0
- mindspore/mindspore_memory_pool.dll +0 -0
- mindspore/mindspore_ms_backend.dll +0 -0
- mindspore/mindspore_ops.dll +0 -0
- mindspore/mindspore_ops_host.dll +0 -0
- mindspore/mindspore_ops_kernel_common.dll +0 -0
- mindspore/mindspore_profiler.dll +0 -0
- mindspore/mindspore_pyboost.dll +0 -0
- mindspore/mindspore_pynative.dll +0 -0
- mindspore/mindspore_res_manager.dll +0 -0
- mindspore/mindspore_runtime_pipeline.dll +0 -0
- mindspore/mint/__init__.py +4 -44
- mindspore/mint/distributed/__init__.py +1 -0
- mindspore/mint/distributed/distributed.py +208 -5
- mindspore/mint/nn/__init__.py +1 -1
- mindspore/mint/nn/functional.py +53 -6
- mindspore/mint/nn/layer/_functions.py +164 -294
- mindspore/mint/nn/layer/activation.py +8 -6
- mindspore/mint/nn/layer/conv.py +122 -98
- mindspore/mint/nn/layer/normalization.py +8 -22
- mindspore/mint/optim/adam.py +19 -18
- mindspore/mint/optim/adamw.py +14 -8
- mindspore/mint/optim/sgd.py +5 -5
- mindspore/nn/cell.py +325 -499
- mindspore/nn/grad/cell_grad.py +11 -12
- mindspore/nn/layer/activation.py +32 -34
- mindspore/nn/layer/basic.py +67 -64
- mindspore/nn/layer/channel_shuffle.py +4 -4
- mindspore/nn/layer/combined.py +4 -2
- mindspore/nn/layer/conv.py +86 -85
- mindspore/nn/layer/dense.py +9 -7
- mindspore/nn/layer/embedding.py +50 -52
- mindspore/nn/layer/image.py +37 -39
- mindspore/nn/layer/math.py +111 -112
- mindspore/nn/layer/normalization.py +56 -44
- mindspore/nn/layer/pooling.py +58 -63
- mindspore/nn/layer/rnn_cells.py +33 -33
- mindspore/nn/layer/rnns.py +56 -56
- mindspore/nn/layer/thor_layer.py +74 -73
- mindspore/nn/layer/transformer.py +11 -1
- mindspore/nn/learning_rate_schedule.py +20 -20
- mindspore/nn/loss/loss.py +79 -81
- mindspore/nn/optim/adam.py +1 -1
- mindspore/nn/optim/adasum.py +2 -2
- mindspore/nn/optim/optimizer.py +1 -1
- mindspore/nn/optim/thor.py +2 -2
- mindspore/nn/probability/distribution/exponential.py +2 -1
- mindspore/nn/probability/distribution/poisson.py +2 -1
- mindspore/nn/sparse/sparse.py +3 -3
- mindspore/nn/wrap/cell_wrapper.py +34 -37
- mindspore/nn/wrap/grad_reducer.py +37 -37
- mindspore/nn/wrap/loss_scale.py +72 -74
- mindspore/numpy/array_creations.py +5 -5
- mindspore/numpy/fft.py +1 -1
- mindspore/numpy/math_ops.py +1 -1
- mindspore/opencv_core452.dll +0 -0
- mindspore/opencv_imgcodecs452.dll +0 -0
- mindspore/opencv_imgproc452.dll +0 -0
- mindspore/ops/_grad_experimental/grad_comm_ops.py +51 -13
- mindspore/ops/_grad_experimental/grad_debug_ops.py +14 -0
- mindspore/ops/_vmap/vmap_array_ops.py +6 -13
- mindspore/ops/_vmap/vmap_nn_ops.py +8 -16
- mindspore/ops/auto_generate/cpp_create_prim_instance_helper.py +17 -8
- mindspore/ops/auto_generate/gen_extend_func.py +1 -51
- mindspore/ops/auto_generate/gen_ops_def.py +463 -257
- mindspore/ops/auto_generate/gen_ops_prim.py +1127 -885
- mindspore/ops/auto_generate/pyboost_inner_prim.py +31 -1
- mindspore/ops/composite/__init__.py +10 -0
- mindspore/ops/composite/base.py +8 -4
- mindspore/ops/composite/multitype_ops/__init__.py +12 -1
- mindspore/ops/composite/multitype_ops/_compile_utils.py +132 -108
- mindspore/ops/composite/multitype_ops/add_impl.py +70 -2
- mindspore/ops/composite/multitype_ops/div_impl.py +49 -0
- mindspore/ops/composite/multitype_ops/floordiv_impl.py +29 -0
- mindspore/ops/composite/multitype_ops/getitem_impl.py +11 -0
- mindspore/ops/composite/multitype_ops/mod_impl.py +5 -3
- mindspore/ops/composite/multitype_ops/mul_impl.py +49 -0
- mindspore/ops/composite/multitype_ops/setitem_impl.py +57 -0
- mindspore/ops/composite/multitype_ops/sub_impl.py +34 -0
- mindspore/ops/composite/multitype_ops/zeros_like_impl.py +14 -0
- mindspore/ops/function/__init__.py +3 -1
- mindspore/ops/function/_add_attr_func.py +11 -6
- mindspore/ops/function/array_func.py +7 -94
- mindspore/ops/function/debug_func.py +4 -3
- mindspore/ops/function/grad/grad_func.py +1 -1
- mindspore/ops/function/math_func.py +21 -367
- mindspore/ops/function/nn_func.py +26 -41
- mindspore/ops/function/other_func.py +4 -1
- mindspore/ops/function/random_func.py +31 -4
- mindspore/ops/functional.py +0 -2
- mindspore/ops/functional_overload.py +463 -6
- mindspore/ops/op_info_register.py +21 -0
- mindspore/ops/operations/__init__.py +5 -2
- mindspore/ops/operations/_custom_ops_utils.py +675 -8
- mindspore/ops/operations/_inner_ops.py +3 -6
- mindspore/ops/operations/_sequence_ops.py +1 -1
- mindspore/ops/operations/comm_ops.py +185 -26
- mindspore/ops/operations/custom_ops.py +235 -172
- mindspore/ops/operations/debug_ops.py +55 -4
- mindspore/ops/operations/image_ops.py +13 -13
- mindspore/ops/operations/manually_defined/ops_def.py +15 -16
- mindspore/ops/operations/math_ops.py +3 -4
- mindspore/ops/operations/nn_ops.py +5 -6
- mindspore/ops/primitive.py +6 -10
- mindspore/ops/tensor_method.py +36 -4
- mindspore/ops_generate/api/cpp_create_prim_instance_helper_generator.py +1 -1
- mindspore/ops_generate/api/functional_map_cpp_generator.py +10 -9
- mindspore/ops_generate/api/functions_cc_generator.py +58 -10
- mindspore/ops_generate/api/tensor_func_reg_cpp_generator.py +1 -1
- mindspore/ops_generate/common/base_generator.py +14 -0
- mindspore/ops_generate/common/gen_constants.py +7 -2
- mindspore/ops_generate/common/gen_utils.py +0 -19
- mindspore/ops_generate/common/op_proto.py +11 -4
- mindspore/ops_generate/common/template.py +88 -11
- mindspore/ops_generate/gen_ops.py +1 -1
- mindspore/ops_generate/op_def/lite_ops_cpp_generator.py +4 -4
- mindspore/ops_generate/op_def/ops_name_h_generator.py +0 -3
- mindspore/ops_generate/op_def/ops_primitive_h_generator.py +0 -4
- mindspore/ops_generate/op_def_py/op_prim_py_generator.py +5 -2
- mindspore/ops_generate/pyboost/auto_grad_impl_cc_generator.py +49 -8
- mindspore/ops_generate/pyboost/auto_grad_reg_cc_generator.py +2 -2
- mindspore/ops_generate/pyboost/gen_pyboost_func.py +31 -0
- mindspore/ops_generate/pyboost/op_template_parser.py +98 -72
- mindspore/ops_generate/pyboost/pyboost_functions_cpp_generator.py +70 -273
- mindspore/ops_generate/pyboost/pyboost_functions_h_generator.py +14 -6
- mindspore/ops_generate/pyboost/pyboost_functions_impl_cpp_generator.py +316 -0
- mindspore/ops_generate/pyboost/pyboost_functions_py_generator.py +1 -1
- mindspore/ops_generate/pyboost/pyboost_grad_function_cpp_generator.py +5 -3
- mindspore/ops_generate/pyboost/pyboost_inner_prim_generator.py +1 -1
- mindspore/ops_generate/pyboost/pyboost_internal_functions_cpp_generator.py +76 -0
- mindspore/ops_generate/pyboost/pyboost_internal_functions_h_generator.py +76 -0
- mindspore/ops_generate/pyboost/pyboost_internal_kernel_info_adapter_generator.py +125 -0
- mindspore/ops_generate/pyboost/pyboost_native_grad_functions_generator.py +4 -3
- mindspore/ops_generate/pyboost/pyboost_op_cpp_code_generator.py +348 -61
- mindspore/ops_generate/pyboost/pyboost_overload_functions_cpp_generator.py +1 -1
- mindspore/ops_generate/pyboost/pyboost_utils.py +118 -9
- mindspore/ops_generate/tensor_py_cc_generator.py +1 -24
- mindspore/parallel/_auto_parallel_context.py +4 -2
- mindspore/parallel/_cell_wrapper.py +106 -40
- mindspore/parallel/_parallel_serialization.py +1 -1
- mindspore/parallel/_ps_context.py +4 -6
- mindspore/parallel/_tensor.py +167 -12
- mindspore/parallel/_transformer/moe.py +1 -1
- mindspore/parallel/_transformer/transformer.py +13 -8
- mindspore/parallel/auto_parallel.py +12 -5
- mindspore/parallel/checkpoint_convert.py +3 -3
- mindspore/parallel/checkpoint_transform.py +3 -1
- mindspore/parallel/cluster/process_entity/_api.py +84 -48
- mindspore/parallel/cluster/process_entity/_utils.py +95 -7
- mindspore/parallel/cluster/run.py +43 -4
- mindspore/parallel/function/__init__.py +8 -1
- mindspore/parallel/function/reshard_func.py +1 -1
- mindspore/parallel/nn/__init__.py +15 -2
- mindspore/parallel/nn/parallel_cell_wrapper.py +9 -10
- mindspore/parallel/nn/parallel_grad_reducer.py +7 -6
- mindspore/parallel/shard.py +2 -2
- mindspore/parallel/transform_safetensors.py +462 -174
- mindspore/profiler/__init__.py +2 -1
- mindspore/profiler/analysis/parser/timeline_assembly_factory/ascend_timeline_assembler.py +7 -7
- mindspore/profiler/analysis/parser/timeline_assembly_factory/base_timeline_assembler.py +3 -0
- mindspore/profiler/analysis/parser/timeline_assembly_factory/trace_view_container.py +3 -0
- mindspore/profiler/analysis/parser/timeline_creator/cpu_op_timeline_creator.py +3 -3
- mindspore/profiler/analysis/parser/timeline_creator/fwk_timeline_creator.py +3 -3
- mindspore/profiler/analysis/parser/timeline_creator/msprof_timeline_creator.py +4 -4
- mindspore/profiler/analysis/parser/timeline_creator/scope_layer_timeline_creator.py +3 -3
- mindspore/profiler/analysis/parser/timeline_event/fwk_event.py +4 -1
- mindspore/profiler/analysis/parser/timeline_event/timeline_event_pool.py +2 -1
- mindspore/profiler/analysis/task_manager.py +1 -1
- mindspore/profiler/analysis/viewer/ascend_communication_viewer.py +5 -1
- mindspore/profiler/analysis/viewer/ascend_integrate_viewer.py +2 -1
- mindspore/profiler/analysis/viewer/ascend_op_memory_viewer.py +42 -22
- mindspore/profiler/analysis/viewer/ascend_step_trace_time_viewer.py +3 -2
- mindspore/profiler/analysis/viewer/ms_minddata_viewer.py +9 -5
- mindspore/profiler/analysis/viewer/ms_operator_details_viewer.py +132 -0
- mindspore/profiler/common/constant.py +16 -0
- mindspore/profiler/common/profiler_context.py +25 -27
- mindspore/profiler/common/profiler_info.py +0 -16
- mindspore/profiler/common/profiler_op_analyse.py +235 -0
- mindspore/profiler/common/profiler_output_path.py +23 -8
- mindspore/profiler/common/profiler_parameters.py +128 -35
- mindspore/profiler/dynamic_profile/__init__.py +0 -0
- mindspore/profiler/dynamic_profile/dynamic_monitor_proxy.py +39 -0
- mindspore/profiler/dynamic_profile/dynamic_profiler_config_context.py +666 -0
- mindspore/profiler/dynamic_profile/dynamic_profiler_utils.py +62 -0
- mindspore/profiler/dynamic_profiler.py +305 -314
- mindspore/profiler/envprofiler.py +12 -7
- mindspore/profiler/experimental_config.py +96 -6
- mindspore/profiler/mstx.py +33 -12
- mindspore/profiler/platform/__init__.py +2 -3
- mindspore/profiler/platform/npu_profiler.py +29 -19
- mindspore/profiler/profiler.py +35 -19
- mindspore/profiler/profiler_action_controller.py +64 -76
- mindspore/profiler/schedule.py +10 -4
- mindspore/rewrite/common/config.py +1 -0
- mindspore/rewrite/common/namer.py +1 -0
- mindspore/rewrite/common/namespace.py +1 -0
- mindspore/rewrite/node/node.py +31 -11
- mindspore/rewrite/parsers/assign_parser.py +1 -1
- mindspore/rewrite/symbol_tree/symbol_tree.py +1 -1
- mindspore/run_check/_check_version.py +7 -10
- mindspore/runtime/__init__.py +5 -5
- mindspore/runtime/event.py +10 -4
- mindspore/runtime/executor.py +60 -45
- mindspore/runtime/memory.py +21 -30
- mindspore/runtime/thread_bind_core.py +298 -164
- mindspore/safeguard/rewrite_obfuscation.py +12 -13
- mindspore/swresample-4.dll +0 -0
- mindspore/swscale-6.dll +0 -0
- mindspore/tinyxml2.dll +0 -0
- mindspore/train/_utils.py +6 -2
- mindspore/train/amp.py +43 -20
- mindspore/train/callback/__init__.py +5 -5
- mindspore/train/callback/_checkpoint.py +3 -6
- mindspore/train/callback/_flops_collector.py +1 -1
- mindspore/train/callback/_landscape.py +0 -1
- mindspore/train/callback/_train_fault_tolerance.py +71 -13
- mindspore/train/data_sink.py +11 -2
- mindspore/train/dataset_helper.py +9 -0
- mindspore/train/model.py +51 -33
- mindspore/train/serialization.py +133 -111
- mindspore/train/summary/summary_record.py +13 -2
- mindspore/turbojpeg.dll +0 -0
- mindspore/utils/__init__.py +3 -2
- mindspore/utils/dryrun.py +0 -6
- mindspore/utils/runtime_execution_order_check.py +162 -78
- mindspore/utils/sdc_detect.py +68 -0
- mindspore/utils/utils.py +6 -9
- mindspore/version.py +1 -1
- {mindspore-2.6.0.dist-info → mindspore-2.7.0rc1.dist-info}/METADATA +5 -4
- {mindspore-2.6.0.dist-info → mindspore-2.7.0rc1.dist-info}/RECORD +329 -367
- mindspore/_deprecated/jit.py +0 -198
- mindspore/experimental/es/__init__.py +0 -22
- mindspore/experimental/es/embedding_service.py +0 -891
- mindspore/experimental/es/embedding_service_layer.py +0 -581
- mindspore/profiler/parser/__init__.py +0 -14
- mindspore/profiler/parser/aicpu_data_parser.py +0 -272
- mindspore/profiler/parser/ascend_analysis/__init__.py +0 -14
- mindspore/profiler/parser/ascend_analysis/constant.py +0 -71
- mindspore/profiler/parser/ascend_analysis/file_manager.py +0 -180
- mindspore/profiler/parser/ascend_analysis/function_event.py +0 -185
- mindspore/profiler/parser/ascend_analysis/fwk_cann_parser.py +0 -136
- mindspore/profiler/parser/ascend_analysis/fwk_file_parser.py +0 -131
- mindspore/profiler/parser/ascend_analysis/msprof_timeline_parser.py +0 -104
- mindspore/profiler/parser/ascend_analysis/path_manager.py +0 -313
- mindspore/profiler/parser/ascend_analysis/profiler_info_parser.py +0 -123
- mindspore/profiler/parser/ascend_analysis/tlv_decoder.py +0 -86
- mindspore/profiler/parser/ascend_analysis/trace_event_manager.py +0 -75
- mindspore/profiler/parser/ascend_cluster_generator.py +0 -116
- mindspore/profiler/parser/ascend_communicate_generator.py +0 -314
- mindspore/profiler/parser/ascend_flops_generator.py +0 -116
- mindspore/profiler/parser/ascend_fpbp_generator.py +0 -82
- mindspore/profiler/parser/ascend_hccl_generator.py +0 -271
- mindspore/profiler/parser/ascend_integrate_generator.py +0 -42
- mindspore/profiler/parser/ascend_memory_generator.py +0 -185
- mindspore/profiler/parser/ascend_msprof_exporter.py +0 -282
- mindspore/profiler/parser/ascend_msprof_generator.py +0 -187
- mindspore/profiler/parser/ascend_op_generator.py +0 -334
- mindspore/profiler/parser/ascend_steptrace_generator.py +0 -94
- mindspore/profiler/parser/ascend_timeline_generator.py +0 -545
- mindspore/profiler/parser/base_timeline_generator.py +0 -483
- mindspore/profiler/parser/container.py +0 -229
- mindspore/profiler/parser/cpu_gpu_timeline_generator.py +0 -697
- mindspore/profiler/parser/flops_parser.py +0 -531
- mindspore/profiler/parser/framework_enum.py +0 -111
- mindspore/profiler/parser/framework_parser.py +0 -464
- mindspore/profiler/parser/framework_struct.py +0 -61
- mindspore/profiler/parser/gpu_analysis/__init__.py +0 -14
- mindspore/profiler/parser/gpu_analysis/function_event.py +0 -44
- mindspore/profiler/parser/gpu_analysis/fwk_file_parser.py +0 -89
- mindspore/profiler/parser/gpu_analysis/profiler_info_parser.py +0 -72
- mindspore/profiler/parser/hccl_parser.py +0 -573
- mindspore/profiler/parser/hwts_log_parser.py +0 -122
- mindspore/profiler/parser/integrator.py +0 -526
- mindspore/profiler/parser/memory_usage_parser.py +0 -277
- mindspore/profiler/parser/minddata_analyzer.py +0 -800
- mindspore/profiler/parser/minddata_parser.py +0 -186
- mindspore/profiler/parser/minddata_pipeline_parser.py +0 -299
- mindspore/profiler/parser/op_intermediate_parser.py +0 -149
- mindspore/profiler/parser/optime_parser.py +0 -250
- mindspore/profiler/parser/profiler_info.py +0 -213
- mindspore/profiler/parser/step_trace_parser.py +0 -666
- {mindspore-2.6.0.dist-info → mindspore-2.7.0rc1.dist-info}/WHEEL +0 -0
- {mindspore-2.6.0.dist-info → mindspore-2.7.0rc1.dist-info}/entry_points.txt +0 -0
- {mindspore-2.6.0.dist-info → mindspore-2.7.0rc1.dist-info}/top_level.txt +0 -0
mindspore/mint/__init__.py
CHANGED
|
@@ -33,7 +33,7 @@ from mindspore.ops.function.array_func import zeros_like_ext as zeros_like
|
|
|
33
33
|
from mindspore.ops.function.array_func import unique_ext as unique
|
|
34
34
|
from mindspore.ops.function.array_func import chunk_ext as chunk
|
|
35
35
|
from mindspore.ops.functional_overload import empty
|
|
36
|
-
from mindspore.ops.
|
|
36
|
+
from mindspore.ops.functional_overload import empty_like
|
|
37
37
|
from mindspore.ops.function.math_func import isclose
|
|
38
38
|
from mindspore.ops.auto_generate import abs
|
|
39
39
|
from mindspore.ops.auto_generate import clone
|
|
@@ -291,7 +291,7 @@ from mindspore.ops.function.nn_func import softmax_ext
|
|
|
291
291
|
# 118
|
|
292
292
|
|
|
293
293
|
# 119
|
|
294
|
-
|
|
294
|
+
from mindspore.ops.functional_overload import any
|
|
295
295
|
# 120
|
|
296
296
|
from mindspore.ops.auto_generate import isneginf_ext as isneginf
|
|
297
297
|
# 121
|
|
@@ -478,7 +478,7 @@ from mindspore.ops.auto_generate import log2_ext as log2
|
|
|
478
478
|
from mindspore.ops.function.math_func import isnan_ext as isnan
|
|
479
479
|
|
|
480
480
|
# 916
|
|
481
|
-
from mindspore.ops.
|
|
481
|
+
from mindspore.ops.functional_overload import index_add
|
|
482
482
|
|
|
483
483
|
# 1007
|
|
484
484
|
from mindspore.ops.auto_generate import t_ext as t
|
|
@@ -490,46 +490,6 @@ from mindspore.ops.auto_generate.gen_ops_prim import equal_ext_op
|
|
|
490
490
|
from mindspore.ops.function.array_func import unbind_ext as unbind
|
|
491
491
|
|
|
492
492
|
|
|
493
|
-
def any(input, dim=None, keepdim=False):
|
|
494
|
-
r"""
|
|
495
|
-
Tests if any element in `input` evaluates to `True` along the given axes.
|
|
496
|
-
|
|
497
|
-
Args:
|
|
498
|
-
input (Tensor): The input tensor.
|
|
499
|
-
dim (Union[int, tuple(int), list(int), Tensor], optional): The dimensions to reduce. If ``None`` ,
|
|
500
|
-
all dimensions are reduced. Default ``None`` .
|
|
501
|
-
keepdim (bool, optional): Whether the output tensor has dim retained or not. Default ``False`` .
|
|
502
|
-
|
|
503
|
-
Returns:
|
|
504
|
-
Tensor
|
|
505
|
-
|
|
506
|
-
Supported Platforms:
|
|
507
|
-
``Ascend`` ``GPU`` ``CPU``
|
|
508
|
-
|
|
509
|
-
Examples:
|
|
510
|
-
>>> import mindspore
|
|
511
|
-
>>> input = mindspore.tensor([[True, False], [True, True]])
|
|
512
|
-
>>>
|
|
513
|
-
>>> # case 1: By default, mindspore.mint.any tests along all the axes.
|
|
514
|
-
>>> mindspore.mint.any(input)
|
|
515
|
-
Tensor(shape=[], dtype=Bool, value= True)
|
|
516
|
-
>>>
|
|
517
|
-
>>> # case 2: Reduces a dimension along dim 1, with keepdim False.
|
|
518
|
-
>>> mindspore.mint.any(input, dim=1)
|
|
519
|
-
Tensor(shape=[2], dtype=Bool, value= [ True, True])
|
|
520
|
-
>>>
|
|
521
|
-
>>> # case 3: Reduces a dimension along dim (0, 1), with keepdim False.
|
|
522
|
-
>>> mindspore.mint.any(input, dim=(0,1))
|
|
523
|
-
Tensor(shape=[], dtype=Bool, value= True)
|
|
524
|
-
>>>
|
|
525
|
-
>>> # case 4: Reduces a dimension along dim [0, 1], with keepdim True.
|
|
526
|
-
>>> mindspore.mint.any(input, dim=[0,1], keepdim=True)
|
|
527
|
-
Tensor(shape=[1, 1], dtype=Bool, value=
|
|
528
|
-
[[ True]])
|
|
529
|
-
"""
|
|
530
|
-
return ops.functional.any(input, dim, keepdim)
|
|
531
|
-
|
|
532
|
-
|
|
533
493
|
def all(input, dim=None, keepdim=False):
|
|
534
494
|
r"""
|
|
535
495
|
all(input) -> Tensor
|
|
@@ -717,7 +677,7 @@ def cummax(input, dim):
|
|
|
717
677
|
\end{array}
|
|
718
678
|
|
|
719
679
|
.. note::
|
|
720
|
-
|
|
680
|
+
GE backend is not supported in Ascend.
|
|
721
681
|
|
|
722
682
|
Args:
|
|
723
683
|
input (Tensor): The input Tensor. Rank of `input` must be greater than 0.
|
|
@@ -24,6 +24,7 @@ Please see the `msrun start up
|
|
|
24
24
|
for more details.
|
|
25
25
|
"""
|
|
26
26
|
from __future__ import absolute_import
|
|
27
|
+
from mindspore.mint.distributed.distributed import TCPStore
|
|
27
28
|
from mindspore.mint.distributed.distributed import (
|
|
28
29
|
init_process_group,
|
|
29
30
|
destroy_process_group,
|
|
@@ -70,6 +70,7 @@ from mindspore.ops.auto_generate.gen_ops_prim import (
|
|
|
70
70
|
dist_comm_barrier_op,
|
|
71
71
|
dist_comm_batch_isend_irecv_op,
|
|
72
72
|
)
|
|
73
|
+
from mindspore._c_expression import TCPStoreClient, GroupOptions
|
|
73
74
|
|
|
74
75
|
_pickler = pickle.Pickler
|
|
75
76
|
_unpickler = pickle.Unpickler
|
|
@@ -137,6 +138,193 @@ def _tensor_to_object(tensor, tensor_size):
|
|
|
137
138
|
return restricted_loads(buf)
|
|
138
139
|
|
|
139
140
|
|
|
141
|
+
class TCPStore:
|
|
142
|
+
"""
|
|
143
|
+
A TCP-based distributed key-value store implementation.
|
|
144
|
+
|
|
145
|
+
Note:
|
|
146
|
+
- The function is implemented by CPU and does not involve any hardware operations related to Ascend.
|
|
147
|
+
- Currently, all parameters provided by the TCPStore class constructor are not supported.
|
|
148
|
+
The master node and port number are uniformly specified by the MindSpore framework.
|
|
149
|
+
The following parameters are provided, currently not supported and settings are invalid.
|
|
150
|
+
- The current TcpStore function is limited and only supports scenarios where the key is
|
|
151
|
+
less than 4k and the value is less than 1G. Complex scenarios are to be supported.
|
|
152
|
+
- The timeout interval for message sending and receiving in the TcpStore function is controlled by
|
|
153
|
+
the `MS_RECEIVE_MSG_TIMEOUT` environment variable, in seconds, with a default value of ``15``.
|
|
154
|
+
If a timeout occurs, the user needs to increase the configuration value.
|
|
155
|
+
|
|
156
|
+
Args:
|
|
157
|
+
host_name (str, invalid, optional): The hostname or IP Address the server store should run on.
|
|
158
|
+
Default is ``None``.
|
|
159
|
+
port (int, invalid, optional): The port on which the server store should listen for incoming requests.
|
|
160
|
+
Default is ``None``.
|
|
161
|
+
world_size (int, invalid, optional): The total number of store users (number of clients + 1 for the server).
|
|
162
|
+
Default is ``None`` (``None`` indicates a non-fixed number of store users).
|
|
163
|
+
is_master (bool, invalid, optional): True when initializing the server store and False for client stores.
|
|
164
|
+
Default is ``False``.
|
|
165
|
+
timeout (timedelta, invalid, optional): Timeout used by the store during initialization, Unit: seconds.
|
|
166
|
+
Default is ``300``.
|
|
167
|
+
wait_for_workers (bool, invalid, optional): Whether to wait for all the workers to connect with the server
|
|
168
|
+
store. This is only applicable when `world_size` is a fixed value. Default is ``True``.
|
|
169
|
+
multi_tenant (bool, invalid, optional): If ``True``, all ``TCPStore`` instances in the current process with
|
|
170
|
+
the same host/port will use the same underlying ``TCPServer``. Default is ``False``.
|
|
171
|
+
master_listen_fd (int, invalid, optional): If specified, the underlying ``TCPServer`` will listen on this file
|
|
172
|
+
descriptor, which must be a socket already bound to ``port``. Useful to avoid port assignment races
|
|
173
|
+
in some scenarios. Default is ``None`` (meaning the server creates a new socket and attempts to bind it
|
|
174
|
+
to ``port``).
|
|
175
|
+
use_libuv (bool, invalid, optional): If True, use libuv for ``TCPServer`` backend. Default is ``True``.
|
|
176
|
+
|
|
177
|
+
Returns:
|
|
178
|
+
TCPStore Object.
|
|
179
|
+
|
|
180
|
+
Supported Platforms:
|
|
181
|
+
``Ascend``
|
|
182
|
+
|
|
183
|
+
Examples:
|
|
184
|
+
.. note::
|
|
185
|
+
Before running the following examples, you need to configure the communication environment variables.
|
|
186
|
+
|
|
187
|
+
For Ascend devices, it is recommended to use the msrun startup method
|
|
188
|
+
without any third-party or configuration file dependencies.
|
|
189
|
+
Please see the `msrun start up
|
|
190
|
+
<https://www.mindspore.cn/tutorials/en/master/parallel/msrun_launcher.html>`_
|
|
191
|
+
for more details.
|
|
192
|
+
|
|
193
|
+
>>> from mindspore.mint.distributed import TCPStore
|
|
194
|
+
>>> store = TCPStore()
|
|
195
|
+
"""
|
|
196
|
+
|
|
197
|
+
def __init__(self, host_name=None, port=None, world_size=None, is_master=False, timeout=300,
|
|
198
|
+
wait_for_workers=True, multi_tenant=False, master_listen_fd=None, use_libuv=True):
|
|
199
|
+
self.instance = TCPStoreClient.get_instance()
|
|
200
|
+
|
|
201
|
+
|
|
202
|
+
def set(self, key, value):
|
|
203
|
+
"""
|
|
204
|
+
Inserts the key-value pair into the store based on the supplied `key` and
|
|
205
|
+
`value`. If `key` already exists in the store, it will overwrite the old
|
|
206
|
+
value with the new supplied `value`.
|
|
207
|
+
|
|
208
|
+
Args:
|
|
209
|
+
key (str): The key to be added to the store.
|
|
210
|
+
value (Union[bytes, str]): The value associated with `key` to be added to the store.
|
|
211
|
+
|
|
212
|
+
Raises:
|
|
213
|
+
TypeError: If `key` is not string.
|
|
214
|
+
TypeError: If `value` is not string or bytes.
|
|
215
|
+
|
|
216
|
+
Supported Platforms:
|
|
217
|
+
``Ascend``
|
|
218
|
+
|
|
219
|
+
Examples:
|
|
220
|
+
.. note::
|
|
221
|
+
Before running the following examples, you need to configure the communication environment variables.
|
|
222
|
+
|
|
223
|
+
For Ascend devices, it is recommended to use the msrun startup method
|
|
224
|
+
without any third-party or configuration file dependencies.
|
|
225
|
+
Please see the `msrun start up
|
|
226
|
+
<https://www.mindspore.cn/tutorials/en/master/parallel/msrun_launcher.html>`_
|
|
227
|
+
for more details.
|
|
228
|
+
|
|
229
|
+
>>> from mindspore.mint.distributed import TCPStore
|
|
230
|
+
>>> store = TCPStore()
|
|
231
|
+
>>> store.set("first_key", "first_value")
|
|
232
|
+
"""
|
|
233
|
+
if not isinstance(key, str):
|
|
234
|
+
raise TypeError(
|
|
235
|
+
"For 'TCPStore.set', the argument 'key' must be type of string, "
|
|
236
|
+
"but got 'key' type : {}.".format(type(key))
|
|
237
|
+
)
|
|
238
|
+
if not isinstance(value, (str, bytes)):
|
|
239
|
+
raise TypeError(
|
|
240
|
+
"For 'TCPStore.set', the argument 'value' must be type of string or bytes, "
|
|
241
|
+
"but got 'value' type : {}.".format(type(value))
|
|
242
|
+
)
|
|
243
|
+
return self.instance.set(key, value)
|
|
244
|
+
|
|
245
|
+
|
|
246
|
+
def get(self, key):
|
|
247
|
+
"""
|
|
248
|
+
Retrieves the value associated with the given `key` in the store. If `key` is not
|
|
249
|
+
present in the store, the function will return "".
|
|
250
|
+
|
|
251
|
+
Args:
|
|
252
|
+
key (str): The function will return the value associated with this key.
|
|
253
|
+
|
|
254
|
+
Returns:
|
|
255
|
+
bytes, Value associated with `key` if `key` is in the store.
|
|
256
|
+
|
|
257
|
+
Raises:
|
|
258
|
+
TypeError: If `key` is not string.
|
|
259
|
+
|
|
260
|
+
Supported Platforms:
|
|
261
|
+
``Ascend``
|
|
262
|
+
|
|
263
|
+
Examples:
|
|
264
|
+
.. note::
|
|
265
|
+
Before running the following examples, you need to configure the communication environment variables.
|
|
266
|
+
|
|
267
|
+
For Ascend devices, it is recommended to use the msrun startup method
|
|
268
|
+
without any third-party or configuration file dependencies.
|
|
269
|
+
Please see the `msrun start up
|
|
270
|
+
<https://www.mindspore.cn/tutorials/en/master/parallel/msrun_launcher.html>`_
|
|
271
|
+
for more details.
|
|
272
|
+
|
|
273
|
+
>>> from mindspore.mint.distributed import TCPStore
|
|
274
|
+
>>> store = TCPStore()
|
|
275
|
+
>>> store.set("first_key", "first_value")
|
|
276
|
+
>>> data = store.get("first_key")
|
|
277
|
+
>>> print(data)
|
|
278
|
+
"""
|
|
279
|
+
if not isinstance(key, str):
|
|
280
|
+
raise TypeError(
|
|
281
|
+
"For 'TCPStore.get', the argument 'key' must be type of string, "
|
|
282
|
+
"but got 'key' type : {}.".format(type(key))
|
|
283
|
+
)
|
|
284
|
+
byte_data = self.instance.get(key)
|
|
285
|
+
return byte_data
|
|
286
|
+
|
|
287
|
+
|
|
288
|
+
def delete_key(self, key):
|
|
289
|
+
"""
|
|
290
|
+
Deletes the key-value pair associated with `key` from the store.
|
|
291
|
+
|
|
292
|
+
Args:
|
|
293
|
+
key (str): The key to be deleted from the store.
|
|
294
|
+
|
|
295
|
+
Returns:
|
|
296
|
+
bool, ``True`` if `key` was deleted, otherwise ``False``.
|
|
297
|
+
|
|
298
|
+
Raises:
|
|
299
|
+
TypeError: If `key` is not string.
|
|
300
|
+
|
|
301
|
+
Supported Platforms:
|
|
302
|
+
``CPU``
|
|
303
|
+
|
|
304
|
+
Examples:
|
|
305
|
+
.. note::
|
|
306
|
+
Before running the following examples, you need to configure the communication environment variables.
|
|
307
|
+
|
|
308
|
+
For Ascend devices, it is recommended to use the msrun startup method
|
|
309
|
+
without any third-party or configuration file dependencies.
|
|
310
|
+
Please see the `msrun start up
|
|
311
|
+
<https://www.mindspore.cn/tutorials/en/master/parallel/msrun_launcher.html>`_
|
|
312
|
+
for more details.
|
|
313
|
+
|
|
314
|
+
>>> from mindspore.mint.distributed import TCPStore
|
|
315
|
+
>>> store = TCPStore()
|
|
316
|
+
>>> store.set("first_key", "first_value")
|
|
317
|
+
>>> # This should return true
|
|
318
|
+
>>> store.delete_key("first_key")
|
|
319
|
+
"""
|
|
320
|
+
if not isinstance(key, str):
|
|
321
|
+
raise TypeError(
|
|
322
|
+
"For 'TCPStore.delete_key', the argument 'key' must be type of string, "
|
|
323
|
+
"but got 'key' type : {}.".format(type(key))
|
|
324
|
+
)
|
|
325
|
+
return self.instance.delete_key(key)
|
|
326
|
+
|
|
327
|
+
|
|
140
328
|
def is_available():
|
|
141
329
|
"""
|
|
142
330
|
Checks if distributed module is available.
|
|
@@ -470,7 +658,19 @@ def new_group(ranks=None,
|
|
|
470
658
|
when backend is ``"hccl"`` will use Huawei Collective Communication Library(HCCL).
|
|
471
659
|
when backend is ``"mccl"`` will use MindSpore Collective Communication Library(MCCL).
|
|
472
660
|
If ``None``, which means ``"hccl"`` in Ascend. Default is ``None``.
|
|
473
|
-
pg_options (
|
|
661
|
+
pg_options (GroupOptions, optional): Additional communication group configuration parameters.
|
|
662
|
+
The backend will automatically select supported parameters and apply them during group
|
|
663
|
+
initialization. i.e. for the ``HCCL`` backend, ``hccl_config`` can be specified so that
|
|
664
|
+
group initialization configurations can be applied. Default is ``None``.
|
|
665
|
+
|
|
666
|
+
`GroupOptions` is defined as a class that can be instantiated as a python object.
|
|
667
|
+
|
|
668
|
+
.. code-block::
|
|
669
|
+
|
|
670
|
+
GroupOptions {
|
|
671
|
+
hccl_config(dict)
|
|
672
|
+
}
|
|
673
|
+
|
|
474
674
|
use_local_synchronization (bool, invalid): Currently it is a reserved parameter.
|
|
475
675
|
group_desc (str, invalid): Currently it is a reserved parameter.
|
|
476
676
|
|
|
@@ -511,8 +711,11 @@ def new_group(ranks=None,
|
|
|
511
711
|
if not isinstance(backend, str) or backend not in ("hccl", "mccl"):
|
|
512
712
|
raise TypeError(f"the input backend must be hccl or mccl, but got {backend}")
|
|
513
713
|
group = backend + "_" + str(len(ranks)) + "_" + hashlib.sha1(bytes("_".join(map(str, ranks)), "utf-8")).hexdigest()
|
|
714
|
+
if pg_options is not None:
|
|
715
|
+
if not isinstance(pg_options, GroupOptions):
|
|
716
|
+
raise TypeError("pg_options must be type GroupOptions, but got {}".format(type(pg_options)))
|
|
514
717
|
try:
|
|
515
|
-
create_group(group, ranks)
|
|
718
|
+
create_group(group, ranks, pg_options)
|
|
516
719
|
except RuntimeError as e:
|
|
517
720
|
logger.warning(e)
|
|
518
721
|
group = ""
|
|
@@ -2018,10 +2221,10 @@ def all_to_all(output_tensor_list, input_tensor_list, group=None, async_op=False
|
|
|
2018
2221
|
recv_shape_list = []
|
|
2019
2222
|
|
|
2020
2223
|
for tensor in input_tensor_list:
|
|
2021
|
-
send_numel_list.append(tensor.
|
|
2224
|
+
send_numel_list.append(tensor.numel())
|
|
2022
2225
|
send_flatten_tensor.append(tensor.reshape(-1))
|
|
2023
2226
|
for tensor in output_tensor_list:
|
|
2024
|
-
recv_numel_list.append(tensor.
|
|
2227
|
+
recv_numel_list.append(tensor.numel())
|
|
2025
2228
|
recv_shape_list.append(tensor.shape)
|
|
2026
2229
|
|
|
2027
2230
|
send_flatten_tensor = cat(send_flatten_tensor)
|
|
@@ -2810,7 +3013,7 @@ def broadcast_object_list(object_list, src=0, group=None, device=None):
|
|
|
2810
3013
|
offset = 0
|
|
2811
3014
|
for i, item in enumerate(object_size):
|
|
2812
3015
|
obj_size = item
|
|
2813
|
-
obj_view = object_tensor[offset
|
|
3016
|
+
obj_view = object_tensor[offset: offset + obj_size]
|
|
2814
3017
|
offset += obj_size
|
|
2815
3018
|
object_list[i] = _tensor_to_object(obj_view, obj_size)
|
|
2816
3019
|
|
mindspore/mint/nn/__init__.py
CHANGED
|
@@ -1189,7 +1189,7 @@ class PixelShuffle(Cell):
|
|
|
1189
1189
|
>>> pixel_shuffle = mint.nn.PixelShuffle(3)
|
|
1190
1190
|
>>> input = mint.randn(1, 9, 4, 4)
|
|
1191
1191
|
>>> output = pixel_shuffle(input)
|
|
1192
|
-
>>> print(output.shape
|
|
1192
|
+
>>> print(output.shape)
|
|
1193
1193
|
[1, 1, 12, 12]
|
|
1194
1194
|
"""
|
|
1195
1195
|
|
mindspore/mint/nn/functional.py
CHANGED
|
@@ -67,7 +67,7 @@ from mindspore.ops.auto_generate import prelu
|
|
|
67
67
|
# 20
|
|
68
68
|
|
|
69
69
|
# 21
|
|
70
|
-
from mindspore.ops.
|
|
70
|
+
from mindspore.ops.functional_overload import conv3d
|
|
71
71
|
# 22
|
|
72
72
|
|
|
73
73
|
# 23
|
|
@@ -118,7 +118,8 @@ from mindspore.ops.auto_generate import soft_margin_loss
|
|
|
118
118
|
# 45
|
|
119
119
|
|
|
120
120
|
# 46
|
|
121
|
-
from mindspore.ops.
|
|
121
|
+
from mindspore.ops.auto_generate import silu as silu_func
|
|
122
|
+
from mindspore.ops.auto_generate import inplace_silu
|
|
122
123
|
# 47
|
|
123
124
|
|
|
124
125
|
# 48
|
|
@@ -283,6 +284,52 @@ from mindspore.ops.functional import adaptive_avg_pool2d_ext as adaptive_avg_poo
|
|
|
283
284
|
from mindspore.ops.function.nn_func import cross_entropy_ext as cross_entropy
|
|
284
285
|
from mindspore.ops.function.nn_func import nll_loss_ext as nll_loss
|
|
285
286
|
|
|
287
|
+
def silu(input, inplace=False):
|
|
288
|
+
r"""
|
|
289
|
+
Computes Sigmoid Linear Unit of input element-wise. The SiLU function is defined as:
|
|
290
|
+
|
|
291
|
+
.. math::
|
|
292
|
+
|
|
293
|
+
\text{SiLU}(x) = x * \sigma(x),
|
|
294
|
+
|
|
295
|
+
where :math:`x` is an element of the input, :math:`\sigma(x)` is Sigmoid function.
|
|
296
|
+
|
|
297
|
+
.. math::
|
|
298
|
+
|
|
299
|
+
\text{sigma}(x_i) = \frac{1}{1 + \exp(-x_i)},
|
|
300
|
+
|
|
301
|
+
SiLU Function Graph:
|
|
302
|
+
|
|
303
|
+
.. image:: ../images/SiLU.png
|
|
304
|
+
:align: center
|
|
305
|
+
|
|
306
|
+
Args:
|
|
307
|
+
input (Tensor): `input` is :math:`x` in the preceding formula. Input with the data type
|
|
308
|
+
float16 or float32.
|
|
309
|
+
inplace (bool, optional): If it is ``True``, enable the in place update function. Default value: ``False``.
|
|
310
|
+
|
|
311
|
+
Returns:
|
|
312
|
+
Tensor, with the same type and shape as the `input`.
|
|
313
|
+
|
|
314
|
+
Raises:
|
|
315
|
+
TypeError: If dtype of `input` is neither float16 nor float32.
|
|
316
|
+
|
|
317
|
+
Supported Platforms:
|
|
318
|
+
``Ascend`` ``GPU`` ``CPU``
|
|
319
|
+
|
|
320
|
+
Examples:
|
|
321
|
+
>>> import mindspore
|
|
322
|
+
>>> from mindspore import Tensor, mint
|
|
323
|
+
>>> import numpy as np
|
|
324
|
+
>>> input = Tensor(np.array([-1, 2, -3, 2, -1]), mindspore.float16)
|
|
325
|
+
>>> output = mint.nn.functional.silu(input, inplace=False)
|
|
326
|
+
>>> print(output)
|
|
327
|
+
[-0.269 1.762 -0.1423 1.762 -0.269]
|
|
328
|
+
"""
|
|
329
|
+
if inplace:
|
|
330
|
+
return inplace_silu(input)
|
|
331
|
+
return silu_func(input)
|
|
332
|
+
|
|
286
333
|
|
|
287
334
|
def elu(input, alpha=1.0, inplace=False):
|
|
288
335
|
r"""
|
|
@@ -511,7 +558,10 @@ def binary_cross_entropy(input, target, weight=None, reduction='mean'):
|
|
|
511
558
|
\end{cases}
|
|
512
559
|
|
|
513
560
|
.. warning::
|
|
514
|
-
|
|
561
|
+
The value of `input` must range from `0` to `l`.
|
|
562
|
+
|
|
563
|
+
.. note::
|
|
564
|
+
Currently, when the platform is Ascend, all gradient calculations are performed on NPU.
|
|
515
565
|
|
|
516
566
|
Args:
|
|
517
567
|
input (Tensor): The predictive value whose data type must be float16 or float32.
|
|
@@ -955,9 +1005,6 @@ def threshold(input, threshold, value, inplace=False): # pylint: disable=W0621
|
|
|
955
1005
|
\text{value}, &\text{ otherwise }
|
|
956
1006
|
\end{cases}
|
|
957
1007
|
|
|
958
|
-
.. warning::
|
|
959
|
-
This is an experimental API that is subject to change or deletion.
|
|
960
|
-
|
|
961
1008
|
Args:
|
|
962
1009
|
input (Tensor): The input Tensor.
|
|
963
1010
|
threshold (Union[int, float]): The value of the threshold.
|