mindspore 2.1.0__cp38-cp38-manylinux1_x86_64.whl → 2.2.0__cp38-cp38-manylinux1_x86_64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of mindspore might be problematic. Click here for more details.
- mindspore/.commit_id +1 -1
- mindspore/__init__.py +4 -1
- mindspore/_akg/akg/build_module.py +5 -6
- mindspore/_akg/akg/composite/build_module.py +49 -16
- mindspore/_akg/akg/composite/split_stitch.py +10 -11
- mindspore/_akg/akg/ms/info_version_adapt.py +67 -1
- mindspore/_akg/akg/tvm/api.py +4 -3
- mindspore/_akg/akg/tvm/autotvm/__init__.py +1 -2
- mindspore/_akg/akg/tvm/autotvm/graph_tuner/base_graph_tuner.py +1 -5
- mindspore/_akg/akg/tvm/autotvm/measure/__init__.py +1 -1
- mindspore/_akg/akg/tvm/autotvm/measure/measure.py +1 -10
- mindspore/_akg/akg/tvm/autotvm/measure/measure_methods.py +1 -372
- mindspore/_akg/akg/tvm/build_module.py +16 -1
- mindspore/_akg/akg/tvm/contrib/graph_runtime.py +0 -53
- mindspore/_akg/akg/tvm/hybrid/parser.py +7 -6
- mindspore/_akg/akg/tvm/ir_builder.py +1 -1
- mindspore/_akg/akg/tvm/module.py +1 -2
- mindspore/_akg/akg/tvm/stmt.py +2 -2
- mindspore/_akg/akg/utils/composite_op_helper.py +9 -10
- mindspore/_akg/akg/utils/kernel_exec.py +58 -260
- mindspore/_akg/akg/utils/result_analysis.py +4 -24
- mindspore/_akg/akg/utils/tbe_codegen_utils.py +198 -0
- mindspore/_c_dataengine.cpython-38-x86_64-linux-gnu.so +0 -0
- mindspore/_c_expression.cpython-38-x86_64-linux-gnu.so +0 -0
- mindspore/_c_mindrecord.cpython-38-x86_64-linux-gnu.so +0 -0
- mindspore/_check_jit_forbidden_api.py +3 -1
- mindspore/_checkparam.py +26 -32
- mindspore/_extends/graph_kernel/__init__.py +0 -1
- mindspore/_extends/graph_kernel/model/model_builder.py +9 -50
- mindspore/_extends/graph_kernel/splitter.py +1 -9
- mindspore/_extends/parallel_compile/akg_compiler/akg_process.py +122 -15
- mindspore/_extends/parallel_compile/akg_compiler/build_tbe_kernel.py +2 -2
- mindspore/_extends/parallel_compile/akg_compiler/tbe_topi.py +4 -2
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_adapter.py +2 -2
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_helper.py +4 -4
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_job.py +1 -1
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_job_manager.py +1 -1
- mindspore/_extends/parse/__init__.py +12 -15
- mindspore/_extends/parse/namespace.py +7 -33
- mindspore/_extends/parse/parser.py +61 -71
- mindspore/_extends/parse/resources.py +1 -1
- mindspore/_extends/parse/standard_method.py +72 -95
- mindspore/_extends/parse/trope.py +1 -1
- mindspore/_extends/remote/kernel_build_server.py +24 -7
- mindspore/_extends/remote/kernel_build_server_akg_v2.py +55 -0
- mindspore/_install_custom.py +43 -0
- mindspore/_mindspore_offline_debug.cpython-38-x86_64-linux-gnu.so +0 -0
- mindspore/amp.py +47 -11
- mindspore/bin/cache_admin +0 -0
- mindspore/bin/cache_server +0 -0
- mindspore/boost/boost.py +1 -8
- mindspore/boost/boost_cell_wrapper.py +3 -2
- mindspore/boost/grad_accumulation.py +1 -1
- mindspore/boost/group_loss_scale_manager.py +8 -7
- mindspore/common/__init__.py +5 -3
- mindspore/common/_jit_fallback_utils.py +6 -0
- mindspore/common/_register_for_adapter.py +2 -0
- mindspore/common/_register_for_tensor.py +2 -2
- mindspore/common/_stub_tensor.py +13 -0
- mindspore/common/_utils.py +13 -0
- mindspore/common/api.py +173 -258
- mindspore/common/auto_dynamic_shape.py +498 -0
- mindspore/common/dtype.py +18 -11
- mindspore/common/dump.py +6 -4
- mindspore/common/initializer.py +14 -14
- mindspore/common/jit_config.py +33 -15
- mindspore/common/lazy_inline.py +126 -7
- mindspore/common/mindir_util.py +101 -0
- mindspore/common/parameter.py +51 -41
- mindspore/common/seed.py +4 -4
- mindspore/common/sparse_tensor.py +13 -14
- mindspore/common/tensor.py +240 -145
- mindspore/communication/__init__.py +7 -4
- mindspore/communication/_comm_helper.py +83 -4
- mindspore/communication/management.py +152 -84
- mindspore/config/op_info.config +13 -2
- mindspore/config/super_bar_config.json +4 -2
- mindspore/context.py +143 -59
- mindspore/dataset/__init__.py +5 -5
- mindspore/dataset/audio/__init__.py +2 -2
- mindspore/dataset/audio/transforms.py +52 -52
- mindspore/dataset/callback/ds_callback.py +16 -2
- mindspore/dataset/core/config.py +68 -51
- mindspore/dataset/engine/cache_client.py +28 -5
- mindspore/dataset/engine/datasets.py +250 -112
- mindspore/dataset/engine/datasets_audio.py +43 -211
- mindspore/dataset/engine/datasets_standard_format.py +11 -35
- mindspore/dataset/engine/datasets_text.py +43 -67
- mindspore/dataset/engine/datasets_user_defined.py +86 -100
- mindspore/dataset/engine/datasets_vision.py +219 -1029
- mindspore/dataset/engine/iterators.py +11 -4
- mindspore/dataset/engine/obs/obs_mindrecord_dataset.py +4 -0
- mindspore/dataset/engine/obs/util.py +3 -0
- mindspore/dataset/engine/samplers.py +1 -1
- mindspore/dataset/engine/validators.py +19 -5
- mindspore/dataset/text/__init__.py +3 -3
- mindspore/dataset/text/transforms.py +101 -127
- mindspore/dataset/text/utils.py +205 -138
- mindspore/dataset/transforms/__init__.py +1 -1
- mindspore/dataset/transforms/py_transforms_util.py +40 -12
- mindspore/dataset/transforms/transforms.py +95 -40
- mindspore/dataset/utils/browse_dataset.py +8 -2
- mindspore/dataset/utils/line_reader.py +17 -19
- mindspore/dataset/vision/__init__.py +3 -3
- mindspore/dataset/vision/c_transforms.py +6 -3
- mindspore/dataset/vision/transforms.py +409 -287
- mindspore/dataset/vision/utils.py +13 -14
- mindspore/dataset/vision/validators.py +11 -1
- mindspore/experimental/map_parameter.py +14 -0
- mindspore/{nn/optim_ex → experimental/optim}/__init__.py +30 -29
- mindspore/{nn/optim_ex → experimental/optim}/adam.py +59 -66
- mindspore/{nn/optim_ex → experimental/optim}/adamw.py +181 -203
- mindspore/experimental/optim/lr_scheduler.py +1427 -0
- mindspore/{nn/optim_ex → experimental/optim}/optimizer.py +252 -259
- mindspore/{nn/optim_ex → experimental/optim}/sgd.py +147 -152
- mindspore/gen_ops.py +273 -0
- mindspore/include/OWNERS +0 -1
- mindspore/include/api/data_type.h +2 -1
- mindspore/include/api/graph.h +0 -15
- mindspore/include/api/kernel.h +2 -0
- mindspore/include/api/kernel_api.h +37 -12
- mindspore/include/api/model.h +0 -14
- mindspore/include/api/types.h +37 -4
- mindspore/include/c_api/ms/abstract.h +67 -0
- mindspore/include/c_api/ms/attribute.h +197 -0
- mindspore/include/c_api/ms/base/handle_types.h +43 -0
- mindspore/include/c_api/ms/base/macros.h +32 -0
- mindspore/include/c_api/ms/base/status.h +33 -0
- mindspore/include/c_api/ms/base/types.h +282 -0
- mindspore/include/c_api/ms/context.h +102 -0
- mindspore/include/c_api/ms/graph.h +160 -0
- mindspore/include/c_api/ms/node.h +606 -0
- mindspore/include/c_api/ms/tensor.h +161 -0
- mindspore/include/c_api/ms/value.h +84 -0
- mindspore/include/dataset/constants.h +6 -5
- mindspore/include/dataset/execute.h +23 -13
- mindspore/include/dataset/text.h +26 -26
- mindspore/include/dataset/transforms.h +13 -13
- mindspore/include/dataset/vision.h +60 -60
- mindspore/include/dataset/vision_ascend.h +5 -6
- mindspore/include/dataset/vision_lite.h +17 -17
- mindspore/include/mindapi/base/type_id.h +1 -0
- mindspore/include/mindapi/base/types.h +1 -0
- mindspore/lib/libdnnl.so.2 +0 -0
- mindspore/lib/libjemalloc.so.2 +0 -0
- mindspore/lib/libmindspore.so +0 -0
- mindspore/lib/libmindspore_backend.so +0 -0
- mindspore/lib/libmindspore_common.so +0 -0
- mindspore/lib/libmindspore_core.so +0 -0
- mindspore/lib/libmindspore_glog.so.0 +0 -0
- mindspore/lib/libmindspore_gpr.so.15 +0 -0
- mindspore/lib/libmindspore_grpc++.so.1 +0 -0
- mindspore/lib/libmindspore_grpc.so.15 +0 -0
- mindspore/lib/libmindspore_shared_lib.so +0 -0
- mindspore/lib/libnnacl.so +0 -0
- mindspore/lib/libopencv_core.so.4.5 +0 -0
- mindspore/lib/libopencv_imgcodecs.so.4.5 +0 -0
- mindspore/lib/libopencv_imgproc.so.4.5 +0 -0
- mindspore/lib/libps_cache.so +0 -0
- mindspore/lib/plugin/ascend/custom_aicpu_ops/op_impl/cpu/aicpu_kernel/impl/libcust_aicpu_kernels.so +0 -0
- mindspore/lib/plugin/ascend/custom_aicpu_ops/op_impl/cpu/aicpu_kernel/impl/libcust_cpu_kernels.so +0 -0
- mindspore/lib/plugin/ascend/custom_aicpu_ops/op_impl/cpu/config/cust_aicpu_kernel.json +9000 -0
- mindspore/lib/plugin/ascend/custom_aicpu_ops/op_proto/libcust_op_proto.so +0 -0
- mindspore/lib/plugin/ascend/libakg.so +0 -0
- mindspore/lib/plugin/ascend/libascend_collective.so +0 -0
- mindspore/lib/plugin/ascend/libdvpp_utils.so +0 -0
- mindspore/lib/plugin/ascend/libhccl_plugin.so +0 -0
- mindspore/lib/plugin/ascend/libmindspore_aicpu_kernels.so +0 -0
- mindspore/lib/plugin/ascend/libmindspore_cpu_kernels.so +0 -0
- mindspore/lib/plugin/cpu/libakg.so +0 -0
- mindspore/lib/plugin/gpu/libcuda_ops.so.10 +0 -0
- mindspore/lib/plugin/gpu/libcuda_ops.so.11 +0 -0
- mindspore/lib/plugin/gpu10.1/libakg.so +0 -0
- mindspore/lib/plugin/gpu10.1/libnccl.so.2 +0 -0
- mindspore/lib/plugin/gpu11.1/libakg.so +0 -0
- mindspore/lib/plugin/gpu11.1/libnccl.so.2 +0 -0
- mindspore/lib/plugin/gpu11.6/libakg.so +0 -0
- mindspore/lib/plugin/gpu11.6/libnccl.so.2 +0 -0
- mindspore/lib/plugin/libmindspore_ascend.so.1 +0 -0
- mindspore/lib/plugin/libmindspore_ascend.so.2 +0 -0
- mindspore/lib/plugin/libmindspore_gpu.so.10.1 +0 -0
- mindspore/lib/plugin/libmindspore_gpu.so.11.1 +0 -0
- mindspore/lib/plugin/libmindspore_gpu.so.11.6 +0 -0
- mindspore/mindrecord/tools/imagenet_to_mr.py +1 -1
- mindspore/mindrecord/tools/mnist_to_mr.py +2 -2
- mindspore/nn/__init__.py +0 -2
- mindspore/nn/cell.py +316 -74
- mindspore/nn/dynamic_lr.py +21 -21
- mindspore/nn/layer/activation.py +21 -28
- mindspore/nn/layer/basic.py +15 -13
- mindspore/nn/layer/channel_shuffle.py +1 -1
- mindspore/nn/layer/container.py +271 -9
- mindspore/nn/layer/conv.py +310 -207
- mindspore/nn/layer/dense.py +8 -5
- mindspore/nn/layer/embedding.py +33 -27
- mindspore/nn/layer/flash_attention.py +82 -41
- mindspore/nn/layer/image.py +8 -6
- mindspore/nn/layer/math.py +13 -18
- mindspore/nn/layer/normalization.py +107 -66
- mindspore/nn/layer/padding.py +1 -1
- mindspore/nn/layer/pooling.py +131 -109
- mindspore/nn/layer/rnn_cells.py +22 -17
- mindspore/nn/layer/rnns.py +13 -16
- mindspore/nn/layer/thor_layer.py +1 -1
- mindspore/nn/layer/transformer.py +221 -154
- mindspore/nn/learning_rate_schedule.py +9 -1
- mindspore/nn/loss/loss.py +235 -174
- mindspore/nn/optim/ada_grad.py +2 -1
- mindspore/nn/optim/adadelta.py +1 -0
- mindspore/nn/optim/adafactor.py +2 -1
- mindspore/nn/optim/adam.py +7 -4
- mindspore/nn/optim/adamax.py +3 -2
- mindspore/nn/optim/adasum.py +2 -2
- mindspore/nn/optim/asgd.py +2 -3
- mindspore/nn/optim/ftrl.py +6 -5
- mindspore/nn/optim/lamb.py +7 -4
- mindspore/nn/optim/lars.py +1 -1
- mindspore/nn/optim/lazyadam.py +5 -3
- mindspore/nn/optim/momentum.py +2 -1
- mindspore/nn/optim/optimizer.py +53 -4
- mindspore/nn/optim/proximal_ada_grad.py +3 -4
- mindspore/nn/optim/rmsprop.py +4 -3
- mindspore/nn/optim/rprop.py +23 -12
- mindspore/nn/optim/sgd.py +26 -11
- mindspore/nn/optim/thor.py +9 -7
- mindspore/nn/probability/bijector/bijector.py +5 -5
- mindspore/nn/probability/bijector/power_transform.py +27 -27
- mindspore/nn/probability/bijector/softplus.py +3 -3
- mindspore/nn/probability/distribution/_utils/custom_ops.py +3 -3
- mindspore/nn/probability/distribution/bernoulli.py +5 -5
- mindspore/nn/probability/distribution/beta.py +3 -3
- mindspore/nn/probability/distribution/categorical.py +7 -7
- mindspore/nn/probability/distribution/cauchy.py +0 -1
- mindspore/nn/probability/distribution/distribution.py +3 -3
- mindspore/nn/probability/distribution/gamma.py +3 -3
- mindspore/nn/probability/distribution/geometric.py +4 -4
- mindspore/nn/probability/distribution/gumbel.py +4 -4
- mindspore/nn/probability/distribution/log_normal.py +2 -2
- mindspore/nn/probability/distribution/logistic.py +2 -2
- mindspore/nn/probability/distribution/poisson.py +4 -4
- mindspore/nn/probability/distribution/transformed_distribution.py +3 -3
- mindspore/nn/probability/distribution/uniform.py +6 -6
- mindspore/nn/wrap/cell_wrapper.py +78 -34
- mindspore/nn/wrap/grad_reducer.py +8 -5
- mindspore/nn/wrap/loss_scale.py +105 -42
- mindspore/numpy/array_creations.py +1 -2
- mindspore/numpy/array_ops.py +3 -2
- mindspore/offline_debug/convert_async.py +2 -2
- mindspore/ops/_grad_experimental/__init__.py +0 -5
- mindspore/ops/_grad_experimental/grad_array_ops.py +1 -2
- mindspore/ops/_grad_experimental/grad_comm_ops.py +15 -2
- mindspore/ops/_grad_experimental/grad_debug_ops.py +0 -37
- mindspore/ops/_grad_experimental/grad_implementations.py +10 -0
- mindspore/ops/_grad_experimental/grad_inner_ops.py +2 -216
- mindspore/ops/_grad_experimental/grad_math_ops.py +0 -181
- mindspore/ops/_grad_experimental/grad_sparse.py +15 -0
- mindspore/ops/_op_impl/_custom_op/dsd_back_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/flash_attention/attention.py +165 -109
- mindspore/ops/_op_impl/_custom_op/flash_attention/flash_attention_bwd.py +144 -86
- mindspore/ops/_op_impl/_custom_op/flash_attention/flash_attention_fwd.py +172 -187
- mindspore/ops/_op_impl/_custom_op/flash_attention/flash_attention_impl.py +51 -57
- mindspore/ops/_op_impl/_custom_op/flash_attention/tik_ops_utils.py +6 -17
- mindspore/ops/_op_impl/_custom_op/flash_attention/tiling_strategy/wukong_tiling.py +1 -1
- mindspore/ops/_op_impl/aicpu/__init__.py +14 -2
- mindspore/ops/_op_impl/aicpu/bias_add_grad.py +0 -1
- mindspore/ops/_op_impl/aicpu/count_nonzero.py +43 -0
- mindspore/ops/_op_impl/aicpu/eps.py +32 -0
- mindspore/ops/_op_impl/aicpu/gamma.py +2 -2
- mindspore/ops/_op_impl/aicpu/log_uniform_candidate_sampler.py +6 -3
- mindspore/ops/_op_impl/aicpu/lu_unpack_grad.py +0 -1
- mindspore/ops/_op_impl/aicpu/multinomial.py +3 -3
- mindspore/ops/_op_impl/aicpu/parameterized_truncated_normal.py +15 -7
- mindspore/ops/_op_impl/aicpu/random_categorical.py +39 -19
- mindspore/ops/_op_impl/aicpu/random_choice_with_mask.py +5 -2
- mindspore/ops/_op_impl/aicpu/random_poisson.py +103 -52
- mindspore/ops/_op_impl/aicpu/random_shuffle.py +17 -15
- mindspore/ops/_op_impl/aicpu/{sparseaddmm.py → sparse_addmm.py} +2 -2
- mindspore/ops/_op_impl/aicpu/{sparsesparsemaximum.py → sparse_sparse_maximum.py} +4 -4
- mindspore/ops/_op_impl/aicpu/standard_laplace.py +5 -5
- mindspore/ops/_op_impl/aicpu/standard_normal.py +5 -5
- mindspore/ops/_op_impl/aicpu/truncated_normal.py +9 -7
- mindspore/ops/_op_impl/aicpu/uniform.py +5 -3
- mindspore/ops/_op_impl/aicpu/uniform_candidate_sampler.py +8 -4
- mindspore/ops/_op_impl/aicpu/uniform_int.py +5 -5
- mindspore/ops/_op_impl/aicpu/uniform_real.py +4 -4
- mindspore/ops/_op_impl/tbe/__init__.py +4 -4
- mindspore/ops/_op_impl/tbe/inplace_index_add.py +7 -3
- mindspore/ops/_op_impl/tbe/trans_data_ds.py +2 -0
- mindspore/ops/_primitive_cache.py +1 -1
- mindspore/ops/_tracefunc.py +45 -13
- mindspore/ops/_utils/utils.py +4 -1
- mindspore/ops/_vmap/vmap_array_ops.py +3 -3
- mindspore/ops/_vmap/vmap_base.py +3 -3
- mindspore/ops/_vmap/vmap_convolution_ops.py +1 -1
- mindspore/ops/_vmap/vmap_grad_math_ops.py +6 -4
- mindspore/ops/_vmap/vmap_math_ops.py +5 -2
- mindspore/ops/_vmap/vmap_nn_ops.py +61 -7
- mindspore/ops/arg_dtype_cast.py +54 -0
- mindspore/ops/composite/base.py +37 -10
- mindspore/ops/composite/math_ops.py +5 -4
- mindspore/ops/composite/multitype_ops/_compile_utils.py +273 -72
- mindspore/ops/composite/multitype_ops/_constexpr_utils.py +16 -9
- mindspore/ops/composite/multitype_ops/add_impl.py +43 -4
- mindspore/ops/composite/multitype_ops/getitem_impl.py +40 -2
- mindspore/ops/composite/multitype_ops/ones_like_impl.py +6 -0
- mindspore/ops/composite/multitype_ops/setitem_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/zeros_like_impl.py +9 -0
- mindspore/ops/deprecated.py +304 -0
- mindspore/ops/function/__init__.py +4 -1
- mindspore/ops/function/array_func.py +167 -189
- mindspore/ops/function/clip_func.py +81 -13
- mindspore/ops/function/debug_func.py +1 -1
- mindspore/ops/function/grad/grad_func.py +18 -8
- mindspore/ops/function/image_func.py +10 -4
- mindspore/ops/function/linalg_func.py +5 -5
- mindspore/ops/function/math_func.py +575 -386
- mindspore/ops/function/nn_func.py +470 -251
- mindspore/ops/function/random_func.py +86 -56
- mindspore/ops/function/sparse_func.py +1 -1
- mindspore/ops/function/sparse_unary_func.py +14 -12
- mindspore/ops/function/vmap_func.py +6 -5
- mindspore/ops/functional.py +15 -10
- mindspore/ops/op_info_register.py +235 -19
- mindspore/ops/operations/__init__.py +25 -17
- mindspore/ops/operations/_grad_ops.py +52 -7
- mindspore/ops/operations/_inner_ops.py +213 -12
- mindspore/ops/operations/_quant_ops.py +4 -8
- mindspore/ops/operations/_sequence_ops.py +42 -0
- mindspore/ops/operations/array_ops.py +64 -280
- mindspore/ops/operations/comm_ops.py +105 -57
- mindspore/ops/operations/custom_ops.py +10 -3
- mindspore/ops/operations/debug_ops.py +8 -4
- mindspore/ops/operations/image_ops.py +18 -12
- mindspore/ops/operations/math_ops.py +185 -138
- mindspore/ops/operations/nn_ops.py +716 -492
- mindspore/ops/operations/other_ops.py +0 -22
- mindspore/ops/operations/random_ops.py +53 -111
- mindspore/ops/operations/sparse_ops.py +3 -1
- mindspore/ops/primitive.py +24 -18
- mindspore/parallel/_auto_parallel_context.py +68 -8
- mindspore/parallel/_cost_model_context.py +2 -2
- mindspore/parallel/_offload_context.py +17 -3
- mindspore/parallel/_parallel_serialization.py +2 -2
- mindspore/parallel/_ps_context.py +12 -0
- mindspore/parallel/_tensor.py +14 -12
- mindspore/parallel/_transformer/layers.py +5 -3
- mindspore/parallel/_transformer/loss.py +1 -0
- mindspore/parallel/_transformer/moe.py +2 -2
- mindspore/parallel/_transformer/op_parallel_config.py +12 -1
- mindspore/parallel/_transformer/transformer.py +23 -3
- mindspore/parallel/_utils.py +11 -7
- mindspore/parallel/algo_parameter_config.py +85 -5
- mindspore/parallel/checkpoint_transform.py +6 -10
- mindspore/parallel/shard.py +4 -4
- mindspore/profiler/common/struct_type.py +3 -3
- mindspore/profiler/common/util.py +3 -2
- mindspore/profiler/envprofiling.py +1 -1
- mindspore/profiler/parser/aicpu_data_parser.py +5 -3
- mindspore/profiler/parser/ascend_flops_generator.py +2 -2
- mindspore/profiler/parser/ascend_fpbp_generator.py +1 -1
- mindspore/profiler/parser/ascend_hccl_generator.py +17 -12
- mindspore/profiler/parser/ascend_msprof_exporter.py +104 -252
- mindspore/profiler/parser/ascend_msprof_generator.py +8 -8
- mindspore/profiler/parser/ascend_op_generator.py +5 -5
- mindspore/profiler/parser/ascend_steptrace_generator.py +6 -4
- mindspore/profiler/parser/ascend_timeline_generator.py +9 -6
- mindspore/profiler/parser/base_timeline_generator.py +9 -7
- mindspore/profiler/parser/cpu_gpu_timeline_generator.py +14 -10
- mindspore/profiler/parser/flops_parser.py +15 -11
- mindspore/profiler/parser/framework_parser.py +37 -21
- mindspore/profiler/parser/hccl_parser.py +16 -12
- mindspore/profiler/parser/integrator.py +22 -11
- mindspore/profiler/parser/memory_usage_parser.py +2 -2
- mindspore/profiler/parser/minddata_analyzer.py +12 -14
- mindspore/profiler/parser/minddata_pipeline_parser.py +1 -1
- mindspore/profiler/parser/msadvisor_parser.py +8 -4
- mindspore/profiler/parser/op_intermediate_parser.py +5 -2
- mindspore/profiler/parser/optime_parser.py +1 -1
- mindspore/profiler/parser/profiler_info.py +2 -2
- mindspore/profiler/parser/step_trace_parser.py +11 -14
- mindspore/profiler/profiling.py +139 -71
- mindspore/rewrite/api/node.py +102 -19
- mindspore/rewrite/api/node_type.py +5 -1
- mindspore/rewrite/api/scoped_value.py +9 -17
- mindspore/rewrite/api/symbol_tree.py +131 -47
- mindspore/rewrite/ast_helpers/__init__.py +2 -1
- mindspore/rewrite/ast_helpers/ast_finder.py +129 -0
- mindspore/rewrite/ast_helpers/ast_modifier.py +116 -104
- mindspore/rewrite/ast_transformers/flatten_recursive_stmt.py +93 -46
- mindspore/rewrite/common/rewrite_elog.py +5 -1
- mindspore/rewrite/namer.py +33 -24
- mindspore/rewrite/namespace.py +14 -5
- mindspore/{_extends/graph_kernel/expanders/complex → rewrite/node}/__init__.py +9 -9
- mindspore/rewrite/node/call_function.py +79 -0
- mindspore/rewrite/node/cell_container.py +135 -0
- mindspore/rewrite/node/control_flow.py +88 -0
- mindspore/rewrite/{node.py → node/node.py} +273 -234
- mindspore/rewrite/node/node_manager.py +254 -0
- mindspore/rewrite/{topological_manager.py → node/node_topological_manager.py} +13 -46
- mindspore/rewrite/parsers/arguments_parser.py +22 -21
- mindspore/rewrite/parsers/assign_parser.py +216 -221
- mindspore/rewrite/parsers/attribute_parser.py +9 -7
- mindspore/rewrite/parsers/class_def_parser.py +174 -113
- mindspore/rewrite/parsers/constant_parser.py +9 -6
- mindspore/rewrite/parsers/container_parser.py +9 -7
- mindspore/rewrite/parsers/for_parser.py +36 -15
- mindspore/rewrite/parsers/function_def_parser.py +24 -16
- mindspore/rewrite/parsers/if_parser.py +28 -24
- mindspore/rewrite/parsers/module_parser.py +196 -25
- mindspore/rewrite/{parser.py → parsers/parser.py} +4 -2
- mindspore/rewrite/{parser_register.py → parsers/parser_register.py} +1 -1
- mindspore/rewrite/parsers/return_parser.py +6 -6
- mindspore/rewrite/sparsify/sparse_transformer.py +12 -3
- mindspore/rewrite/sparsify/utils.py +1 -1
- mindspore/rewrite/symbol_tree.py +525 -577
- mindspore/rewrite/symbol_tree_builder.py +9 -193
- mindspore/rewrite/symbol_tree_dumper.py +2 -2
- mindspore/run_check/_check_version.py +2 -2
- mindspore/{ops/bprop_mindir → safeguard}/__init__.py +4 -3
- mindspore/safeguard/rewrite_obfuscation.py +517 -0
- mindspore/scipy/linalg.py +1 -1
- mindspore/scipy/optimize/minimize.py +7 -3
- mindspore/train/_utils.py +7 -3
- mindspore/train/amp.py +323 -123
- mindspore/train/anf_ir_pb2.py +14 -2
- mindspore/train/callback/_backup_and_restore.py +2 -12
- mindspore/train/callback/_callback.py +29 -4
- mindspore/train/callback/_checkpoint.py +23 -8
- mindspore/train/callback/_early_stop.py +2 -2
- mindspore/train/callback/_landscape.py +4 -4
- mindspore/train/callback/_loss_monitor.py +2 -2
- mindspore/train/callback/_on_request_exit.py +2 -2
- mindspore/train/callback/_reduce_lr_on_plateau.py +3 -4
- mindspore/train/callback/_summary_collector.py +14 -7
- mindspore/train/callback/_time_monitor.py +58 -5
- mindspore/train/data_sink.py +5 -11
- mindspore/train/dataset_helper.py +83 -57
- mindspore/train/loss_scale_manager.py +2 -2
- mindspore/train/metrics/__init__.py +3 -3
- mindspore/train/metrics/cosine_similarity.py +1 -1
- mindspore/train/metrics/hausdorff_distance.py +3 -2
- mindspore/train/metrics/mean_surface_distance.py +3 -2
- mindspore/train/metrics/metric.py +39 -19
- mindspore/train/metrics/roc.py +2 -2
- mindspore/train/metrics/root_mean_square_surface_distance.py +4 -3
- mindspore/train/mind_ir_pb2.py +85 -36
- mindspore/train/model.py +185 -45
- mindspore/train/serialization.py +390 -150
- mindspore/train/summary/_writer_pool.py +3 -2
- mindspore/train/summary/summary_record.py +14 -10
- mindspore/train/train_thor/convert_utils.py +3 -3
- mindspore/train/train_thor/dataset_helper.py +1 -1
- mindspore/version.py +1 -1
- {mindspore-2.1.0.dist-info → mindspore-2.2.0.dist-info}/METADATA +6 -7
- {mindspore-2.1.0.dist-info → mindspore-2.2.0.dist-info}/RECORD +458 -518
- {mindspore-2.1.0.dist-info → mindspore-2.2.0.dist-info}/entry_points.txt +0 -1
- mindspore/_akg/akg/tvm/contrib/debugger/__init__.py +0 -16
- mindspore/_akg/akg/tvm/contrib/debugger/debug_result.py +0 -274
- mindspore/_akg/akg/tvm/contrib/debugger/debug_runtime.py +0 -259
- mindspore/_akg/akg/tvm/contrib/peak.py +0 -341
- mindspore/_akg/akg/tvm/contrib/rpc.py +0 -25
- mindspore/_akg/akg/tvm/contrib/xcode.py +0 -257
- mindspore/_akg/akg/tvm/exec/__init__.py +0 -17
- mindspore/_akg/akg/tvm/exec/autotvm_log_editor.py +0 -60
- mindspore/_akg/akg/tvm/exec/measure_peak.py +0 -48
- mindspore/_akg/akg/tvm/exec/query_rpc_tracker.py +0 -48
- mindspore/_akg/akg/tvm/exec/rpc_proxy.py +0 -98
- mindspore/_akg/akg/tvm/exec/rpc_server.py +0 -88
- mindspore/_akg/akg/tvm/exec/rpc_tracker.py +0 -62
- mindspore/_akg/akg/tvm/rpc/__init__.py +0 -29
- mindspore/_akg/akg/tvm/rpc/base.py +0 -182
- mindspore/_akg/akg/tvm/rpc/client.py +0 -436
- mindspore/_akg/akg/tvm/rpc/proxy.py +0 -595
- mindspore/_akg/akg/tvm/rpc/server.py +0 -413
- mindspore/_akg/akg/tvm/rpc/tornado_util.py +0 -121
- mindspore/_akg/akg/tvm/rpc/tracker.py +0 -431
- mindspore/_extends/graph_kernel/expander.py +0 -80
- mindspore/_extends/graph_kernel/expanders/__init__.py +0 -54
- mindspore/_extends/graph_kernel/expanders/_utils.py +0 -269
- mindspore/_extends/graph_kernel/expanders/addn.py +0 -33
- mindspore/_extends/graph_kernel/expanders/batchnorm.py +0 -152
- mindspore/_extends/graph_kernel/expanders/batchnorm_grad.py +0 -105
- mindspore/_extends/graph_kernel/expanders/clip_by_norm_no_div_sum.py +0 -33
- mindspore/_extends/graph_kernel/expanders/complex/abs.py +0 -30
- mindspore/_extends/graph_kernel/expanders/complex/add.py +0 -44
- mindspore/_extends/graph_kernel/expanders/complex/div.py +0 -62
- mindspore/_extends/graph_kernel/expanders/complex/mul.py +0 -52
- mindspore/_extends/graph_kernel/expanders/complex/real_div.py +0 -62
- mindspore/_extends/graph_kernel/expanders/complex/sub.py +0 -45
- mindspore/_extends/graph_kernel/expanders/conv2d.py +0 -200
- mindspore/_extends/graph_kernel/expanders/dropout_grad.py +0 -30
- mindspore/_extends/graph_kernel/expanders/equal_count.py +0 -50
- mindspore/_extends/graph_kernel/expanders/erfc.py +0 -35
- mindspore/_extends/graph_kernel/expanders/expand_dims.py +0 -50
- mindspore/_extends/graph_kernel/expanders/fused_adam.py +0 -44
- mindspore/_extends/graph_kernel/expanders/fused_adam_weight_decay.py +0 -47
- mindspore/_extends/graph_kernel/expanders/fused_mul_add.py +0 -28
- mindspore/_extends/graph_kernel/expanders/gelu_grad.py +0 -70
- mindspore/_extends/graph_kernel/expanders/gkdropout.py +0 -40
- mindspore/_extends/graph_kernel/expanders/identity.py +0 -25
- mindspore/_extends/graph_kernel/expanders/layernorm.py +0 -93
- mindspore/_extends/graph_kernel/expanders/layernorm_grad.py +0 -113
- mindspore/_extends/graph_kernel/expanders/logsoftmax.py +0 -46
- mindspore/_extends/graph_kernel/expanders/logsoftmax_grad.py +0 -36
- mindspore/_extends/graph_kernel/expanders/matmul.py +0 -80
- mindspore/_extends/graph_kernel/expanders/maximum_grad.py +0 -59
- mindspore/_extends/graph_kernel/expanders/minimum_grad.py +0 -80
- mindspore/_extends/graph_kernel/expanders/oneslike.py +0 -26
- mindspore/_extends/graph_kernel/expanders/reduce_mean.py +0 -43
- mindspore/_extends/graph_kernel/expanders/relu_grad.py +0 -32
- mindspore/_extends/graph_kernel/expanders/sigmoid_cross_entropy_with_logits.py +0 -41
- mindspore/_extends/graph_kernel/expanders/sigmoid_cross_entropy_with_logits_grad.py +0 -35
- mindspore/_extends/graph_kernel/expanders/sigmoid_grad.py +0 -31
- mindspore/_extends/graph_kernel/expanders/slice.py +0 -35
- mindspore/_extends/graph_kernel/expanders/softmax_cross_entropy_with_logits.py +0 -42
- mindspore/_extends/graph_kernel/expanders/softmax_grad_ext.py +0 -41
- mindspore/_extends/graph_kernel/expanders/softsign.py +0 -28
- mindspore/_extends/graph_kernel/expanders/sqrt_grad.py +0 -29
- mindspore/_extends/graph_kernel/expanders/square_sum_all.py +0 -44
- mindspore/_extends/graph_kernel/expanders/square_sum_v1.py +0 -37
- mindspore/_extends/graph_kernel/expanders/squared_difference.py +0 -43
- mindspore/_extends/graph_kernel/expanders/tanh_grad.py +0 -31
- mindspore/_extends/graph_kernel/model/op_infer.py +0 -506
- mindspore/dataset/datapreprocess/__init__.py +0 -20
- mindspore/dataset/datapreprocess/preprocess_imagenet_validate_dataset.py +0 -54
- mindspore/include/api/net.h +0 -142
- mindspore/nn/lr_scheduler.py +0 -262
- mindspore/ops/_grad_experimental/grad_image_ops.py +0 -248
- mindspore/ops/_grad_experimental/grad_linalg_ops.py +0 -181
- mindspore/ops/_grad_experimental/grad_other_ops.py +0 -72
- mindspore/ops/_grad_experimental/grad_scalar_ops.py +0 -112
- mindspore/ops/_grad_experimental/grad_sequence_ops.py +0 -351
- mindspore/ops/bprop_mindir/BNTrainingReduce_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Broadcast_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Depend_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/DepthwiseConv2dNative_bprop.mindir +0 -138
- mindspore/ops/bprop_mindir/EmbeddingLookup_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Load_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/ScatterNonAliasingAdd_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/SparseGatherV2_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/SparseSoftmaxCrossEntropyWithLogits_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Switch_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/TransShape_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/TupleGetItem_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Unique_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Unstack_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/generate_mindir.py +0 -114
- mindspore/rewrite/node_visitor.py +0 -44
- {mindspore-2.1.0.dist-info → mindspore-2.2.0.dist-info}/WHEEL +0 -0
- {mindspore-2.1.0.dist-info → mindspore-2.2.0.dist-info}/top_level.txt +0 -0
|
@@ -25,7 +25,6 @@ from tbe.common.platform import get_soc_spec
|
|
|
25
25
|
from mindspore.ops._op_impl._custom_op.flash_attention.constants import FP16
|
|
26
26
|
from mindspore.ops._op_impl._custom_op.flash_attention.constants import FP32
|
|
27
27
|
from mindspore.ops._op_impl._custom_op.flash_attention.constants import GM
|
|
28
|
-
from mindspore.ops._op_impl._custom_op.flash_attention.constants import INT8
|
|
29
28
|
from mindspore.ops._op_impl._custom_op.flash_attention.constants import MASK_FILL_VALUE
|
|
30
29
|
from mindspore.ops._op_impl._custom_op.flash_attention.constants import UB
|
|
31
30
|
from mindspore.ops._op_impl._custom_op.flash_attention.tik_ops_utils import TikOpsUtils
|
|
@@ -37,7 +36,7 @@ from mindspore.ops._op_impl._custom_op.flash_attention.tiling_strategy.sparse_ti
|
|
|
37
36
|
class FlashAttention(metaclass=ABCMeta):
|
|
38
37
|
"""The base class of FlashAttention"""
|
|
39
38
|
|
|
40
|
-
def __init__(self, q, k, v,
|
|
39
|
+
def __init__(self, q, k, v, attn_mask, dropout_mask, alibi_mask, kernel_name,
|
|
41
40
|
tiling_stgy_cls,
|
|
42
41
|
prev_block_num=65536,
|
|
43
42
|
next_block_num=65536,
|
|
@@ -48,7 +47,6 @@ class FlashAttention(metaclass=ABCMeta):
|
|
|
48
47
|
:param q: with shape: (B, h, N, d)
|
|
49
48
|
:param k: with shape: (B, h, N, d)
|
|
50
49
|
:param v: with shape: (B, h, N, d)
|
|
51
|
-
:param dim_mask: with shape: (x, ) x equals length last dim that not padded to 16.
|
|
52
50
|
:param attn_mask: with shape: (1, N, N) or (B, N, N)
|
|
53
51
|
:param dropout_mask: with shape: (B, h, N, N)
|
|
54
52
|
:param alibi_mask: with shape: (B, h, 1, N)
|
|
@@ -66,19 +64,25 @@ class FlashAttention(metaclass=ABCMeta):
|
|
|
66
64
|
src_stride=0,
|
|
67
65
|
dst_stride=0)
|
|
68
66
|
self.tik_ops_utils = TikOpsUtils(self.tik_instance)
|
|
69
|
-
self.
|
|
70
|
-
|
|
71
|
-
|
|
67
|
+
self.parse_input_shape(alibi_mask, attn_mask, dropout_mask, k, q, v)
|
|
68
|
+
# NZ
|
|
69
|
+
_, _, N1, M1, M0, N0 = self.q_shape
|
|
70
|
+
self.M1 = M1
|
|
71
|
+
self.N1 = N1
|
|
72
|
+
self.M0 = M0
|
|
73
|
+
self.N0 = N0
|
|
74
|
+
self.d = N1 * N0
|
|
75
|
+
# ND
|
|
76
|
+
batch_size, h, Nq, actual_d = self.q_ori_shape
|
|
72
77
|
self.head_num = h
|
|
73
|
-
self.B, self.Nq
|
|
74
|
-
self.N = self.
|
|
78
|
+
self.B, self.Nq = batch_size * h, Nq
|
|
79
|
+
self.N = self.k_ori_shape[2]
|
|
80
|
+
self.actual_d = actual_d
|
|
75
81
|
|
|
76
82
|
self.l_shape = [batch_size, h, self.Nq]
|
|
77
83
|
self.m_shape = [batch_size, h, self.Nq]
|
|
78
|
-
self.O_shape =
|
|
79
|
-
self.actual_d = self.dim_mask_shape[0]
|
|
84
|
+
self.O_shape = self.q_shape
|
|
80
85
|
|
|
81
|
-
self.K0 = 16
|
|
82
86
|
self.prev_block_num = prev_block_num
|
|
83
87
|
self.next_block_num = next_block_num
|
|
84
88
|
self.high_precision = high_precision
|
|
@@ -86,7 +90,6 @@ class FlashAttention(metaclass=ABCMeta):
|
|
|
86
90
|
self.precision_type = FP32
|
|
87
91
|
else:
|
|
88
92
|
self.precision_type = FP16
|
|
89
|
-
|
|
90
93
|
if tiling_stgy_cls is None:
|
|
91
94
|
self.tiling_stgy = SparseTiling(self.Nq, self.N, self.d)
|
|
92
95
|
else:
|
|
@@ -106,9 +109,9 @@ class FlashAttention(metaclass=ABCMeta):
|
|
|
106
109
|
self.alibi_mask_gm = None
|
|
107
110
|
|
|
108
111
|
@staticmethod
|
|
109
|
-
def
|
|
110
|
-
"""get gm offset"""
|
|
111
|
-
gm_offset = (batch_start + batch_idx) * h
|
|
112
|
+
def get_l_m_gm_offset(batch_start, batch_idx, h, block_h, block_idx):
|
|
113
|
+
"""get l m gm offset"""
|
|
114
|
+
gm_offset = (batch_start + batch_idx) * h + block_idx * block_h
|
|
112
115
|
return gm_offset
|
|
113
116
|
|
|
114
117
|
@staticmethod
|
|
@@ -143,108 +146,159 @@ class FlashAttention(metaclass=ABCMeta):
|
|
|
143
146
|
"""collect outputs"""
|
|
144
147
|
raise NotImplementedError
|
|
145
148
|
|
|
146
|
-
|
|
147
|
-
|
|
148
|
-
|
|
149
|
-
|
|
150
|
-
core_m_batch_num]]), Tensor([[[core_1_batch_1_Tr_start, core_1_batch_1_Tr_end],...[core_1_batch_n_Tr_start,
|
|
151
|
-
core_1_batch_n_Tr_end]],...,[[core_m_batch_1_Tr_start, core_m_batch_1_Tr_end],...[core_m_batch_n_Tr_start,
|
|
152
|
-
core_m_batch_n_Tr_end]]
|
|
153
|
-
"""
|
|
154
|
-
if self.core_num > self.B * self.Tr:
|
|
155
|
-
self.core_num = self.B * self.Tr
|
|
156
|
-
|
|
157
|
-
task_idx_to_batch_tr_idx = dict()
|
|
158
|
-
for task_idx in range(self.B * self.Tr):
|
|
159
|
-
batch_idx = task_idx // self.Tr
|
|
160
|
-
tr_idx = task_idx % self.Tr
|
|
161
|
-
task_idx_to_batch_tr_idx[task_idx] = [batch_idx, tr_idx]
|
|
149
|
+
@abstractmethod
|
|
150
|
+
def compute_one_core(self, batch_start_s, batch_num_s, core_idx_to_tr_info, core_idx):
|
|
151
|
+
"""compute one core"""
|
|
152
|
+
raise NotImplementedError
|
|
162
153
|
|
|
163
|
-
|
|
164
|
-
|
|
165
|
-
|
|
166
|
-
|
|
154
|
+
@abstractmethod
|
|
155
|
+
def prepare_global_ones(self):
|
|
156
|
+
"""prepare global ones"""
|
|
157
|
+
raise NotImplementedError
|
|
167
158
|
|
|
168
|
-
|
|
169
|
-
|
|
170
|
-
|
|
171
|
-
|
|
172
|
-
task_end = task_start + cur_core_task_num
|
|
173
|
-
for task_idx in range(task_start, task_end):
|
|
174
|
-
try:
|
|
175
|
-
batch_idx, tr_idx = task_idx_to_batch_tr_idx[task_idx]
|
|
176
|
-
except KeyError:
|
|
177
|
-
raise ValueError("The argument 'task_idx' is not valid.")
|
|
178
|
-
batch_start_end_pair = core_idx_to_batch_idx[core_idx]
|
|
179
|
-
if batch_idx < batch_start_end_pair[0]:
|
|
180
|
-
batch_start_end_pair[0] = batch_idx
|
|
181
|
-
if batch_idx > batch_start_end_pair[1]:
|
|
182
|
-
batch_start_end_pair[1] = batch_idx
|
|
183
|
-
tr_start_end_pair = core_idx_to_tr_idx[core_idx][batch_idx]
|
|
184
|
-
if tr_idx < tr_start_end_pair[0]:
|
|
185
|
-
tr_start_end_pair[0] = tr_idx
|
|
186
|
-
if tr_idx > tr_start_end_pair[1]:
|
|
187
|
-
tr_start_end_pair[1] = tr_idx
|
|
188
|
-
task_start = task_end
|
|
159
|
+
def get_gm_offset(self, batch_start, batch_idx, h, w, block_h, block_idx):
|
|
160
|
+
"""get gm offset"""
|
|
161
|
+
gm_offset = (batch_start + batch_idx) * h * w + block_idx * block_h * self.N0
|
|
162
|
+
return gm_offset
|
|
189
163
|
|
|
164
|
+
def get_cur_tr_block_num(self, tr_idx):
|
|
165
|
+
"""get cur tr block_num"""
|
|
166
|
+
cur_prev_block_num = min(tr_idx, self.prev_block_num)
|
|
167
|
+
cur_next_block_num = min(self.next_block_num, self.Tc - tr_idx - 1)
|
|
168
|
+
block_num = cur_prev_block_num + 1 + cur_next_block_num
|
|
169
|
+
return block_num
|
|
170
|
+
|
|
171
|
+
def get_total_block_num(self):
|
|
172
|
+
"""get total block num"""
|
|
173
|
+
block_num = 0
|
|
174
|
+
for b_idx in range(self.B):
|
|
175
|
+
for tr_idx in range(self.Tr):
|
|
176
|
+
block_num += self.get_cur_tr_block_num(tr_idx)
|
|
177
|
+
return block_num
|
|
178
|
+
|
|
179
|
+
def update_core_task_map(self,
|
|
180
|
+
core_b_map,
|
|
181
|
+
core_b_tr_map,
|
|
182
|
+
core_idx,
|
|
183
|
+
b_start,
|
|
184
|
+
b_end,
|
|
185
|
+
tr_start,
|
|
186
|
+
tr_end):
|
|
187
|
+
"""update core task map"""
|
|
188
|
+
core_b_map[core_idx][0] = min(core_b_map[core_idx][0], b_start)
|
|
189
|
+
if tr_end == 0: # 跨head,但跨过的head不会被当前的core处理
|
|
190
|
+
core_b_map[core_idx][1] = max(core_b_map[core_idx][1], b_end - 1)
|
|
191
|
+
else:
|
|
192
|
+
core_b_map[core_idx][1] = max(core_b_map[core_idx][1], b_end)
|
|
193
|
+
for b_idx in range(b_start, b_end + 1):
|
|
194
|
+
if b_idx == b_end and tr_end == 0: # 跨head,但跨过的head不会被当前的core处理
|
|
195
|
+
break
|
|
196
|
+
elif b_idx == b_start and b_idx == b_end: # 没跨head
|
|
197
|
+
core_b_tr_map[core_idx][b_idx] = (tr_start, tr_end)
|
|
198
|
+
elif b_idx == b_start: # 跨head,第一个head
|
|
199
|
+
core_b_tr_map[core_idx][b_idx] = (tr_start, self.Tr)
|
|
200
|
+
elif b_idx == b_end: # 跨head,最后一个head
|
|
201
|
+
core_b_tr_map[core_idx][b_idx] = (0, tr_end)
|
|
202
|
+
else: # 跨head,中间的head
|
|
203
|
+
core_b_tr_map[core_idx][b_idx] = (0, self.Tr)
|
|
204
|
+
|
|
205
|
+
def convert_py_dict_to_tik_tensor(self, core_b_map, core_b_tr_map):
|
|
206
|
+
"""convert py dict to tik tensor"""
|
|
207
|
+
# python dict -> tik tensor
|
|
208
|
+
# [batch_start, batch_idx_end] -> [batch_start, batch_num]
|
|
209
|
+
# [tr_start, tr_idx_end] -> [tr_start, tr_idx_end)
|
|
190
210
|
core_idx_to_batch_info = self.tik_instance.Tensor(
|
|
191
211
|
"int32", (self.core_num, 2), name="core_idx_to_batch_info", scope=UB
|
|
192
212
|
)
|
|
193
213
|
core_idx_to_tr_info = self.tik_instance.Tensor(
|
|
194
214
|
"int32", (self.core_num, self.B, 2), name="core_idx_to_tr_info", scope=UB
|
|
195
215
|
)
|
|
196
|
-
for core_idx in
|
|
197
|
-
batch_start, batch_end =
|
|
216
|
+
for core_idx in core_b_map.keys():
|
|
217
|
+
batch_start, batch_end = core_b_map[core_idx]
|
|
198
218
|
core_idx_to_batch_info[core_idx, 0] = batch_start
|
|
199
219
|
core_idx_to_batch_info[core_idx, 1] = batch_end - batch_start + 1
|
|
200
|
-
for batch_idx in
|
|
201
|
-
tr_start, tr_end =
|
|
220
|
+
for batch_idx in core_b_tr_map[core_idx].keys():
|
|
221
|
+
tr_start, tr_end = core_b_tr_map[core_idx][batch_idx]
|
|
202
222
|
core_idx_to_tr_info[core_idx, batch_idx, 0] = tr_start
|
|
203
|
-
core_idx_to_tr_info[core_idx, batch_idx, 1] = tr_end
|
|
223
|
+
core_idx_to_tr_info[core_idx, batch_idx, 1] = tr_end
|
|
224
|
+
|
|
225
|
+
return core_idx_to_batch_info, core_idx_to_tr_info
|
|
226
|
+
|
|
227
|
+
def get_core_task_info(self):
|
|
228
|
+
"""
|
|
229
|
+
Get batch start and batch number of each NPU core.
|
|
230
|
+
:return: Tensor([[core_1_batch_start, core_1_batch_num],...,[core_m_batch_start,
|
|
231
|
+
core_m_batch_num]]), Tensor([[[core_1_batch_1_Tr_start, core_1_batch_1_Tr_end],...[core_1_batch_n_Tr_start,
|
|
232
|
+
core_1_batch_n_Tr_end]],...,[[core_m_batch_1_Tr_start, core_m_batch_1_Tr_end],...[core_m_batch_n_Tr_start,
|
|
233
|
+
core_m_batch_n_Tr_end]]
|
|
234
|
+
"""
|
|
235
|
+
if self.core_num > self.B * self.Tr:
|
|
236
|
+
self.core_num = self.B * self.Tr
|
|
237
|
+
|
|
238
|
+
total_blk_num = self.get_total_block_num()
|
|
239
|
+
b_start = 0
|
|
240
|
+
tr_start = 0
|
|
241
|
+
remain_blk_num = total_blk_num
|
|
242
|
+
core_b_map = defaultdict(lambda: [100000, -1])
|
|
243
|
+
core_b_tr_map = defaultdict(lambda: defaultdict(list))
|
|
244
|
+
for core_idx in range(self.core_num):
|
|
245
|
+
cur_core_blk_num = 0
|
|
246
|
+
cur_each_core_blk_num = remain_blk_num // (self.core_num - core_idx)
|
|
247
|
+
cur_core_finished = False
|
|
248
|
+
b_end = b_start
|
|
249
|
+
tr_end = tr_start
|
|
250
|
+
while b_end < self.B:
|
|
251
|
+
while tr_end < self.Tr:
|
|
252
|
+
cur_tr_blk_num = self.get_cur_tr_block_num(tr_end)
|
|
253
|
+
if abs(cur_core_blk_num - cur_each_core_blk_num) <= \
|
|
254
|
+
(cur_core_blk_num + cur_tr_blk_num - cur_each_core_blk_num):
|
|
255
|
+
self.update_core_task_map(core_b_map, core_b_tr_map, core_idx, b_start, b_end, tr_start, tr_end)
|
|
256
|
+
remain_blk_num -= cur_core_blk_num
|
|
257
|
+
cur_core_finished = True
|
|
258
|
+
break
|
|
259
|
+
else:
|
|
260
|
+
cur_core_blk_num += cur_tr_blk_num
|
|
261
|
+
tr_end += 1
|
|
262
|
+
if tr_end == self.Tr:
|
|
263
|
+
tr_end = 0
|
|
264
|
+
b_end += 1
|
|
265
|
+
if cur_core_finished:
|
|
266
|
+
b_start = b_end
|
|
267
|
+
tr_start = tr_end
|
|
268
|
+
break
|
|
269
|
+
core_idx_to_batch_info, core_idx_to_tr_info = self.convert_py_dict_to_tik_tensor(core_b_map, core_b_tr_map)
|
|
204
270
|
return core_idx_to_batch_info, core_idx_to_tr_info
|
|
205
271
|
|
|
206
272
|
def get_attn_mask_gm_offset(self, batch_start, batch_idx, h, w, block_h, block_h_idx, block_w, block_w_idx):
|
|
207
273
|
"""get attn mask gm offset"""
|
|
208
274
|
if self.att_mask_shape[0] == 1:
|
|
209
|
-
gm_offset =
|
|
275
|
+
gm_offset = block_w_idx * (h * block_w) + block_h_idx * block_h * self.N0
|
|
210
276
|
else:
|
|
211
277
|
gm_offset = ((batch_start + batch_idx) // self.head_num) * h * w \
|
|
212
|
-
+
|
|
278
|
+
+ block_w_idx * (h * block_w) + block_h_idx * block_h * self.N0
|
|
213
279
|
return gm_offset
|
|
214
280
|
|
|
215
|
-
def
|
|
281
|
+
def parse_input_shape(self, alibi_mask, attn_mask, dropout_mask, k, q, v):
|
|
216
282
|
"""parser input shape"""
|
|
217
283
|
self.has_attn_mask = False
|
|
218
284
|
self.has_drop_mask = False
|
|
219
285
|
self.has_alibi_mask = False
|
|
220
|
-
|
|
221
|
-
|
|
222
|
-
|
|
223
|
-
|
|
224
|
-
|
|
225
|
-
|
|
226
|
-
|
|
227
|
-
|
|
228
|
-
|
|
229
|
-
|
|
230
|
-
|
|
231
|
-
|
|
232
|
-
|
|
233
|
-
|
|
234
|
-
|
|
235
|
-
self.
|
|
236
|
-
self.k_shape = k.shape
|
|
237
|
-
self.v_shape = v.shape
|
|
238
|
-
self.dim_mask_shape = dim_mask.shape
|
|
239
|
-
if attn_mask is not None:
|
|
240
|
-
self.has_attn_mask = True
|
|
241
|
-
self.att_mask_shape = attn_mask.shape
|
|
242
|
-
if dropout_mask is not None:
|
|
243
|
-
self.has_drop_mask = True
|
|
244
|
-
self.drop_mask_shape = dropout_mask.shape
|
|
245
|
-
if alibi_mask is not None:
|
|
246
|
-
self.has_alibi_mask = True
|
|
247
|
-
self.alibi_mask_shape = alibi_mask.shape
|
|
286
|
+
# NZ
|
|
287
|
+
self.q_shape = q["shape"]
|
|
288
|
+
self.k_shape = k["shape"]
|
|
289
|
+
self.v_shape = v["shape"]
|
|
290
|
+
# ND
|
|
291
|
+
self.q_ori_shape = q["ori_shape"]
|
|
292
|
+
self.k_ori_shape = k["ori_shape"]
|
|
293
|
+
if attn_mask is not None:
|
|
294
|
+
self.has_attn_mask = True
|
|
295
|
+
self.att_mask_shape = attn_mask["shape"]
|
|
296
|
+
if dropout_mask is not None:
|
|
297
|
+
self.has_drop_mask = True
|
|
298
|
+
self.drop_mask_shape = dropout_mask["shape"]
|
|
299
|
+
if alibi_mask is not None:
|
|
300
|
+
self.has_alibi_mask = True
|
|
301
|
+
self.alibi_mask_shape = alibi_mask["shape"]
|
|
248
302
|
|
|
249
303
|
def define_inputs_outputs(self):
|
|
250
304
|
"""define inputs outputs"""
|
|
@@ -272,8 +326,6 @@ class FlashAttention(metaclass=ABCMeta):
|
|
|
272
326
|
self.Q_gm = self.tik_instance.Tensor(FP16, self.q_shape, name="Q_gm", scope=GM)
|
|
273
327
|
self.K_gm = self.tik_instance.Tensor(FP16, self.k_shape, name="K_gm", scope=GM)
|
|
274
328
|
self.V_gm = self.tik_instance.Tensor(FP16, self.v_shape, name="V_gm", scope=GM)
|
|
275
|
-
self.dim_mask_gm = self.tik_instance.Tensor(INT8, self.dim_mask_shape, name="mask_gm",
|
|
276
|
-
scope=GM)
|
|
277
329
|
if self.has_attn_mask:
|
|
278
330
|
self.att_mask_gm = self.tik_instance.Tensor(FP16, self.att_mask_shape,
|
|
279
331
|
name="att_mask_gm", scope=GM)
|
|
@@ -294,43 +346,47 @@ class FlashAttention(metaclass=ABCMeta):
|
|
|
294
346
|
alibi_mask_ub_broadcast = self.tik_ops_utils.broadcast_row(alibi_mask_ub, (m_aligned, n_aligned))
|
|
295
347
|
self.tik_instance.h_add(Sij_ub, Sij_ub, alibi_mask_ub_broadcast)
|
|
296
348
|
|
|
297
|
-
def do_att_mask(self,
|
|
349
|
+
def do_att_mask(self, Sij_ub_N1MN0, attn_mask_gm_offset, q_blk_height, kv_blk_height,
|
|
298
350
|
q_blk_h_aligned, kv_blk_h_aligned):
|
|
299
351
|
"""load attn mask from gm to ub, then mul it by MASK_FILL_VALUE and add Sij"""
|
|
300
352
|
with self.tik_instance.new_stmt_scope(disable_sync=False):
|
|
301
|
-
att_mask_ub = self.tik_instance.Tensor(FP16, (q_blk_h_aligned,
|
|
353
|
+
att_mask_ub = self.tik_instance.Tensor(FP16, (kv_blk_h_aligned // self.N0, q_blk_h_aligned, self.N0),
|
|
302
354
|
scope=UB, name="att_mask_ub")
|
|
303
355
|
self.tik_instance.data_move(att_mask_ub, self.att_mask_gm[attn_mask_gm_offset], 0,
|
|
304
|
-
|
|
356
|
+
kv_blk_height // self.N0, q_blk_height * self.N0 // 16,
|
|
357
|
+
(self.Nq - q_blk_height) * self.N0 // 16, 0)
|
|
305
358
|
self.tik_instance.h_mul(att_mask_ub, att_mask_ub, MASK_FILL_VALUE)
|
|
306
|
-
self.tik_instance.h_add(
|
|
359
|
+
self.tik_instance.h_add(Sij_ub_N1MN0, Sij_ub_N1MN0, att_mask_ub)
|
|
307
360
|
|
|
308
361
|
def do_dropout_mask(self, Pij_ub, dropout_mask_gm_offset, kv_blk_h_aligned, kv_blk_height,
|
|
309
|
-
q_blk_h_aligned, q_blk_height, precision_type=FP16):
|
|
362
|
+
q_blk_h_aligned, q_blk_height, precision_type=FP16, workspace=None):
|
|
310
363
|
"""load drop mask from gm to ub, then mul it by Pij"""
|
|
311
364
|
with self.tik_instance.new_stmt_scope(disable_sync=False):
|
|
312
365
|
dropout_mask_ub = self.tik_instance.Tensor(FP16, (q_blk_h_aligned, kv_blk_h_aligned),
|
|
313
366
|
scope=UB, name="drop_mask_ub")
|
|
314
367
|
self.tik_instance.data_move(dropout_mask_ub, self.drop_mask_gm[dropout_mask_gm_offset], 0,
|
|
315
368
|
q_blk_height, kv_blk_height // 16, (self.N - kv_blk_height) // 16, 0)
|
|
369
|
+
dropout_mask_ub = dropout_mask_ub.reshape((kv_blk_height // self.N0, q_blk_height, self.N0))
|
|
316
370
|
if precision_type == FP32:
|
|
317
|
-
dropout_mask_ub_fp32 = self.tik_instance.Tensor(FP32,
|
|
371
|
+
dropout_mask_ub_fp32 = self.tik_instance.Tensor(FP32,
|
|
372
|
+
(kv_blk_h_aligned // self.N0, q_blk_h_aligned, self.N0),
|
|
318
373
|
scope=UB, name="dropout_mask_ub_fp32")
|
|
319
374
|
self.tik_instance.h_cast(dropout_mask_ub_fp32, dropout_mask_ub, "none")
|
|
320
|
-
|
|
375
|
+
if workspace is None:
|
|
376
|
+
self.tik_instance.h_mul(Pij_ub, Pij_ub, dropout_mask_ub_fp32)
|
|
377
|
+
else:
|
|
378
|
+
self.tik_instance.h_mul(workspace, Pij_ub, dropout_mask_ub_fp32)
|
|
321
379
|
else:
|
|
322
|
-
|
|
323
|
-
|
|
324
|
-
|
|
325
|
-
|
|
326
|
-
"""compute one core"""
|
|
327
|
-
raise NotImplementedError
|
|
380
|
+
if workspace is None:
|
|
381
|
+
self.tik_instance.h_mul(Pij_ub, Pij_ub, dropout_mask_ub)
|
|
382
|
+
else:
|
|
383
|
+
self.tik_instance.h_mul(workspace, Pij_ub, dropout_mask_ub)
|
|
328
384
|
|
|
329
385
|
def compute_process(self):
|
|
330
386
|
"""The compute process of FlashAttention"""
|
|
331
387
|
self.init()
|
|
332
|
-
|
|
333
|
-
core_idx_to_batch_info, core_idx_to_tr_info = self.
|
|
388
|
+
self.prepare_global_ones()
|
|
389
|
+
core_idx_to_batch_info, core_idx_to_tr_info = self.get_core_task_info()
|
|
334
390
|
with self.tik_instance.for_range(begint=0, endt=self.core_num, name="core_index",
|
|
335
391
|
block_num=self.core_num) as core_idx:
|
|
336
392
|
batch_start_s = self.tik_instance.Scalar("int32", name="batch_start_s")
|