mindspore 2.5.0__cp39-cp39-win_amd64.whl → 2.6.0rc1__cp39-cp39-win_amd64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of mindspore might be problematic. Click here for more details.
- mindspore/.commit_id +1 -1
- mindspore/Microsoft.VisualStudio.Telemetry.dll +0 -0
- mindspore/Newtonsoft.Json.dll +0 -0
- mindspore/__init__.py +6 -4
- mindspore/_c_dataengine.cp39-win_amd64.pyd +0 -0
- mindspore/_c_expression.cp39-win_amd64.pyd +0 -0
- mindspore/_c_mindrecord.cp39-win_amd64.pyd +0 -0
- mindspore/_check_jit_forbidden_api.py +3 -0
- mindspore/_checkparam.py +3 -33
- mindspore/_deprecated/__init__.py +17 -0
- mindspore/_deprecated/jit.py +198 -0
- mindspore/_extends/builtin_operations.py +1 -1
- mindspore/_extends/parse/__init__.py +6 -7
- mindspore/_extends/parse/compile_config.py +19 -0
- mindspore/_extends/parse/deprecated/deprecated_tensor_method.py +22 -3
- mindspore/_extends/parse/jit_fallback_modules/__init__.py +0 -0
- mindspore/_extends/parse/jit_fallback_modules/check_utils.py +123 -0
- mindspore/_extends/parse/jit_fallback_modules/third_party_modules.py +50 -0
- mindspore/_extends/parse/parser.py +24 -193
- mindspore/_extends/parse/resources.py +1 -5
- mindspore/_extends/parse/standard_method.py +97 -74
- mindspore/_extends/pijit/__init__.py +2 -2
- mindspore/_extends/pijit/pijit_func_white_list.py +16 -11
- mindspore/_extends/pijit/tensor_func_list.py +27 -0
- mindspore/_extends/utils.py +1 -1
- mindspore/amp.py +4 -4
- mindspore/atlprov.dll +0 -0
- mindspore/avcodec-59.dll +0 -0
- mindspore/avdevice-59.dll +0 -0
- mindspore/avfilter-8.dll +0 -0
- mindspore/avformat-59.dll +0 -0
- mindspore/avutil-57.dll +0 -0
- mindspore/boost/__init__.py +2 -2
- mindspore/boost/base.py +3 -7
- mindspore/boost/boost_cell_wrapper.py +2 -2
- mindspore/c1.dll +0 -0
- mindspore/c1xx.dll +0 -0
- mindspore/c2.dll +0 -0
- mindspore/common/__init__.py +4 -3
- mindspore/common/_grad_function.py +56 -0
- mindspore/common/_pijit_context.py +14 -5
- mindspore/common/_register_for_tensor.py +1 -1
- mindspore/common/_stub_tensor.py +5 -10
- mindspore/common/_tensor_cpp_method.py +1 -1
- mindspore/common/_tensor_docs.py +1915 -3287
- mindspore/common/api.py +341 -354
- mindspore/common/auto_dynamic_shape.py +41 -44
- mindspore/common/dtype.py +5 -2
- mindspore/common/dump.py +7 -5
- mindspore/common/file_system.py +3 -0
- mindspore/common/hook_handle.py +5 -3
- mindspore/common/initializer.py +10 -6
- mindspore/common/jit_begin_end.py +94 -0
- mindspore/common/jit_config.py +6 -1
- mindspore/common/jit_context.py +76 -0
- mindspore/common/jit_trace.py +378 -0
- mindspore/common/lazy_inline.py +2 -2
- mindspore/common/mutable.py +5 -4
- mindspore/common/parameter.py +106 -39
- mindspore/common/seed.py +2 -2
- mindspore/common/sparse_tensor.py +23 -17
- mindspore/common/tensor.py +297 -714
- mindspore/communication/__init__.py +7 -5
- mindspore/communication/_comm_helper.py +47 -2
- mindspore/communication/comm_func.py +70 -53
- mindspore/communication/management.py +83 -17
- mindspore/context.py +214 -560
- mindspore/dataset/__init__.py +44 -20
- mindspore/dataset/audio/__init__.py +2 -8
- mindspore/dataset/audio/transforms.py +3 -17
- mindspore/dataset/core/config.py +3 -3
- mindspore/dataset/engine/cache_client.py +1 -1
- mindspore/dataset/engine/datasets.py +102 -120
- mindspore/dataset/engine/datasets_audio.py +22 -22
- mindspore/dataset/engine/datasets_standard_format.py +43 -24
- mindspore/dataset/engine/datasets_text.py +78 -85
- mindspore/dataset/engine/datasets_user_defined.py +108 -76
- mindspore/dataset/engine/datasets_vision.py +111 -108
- mindspore/dataset/engine/iterators.py +5 -3
- mindspore/dataset/engine/obs/obs_mindrecord_dataset.py +1 -1
- mindspore/dataset/engine/samplers.py +279 -57
- mindspore/dataset/engine/serializer_deserializer.py +2 -1
- mindspore/dataset/engine/validators.py +10 -0
- mindspore/dataset/text/__init__.py +7 -6
- mindspore/dataset/text/transforms.py +6 -5
- mindspore/dataset/text/utils.py +3 -3
- mindspore/dataset/transforms/__init__.py +0 -9
- mindspore/dataset/transforms/transforms.py +3 -3
- mindspore/dataset/utils/browse_dataset.py +1 -1
- mindspore/dataset/vision/__init__.py +2 -9
- mindspore/dataset/vision/transforms.py +202 -158
- mindspore/dataset/vision/utils.py +7 -5
- mindspore/device_context/ascend/op_debug.py +60 -1
- mindspore/device_context/ascend/op_tuning.py +0 -4
- mindspore/device_manager.py +39 -3
- mindspore/dnnl.dll +0 -0
- mindspore/dpcmi.dll +0 -0
- mindspore/experimental/es/embedding_service.py +35 -27
- mindspore/experimental/map_parameter.py +4 -4
- mindspore/experimental/optim/adadelta.py +22 -26
- mindspore/experimental/optim/adagrad.py +4 -4
- mindspore/experimental/optim/adam.py +4 -0
- mindspore/experimental/optim/adamax.py +4 -4
- mindspore/experimental/optim/adamw.py +4 -0
- mindspore/experimental/optim/asgd.py +1 -1
- mindspore/experimental/optim/lr_scheduler.py +40 -22
- mindspore/experimental/optim/radam.py +5 -5
- mindspore/experimental/optim/rprop.py +1 -1
- mindspore/experimental/optim/sgd.py +1 -1
- mindspore/hal/contiguous_tensors_handle.py +6 -10
- mindspore/hal/device.py +55 -81
- mindspore/hal/event.py +38 -55
- mindspore/hal/memory.py +93 -144
- mindspore/hal/stream.py +81 -125
- mindspore/include/dataset/constants.h +7 -4
- mindspore/include/dataset/execute.h +2 -2
- mindspore/jpeg62.dll +0 -0
- mindspore/log.py +40 -2
- mindspore/mindrecord/__init__.py +20 -7
- mindspore/mindspore_backend_common.dll +0 -0
- mindspore/mindspore_backend_manager.dll +0 -0
- mindspore/mindspore_common.dll +0 -0
- mindspore/mindspore_core.dll +0 -0
- mindspore/mindspore_dump.dll +0 -0
- mindspore/mindspore_frontend.dll +0 -0
- mindspore/mindspore_glog.dll +0 -0
- mindspore/mindspore_memory_pool.dll +0 -0
- mindspore/mindspore_ms_backend.dll +0 -0
- mindspore/mindspore_ops.dll +0 -0
- mindspore/{mindspore_backend.dll → mindspore_ops_host.dll} +0 -0
- mindspore/mindspore_ops_kernel_common.dll +0 -0
- mindspore/mindspore_profiler.dll +0 -0
- mindspore/mindspore_pyboost.dll +0 -0
- mindspore/mindspore_pynative.dll +0 -0
- mindspore/mindspore_res_manager.dll +0 -0
- mindspore/mindspore_runtime_pipeline.dll +0 -0
- mindspore/mint/__init__.py +131 -700
- mindspore/mint/distributed/__init__.py +5 -1
- mindspore/mint/distributed/distributed.py +194 -109
- mindspore/mint/linalg/__init__.py +2 -0
- mindspore/mint/nn/__init__.py +280 -18
- mindspore/mint/nn/functional.py +282 -64
- mindspore/mint/nn/layer/__init__.py +4 -0
- mindspore/mint/nn/layer/_functions.py +7 -3
- mindspore/mint/nn/layer/activation.py +120 -13
- mindspore/mint/nn/layer/conv.py +218 -24
- mindspore/mint/nn/layer/normalization.py +15 -16
- mindspore/mint/nn/layer/padding.py +1 -1
- mindspore/mint/nn/layer/pooling.py +66 -1
- mindspore/mint/optim/__init__.py +2 -1
- mindspore/mint/optim/sgd.py +171 -0
- mindspore/msobj140.dll +0 -0
- mindspore/mspdb140.dll +0 -0
- mindspore/mspdbcore.dll +0 -0
- mindspore/mspdbst.dll +0 -0
- mindspore/mspft140.dll +0 -0
- mindspore/msvcdis140.dll +0 -0
- mindspore/msvcp140_1.dll +0 -0
- mindspore/msvcp140_2.dll +0 -0
- mindspore/msvcp140_atomic_wait.dll +0 -0
- mindspore/msvcp140_codecvt_ids.dll +0 -0
- mindspore/nn/__init__.py +4 -1
- mindspore/nn/cell.py +1250 -176
- mindspore/nn/layer/activation.py +23 -21
- mindspore/nn/layer/basic.py +22 -16
- mindspore/nn/layer/container.py +1 -1
- mindspore/nn/layer/conv.py +22 -17
- mindspore/nn/layer/embedding.py +9 -8
- mindspore/nn/layer/normalization.py +48 -42
- mindspore/nn/layer/pooling.py +75 -31
- mindspore/nn/layer/transformer.py +11 -10
- mindspore/nn/learning_rate_schedule.py +4 -2
- mindspore/nn/loss/loss.py +27 -19
- mindspore/nn/optim/ada_grad.py +6 -5
- mindspore/nn/optim/adadelta.py +9 -7
- mindspore/nn/optim/adafactor.py +1 -1
- mindspore/nn/optim/adam.py +16 -12
- mindspore/nn/optim/adamax.py +8 -7
- mindspore/nn/optim/adasum.py +5 -5
- mindspore/nn/optim/asgd.py +1 -1
- mindspore/nn/optim/ftrl.py +11 -9
- mindspore/nn/optim/lamb.py +1 -1
- mindspore/nn/optim/lazyadam.py +12 -10
- mindspore/nn/optim/momentum.py +7 -6
- mindspore/nn/optim/optimizer.py +2 -2
- mindspore/nn/optim/proximal_ada_grad.py +12 -10
- mindspore/nn/optim/rmsprop.py +13 -12
- mindspore/nn/optim/rprop.py +9 -7
- mindspore/nn/optim/sgd.py +9 -6
- mindspore/nn/optim/tft_wrapper.py +5 -2
- mindspore/nn/probability/bijector/bijector.py +17 -11
- mindspore/nn/probability/bijector/gumbel_cdf.py +5 -5
- mindspore/nn/probability/bijector/invert.py +2 -2
- mindspore/nn/probability/bijector/scalar_affine.py +3 -3
- mindspore/nn/probability/bijector/softplus.py +3 -2
- mindspore/nn/probability/distribution/beta.py +3 -3
- mindspore/nn/probability/distribution/categorical.py +1 -1
- mindspore/nn/probability/distribution/cauchy.py +4 -2
- mindspore/nn/probability/distribution/exponential.py +6 -7
- mindspore/nn/probability/distribution/gamma.py +2 -2
- mindspore/nn/probability/distribution/gumbel.py +2 -2
- mindspore/nn/probability/distribution/half_normal.py +5 -3
- mindspore/nn/probability/distribution/logistic.py +5 -3
- mindspore/nn/probability/distribution/poisson.py +1 -1
- mindspore/nn/probability/distribution/uniform.py +5 -3
- mindspore/nn/reinforcement/_tensors_queue.py +1 -1
- mindspore/nn/reinforcement/tensor_array.py +1 -1
- mindspore/nn/wrap/__init__.py +6 -6
- mindspore/nn/wrap/cell_wrapper.py +178 -117
- mindspore/nn/wrap/grad_reducer.py +45 -36
- mindspore/nn/wrap/loss_scale.py +3 -3
- mindspore/numpy/array_creations.py +3 -3
- mindspore/numpy/array_ops.py +1 -1
- mindspore/numpy/math_ops.py +4 -4
- mindspore/numpy/utils.py +1 -2
- mindspore/numpy/utils_const.py +1 -2
- mindspore/opencv_core452.dll +0 -0
- mindspore/opencv_imgcodecs452.dll +0 -0
- mindspore/opencv_imgproc452.dll +0 -0
- mindspore/ops/__init__.py +3 -2
- mindspore/ops/_grad_experimental/grad_comm_ops.py +18 -3
- mindspore/ops/_grad_experimental/grad_debug_ops.py +8 -1
- mindspore/ops/_grad_experimental/taylor_rule.py +29 -0
- mindspore/ops/_register_for_op.py +0 -11
- mindspore/{ops_generate → ops/_utils}/arg_dtype_cast.py +123 -4
- mindspore/{ops_generate → ops/_utils}/arg_handler.py +3 -4
- mindspore/ops/_vmap/vmap_array_ops.py +7 -6
- mindspore/ops/_vmap/vmap_grad_nn_ops.py +2 -1
- mindspore/ops/_vmap/vmap_math_ops.py +4 -7
- mindspore/ops/_vmap/vmap_nn_ops.py +9 -8
- mindspore/ops/auto_generate/__init__.py +4 -3
- mindspore/ops/auto_generate/cpp_create_prim_instance_helper.py +102 -49
- mindspore/ops/auto_generate/gen_extend_func.py +281 -135
- mindspore/ops/auto_generate/gen_ops_def.py +2574 -2326
- mindspore/ops/auto_generate/gen_ops_prim.py +8566 -2755
- mindspore/ops/auto_generate/pyboost_inner_prim.py +106 -76
- mindspore/ops/composite/__init__.py +2 -1
- mindspore/ops/composite/base.py +19 -24
- mindspore/ops/composite/math_ops.py +6 -16
- mindspore/ops/composite/multitype_ops/__init__.py +5 -2
- mindspore/ops/composite/multitype_ops/_compile_utils.py +2 -3
- mindspore/ops/composite/multitype_ops/_constexpr_utils.py +1 -2
- mindspore/ops/composite/multitype_ops/add_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/bitwise_and_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/bitwise_or_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/bitwise_xor_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/div_impl.py +6 -4
- mindspore/ops/composite/multitype_ops/equal_impl.py +4 -3
- mindspore/ops/composite/multitype_ops/floordiv_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/getitem_impl.py +3 -2
- mindspore/ops/composite/multitype_ops/greater_equal_impl.py +4 -3
- mindspore/ops/composite/multitype_ops/greater_impl.py +4 -3
- mindspore/ops/composite/multitype_ops/in_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/invert_impl.py +50 -0
- mindspore/ops/composite/multitype_ops/left_shift_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/less_equal_impl.py +4 -3
- mindspore/ops/composite/multitype_ops/less_impl.py +4 -3
- mindspore/ops/composite/multitype_ops/logic_not_impl.py +3 -2
- mindspore/ops/composite/multitype_ops/logical_and_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/logical_or_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/mod_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/mul_impl.py +3 -2
- mindspore/ops/composite/multitype_ops/negative_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/not_equal_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/not_in_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/ones_like_impl.py +18 -0
- mindspore/ops/composite/multitype_ops/pow_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/right_shift_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/setitem_impl.py +2 -1
- mindspore/ops/composite/multitype_ops/sub_impl.py +2 -1
- mindspore/ops/function/__init__.py +28 -2
- mindspore/ops/function/_add_attr_func.py +58 -0
- mindspore/ops/function/array_func.py +1629 -2345
- mindspore/ops/function/clip_func.py +38 -45
- mindspore/ops/function/debug_func.py +36 -44
- mindspore/ops/function/grad/__init__.py +1 -0
- mindspore/ops/function/grad/grad_func.py +104 -71
- mindspore/ops/function/image_func.py +1 -1
- mindspore/ops/function/linalg_func.py +46 -78
- mindspore/ops/function/math_func.py +3035 -3705
- mindspore/ops/function/nn_func.py +676 -241
- mindspore/ops/function/other_func.py +159 -1
- mindspore/ops/function/parameter_func.py +17 -30
- mindspore/ops/function/random_func.py +204 -361
- mindspore/ops/function/reshard_func.py +4 -70
- mindspore/ops/function/sparse_func.py +3 -3
- mindspore/ops/function/sparse_unary_func.py +5 -5
- mindspore/ops/function/spectral_func.py +25 -58
- mindspore/ops/function/vmap_func.py +24 -17
- mindspore/ops/functional.py +6 -4
- mindspore/ops/functional_overload.py +547 -4
- mindspore/ops/op_info_register.py +32 -244
- mindspore/ops/operations/__init__.py +10 -5
- mindspore/ops/operations/_custom_ops_utils.py +247 -0
- mindspore/ops/operations/_grad_ops.py +1 -10
- mindspore/ops/operations/_inner_ops.py +5 -76
- mindspore/ops/operations/_ms_kernel.py +4 -10
- mindspore/ops/operations/_rl_inner_ops.py +1 -1
- mindspore/ops/operations/_scalar_ops.py +3 -2
- mindspore/ops/operations/_sequence_ops.py +1 -1
- mindspore/ops/operations/_tensor_array.py +1 -1
- mindspore/ops/operations/array_ops.py +37 -22
- mindspore/ops/operations/comm_ops.py +150 -107
- mindspore/ops/operations/custom_ops.py +221 -23
- mindspore/ops/operations/debug_ops.py +115 -16
- mindspore/ops/operations/inner_ops.py +1 -1
- mindspore/ops/operations/linalg_ops.py +1 -58
- mindspore/ops/operations/manually_defined/_inner.py +1 -1
- mindspore/ops/operations/manually_defined/ops_def.py +746 -79
- mindspore/ops/operations/math_ops.py +21 -18
- mindspore/ops/operations/nn_ops.py +65 -191
- mindspore/ops/operations/other_ops.py +62 -9
- mindspore/ops/operations/random_ops.py +13 -7
- mindspore/ops/operations/reshard_ops.py +1 -1
- mindspore/ops/operations/sparse_ops.py +2 -2
- mindspore/ops/primitive.py +43 -32
- mindspore/ops/tensor_method.py +232 -13
- mindspore/ops_generate/__init__.py +0 -5
- mindspore/ops_generate/aclnn/__init__.py +0 -0
- mindspore/ops_generate/{aclnn_kernel_register_auto_cc_generator.py → aclnn/aclnn_kernel_register_auto_cc_generator.py} +43 -18
- mindspore/ops_generate/{gen_aclnn_implement.py → aclnn/gen_aclnn_implement.py} +49 -51
- mindspore/ops_generate/api/__init__.py +0 -0
- mindspore/ops_generate/{add_tensor_docs_generator.py → api/add_tensor_docs_generator.py} +9 -7
- mindspore/ops_generate/{cpp_create_prim_instance_helper_generator.py → api/cpp_create_prim_instance_helper_generator.py} +6 -9
- mindspore/ops_generate/{functional_map_cpp_generator.py → api/functional_map_cpp_generator.py} +25 -12
- mindspore/ops_generate/{functional_overload_py_generator.py → api/functional_overload_py_generator.py} +8 -6
- mindspore/ops_generate/{functions_cc_generator.py → api/functions_cc_generator.py} +14 -10
- mindspore/ops_generate/api/gen_api.py +103 -0
- mindspore/ops_generate/{op_api_proto.py → api/op_api_proto.py} +98 -69
- mindspore/ops_generate/{tensor_func_reg_cpp_generator.py → api/tensor_func_reg_cpp_generator.py} +82 -43
- mindspore/ops_generate/common/__init__.py +0 -0
- mindspore/ops_generate/common/gen_constants.py +91 -0
- mindspore/ops_generate/{gen_utils.py → common/gen_utils.py} +72 -19
- mindspore/ops_generate/{op_proto.py → common/op_proto.py} +64 -1
- mindspore/ops_generate/{template.py → common/template.py} +96 -84
- mindspore/ops_generate/gen_ops.py +23 -325
- mindspore/ops_generate/op_def/__init__.py +0 -0
- mindspore/ops_generate/op_def/gen_op_def.py +90 -0
- mindspore/ops_generate/{lite_ops_cpp_generator.py → op_def/lite_ops_cpp_generator.py} +47 -11
- mindspore/ops_generate/{ops_def_cc_generator.py → op_def/ops_def_cc_generator.py} +18 -7
- mindspore/ops_generate/{ops_def_h_generator.py → op_def/ops_def_h_generator.py} +5 -5
- mindspore/ops_generate/{ops_name_h_generator.py → op_def/ops_name_h_generator.py} +30 -15
- mindspore/ops_generate/op_def/ops_primitive_h_generator.py +125 -0
- mindspore/ops_generate/op_def_py/__init__.py +0 -0
- mindspore/ops_generate/op_def_py/gen_op_def_py.py +47 -0
- mindspore/ops_generate/{op_def_py_generator.py → op_def_py/op_def_py_generator.py} +6 -5
- mindspore/ops_generate/{op_prim_py_generator.py → op_def_py/op_prim_py_generator.py} +24 -15
- mindspore/ops_generate/pyboost/__init__.py +0 -0
- mindspore/ops_generate/{auto_grad_impl_cc_generator.py → pyboost/auto_grad_impl_cc_generator.py} +11 -7
- mindspore/ops_generate/{auto_grad_reg_cc_generator.py → pyboost/auto_grad_reg_cc_generator.py} +7 -7
- mindspore/ops_generate/{gen_pyboost_func.py → pyboost/gen_pyboost_func.py} +40 -16
- mindspore/ops_generate/{op_template_parser.py → pyboost/op_template_parser.py} +105 -24
- mindspore/ops_generate/{pyboost_functions_cpp_generator.py → pyboost/pyboost_functions_cpp_generator.py} +55 -18
- mindspore/ops_generate/{pyboost_functions_h_generator.py → pyboost/pyboost_functions_h_generator.py} +42 -10
- mindspore/ops_generate/{pyboost_functions_py_generator.py → pyboost/pyboost_functions_py_generator.py} +6 -6
- mindspore/ops_generate/{pyboost_grad_function_cpp_generator.py → pyboost/pyboost_grad_function_cpp_generator.py} +11 -10
- mindspore/ops_generate/{pyboost_inner_prim_generator.py → pyboost/pyboost_inner_prim_generator.py} +8 -7
- mindspore/ops_generate/{pyboost_native_grad_functions_generator.py → pyboost/pyboost_native_grad_functions_generator.py} +14 -10
- mindspore/ops_generate/{pyboost_op_cpp_code_generator.py → pyboost/pyboost_op_cpp_code_generator.py} +140 -53
- mindspore/ops_generate/{pyboost_overload_functions_cpp_generator.py → pyboost/pyboost_overload_functions_cpp_generator.py} +28 -15
- mindspore/ops_generate/{pyboost_utils.py → pyboost/pyboost_utils.py} +88 -4
- mindspore/ops_generate/resources/__init__.py +0 -0
- mindspore/ops_generate/resources/resource_list.py +30 -0
- mindspore/ops_generate/resources/resource_loader.py +36 -0
- mindspore/ops_generate/resources/resource_manager.py +64 -0
- mindspore/ops_generate/resources/yaml_loader.py +88 -0
- mindspore/ops_generate/tensor_py_cc_generator.py +122 -0
- mindspore/parallel/__init__.py +6 -2
- mindspore/parallel/_auto_parallel_context.py +133 -6
- mindspore/parallel/_cell_wrapper.py +130 -15
- mindspore/parallel/_parallel_serialization.py +95 -4
- mindspore/parallel/_ps_context.py +1 -1
- mindspore/parallel/_recovery_context.py +7 -2
- mindspore/parallel/_tensor.py +142 -18
- mindspore/parallel/_utils.py +198 -25
- mindspore/parallel/algo_parameter_config.py +3 -3
- mindspore/parallel/auto_parallel.py +732 -0
- mindspore/parallel/checkpoint_convert.py +159 -0
- mindspore/parallel/checkpoint_transform.py +656 -37
- mindspore/parallel/cluster/process_entity/_api.py +151 -19
- mindspore/parallel/cluster/run.py +1 -1
- mindspore/parallel/function/__init__.py +24 -0
- mindspore/parallel/function/reshard_func.py +259 -0
- mindspore/parallel/nn/__init__.py +25 -0
- mindspore/parallel/nn/parallel_cell_wrapper.py +263 -0
- mindspore/parallel/nn/parallel_grad_reducer.py +169 -0
- mindspore/parallel/parameter_broadcast.py +24 -13
- mindspore/parallel/shard.py +137 -61
- mindspore/parallel/transform_safetensors.py +287 -95
- mindspore/pgodb140.dll +0 -0
- mindspore/pgort140.dll +0 -0
- mindspore/profiler/__init__.py +9 -5
- mindspore/profiler/analysis/parser/ascend_cann_parser.py +6 -2
- mindspore/profiler/analysis/parser/ms_framework_parser.py +4 -4
- mindspore/profiler/analysis/parser/timeline_assembly_factory/ascend_timeline_assembler.py +7 -4
- mindspore/profiler/analysis/parser/timeline_assembly_factory/trace_view_container.py +22 -0
- mindspore/profiler/analysis/parser/timeline_creator/fwk_timeline_creator.py +3 -3
- mindspore/profiler/analysis/parser/timeline_event/fwk_event.py +241 -86
- mindspore/profiler/analysis/viewer/ascend_communication_viewer.py +41 -2
- mindspore/profiler/analysis/viewer/ascend_kernel_details_viewer.py +33 -35
- mindspore/profiler/analysis/viewer/ascend_memory_viewer.py +7 -0
- mindspore/profiler/analysis/viewer/ascend_op_memory_viewer.py +8 -3
- mindspore/profiler/analysis/viewer/ascend_step_trace_time_viewer.py +141 -30
- mindspore/profiler/analysis/viewer/ms_dataset_viewer.py +5 -6
- mindspore/profiler/common/ascend_msprof_exporter.py +5 -4
- mindspore/profiler/common/constant.py +12 -0
- mindspore/profiler/common/msprof_cmd_tool.py +42 -23
- mindspore/profiler/common/path_manager.py +24 -0
- mindspore/profiler/common/profiler_context.py +26 -2
- mindspore/profiler/common/profiler_meta_data.py +74 -0
- mindspore/profiler/common/profiler_parameters.py +59 -18
- mindspore/profiler/common/profiler_path_manager.py +66 -7
- mindspore/profiler/dynamic_profiler.py +112 -79
- mindspore/profiler/envprofiler.py +26 -1
- mindspore/profiler/experimental_config.py +197 -0
- mindspore/profiler/mstx.py +57 -14
- mindspore/profiler/platform/npu_profiler.py +33 -7
- mindspore/profiler/profiler.py +541 -45
- mindspore/profiler/profiler_action_controller.py +1 -1
- mindspore/profiler/profiler_interface.py +4 -0
- mindspore/profiler/schedule.py +57 -22
- mindspore/rewrite/api/node.py +15 -13
- mindspore/rewrite/api/symbol_tree.py +1 -1
- mindspore/run_check/_check_version.py +25 -14
- mindspore/run_check/run_check.py +1 -1
- mindspore/runtime/__init__.py +2 -2
- mindspore/runtime/executor.py +40 -11
- mindspore/runtime/memory.py +25 -8
- mindspore/safeguard/rewrite_obfuscation.py +12 -9
- mindspore/swresample-4.dll +0 -0
- mindspore/swscale-6.dll +0 -0
- mindspore/tbbmalloc.dll +0 -0
- mindspore/tinyxml2.dll +0 -0
- mindspore/train/__init__.py +8 -8
- mindspore/train/_utils.py +35 -7
- mindspore/train/amp.py +1 -1
- mindspore/train/callback/__init__.py +2 -2
- mindspore/train/callback/_callback.py +2 -16
- mindspore/train/callback/_checkpoint.py +24 -40
- mindspore/train/callback/_cluster_monitor.py +14 -18
- mindspore/train/callback/_flops_collector.py +2 -3
- mindspore/train/callback/_history.py +7 -4
- mindspore/train/callback/_lambda_callback.py +2 -2
- mindspore/train/callback/_landscape.py +0 -3
- mindspore/train/callback/_loss_monitor.py +2 -1
- mindspore/train/callback/_on_request_exit.py +6 -5
- mindspore/train/callback/_reduce_lr_on_plateau.py +11 -6
- mindspore/train/callback/_summary_collector.py +8 -13
- mindspore/train/callback/_time_monitor.py +2 -1
- mindspore/train/callback/{_tft_register.py → _train_fault_tolerance.py} +179 -103
- mindspore/train/data_sink.py +25 -2
- mindspore/train/dataset_helper.py +4 -5
- mindspore/train/loss_scale_manager.py +8 -7
- mindspore/train/metrics/accuracy.py +3 -3
- mindspore/train/metrics/confusion_matrix.py +9 -9
- mindspore/train/metrics/error.py +3 -3
- mindspore/train/metrics/hausdorff_distance.py +4 -4
- mindspore/train/metrics/mean_surface_distance.py +3 -3
- mindspore/train/metrics/metric.py +0 -12
- mindspore/train/metrics/occlusion_sensitivity.py +4 -2
- mindspore/train/metrics/precision.py +8 -6
- mindspore/train/metrics/recall.py +9 -9
- mindspore/train/metrics/root_mean_square_surface_distance.py +2 -2
- mindspore/train/mind_ir_pb2.py +19 -12
- mindspore/train/model.py +176 -103
- mindspore/train/serialization.py +246 -988
- mindspore/train/summary/_summary_adapter.py +2 -2
- mindspore/train/summary/summary_record.py +1 -1
- mindspore/turbojpeg.dll +0 -0
- mindspore/utils/__init__.py +3 -2
- mindspore/utils/dryrun.py +4 -2
- mindspore/utils/hooks.py +81 -0
- mindspore/utils/utils.py +138 -4
- mindspore/vcmeta.dll +0 -0
- mindspore/vcruntime140.dll +0 -0
- mindspore/vcruntime140_1.dll +0 -0
- mindspore/version.py +1 -1
- {mindspore-2.5.0.dist-info → mindspore-2.6.0rc1.dist-info}/METADATA +2 -1
- {mindspore-2.5.0.dist-info → mindspore-2.6.0rc1.dist-info}/RECORD +483 -438
- mindspore/_install_custom.py +0 -43
- mindspore/common/_register_for_adapter.py +0 -74
- mindspore/ops/auto_generate/gen_arg_dtype_cast.py +0 -252
- mindspore/ops/auto_generate/gen_arg_handler.py +0 -136
- mindspore/ops/operations/_opaque_predicate_registry.py +0 -41
- mindspore/ops_generate/gen_constants.py +0 -190
- mindspore/ops_generate/gen_ops_inner_prim.py +0 -131
- mindspore/ops_generate/ops_primitive_h_generator.py +0 -81
- /mindspore/ops_generate/{base_generator.py → common/base_generator.py} +0 -0
- {mindspore-2.5.0.dist-info → mindspore-2.6.0rc1.dist-info}/WHEEL +0 -0
- {mindspore-2.5.0.dist-info → mindspore-2.6.0rc1.dist-info}/entry_points.txt +0 -0
- {mindspore-2.5.0.dist-info → mindspore-2.6.0rc1.dist-info}/top_level.txt +0 -0
|
@@ -25,7 +25,8 @@ After declaring the dataset object, you can further apply dataset operations
|
|
|
25
25
|
"""
|
|
26
26
|
import mindspore._c_dataengine as cde
|
|
27
27
|
|
|
28
|
-
from .
|
|
28
|
+
from .samplers import Shuffle
|
|
29
|
+
from .datasets import TextBaseDataset, SourceDataset, MappableDataset
|
|
29
30
|
from .validators import check_imdb_dataset, check_iwslt2016_dataset, check_iwslt2017_dataset, \
|
|
30
31
|
check_penn_treebank_dataset, check_ag_news_dataset, check_amazon_review_dataset, check_udpos_dataset, \
|
|
31
32
|
check_wiki_text_dataset, check_conll2000_dataset, check_cluedataset, \
|
|
@@ -61,18 +62,17 @@ class AGNewsDataset(SourceDataset, TextBaseDataset):
|
|
|
61
62
|
Set the mode of data shuffling by passing in enumeration variables:
|
|
62
63
|
|
|
63
64
|
- ``Shuffle.GLOBAL``: Shuffle both the files and samples.
|
|
64
|
-
|
|
65
65
|
- ``Shuffle.FILES``: Shuffle files only.
|
|
66
66
|
|
|
67
67
|
num_shards (int, optional): Number of shards that the dataset will be divided into.
|
|
68
68
|
Default: ``None``. When this argument is specified, `num_samples` reflects the
|
|
69
69
|
max sample number of per shard.
|
|
70
|
-
Used in `data parallel training <https://www.mindspore.cn/
|
|
71
|
-
parallel/data_parallel.html#
|
|
70
|
+
Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
|
|
71
|
+
parallel/data_parallel.html#loading-datasets>`_ .
|
|
72
72
|
shard_id (int, optional): The shard ID within `num_shards` . This
|
|
73
73
|
argument can only be specified when `num_shards` is also specified. Default: ``None``.
|
|
74
74
|
cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
|
|
75
|
-
`Single-Node Data Cache <https://www.mindspore.cn/
|
|
75
|
+
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
|
|
76
76
|
Default: ``None``, which means no cache is used.
|
|
77
77
|
|
|
78
78
|
Raises:
|
|
@@ -90,9 +90,9 @@ class AGNewsDataset(SourceDataset, TextBaseDataset):
|
|
|
90
90
|
>>> ag_news_dataset_dir = "/path/to/ag_news_dataset_file"
|
|
91
91
|
>>> dataset = ds.AGNewsDataset(dataset_dir=ag_news_dataset_dir, usage='all')
|
|
92
92
|
|
|
93
|
-
About
|
|
93
|
+
About AG News dataset:
|
|
94
94
|
|
|
95
|
-
AG is a collection of over 1 million news articles. The news articles were collected
|
|
95
|
+
AG News is a collection of over 1 million news articles. The news articles were collected
|
|
96
96
|
by ComeToMyHead from over 2,000 news sources in over 1 year of activity. ComeToMyHead
|
|
97
97
|
is an academic news search engine that has been in operation since July 2004.
|
|
98
98
|
The dataset is provided by academics for research purposes such as data mining
|
|
@@ -172,17 +172,16 @@ class AmazonReviewDataset(SourceDataset, TextBaseDataset):
|
|
|
172
172
|
Set the mode of data shuffling by passing in enumeration variables:
|
|
173
173
|
|
|
174
174
|
- ``Shuffle.GLOBAL`` : Shuffle both the files and samples.
|
|
175
|
-
|
|
176
175
|
- ``Shuffle.FILES`` : Shuffle files only.
|
|
177
176
|
|
|
178
177
|
num_shards (int, optional): Number of shards that the dataset will be divided into. Default: ``None`` .
|
|
179
178
|
When this argument is specified, `num_samples` reflects the max sample number of per shard.
|
|
180
|
-
Used in `data parallel training <https://www.mindspore.cn/
|
|
181
|
-
parallel/data_parallel.html#
|
|
179
|
+
Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
|
|
180
|
+
parallel/data_parallel.html#loading-datasets>`_ .
|
|
182
181
|
shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
|
|
183
182
|
argument can only be specified when `num_shards` is also specified.
|
|
184
183
|
cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
|
|
185
|
-
`Single-Node Data Cache <https://www.mindspore.cn/
|
|
184
|
+
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
|
|
186
185
|
Default: ``None`` , which means no cache is used.
|
|
187
186
|
|
|
188
187
|
Raises:
|
|
@@ -272,17 +271,16 @@ class CLUEDataset(SourceDataset, TextBaseDataset):
|
|
|
272
271
|
There are three levels of shuffling, desired shuffle enum defined by :class:`mindspore.dataset.Shuffle` .
|
|
273
272
|
|
|
274
273
|
- ``Shuffle.GLOBAL`` : Shuffle both the files and samples, same as setting `shuffle` to ``True``.
|
|
275
|
-
|
|
276
274
|
- ``Shuffle.FILES`` : Shuffle files only.
|
|
277
275
|
|
|
278
276
|
num_shards (int, optional): Number of shards that the dataset will be divided into. Default: ``None`` .
|
|
279
277
|
When this argument is specified, `num_samples` reflects the maximum sample number of per shard.
|
|
280
|
-
Used in `data parallel training <https://www.mindspore.cn/
|
|
281
|
-
parallel/data_parallel.html#
|
|
278
|
+
Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
|
|
279
|
+
parallel/data_parallel.html#loading-datasets>`_ .
|
|
282
280
|
shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
|
|
283
281
|
argument can only be specified when `num_shards` is also specified.
|
|
284
282
|
cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
|
|
285
|
-
`Single-Node Data Cache <https://www.mindspore.cn/
|
|
283
|
+
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
|
|
286
284
|
Default: ``None`` , which means no cache is used.
|
|
287
285
|
|
|
288
286
|
The generated dataset with different task setting has different output columns:
|
|
@@ -521,15 +519,15 @@ class CoNLL2000Dataset(SourceDataset, TextBaseDataset):
|
|
|
521
519
|
|
|
522
520
|
num_shards (int, optional): Number of shards that the dataset will be divided into.
|
|
523
521
|
When this argument is specified, `num_samples` reflects the max sample number of per shard.
|
|
524
|
-
Default: ``None`` . Used in `data parallel training <https://www.mindspore.cn/
|
|
525
|
-
parallel/data_parallel.html#
|
|
522
|
+
Default: ``None`` . Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
|
|
523
|
+
parallel/data_parallel.html#loading-datasets>`_ .
|
|
526
524
|
shard_id (int, optional): The shard ID within `num_shards` . This
|
|
527
525
|
argument can only be specified when `num_shards` is also specified. Default: ``None`` .
|
|
528
526
|
num_parallel_workers (int, optional): Number of worker threads to read the data.
|
|
529
527
|
Default: ``None`` , will use global default workers(8), it can be set
|
|
530
528
|
by :func:`mindspore.dataset.config.set_num_parallel_workers` .
|
|
531
529
|
cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
|
|
532
|
-
`Single-Node Data Cache <https://www.mindspore.cn/
|
|
530
|
+
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
|
|
533
531
|
Default: ``None`` , which means no cache is used.
|
|
534
532
|
|
|
535
533
|
Raises:
|
|
@@ -620,17 +618,16 @@ class DBpediaDataset(SourceDataset, TextBaseDataset):
|
|
|
620
618
|
Set the mode of data shuffling by passing in enumeration variables:
|
|
621
619
|
|
|
622
620
|
- ``Shuffle.GLOBAL`` : Shuffle both the files and samples.
|
|
623
|
-
|
|
624
621
|
- ``Shuffle.FILES`` : Shuffle files only.
|
|
625
622
|
|
|
626
623
|
num_shards (int, optional): Number of shards that the dataset will be divided into. Default: ``None`` .
|
|
627
624
|
When this argument is specified, `num_samples` reflects the maximum sample number of per shard.
|
|
628
|
-
Used in `data parallel training <https://www.mindspore.cn/
|
|
629
|
-
parallel/data_parallel.html#
|
|
625
|
+
Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
|
|
626
|
+
parallel/data_parallel.html#loading-datasets>`_ .
|
|
630
627
|
shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
|
|
631
628
|
argument can only be specified when `num_shards` is also specified.
|
|
632
629
|
cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
|
|
633
|
-
`Single-Node Data Cache <https://www.mindspore.cn/
|
|
630
|
+
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
|
|
634
631
|
Default: ``None`` , which means no cache is used.
|
|
635
632
|
|
|
636
633
|
Raises:
|
|
@@ -721,17 +718,16 @@ class EnWik9Dataset(SourceDataset, TextBaseDataset):
|
|
|
721
718
|
Set the mode of data shuffling by passing in enumeration variables:
|
|
722
719
|
|
|
723
720
|
- ``Shuffle.GLOBAL`` : Shuffle both the files and samples.
|
|
724
|
-
|
|
725
721
|
- ``Shuffle.FILES`` : Shuffle files only.
|
|
726
722
|
|
|
727
723
|
num_shards (int, optional): Number of shards that the dataset will be divided into. Default: ``None`` .
|
|
728
724
|
When this argument is specified, `num_samples` reflects the maximum sample number of per shard.
|
|
729
|
-
Used in `data parallel training <https://www.mindspore.cn/
|
|
730
|
-
parallel/data_parallel.html#
|
|
725
|
+
Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
|
|
726
|
+
parallel/data_parallel.html#loading-datasets>`_ .
|
|
731
727
|
shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
|
|
732
728
|
argument can only be specified when `num_shards` is also specified.
|
|
733
729
|
cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
|
|
734
|
-
`Single-Node Data Cache <https://www.mindspore.cn/
|
|
730
|
+
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
|
|
735
731
|
Default: ``None`` , which means no cache is used.
|
|
736
732
|
|
|
737
733
|
Raises:
|
|
@@ -816,12 +812,12 @@ class IMDBDataset(MappableDataset, TextBaseDataset):
|
|
|
816
812
|
num_shards (int, optional): Number of shards that the dataset will be divided
|
|
817
813
|
into. Default: ``None`` . When this argument is specified, `num_samples` reflects
|
|
818
814
|
the maximum sample number of per shard.
|
|
819
|
-
Used in `data parallel training <https://www.mindspore.cn/
|
|
820
|
-
parallel/data_parallel.html#
|
|
815
|
+
Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
|
|
816
|
+
parallel/data_parallel.html#loading-datasets>`_ .
|
|
821
817
|
shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
|
|
822
818
|
argument can only be specified when `num_shards` is also specified.
|
|
823
819
|
cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
|
|
824
|
-
`Single-Node Data Cache <https://www.mindspore.cn/
|
|
820
|
+
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
|
|
825
821
|
Default: ``None`` , which means no cache is used.
|
|
826
822
|
|
|
827
823
|
Raises:
|
|
@@ -949,20 +945,19 @@ class IWSLT2016Dataset(SourceDataset, TextBaseDataset):
|
|
|
949
945
|
Set the mode of data shuffling by passing in enumeration variables:
|
|
950
946
|
|
|
951
947
|
- ``Shuffle.GLOBAL`` : Shuffle both the files and samples.
|
|
952
|
-
|
|
953
948
|
- ``Shuffle.FILES`` : Shuffle files only.
|
|
954
949
|
|
|
955
950
|
num_shards (int, optional): Number of shards that the dataset will be divided into. Default: ``None`` .
|
|
956
951
|
When this argument is specified, `num_samples` reflects the max sample number of per shard.
|
|
957
|
-
Used in `data parallel training <https://www.mindspore.cn/
|
|
958
|
-
parallel/data_parallel.html#
|
|
952
|
+
Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
|
|
953
|
+
parallel/data_parallel.html#loading-datasets>`_ .
|
|
959
954
|
shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
|
|
960
955
|
argument can only be specified when `num_shards` is also specified.
|
|
961
956
|
num_parallel_workers (int, optional): Number of worker threads to read the data.
|
|
962
957
|
Default: ``None`` , will use global default workers(8), it can be set
|
|
963
958
|
by :func:`mindspore.dataset.config.set_num_parallel_workers` .
|
|
964
959
|
cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
|
|
965
|
-
`Single-Node Data Cache <https://www.mindspore.cn/
|
|
960
|
+
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
|
|
966
961
|
Default: ``None`` , which means no cache is used.
|
|
967
962
|
|
|
968
963
|
Raises:
|
|
@@ -1083,20 +1078,19 @@ class IWSLT2017Dataset(SourceDataset, TextBaseDataset):
|
|
|
1083
1078
|
Set the mode of data shuffling by passing in enumeration variables:
|
|
1084
1079
|
|
|
1085
1080
|
- ``Shuffle.GLOBAL`` : Shuffle both the files and samples.
|
|
1086
|
-
|
|
1087
1081
|
- ``Shuffle.FILES`` : Shuffle files only.
|
|
1088
1082
|
|
|
1089
1083
|
num_shards (int, optional): Number of shards that the dataset will be divided into. Default: ``None`` .
|
|
1090
1084
|
When this argument is specified, `num_samples` reflects the max sample number of per shard.
|
|
1091
|
-
Used in `data parallel training <https://www.mindspore.cn/
|
|
1092
|
-
parallel/data_parallel.html#
|
|
1085
|
+
Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
|
|
1086
|
+
parallel/data_parallel.html#loading-datasets>`_ .
|
|
1093
1087
|
shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
|
|
1094
1088
|
argument can only be specified when `num_shards` is also specified.
|
|
1095
1089
|
num_parallel_workers (int, optional): Number of worker threads to read the data.
|
|
1096
1090
|
Default: ``None`` , will use global default workers(8), it can be set
|
|
1097
1091
|
by :func:`mindspore.dataset.config.set_num_parallel_workers` .
|
|
1098
1092
|
cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
|
|
1099
|
-
`Single-Node Data Cache <https://www.mindspore.cn/
|
|
1093
|
+
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
|
|
1100
1094
|
Default: ``None`` , which means no cache is used.
|
|
1101
1095
|
|
|
1102
1096
|
Raises:
|
|
@@ -1198,12 +1192,12 @@ class Multi30kDataset(SourceDataset, TextBaseDataset):
|
|
|
1198
1192
|
num_shards (int, optional): Number of shards that the dataset will be divided
|
|
1199
1193
|
into. Default: ``None`` . When this argument is specified, `num_samples` reflects
|
|
1200
1194
|
the max sample number of per shard.
|
|
1201
|
-
Used in `data parallel training <https://www.mindspore.cn/
|
|
1202
|
-
parallel/data_parallel.html#
|
|
1195
|
+
Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
|
|
1196
|
+
parallel/data_parallel.html#loading-datasets>`_ .
|
|
1203
1197
|
shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
|
|
1204
1198
|
argument can only be specified when `num_shards` is also specified.
|
|
1205
1199
|
cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
|
|
1206
|
-
`Single-Node Data Cache <https://www.mindspore.cn/
|
|
1200
|
+
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
|
|
1207
1201
|
Default: ``None`` , which means no cache is used.
|
|
1208
1202
|
|
|
1209
1203
|
Raises:
|
|
@@ -1229,7 +1223,7 @@ class Multi30kDataset(SourceDataset, TextBaseDataset):
|
|
|
1229
1223
|
|
|
1230
1224
|
Multi30K is a multilingual dataset that features approximately 31,000 standardized images
|
|
1231
1225
|
described in multiple languages. The images are sourced from Flickr and each image comes
|
|
1232
|
-
with sentence
|
|
1226
|
+
with sentence descriptions in both English and German, as well as descriptions in other
|
|
1233
1227
|
languages. Multi30k is used primarily for training and testing in tasks such as image
|
|
1234
1228
|
captioning, machine translation, and visual question answering.
|
|
1235
1229
|
|
|
@@ -1304,17 +1298,16 @@ class PennTreebankDataset(SourceDataset, TextBaseDataset):
|
|
|
1304
1298
|
Set the mode of data shuffling by passing in enumeration variables:
|
|
1305
1299
|
|
|
1306
1300
|
- ``Shuffle.GLOBAL`` : Shuffle both the files and samples.
|
|
1307
|
-
|
|
1308
1301
|
- ``Shuffle.FILES`` : Shuffle files only.
|
|
1309
1302
|
|
|
1310
1303
|
num_shards (int, optional): Number of shards that the dataset will be divided into. Default: ``None`` .
|
|
1311
1304
|
When this argument is specified, `num_samples` reflects the max sample number of per shard.
|
|
1312
|
-
Used in `data parallel training <https://www.mindspore.cn/
|
|
1313
|
-
parallel/data_parallel.html#
|
|
1305
|
+
Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
|
|
1306
|
+
parallel/data_parallel.html#loading-datasets>`_ .
|
|
1314
1307
|
shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
|
|
1315
1308
|
argument can only be specified when `num_shards` is also specified.
|
|
1316
1309
|
cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
|
|
1317
|
-
`Single-Node Data Cache <https://www.mindspore.cn/
|
|
1310
|
+
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
|
|
1318
1311
|
Default: ``None`` , which means no cache is used.
|
|
1319
1312
|
|
|
1320
1313
|
Raises:
|
|
@@ -1335,7 +1328,7 @@ class PennTreebankDataset(SourceDataset, TextBaseDataset):
|
|
|
1335
1328
|
About PennTreebank dataset:
|
|
1336
1329
|
|
|
1337
1330
|
Penn Treebank (PTB) dataset, is widely used in machine learning for NLP (Natural Language Processing)
|
|
1338
|
-
research. Word-level PTB does not contain capital letters, numbers, and
|
|
1331
|
+
research. Word-level PTB does not contain capital letters, numbers, and punctuation, and the vocabulary
|
|
1339
1332
|
is capped at 10k unique words, which is relatively small in comparison to most modern datasets which
|
|
1340
1333
|
can result in a larger number of out of vocabulary tokens.
|
|
1341
1334
|
|
|
@@ -1406,19 +1399,19 @@ class SogouNewsDataset(SourceDataset, TextBaseDataset):
|
|
|
1406
1399
|
Set the mode of data shuffling by passing in enumeration variables:
|
|
1407
1400
|
|
|
1408
1401
|
- ``Shuffle.GLOBAL`` : Shuffle both the files and samples, same as setting shuffle to True.
|
|
1409
|
-
|
|
1410
1402
|
- ``Shuffle.FILES`` : Shuffle files only.
|
|
1403
|
+
|
|
1411
1404
|
num_shards (int, optional): Number of shards that the dataset will be divided into. Default: ``None`` .
|
|
1412
1405
|
When this argument is specified, `num_samples` reflects the max sample number of per shard.
|
|
1413
|
-
Used in `data parallel training <https://www.mindspore.cn/
|
|
1414
|
-
parallel/data_parallel.html#
|
|
1406
|
+
Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
|
|
1407
|
+
parallel/data_parallel.html#loading-datasets>`_ .
|
|
1415
1408
|
shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
|
|
1416
1409
|
argument can only be specified when `num_shards` is also specified.
|
|
1417
1410
|
num_parallel_workers (int, optional): Number of worker threads to read the data.
|
|
1418
1411
|
Default: ``None`` , will use global default workers(8), it can be set
|
|
1419
1412
|
by :func:`mindspore.dataset.config.set_num_parallel_workers` .
|
|
1420
1413
|
cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
|
|
1421
|
-
`Single-Node Data Cache <https://www.mindspore.cn/
|
|
1414
|
+
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
|
|
1422
1415
|
Default: ``None`` , which means no cache is used.
|
|
1423
1416
|
|
|
1424
1417
|
Raises:
|
|
@@ -1436,11 +1429,11 @@ class SogouNewsDataset(SourceDataset, TextBaseDataset):
|
|
|
1436
1429
|
>>> sogou_news_dataset_dir = "/path/to/sogou_news_dataset_dir"
|
|
1437
1430
|
>>> dataset = ds.SogouNewsDataset(dataset_dir=sogou_news_dataset_dir, usage='all')
|
|
1438
1431
|
|
|
1439
|
-
About
|
|
1432
|
+
About Sogou News Dataset:
|
|
1440
1433
|
|
|
1441
|
-
|
|
1434
|
+
Sogou News dataset includes 3 columns, corresponding to class index (1 to 5), title and content. The title and
|
|
1442
1435
|
content are escaped using double quotes ("), and any internal double quote is escaped by 2 double quotes ("").
|
|
1443
|
-
New lines are escaped by a backslash followed with an "n" character, that is "
|
|
1436
|
+
New lines are escaped by a backslash followed with an "n" character, that is "\\n".
|
|
1444
1437
|
|
|
1445
1438
|
You can unzip the dataset files into the following structure and read by MindSpore's API:
|
|
1446
1439
|
|
|
@@ -1513,12 +1506,12 @@ class SQuADDataset(SourceDataset, TextBaseDataset):
|
|
|
1513
1506
|
|
|
1514
1507
|
num_shards (int, optional): Number of shards that the dataset will be divided into. Default: ``None`` .
|
|
1515
1508
|
When this argument is specified, `num_samples` reflects the maximum sample number of per shard.
|
|
1516
|
-
Used in `data parallel training <https://www.mindspore.cn/
|
|
1517
|
-
parallel/data_parallel.html#
|
|
1509
|
+
Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
|
|
1510
|
+
parallel/data_parallel.html#loading-datasets>`_ .
|
|
1518
1511
|
shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
|
|
1519
1512
|
argument can only be specified when `num_shards` is also specified.
|
|
1520
1513
|
cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
|
|
1521
|
-
`Single-Node Data Cache <https://www.mindspore.cn/
|
|
1514
|
+
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
|
|
1522
1515
|
Default: ``None`` , which means no cache is used.
|
|
1523
1516
|
|
|
1524
1517
|
Raises:
|
|
@@ -1629,16 +1622,17 @@ class SST2Dataset(SourceDataset, TextBaseDataset):
|
|
|
1629
1622
|
If `shuffle` is ``True`` , the behavior is the same as setting shuffle to be Shuffle.GLOBAL
|
|
1630
1623
|
Set the mode of data shuffling by passing in enumeration variables:
|
|
1631
1624
|
|
|
1632
|
-
- ``Shuffle.GLOBAL`` : Shuffle the samples.
|
|
1625
|
+
- ``Shuffle.GLOBAL`` : Shuffle both the files and samples.
|
|
1626
|
+
- ``Shuffle.FILES`` : Shuffle files only.
|
|
1633
1627
|
|
|
1634
1628
|
num_shards (int, optional): Number of shards that the dataset will be divided into. Default: ``None`` .
|
|
1635
1629
|
When this argument is specified, `num_samples` reflects the maximum sample number of per shard.
|
|
1636
|
-
Used in `data parallel training <https://www.mindspore.cn/
|
|
1637
|
-
parallel/data_parallel.html#
|
|
1630
|
+
Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
|
|
1631
|
+
parallel/data_parallel.html#loading-datasets>`_ .
|
|
1638
1632
|
shard_id (int, optional): The shard ID within `num_shards`. This argument can only be specified when
|
|
1639
1633
|
`num_shards` is also specified. Default: ``None`` .
|
|
1640
1634
|
cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
|
|
1641
|
-
`Single-Node Data Cache <https://www.mindspore.cn/
|
|
1635
|
+
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
|
|
1642
1636
|
Default: ``None`` , which means no cache is used.
|
|
1643
1637
|
|
|
1644
1638
|
Raises:
|
|
@@ -1733,17 +1727,16 @@ class TextFileDataset(SourceDataset, TextBaseDataset):
|
|
|
1733
1727
|
There are three levels of shuffling, desired shuffle enum defined by :class:`mindspore.dataset.Shuffle` .
|
|
1734
1728
|
|
|
1735
1729
|
- ``Shuffle.GLOBAL`` : Shuffle both the files and samples, same as setting shuffle to True.
|
|
1736
|
-
|
|
1737
1730
|
- ``Shuffle.FILES`` : Shuffle files only.
|
|
1738
1731
|
|
|
1739
1732
|
num_shards (int, optional): Number of shards that the dataset will be divided into. Default: ``None`` .
|
|
1740
1733
|
When this argument is specified, `num_samples` reflects the maximum sample number of per shard.
|
|
1741
|
-
Used in `data parallel training <https://www.mindspore.cn/
|
|
1742
|
-
parallel/data_parallel.html#
|
|
1734
|
+
Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
|
|
1735
|
+
parallel/data_parallel.html#loading-datasets>`_ .
|
|
1743
1736
|
shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
|
|
1744
1737
|
argument can only be specified when `num_shards` is also specified.
|
|
1745
1738
|
cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
|
|
1746
|
-
`Single-Node Data Cache <https://www.mindspore.cn/
|
|
1739
|
+
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
|
|
1747
1740
|
Default: ``None`` , which means no cache is used.
|
|
1748
1741
|
|
|
1749
1742
|
Raises:
|
|
@@ -1799,20 +1792,19 @@ class UDPOSDataset(SourceDataset, TextBaseDataset):
|
|
|
1799
1792
|
Set the mode of data shuffling by passing in enumeration variables:
|
|
1800
1793
|
|
|
1801
1794
|
- ``Shuffle.GLOBAL`` : Shuffle both the files and samples.
|
|
1802
|
-
|
|
1803
1795
|
- ``Shuffle.FILES`` : Shuffle files only.
|
|
1804
1796
|
|
|
1805
1797
|
num_shards (int, optional): Number of shards that the dataset will be divided into. Default: ``None`` .
|
|
1806
1798
|
When this argument is specified, `num_samples` reflects the max sample number of per shard.
|
|
1807
|
-
Used in `data parallel training <https://www.mindspore.cn/
|
|
1808
|
-
parallel/data_parallel.html#
|
|
1799
|
+
Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
|
|
1800
|
+
parallel/data_parallel.html#loading-datasets>`_ .
|
|
1809
1801
|
shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
|
|
1810
1802
|
argument can only be specified when `num_shards` is also specified.
|
|
1811
1803
|
num_parallel_workers (int, optional): Number of worker threads to read the data.
|
|
1812
1804
|
Default: ``None`` , will use global default workers(8), it can be set
|
|
1813
1805
|
by :func:`mindspore.dataset.config.set_num_parallel_workers` .
|
|
1814
1806
|
cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
|
|
1815
|
-
`Single-Node Data Cache <https://www.mindspore.cn/
|
|
1807
|
+
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
|
|
1816
1808
|
Default: ``None`` , which means no cache is used.
|
|
1817
1809
|
|
|
1818
1810
|
Raises:
|
|
@@ -1887,17 +1879,16 @@ class WikiTextDataset(SourceDataset, TextBaseDataset):
|
|
|
1887
1879
|
Set the mode of data shuffling by passing in enumeration variables:
|
|
1888
1880
|
|
|
1889
1881
|
- ``Shuffle.GLOBAL`` : Shuffle both the files and samples.
|
|
1890
|
-
|
|
1891
1882
|
- ``Shuffle.FILES`` : Shuffle files only.
|
|
1892
1883
|
|
|
1893
1884
|
num_shards (int, optional): Number of shards that the dataset will be divided into. Default: ``None`` .
|
|
1894
1885
|
When this argument is specified, `num_samples` reflects the max sample number of per shard.
|
|
1895
|
-
Used in `data parallel training <https://www.mindspore.cn/
|
|
1896
|
-
parallel/data_parallel.html#
|
|
1886
|
+
Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
|
|
1887
|
+
parallel/data_parallel.html#loading-datasets>`_ .
|
|
1897
1888
|
shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
|
|
1898
1889
|
argument can only be specified when `num_shards` is also specified.
|
|
1899
1890
|
cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
|
|
1900
|
-
`Single-Node Data Cache <https://www.mindspore.cn/
|
|
1891
|
+
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
|
|
1901
1892
|
Default: ``None`` , which means no cache is used.
|
|
1902
1893
|
|
|
1903
1894
|
Raises:
|
|
@@ -1986,17 +1977,16 @@ class YahooAnswersDataset(SourceDataset, TextBaseDataset):
|
|
|
1986
1977
|
Set the mode of data shuffling by passing in enumeration variables:
|
|
1987
1978
|
|
|
1988
1979
|
- ``Shuffle.GLOBAL`` : Shuffle both the files and samples.
|
|
1989
|
-
|
|
1990
1980
|
- ``Shuffle.FILES`` : Shuffle files only.
|
|
1991
1981
|
|
|
1992
1982
|
num_shards (int, optional): Number of shards that the dataset will be divided into. Default: ``None`` .
|
|
1993
1983
|
When this argument is specified, `num_samples` reflects the maximum sample number of per shard.
|
|
1994
|
-
Used in `data parallel training <https://www.mindspore.cn/
|
|
1995
|
-
parallel/data_parallel.html#
|
|
1984
|
+
Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
|
|
1985
|
+
parallel/data_parallel.html#loading-datasets>`_ .
|
|
1996
1986
|
shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
|
|
1997
1987
|
argument can only be specified when `num_shards` is also specified.
|
|
1998
1988
|
cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
|
|
1999
|
-
`Single-Node Data Cache <https://www.mindspore.cn/
|
|
1989
|
+
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
|
|
2000
1990
|
Default: ``None`` , which means no cache is used.
|
|
2001
1991
|
|
|
2002
1992
|
Raises:
|
|
@@ -2074,11 +2064,14 @@ class YelpReviewDataset(SourceDataset, TextBaseDataset):
|
|
|
2074
2064
|
Args:
|
|
2075
2065
|
dataset_dir (str): Path to the root directory that contains the dataset.
|
|
2076
2066
|
usage (str, optional): Usage of this dataset, can be ``'train'`` , ``'test'`` or ``'all'`` .
|
|
2077
|
-
|
|
2078
|
-
|
|
2079
|
-
``'
|
|
2080
|
-
|
|
2081
|
-
|
|
2067
|
+
Default: ``None`` , all samples.
|
|
2068
|
+
|
|
2069
|
+
- For Polarity, ``'train'`` will read from 560,000 train samples,
|
|
2070
|
+
``'test'`` will read from 38,000 test samples,
|
|
2071
|
+
``'all'`` will read from all 598,000 samples.
|
|
2072
|
+
- For Full, ``'train'`` will read from 650,000 train samples, ``'test'`` will read from 50,000 test samples,
|
|
2073
|
+
``'all'`` will read from all 700,000 samples.
|
|
2074
|
+
|
|
2082
2075
|
num_samples (int, optional): Number of samples (rows) to read. Default: ``None`` , reads all samples.
|
|
2083
2076
|
shuffle (Union[bool, Shuffle], optional): Perform reshuffling of the data every epoch.
|
|
2084
2077
|
Bool type and Shuffle enum are both supported to pass in.
|
|
@@ -2089,19 +2082,19 @@ class YelpReviewDataset(SourceDataset, TextBaseDataset):
|
|
|
2089
2082
|
Set the mode of data shuffling by passing in enumeration variables:
|
|
2090
2083
|
|
|
2091
2084
|
- ``Shuffle.GLOBAL`` : Shuffle both the files and samples.
|
|
2092
|
-
|
|
2093
2085
|
- ``Shuffle.FILES`` : Shuffle files only.
|
|
2086
|
+
|
|
2094
2087
|
num_shards (int, optional): Number of shards that the dataset will be divided into. Default: ``None`` .
|
|
2095
2088
|
When this argument is specified, `num_samples` reflects the max sample number of per shard.
|
|
2096
|
-
Used in `data parallel training <https://www.mindspore.cn/
|
|
2097
|
-
parallel/data_parallel.html#
|
|
2089
|
+
Used in `data parallel training <https://www.mindspore.cn/tutorials/en/master/
|
|
2090
|
+
parallel/data_parallel.html#loading-datasets>`_ .
|
|
2098
2091
|
shard_id (int, optional): The shard ID within `num_shards` . Default: ``None`` . This
|
|
2099
2092
|
argument can only be specified when `num_shards` is also specified.
|
|
2100
2093
|
num_parallel_workers (int, optional): Number of worker threads to read the data.
|
|
2101
2094
|
Default: ``None`` , will use global default workers(8), it can be set
|
|
2102
2095
|
by :func:`mindspore.dataset.config.set_num_parallel_workers` .
|
|
2103
2096
|
cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
|
|
2104
|
-
`Single-Node Data Cache <https://www.mindspore.cn/
|
|
2097
|
+
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/en/master/dataset/cache.html>`_ .
|
|
2105
2098
|
Default: ``None`` , which means no cache is used.
|
|
2106
2099
|
|
|
2107
2100
|
Raises:
|