mindspore 2.0.0a0__cp39-cp39-win_amd64.whl → 2.0.0rc1__cp39-cp39-win_amd64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of mindspore might be problematic. Click here for more details.
- mindspore/.commit_id +1 -1
- mindspore/__init__.py +4 -2
- mindspore/_c_dataengine.cp39-win_amd64.pyd +0 -0
- mindspore/_c_expression.cp39-win_amd64.pyd +0 -0
- mindspore/_c_mindrecord.cp39-win_amd64.pyd +0 -0
- mindspore/_check_jit_forbidden_api.py +102 -0
- mindspore/_checkparam.py +1066 -1001
- mindspore/_extends/parallel_compile/akg_compiler/akg_process.py +4 -3
- mindspore/_extends/parallel_compile/akg_compiler/tbe_topi.py +50 -48
- mindspore/_extends/parallel_compile/akg_compiler/util.py +9 -4
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_adapter.py +4 -4
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_helper.py +9 -4
- mindspore/_extends/parse/__init__.py +5 -3
- mindspore/_extends/parse/namespace.py +16 -1
- mindspore/_extends/parse/parser.py +107 -22
- mindspore/_extends/parse/resources.py +0 -7
- mindspore/_extends/parse/standard_method.py +885 -413
- mindspore/amp.py +52 -57
- mindspore/boost/boost.py +2 -2
- mindspore/boost/boost_cell_wrapper.py +38 -20
- mindspore/boost/dim_reduce.py +3 -3
- mindspore/boost/group_loss_scale_manager.py +1 -1
- mindspore/common/__init__.py +4 -6
- mindspore/common/_decorator.py +2 -0
- mindspore/common/_register_for_adapter.py +55 -0
- mindspore/common/_stub_tensor.py +201 -0
- mindspore/common/_utils.py +41 -7
- mindspore/common/api.py +215 -141
- mindspore/common/dtype.py +8 -1
- mindspore/common/dump.py +2 -2
- mindspore/common/initializer.py +4 -2
- mindspore/common/jit_config.py +17 -13
- mindspore/common/mutable.py +33 -13
- mindspore/common/parameter.py +23 -21
- mindspore/common/seed.py +8 -24
- mindspore/common/sparse_tensor.py +62 -41
- mindspore/common/tensor.py +852 -1154
- mindspore/communication/__init__.py +2 -2
- mindspore/communication/_comm_helper.py +11 -4
- mindspore/communication/management.py +22 -21
- mindspore/config/op_info.config +501 -1008
- mindspore/context.py +201 -23
- mindspore/dataset/__init__.py +6 -6
- mindspore/dataset/audio/__init__.py +7 -7
- mindspore/dataset/audio/transforms.py +670 -30
- mindspore/dataset/audio/utils.py +47 -4
- mindspore/dataset/audio/validators.py +223 -1
- mindspore/dataset/callback/ds_callback.py +2 -2
- mindspore/dataset/core/config.py +210 -14
- mindspore/dataset/core/validator_helpers.py +2 -2
- mindspore/{parallel/nn/layers.py → dataset/debug/__init__.py} +7 -8
- mindspore/dataset/debug/debug_hook.py +65 -0
- mindspore/dataset/debug/pre_defined_hook.py +67 -0
- mindspore/dataset/engine/__init__.py +7 -3
- mindspore/dataset/engine/cache_client.py +1 -1
- mindspore/dataset/engine/datasets.py +322 -66
- mindspore/dataset/engine/datasets_audio.py +80 -76
- mindspore/dataset/engine/datasets_standard_format.py +51 -38
- mindspore/dataset/engine/datasets_text.py +232 -118
- mindspore/dataset/engine/datasets_user_defined.py +41 -17
- mindspore/dataset/engine/datasets_vision.py +746 -225
- mindspore/dataset/engine/graphdata.py +75 -10
- mindspore/dataset/engine/iterators.py +45 -5
- mindspore/dataset/engine/offload.py +48 -28
- mindspore/dataset/engine/validators.py +117 -8
- mindspore/dataset/text/__init__.py +6 -5
- mindspore/dataset/text/transforms.py +86 -3
- mindspore/dataset/text/utils.py +6 -4
- mindspore/dataset/text/validators.py +25 -0
- mindspore/dataset/transforms/__init__.py +3 -2
- mindspore/dataset/transforms/c_transforms.py +1 -1
- mindspore/dataset/transforms/transforms.py +2 -2
- mindspore/dataset/utils/__init__.py +2 -1
- mindspore/dataset/utils/line_reader.py +121 -0
- mindspore/dataset/vision/__init__.py +2 -3
- mindspore/dataset/vision/c_transforms.py +9 -9
- mindspore/dataset/vision/py_transforms.py +5 -5
- mindspore/dataset/vision/py_transforms_util.py +2 -0
- mindspore/dataset/vision/transforms.py +160 -161
- mindspore/dataset/vision/utils.py +3 -3
- mindspore/experimental/map_parameter.py +38 -26
- mindspore/include/OWNERS +0 -1
- mindspore/include/api/callback/callback.h +9 -13
- mindspore/include/api/callback/ckpt_saver.h +2 -2
- mindspore/include/api/callback/loss_monitor.h +2 -2
- mindspore/include/api/callback/lr_scheduler.h +5 -5
- mindspore/include/api/callback/time_monitor.h +2 -2
- mindspore/include/api/callback/train_accuracy.h +4 -6
- mindspore/include/api/cfg.h +19 -6
- mindspore/include/api/context.h +44 -9
- mindspore/include/api/delegate.h +1 -1
- mindspore/include/api/metrics/accuracy.h +2 -2
- mindspore/include/api/metrics/metrics.h +4 -3
- mindspore/include/api/model.h +9 -4
- mindspore/include/api/model_parallel_runner.h +2 -2
- mindspore/include/api/net.h +12 -11
- mindspore/include/api/serialization.h +19 -3
- mindspore/include/api/types.h +3 -3
- mindspore/include/dataset/constants.h +7 -0
- mindspore/include/dataset/text.h +59 -0
- mindspore/jpeg62.dll +0 -0
- mindspore/log.py +1 -1
- mindspore/mindrecord/filereader.py +18 -0
- mindspore/mindrecord/filewriter.py +197 -34
- mindspore/mindrecord/shardreader.py +9 -0
- mindspore/mindrecord/shardwriter.py +1 -1
- mindspore/mindrecord/tools/cifar100_to_mr.py +3 -3
- mindspore/mindrecord/tools/cifar10_to_mr.py +3 -3
- mindspore/mindrecord/tools/csv_to_mr.py +3 -3
- mindspore/mindrecord/tools/imagenet_to_mr.py +16 -11
- mindspore/mindrecord/tools/mnist_to_mr.py +2 -2
- mindspore/mindrecord/tools/tfrecord_to_mr.py +6 -6
- mindspore/mindspore_backend.dll +0 -0
- mindspore/mindspore_common.dll +0 -0
- mindspore/mindspore_core.dll +0 -0
- mindspore/mindspore_glog.dll +0 -0
- mindspore/mindspore_shared_lib.dll +0 -0
- mindspore/nn/__init__.py +0 -4
- mindspore/nn/cell.py +204 -132
- mindspore/nn/dynamic_lr.py +1 -1
- mindspore/nn/grad/cell_grad.py +7 -6
- mindspore/nn/layer/__init__.py +5 -4
- mindspore/nn/layer/activation.py +40 -89
- mindspore/nn/layer/basic.py +255 -624
- mindspore/nn/layer/channel_shuffle.py +7 -6
- mindspore/nn/layer/combined.py +1 -1
- mindspore/nn/layer/container.py +41 -4
- mindspore/nn/layer/conv.py +64 -28
- mindspore/nn/layer/dense.py +9 -8
- mindspore/nn/layer/embedding.py +27 -25
- mindspore/nn/layer/image.py +53 -46
- mindspore/nn/layer/math.py +97 -105
- mindspore/nn/layer/normalization.py +117 -86
- mindspore/nn/layer/padding.py +185 -95
- mindspore/nn/layer/pooling.py +817 -414
- mindspore/nn/layer/rnn_cells.py +10 -15
- mindspore/nn/layer/rnns.py +37 -38
- mindspore/nn/layer/thor_layer.py +11 -12
- mindspore/nn/layer/timedistributed.py +5 -5
- mindspore/nn/layer/transformer.py +701 -0
- mindspore/nn/learning_rate_schedule.py +8 -8
- mindspore/nn/loss/__init__.py +5 -4
- mindspore/nn/loss/loss.py +334 -199
- mindspore/nn/optim/ada_grad.py +6 -6
- mindspore/nn/optim/adadelta.py +2 -3
- mindspore/nn/optim/adafactor.py +4 -5
- mindspore/nn/optim/adam.py +126 -62
- mindspore/nn/optim/adamax.py +3 -4
- mindspore/nn/optim/adasum.py +6 -6
- mindspore/nn/optim/asgd.py +2 -2
- mindspore/nn/optim/ftrl.py +67 -38
- mindspore/nn/optim/lamb.py +4 -5
- mindspore/nn/optim/lars.py +2 -2
- mindspore/nn/optim/lazyadam.py +43 -4
- mindspore/nn/optim/momentum.py +6 -5
- mindspore/nn/optim/optimizer.py +3 -1
- mindspore/nn/optim/proximal_ada_grad.py +2 -2
- mindspore/nn/optim/rmsprop.py +1 -1
- mindspore/nn/optim/rprop.py +8 -9
- mindspore/nn/optim/sgd.py +19 -13
- mindspore/nn/optim/thor.py +10 -15
- mindspore/nn/probability/__init__.py +0 -2
- mindspore/nn/probability/bijector/bijector.py +4 -4
- mindspore/nn/probability/bijector/invert.py +1 -1
- mindspore/nn/probability/bijector/softplus.py +2 -2
- mindspore/nn/probability/bnn_layers/dense_variational.py +1 -1
- mindspore/nn/probability/bnn_layers/layer_distribution.py +2 -2
- mindspore/nn/probability/distribution/_utils/utils.py +9 -15
- mindspore/nn/probability/distribution/bernoulli.py +3 -3
- mindspore/nn/probability/distribution/beta.py +1 -1
- mindspore/nn/probability/distribution/categorical.py +5 -7
- mindspore/nn/probability/distribution/cauchy.py +3 -3
- mindspore/nn/probability/distribution/distribution.py +2 -2
- mindspore/nn/probability/distribution/exponential.py +2 -2
- mindspore/nn/probability/distribution/gamma.py +3 -3
- mindspore/nn/probability/distribution/geometric.py +1 -1
- mindspore/nn/probability/distribution/gumbel.py +3 -3
- mindspore/nn/probability/distribution/half_normal.py +15 -11
- mindspore/nn/probability/distribution/laplace.py +16 -13
- mindspore/nn/probability/distribution/logistic.py +2 -2
- mindspore/nn/probability/distribution/normal.py +1 -1
- mindspore/nn/probability/distribution/poisson.py +1 -1
- mindspore/nn/probability/distribution/student_t.py +20 -15
- mindspore/nn/probability/distribution/transformed_distribution.py +4 -4
- mindspore/nn/probability/distribution/uniform.py +2 -2
- mindspore/nn/reinforcement/_tensors_queue.py +3 -3
- mindspore/nn/reinforcement/tensor_array.py +2 -2
- mindspore/nn/sparse/sparse.py +2 -2
- mindspore/nn/wrap/cell_wrapper.py +27 -10
- mindspore/nn/wrap/grad_reducer.py +2 -2
- mindspore/nn/wrap/loss_scale.py +40 -24
- mindspore/numpy/array_creations.py +33 -22
- mindspore/numpy/array_ops.py +35 -30
- mindspore/numpy/logic_ops.py +6 -27
- mindspore/numpy/math_ops.py +22 -19
- mindspore/numpy/utils.py +1 -1
- mindspore/numpy/utils_const.py +108 -58
- mindspore/opencv_core452.dll +0 -0
- mindspore/opencv_imgcodecs452.dll +0 -0
- mindspore/opencv_imgproc452.dll +0 -0
- mindspore/ops/_constants.py +0 -6
- mindspore/ops/_grad/__init__.py +2 -1
- mindspore/ops/_grad/grad_array_ops.py +86 -117
- mindspore/ops/_grad/grad_base.py +23 -1
- mindspore/ops/_grad/grad_clip_ops.py +2 -3
- mindspore/ops/_grad/grad_comm_ops.py +34 -24
- mindspore/ops/_grad/grad_implementations.py +9 -45
- mindspore/ops/_grad/grad_inner_ops.py +47 -4
- mindspore/ops/_grad/grad_math_ops.py +142 -117
- mindspore/ops/_grad/grad_nn_ops.py +71 -165
- mindspore/ops/_grad/grad_sequence_ops.py +296 -0
- mindspore/ops/_grad/grad_sparse.py +7 -6
- mindspore/ops/_grad_experimental/__init__.py +1 -0
- mindspore/ops/_grad_experimental/grad_array_ops.py +150 -15
- mindspore/ops/_grad_experimental/grad_image_ops.py +16 -7
- mindspore/ops/_grad_experimental/grad_inner_ops.py +1 -22
- mindspore/ops/_grad_experimental/grad_linalg_ops.py +4 -11
- mindspore/ops/_grad_experimental/grad_math_ops.py +210 -89
- mindspore/ops/_grad_experimental/grad_nn_ops.py +26 -22
- mindspore/ops/_grad_experimental/grad_scalar_ops.py +112 -0
- mindspore/ops/_grad_experimental/grad_sparse_ops.py +49 -8
- mindspore/ops/_op_impl/_custom_op/batch_matmul_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/batchnorm_fold.py +2 -2
- mindspore/ops/_op_impl/_custom_op/batchnorm_fold2.py +2 -2
- mindspore/ops/_op_impl/_custom_op/batchnorm_fold2_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/batchnorm_fold2_grad_reduce.py +4 -4
- mindspore/ops/_op_impl/_custom_op/batchnorm_fold_grad.py +3 -3
- mindspore/ops/_op_impl/_custom_op/cholesky_trsm_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/correction_mul.py +2 -2
- mindspore/ops/_op_impl/_custom_op/correction_mul_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/dsd_back_impl.py +1 -5
- mindspore/ops/_op_impl/_custom_op/dsd_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perchannel.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perchannel_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perchannel_grad_reduce.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perlayer.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perlayer_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_learned_scale_quant_perlayer_grad_reduce.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_quant_perchannel.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_quant_perchannel_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_quant_perlayer.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fake_quant_perlayer_grad.py +2 -2
- mindspore/ops/_op_impl/_custom_op/fused_abs_max1_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/img2col_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/matmul_cube_dense_left_impl.py +2 -2
- mindspore/ops/_op_impl/_custom_op/matmul_cube_dense_right_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/matmul_cube_fracz_left_cast_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/matmul_cube_fracz_right_mul_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/matmul_cube_impl.py +2 -2
- mindspore/ops/_op_impl/_custom_op/matmul_dds_impl.py +0 -4
- mindspore/ops/_op_impl/_custom_op/matrix_combine_impl.py +1 -1
- mindspore/ops/_op_impl/_custom_op/minmax_update_perchannel.py +2 -2
- mindspore/ops/_op_impl/_custom_op/minmax_update_perlayer.py +2 -2
- mindspore/ops/_op_impl/_custom_op/transpose02314_impl.py +1 -1
- mindspore/ops/_op_impl/aicpu/__init__.py +236 -4
- mindspore/ops/_op_impl/aicpu/abs.py +36 -0
- mindspore/ops/_op_impl/aicpu/{adaptive_avg_pool_2d_v1.py → adaptive_avg_pool_2d.py} +6 -5
- mindspore/ops/_op_impl/aicpu/adaptive_avg_pool_2d_grad.py +34 -0
- mindspore/ops/_op_impl/aicpu/add.py +43 -0
- mindspore/ops/_op_impl/aicpu/addcdiv.py +0 -32
- mindspore/ops/_op_impl/aicpu/addcmul.py +0 -84
- mindspore/ops/_op_impl/aicpu/affine_grid_grad.py +35 -0
- mindspore/ops/_op_impl/aicpu/batch_matmul.py +43 -43
- mindspore/ops/_op_impl/aicpu/bernoulli.py +48 -0
- mindspore/{compression/common/__init__.py → ops/_op_impl/aicpu/bessel_i0.py} +15 -8
- mindspore/ops/_op_impl/aicpu/channel_shuffle.py +40 -0
- mindspore/ops/_op_impl/aicpu/conj.py +11 -0
- mindspore/ops/_op_impl/aicpu/cumulative_logsumexp.py +0 -3
- mindspore/ops/_op_impl/aicpu/deformable_offsets.py +38 -0
- mindspore/ops/_op_impl/aicpu/deformable_offsets_grad.py +43 -0
- mindspore/ops/_op_impl/aicpu/{adaptive_avg_pool_2d_grad_v1.py → digamma.py} +7 -9
- mindspore/ops/_op_impl/aicpu/flatten.py +1 -0
- mindspore/ops/_op_impl/aicpu/fmax.py +36 -0
- mindspore/ops/_op_impl/aicpu/fmin.py +37 -0
- mindspore/ops/_op_impl/aicpu/fractional_max_pool3d_with_fixed_ksize.py +1 -1
- mindspore/ops/_op_impl/aicpu/fse_decode.py +43 -0
- mindspore/ops/_op_impl/aicpu/greater.py +41 -0
- mindspore/ops/_op_impl/aicpu/greater_equal.py +41 -0
- mindspore/ops/_op_impl/aicpu/index_put.py +50 -0
- mindspore/ops/_op_impl/aicpu/less.py +41 -0
- mindspore/{nn/probability/infer/variational/__init__.py → ops/_op_impl/aicpu/lgamma.py} +16 -10
- mindspore/ops/_op_impl/aicpu/mirror_pad.py +0 -4
- mindspore/ops/_op_impl/aicpu/mirror_pad_grad.py +0 -4
- mindspore/ops/_op_impl/aicpu/mul.py +3 -1
- mindspore/ops/_op_impl/aicpu/multinomial.py +14 -6
- mindspore/ops/_op_impl/aicpu/nllloss.py +38 -0
- mindspore/ops/_op_impl/aicpu/nllloss_grad.py +39 -0
- mindspore/ops/_op_impl/aicpu/ones_like.py +0 -2
- mindspore/ops/_op_impl/aicpu/polar.py +32 -0
- mindspore/ops/_op_impl/aicpu/polygamma.py +34 -0
- mindspore/ops/_op_impl/aicpu/quant_dtype_cast.py +40 -0
- mindspore/ops/_op_impl/aicpu/quantile.py +35 -0
- mindspore/ops/_op_impl/aicpu/ragged_tensor_to_sparse.py +73 -0
- mindspore/ops/_op_impl/aicpu/randperm_v2.py +41 -0
- mindspore/ops/_op_impl/aicpu/resize_bicubic.py +2 -8
- mindspore/ops/_op_impl/aicpu/resize_bicubic_grad.py +1 -1
- mindspore/ops/_op_impl/aicpu/resize_v2.py +68 -0
- mindspore/ops/_op_impl/aicpu/resize_v2_grad.py +68 -0
- mindspore/ops/_op_impl/aicpu/scatter_elements.py +4 -0
- mindspore/ops/_op_impl/aicpu/scatter_nd_update.py +2 -0
- mindspore/ops/_op_impl/aicpu/sequence_add.py +34 -0
- mindspore/ops/_op_impl/aicpu/sequence_add_offset.py +34 -0
- mindspore/ops/_op_impl/aicpu/sequence_addn.py +38 -0
- mindspore/ops/_op_impl/aicpu/smooth_l1_loss.py +35 -0
- mindspore/ops/_op_impl/aicpu/smooth_l1_loss_grad.py +37 -0
- mindspore/ops/_op_impl/aicpu/sparse_apply_adagrad_da.py +0 -24
- mindspore/ops/_op_impl/aicpu/sparse_cross.py +42 -0
- mindspore/ops/_op_impl/aicpu/sparse_slice.py +4 -0
- mindspore/ops/_op_impl/aicpu/sparse_slice_grad.py +6 -0
- mindspore/ops/_op_impl/aicpu/tensor_scatter_update.py +59 -0
- mindspore/ops/_op_impl/aicpu/trans_data.py +1 -0
- mindspore/ops/_op_impl/aicpu/tril_indices.py +34 -0
- mindspore/ops/_op_impl/aicpu/uniform.py +34 -0
- mindspore/ops/_op_impl/aicpu/uniform_candidate_sampler.py +1 -0
- mindspore/ops/_op_impl/aicpu/unique_consecutive.py +10 -2
- mindspore/ops/_op_impl/cpu/dynamic_shape.py +5 -1
- mindspore/ops/_op_impl/cpu/sparse_slice.py +4 -0
- mindspore/ops/_op_impl/cpu/sparse_slice_grad.py +6 -0
- mindspore/ops/_op_impl/cpu/tensor_shape.py +5 -1
- mindspore/ops/_op_impl/tbe/__init__.py +27 -611
- mindspore/ops/_op_impl/tbe/assign_add_ds.py +1 -0
- mindspore/ops/_op_impl/tbe/atomic_addr_clean.py +1 -1
- mindspore/ops/_op_impl/tbe/avg_pool_3d_grad.py +1 -1
- mindspore/ops/_op_impl/tbe/batch_matmul_ds.py +1 -0
- mindspore/ops/_op_impl/tbe/batch_to_space.py +1 -1
- mindspore/ops/_op_impl/tbe/batch_to_space_nd.py +1 -1
- mindspore/ops/_op_impl/tbe/bn_infer_grad.py +4 -2
- mindspore/ops/_op_impl/tbe/bn_training_update.py +0 -1
- mindspore/ops/_op_impl/tbe/bn_training_update_ds.py +0 -1
- mindspore/ops/_op_impl/tbe/broadcast_to_ds.py +6 -4
- mindspore/ops/_op_impl/tbe/cast.py +0 -2
- mindspore/ops/_op_impl/tbe/cast_ds.py +3 -3
- mindspore/ops/_op_impl/tbe/data_format_dim_map_ds.py +1 -0
- mindspore/ops/_op_impl/tbe/depthwise_conv2d.py +2 -2
- mindspore/ops/_op_impl/tbe/dynamic_atomic_addr_clean.py +1 -1
- mindspore/ops/_op_impl/tbe/gather_nd.py +1 -0
- mindspore/ops/_op_impl/tbe/{index_add.py → inplace_index_add.py} +3 -6
- mindspore/ops/_op_impl/tbe/matmul_ds.py +2 -0
- mindspore/ops/_op_impl/tbe/npu_clear_float_status_v2.py +35 -0
- mindspore/ops/_op_impl/tbe/npu_get_float_status_v2.py +35 -0
- mindspore/ops/_op_impl/tbe/scatter_mul.py +2 -0
- mindspore/ops/_op_impl/tbe/scatter_nd_add.py +0 -2
- mindspore/ops/_op_impl/tbe/space_to_batch.py +1 -1
- mindspore/ops/_op_impl/tbe/space_to_batch_nd.py +1 -1
- mindspore/ops/_op_impl/tbe/trans_data_ds.py +15 -5
- mindspore/ops/_register_for_op.py +1 -0
- mindspore/ops/_utils/__init__.py +1 -2
- mindspore/ops/_utils/utils.py +19 -40
- mindspore/ops/_vmap/vmap_array_ops.py +116 -38
- mindspore/ops/_vmap/vmap_base.py +16 -9
- mindspore/ops/_vmap/vmap_convolution_ops.py +7 -10
- mindspore/ops/_vmap/vmap_grad_math_ops.py +4 -4
- mindspore/ops/_vmap/vmap_grad_nn_ops.py +7 -5
- mindspore/ops/_vmap/vmap_image_ops.py +12 -5
- mindspore/ops/_vmap/vmap_math_ops.py +46 -5
- mindspore/ops/_vmap/vmap_nn_ops.py +15 -21
- mindspore/ops/_vmap/vmap_random_ops.py +1 -1
- mindspore/ops/bprop_mindir/AdaptiveAvgPool2D_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/AdaptiveMaxPool2D_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/AvgPool3D_bprop.mindir +150 -0
- mindspore/ops/bprop_mindir/AvgPool_bprop.mindir +66 -0
- mindspore/ops/bprop_mindir/BCEWithLogitsLoss_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/BatchNormGrad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/BiasAddGrad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/BinaryCrossEntropy_bprop.mindir +33 -0
- mindspore/ops/bprop_mindir/BroadcastTo_bprop.mindir +220 -106
- mindspore/ops/bprop_mindir/CTCLoss_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Conv2DBackpropFilter_bprop.mindir +240 -0
- mindspore/ops/bprop_mindir/Conv2DBackpropInput_bprop.mindir +247 -0
- mindspore/ops/bprop_mindir/Conv2DTranspose_bprop.mindir +247 -0
- mindspore/ops/bprop_mindir/Conv3DTranspose_bprop.mindir +315 -0
- mindspore/ops/bprop_mindir/Conv3D_bprop.mindir +278 -0
- mindspore/ops/bprop_mindir/DeformableOffsets_bprop.mindir +58 -0
- mindspore/ops/bprop_mindir/DepthwiseConv2dNative_bprop.mindir +138 -0
- mindspore/ops/bprop_mindir/Dropout2D_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Dropout3D_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/DropoutDoMask_bprop.mindir +22 -23
- mindspore/ops/bprop_mindir/DropoutGenMask_bprop.mindir +16 -17
- mindspore/ops/bprop_mindir/DropoutGrad_bprop.mindir +27 -0
- mindspore/ops/bprop_mindir/Dropout_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/DynamicGRUV2_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/DynamicRNN_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Elu_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/EmbeddingLookup_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/ExpandDims_bprop.mindir +39 -41
- mindspore/ops/bprop_mindir/FastGeLU_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/Flatten_bprop.mindir +41 -43
- mindspore/ops/bprop_mindir/GatherNd_bprop.mindir +51 -57
- mindspore/ops/bprop_mindir/Gather_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/HSigmoid_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/HSwish_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/InstanceNorm_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/KLDivLoss_bprop.mindir +126 -0
- mindspore/ops/bprop_mindir/L2Loss_bprop.mindir +15 -0
- mindspore/ops/bprop_mindir/L2Normalize_bprop.mindir +30 -0
- mindspore/ops/bprop_mindir/LRN_bprop.mindir +43 -0
- mindspore/ops/bprop_mindir/LayerNormGrad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/LogSoftmax_bprop.mindir +23 -0
- mindspore/ops/bprop_mindir/MaxPool3DGradGrad_bprop.mindir +74 -0
- mindspore/ops/bprop_mindir/MaxPool3DGrad_bprop.mindir +74 -0
- mindspore/ops/bprop_mindir/MaxPool3D_bprop.mindir +75 -0
- mindspore/ops/bprop_mindir/MaxPoolGradGrad_bprop.mindir +65 -0
- mindspore/ops/bprop_mindir/MaxPoolWithArgmax_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/MirrorPad_bprop.mindir +27 -0
- mindspore/ops/bprop_mindir/Mish_bprop.mindir +35 -0
- mindspore/ops/bprop_mindir/MulNoNan_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/NLLLoss_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/OneHot_bprop.mindir +24 -25
- mindspore/ops/bprop_mindir/PReLU_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Pad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Padding_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/RNNTLoss_bprop.mindir +29 -0
- mindspore/ops/bprop_mindir/ROIAlign_bprop.mindir +82 -0
- mindspore/ops/bprop_mindir/ReLU6_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/ReLUV2_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/ReluGrad_bprop.mindir +18 -19
- mindspore/ops/bprop_mindir/Reshape_bprop.mindir +53 -53
- mindspore/ops/bprop_mindir/ResizeBilinear_bprop.mindir +29 -0
- mindspore/ops/bprop_mindir/ResizeNearestNeighbor_bprop.mindir +77 -85
- mindspore/ops/bprop_mindir/SeLU_bprop.mindir +21 -0
- mindspore/ops/bprop_mindir/SigmoidCrossEntropyWithLogits_bprop.mindir +21 -0
- mindspore/ops/bprop_mindir/SigmoidGrad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Sigmoid_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/SmoothL1Loss_bprop.mindir +36 -0
- mindspore/ops/bprop_mindir/SoftmaxCrossEntropyWithLogits_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Softplus_bprop.mindir +16 -0
- mindspore/ops/bprop_mindir/Softsign_bprop.mindir +33 -0
- mindspore/ops/bprop_mindir/SparseSoftmaxCrossEntropyWithLogits_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Squeeze_bprop.mindir +37 -39
- mindspore/ops/bprop_mindir/StridedSlice_bprop.mindir +70 -72
- mindspore/ops/bprop_mindir/TanhGrad_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/Tanh_bprop.mindir +66 -0
- mindspore/ops/bprop_mindir/Tile_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/TopK_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/TupleGetItem_bprop.mindir +17 -17
- mindspore/ops/bprop_mindir/UpsampleNearest3D_bprop.mindir +32 -0
- mindspore/ops/bprop_mindir/UpsampleTrilinear3D_bprop.mindir +38 -0
- mindspore/ops/bprop_mindir/generate_mindir.py +2 -0
- mindspore/ops/composite/__init__.py +7 -8
- mindspore/ops/composite/base.py +101 -47
- mindspore/ops/composite/math_ops.py +188 -158
- mindspore/ops/composite/multitype_ops/_compile_utils.py +415 -170
- mindspore/ops/composite/multitype_ops/_constexpr_utils.py +142 -87
- mindspore/ops/composite/multitype_ops/add_impl.py +6 -1
- mindspore/ops/composite/multitype_ops/div_impl.py +2 -3
- mindspore/ops/composite/multitype_ops/getitem_impl.py +31 -3
- mindspore/ops/composite/multitype_ops/greater_equal_impl.py +31 -0
- mindspore/ops/composite/multitype_ops/greater_impl.py +31 -0
- mindspore/ops/composite/multitype_ops/in_impl.py +9 -0
- mindspore/ops/composite/multitype_ops/less_equal_impl.py +31 -0
- mindspore/ops/composite/multitype_ops/less_impl.py +31 -0
- mindspore/ops/composite/multitype_ops/mul_impl.py +21 -5
- mindspore/ops/composite/multitype_ops/not_in_impl.py +9 -0
- mindspore/ops/composite/multitype_ops/ones_like_impl.py +2 -4
- mindspore/ops/composite/multitype_ops/setitem_impl.py +21 -3
- mindspore/ops/composite/multitype_ops/sub_impl.py +1 -1
- mindspore/ops/composite/multitype_ops/zeros_like_impl.py +35 -4
- mindspore/ops/function/__init__.py +152 -8
- mindspore/ops/function/array_func.py +2555 -674
- mindspore/ops/function/clip_func.py +209 -13
- mindspore/ops/function/debug_func.py +2 -2
- mindspore/ops/function/grad/__init__.py +2 -1
- mindspore/ops/function/grad/grad_func.py +147 -62
- mindspore/ops/function/image_func.py +54 -38
- mindspore/ops/function/linalg_func.py +167 -16
- mindspore/ops/function/math_func.py +4849 -1492
- mindspore/ops/function/nn_func.py +2573 -988
- mindspore/ops/function/other_func.py +115 -0
- mindspore/ops/function/parameter_func.py +3 -3
- mindspore/ops/function/random_func.py +790 -73
- mindspore/ops/function/sparse_func.py +98 -78
- mindspore/ops/function/sparse_unary_func.py +54 -53
- mindspore/ops/function/spectral_func.py +27 -24
- mindspore/ops/function/vmap_func.py +22 -2
- mindspore/ops/functional.py +97 -37
- mindspore/ops/op_info_register.py +70 -28
- mindspore/ops/operations/__init__.py +47 -14
- mindspore/ops/operations/_csr_ops.py +7 -7
- mindspore/ops/operations/_embedding_cache_ops.py +5 -5
- mindspore/ops/operations/_grad_ops.py +276 -187
- mindspore/ops/operations/_inner_ops.py +319 -113
- mindspore/ops/operations/_ms_kernel.py +10 -8
- mindspore/ops/operations/_ocr_ops.py +9 -9
- mindspore/ops/operations/_opaque_predicate_registry.py +4 -0
- mindspore/ops/operations/_quant_ops.py +137 -102
- mindspore/ops/operations/_rl_inner_ops.py +121 -60
- mindspore/ops/operations/_scalar_ops.py +466 -0
- mindspore/ops/operations/_sequence_ops.py +1004 -2
- mindspore/ops/operations/_tensor_array.py +10 -11
- mindspore/ops/operations/_thor_ops.py +1 -1
- mindspore/ops/operations/array_ops.py +801 -466
- mindspore/ops/operations/comm_ops.py +51 -49
- mindspore/ops/operations/control_ops.py +2 -2
- mindspore/ops/operations/custom_ops.py +123 -44
- mindspore/ops/operations/debug_ops.py +24 -24
- mindspore/ops/operations/image_ops.py +240 -153
- mindspore/ops/operations/inner_ops.py +34 -50
- mindspore/ops/operations/linalg_ops.py +31 -9
- mindspore/ops/operations/math_ops.py +988 -757
- mindspore/ops/operations/nn_ops.py +965 -819
- mindspore/ops/operations/other_ops.py +51 -40
- mindspore/ops/operations/random_ops.py +204 -122
- mindspore/ops/operations/rl_ops.py +8 -9
- mindspore/ops/operations/sparse_ops.py +254 -93
- mindspore/ops/operations/spectral_ops.py +35 -3
- mindspore/ops/primitive.py +111 -9
- mindspore/parallel/_auto_parallel_context.py +189 -83
- mindspore/parallel/_offload_context.py +185 -0
- mindspore/parallel/_parallel_serialization.py +99 -7
- mindspore/parallel/_ps_context.py +9 -5
- mindspore/parallel/_recovery_context.py +1 -1
- mindspore/parallel/_tensor.py +7 -1
- mindspore/{nn/transformer → parallel/_transformer}/__init__.py +6 -6
- mindspore/{nn/transformer → parallel/_transformer}/layers.py +6 -37
- mindspore/{nn/transformer → parallel/_transformer}/loss.py +4 -7
- mindspore/{nn/transformer → parallel/_transformer}/moe.py +20 -16
- mindspore/{nn/transformer → parallel/_transformer}/op_parallel_config.py +3 -3
- mindspore/{nn/transformer → parallel/_transformer}/transformer.py +48 -111
- mindspore/parallel/_utils.py +1 -2
- mindspore/parallel/algo_parameter_config.py +1 -1
- mindspore/parallel/checkpoint_transform.py +37 -34
- mindspore/parallel/shard.py +17 -18
- mindspore/profiler/common/validator/validate_path.py +2 -2
- mindspore/profiler/envprofiling.py +69 -47
- mindspore/profiler/parser/ascend_timeline_generator.py +49 -42
- mindspore/profiler/parser/base_timeline_generator.py +49 -56
- mindspore/profiler/parser/cpu_gpu_timeline_generator.py +98 -78
- mindspore/profiler/parser/hwts_log_parser.py +1 -1
- mindspore/profiler/parser/integrator.py +15 -14
- mindspore/profiler/parser/minddata_analyzer.py +2 -2
- mindspore/profiler/parser/msadvisor_analyzer.py +12 -25
- mindspore/profiler/parser/msadvisor_parser.py +2 -4
- mindspore/profiler/parser/optime_parser.py +17 -18
- mindspore/profiler/parser/profiler_info.py +2 -1
- mindspore/profiler/profiling.py +218 -186
- mindspore/rewrite/__init__.py +3 -1
- mindspore/rewrite/api/node.py +1 -114
- mindspore/rewrite/api/node_type.py +3 -0
- mindspore/rewrite/api/pattern_engine.py +31 -1
- mindspore/rewrite/api/scoped_value.py +4 -4
- mindspore/rewrite/api/symbol_tree.py +3 -78
- mindspore/rewrite/api/tree_node_helper.py +1 -1
- mindspore/rewrite/ast_creator_register.py +1 -0
- mindspore/rewrite/ast_helpers/__init__.py +2 -2
- mindspore/rewrite/ast_helpers/ast_creator.py +1 -2
- mindspore/rewrite/ast_helpers/ast_finder.py +65 -0
- mindspore/rewrite/ast_helpers/ast_modifier.py +11 -3
- mindspore/rewrite/ast_transformers/flatten_recursive_stmt.py +18 -2
- mindspore/rewrite/namespace.py +0 -2
- mindspore/rewrite/node.py +157 -11
- mindspore/rewrite/parsers/assign_parser.py +231 -53
- mindspore/rewrite/parsers/class_def_parser.py +187 -109
- mindspore/rewrite/parsers/for_parser.py +24 -14
- mindspore/rewrite/parsers/function_def_parser.py +21 -4
- mindspore/rewrite/parsers/if_parser.py +6 -2
- mindspore/rewrite/sparsify/__init__.py +0 -0
- mindspore/rewrite/sparsify/sparse_transformer.py +448 -0
- mindspore/rewrite/sparsify/sparsify.py +109 -0
- mindspore/rewrite/sparsify/utils.py +173 -0
- mindspore/rewrite/symbol_tree.py +256 -133
- mindspore/rewrite/symbol_tree_builder.py +38 -1
- mindspore/run_check/_check_version.py +69 -63
- mindspore/run_check/run_check.py +2 -1
- mindspore/tinyxml2.dll +0 -0
- mindspore/train/__init__.py +1 -1
- mindspore/train/_utils.py +28 -5
- mindspore/train/amp.py +273 -102
- mindspore/train/callback/_backup_and_restore.py +5 -5
- mindspore/train/callback/_callback.py +2 -2
- mindspore/train/callback/_checkpoint.py +3 -3
- mindspore/train/callback/_early_stop.py +3 -3
- mindspore/train/callback/_lambda_callback.py +2 -2
- mindspore/train/callback/_landscape.py +29 -31
- mindspore/train/callback/_loss_monitor.py +3 -3
- mindspore/train/callback/_on_request_exit.py +3 -3
- mindspore/train/callback/_reduce_lr_on_plateau.py +4 -4
- mindspore/train/callback/_summary_collector.py +23 -16
- mindspore/train/callback/_time_monitor.py +3 -3
- mindspore/train/checkpoint_pb2.py +68 -8
- mindspore/train/data_sink.py +15 -3
- mindspore/train/dataset_helper.py +10 -15
- mindspore/train/loss_scale_manager.py +8 -11
- mindspore/train/metrics/__init__.py +1 -1
- mindspore/train/metrics/bleu_score.py +1 -1
- mindspore/train/metrics/confusion_matrix.py +1 -1
- mindspore/train/metrics/cosine_similarity.py +1 -1
- mindspore/train/metrics/dice.py +2 -2
- mindspore/train/metrics/fbeta.py +1 -1
- mindspore/train/metrics/hausdorff_distance.py +4 -3
- mindspore/train/metrics/mean_surface_distance.py +2 -2
- mindspore/train/metrics/occlusion_sensitivity.py +1 -1
- mindspore/train/metrics/perplexity.py +1 -1
- mindspore/train/metrics/precision.py +1 -1
- mindspore/train/metrics/recall.py +1 -1
- mindspore/train/metrics/roc.py +2 -2
- mindspore/train/metrics/root_mean_square_surface_distance.py +2 -2
- mindspore/train/mind_ir_pb2.py +116 -37
- mindspore/train/model.py +45 -28
- mindspore/train/serialization.py +295 -188
- mindspore/train/summary/_summary_adapter.py +1 -1
- mindspore/train/summary/summary_record.py +43 -13
- mindspore/train/train_thor/convert_utils.py +2 -2
- mindspore/train/train_thor/dataset_helper.py +3 -3
- mindspore/turbojpeg.dll +0 -0
- mindspore/version.py +1 -1
- {mindspore-2.0.0a0.dist-info → mindspore-2.0.0rc1.dist-info}/METADATA +3 -2
- {mindspore-2.0.0a0.dist-info → mindspore-2.0.0rc1.dist-info}/RECORD +610 -541
- mindspore/compression/__init__.py +0 -19
- mindspore/compression/common/constant.py +0 -124
- mindspore/compression/export/__init__.py +0 -19
- mindspore/compression/export/quant_export.py +0 -515
- mindspore/compression/quant/__init__.py +0 -28
- mindspore/compression/quant/qat.py +0 -634
- mindspore/compression/quant/quant_utils.py +0 -462
- mindspore/compression/quant/quantizer.py +0 -68
- mindspore/nn/layer/quant.py +0 -1868
- mindspore/nn/layer/rnn_utils.py +0 -90
- mindspore/nn/probability/dpn/__init__.py +0 -22
- mindspore/nn/probability/dpn/vae/__init__.py +0 -25
- mindspore/nn/probability/dpn/vae/cvae.py +0 -140
- mindspore/nn/probability/dpn/vae/vae.py +0 -124
- mindspore/nn/probability/infer/__init__.py +0 -22
- mindspore/nn/probability/infer/variational/elbo.py +0 -70
- mindspore/nn/probability/infer/variational/svi.py +0 -84
- mindspore/nn/probability/toolbox/__init__.py +0 -22
- mindspore/nn/probability/toolbox/anomaly_detection.py +0 -99
- mindspore/nn/probability/toolbox/uncertainty_evaluation.py +0 -364
- mindspore/nn/probability/transforms/__init__.py +0 -22
- mindspore/nn/probability/transforms/transform_bnn.py +0 -262
- mindspore/nn/probability/zhusuan/__init__.py +0 -18
- mindspore/nn/probability/zhusuan/framework/__init__.py +0 -18
- mindspore/nn/probability/zhusuan/framework/bn.py +0 -95
- mindspore/nn/probability/zhusuan/variational/__init__.py +0 -18
- mindspore/nn/probability/zhusuan/variational/elbo.py +0 -46
- mindspore/ops/_op_impl/aicpu/parallel_concat.py +0 -42
- mindspore/ops/_op_impl/tbe/gather_v2.py +0 -56
- mindspore/ops/bprop_mindir/AssignAdd_bprop.mindir +0 -19
- mindspore/ops/bprop_mindir/Cast_bprop.mindir +0 -19
- mindspore/ops/bprop_mindir/LogicalOr_bprop.mindir +0 -19
- mindspore/ops/bprop_mindir/MatMul_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/ReLU_bprop.mindir +0 -17
- mindspore/ops/bprop_mindir/Transpose_bprop.mindir +0 -0
- mindspore/ops/bprop_mindir/UpdateState_bprop.mindir +0 -15
- mindspore/ops/composite/array_ops.py +0 -241
- mindspore/ops/composite/clip_ops.py +0 -134
- mindspore/ops/composite/random_ops.py +0 -426
- mindspore/ops/composite/vmap_ops.py +0 -38
- mindspore/parallel/nn/__init__.py +0 -42
- mindspore/parallel/nn/loss.py +0 -22
- mindspore/parallel/nn/moe.py +0 -21
- mindspore/parallel/nn/op_parallel_config.py +0 -22
- mindspore/parallel/nn/transformer.py +0 -31
- {mindspore-2.0.0a0.dist-info → mindspore-2.0.0rc1.dist-info}/WHEEL +0 -0
- {mindspore-2.0.0a0.dist-info → mindspore-2.0.0rc1.dist-info}/entry_points.txt +0 -0
- {mindspore-2.0.0a0.dist-info → mindspore-2.0.0rc1.dist-info}/top_level.txt +0 -0
|
@@ -34,9 +34,9 @@ from ..core.validator_helpers import replace_none
|
|
|
34
34
|
|
|
35
35
|
class CMUArcticDataset(MappableDataset, AudioBaseDataset):
|
|
36
36
|
"""
|
|
37
|
-
|
|
37
|
+
CMU Arctic dataset.
|
|
38
38
|
|
|
39
|
-
The generated dataset has four columns: :py:obj:`[
|
|
39
|
+
The generated dataset has four columns: :py:obj:`[waveform, sample_rate, transcript, utterance_id]` .
|
|
40
40
|
The tensor of column :py:obj:`waveform` is of the float32 type.
|
|
41
41
|
The tensor of column :py:obj:`sample_rate` is of a scalar of uint32 type.
|
|
42
42
|
The tensor of column :py:obj:`transcript` is of a scalar of string type.
|
|
@@ -46,63 +46,63 @@ class CMUArcticDataset(MappableDataset, AudioBaseDataset):
|
|
|
46
46
|
dataset_dir (str): Path to the root directory that contains the dataset.
|
|
47
47
|
name (str, optional): Part of this dataset, can be 'aew', 'ahw', 'aup', 'awb', 'axb', 'bdl',
|
|
48
48
|
'clb', 'eey', 'fem', 'gka', 'jmk', 'ksp', 'ljm', 'lnh', 'rms', 'rxr', 'slp' or 'slt'.
|
|
49
|
-
Default: None,
|
|
49
|
+
Default: None, means 'aew'.
|
|
50
50
|
num_samples (int, optional): The number of audio to be included in the dataset.
|
|
51
51
|
Default: None, will read all audio.
|
|
52
|
-
num_parallel_workers (int, optional): Number of
|
|
53
|
-
Default: None, will use
|
|
52
|
+
num_parallel_workers (int, optional): Number of worker threads to read the data.
|
|
53
|
+
Default: None, will use global default workers(8), it can be set
|
|
54
|
+
by `mindspore.dataset.config.set_num_parallel_workers` .
|
|
54
55
|
shuffle (bool, optional): Whether or not to perform shuffle on the dataset.
|
|
55
56
|
Default: None, expected order behavior shown in the table below.
|
|
56
57
|
sampler (Sampler, optional): Object used to choose samples from the
|
|
57
58
|
dataset. Default: None, expected order behavior shown in the table below.
|
|
58
|
-
num_shards (int, optional): Number of shards that the dataset will be divided into. Default: None.
|
|
59
|
+
num_shards (int, optional): Number of shards that the dataset will be divided into. Default: None, no dividing.
|
|
59
60
|
When this argument is specified, `num_samples` reflects the max sample number of per shard.
|
|
60
|
-
shard_id (int, optional): The shard ID within `num_shards` . Default: None. This
|
|
61
|
+
shard_id (int, optional): The shard ID within `num_shards` . Default: None, will use 0. This
|
|
61
62
|
argument can only be specified when `num_shards` is also specified.
|
|
62
63
|
cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
|
|
63
|
-
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/experts/en/r2.0
|
|
64
|
+
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/experts/en/r2.0/dataset/cache.html>`_ .
|
|
64
65
|
Default: None, which means no cache is used.
|
|
65
66
|
|
|
66
67
|
Raises:
|
|
67
|
-
RuntimeError: If source raises an exception during execution.
|
|
68
68
|
RuntimeError: If `dataset_dir` does not contain data files.
|
|
69
69
|
ValueError: If `num_parallel_workers` exceeds the max thread numbers.
|
|
70
70
|
RuntimeError: If `sampler` and `shuffle` are specified at the same time.
|
|
71
71
|
RuntimeError: If `sampler` and `num_shards`/`shard_id` are specified at the same time.
|
|
72
72
|
RuntimeError: If `num_shards` is specified but `shard_id` is None.
|
|
73
73
|
RuntimeError: If `shard_id` is specified but `num_shards` is None.
|
|
74
|
-
ValueError: If `shard_id` is
|
|
74
|
+
ValueError: If `shard_id` is not in range of [0, `num_shards` ).
|
|
75
75
|
|
|
76
76
|
Note:
|
|
77
|
-
-
|
|
77
|
+
- Not support `mindspore.dataset.PKSampler` for `sampler` parameter yet.
|
|
78
78
|
- This dataset can take in a `sampler` . `sampler` and `shuffle` are mutually exclusive.
|
|
79
79
|
The table below shows what input arguments are allowed and their expected behavior.
|
|
80
80
|
|
|
81
|
-
.. list-table:: Expected Order Behavior of Using
|
|
81
|
+
.. list-table:: Expected Order Behavior of Using `sampler` and `shuffle`
|
|
82
82
|
:widths: 25 25 50
|
|
83
83
|
:header-rows: 1
|
|
84
84
|
|
|
85
|
-
|
|
86
|
-
|
|
87
|
-
|
|
88
|
-
|
|
89
|
-
|
|
90
|
-
|
|
91
|
-
|
|
92
|
-
|
|
93
|
-
|
|
94
|
-
|
|
95
|
-
|
|
96
|
-
|
|
97
|
-
|
|
98
|
-
|
|
99
|
-
|
|
100
|
-
|
|
101
|
-
|
|
102
|
-
|
|
103
|
-
|
|
104
|
-
|
|
105
|
-
|
|
85
|
+
* - Parameter `sampler`
|
|
86
|
+
- Parameter `shuffle`
|
|
87
|
+
- Expected Order Behavior
|
|
88
|
+
* - None
|
|
89
|
+
- None
|
|
90
|
+
- random order
|
|
91
|
+
* - None
|
|
92
|
+
- True
|
|
93
|
+
- random order
|
|
94
|
+
* - None
|
|
95
|
+
- False
|
|
96
|
+
- sequential order
|
|
97
|
+
* - Sampler object
|
|
98
|
+
- None
|
|
99
|
+
- order defined by sampler
|
|
100
|
+
* - Sampler object
|
|
101
|
+
- True
|
|
102
|
+
- not allowed
|
|
103
|
+
* - Sampler object
|
|
104
|
+
- False
|
|
105
|
+
- not allowed
|
|
106
106
|
|
|
107
107
|
Examples:
|
|
108
108
|
>>> cmu_arctic_dataset_directory = "/path/to/cmu_arctic_dataset_directory"
|
|
@@ -115,7 +115,7 @@ class CMUArcticDataset(MappableDataset, AudioBaseDataset):
|
|
|
115
115
|
|
|
116
116
|
About CMUArctic dataset:
|
|
117
117
|
|
|
118
|
-
The CMU
|
|
118
|
+
The CMU Arctic databases are designed for the purpose of speech synthesis research.
|
|
119
119
|
These single speaker speech databases have been carefully recorded under studio conditions
|
|
120
120
|
and consist of approximately 1200 phonetically balanced English utterances. In addition to wavefiles,
|
|
121
121
|
the databases provide complete support for the Festival Speech Synthesis System, including pre-built
|
|
@@ -172,9 +172,9 @@ class CMUArcticDataset(MappableDataset, AudioBaseDataset):
|
|
|
172
172
|
|
|
173
173
|
class GTZANDataset(MappableDataset, AudioBaseDataset):
|
|
174
174
|
"""
|
|
175
|
-
|
|
175
|
+
GTZAN dataset.
|
|
176
176
|
|
|
177
|
-
The generated dataset has three columns: :py:obj:`[
|
|
177
|
+
The generated dataset has three columns: :py:obj:`[waveform, sample_rate, label]` .
|
|
178
178
|
The tensor of column :py:obj:`waveform` is of the float32 type.
|
|
179
179
|
The tensor of column :py:obj:`sample_rate` is of a scalar of uint32 type.
|
|
180
180
|
The tensor of column :py:obj:`label` is of a scalar of string type.
|
|
@@ -182,11 +182,12 @@ class GTZANDataset(MappableDataset, AudioBaseDataset):
|
|
|
182
182
|
Args:
|
|
183
183
|
dataset_dir (str): Path to the root directory that contains the dataset.
|
|
184
184
|
usage (str, optional): Usage of this dataset, can be 'train', 'valid', 'test' or 'all'.
|
|
185
|
-
Default: None, all samples.
|
|
185
|
+
Default: None, will read all samples.
|
|
186
186
|
num_samples (int, optional): The number of audio to be included in the dataset.
|
|
187
187
|
Default: None, will read all audio.
|
|
188
|
-
num_parallel_workers (int, optional): Number of
|
|
189
|
-
Default: None, will use
|
|
188
|
+
num_parallel_workers (int, optional): Number of worker threads to read the data.
|
|
189
|
+
Default: None, will use global default workers(8), it can be set
|
|
190
|
+
by `mindspore.dataset.config.set_num_parallel_workers` .
|
|
190
191
|
shuffle (bool, optional): Whether or not to perform shuffle on the dataset.
|
|
191
192
|
Default: None, expected order behavior shown in the table below.
|
|
192
193
|
sampler (Sampler, optional): Object used to choose samples from the
|
|
@@ -196,25 +197,24 @@ class GTZANDataset(MappableDataset, AudioBaseDataset):
|
|
|
196
197
|
shard_id (int, optional): The shard ID within `num_shards` . Default: None. This
|
|
197
198
|
argument can only be specified when `num_shards` is also specified.
|
|
198
199
|
cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
|
|
199
|
-
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/experts/en/r2.0
|
|
200
|
+
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/experts/en/r2.0/dataset/cache.html>`_ .
|
|
200
201
|
Default: None, which means no cache is used.
|
|
201
202
|
|
|
202
203
|
Raises:
|
|
203
|
-
RuntimeError: If source raises an exception during execution.
|
|
204
204
|
RuntimeError: If `dataset_dir` does not contain data files.
|
|
205
205
|
ValueError: If `num_parallel_workers` exceeds the max thread numbers.
|
|
206
206
|
RuntimeError: If `sampler` and `shuffle` are specified at the same time.
|
|
207
207
|
RuntimeError: If `sampler` and `num_shards`/`shard_id` are specified at the same time.
|
|
208
208
|
RuntimeError: If `num_shards` is specified but `shard_id` is None.
|
|
209
209
|
RuntimeError: If `shard_id` is specified but `num_shards` is None.
|
|
210
|
-
ValueError: If `shard_id` is
|
|
210
|
+
ValueError: If `shard_id` is not in range of [0, `num_shards` ).
|
|
211
211
|
|
|
212
212
|
Note:
|
|
213
|
-
-
|
|
213
|
+
- Not support `mindspore.dataset.PKSampler` for `sampler` parameter yet.
|
|
214
214
|
- This dataset can take in a `sampler` . `sampler` and `shuffle` are mutually exclusive.
|
|
215
215
|
The table below shows what input arguments are allowed and their expected behavior.
|
|
216
216
|
|
|
217
|
-
.. list-table:: Expected Order Behavior of Using
|
|
217
|
+
.. list-table:: Expected Order Behavior of Using `sampler` and `shuffle`
|
|
218
218
|
:widths: 25 25 50
|
|
219
219
|
:header-rows: 1
|
|
220
220
|
|
|
@@ -254,7 +254,7 @@ class GTZANDataset(MappableDataset, AudioBaseDataset):
|
|
|
254
254
|
The GTZAN dataset appears in at least 100 published works and is the most commonly used
|
|
255
255
|
public dataset for evaluation in machine listening research for music genre recognition.
|
|
256
256
|
It consists of 1000 audio tracks, each of which is 30 seconds long. It contains 10 genres (blues,
|
|
257
|
-
classical, country, disco, hiphop, jazz, metal, pop, reggae and
|
|
257
|
+
classical, country, disco, hiphop, jazz, metal, pop, reggae and rock), each of which is
|
|
258
258
|
represented by 100 tracks. The tracks are all 22050Hz Mono 16-bit audio files in .wav format.
|
|
259
259
|
|
|
260
260
|
You can construct the following directory structure from GTZAN dataset and read by MindSpore's API.
|
|
@@ -303,10 +303,10 @@ class GTZANDataset(MappableDataset, AudioBaseDataset):
|
|
|
303
303
|
|
|
304
304
|
class LibriTTSDataset(MappableDataset, AudioBaseDataset):
|
|
305
305
|
"""
|
|
306
|
-
|
|
306
|
+
LibriTTS dataset.
|
|
307
307
|
|
|
308
|
-
The generated dataset has seven columns :py:obj:`[
|
|
309
|
-
|
|
308
|
+
The generated dataset has seven columns :py:obj:`[waveform, sample_rate, original_text, normalized_text,
|
|
309
|
+
speaker_id, chapter_id, utterance_id]` .
|
|
310
310
|
The tensor of column :py:obj:`waveform` is of the float32 type.
|
|
311
311
|
The tensor of column :py:obj:`sample_rate` is of a scalar of uint32 type.
|
|
312
312
|
The tensor of column :py:obj:`original_text` is of a scalar of string type.
|
|
@@ -318,11 +318,12 @@ class LibriTTSDataset(MappableDataset, AudioBaseDataset):
|
|
|
318
318
|
Args:
|
|
319
319
|
dataset_dir (str): Path to the root directory that contains the dataset.
|
|
320
320
|
usage (str, optional): Part of this dataset, can be 'dev-clean', 'dev-other', 'test-clean', 'test-other',
|
|
321
|
-
'train-clean-100', 'train-clean-360', 'train-other-500', or 'all'. Default: None,
|
|
321
|
+
'train-clean-100', 'train-clean-360', 'train-other-500', or 'all'. Default: None, means 'all'.
|
|
322
322
|
num_samples (int, optional): The number of images to be included in the dataset.
|
|
323
323
|
Default: None, will read all audio.
|
|
324
|
-
num_parallel_workers (int, optional): Number of
|
|
325
|
-
Default: None, will use
|
|
324
|
+
num_parallel_workers (int, optional): Number of worker threads to read the data.
|
|
325
|
+
Default: None, will use global default workers(8), it can be set
|
|
326
|
+
by `mindspore.dataset.config.set_num_parallel_workers` .
|
|
326
327
|
shuffle (bool, optional): Whether or not to perform shuffle on the dataset.
|
|
327
328
|
Default: None, expected order behavior shown in the table below.
|
|
328
329
|
sampler (Sampler, optional): Object used to choose samples from the
|
|
@@ -332,25 +333,24 @@ class LibriTTSDataset(MappableDataset, AudioBaseDataset):
|
|
|
332
333
|
shard_id (int, optional): The shard ID within `num_shards` . Default: None. This
|
|
333
334
|
argument can only be specified when `num_shards` is also specified.
|
|
334
335
|
cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
|
|
335
|
-
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/experts/en/r2.0
|
|
336
|
+
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/experts/en/r2.0/dataset/cache.html>`_ .
|
|
336
337
|
Default: None, which means no cache is used.
|
|
337
338
|
|
|
338
339
|
Raises:
|
|
339
|
-
RuntimeError: If source raises an exception during execution.
|
|
340
340
|
RuntimeError: If `dataset_dir` does not contain data files.
|
|
341
341
|
ValueError: If `num_parallel_workers` exceeds the max thread numbers.
|
|
342
342
|
RuntimeError: If `sampler` and `shuffle` are specified at the same time.
|
|
343
343
|
RuntimeError: If `sampler` and `num_shards`/`shard_id` are specified at the same time.
|
|
344
344
|
RuntimeError: If `num_shards` is specified but `shard_id` is None.
|
|
345
345
|
RuntimeError: If `shard_id` is specified but `num_shards` is None.
|
|
346
|
-
ValueError: If `shard_id` is
|
|
346
|
+
ValueError: If `shard_id` is not in range of [0, `num_shards` ).
|
|
347
347
|
|
|
348
348
|
Note:
|
|
349
|
-
-
|
|
349
|
+
- Not support `mindspore.dataset.PKSampler` for `sampler` parameter yet.
|
|
350
350
|
- This dataset can take in a `sampler` . `sampler` and `shuffle` are mutually exclusive.
|
|
351
351
|
The table below shows what input arguments are allowed and their expected behavior.
|
|
352
352
|
|
|
353
|
-
.. list-table:: Expected Order Behavior of Using
|
|
353
|
+
.. list-table:: Expected Order Behavior of Using `sampler` and `shuffle`
|
|
354
354
|
:widths: 25 25 50
|
|
355
355
|
:header-rows: 1
|
|
356
356
|
|
|
@@ -454,7 +454,7 @@ class LibriTTSDataset(MappableDataset, AudioBaseDataset):
|
|
|
454
454
|
|
|
455
455
|
class LJSpeechDataset(MappableDataset, AudioBaseDataset):
|
|
456
456
|
"""
|
|
457
|
-
|
|
457
|
+
LJSpeech dataset.
|
|
458
458
|
|
|
459
459
|
The generated dataset has four columns :py:obj:`[waveform, sample_rate, transcription, normalized_transcript]` .
|
|
460
460
|
The column :py:obj:`waveform` is a tensor of the float32 type.
|
|
@@ -466,8 +466,9 @@ class LJSpeechDataset(MappableDataset, AudioBaseDataset):
|
|
|
466
466
|
dataset_dir (str): Path to the root directory that contains the dataset.
|
|
467
467
|
num_samples (int, optional): The number of audios to be included in the dataset.
|
|
468
468
|
Default: None, all audios.
|
|
469
|
-
num_parallel_workers (int, optional): Number of
|
|
470
|
-
Default: None,
|
|
469
|
+
num_parallel_workers (int, optional): Number of worker threads to read the data.
|
|
470
|
+
Default: None, will use global default workers(8), it can be set
|
|
471
|
+
by `mindspore.dataset.config.set_num_parallel_workers` .
|
|
471
472
|
shuffle (bool, optional): Whether to perform shuffle on the dataset. Default: None, expected
|
|
472
473
|
order behavior shown in the table below.
|
|
473
474
|
sampler (Sampler, optional): Object used to choose samples from the dataset.
|
|
@@ -478,7 +479,7 @@ class LJSpeechDataset(MappableDataset, AudioBaseDataset):
|
|
|
478
479
|
shard_id (int, optional): The shard ID within `num_shards` . Default: None. This
|
|
479
480
|
argument can only be specified when `num_shards` is also specified.
|
|
480
481
|
cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
|
|
481
|
-
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/experts/en/r2.0
|
|
482
|
+
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/experts/en/r2.0/dataset/cache.html>`_ .
|
|
482
483
|
Default: None, which means no cache is used.
|
|
483
484
|
|
|
484
485
|
Raises:
|
|
@@ -488,7 +489,7 @@ class LJSpeechDataset(MappableDataset, AudioBaseDataset):
|
|
|
488
489
|
RuntimeError: If `num_shards` is specified but `shard_id` is None.
|
|
489
490
|
RuntimeError: If `shard_id` is specified but `num_shards` is None.
|
|
490
491
|
ValueError: If `num_parallel_workers` exceeds the max thread numbers.
|
|
491
|
-
ValueError: If `shard_id` is
|
|
492
|
+
ValueError: If `shard_id` is not in range of [0, `num_shards` ).
|
|
492
493
|
|
|
493
494
|
Note:
|
|
494
495
|
- This dataset can take in a `sampler` . `sampler` and `shuffle` are mutually exclusive.
|
|
@@ -591,7 +592,7 @@ class LJSpeechDataset(MappableDataset, AudioBaseDataset):
|
|
|
591
592
|
|
|
592
593
|
class SpeechCommandsDataset(MappableDataset, AudioBaseDataset):
|
|
593
594
|
"""
|
|
594
|
-
|
|
595
|
+
Speech Commands dataset.
|
|
595
596
|
|
|
596
597
|
The generated dataset has five columns :py:obj:`[waveform, sample_rate, label, speaker_id, utterance_number]` .
|
|
597
598
|
The tensor of column :py:obj:`waveform` is a vector of the float32 type.
|
|
@@ -607,8 +608,9 @@ class SpeechCommandsDataset(MappableDataset, AudioBaseDataset):
|
|
|
607
608
|
test samples and 'all' will read from all 105,829 samples. Default: None, will read all samples.
|
|
608
609
|
num_samples (int, optional): The number of samples to be included in the dataset.
|
|
609
610
|
Default: None, will read all samples.
|
|
610
|
-
num_parallel_workers (int, optional): Number of
|
|
611
|
-
Default: None, will use
|
|
611
|
+
num_parallel_workers (int, optional): Number of worker threads to read the data.
|
|
612
|
+
Default: None, will use global default workers(8), it can be set
|
|
613
|
+
by `mindspore.dataset.config.set_num_parallel_workers` .
|
|
612
614
|
shuffle (bool, optional): Whether or not to perform shuffle on the dataset.
|
|
613
615
|
Default: None, expected order behavior shown in the table below.
|
|
614
616
|
sampler (Sampler, optional): Object used to choose samples from the dataset.
|
|
@@ -618,7 +620,7 @@ class SpeechCommandsDataset(MappableDataset, AudioBaseDataset):
|
|
|
618
620
|
shard_id (int, optional): The shard ID within `num_shards` . Default: None. This argument can only be specified
|
|
619
621
|
when `num_shards` is also specified.
|
|
620
622
|
cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
|
|
621
|
-
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/experts/en/r2.0
|
|
623
|
+
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/experts/en/r2.0/dataset/cache.html>`_ .
|
|
622
624
|
Default: None, which means no cache is used.
|
|
623
625
|
|
|
624
626
|
Raises:
|
|
@@ -628,7 +630,7 @@ class SpeechCommandsDataset(MappableDataset, AudioBaseDataset):
|
|
|
628
630
|
RuntimeError: If `num_shards` is specified but `shard_id` is None.
|
|
629
631
|
RuntimeError: If `shard_id` is specified but `num_shards` is None.
|
|
630
632
|
ValueError: If `num_parallel_workers` exceeds the max thread numbers.
|
|
631
|
-
ValueError: If `shard_id` is
|
|
633
|
+
ValueError: If `shard_id` is not in range of [0, `num_shards` ).
|
|
632
634
|
|
|
633
635
|
Note:
|
|
634
636
|
- This dataset can take in a `sampler` . `sampler` and `shuffle` are mutually exclusive.
|
|
@@ -714,7 +716,7 @@ class SpeechCommandsDataset(MappableDataset, AudioBaseDataset):
|
|
|
714
716
|
|
|
715
717
|
class TedliumDataset(MappableDataset, AudioBaseDataset):
|
|
716
718
|
"""
|
|
717
|
-
|
|
719
|
+
Tedlium dataset.
|
|
718
720
|
The columns of generated dataset depend on the source SPH files and the corresponding STM files.
|
|
719
721
|
|
|
720
722
|
The generated dataset has six columns :py:obj:`[waveform, sample_rate, transcript, talk_id, speaker_id,
|
|
@@ -737,8 +739,9 @@ class TedliumDataset(MappableDataset, AudioBaseDataset):
|
|
|
737
739
|
Default: None, ".sph".
|
|
738
740
|
num_samples (int, optional): The number of audio samples to be included in the dataset.
|
|
739
741
|
Default: None, all samples.
|
|
740
|
-
num_parallel_workers (int, optional): Number of
|
|
741
|
-
Default: None,
|
|
742
|
+
num_parallel_workers (int, optional): Number of worker threads to read the data.
|
|
743
|
+
Default: None, will use global default workers(8), it can be set
|
|
744
|
+
by `mindspore.dataset.config.set_num_parallel_workers` .
|
|
742
745
|
shuffle (bool, optional): Whether to perform shuffle on the dataset. Default: None, expected
|
|
743
746
|
order behavior shown in the table below.
|
|
744
747
|
sampler (Sampler, optional): Object used to choose samples from the
|
|
@@ -749,7 +752,7 @@ class TedliumDataset(MappableDataset, AudioBaseDataset):
|
|
|
749
752
|
shard_id (int, optional): The shard ID within `num_shards` . Default: None. This
|
|
750
753
|
argument can only be specified when `num_shards` is also specified.
|
|
751
754
|
cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
|
|
752
|
-
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/experts/en/r2.0
|
|
755
|
+
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/experts/en/r2.0/dataset/cache.html>`_ .
|
|
753
756
|
Default: None, which means no cache is used.
|
|
754
757
|
|
|
755
758
|
Raises:
|
|
@@ -759,7 +762,7 @@ class TedliumDataset(MappableDataset, AudioBaseDataset):
|
|
|
759
762
|
RuntimeError: If `num_shards` is specified but `shard_id` is None.
|
|
760
763
|
RuntimeError: If `shard_id` is specified but `num_shards` is None.
|
|
761
764
|
ValueError: If `num_parallel_workers` exceeds the max thread numbers.
|
|
762
|
-
ValueError: If `shard_id` is
|
|
765
|
+
ValueError: If `shard_id` is not in range of [0, `num_shards` ).
|
|
763
766
|
|
|
764
767
|
Note:
|
|
765
768
|
- This dataset can take in a `sampler` . `sampler` and `shuffle` are mutually exclusive.
|
|
@@ -925,7 +928,7 @@ class TedliumDataset(MappableDataset, AudioBaseDataset):
|
|
|
925
928
|
|
|
926
929
|
class YesNoDataset(MappableDataset, AudioBaseDataset):
|
|
927
930
|
"""
|
|
928
|
-
|
|
931
|
+
YesNo dataset.
|
|
929
932
|
|
|
930
933
|
The generated dataset has three columns :py:obj:`[waveform, sample_rate, labels]` .
|
|
931
934
|
The tensor of column :py:obj:`waveform` is a vector of the float32 type.
|
|
@@ -936,8 +939,9 @@ class YesNoDataset(MappableDataset, AudioBaseDataset):
|
|
|
936
939
|
dataset_dir (str): Path to the root directory that contains the dataset.
|
|
937
940
|
num_samples (int, optional): The number of images to be included in the dataset.
|
|
938
941
|
Default: None, will read all images.
|
|
939
|
-
num_parallel_workers (int, optional): Number of
|
|
940
|
-
Default: None, will use
|
|
942
|
+
num_parallel_workers (int, optional): Number of worker threads to read the data.
|
|
943
|
+
Default: None, will use global default workers(8), it can be set
|
|
944
|
+
by `mindspore.dataset.config.set_num_parallel_workers` .
|
|
941
945
|
shuffle (bool, optional): Whether or not to perform shuffle on the dataset.
|
|
942
946
|
Default: None, expected order behavior shown in the table below.
|
|
943
947
|
sampler (Sampler, optional): Object used to choose samples from the
|
|
@@ -947,7 +951,7 @@ class YesNoDataset(MappableDataset, AudioBaseDataset):
|
|
|
947
951
|
shard_id (int, optional): The shard ID within `num_shards` . Default: None. This argument can only
|
|
948
952
|
be specified when `num_shards` is also specified.
|
|
949
953
|
cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
|
|
950
|
-
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/experts/en/r2.0
|
|
954
|
+
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/experts/en/r2.0/dataset/cache.html>`_ .
|
|
951
955
|
Default: None, which means no cache is used.
|
|
952
956
|
|
|
953
957
|
Raises:
|
|
@@ -957,7 +961,7 @@ class YesNoDataset(MappableDataset, AudioBaseDataset):
|
|
|
957
961
|
RuntimeError: If `sampler` and `num_shards`/`shard_id` are specified at the same time.
|
|
958
962
|
RuntimeError: If `num_shards` is specified but `shard_id` is None.
|
|
959
963
|
RuntimeError: If `shard_id` is specified but `num_shards` is None.
|
|
960
|
-
ValueError: If `shard_id` is
|
|
964
|
+
ValueError: If `shard_id` is not in range of [0, `num_shards` ).
|
|
961
965
|
|
|
962
966
|
Note:
|
|
963
967
|
- This dataset can take in a `sampler` . `sampler` and `shuffle` are mutually exclusive.
|
|
@@ -1,4 +1,4 @@
|
|
|
1
|
-
# Copyright 2019-
|
|
1
|
+
# Copyright 2019-2023 Huawei Technologies Co., Ltd
|
|
2
2
|
#
|
|
3
3
|
# Licensed under the Apache License, Version 2.0 (the "License");
|
|
4
4
|
# you may not use this file except in compliance with the License.
|
|
@@ -26,6 +26,7 @@ import platform
|
|
|
26
26
|
import numpy as np
|
|
27
27
|
|
|
28
28
|
import mindspore._c_dataengine as cde
|
|
29
|
+
from mindspore import log as logger
|
|
29
30
|
|
|
30
31
|
from .datasets import UnionBaseDataset, SourceDataset, MappableDataset, Shuffle, Schema, \
|
|
31
32
|
shuffle_to_shuffle_mode, shuffle_to_bool
|
|
@@ -56,8 +57,9 @@ class CSVDataset(SourceDataset, UnionBaseDataset):
|
|
|
56
57
|
is not provided, infers the column_names from the first row of CSV file.
|
|
57
58
|
num_samples (int, optional): The number of samples to be included in the dataset.
|
|
58
59
|
Default: None, will include all images.
|
|
59
|
-
num_parallel_workers (int, optional): Number of
|
|
60
|
-
Default: None,
|
|
60
|
+
num_parallel_workers (int, optional): Number of worker threads to read the data.
|
|
61
|
+
Default: None, will use global default workers(8), it can be set
|
|
62
|
+
by `mindspore.dataset.config.set_num_parallel_workers` .
|
|
61
63
|
shuffle (Union[bool, Shuffle], optional): Perform reshuffling of the data every epoch.
|
|
62
64
|
Default: Shuffle.GLOBAL. Bool type and Shuffle enum are both supported to pass in.
|
|
63
65
|
If shuffle is False, no shuffling will be performed.
|
|
@@ -73,7 +75,7 @@ class CSVDataset(SourceDataset, UnionBaseDataset):
|
|
|
73
75
|
shard_id (int, optional): The shard ID within `num_shards` . Default: None. This
|
|
74
76
|
argument can only be specified when `num_shards` is also specified.
|
|
75
77
|
cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
|
|
76
|
-
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/experts/en/r2.0
|
|
78
|
+
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/experts/en/r2.0/dataset/cache.html>`_ .
|
|
77
79
|
Default: None, which means no cache is used.
|
|
78
80
|
|
|
79
81
|
Raises:
|
|
@@ -82,7 +84,7 @@ class CSVDataset(SourceDataset, UnionBaseDataset):
|
|
|
82
84
|
ValueError: If `num_parallel_workers` exceeds the max thread numbers.
|
|
83
85
|
RuntimeError: If `num_shards` is specified but `shard_id` is None.
|
|
84
86
|
RuntimeError: If `shard_id` is specified but `num_shards` is None.
|
|
85
|
-
ValueError: If `shard_id` is
|
|
87
|
+
ValueError: If `shard_id` is not in range of [0, `num_shards` ).
|
|
86
88
|
|
|
87
89
|
Examples:
|
|
88
90
|
>>> csv_dataset_dir = ["/path/to/csv_dataset_file"] # contains 1 or multiple csv files
|
|
@@ -116,8 +118,10 @@ class MindDataset(MappableDataset, UnionBaseDataset):
|
|
|
116
118
|
a file name of one component of a mindrecord source, other files with identical source
|
|
117
119
|
in the same path will be found and loaded automatically. If dataset_file is a list,
|
|
118
120
|
it represents for a list of dataset files to be read directly.
|
|
119
|
-
columns_list (list[str], optional): List of columns to be read. Default: None.
|
|
120
|
-
num_parallel_workers (int, optional):
|
|
121
|
+
columns_list (list[str], optional): List of columns to be read. Default: None, read all columns.
|
|
122
|
+
num_parallel_workers (int, optional): Number of worker threads to read the data.
|
|
123
|
+
Default: None, will use global default workers(8), it can be set
|
|
124
|
+
by `mindspore.dataset.config.set_num_parallel_workers` .
|
|
121
125
|
shuffle (Union[bool, Shuffle], optional): Perform reshuffling of the data every epoch.
|
|
122
126
|
Default: None, performs global shuffle. Bool type and Shuffle enum are both supported to pass in.
|
|
123
127
|
If shuffle is False, no shuffling will be performed.
|
|
@@ -131,7 +135,7 @@ class MindDataset(MappableDataset, UnionBaseDataset):
|
|
|
131
135
|
- Shuffle.INFILE: Keep the file sequence the same but shuffle the data within each file.
|
|
132
136
|
|
|
133
137
|
num_shards (int, optional): Number of shards that the dataset will be divided into. Default: None.
|
|
134
|
-
When this argument is specified,
|
|
138
|
+
When this argument is specified, `num_samples` reflects the maximum sample number of per shard.
|
|
135
139
|
shard_id (int, optional): The shard ID within `num_shards` . Default: None. This
|
|
136
140
|
argument can only be specified when `num_shards` is also specified.
|
|
137
141
|
sampler (Sampler, optional): Object used to choose samples from the
|
|
@@ -145,7 +149,7 @@ class MindDataset(MappableDataset, UnionBaseDataset):
|
|
|
145
149
|
num_samples (int, optional): The number of samples to be included in the dataset.
|
|
146
150
|
Default: None, all samples.
|
|
147
151
|
cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
|
|
148
|
-
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/experts/en/r2.0
|
|
152
|
+
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/experts/en/r2.0/dataset/cache.html>`_ .
|
|
149
153
|
Default: None, which means no cache is used.
|
|
150
154
|
|
|
151
155
|
Raises:
|
|
@@ -153,7 +157,7 @@ class MindDataset(MappableDataset, UnionBaseDataset):
|
|
|
153
157
|
ValueError: If `num_parallel_workers` exceeds the max thread numbers.
|
|
154
158
|
RuntimeError: If `num_shards` is specified but `shard_id` is None.
|
|
155
159
|
RuntimeError: If `shard_id` is specified but `num_shards` is None.
|
|
156
|
-
ValueError: If `shard_id` is
|
|
160
|
+
ValueError: If `shard_id` is not in range of [0, `num_shards` ).
|
|
157
161
|
|
|
158
162
|
Note:
|
|
159
163
|
- This dataset can take in a `sampler` . `sampler` and `shuffle` are mutually exclusive.
|
|
@@ -247,56 +251,60 @@ class TFRecordDataset(SourceDataset, UnionBaseDataset):
|
|
|
247
251
|
|
|
248
252
|
Args:
|
|
249
253
|
dataset_files (Union[str, list[str]]): String or list of files to be read or glob strings to search for a
|
|
250
|
-
pattern of files. The list will be sorted in
|
|
254
|
+
pattern of files. The list will be sorted in lexicographical order.
|
|
251
255
|
schema (Union[str, Schema], optional): Data format policy, which specifies the data types and shapes of the data
|
|
252
256
|
column to be read. Both JSON file path and objects constructed by mindspore.dataset.Schema are acceptable.
|
|
253
257
|
Default: None.
|
|
254
258
|
columns_list (list[str], optional): List of columns to be read. Default: None, read all columns.
|
|
255
|
-
num_samples (int, optional): The number of samples (rows) to be included in the dataset.
|
|
256
|
-
|
|
257
|
-
|
|
258
|
-
If
|
|
259
|
-
|
|
260
|
-
|
|
261
|
-
|
|
262
|
-
|
|
259
|
+
num_samples (int, optional): The number of samples (rows) to be included in the dataset.
|
|
260
|
+
Default: None.
|
|
261
|
+
Processing priority for `num_samples` is as the following:
|
|
262
|
+
1. If `num_samples` is greater than 0, read `num_samples` rows.
|
|
263
|
+
2. Otherwise, if numRows (parsed from `schema` ) is greater than 0, read numRows rows.
|
|
264
|
+
3. Otherwise, read the full dataset.
|
|
265
|
+
`num_samples` or numRows (parsed from `schema` ) will be interpreted as number of rows per shard.
|
|
266
|
+
It is highly recommended to provide `num_samples` or numRows (parsed from `schema` )
|
|
267
|
+
when `compression_type` is "GZIP" or "ZLIB" to avoid performance degradation due to multiple
|
|
268
|
+
decompressions of the same file to obtain the file size.
|
|
269
|
+
num_parallel_workers (int, optional): Number of worker threads to read the data.
|
|
270
|
+
Default: None, will use global default workers(8), it can be set
|
|
271
|
+
by `mindspore.dataset.config.set_num_parallel_workers` .
|
|
263
272
|
shuffle (Union[bool, Shuffle], optional): Perform reshuffling of the data every epoch.
|
|
264
273
|
Default: Shuffle.GLOBAL. Bool type and Shuffle enum are both supported to pass in.
|
|
265
|
-
If shuffle is False, no shuffling will be performed.
|
|
266
|
-
If shuffle is True,
|
|
274
|
+
If `shuffle` is False, no shuffling will be performed.
|
|
275
|
+
If `shuffle` is True, perform global shuffle.
|
|
267
276
|
There are three levels of shuffling, desired shuffle enum defined by mindspore.dataset.Shuffle.
|
|
268
277
|
|
|
269
|
-
- Shuffle.GLOBAL: Shuffle both the files and samples, same as setting shuffle to True.
|
|
278
|
+
- Shuffle.GLOBAL: Shuffle both the files and samples, same as setting `shuffle` to True.
|
|
270
279
|
|
|
271
280
|
- Shuffle.FILES: Shuffle files only.
|
|
272
281
|
|
|
273
282
|
num_shards (int, optional): Number of shards that the dataset will be divided
|
|
274
283
|
into. Default: None. When this argument is specified, `num_samples` reflects
|
|
275
|
-
the maximum sample number
|
|
284
|
+
the maximum sample number per shard.
|
|
276
285
|
shard_id (int, optional): The shard ID within `num_shards` . Default: None. This
|
|
277
286
|
argument can only be specified when `num_shards` is also specified.
|
|
278
287
|
shard_equal_rows (bool, optional): Get equal rows for all shards. Default: False. If `shard_equal_rows`
|
|
279
|
-
is False, number of rows of each shard may be
|
|
280
|
-
When the number of samples
|
|
288
|
+
is False, the number of rows of each shard may not be equal, and may lead to a failure in distributed
|
|
289
|
+
training. When the number of samples per TFRecord file are not equal, it is suggested to set it to True.
|
|
281
290
|
This argument should only be specified when `num_shards` is also specified.
|
|
282
|
-
When `compression_type` is
|
|
291
|
+
When `compression_type` is not None, and `num_samples` or numRows (parsed from `schema` ) is provided,
|
|
292
|
+
`shard_equal_rows` will be implied as true.
|
|
283
293
|
cache (DatasetCache, optional): Use tensor caching service to speed up dataset processing. More details:
|
|
284
|
-
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/experts/en/r2.0
|
|
294
|
+
`Single-Node Data Cache <https://www.mindspore.cn/tutorials/experts/en/r2.0/dataset/cache.html>`_ .
|
|
285
295
|
Default: None, which means no cache is used.
|
|
286
|
-
compression_type (str, optional):
|
|
287
|
-
Default: None, as in empty string.
|
|
288
|
-
This will automatically get equal rows for all shards (`shard_equal_rows` considered to be True) and thus
|
|
289
|
-
cannot have the case where `num_samples` is None.
|
|
296
|
+
compression_type (str, optional): The type of compression used for all files, must be either '', 'GZIP', or
|
|
297
|
+
'ZLIB'. Default: None, as in empty string.
|
|
290
298
|
|
|
291
299
|
Raises:
|
|
292
300
|
ValueError: If dataset_files are not valid or do not exist.
|
|
293
301
|
ValueError: If `num_parallel_workers` exceeds the max thread numbers.
|
|
294
302
|
RuntimeError: If `num_shards` is specified but `shard_id` is None.
|
|
295
303
|
RuntimeError: If `shard_id` is specified but `num_shards` is None.
|
|
296
|
-
ValueError: If `shard_id` is
|
|
304
|
+
ValueError: If `shard_id` is not in range of [0, `num_shards` ).
|
|
297
305
|
ValueError: If `compression_type` is invalid (other than '', 'GZIP', or 'ZLIB').
|
|
298
|
-
ValueError: If `compression_type` is provided but number of dataset files < `num_shards` .
|
|
299
|
-
ValueError: If `
|
|
306
|
+
ValueError: If `compression_type` is provided, but the number of dataset files < `num_shards` .
|
|
307
|
+
ValueError: If `num_samples` < 0.
|
|
300
308
|
|
|
301
309
|
Examples:
|
|
302
310
|
>>> from mindspore import dtype as mstype
|
|
@@ -313,7 +321,7 @@ class TFRecordDataset(SourceDataset, UnionBaseDataset):
|
|
|
313
321
|
>>> schema.add_column(name='col_1d', de_type=mstype.int64, shape=[2])
|
|
314
322
|
>>> dataset = ds.TFRecordDataset(dataset_files=tfrecord_dataset_dir, schema=schema)
|
|
315
323
|
>>>
|
|
316
|
-
>>> # 3) Get all rows from tfrecord_dataset_dir with schema file.
|
|
324
|
+
>>> # 3) Get all rows from tfrecord_dataset_dir with the schema file.
|
|
317
325
|
>>> dataset = ds.TFRecordDataset(dataset_files=tfrecord_dataset_dir, schema=tfrecord_schema_file)
|
|
318
326
|
"""
|
|
319
327
|
|
|
@@ -331,9 +339,14 @@ class TFRecordDataset(SourceDataset, UnionBaseDataset):
|
|
|
331
339
|
self.shard_equal_rows = replace_none(shard_equal_rows, False)
|
|
332
340
|
self.compression_type = replace_none(compression_type, "")
|
|
333
341
|
|
|
342
|
+
# Only take numRows from schema when num_samples is not provided
|
|
334
343
|
if self.schema is not None and (self.num_samples is None or self.num_samples == 0):
|
|
335
344
|
self.num_samples = Schema.get_num_rows(self.schema)
|
|
336
345
|
|
|
346
|
+
if self.compression_type in ['ZLIB', 'GZIP'] and (self.num_samples is None or self.num_samples == 0):
|
|
347
|
+
logger.warning("Since compression_type is set, but neither num_samples nor numRows (from schema file) " +
|
|
348
|
+
"is provided, performance might be degraded.")
|
|
349
|
+
|
|
337
350
|
def parse(self, children=None):
|
|
338
351
|
schema = self.schema.cpp_schema if isinstance(self.schema, Schema) else self.schema
|
|
339
352
|
return cde.TFRecordNode(self.dataset_files, schema, self.columns_list, self.num_samples, self.shuffle_flag,
|
|
@@ -375,18 +388,18 @@ class OBSMindDataset(GeneratorDataset):
|
|
|
375
388
|
num_shards (int, optional): Number of shards that the dataset will be divided
|
|
376
389
|
into. Default: None.
|
|
377
390
|
shard_id (int, optional): The shard ID within num_shards. Default: None. This
|
|
378
|
-
argument can only be specified when num_shards is also specified.
|
|
391
|
+
argument can only be specified when `num_shards` is also specified.
|
|
379
392
|
shard_equal_rows (bool, optional): Get equal rows for all shards. Default: True. If shard_equal_rows
|
|
380
393
|
is false, number of rows of each shard may be not equal, and may lead to a failure in distributed training.
|
|
381
394
|
When the number of samples of per MindRecord file are not equal, it is suggested to set to true.
|
|
382
|
-
This argument should only be specified when num_shards is also specified.
|
|
395
|
+
This argument should only be specified when `num_shards` is also specified.
|
|
383
396
|
|
|
384
397
|
Raises:
|
|
385
398
|
RuntimeError: If `sync_obs_path` do not exist.
|
|
386
399
|
ValueError: If `columns_list` is invalid.
|
|
387
400
|
RuntimeError: If `num_shards` is specified but `shard_id` is None.
|
|
388
401
|
RuntimeError: If `shard_id` is specified but `num_shards` is None.
|
|
389
|
-
ValueError: If `shard_id` is
|
|
402
|
+
ValueError: If `shard_id` is not in range of [0, `num_shards` ).
|
|
390
403
|
|
|
391
404
|
Note:
|
|
392
405
|
- It's necessary to create a synchronization directory on cloud storage in
|