mindspore 2.2.0__cp39-none-any.whl → 2.2.10__cp39-none-any.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Potentially problematic release.
This version of mindspore might be problematic. Click here for more details.
- mindspore/.commit_id +1 -1
- mindspore/_akg/akg/composite/build_module.py +9 -15
- mindspore/_akg/akg/utils/ascend_profilier/__init__.py +0 -0
- mindspore/_akg/akg/utils/ascend_profilier/cann_file_parser.py +76 -0
- mindspore/_akg/akg/utils/ascend_profilier/file_manager.py +56 -0
- mindspore/_akg/akg/utils/ascend_profilier/op_summary_bean.py +23 -0
- mindspore/_akg/akg/utils/ascend_profilier/op_summary_headers.py +8 -0
- mindspore/_akg/akg/utils/ascend_profilier/op_summary_parser.py +42 -0
- mindspore/_akg/akg/utils/ascend_profilier/path_manager.py +65 -0
- mindspore/_akg/akg/utils/kernel_exec.py +41 -15
- mindspore/_akg/akg/utils/tbe_codegen_utils.py +27 -6
- mindspore/_akg/akg/utils/util.py +38 -0
- mindspore/_c_dataengine.cpython-39-aarch64-linux-gnu.so +0 -0
- mindspore/_c_expression.cpython-39-aarch64-linux-gnu.so +0 -0
- mindspore/_checkparam.py +3 -3
- mindspore/_extends/graph_kernel/model/graph_split.py +84 -76
- mindspore/_extends/graph_kernel/splitter.py +3 -2
- mindspore/_extends/parallel_compile/akg_compiler/build_tbe_kernel.py +83 -66
- mindspore/_extends/parallel_compile/akg_compiler/tbe_topi.py +4 -4
- mindspore/_extends/parallel_compile/akg_compiler/util.py +10 -7
- mindspore/_extends/parallel_compile/tbe_compiler/tbe_helper.py +2 -1
- mindspore/_extends/parse/standard_method.py +2 -9
- mindspore/_extends/remote/kernel_build_server.py +2 -1
- mindspore/_mindspore_offline_debug.cpython-39-aarch64-linux-gnu.so +0 -0
- mindspore/bin/cache_admin +0 -0
- mindspore/bin/cache_server +0 -0
- mindspore/common/api.py +1 -1
- mindspore/common/auto_dynamic_shape.py +81 -85
- mindspore/common/dump.py +1 -1
- mindspore/common/tensor.py +3 -20
- mindspore/config/op_info.config +1 -1
- mindspore/context.py +11 -4
- mindspore/dataset/engine/datasets_standard_format.py +5 -0
- mindspore/dataset/vision/transforms.py +21 -21
- mindspore/experimental/optim/adam.py +1 -1
- mindspore/gen_ops.py +1 -1
- mindspore/include/api/model.h +17 -0
- mindspore/include/api/status.h +8 -3
- mindspore/lib/libdnnl.so.2 +0 -0
- mindspore/lib/libmindspore.so +0 -0
- mindspore/lib/libmindspore_backend.so +0 -0
- mindspore/lib/libmindspore_common.so +0 -0
- mindspore/lib/libmindspore_core.so +0 -0
- mindspore/lib/libmindspore_glog.so.0 +0 -0
- mindspore/lib/libmindspore_gpr.so.15 +0 -0
- mindspore/lib/libmindspore_grpc++.so.1 +0 -0
- mindspore/lib/libmindspore_grpc.so.15 +0 -0
- mindspore/lib/libmindspore_shared_lib.so +0 -0
- mindspore/lib/libnnacl.so +0 -0
- mindspore/lib/libopencv_core.so.4.5 +0 -0
- mindspore/lib/libopencv_imgcodecs.so.4.5 +0 -0
- mindspore/lib/libopencv_imgproc.so.4.5 +0 -0
- mindspore/lib/plugin/ascend/custom_aicore_ops/op_impl/ai_core/tbe/config/ascend310/aic-ascend310-ops-info.json +123 -0
- mindspore/lib/plugin/ascend/custom_aicore_ops/op_impl/ai_core/tbe/config/ascend310p/aic-ascend310p-ops-info.json +123 -0
- mindspore/lib/plugin/ascend/custom_aicore_ops/op_impl/ai_core/tbe/config/ascend910/aic-ascend910-ops-info.json +158 -0
- mindspore/lib/plugin/ascend/custom_aicore_ops/op_impl/ai_core/tbe/config/ascend910b/aic-ascend910b-ops-info.json +37 -0
- mindspore/lib/plugin/ascend/custom_aicore_ops/op_impl/ai_core/tbe/custom_aicore_ops_impl/add_dsl.py +46 -0
- mindspore/lib/plugin/ascend/custom_aicore_ops/op_impl/ai_core/tbe/custom_aicore_ops_impl/add_tik.py +51 -0
- mindspore/lib/plugin/ascend/custom_aicore_ops/op_impl/ai_core/tbe/custom_aicore_ops_impl/kv_cache_mgr.py +241 -0
- mindspore/lib/plugin/ascend/custom_aicore_ops/op_impl/ai_core/tbe/custom_aicore_ops_impl/matmul_tik.py +212 -0
- mindspore/lib/plugin/ascend/custom_aicore_ops/op_impl/vector_core/tbe/custom_aicore_ops_impl/add_dsl.py +46 -0
- mindspore/lib/plugin/ascend/custom_aicore_ops/op_impl/vector_core/tbe/custom_aicore_ops_impl/add_tik.py +51 -0
- mindspore/lib/plugin/ascend/custom_aicore_ops/op_impl/vector_core/tbe/custom_aicore_ops_impl/kv_cache_mgr.py +241 -0
- mindspore/lib/plugin/ascend/custom_aicore_ops/op_impl/vector_core/tbe/custom_aicore_ops_impl/matmul_tik.py +212 -0
- mindspore/lib/plugin/ascend/custom_aicore_ops/op_proto/libop_proto.so +0 -0
- mindspore/lib/plugin/ascend/custom_aicpu_ops/op_impl/cpu/aicpu_kernel/impl/libcust_aicpu_kernels.so +0 -0
- mindspore/lib/plugin/ascend/custom_aicpu_ops/op_impl/cpu/aicpu_kernel/impl/libcust_cpu_kernels.so +0 -0
- mindspore/lib/plugin/ascend/custom_aicpu_ops/op_impl/cpu/config/cust_aicpu_kernel.json +8 -80
- mindspore/lib/plugin/ascend/custom_aicpu_ops/op_proto/libcust_op_proto.so +0 -0
- mindspore/lib/plugin/ascend/libakg.so +0 -0
- mindspore/lib/plugin/ascend/libhccl_plugin.so +0 -0
- mindspore/lib/plugin/ascend/libmindspore_aicpu_kernels.so +0 -0
- mindspore/lib/plugin/ascend/libmindspore_cpu_kernels.so +0 -0
- mindspore/lib/plugin/cpu/libakg.so +0 -0
- mindspore/lib/plugin/libmindspore_ascend.so.1 +0 -0
- mindspore/lib/plugin/libmindspore_ascend.so.2 +0 -0
- mindspore/nn/cell.py +0 -3
- mindspore/nn/layer/activation.py +4 -5
- mindspore/nn/layer/conv.py +39 -23
- mindspore/nn/layer/flash_attention.py +90 -78
- mindspore/nn/layer/math.py +3 -7
- mindspore/nn/layer/rnn_cells.py +5 -5
- mindspore/nn/wrap/cell_wrapper.py +6 -0
- mindspore/numpy/utils_const.py +5 -5
- mindspore/ops/_grad_experimental/grad_array_ops.py +1 -1
- mindspore/ops/_grad_experimental/grad_implementations.py +2 -2
- mindspore/ops/_grad_experimental/grad_math_ops.py +19 -18
- mindspore/ops/_grad_experimental/grad_sparse_ops.py +3 -3
- mindspore/ops/_op_impl/aicpu/add.py +3 -3
- mindspore/ops/_utils/utils.py +2 -0
- mindspore/ops/composite/multitype_ops/_compile_utils.py +2 -1
- mindspore/ops/composite/multitype_ops/getitem_impl.py +2 -2
- mindspore/ops/function/array_func.py +10 -7
- mindspore/ops/function/grad/grad_func.py +0 -1
- mindspore/ops/function/nn_func.py +98 -9
- mindspore/ops/function/random_func.py +2 -1
- mindspore/ops/op_info_register.py +24 -21
- mindspore/ops/operations/__init__.py +3 -2
- mindspore/ops/operations/_grad_ops.py +24 -4
- mindspore/ops/operations/_inner_ops.py +155 -23
- mindspore/ops/operations/array_ops.py +9 -7
- mindspore/ops/operations/comm_ops.py +2 -2
- mindspore/ops/operations/custom_ops.py +85 -68
- mindspore/ops/operations/inner_ops.py +26 -3
- mindspore/ops/operations/math_ops.py +4 -3
- mindspore/ops/operations/nn_ops.py +109 -28
- mindspore/parallel/_parallel_serialization.py +10 -3
- mindspore/parallel/_tensor.py +4 -1
- mindspore/parallel/checkpoint_transform.py +13 -2
- mindspore/parallel/shard.py +17 -10
- mindspore/profiler/common/util.py +1 -0
- mindspore/profiler/parser/ascend_hccl_generator.py +232 -0
- mindspore/profiler/parser/ascend_msprof_exporter.py +86 -43
- mindspore/profiler/parser/ascend_msprof_generator.py +196 -9
- mindspore/profiler/parser/ascend_op_generator.py +1 -1
- mindspore/profiler/parser/ascend_timeline_generator.py +6 -182
- mindspore/profiler/parser/base_timeline_generator.py +1 -1
- mindspore/profiler/parser/cpu_gpu_timeline_generator.py +2 -2
- mindspore/profiler/parser/framework_parser.py +1 -1
- mindspore/profiler/parser/profiler_info.py +19 -0
- mindspore/profiler/profiling.py +46 -24
- mindspore/rewrite/api/pattern_engine.py +1 -1
- mindspore/rewrite/parsers/for_parser.py +1 -1
- mindspore/rewrite/symbol_tree.py +1 -4
- mindspore/run_check/_check_version.py +5 -3
- mindspore/safeguard/rewrite_obfuscation.py +52 -28
- mindspore/train/callback/_summary_collector.py +1 -1
- mindspore/train/dataset_helper.py +1 -0
- mindspore/train/model.py +2 -2
- mindspore/train/serialization.py +97 -11
- mindspore/train/summary/_summary_adapter.py +1 -1
- mindspore/train/summary/summary_record.py +23 -7
- mindspore/version.py +1 -1
- {mindspore-2.2.0.dist-info → mindspore-2.2.10.dist-info}/METADATA +1 -1
- {mindspore-2.2.0.dist-info → mindspore-2.2.10.dist-info}/RECORD +138 -118
- {mindspore-2.2.0.dist-info → mindspore-2.2.10.dist-info}/WHEEL +0 -0
- {mindspore-2.2.0.dist-info → mindspore-2.2.10.dist-info}/entry_points.txt +0 -0
- {mindspore-2.2.0.dist-info → mindspore-2.2.10.dist-info}/top_level.txt +0 -0
|
@@ -27,6 +27,7 @@ SHAPE_RANK_ANY = -2
|
|
|
27
27
|
|
|
28
28
|
auto_dynamic_shepe_dict = {}
|
|
29
29
|
|
|
30
|
+
|
|
30
31
|
class _AutoDynamicShapeManager:
|
|
31
32
|
"""
|
|
32
33
|
Represents a function to manage auto identify dynamic shape.
|
|
@@ -101,14 +102,14 @@ class _AutoDynamicShapeManager:
|
|
|
101
102
|
|
|
102
103
|
def get_compile_args_shape_without_sink(self, input_args, res_shape):
|
|
103
104
|
"""get compile args shape with out sink mode"""
|
|
104
|
-
for
|
|
105
|
-
if isinstance(
|
|
106
|
-
res_shape.append(
|
|
107
|
-
elif isinstance(
|
|
105
|
+
for arg in input_args:
|
|
106
|
+
if isinstance(arg, Tensor):
|
|
107
|
+
res_shape.append(arg.shape)
|
|
108
|
+
elif isinstance(arg, (int, float)):
|
|
108
109
|
res_shape.append([])
|
|
109
|
-
elif isinstance(
|
|
110
|
+
elif isinstance(arg, (tuple, list)):
|
|
110
111
|
tmp_shape = []
|
|
111
|
-
self.get_compile_args_shape_without_sink(
|
|
112
|
+
self.get_compile_args_shape_without_sink(arg, tmp_shape)
|
|
112
113
|
res_shape.append(tmp_shape)
|
|
113
114
|
|
|
114
115
|
|
|
@@ -179,14 +180,6 @@ class _AutoDynamicShapeManager:
|
|
|
179
180
|
self.generalize_shape_cache.append(compile_args)
|
|
180
181
|
|
|
181
182
|
|
|
182
|
-
def check_tuple_of_scalar(self, input):
|
|
183
|
-
"""check tuple of scalar"""
|
|
184
|
-
for elem in input:
|
|
185
|
-
if not isinstance(elem, int):
|
|
186
|
-
return False
|
|
187
|
-
return True
|
|
188
|
-
|
|
189
|
-
|
|
190
183
|
def _compare_input_args_and_cache_args(self, input_args, cache_args):
|
|
191
184
|
"""compare input args and cache args"""
|
|
192
185
|
for (input, cache) in zip(input_args, cache_args):
|
|
@@ -233,23 +226,25 @@ class _AutoIdentifyDynamicShape:
|
|
|
233
226
|
|
|
234
227
|
def _check_input_tensor_type(self, args_list, cache_list):
|
|
235
228
|
"""check input args type"""
|
|
236
|
-
for (
|
|
237
|
-
if isinstance(
|
|
238
|
-
if
|
|
239
|
-
logger.debug((f'input tensor type = {
|
|
229
|
+
for (arg, cache) in zip(args_list, cache_list):
|
|
230
|
+
if isinstance(arg, Tensor) and isinstance(cache, Tensor):
|
|
231
|
+
if arg.dtype != cache.dtype:
|
|
232
|
+
logger.debug((f'input tensor type = {arg.dtype}, cache tensor type = {cache.dtype}, '
|
|
240
233
|
f'tensor types are not same.'))
|
|
241
234
|
return False
|
|
242
|
-
elif isinstance(
|
|
243
|
-
res = self._check_input_tensor_type(
|
|
235
|
+
elif isinstance(arg, (tuple, list)) and isinstance(cache, (tuple, list)):
|
|
236
|
+
res = self._check_input_tensor_type(arg, cache)
|
|
244
237
|
if not res:
|
|
245
238
|
return False
|
|
246
|
-
elif (isinstance(
|
|
247
|
-
(isinstance(
|
|
248
|
-
if
|
|
239
|
+
elif (isinstance(arg, int) and isinstance(cache, int)) or \
|
|
240
|
+
(isinstance(arg, float) and isinstance(cache, float)):
|
|
241
|
+
if arg != cache:
|
|
249
242
|
return False
|
|
250
|
-
elif
|
|
251
|
-
|
|
252
|
-
|
|
243
|
+
elif isinstance(arg, Tensor) and not isinstance(cache, Tensor):
|
|
244
|
+
return False
|
|
245
|
+
elif isinstance(arg, (int, float)) and not isinstance(cache, (int, float)):
|
|
246
|
+
return False
|
|
247
|
+
elif isinstance(arg, (tuple, list)) and not isinstance(cache, (tuple, list)):
|
|
253
248
|
return False
|
|
254
249
|
return True
|
|
255
250
|
|
|
@@ -292,21 +287,21 @@ class _AutoIdentifyDynamicShape:
|
|
|
292
287
|
|
|
293
288
|
|
|
294
289
|
@staticmethod
|
|
295
|
-
def _do_generalize_in_sink(
|
|
290
|
+
def _do_generalize_in_sink(arg, cache, input_index, cache_index, cache_type):
|
|
296
291
|
"""do generalize in sink, input rank must be 2"""
|
|
297
|
-
if not
|
|
292
|
+
if not arg:
|
|
298
293
|
raise ValueError("In sink mode, cell input can not be scalar.")
|
|
299
294
|
|
|
300
|
-
if
|
|
295
|
+
if arg == cache:
|
|
301
296
|
return cache
|
|
302
297
|
|
|
303
298
|
shape_value = []
|
|
304
|
-
if len(
|
|
299
|
+
if len(arg) != len(cache):
|
|
305
300
|
shape_value.append(SHAPE_RANK_ANY)
|
|
306
301
|
else:
|
|
307
|
-
for _ in
|
|
302
|
+
for _ in arg:
|
|
308
303
|
shape_value.append(SHAPE_DIM_ANY)
|
|
309
|
-
logger.info((f'In the {cache_type} cache[{cache_index}], the {input_index}th input tensor shape is {
|
|
304
|
+
logger.info((f'In the {cache_type} cache[{cache_index}], the {input_index}th input tensor shape is {arg},'
|
|
310
305
|
f'cache shape is {cache}, not equal, need generalize to {shape_value}.'))
|
|
311
306
|
return shape_value
|
|
312
307
|
|
|
@@ -318,6 +313,16 @@ class _AutoIdentifyDynamicShape:
|
|
|
318
313
|
is_sink_mode, aux)
|
|
319
314
|
|
|
320
315
|
|
|
316
|
+
def _generate_with_generalize_shape(self, generalize_shape_args, is_sink_mode, args_list):
|
|
317
|
+
"""generate with generalize_shape """
|
|
318
|
+
new_generalize_shape, can_generalize = self._do_generalize_shape("generalize", generalize_shape_args,
|
|
319
|
+
is_sink_mode)
|
|
320
|
+
if not can_generalize:
|
|
321
|
+
return args_list
|
|
322
|
+
|
|
323
|
+
res_shape = self.auto_dynamic_shape_manager.get_compile_args_shape(new_generalize_shape, is_sink_mode)
|
|
324
|
+
logger.info((f'generalize with generalize shape cache, compile args shape = {res_shape}'))
|
|
325
|
+
return new_generalize_shape
|
|
321
326
|
|
|
322
327
|
def auto_dynamic_generate_compile_args(self, args_list, is_sink_mode):
|
|
323
328
|
"""generate compile args in auto dynamic shape"""
|
|
@@ -331,15 +336,7 @@ class _AutoIdentifyDynamicShape:
|
|
|
331
336
|
logger.debug((f'input args list shape = {res_shape}.'))
|
|
332
337
|
|
|
333
338
|
# step1: find cache in real_shape_cache.
|
|
334
|
-
|
|
335
|
-
if real_cache_number < 2:
|
|
336
|
-
logger.info((f'real shape cache cap is {real_cache_number}, smaller than 2, '
|
|
337
|
-
f'compile args shape={res_shape}.'))
|
|
338
|
-
return args_list
|
|
339
|
-
|
|
340
|
-
is_real_shape_exist = self.auto_dynamic_shape_manager.find_compile_args_in_shape_cache(args_list, "real")
|
|
341
|
-
if is_real_shape_exist:
|
|
342
|
-
logger.debug((f'find compile args in real shape cache, compile args shape={res_shape}'))
|
|
339
|
+
if self._check_real_shape_cache(res_shape, args_list):
|
|
343
340
|
return args_list
|
|
344
341
|
|
|
345
342
|
# step2: if can not find cache in real_shape_cache, then generate it
|
|
@@ -358,14 +355,7 @@ class _AutoIdentifyDynamicShape:
|
|
|
358
355
|
|
|
359
356
|
# step 4: if can not find cache in generalize_shape_cache, then generate it again
|
|
360
357
|
if not is_generalize_shape_exist:
|
|
361
|
-
|
|
362
|
-
is_sink_mode)
|
|
363
|
-
if not can_generalize:
|
|
364
|
-
return args_list
|
|
365
|
-
|
|
366
|
-
res_shape = self.auto_dynamic_shape_manager.get_compile_args_shape(new_generalize_shape, is_sink_mode)
|
|
367
|
-
logger.info((f'generalize with generalize shape cache, compile args shape = {res_shape}'))
|
|
368
|
-
return new_generalize_shape
|
|
358
|
+
return self._generate_with_generalize_shape(generalize_shape_args, is_sink_mode, args_list)
|
|
369
359
|
|
|
370
360
|
res_shape = self.auto_dynamic_shape_manager.get_compile_args_shape(generalize_shape_args, is_sink_mode)
|
|
371
361
|
logger.debug((f'find compile args in generalize shape cache, compile args shape={res_shape}'))
|
|
@@ -396,54 +386,61 @@ class _AutoIdentifyDynamicShape:
|
|
|
396
386
|
|
|
397
387
|
def _do_generalize_one_input_shape(self, input_args, cache_args, cache_type, index, is_sink_mode):
|
|
398
388
|
"""do generalize shape one input by cache"""
|
|
389
|
+
def generalize_tensor(arg, cache, i):
|
|
390
|
+
if self.auto_dynamic_shape_manager.is_tensor_equal(arg, cache):
|
|
391
|
+
return arg
|
|
392
|
+
|
|
393
|
+
shape_value = []
|
|
394
|
+
if len(arg.shape) != len(cache.shape):
|
|
395
|
+
shape_value.append(SHAPE_RANK_ANY)
|
|
396
|
+
else:
|
|
397
|
+
shape_value = [SHAPE_DIM_ANY for _ in range(len(arg.shape))]
|
|
398
|
+
shape_tuple = tuple(shape_value)
|
|
399
|
+
logger.info((f'In the {cache_type} cache[{index}], the {i}th input tensor shape is {arg.shape},'
|
|
400
|
+
f'cache shape is {cache.shape}, not equal, need generalize to {shape_tuple}.'))
|
|
401
|
+
return Tensor(shape=shape_tuple, dtype=arg.dtype)
|
|
402
|
+
|
|
403
|
+
def generalize_sequence(arg, cache, i):
|
|
404
|
+
if is_sink_mode:
|
|
405
|
+
# when is_sink_mode=True, input must be the shape of Tensor.
|
|
406
|
+
res = self._do_generalize_in_sink(arg, cache, i, index, cache_type)
|
|
407
|
+
return res
|
|
408
|
+
|
|
409
|
+
res = self._do_generalize_one_input_shape(arg, cache, cache_type, index, is_sink_mode)
|
|
410
|
+
return res
|
|
411
|
+
|
|
399
412
|
generalize_one_shape = []
|
|
400
|
-
for i, (
|
|
401
|
-
if isinstance(
|
|
402
|
-
if self.auto_dynamic_shape_manager.is_tensor_equal(
|
|
403
|
-
generalize_one_shape.append(
|
|
413
|
+
for i, (arg, cache) in enumerate(zip(input_args, cache_args)):
|
|
414
|
+
if isinstance(arg, Parameter) and isinstance(cache, Parameter):
|
|
415
|
+
if self.auto_dynamic_shape_manager.is_tensor_equal(arg, cache):
|
|
416
|
+
generalize_one_shape.append(arg)
|
|
404
417
|
continue
|
|
405
|
-
else:
|
|
406
|
-
logger.info("In auto dynamic shape mode, parameter must be equal, it can not be generalize.")
|
|
407
|
-
return input_args, False
|
|
408
418
|
|
|
409
|
-
|
|
410
|
-
|
|
411
|
-
generalize_one_shape.append(input)
|
|
412
|
-
else:
|
|
413
|
-
shape_value = []
|
|
414
|
-
if len(input.shape) != len(cache.shape):
|
|
415
|
-
shape_value.append(SHAPE_RANK_ANY)
|
|
416
|
-
else:
|
|
417
|
-
for _ in range(len(input.shape)):
|
|
418
|
-
shape_value.append(SHAPE_DIM_ANY)
|
|
419
|
-
shape_tuple = tuple(shape_value)
|
|
420
|
-
generalize_one_shape.append(Tensor(shape=shape_tuple, dtype=input.dtype))
|
|
421
|
-
logger.info((f'In the {cache_type} cache[{index}], the {i}th input tensor shape is {input.shape},'
|
|
422
|
-
f'cache shape is {cache.shape}, not equal, need generalize to {shape_tuple}.'))
|
|
419
|
+
logger.info("In auto dynamic shape mode, parameter must be equal, it can not be generalize.")
|
|
420
|
+
return input_args, False
|
|
423
421
|
|
|
424
|
-
|
|
425
|
-
|
|
426
|
-
|
|
427
|
-
|
|
428
|
-
|
|
429
|
-
|
|
430
|
-
|
|
431
|
-
generalize_one_shape.append(res)
|
|
432
|
-
elif isinstance(input, int) and isinstance(cache, int):
|
|
422
|
+
if isinstance(arg, Tensor) and isinstance(cache, Tensor):
|
|
423
|
+
res = generalize_tensor(arg, cache, i)
|
|
424
|
+
generalize_one_shape.append(res)
|
|
425
|
+
elif isinstance(arg, (tuple, list)) and isinstance(cache, (tuple, list)):
|
|
426
|
+
res = generalize_sequence(arg, cache, i)
|
|
427
|
+
generalize_one_shape.append(res)
|
|
428
|
+
elif isinstance(arg, int) and isinstance(cache, int):
|
|
433
429
|
# when is_sink_mode=False, the input must may be scalar, or the value of list/tuple.
|
|
434
430
|
# is_sink_mode can not be True
|
|
435
|
-
if
|
|
436
|
-
generalize_one_shape.append(
|
|
431
|
+
if arg == cache:
|
|
432
|
+
generalize_one_shape.append(arg)
|
|
437
433
|
else:
|
|
438
434
|
logger.info("In auto dynamic shape mode, scalar/tuple/list must be equal, it can not be " \
|
|
439
435
|
"generalize.")
|
|
440
436
|
return input_args, False
|
|
441
|
-
elif
|
|
442
|
-
generalize_one_shape.append(
|
|
437
|
+
elif arg is None and cache is None:
|
|
438
|
+
generalize_one_shape.append(arg)
|
|
443
439
|
|
|
444
440
|
return generalize_one_shape, True
|
|
445
441
|
|
|
446
442
|
|
|
443
|
+
|
|
447
444
|
def _do_generalize_shape(self, cache_type, input_args, is_sink_mode):
|
|
448
445
|
"""do generalize shape by cache"""
|
|
449
446
|
shape_cache = self.auto_dynamic_shape_manager.get_cache_by_type(cache_type)
|
|
@@ -464,8 +461,7 @@ class _AutoIdentifyDynamicShape:
|
|
|
464
461
|
|
|
465
462
|
keys = list(unknown_shape_dict.keys())
|
|
466
463
|
keys.sort(key=lambda x: (x[0], x[1]))
|
|
467
|
-
|
|
468
|
-
return unknown_shape_dict.get(index), True
|
|
464
|
+
return unknown_shape_dict.get(keys[0]), True
|
|
469
465
|
|
|
470
466
|
_auto_dynamic_shape = _AutoIdentifyDynamicShape()
|
|
471
467
|
|
mindspore/common/dump.py
CHANGED
|
@@ -57,7 +57,7 @@ def set_dump(target, enabled=True):
|
|
|
57
57
|
``Ascend``
|
|
58
58
|
|
|
59
59
|
Examples:
|
|
60
|
-
..
|
|
60
|
+
.. note::
|
|
61
61
|
Please set environment variable `MINDSPORE_DUMP_CONFIG` to the dump config file and set `dump_mode` field
|
|
62
62
|
in dump config file to 2 before running this example.
|
|
63
63
|
See `dump document <https://www.mindspore.cn/tutorials/experts/en/r2.2/debug/dump.html>`_ for details.
|
mindspore/common/tensor.py
CHANGED
|
@@ -2186,7 +2186,7 @@ class Tensor(Tensor_, metaclass=_TensorMeta):
|
|
|
2186
2186
|
[0.7] [3]
|
|
2187
2187
|
"""
|
|
2188
2188
|
if self.shape == ():
|
|
2189
|
-
return (Tensor(0)
|
|
2189
|
+
return (self, Tensor(0))
|
|
2190
2190
|
self._init_check()
|
|
2191
2191
|
return tensor_operator_registry.get('argmax_with_value')(self, axis, keep_dims)
|
|
2192
2192
|
|
|
@@ -2234,7 +2234,7 @@ class Tensor(Tensor_, metaclass=_TensorMeta):
|
|
|
2234
2234
|
[0.0] [0]
|
|
2235
2235
|
"""
|
|
2236
2236
|
if self.shape == ():
|
|
2237
|
-
return (Tensor(0)
|
|
2237
|
+
return (self, Tensor(0))
|
|
2238
2238
|
self._init_check()
|
|
2239
2239
|
return tensor_operator_registry.get('argmin_with_value')(self, axis, keep_dims)
|
|
2240
2240
|
|
|
@@ -2683,7 +2683,6 @@ class Tensor(Tensor_, metaclass=_TensorMeta):
|
|
|
2683
2683
|
def _init_check(self):
|
|
2684
2684
|
if self.has_init:
|
|
2685
2685
|
self.init_data()
|
|
2686
|
-
return self
|
|
2687
2686
|
|
|
2688
2687
|
def init_data(self, slice_index=None, shape=None, opt_shard_group=None):
|
|
2689
2688
|
"""
|
|
@@ -4631,23 +4630,7 @@ class Tensor(Tensor_, metaclass=_TensorMeta):
|
|
|
4631
4630
|
|
|
4632
4631
|
def imag(self):
|
|
4633
4632
|
r"""
|
|
4634
|
-
|
|
4635
|
-
If input tensor is real, it will return zeros.
|
|
4636
|
-
|
|
4637
|
-
Returns:
|
|
4638
|
-
Tensor, the shape is the same as the input tensor.
|
|
4639
|
-
|
|
4640
|
-
Supported Platforms:
|
|
4641
|
-
``GPU`` ``CPU``
|
|
4642
|
-
|
|
4643
|
-
Examples:
|
|
4644
|
-
>>> import numpy as np
|
|
4645
|
-
>>> import mindspore
|
|
4646
|
-
>>> from mindspore import Tensor
|
|
4647
|
-
>>> x = Tensor(np.asarray(np.complex(1.3 + 0.4j)), mindspore.complex64)
|
|
4648
|
-
>>> output = x.imag()
|
|
4649
|
-
>>> print(output)
|
|
4650
|
-
0.4
|
|
4633
|
+
For details, please refer to :func:`mindspore.ops.imag`.
|
|
4651
4634
|
"""
|
|
4652
4635
|
self._init_check()
|
|
4653
4636
|
return tensor_operator_registry.get('imag')(self)
|
mindspore/config/op_info.config
CHANGED
|
@@ -9,7 +9,7 @@
|
|
|
9
9
|
{"op_name": "Less", "inputs": [{"index": 0, "name": "x1", "paramType": "required"}, {"index": 1, "name": "x2", "paramType": "required"}], "outputs": [{"index": 0, "name": "y", "paramType": "required"}], "attr": [], "fusion_type": "OPAQUE", "dtype_format": [[["int8", "DefaultFormat"], ["int8", "DefaultFormat"], ["bool", "DefaultFormat"]], [["int16", "DefaultFormat"], ["int16", "DefaultFormat"], ["bool", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["bool", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["bool", "DefaultFormat"]], [["uint8", "DefaultFormat"], ["uint8", "DefaultFormat"], ["bool", "DefaultFormat"]], [["uint16", "DefaultFormat"], ["uint16", "DefaultFormat"], ["bool", "DefaultFormat"]], [["uint32", "DefaultFormat"], ["uint32", "DefaultFormat"], ["bool", "DefaultFormat"]], [["uint64", "DefaultFormat"], ["uint64", "DefaultFormat"], ["bool", "DefaultFormat"]], [["float16", "DefaultFormat"], ["float16", "DefaultFormat"], ["bool", "DefaultFormat"]], [["float32", "DefaultFormat"], ["float32", "DefaultFormat"], ["bool", "DefaultFormat"]], [["float64", "DefaultFormat"], ["float64", "DefaultFormat"], ["bool", "DefaultFormat"]]], "imply_type": "AiCPU"}
|
|
10
10
|
{"op_name": "Lstsq", "inputs": [{"index": 0, "name": "matrix", "paramType": "required"}, {"index": 1, "name": "rhs", "paramType": "required"}], "outputs": [{"index": 0, "name": "y", "paramType": "required"}], "attr": [{"name": "l2_regularizer", "type": "float", "value": "0.0"}, {"name": "fast", "type": "bool", "value": "True"}], "fusion_type": "OPAQUE", "dtype_format": [[["float16", "DefaultFormat"], ["float16", "DefaultFormat"], ["float16", "DefaultFormat"]], [["float32", "DefaultFormat"], ["float32", "DefaultFormat"], ["float32", "DefaultFormat"]], [["float64", "DefaultFormat"], ["float64", "DefaultFormat"], ["float64", "DefaultFormat"]]], "imply_type": "AiCPU"}
|
|
11
11
|
{"op_name": "LeftShift", "inputs": [{"index": 0, "name": "x1", "paramType": "required"}, {"index": 1, "name": "x2", "paramType": "required"}], "outputs": [{"index": 0, "name": "y", "paramType": "required"}], "attr": [], "fusion_type": "OPAQUE", "dtype_format": [[["int8", "DefaultFormat"], ["int8", "DefaultFormat"], ["int8", "DefaultFormat"]], [["int16", "DefaultFormat"], ["int16", "DefaultFormat"], ["int16", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"]], [["uint8", "DefaultFormat"], ["uint8", "DefaultFormat"], ["uint8", "DefaultFormat"]], [["uint16", "DefaultFormat"], ["uint16", "DefaultFormat"], ["uint16", "DefaultFormat"]], [["uint32", "DefaultFormat"], ["uint32", "DefaultFormat"], ["uint32", "DefaultFormat"]], [["uint64", "DefaultFormat"], ["uint64", "DefaultFormat"], ["uint64", "DefaultFormat"]]], "imply_type": "AiCPU"}
|
|
12
|
-
{"op_name": "Add", "inputs": [{"index": 0, "name": "x", "paramType": "required"}, {"index": 1, "name": "y", "paramType": "required"}], "outputs": [{"index": 0, "name": "output", "paramType": "required"}], "attr": [], "fusion_type": "OPAQUE", "dtype_format": [[["float16", "DefaultFormat"], ["float16", "DefaultFormat"], ["float16", "DefaultFormat"]], [["float32", "DefaultFormat"], ["float32", "DefaultFormat"], ["float32", "DefaultFormat"]], [["float64", "DefaultFormat"], ["float64", "DefaultFormat"], ["float64", "DefaultFormat"]], [["int8", "DefaultFormat"], ["int8", "DefaultFormat"], ["int8", "DefaultFormat"]], [["int16", "DefaultFormat"], ["int16", "DefaultFormat"], ["int16", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"]], [["uint8", "DefaultFormat"], ["uint8", "DefaultFormat"], ["uint8", "DefaultFormat"]], [["uint16", "DefaultFormat"], ["
|
|
12
|
+
{"op_name": "Add", "inputs": [{"index": 0, "name": "x", "paramType": "required"}, {"index": 1, "name": "y", "paramType": "required"}], "outputs": [{"index": 0, "name": "output", "paramType": "required"}], "attr": [], "fusion_type": "OPAQUE", "dtype_format": [[["float16", "DefaultFormat"], ["float16", "DefaultFormat"], ["float16", "DefaultFormat"]], [["float32", "DefaultFormat"], ["float32", "DefaultFormat"], ["float32", "DefaultFormat"]], [["float64", "DefaultFormat"], ["float64", "DefaultFormat"], ["float64", "DefaultFormat"]], [["int8", "DefaultFormat"], ["int8", "DefaultFormat"], ["int8", "DefaultFormat"]], [["int16", "DefaultFormat"], ["int16", "DefaultFormat"], ["int16", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"]], [["uint8", "DefaultFormat"], ["uint8", "DefaultFormat"], ["uint8", "DefaultFormat"]], [["uint16", "DefaultFormat"], ["uint16", "DefaultFormat"], ["uint16", "DefaultFormat"]], [["uint32", "DefaultFormat"], ["uint32", "DefaultFormat"], ["uint32", "DefaultFormat"]], [["uint64", "DefaultFormat"], ["uint64", "DefaultFormat"], ["uint64", "DefaultFormat"]], [["complex64", "DefaultFormat"], ["complex64", "DefaultFormat"], ["complex64", "DefaultFormat"]], [["complex128", "DefaultFormat"], ["complex128", "DefaultFormat"], ["complex128", "DefaultFormat"]]], "imply_type": "AiCPU"}
|
|
13
13
|
{"op_name": "SparseMatrixTranspose", "inputs": [{"index": 0, "name": "x_dense_shape", "paramType": "required"}, {"index": 1, "name": "x_batch_pointers", "paramType": "required"}, {"index": 2, "name": "x_row_pointers", "paramType": "required"}, {"index": 3, "name": "x_col_indices", "paramType": "required"}, {"index": 4, "name": "x_values", "paramType": "required"}], "outputs": [{"index": 0, "name": "y_dense_shape", "paramType": "required"}, {"index": 1, "name": "y_batch_pointers", "paramType": "required"}, {"index": 2, "name": "y_row_pointers", "paramType": "required"}, {"index": 3, "name": "y_col_indices", "paramType": "required"}, {"index": 4, "name": "y_values", "paramType": "required"}], "attr": [{"name": "conjugate", "type": "bool"}], "fusion_type": "OPAQUE", "dtype_format": [[["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int8", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int8", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["uint8", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["uint8", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int16", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int16", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["uint16", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["uint16", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["uint32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["uint32", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int64", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int64", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["uint64", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["uint64", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["float16", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["float16", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["float32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["float32", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["float64", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["float64", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["complex64", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["complex64", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["complex128", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["complex128", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int8", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int8", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["uint8", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["uint8", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int16", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int16", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["uint16", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["uint16", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int32", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["uint32", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["uint32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["uint64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["uint64", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["float16", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["float16", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["float32", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["float32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["float64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["float64", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["complex64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["complex64", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["complex128", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["complex128", "DefaultFormat"]]], "imply_type": "AiCPU"}
|
|
14
14
|
{"op_name": "SparseMatrixNNZ", "inputs": [{"index": 0, "name": "x_dense_shape", "paramType": "required"}, {"index": 1, "name": "x_batch_pointers", "paramType": "required"}, {"index": 2, "name": "x_row_pointers", "paramType": "required"}, {"index": 3, "name": "x_col_indices", "paramType": "required"}, {"index": 4, "name": "x_values", "paramType": "required"}], "outputs": [{"index": 0, "name": "y", "paramType": "required"}], "attr": [], "fusion_type": "OPAQUE", "dtype_format": [[["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int8", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["uint8", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int16", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["uint16", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int64", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["float16", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["float32", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["float64", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["bool", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["complex64", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["complex128", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int8", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["uint8", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int16", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["uint16", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["float16", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["float32", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["float64", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["bool", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["complex64", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["complex128", "DefaultFormat"], ["int32", "DefaultFormat"]]], "imply_type": "AiCPU"}
|
|
15
15
|
{"op_name": "SparseDenseCwiseMul", "inputs": [{"index": 0, "name": "x1_indices", "paramType": "required"}, {"index": 1, "name": "x1_values", "paramType": "required"}, {"index": 2, "name": "x1_sparse", "paramType": "required"}, {"index": 3, "name": "x2", "paramType": "required"}], "outputs": [{"index": 0, "name": "y", "paramType": "required"}], "attr": [], "fusion_type": "OPAQUE", "dtype_format": [[["int64", "DefaultFormat"], ["int8", "DefaultFormat"], ["int64", "DefaultFormat"], ["int8", "DefaultFormat"], ["int8", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int16", "DefaultFormat"], ["int64", "DefaultFormat"], ["int16", "DefaultFormat"], ["int16", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int32", "DefaultFormat"], ["int64", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"]], [["int64", "DefaultFormat"], ["uint8", "DefaultFormat"], ["int64", "DefaultFormat"], ["uint8", "DefaultFormat"], ["uint8", "DefaultFormat"]], [["int64", "DefaultFormat"], ["uint16", "DefaultFormat"], ["int64", "DefaultFormat"], ["uint16", "DefaultFormat"], ["uint16", "DefaultFormat"]], [["int64", "DefaultFormat"], ["uint32", "DefaultFormat"], ["int64", "DefaultFormat"], ["uint32", "DefaultFormat"], ["uint32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["uint64", "DefaultFormat"], ["int64", "DefaultFormat"], ["uint64", "DefaultFormat"], ["uint64", "DefaultFormat"]], [["int64", "DefaultFormat"], ["float16", "DefaultFormat"], ["int64", "DefaultFormat"], ["float16", "DefaultFormat"], ["float16", "DefaultFormat"]], [["int64", "DefaultFormat"], ["float32", "DefaultFormat"], ["int64", "DefaultFormat"], ["float32", "DefaultFormat"], ["float32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["float64", "DefaultFormat"], ["int64", "DefaultFormat"], ["float64", "DefaultFormat"], ["float64", "DefaultFormat"]], [["int64", "DefaultFormat"], ["complex64", "DefaultFormat"], ["int64", "DefaultFormat"], ["complex64", "DefaultFormat"], ["complex64", "DefaultFormat"]], [["int64", "DefaultFormat"], ["complex128", "DefaultFormat"], ["int64", "DefaultFormat"], ["complex128", "DefaultFormat"], ["complex128", "DefaultFormat"]]], "imply_type": "AiCPU"}
|
mindspore/context.py
CHANGED
|
@@ -268,6 +268,8 @@ class _Context:
|
|
|
268
268
|
"allow_mix_precision_fp16" and "allow_mix_precision_bf16".
|
|
269
269
|
- jit_compile (bool): ``False`` and ``True``.
|
|
270
270
|
- atomic_clean_policy (int): ``0`` and ``1``. Default: ``1`` .
|
|
271
|
+
- exception_dump (str): Enable exception dump for Ascend operators. ``"0"`` , ``"1"`` and ``"2"``.
|
|
272
|
+
Default: ``"2"`` .
|
|
271
273
|
- op_precision_mode (str): config file path.
|
|
272
274
|
- parallel_speed_up_json_path(Union[str, None]): The path to the parallel speed up json file.
|
|
273
275
|
If its value is None or '', it does not take effect. Default None.
|
|
@@ -280,6 +282,7 @@ class _Context:
|
|
|
280
282
|
'atomic_clean_policy': [0, 1],
|
|
281
283
|
'matmul_allow_hf32': [True, False],
|
|
282
284
|
'conv_allow_hf32': [True, False],
|
|
285
|
+
'exception_dump': ["0", "1", "2"],
|
|
283
286
|
'op_precision_mode': (str,),
|
|
284
287
|
'parallel_speed_up_json_path': (str, None)
|
|
285
288
|
}
|
|
@@ -289,6 +292,7 @@ class _Context:
|
|
|
289
292
|
'atomic_clean_policy': self._get_ascend_config_setter('atomic_clean_policy', str),
|
|
290
293
|
'matmul_allow_hf32': self._get_ascend_config_setter('matmul_allow_hf32', lambda v: "1" if v else "0"),
|
|
291
294
|
'conv_allow_hf32': self._get_ascend_config_setter('conv_allow_hf32', lambda v: "1" if v else "0"),
|
|
295
|
+
'exception_dump': self._get_ascend_config_setter('exception_dump'),
|
|
292
296
|
'op_precision_mode': self._set_op_precision_mode,
|
|
293
297
|
'parallel_speed_up_json_path': self._set_speedup_config_path
|
|
294
298
|
}
|
|
@@ -629,6 +633,7 @@ class _Context:
|
|
|
629
633
|
"enable_task_opt": ms_ctx_param.enable_task_opt,
|
|
630
634
|
"enable_grad_comm_opt": ms_ctx_param.enable_grad_comm_opt,
|
|
631
635
|
"interleaved_matmul_comm": ms_ctx_param.interleaved_matmul_comm,
|
|
636
|
+
"enable_opt_shard_comm_opt": ms_ctx_param.enable_opt_shard_comm_opt,
|
|
632
637
|
"interleaved_layernorm_comm": ms_ctx_param.interleaved_layernorm_comm}
|
|
633
638
|
with open(speedup_config_real_path, 'r') as f:
|
|
634
639
|
speedup_config = json.load(f)
|
|
@@ -1277,12 +1282,10 @@ def set_context(**kwargs):
|
|
|
1277
1282
|
memory_optimize_level is set 'O1'.
|
|
1278
1283
|
- OFF: Turn off the memory Offload function.
|
|
1279
1284
|
ascend_config (dict): Set the parameters specific to Ascend hardware platform. It is not set by default.
|
|
1280
|
-
|
|
1281
|
-
supported on Ascend910B hardware platform. The default value of `precision_mode`, `jit_compile` and
|
|
1285
|
+
The default value of `precision_mode`, `jit_compile` and
|
|
1282
1286
|
`atomic_clean_policy` are experimental parameters, may change in the future.
|
|
1283
1287
|
|
|
1284
|
-
- precision_mode (str): Mixed precision mode setting,
|
|
1285
|
-
value of training network is based on the value of CANN, and the default value of inference network
|
|
1288
|
+
- precision_mode (str): Mixed precision mode setting, and the default value of inference network
|
|
1286
1289
|
is ``force_fp16`` . The value range is as follows:
|
|
1287
1290
|
|
|
1288
1291
|
- force_fp16: When the operator supports both float16 and float32, select float16 directly.
|
|
@@ -1315,6 +1318,10 @@ def set_context(**kwargs):
|
|
|
1315
1318
|
- conv_allow_hf32 (bool): Whether to convert FP32 to HF32 for Conv operators. Default value: ``True``.
|
|
1316
1319
|
This is an experimental prototype that is subject to change and/or deletion.
|
|
1317
1320
|
For detailed information, please refer to `Ascend community <https://www.hiascend.com/>`_ .
|
|
1321
|
+
- exception_dump (str): Enable exception dump for Ascend operators, providing the input and output data for
|
|
1322
|
+
failing Ascend operators. The value can be ``"0"`` , ``"1"`` and ``"2"``. For ``"0"`` , exception dump is
|
|
1323
|
+
turned off; for ``"1"``, all inputs and outputs will be dumped for AICore and AICPU exception operators;
|
|
1324
|
+
for ``"2"``, inputs will be dumped for AICore exception operators. Default: ``"2"`` .
|
|
1318
1325
|
- op_precision_mode (str): Path to config file of op precision mode. For detailed information, please refer
|
|
1319
1326
|
to `Ascend community <https://www.hiascend.com/>`_ .
|
|
1320
1327
|
- parallel_speed_up_json_path(Union[str, None]): The path to the parallel speed up json file, configuration
|
|
@@ -232,6 +232,9 @@ class TFRecordDataset(SourceDataset, UnionBaseDataset):
|
|
|
232
232
|
|
|
233
233
|
The columns of generated dataset depend on the source TFRecord files.
|
|
234
234
|
|
|
235
|
+
Note:
|
|
236
|
+
'TFRecordDataset' is not support on Windows platform yet.
|
|
237
|
+
|
|
235
238
|
Args:
|
|
236
239
|
dataset_files (Union[str, list[str]]): String or list of files to be read or glob strings to search for a
|
|
237
240
|
pattern of files. The list will be sorted in lexicographical order.
|
|
@@ -318,6 +321,8 @@ class TFRecordDataset(SourceDataset, UnionBaseDataset):
|
|
|
318
321
|
cache=None, compression_type=None):
|
|
319
322
|
super().__init__(num_parallel_workers=num_parallel_workers, num_samples=num_samples, shuffle=shuffle,
|
|
320
323
|
num_shards=num_shards, shard_id=shard_id, cache=cache)
|
|
324
|
+
if platform.system().lower() == "windows":
|
|
325
|
+
raise NotImplementedError("TFRecordDataset is not supported for windows.")
|
|
321
326
|
self.dataset_files = self._find_files(dataset_files)
|
|
322
327
|
self.dataset_files.sort()
|
|
323
328
|
|
|
@@ -144,14 +144,14 @@ class AdjustBrightness(ImageTensorOperation, PyTensorOperation):
|
|
|
144
144
|
|
|
145
145
|
Args:
|
|
146
146
|
device_target (str, optional): The operator will be executed on this device. Currently supports
|
|
147
|
-
``CPU``
|
|
147
|
+
``CPU`` . Default: ``CPU`` .
|
|
148
148
|
|
|
149
149
|
Raises:
|
|
150
150
|
TypeError: If `device_target` is not of type str.
|
|
151
|
-
ValueError: If `device_target` is not
|
|
151
|
+
ValueError: If `device_target` is not ``CPU`` .
|
|
152
152
|
|
|
153
153
|
Supported Platforms:
|
|
154
|
-
``CPU``
|
|
154
|
+
``CPU``
|
|
155
155
|
|
|
156
156
|
Examples:
|
|
157
157
|
>>> import mindspore.dataset as ds
|
|
@@ -227,14 +227,14 @@ class AdjustContrast(ImageTensorOperation, PyTensorOperation):
|
|
|
227
227
|
|
|
228
228
|
Args:
|
|
229
229
|
device_target (str, optional): The operator will be executed on this device. Currently supports
|
|
230
|
-
``CPU``
|
|
230
|
+
``CPU`` . Default: ``CPU`` .
|
|
231
231
|
|
|
232
232
|
Raises:
|
|
233
233
|
TypeError: If `device_target` is not of type str.
|
|
234
|
-
ValueError: If `device_target` is not
|
|
234
|
+
ValueError: If `device_target` is not ``CPU`` .
|
|
235
235
|
|
|
236
236
|
Supported Platforms:
|
|
237
|
-
``CPU``
|
|
237
|
+
``CPU``
|
|
238
238
|
|
|
239
239
|
Examples:
|
|
240
240
|
>>> import mindspore.dataset as ds
|
|
@@ -373,14 +373,14 @@ class AdjustHue(ImageTensorOperation, PyTensorOperation):
|
|
|
373
373
|
|
|
374
374
|
Args:
|
|
375
375
|
device_target (str, optional): The operator will be executed on this device. Currently supports
|
|
376
|
-
``CPU``
|
|
376
|
+
``CPU`` . Default: ``CPU`` .
|
|
377
377
|
|
|
378
378
|
Raises:
|
|
379
379
|
TypeError: If `device_target` is not of type str.
|
|
380
|
-
ValueError: If `device_target` is not
|
|
380
|
+
ValueError: If `device_target` is not ``CPU`` .
|
|
381
381
|
|
|
382
382
|
Supported Platforms:
|
|
383
|
-
``CPU``
|
|
383
|
+
``CPU``
|
|
384
384
|
|
|
385
385
|
Examples:
|
|
386
386
|
>>> import mindspore.dataset as ds
|
|
@@ -457,14 +457,14 @@ class AdjustSaturation(ImageTensorOperation, PyTensorOperation):
|
|
|
457
457
|
|
|
458
458
|
Args:
|
|
459
459
|
device_target (str, optional): The operator will be executed on this device. Currently supports
|
|
460
|
-
``CPU``
|
|
460
|
+
``CPU`` . Default: ``CPU`` .
|
|
461
461
|
|
|
462
462
|
Raises:
|
|
463
463
|
TypeError: If `device_target` is not of type str.
|
|
464
|
-
ValueError: If `device_target` is not
|
|
464
|
+
ValueError: If `device_target` is not ``CPU`` .
|
|
465
465
|
|
|
466
466
|
Supported Platforms:
|
|
467
|
-
``CPU``
|
|
467
|
+
``CPU``
|
|
468
468
|
|
|
469
469
|
Examples:
|
|
470
470
|
>>> import mindspore.dataset as ds
|
|
@@ -1159,14 +1159,14 @@ class Decode(ImageTensorOperation, PyTensorOperation):
|
|
|
1159
1159
|
|
|
1160
1160
|
Args:
|
|
1161
1161
|
device_target (str, optional): The operator will be executed on this device. Currently supports
|
|
1162
|
-
``CPU``
|
|
1162
|
+
``CPU`` . Default: ``CPU`` .
|
|
1163
1163
|
|
|
1164
1164
|
Raises:
|
|
1165
1165
|
TypeError: If `device_target` is not of type str.
|
|
1166
|
-
ValueError: If `device_target` is not
|
|
1166
|
+
ValueError: If `device_target` is not ``CPU`` .
|
|
1167
1167
|
|
|
1168
1168
|
Supported Platforms:
|
|
1169
|
-
``CPU``
|
|
1169
|
+
``CPU``
|
|
1170
1170
|
|
|
1171
1171
|
Examples:
|
|
1172
1172
|
>>> import mindspore.dataset as ds
|
|
@@ -1908,14 +1908,14 @@ class Normalize(ImageTensorOperation):
|
|
|
1908
1908
|
|
|
1909
1909
|
Args:
|
|
1910
1910
|
device_target (str, optional): The operator will be executed on this device. Currently supports
|
|
1911
|
-
``CPU``
|
|
1911
|
+
``CPU`` . Default: ``CPU`` .
|
|
1912
1912
|
|
|
1913
1913
|
Raises:
|
|
1914
1914
|
TypeError: If `device_target` is not of type str.
|
|
1915
|
-
ValueError: If `device_target` is not
|
|
1915
|
+
ValueError: If `device_target` is not ``CPU`` .
|
|
1916
1916
|
|
|
1917
1917
|
Supported Platforms:
|
|
1918
|
-
``CPU``
|
|
1918
|
+
``CPU``
|
|
1919
1919
|
|
|
1920
1920
|
Examples:
|
|
1921
1921
|
>>> import mindspore.dataset as ds
|
|
@@ -4182,14 +4182,14 @@ class Resize(ImageTensorOperation, PyTensorOperation):
|
|
|
4182
4182
|
|
|
4183
4183
|
Args:
|
|
4184
4184
|
device_target (str, optional): The operator will be executed on this device. Currently supports
|
|
4185
|
-
``CPU``
|
|
4185
|
+
``CPU`` . Default: ``CPU`` .
|
|
4186
4186
|
|
|
4187
4187
|
Raises:
|
|
4188
4188
|
TypeError: If `device_target` is not of type str.
|
|
4189
|
-
ValueError: If `device_target` is not
|
|
4189
|
+
ValueError: If `device_target` is not ``CPU`` .
|
|
4190
4190
|
|
|
4191
4191
|
Supported Platforms:
|
|
4192
|
-
``CPU``
|
|
4192
|
+
``CPU``
|
|
4193
4193
|
|
|
4194
4194
|
Examples:
|
|
4195
4195
|
>>> import mindspore.dataset as ds
|
mindspore/gen_ops.py
CHANGED
|
@@ -120,7 +120,7 @@ def generate_py_primitive(yaml_data):
|
|
|
120
120
|
assign_str += arg_name
|
|
121
121
|
args_assign.append(assign_str)
|
|
122
122
|
|
|
123
|
-
args_assign = '\n'.join(assign for assign in args_assign)
|
|
123
|
+
args_assign = '\n'.join([assign for assign in args_assign])
|
|
124
124
|
primitive_code = f"""
|
|
125
125
|
class {class_name}(Primitive):
|
|
126
126
|
def __init__(self, {', '.join(init_args_with_default)}):
|