mindspore 2.2.0__cp39-cp39-win_amd64.whl → 2.2.10__cp39-cp39-win_amd64.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.

Potentially problematic release.


This version of mindspore might be problematic. Click here for more details.

Files changed (122) hide show
  1. mindspore/.commit_id +1 -1
  2. mindspore/Microsoft.VisualStudio.Telemetry.dll +0 -0
  3. mindspore/Newtonsoft.Json.dll +0 -0
  4. mindspore/_c_dataengine.cp39-win_amd64.pyd +0 -0
  5. mindspore/_c_expression.cp39-win_amd64.pyd +0 -0
  6. mindspore/_c_mindrecord.cp39-win_amd64.pyd +0 -0
  7. mindspore/_checkparam.py +3 -3
  8. mindspore/_extends/graph_kernel/model/graph_split.py +84 -76
  9. mindspore/_extends/graph_kernel/splitter.py +3 -2
  10. mindspore/_extends/parallel_compile/akg_compiler/build_tbe_kernel.py +83 -66
  11. mindspore/_extends/parallel_compile/akg_compiler/tbe_topi.py +4 -4
  12. mindspore/_extends/parallel_compile/akg_compiler/util.py +10 -7
  13. mindspore/_extends/parallel_compile/tbe_compiler/tbe_helper.py +2 -1
  14. mindspore/_extends/parse/standard_method.py +2 -9
  15. mindspore/_extends/remote/kernel_build_server.py +2 -1
  16. mindspore/atlprov.dll +0 -0
  17. mindspore/c1.dll +0 -0
  18. mindspore/c1xx.dll +0 -0
  19. mindspore/c2.dll +0 -0
  20. mindspore/common/api.py +1 -1
  21. mindspore/common/auto_dynamic_shape.py +81 -85
  22. mindspore/common/dump.py +1 -1
  23. mindspore/common/tensor.py +3 -20
  24. mindspore/config/op_info.config +1 -1
  25. mindspore/context.py +11 -4
  26. mindspore/dataset/engine/datasets_standard_format.py +5 -0
  27. mindspore/dataset/vision/transforms.py +21 -21
  28. mindspore/dnnl.dll +0 -0
  29. mindspore/dpcmi.dll +0 -0
  30. mindspore/experimental/optim/adam.py +1 -1
  31. mindspore/gen_ops.py +1 -1
  32. mindspore/include/api/model.h +17 -0
  33. mindspore/include/api/status.h +8 -3
  34. mindspore/jpeg62.dll +0 -0
  35. mindspore/mindspore_backend.dll +0 -0
  36. mindspore/mindspore_common.dll +0 -0
  37. mindspore/mindspore_core.dll +0 -0
  38. mindspore/mindspore_glog.dll +0 -0
  39. mindspore/mindspore_shared_lib.dll +0 -0
  40. mindspore/msobj140.dll +0 -0
  41. mindspore/mspdb140.dll +0 -0
  42. mindspore/mspdbcore.dll +0 -0
  43. mindspore/mspdbst.dll +0 -0
  44. mindspore/mspft140.dll +0 -0
  45. mindspore/msvcdis140.dll +0 -0
  46. mindspore/msvcp140_1.dll +0 -0
  47. mindspore/msvcp140_2.dll +0 -0
  48. mindspore/msvcp140_atomic_wait.dll +0 -0
  49. mindspore/msvcp140_codecvt_ids.dll +0 -0
  50. mindspore/nn/cell.py +0 -3
  51. mindspore/nn/layer/activation.py +4 -5
  52. mindspore/nn/layer/conv.py +39 -23
  53. mindspore/nn/layer/flash_attention.py +90 -78
  54. mindspore/nn/layer/math.py +3 -7
  55. mindspore/nn/layer/rnn_cells.py +5 -5
  56. mindspore/nn/wrap/cell_wrapper.py +6 -0
  57. mindspore/numpy/utils_const.py +5 -5
  58. mindspore/opencv_core452.dll +0 -0
  59. mindspore/opencv_imgcodecs452.dll +0 -0
  60. mindspore/opencv_imgproc452.dll +0 -0
  61. mindspore/ops/_grad_experimental/grad_array_ops.py +1 -1
  62. mindspore/ops/_grad_experimental/grad_implementations.py +2 -2
  63. mindspore/ops/_grad_experimental/grad_math_ops.py +19 -18
  64. mindspore/ops/_grad_experimental/grad_sparse_ops.py +3 -3
  65. mindspore/ops/_op_impl/aicpu/add.py +3 -3
  66. mindspore/ops/_utils/utils.py +2 -0
  67. mindspore/ops/composite/multitype_ops/_compile_utils.py +2 -1
  68. mindspore/ops/composite/multitype_ops/getitem_impl.py +2 -2
  69. mindspore/ops/function/array_func.py +10 -7
  70. mindspore/ops/function/grad/grad_func.py +0 -1
  71. mindspore/ops/function/nn_func.py +98 -9
  72. mindspore/ops/function/random_func.py +2 -1
  73. mindspore/ops/op_info_register.py +24 -21
  74. mindspore/ops/operations/__init__.py +3 -2
  75. mindspore/ops/operations/_grad_ops.py +24 -4
  76. mindspore/ops/operations/_inner_ops.py +155 -23
  77. mindspore/ops/operations/array_ops.py +9 -7
  78. mindspore/ops/operations/comm_ops.py +2 -2
  79. mindspore/ops/operations/custom_ops.py +85 -68
  80. mindspore/ops/operations/inner_ops.py +26 -3
  81. mindspore/ops/operations/math_ops.py +4 -3
  82. mindspore/ops/operations/nn_ops.py +109 -28
  83. mindspore/parallel/_parallel_serialization.py +10 -3
  84. mindspore/parallel/_tensor.py +4 -1
  85. mindspore/parallel/checkpoint_transform.py +13 -2
  86. mindspore/parallel/shard.py +17 -10
  87. mindspore/pgodb140.dll +0 -0
  88. mindspore/pgort140.dll +0 -0
  89. mindspore/profiler/common/util.py +1 -0
  90. mindspore/profiler/parser/ascend_hccl_generator.py +232 -0
  91. mindspore/profiler/parser/ascend_msprof_exporter.py +86 -43
  92. mindspore/profiler/parser/ascend_msprof_generator.py +196 -9
  93. mindspore/profiler/parser/ascend_op_generator.py +1 -1
  94. mindspore/profiler/parser/ascend_timeline_generator.py +6 -182
  95. mindspore/profiler/parser/base_timeline_generator.py +1 -1
  96. mindspore/profiler/parser/cpu_gpu_timeline_generator.py +2 -2
  97. mindspore/profiler/parser/framework_parser.py +1 -1
  98. mindspore/profiler/parser/profiler_info.py +19 -0
  99. mindspore/profiler/profiling.py +46 -24
  100. mindspore/rewrite/api/pattern_engine.py +1 -1
  101. mindspore/rewrite/parsers/for_parser.py +1 -1
  102. mindspore/rewrite/symbol_tree.py +1 -4
  103. mindspore/run_check/_check_version.py +5 -3
  104. mindspore/safeguard/rewrite_obfuscation.py +52 -28
  105. mindspore/tbbmalloc.dll +0 -0
  106. mindspore/tinyxml2.dll +0 -0
  107. mindspore/train/callback/_summary_collector.py +1 -1
  108. mindspore/train/dataset_helper.py +1 -0
  109. mindspore/train/model.py +2 -2
  110. mindspore/train/serialization.py +97 -11
  111. mindspore/train/summary/_summary_adapter.py +1 -1
  112. mindspore/train/summary/summary_record.py +23 -7
  113. mindspore/turbojpeg.dll +0 -0
  114. mindspore/vcmeta.dll +0 -0
  115. mindspore/vcruntime140.dll +0 -0
  116. mindspore/vcruntime140_1.dll +0 -0
  117. mindspore/version.py +1 -1
  118. {mindspore-2.2.0.dist-info → mindspore-2.2.10.dist-info}/METADATA +1 -1
  119. {mindspore-2.2.0.dist-info → mindspore-2.2.10.dist-info}/RECORD +122 -122
  120. {mindspore-2.2.0.dist-info → mindspore-2.2.10.dist-info}/WHEEL +0 -0
  121. {mindspore-2.2.0.dist-info → mindspore-2.2.10.dist-info}/entry_points.txt +0 -0
  122. {mindspore-2.2.0.dist-info → mindspore-2.2.10.dist-info}/top_level.txt +0 -0
@@ -27,6 +27,7 @@ SHAPE_RANK_ANY = -2
27
27
 
28
28
  auto_dynamic_shepe_dict = {}
29
29
 
30
+
30
31
  class _AutoDynamicShapeManager:
31
32
  """
32
33
  Represents a function to manage auto identify dynamic shape.
@@ -101,14 +102,14 @@ class _AutoDynamicShapeManager:
101
102
 
102
103
  def get_compile_args_shape_without_sink(self, input_args, res_shape):
103
104
  """get compile args shape with out sink mode"""
104
- for input in input_args:
105
- if isinstance(input, Tensor):
106
- res_shape.append(input.shape)
107
- elif isinstance(input, (int, float)):
105
+ for arg in input_args:
106
+ if isinstance(arg, Tensor):
107
+ res_shape.append(arg.shape)
108
+ elif isinstance(arg, (int, float)):
108
109
  res_shape.append([])
109
- elif isinstance(input, (tuple, list)):
110
+ elif isinstance(arg, (tuple, list)):
110
111
  tmp_shape = []
111
- self.get_compile_args_shape_without_sink(input, tmp_shape)
112
+ self.get_compile_args_shape_without_sink(arg, tmp_shape)
112
113
  res_shape.append(tmp_shape)
113
114
 
114
115
 
@@ -179,14 +180,6 @@ class _AutoDynamicShapeManager:
179
180
  self.generalize_shape_cache.append(compile_args)
180
181
 
181
182
 
182
- def check_tuple_of_scalar(self, input):
183
- """check tuple of scalar"""
184
- for elem in input:
185
- if not isinstance(elem, int):
186
- return False
187
- return True
188
-
189
-
190
183
  def _compare_input_args_and_cache_args(self, input_args, cache_args):
191
184
  """compare input args and cache args"""
192
185
  for (input, cache) in zip(input_args, cache_args):
@@ -233,23 +226,25 @@ class _AutoIdentifyDynamicShape:
233
226
 
234
227
  def _check_input_tensor_type(self, args_list, cache_list):
235
228
  """check input args type"""
236
- for (input, cache) in zip(args_list, cache_list):
237
- if isinstance(input, Tensor) and isinstance(cache, Tensor):
238
- if input.dtype != cache.dtype:
239
- logger.debug((f'input tensor type = {input.dtype}, cache tensor type = {cache.dtype}, '
229
+ for (arg, cache) in zip(args_list, cache_list):
230
+ if isinstance(arg, Tensor) and isinstance(cache, Tensor):
231
+ if arg.dtype != cache.dtype:
232
+ logger.debug((f'input tensor type = {arg.dtype}, cache tensor type = {cache.dtype}, '
240
233
  f'tensor types are not same.'))
241
234
  return False
242
- elif isinstance(input, (tuple, list)) and isinstance(cache, (tuple, list)):
243
- res = self._check_input_tensor_type(input, cache)
235
+ elif isinstance(arg, (tuple, list)) and isinstance(cache, (tuple, list)):
236
+ res = self._check_input_tensor_type(arg, cache)
244
237
  if not res:
245
238
  return False
246
- elif (isinstance(input, int) and isinstance(cache, int)) or \
247
- (isinstance(input, float) and isinstance(cache, float)):
248
- if input != cache:
239
+ elif (isinstance(arg, int) and isinstance(cache, int)) or \
240
+ (isinstance(arg, float) and isinstance(cache, float)):
241
+ if arg != cache:
249
242
  return False
250
- elif (isinstance(input, Tensor) and not isinstance(cache, Tensor)) or \
251
- (isinstance(input, (int, float)) and not isinstance(cache, (int, float))) or \
252
- (isinstance(input, (tuple, list))) and not isinstance(cache, (tuple, list)):
243
+ elif isinstance(arg, Tensor) and not isinstance(cache, Tensor):
244
+ return False
245
+ elif isinstance(arg, (int, float)) and not isinstance(cache, (int, float)):
246
+ return False
247
+ elif isinstance(arg, (tuple, list)) and not isinstance(cache, (tuple, list)):
253
248
  return False
254
249
  return True
255
250
 
@@ -292,21 +287,21 @@ class _AutoIdentifyDynamicShape:
292
287
 
293
288
 
294
289
  @staticmethod
295
- def _do_generalize_in_sink(input, cache, input_index, cache_index, cache_type):
290
+ def _do_generalize_in_sink(arg, cache, input_index, cache_index, cache_type):
296
291
  """do generalize in sink, input rank must be 2"""
297
- if not input:
292
+ if not arg:
298
293
  raise ValueError("In sink mode, cell input can not be scalar.")
299
294
 
300
- if input == cache:
295
+ if arg == cache:
301
296
  return cache
302
297
 
303
298
  shape_value = []
304
- if len(input) != len(cache):
299
+ if len(arg) != len(cache):
305
300
  shape_value.append(SHAPE_RANK_ANY)
306
301
  else:
307
- for _ in input:
302
+ for _ in arg:
308
303
  shape_value.append(SHAPE_DIM_ANY)
309
- logger.info((f'In the {cache_type} cache[{cache_index}], the {input_index}th input tensor shape is {input},'
304
+ logger.info((f'In the {cache_type} cache[{cache_index}], the {input_index}th input tensor shape is {arg},'
310
305
  f'cache shape is {cache}, not equal, need generalize to {shape_value}.'))
311
306
  return shape_value
312
307
 
@@ -318,6 +313,16 @@ class _AutoIdentifyDynamicShape:
318
313
  is_sink_mode, aux)
319
314
 
320
315
 
316
+ def _generate_with_generalize_shape(self, generalize_shape_args, is_sink_mode, args_list):
317
+ """generate with generalize_shape """
318
+ new_generalize_shape, can_generalize = self._do_generalize_shape("generalize", generalize_shape_args,
319
+ is_sink_mode)
320
+ if not can_generalize:
321
+ return args_list
322
+
323
+ res_shape = self.auto_dynamic_shape_manager.get_compile_args_shape(new_generalize_shape, is_sink_mode)
324
+ logger.info((f'generalize with generalize shape cache, compile args shape = {res_shape}'))
325
+ return new_generalize_shape
321
326
 
322
327
  def auto_dynamic_generate_compile_args(self, args_list, is_sink_mode):
323
328
  """generate compile args in auto dynamic shape"""
@@ -331,15 +336,7 @@ class _AutoIdentifyDynamicShape:
331
336
  logger.debug((f'input args list shape = {res_shape}.'))
332
337
 
333
338
  # step1: find cache in real_shape_cache.
334
- real_cache_number = self.auto_dynamic_shape_manager.get_real_shape_cache_number()
335
- if real_cache_number < 2:
336
- logger.info((f'real shape cache cap is {real_cache_number}, smaller than 2, '
337
- f'compile args shape={res_shape}.'))
338
- return args_list
339
-
340
- is_real_shape_exist = self.auto_dynamic_shape_manager.find_compile_args_in_shape_cache(args_list, "real")
341
- if is_real_shape_exist:
342
- logger.debug((f'find compile args in real shape cache, compile args shape={res_shape}'))
339
+ if self._check_real_shape_cache(res_shape, args_list):
343
340
  return args_list
344
341
 
345
342
  # step2: if can not find cache in real_shape_cache, then generate it
@@ -358,14 +355,7 @@ class _AutoIdentifyDynamicShape:
358
355
 
359
356
  # step 4: if can not find cache in generalize_shape_cache, then generate it again
360
357
  if not is_generalize_shape_exist:
361
- new_generalize_shape, can_generalize = self._do_generalize_shape("generalize", generalize_shape_args,
362
- is_sink_mode)
363
- if not can_generalize:
364
- return args_list
365
-
366
- res_shape = self.auto_dynamic_shape_manager.get_compile_args_shape(new_generalize_shape, is_sink_mode)
367
- logger.info((f'generalize with generalize shape cache, compile args shape = {res_shape}'))
368
- return new_generalize_shape
358
+ return self._generate_with_generalize_shape(generalize_shape_args, is_sink_mode, args_list)
369
359
 
370
360
  res_shape = self.auto_dynamic_shape_manager.get_compile_args_shape(generalize_shape_args, is_sink_mode)
371
361
  logger.debug((f'find compile args in generalize shape cache, compile args shape={res_shape}'))
@@ -396,54 +386,61 @@ class _AutoIdentifyDynamicShape:
396
386
 
397
387
  def _do_generalize_one_input_shape(self, input_args, cache_args, cache_type, index, is_sink_mode):
398
388
  """do generalize shape one input by cache"""
389
+ def generalize_tensor(arg, cache, i):
390
+ if self.auto_dynamic_shape_manager.is_tensor_equal(arg, cache):
391
+ return arg
392
+
393
+ shape_value = []
394
+ if len(arg.shape) != len(cache.shape):
395
+ shape_value.append(SHAPE_RANK_ANY)
396
+ else:
397
+ shape_value = [SHAPE_DIM_ANY for _ in range(len(arg.shape))]
398
+ shape_tuple = tuple(shape_value)
399
+ logger.info((f'In the {cache_type} cache[{index}], the {i}th input tensor shape is {arg.shape},'
400
+ f'cache shape is {cache.shape}, not equal, need generalize to {shape_tuple}.'))
401
+ return Tensor(shape=shape_tuple, dtype=arg.dtype)
402
+
403
+ def generalize_sequence(arg, cache, i):
404
+ if is_sink_mode:
405
+ # when is_sink_mode=True, input must be the shape of Tensor.
406
+ res = self._do_generalize_in_sink(arg, cache, i, index, cache_type)
407
+ return res
408
+
409
+ res = self._do_generalize_one_input_shape(arg, cache, cache_type, index, is_sink_mode)
410
+ return res
411
+
399
412
  generalize_one_shape = []
400
- for i, (input, cache) in enumerate(zip(input_args, cache_args)):
401
- if isinstance(input, Parameter) and isinstance(cache, Parameter):
402
- if self.auto_dynamic_shape_manager.is_tensor_equal(input, cache):
403
- generalize_one_shape.append(input)
413
+ for i, (arg, cache) in enumerate(zip(input_args, cache_args)):
414
+ if isinstance(arg, Parameter) and isinstance(cache, Parameter):
415
+ if self.auto_dynamic_shape_manager.is_tensor_equal(arg, cache):
416
+ generalize_one_shape.append(arg)
404
417
  continue
405
- else:
406
- logger.info("In auto dynamic shape mode, parameter must be equal, it can not be generalize.")
407
- return input_args, False
408
418
 
409
- if isinstance(input, Tensor) and isinstance(cache, Tensor):
410
- if self.auto_dynamic_shape_manager.is_tensor_equal(input, cache):
411
- generalize_one_shape.append(input)
412
- else:
413
- shape_value = []
414
- if len(input.shape) != len(cache.shape):
415
- shape_value.append(SHAPE_RANK_ANY)
416
- else:
417
- for _ in range(len(input.shape)):
418
- shape_value.append(SHAPE_DIM_ANY)
419
- shape_tuple = tuple(shape_value)
420
- generalize_one_shape.append(Tensor(shape=shape_tuple, dtype=input.dtype))
421
- logger.info((f'In the {cache_type} cache[{index}], the {i}th input tensor shape is {input.shape},'
422
- f'cache shape is {cache.shape}, not equal, need generalize to {shape_tuple}.'))
419
+ logger.info("In auto dynamic shape mode, parameter must be equal, it can not be generalize.")
420
+ return input_args, False
423
421
 
424
- elif isinstance(input, (tuple, list)) and isinstance(cache, (tuple, list)):
425
- if is_sink_mode:
426
- # when is_sink_mode=True, input must be the shape of Tensor.
427
- res = self._do_generalize_in_sink(input, cache, i, index, cache_type)
428
- generalize_one_shape.append(res)
429
- else:
430
- res = self._do_generalize_one_input_shape(input, cache, cache_type, index, is_sink_mode)
431
- generalize_one_shape.append(res)
432
- elif isinstance(input, int) and isinstance(cache, int):
422
+ if isinstance(arg, Tensor) and isinstance(cache, Tensor):
423
+ res = generalize_tensor(arg, cache, i)
424
+ generalize_one_shape.append(res)
425
+ elif isinstance(arg, (tuple, list)) and isinstance(cache, (tuple, list)):
426
+ res = generalize_sequence(arg, cache, i)
427
+ generalize_one_shape.append(res)
428
+ elif isinstance(arg, int) and isinstance(cache, int):
433
429
  # when is_sink_mode=False, the input must may be scalar, or the value of list/tuple.
434
430
  # is_sink_mode can not be True
435
- if input == cache:
436
- generalize_one_shape.append(input)
431
+ if arg == cache:
432
+ generalize_one_shape.append(arg)
437
433
  else:
438
434
  logger.info("In auto dynamic shape mode, scalar/tuple/list must be equal, it can not be " \
439
435
  "generalize.")
440
436
  return input_args, False
441
- elif input is None and cache is None:
442
- generalize_one_shape.append(input)
437
+ elif arg is None and cache is None:
438
+ generalize_one_shape.append(arg)
443
439
 
444
440
  return generalize_one_shape, True
445
441
 
446
442
 
443
+
447
444
  def _do_generalize_shape(self, cache_type, input_args, is_sink_mode):
448
445
  """do generalize shape by cache"""
449
446
  shape_cache = self.auto_dynamic_shape_manager.get_cache_by_type(cache_type)
@@ -464,8 +461,7 @@ class _AutoIdentifyDynamicShape:
464
461
 
465
462
  keys = list(unknown_shape_dict.keys())
466
463
  keys.sort(key=lambda x: (x[0], x[1]))
467
- index = keys[0]
468
- return unknown_shape_dict.get(index), True
464
+ return unknown_shape_dict.get(keys[0]), True
469
465
 
470
466
  _auto_dynamic_shape = _AutoIdentifyDynamicShape()
471
467
 
mindspore/common/dump.py CHANGED
@@ -57,7 +57,7 @@ def set_dump(target, enabled=True):
57
57
  ``Ascend``
58
58
 
59
59
  Examples:
60
- .. node::
60
+ .. note::
61
61
  Please set environment variable `MINDSPORE_DUMP_CONFIG` to the dump config file and set `dump_mode` field
62
62
  in dump config file to 2 before running this example.
63
63
  See `dump document <https://www.mindspore.cn/tutorials/experts/en/r2.2/debug/dump.html>`_ for details.
@@ -2186,7 +2186,7 @@ class Tensor(Tensor_, metaclass=_TensorMeta):
2186
2186
  [0.7] [3]
2187
2187
  """
2188
2188
  if self.shape == ():
2189
- return (Tensor(0), self)
2189
+ return (self, Tensor(0))
2190
2190
  self._init_check()
2191
2191
  return tensor_operator_registry.get('argmax_with_value')(self, axis, keep_dims)
2192
2192
 
@@ -2234,7 +2234,7 @@ class Tensor(Tensor_, metaclass=_TensorMeta):
2234
2234
  [0.0] [0]
2235
2235
  """
2236
2236
  if self.shape == ():
2237
- return (Tensor(0), self)
2237
+ return (self, Tensor(0))
2238
2238
  self._init_check()
2239
2239
  return tensor_operator_registry.get('argmin_with_value')(self, axis, keep_dims)
2240
2240
 
@@ -2683,7 +2683,6 @@ class Tensor(Tensor_, metaclass=_TensorMeta):
2683
2683
  def _init_check(self):
2684
2684
  if self.has_init:
2685
2685
  self.init_data()
2686
- return self
2687
2686
 
2688
2687
  def init_data(self, slice_index=None, shape=None, opt_shard_group=None):
2689
2688
  """
@@ -4631,23 +4630,7 @@ class Tensor(Tensor_, metaclass=_TensorMeta):
4631
4630
 
4632
4631
  def imag(self):
4633
4632
  r"""
4634
- Returns a new tensor containing imaginary value of the input tensor.
4635
- If input tensor is real, it will return zeros.
4636
-
4637
- Returns:
4638
- Tensor, the shape is the same as the input tensor.
4639
-
4640
- Supported Platforms:
4641
- ``GPU`` ``CPU``
4642
-
4643
- Examples:
4644
- >>> import numpy as np
4645
- >>> import mindspore
4646
- >>> from mindspore import Tensor
4647
- >>> x = Tensor(np.asarray(np.complex(1.3 + 0.4j)), mindspore.complex64)
4648
- >>> output = x.imag()
4649
- >>> print(output)
4650
- 0.4
4633
+ For details, please refer to :func:`mindspore.ops.imag`.
4651
4634
  """
4652
4635
  self._init_check()
4653
4636
  return tensor_operator_registry.get('imag')(self)
@@ -9,7 +9,7 @@
9
9
  {"op_name": "Less", "inputs": [{"index": 0, "name": "x1", "paramType": "required"}, {"index": 1, "name": "x2", "paramType": "required"}], "outputs": [{"index": 0, "name": "y", "paramType": "required"}], "attr": [], "fusion_type": "OPAQUE", "dtype_format": [[["int8", "DefaultFormat"], ["int8", "DefaultFormat"], ["bool", "DefaultFormat"]], [["int16", "DefaultFormat"], ["int16", "DefaultFormat"], ["bool", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["bool", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["bool", "DefaultFormat"]], [["uint8", "DefaultFormat"], ["uint8", "DefaultFormat"], ["bool", "DefaultFormat"]], [["uint16", "DefaultFormat"], ["uint16", "DefaultFormat"], ["bool", "DefaultFormat"]], [["uint32", "DefaultFormat"], ["uint32", "DefaultFormat"], ["bool", "DefaultFormat"]], [["uint64", "DefaultFormat"], ["uint64", "DefaultFormat"], ["bool", "DefaultFormat"]], [["float16", "DefaultFormat"], ["float16", "DefaultFormat"], ["bool", "DefaultFormat"]], [["float32", "DefaultFormat"], ["float32", "DefaultFormat"], ["bool", "DefaultFormat"]], [["float64", "DefaultFormat"], ["float64", "DefaultFormat"], ["bool", "DefaultFormat"]]], "imply_type": "AiCPU"}
10
10
  {"op_name": "Lstsq", "inputs": [{"index": 0, "name": "matrix", "paramType": "required"}, {"index": 1, "name": "rhs", "paramType": "required"}], "outputs": [{"index": 0, "name": "y", "paramType": "required"}], "attr": [{"name": "l2_regularizer", "type": "float", "value": "0.0"}, {"name": "fast", "type": "bool", "value": "True"}], "fusion_type": "OPAQUE", "dtype_format": [[["float16", "DefaultFormat"], ["float16", "DefaultFormat"], ["float16", "DefaultFormat"]], [["float32", "DefaultFormat"], ["float32", "DefaultFormat"], ["float32", "DefaultFormat"]], [["float64", "DefaultFormat"], ["float64", "DefaultFormat"], ["float64", "DefaultFormat"]]], "imply_type": "AiCPU"}
11
11
  {"op_name": "LeftShift", "inputs": [{"index": 0, "name": "x1", "paramType": "required"}, {"index": 1, "name": "x2", "paramType": "required"}], "outputs": [{"index": 0, "name": "y", "paramType": "required"}], "attr": [], "fusion_type": "OPAQUE", "dtype_format": [[["int8", "DefaultFormat"], ["int8", "DefaultFormat"], ["int8", "DefaultFormat"]], [["int16", "DefaultFormat"], ["int16", "DefaultFormat"], ["int16", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"]], [["uint8", "DefaultFormat"], ["uint8", "DefaultFormat"], ["uint8", "DefaultFormat"]], [["uint16", "DefaultFormat"], ["uint16", "DefaultFormat"], ["uint16", "DefaultFormat"]], [["uint32", "DefaultFormat"], ["uint32", "DefaultFormat"], ["uint32", "DefaultFormat"]], [["uint64", "DefaultFormat"], ["uint64", "DefaultFormat"], ["uint64", "DefaultFormat"]]], "imply_type": "AiCPU"}
12
- {"op_name": "Add", "inputs": [{"index": 0, "name": "x", "paramType": "required"}, {"index": 1, "name": "y", "paramType": "required"}], "outputs": [{"index": 0, "name": "output", "paramType": "required"}], "attr": [], "fusion_type": "OPAQUE", "dtype_format": [[["float16", "DefaultFormat"], ["float16", "DefaultFormat"], ["float16", "DefaultFormat"]], [["float32", "DefaultFormat"], ["float32", "DefaultFormat"], ["float32", "DefaultFormat"]], [["float64", "DefaultFormat"], ["float64", "DefaultFormat"], ["float64", "DefaultFormat"]], [["int8", "DefaultFormat"], ["int8", "DefaultFormat"], ["int8", "DefaultFormat"]], [["int16", "DefaultFormat"], ["int16", "DefaultFormat"], ["int16", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"]], [["uint8", "DefaultFormat"], ["uint8", "DefaultFormat"], ["uint8", "DefaultFormat"]], [["uint16", "DefaultFormat"], ["int16", "DefaultFormat"], ["int16", "DefaultFormat"]], [["uint32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"]], [["uint64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"]], [["complex64", "DefaultFormat"], ["complex64", "DefaultFormat"], ["complex64", "DefaultFormat"]], [["complex128", "DefaultFormat"], ["complex128", "DefaultFormat"], ["complex128", "DefaultFormat"]]], "imply_type": "AiCPU"}
12
+ {"op_name": "Add", "inputs": [{"index": 0, "name": "x", "paramType": "required"}, {"index": 1, "name": "y", "paramType": "required"}], "outputs": [{"index": 0, "name": "output", "paramType": "required"}], "attr": [], "fusion_type": "OPAQUE", "dtype_format": [[["float16", "DefaultFormat"], ["float16", "DefaultFormat"], ["float16", "DefaultFormat"]], [["float32", "DefaultFormat"], ["float32", "DefaultFormat"], ["float32", "DefaultFormat"]], [["float64", "DefaultFormat"], ["float64", "DefaultFormat"], ["float64", "DefaultFormat"]], [["int8", "DefaultFormat"], ["int8", "DefaultFormat"], ["int8", "DefaultFormat"]], [["int16", "DefaultFormat"], ["int16", "DefaultFormat"], ["int16", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"]], [["uint8", "DefaultFormat"], ["uint8", "DefaultFormat"], ["uint8", "DefaultFormat"]], [["uint16", "DefaultFormat"], ["uint16", "DefaultFormat"], ["uint16", "DefaultFormat"]], [["uint32", "DefaultFormat"], ["uint32", "DefaultFormat"], ["uint32", "DefaultFormat"]], [["uint64", "DefaultFormat"], ["uint64", "DefaultFormat"], ["uint64", "DefaultFormat"]], [["complex64", "DefaultFormat"], ["complex64", "DefaultFormat"], ["complex64", "DefaultFormat"]], [["complex128", "DefaultFormat"], ["complex128", "DefaultFormat"], ["complex128", "DefaultFormat"]]], "imply_type": "AiCPU"}
13
13
  {"op_name": "SparseMatrixTranspose", "inputs": [{"index": 0, "name": "x_dense_shape", "paramType": "required"}, {"index": 1, "name": "x_batch_pointers", "paramType": "required"}, {"index": 2, "name": "x_row_pointers", "paramType": "required"}, {"index": 3, "name": "x_col_indices", "paramType": "required"}, {"index": 4, "name": "x_values", "paramType": "required"}], "outputs": [{"index": 0, "name": "y_dense_shape", "paramType": "required"}, {"index": 1, "name": "y_batch_pointers", "paramType": "required"}, {"index": 2, "name": "y_row_pointers", "paramType": "required"}, {"index": 3, "name": "y_col_indices", "paramType": "required"}, {"index": 4, "name": "y_values", "paramType": "required"}], "attr": [{"name": "conjugate", "type": "bool"}], "fusion_type": "OPAQUE", "dtype_format": [[["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int8", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int8", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["uint8", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["uint8", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int16", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int16", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["uint16", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["uint16", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["uint32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["uint32", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int64", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int64", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["uint64", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["uint64", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["float16", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["float16", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["float32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["float32", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["float64", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["float64", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["complex64", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["complex64", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["complex128", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["complex128", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int8", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int8", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["uint8", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["uint8", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int16", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int16", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["uint16", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["uint16", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int32", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["uint32", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["uint32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["uint64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["uint64", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["float16", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["float16", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["float32", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["float32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["float64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["float64", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["complex64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["complex64", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["complex128", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["complex128", "DefaultFormat"]]], "imply_type": "AiCPU"}
14
14
  {"op_name": "SparseMatrixNNZ", "inputs": [{"index": 0, "name": "x_dense_shape", "paramType": "required"}, {"index": 1, "name": "x_batch_pointers", "paramType": "required"}, {"index": 2, "name": "x_row_pointers", "paramType": "required"}, {"index": 3, "name": "x_col_indices", "paramType": "required"}, {"index": 4, "name": "x_values", "paramType": "required"}], "outputs": [{"index": 0, "name": "y", "paramType": "required"}], "attr": [], "fusion_type": "OPAQUE", "dtype_format": [[["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int8", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["uint8", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int16", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["uint16", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int64", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["float16", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["float32", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["float64", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["bool", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["complex64", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"], ["complex128", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int8", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["uint8", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int16", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["uint16", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["float16", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["float32", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["float64", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["bool", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["complex64", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["complex128", "DefaultFormat"], ["int32", "DefaultFormat"]]], "imply_type": "AiCPU"}
15
15
  {"op_name": "SparseDenseCwiseMul", "inputs": [{"index": 0, "name": "x1_indices", "paramType": "required"}, {"index": 1, "name": "x1_values", "paramType": "required"}, {"index": 2, "name": "x1_sparse", "paramType": "required"}, {"index": 3, "name": "x2", "paramType": "required"}], "outputs": [{"index": 0, "name": "y", "paramType": "required"}], "attr": [], "fusion_type": "OPAQUE", "dtype_format": [[["int64", "DefaultFormat"], ["int8", "DefaultFormat"], ["int64", "DefaultFormat"], ["int8", "DefaultFormat"], ["int8", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int16", "DefaultFormat"], ["int64", "DefaultFormat"], ["int16", "DefaultFormat"], ["int16", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int32", "DefaultFormat"], ["int64", "DefaultFormat"], ["int32", "DefaultFormat"], ["int32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"], ["int64", "DefaultFormat"]], [["int64", "DefaultFormat"], ["uint8", "DefaultFormat"], ["int64", "DefaultFormat"], ["uint8", "DefaultFormat"], ["uint8", "DefaultFormat"]], [["int64", "DefaultFormat"], ["uint16", "DefaultFormat"], ["int64", "DefaultFormat"], ["uint16", "DefaultFormat"], ["uint16", "DefaultFormat"]], [["int64", "DefaultFormat"], ["uint32", "DefaultFormat"], ["int64", "DefaultFormat"], ["uint32", "DefaultFormat"], ["uint32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["uint64", "DefaultFormat"], ["int64", "DefaultFormat"], ["uint64", "DefaultFormat"], ["uint64", "DefaultFormat"]], [["int64", "DefaultFormat"], ["float16", "DefaultFormat"], ["int64", "DefaultFormat"], ["float16", "DefaultFormat"], ["float16", "DefaultFormat"]], [["int64", "DefaultFormat"], ["float32", "DefaultFormat"], ["int64", "DefaultFormat"], ["float32", "DefaultFormat"], ["float32", "DefaultFormat"]], [["int64", "DefaultFormat"], ["float64", "DefaultFormat"], ["int64", "DefaultFormat"], ["float64", "DefaultFormat"], ["float64", "DefaultFormat"]], [["int64", "DefaultFormat"], ["complex64", "DefaultFormat"], ["int64", "DefaultFormat"], ["complex64", "DefaultFormat"], ["complex64", "DefaultFormat"]], [["int64", "DefaultFormat"], ["complex128", "DefaultFormat"], ["int64", "DefaultFormat"], ["complex128", "DefaultFormat"], ["complex128", "DefaultFormat"]]], "imply_type": "AiCPU"}
mindspore/context.py CHANGED
@@ -268,6 +268,8 @@ class _Context:
268
268
  "allow_mix_precision_fp16" and "allow_mix_precision_bf16".
269
269
  - jit_compile (bool): ``False`` and ``True``.
270
270
  - atomic_clean_policy (int): ``0`` and ``1``. Default: ``1`` .
271
+ - exception_dump (str): Enable exception dump for Ascend operators. ``"0"`` , ``"1"`` and ``"2"``.
272
+ Default: ``"2"`` .
271
273
  - op_precision_mode (str): config file path.
272
274
  - parallel_speed_up_json_path(Union[str, None]): The path to the parallel speed up json file.
273
275
  If its value is None or '', it does not take effect. Default None.
@@ -280,6 +282,7 @@ class _Context:
280
282
  'atomic_clean_policy': [0, 1],
281
283
  'matmul_allow_hf32': [True, False],
282
284
  'conv_allow_hf32': [True, False],
285
+ 'exception_dump': ["0", "1", "2"],
283
286
  'op_precision_mode': (str,),
284
287
  'parallel_speed_up_json_path': (str, None)
285
288
  }
@@ -289,6 +292,7 @@ class _Context:
289
292
  'atomic_clean_policy': self._get_ascend_config_setter('atomic_clean_policy', str),
290
293
  'matmul_allow_hf32': self._get_ascend_config_setter('matmul_allow_hf32', lambda v: "1" if v else "0"),
291
294
  'conv_allow_hf32': self._get_ascend_config_setter('conv_allow_hf32', lambda v: "1" if v else "0"),
295
+ 'exception_dump': self._get_ascend_config_setter('exception_dump'),
292
296
  'op_precision_mode': self._set_op_precision_mode,
293
297
  'parallel_speed_up_json_path': self._set_speedup_config_path
294
298
  }
@@ -629,6 +633,7 @@ class _Context:
629
633
  "enable_task_opt": ms_ctx_param.enable_task_opt,
630
634
  "enable_grad_comm_opt": ms_ctx_param.enable_grad_comm_opt,
631
635
  "interleaved_matmul_comm": ms_ctx_param.interleaved_matmul_comm,
636
+ "enable_opt_shard_comm_opt": ms_ctx_param.enable_opt_shard_comm_opt,
632
637
  "interleaved_layernorm_comm": ms_ctx_param.interleaved_layernorm_comm}
633
638
  with open(speedup_config_real_path, 'r') as f:
634
639
  speedup_config = json.load(f)
@@ -1277,12 +1282,10 @@ def set_context(**kwargs):
1277
1282
  memory_optimize_level is set 'O1'.
1278
1283
  - OFF: Turn off the memory Offload function.
1279
1284
  ascend_config (dict): Set the parameters specific to Ascend hardware platform. It is not set by default.
1280
- Currently, configurations except `parallel_speed_up_json_path` and `precision_mode.force_fp32` are only
1281
- supported on Ascend910B hardware platform. The default value of `precision_mode`, `jit_compile` and
1285
+ The default value of `precision_mode`, `jit_compile` and
1282
1286
  `atomic_clean_policy` are experimental parameters, may change in the future.
1283
1287
 
1284
- - precision_mode (str): Mixed precision mode setting, on Ascend910B hardware platform, the default
1285
- value of training network is based on the value of CANN, and the default value of inference network
1288
+ - precision_mode (str): Mixed precision mode setting, and the default value of inference network
1286
1289
  is ``force_fp16`` . The value range is as follows:
1287
1290
 
1288
1291
  - force_fp16: When the operator supports both float16 and float32, select float16 directly.
@@ -1315,6 +1318,10 @@ def set_context(**kwargs):
1315
1318
  - conv_allow_hf32 (bool): Whether to convert FP32 to HF32 for Conv operators. Default value: ``True``.
1316
1319
  This is an experimental prototype that is subject to change and/or deletion.
1317
1320
  For detailed information, please refer to `Ascend community <https://www.hiascend.com/>`_ .
1321
+ - exception_dump (str): Enable exception dump for Ascend operators, providing the input and output data for
1322
+ failing Ascend operators. The value can be ``"0"`` , ``"1"`` and ``"2"``. For ``"0"`` , exception dump is
1323
+ turned off; for ``"1"``, all inputs and outputs will be dumped for AICore and AICPU exception operators;
1324
+ for ``"2"``, inputs will be dumped for AICore exception operators. Default: ``"2"`` .
1318
1325
  - op_precision_mode (str): Path to config file of op precision mode. For detailed information, please refer
1319
1326
  to `Ascend community <https://www.hiascend.com/>`_ .
1320
1327
  - parallel_speed_up_json_path(Union[str, None]): The path to the parallel speed up json file, configuration
@@ -232,6 +232,9 @@ class TFRecordDataset(SourceDataset, UnionBaseDataset):
232
232
 
233
233
  The columns of generated dataset depend on the source TFRecord files.
234
234
 
235
+ Note:
236
+ 'TFRecordDataset' is not support on Windows platform yet.
237
+
235
238
  Args:
236
239
  dataset_files (Union[str, list[str]]): String or list of files to be read or glob strings to search for a
237
240
  pattern of files. The list will be sorted in lexicographical order.
@@ -318,6 +321,8 @@ class TFRecordDataset(SourceDataset, UnionBaseDataset):
318
321
  cache=None, compression_type=None):
319
322
  super().__init__(num_parallel_workers=num_parallel_workers, num_samples=num_samples, shuffle=shuffle,
320
323
  num_shards=num_shards, shard_id=shard_id, cache=cache)
324
+ if platform.system().lower() == "windows":
325
+ raise NotImplementedError("TFRecordDataset is not supported for windows.")
321
326
  self.dataset_files = self._find_files(dataset_files)
322
327
  self.dataset_files.sort()
323
328
 
@@ -144,14 +144,14 @@ class AdjustBrightness(ImageTensorOperation, PyTensorOperation):
144
144
 
145
145
  Args:
146
146
  device_target (str, optional): The operator will be executed on this device. Currently supports
147
- ``CPU`` and ``Ascend`` , where ``Ascend`` refers to Ascend910B device. Default: ``CPU`` .
147
+ ``CPU`` . Default: ``CPU`` .
148
148
 
149
149
  Raises:
150
150
  TypeError: If `device_target` is not of type str.
151
- ValueError: If `device_target` is not within the valid set of ['CPU', 'Ascend'].
151
+ ValueError: If `device_target` is not ``CPU`` .
152
152
 
153
153
  Supported Platforms:
154
- ``CPU`` ``Ascend``
154
+ ``CPU``
155
155
 
156
156
  Examples:
157
157
  >>> import mindspore.dataset as ds
@@ -227,14 +227,14 @@ class AdjustContrast(ImageTensorOperation, PyTensorOperation):
227
227
 
228
228
  Args:
229
229
  device_target (str, optional): The operator will be executed on this device. Currently supports
230
- ``CPU`` and ``Ascend`` , where ``Ascend`` refers to Ascend910B device. Default: ``CPU`` .
230
+ ``CPU`` . Default: ``CPU`` .
231
231
 
232
232
  Raises:
233
233
  TypeError: If `device_target` is not of type str.
234
- ValueError: If `device_target` is not within the valid set of ['CPU', 'Ascend'].
234
+ ValueError: If `device_target` is not ``CPU`` .
235
235
 
236
236
  Supported Platforms:
237
- ``CPU`` ``Ascend``
237
+ ``CPU``
238
238
 
239
239
  Examples:
240
240
  >>> import mindspore.dataset as ds
@@ -373,14 +373,14 @@ class AdjustHue(ImageTensorOperation, PyTensorOperation):
373
373
 
374
374
  Args:
375
375
  device_target (str, optional): The operator will be executed on this device. Currently supports
376
- ``CPU`` and ``Ascend`` , where ``Ascend`` refers to Ascend910B device. Default: ``CPU`` .
376
+ ``CPU`` . Default: ``CPU`` .
377
377
 
378
378
  Raises:
379
379
  TypeError: If `device_target` is not of type str.
380
- ValueError: If `device_target` is not within the valid set of ['CPU', 'Ascend'].
380
+ ValueError: If `device_target` is not ``CPU`` .
381
381
 
382
382
  Supported Platforms:
383
- ``CPU`` ``Ascend``
383
+ ``CPU``
384
384
 
385
385
  Examples:
386
386
  >>> import mindspore.dataset as ds
@@ -457,14 +457,14 @@ class AdjustSaturation(ImageTensorOperation, PyTensorOperation):
457
457
 
458
458
  Args:
459
459
  device_target (str, optional): The operator will be executed on this device. Currently supports
460
- ``CPU`` and ``Ascend`` , where ``Ascend`` refers to Ascend910B device. Default: ``CPU`` .
460
+ ``CPU`` . Default: ``CPU`` .
461
461
 
462
462
  Raises:
463
463
  TypeError: If `device_target` is not of type str.
464
- ValueError: If `device_target` is not within the valid set of ['CPU', 'Ascend'].
464
+ ValueError: If `device_target` is not ``CPU`` .
465
465
 
466
466
  Supported Platforms:
467
- ``CPU`` ``Ascend``
467
+ ``CPU``
468
468
 
469
469
  Examples:
470
470
  >>> import mindspore.dataset as ds
@@ -1159,14 +1159,14 @@ class Decode(ImageTensorOperation, PyTensorOperation):
1159
1159
 
1160
1160
  Args:
1161
1161
  device_target (str, optional): The operator will be executed on this device. Currently supports
1162
- ``CPU`` and ``Ascend`` , where ``Ascend`` refers to Ascend910B device. Default: ``CPU`` .
1162
+ ``CPU`` . Default: ``CPU`` .
1163
1163
 
1164
1164
  Raises:
1165
1165
  TypeError: If `device_target` is not of type str.
1166
- ValueError: If `device_target` is not within the valid set of ['CPU', 'Ascend'].
1166
+ ValueError: If `device_target` is not ``CPU`` .
1167
1167
 
1168
1168
  Supported Platforms:
1169
- ``CPU`` ``Ascend``
1169
+ ``CPU``
1170
1170
 
1171
1171
  Examples:
1172
1172
  >>> import mindspore.dataset as ds
@@ -1908,14 +1908,14 @@ class Normalize(ImageTensorOperation):
1908
1908
 
1909
1909
  Args:
1910
1910
  device_target (str, optional): The operator will be executed on this device. Currently supports
1911
- ``CPU`` and ``Ascend`` , where ``Ascend`` refers to Ascend910B device. Default: ``CPU`` .
1911
+ ``CPU`` . Default: ``CPU`` .
1912
1912
 
1913
1913
  Raises:
1914
1914
  TypeError: If `device_target` is not of type str.
1915
- ValueError: If `device_target` is not within the valid set of ['CPU', 'Ascend'].
1915
+ ValueError: If `device_target` is not ``CPU`` .
1916
1916
 
1917
1917
  Supported Platforms:
1918
- ``CPU`` ``Ascend``
1918
+ ``CPU``
1919
1919
 
1920
1920
  Examples:
1921
1921
  >>> import mindspore.dataset as ds
@@ -4182,14 +4182,14 @@ class Resize(ImageTensorOperation, PyTensorOperation):
4182
4182
 
4183
4183
  Args:
4184
4184
  device_target (str, optional): The operator will be executed on this device. Currently supports
4185
- ``CPU`` and ``Ascend`` , where ``Ascend`` refers to Ascend910B device. Default: ``CPU`` .
4185
+ ``CPU`` . Default: ``CPU`` .
4186
4186
 
4187
4187
  Raises:
4188
4188
  TypeError: If `device_target` is not of type str.
4189
- ValueError: If `device_target` is not within the valid set of ['CPU', 'Ascend'].
4189
+ ValueError: If `device_target` is not ``CPU`` .
4190
4190
 
4191
4191
  Supported Platforms:
4192
- ``CPU`` ``Ascend``
4192
+ ``CPU``
4193
4193
 
4194
4194
  Examples:
4195
4195
  >>> import mindspore.dataset as ds
mindspore/dnnl.dll CHANGED
Binary file
mindspore/dpcmi.dll CHANGED
Binary file
@@ -43,7 +43,7 @@ def _run_adam_with_amsgrad_opt(opt, beta1_power, beta2_power, lr, gradient, para
43
43
 
44
44
  class Adam(Optimizer):
45
45
  r"""
46
- Implements Adam algorithm..
46
+ Implements Adam algorithm.
47
47
 
48
48
  The updating formulas are as follows:
49
49
 
mindspore/gen_ops.py CHANGED
@@ -120,7 +120,7 @@ def generate_py_primitive(yaml_data):
120
120
  assign_str += arg_name
121
121
  args_assign.append(assign_str)
122
122
 
123
- args_assign = '\n'.join(assign for assign in args_assign)
123
+ args_assign = '\n'.join([assign for assign in args_assign])
124
124
  primitive_code = f"""
125
125
  class {class_name}(Primitive):
126
126
  def __init__(self, {', '.join(init_args_with_default)}):