xmos-ai-tools 1.3.2.dev19__py3-none-macosx_11_0_arm64.whl → 1.3.2.dev37__py3-none-macosx_11_0_arm64.whl

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (206) hide show
  1. xmos_ai_tools/runtime/include/lib_tflite_micro/api/inference_engine.h +9 -9
  2. xmos_ai_tools/runtime/include/lib_tflite_micro/src/tflite-xcore-kernels/xcore_common.h +2 -2
  3. xmos_ai_tools/runtime/include/lib_tflite_micro/src/tflite-xcore-kernels/xcore_custom_options.h +2 -2
  4. xmos_ai_tools/runtime/include/lib_tflite_micro/src/tflite-xcore-kernels/xcore_error_reporter.h +3 -3
  5. xmos_ai_tools/runtime/include/lib_tflite_micro/src/tflite-xcore-kernels/xcore_interpreter.h +8 -8
  6. xmos_ai_tools/runtime/include/lib_tflite_micro/src/tflite-xcore-kernels/xcore_ops.h +3 -3
  7. xmos_ai_tools/runtime/include/lib_tflite_micro/src/tflite-xcore-kernels/xcore_profiler.h +4 -4
  8. xmos_ai_tools/runtime/include/lib_tflite_micro/src/tflite-xcore-kernels/xcore_utils.h +5 -5
  9. xmos_ai_tools/runtime/include/lib_xud/lib_xud/api/xud.h +3 -3
  10. xmos_ai_tools/runtime/include/tensorflow/lite/array.h +4 -4
  11. xmos_ai_tools/runtime/include/tensorflow/lite/context_util.h +2 -2
  12. xmos_ai_tools/runtime/include/tensorflow/lite/core/api/error_reporter.h +3 -3
  13. xmos_ai_tools/runtime/include/tensorflow/lite/core/api/flatbuffer_conversions.h +2 -2
  14. xmos_ai_tools/runtime/include/tensorflow/lite/core/api/tensor_utils.h +2 -2
  15. xmos_ai_tools/runtime/include/tensorflow/lite/core/c/c_api_types.h +3 -3
  16. xmos_ai_tools/runtime/include/tensorflow/lite/core/c/common.h +17 -17
  17. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/common.h +2 -2
  18. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/cppmath.h +2 -2
  19. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/max.h +2 -2
  20. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/min.h +2 -2
  21. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/portable_tensor.h +2 -2
  22. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/portable_tensor_utils.h +2 -2
  23. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/quantization_util.h +2 -2
  24. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/add.h +3 -3
  25. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/add_n.h +2 -2
  26. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/arg_min_max.h +2 -2
  27. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/batch_matmul.h +2 -2
  28. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/batch_to_space_nd.h +2 -2
  29. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/binary_function.h +2 -2
  30. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/broadcast_args.h +2 -2
  31. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/broadcast_to.h +2 -2
  32. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/ceil.h +2 -2
  33. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/comparisons.h +2 -2
  34. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/concatenation.h +3 -3
  35. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/conv.h +2 -2
  36. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/cumsum.h +2 -2
  37. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/depth_to_space.h +3 -3
  38. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/depthwiseconv_float.h +1 -1
  39. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/depthwiseconv_uint8.h +1 -1
  40. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/dequantize.h +4 -4
  41. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/div.h +2 -2
  42. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/elu.h +2 -2
  43. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/exp.h +2 -2
  44. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/fill.h +2 -2
  45. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/floor.h +2 -2
  46. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/floor_div.h +2 -2
  47. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/floor_mod.h +2 -2
  48. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/fully_connected.h +2 -2
  49. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/hard_swish.h +2 -2
  50. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/integer_ops/add.h +2 -2
  51. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/integer_ops/conv.h +2 -2
  52. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/integer_ops/depthwise_conv.h +2 -2
  53. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/integer_ops/fully_connected.h +2 -2
  54. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/integer_ops/l2normalization.h +2 -2
  55. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/integer_ops/logistic.h +2 -2
  56. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/integer_ops/mul.h +2 -2
  57. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/integer_ops/pooling.h +2 -2
  58. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/integer_ops/tanh.h +2 -2
  59. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/integer_ops/transpose_conv.h +2 -2
  60. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/l2normalization.h +4 -4
  61. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/leaky_relu.h +3 -3
  62. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/log_softmax.h +2 -2
  63. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/logistic.h +2 -2
  64. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/lstm_cell.h +5 -5
  65. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/maximum_minimum.h +2 -2
  66. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/mul.h +2 -2
  67. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/neg.h +2 -2
  68. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/pad.h +8 -8
  69. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/pooling.h +2 -2
  70. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/portable_tensor_utils.h +2 -2
  71. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/portable_tensor_utils_impl.h +2 -2
  72. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/prelu.h +2 -2
  73. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/process_broadcast_shapes.h +3 -3
  74. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/quantize.h +4 -4
  75. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/reduce.h +3 -3
  76. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/requantize.h +2 -2
  77. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/resize_bilinear.h +4 -4
  78. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/resize_nearest_neighbor.h +3 -3
  79. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/round.h +2 -2
  80. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/select.h +2 -2
  81. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/slice.h +5 -5
  82. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/softmax.h +2 -2
  83. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/space_to_batch_nd.h +2 -2
  84. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/space_to_depth.h +3 -3
  85. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/strided_slice.h +6 -6
  86. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/sub.h +2 -2
  87. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/tanh.h +2 -2
  88. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/transpose.h +2 -2
  89. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/transpose_conv.h +2 -2
  90. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/runtime_shape.h +2 -2
  91. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/strided_slice_logic.h +9 -9
  92. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/tensor_ctypes.h +2 -2
  93. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/types.h +2 -2
  94. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/kernel_util.h +2 -2
  95. xmos_ai_tools/runtime/include/tensorflow/lite/kernels/padding.h +2 -2
  96. xmos_ai_tools/runtime/include/tensorflow/lite/micro/arena_allocator/ibuffer_allocator.h +2 -2
  97. xmos_ai_tools/runtime/include/tensorflow/lite/micro/arena_allocator/non_persistent_arena_buffer_allocator.h +2 -2
  98. xmos_ai_tools/runtime/include/tensorflow/lite/micro/arena_allocator/persistent_arena_buffer_allocator.h +2 -2
  99. xmos_ai_tools/runtime/include/tensorflow/lite/micro/arena_allocator/recording_single_arena_buffer_allocator.h +2 -2
  100. xmos_ai_tools/runtime/include/tensorflow/lite/micro/arena_allocator/single_arena_buffer_allocator.h +2 -2
  101. xmos_ai_tools/runtime/include/tensorflow/lite/micro/benchmarks/micro_benchmark.h +7 -7
  102. xmos_ai_tools/runtime/include/tensorflow/lite/micro/fake_micro_context.h +2 -2
  103. xmos_ai_tools/runtime/include/tensorflow/lite/micro/flatbuffer_utils.h +2 -2
  104. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/activation_utils.h +2 -2
  105. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/activations.h +2 -2
  106. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/add.h +2 -2
  107. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/arc_mli/mli_function_specializations.h +2 -2
  108. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/arc_mli/mli_interface.h +2 -2
  109. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/arc_mli/mli_slicers.h +2 -2
  110. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/arc_mli/mli_tf_utils.h +4 -4
  111. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/arc_mli/scratch_buf_mgr.h +2 -2
  112. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/arc_mli/scratch_buffers.h +2 -2
  113. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/ceva/types.h +6 -6
  114. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/circular_buffer.h +2 -2
  115. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/conv.h +2 -2
  116. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/conv_test.h +2 -2
  117. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/depthwise_conv.h +2 -2
  118. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/dequantize.h +3 -3
  119. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/ethosu.h +2 -2
  120. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/fully_connected.h +2 -2
  121. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/hard_swish.h +2 -2
  122. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/kernel_runner.h +2 -2
  123. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/kernel_util.h +2 -2
  124. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/leaky_relu.h +2 -2
  125. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/logical.h +2 -2
  126. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/logistic.h +2 -2
  127. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/lstm_eval.h +48 -48
  128. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/lstm_eval_test.h +57 -57
  129. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/lstm_shared.h +2 -2
  130. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/micro_ops.h +2 -2
  131. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/micro_tensor_utils.h +2 -2
  132. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/mul.h +2 -2
  133. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/pad.h +2 -2
  134. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/pooling.h +15 -15
  135. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/prelu.h +2 -2
  136. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/quantize.h +3 -3
  137. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/reduce.h +2 -2
  138. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/reshape.h +2 -2
  139. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/softmax.h +2 -2
  140. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/strided_slice.h +2 -2
  141. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/sub.h +2 -2
  142. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/svdf.h +2 -2
  143. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/testdata/conv_test_data.h +2 -2
  144. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/testdata/lstm_test_data.h +7 -7
  145. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/unidirectional_sequence_lstm.h +2 -2
  146. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/xtensa/hifimini/fixedpoint_utils.h +2 -2
  147. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/xtensa/lstm_eval.h +2 -2
  148. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/xtensa/lstm_shared.h +2 -2
  149. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/xtensa/xtensa_add.h +2 -2
  150. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/xtensa/xtensa_conv.h +2 -2
  151. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/xtensa/xtensa_depthwise_conv.h +2 -2
  152. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/xtensa/xtensa_fully_connected.h +2 -2
  153. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/xtensa/xtensa_pad.h +2 -2
  154. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/xtensa/xtensa_pooling.h +2 -2
  155. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/xtensa/xtensa_reduce.h +2 -2
  156. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/xtensa/xtensa_reshape.h +2 -2
  157. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/xtensa/xtensa_softmax.h +2 -2
  158. xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/xtensa/xtensa_svdf.h +2 -2
  159. xmos_ai_tools/runtime/include/tensorflow/lite/micro/memory_helpers.h +3 -3
  160. xmos_ai_tools/runtime/include/tensorflow/lite/micro/memory_planner/greedy_memory_planner.h +2 -2
  161. xmos_ai_tools/runtime/include/tensorflow/lite/micro/memory_planner/linear_memory_planner.h +2 -2
  162. xmos_ai_tools/runtime/include/tensorflow/lite/micro/memory_planner/memory_plan_struct.h +2 -2
  163. xmos_ai_tools/runtime/include/tensorflow/lite/micro/memory_planner/micro_memory_planner.h +2 -2
  164. xmos_ai_tools/runtime/include/tensorflow/lite/micro/memory_planner/non_persistent_buffer_planner_shim.h +2 -2
  165. xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_allocation_info.h +3 -3
  166. xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_allocator.h +3 -3
  167. xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_arena_constants.h +2 -2
  168. xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_context.h +3 -3
  169. xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_graph.h +4 -4
  170. xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_interpreter.h +2 -2
  171. xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_interpreter_context.h +2 -2
  172. xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_interpreter_graph.h +4 -4
  173. xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_log.h +4 -4
  174. xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_mutable_op_resolver.h +50 -50
  175. xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_op_resolver.h +2 -2
  176. xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_profiler.h +2 -2
  177. xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_profiler_interface.h +2 -2
  178. xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_resource_variable.h +2 -2
  179. xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_time.h +2 -2
  180. xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_utils.h +2 -2
  181. xmos_ai_tools/runtime/include/tensorflow/lite/micro/mock_micro_graph.h +2 -2
  182. xmos_ai_tools/runtime/include/tensorflow/lite/micro/python/tflite_size/src/flatbuffer_size.h +2 -2
  183. xmos_ai_tools/runtime/include/tensorflow/lite/micro/python/tflite_size/src/flatbuffer_size_wrapper.h +2 -2
  184. xmos_ai_tools/runtime/include/tensorflow/lite/micro/recording_micro_allocator.h +2 -2
  185. xmos_ai_tools/runtime/include/tensorflow/lite/micro/recording_micro_interpreter.h +2 -2
  186. xmos_ai_tools/runtime/include/tensorflow/lite/micro/system_setup.h +2 -2
  187. xmos_ai_tools/runtime/include/tensorflow/lite/micro/test_helper_custom_ops.h +2 -2
  188. xmos_ai_tools/runtime/include/tensorflow/lite/micro/test_helpers.h +6 -6
  189. xmos_ai_tools/runtime/include/tensorflow/lite/micro/testing/micro_test.h +4 -4
  190. xmos_ai_tools/runtime/include/tensorflow/lite/micro/tflite_bridge/flatbuffer_conversions_bridge.h +2 -2
  191. xmos_ai_tools/runtime/include/tensorflow/lite/micro/tflite_bridge/micro_error_reporter.h +2 -2
  192. xmos_ai_tools/runtime/include/tensorflow/lite/micro/tools/benchmarking/log_utils.h +2 -2
  193. xmos_ai_tools/runtime/include/tensorflow/lite/micro/tools/benchmarking/metrics.h +3 -3
  194. xmos_ai_tools/runtime/include/tensorflow/lite/micro/tools/benchmarking/op_resolver.h +3 -3
  195. xmos_ai_tools/runtime/include/tensorflow/lite/portable_type_to_tflitetype.h +2 -2
  196. xmos_ai_tools/runtime/include/tensorflow/lite/schema/schema_generated.h +2731 -2731
  197. xmos_ai_tools/runtime/include/tensorflow/lite/schema/schema_utils.h +2 -2
  198. xmos_ai_tools/runtime/lib/libhost_xtflitemicro.a +0 -0
  199. xmos_ai_tools/runtime/lib/libxtflitemicro.a +0 -0
  200. xmos_ai_tools/xinterpreters/libs/macos/xtflm_python.1.0.1.dylib +0 -0
  201. xmos_ai_tools/xinterpreters/libs/macos/xtflm_python.dylib +0 -0
  202. {xmos_ai_tools-1.3.2.dev19.data → xmos_ai_tools-1.3.2.dev37.data}/data/bin/xcore-opt +0 -0
  203. {xmos_ai_tools-1.3.2.dev19.dist-info → xmos_ai_tools-1.3.2.dev37.dist-info}/METADATA +2 -2
  204. {xmos_ai_tools-1.3.2.dev19.dist-info → xmos_ai_tools-1.3.2.dev37.dist-info}/RECORD +206 -206
  205. {xmos_ai_tools-1.3.2.dev19.dist-info → xmos_ai_tools-1.3.2.dev37.dist-info}/WHEEL +0 -0
  206. {xmos_ai_tools-1.3.2.dev19.dist-info → xmos_ai_tools-1.3.2.dev37.dist-info}/top_level.txt +0 -0
@@ -18,7 +18,7 @@ limitations under the License.
18
18
  #include "tensorflow/lite/c/builtin_op_data.h"
19
19
  #include "tensorflow/lite/micro/micro_common.h"
20
20
 
21
- namespace tflite {
21
+ namespace tflite_micro {
22
22
 
23
23
  struct OpDataSvdf {
24
24
  int32_t effective_scale_1_a;
@@ -95,6 +95,6 @@ TFLMRegistration Register_SVDF_INT8();
95
95
  inline TFLMRegistration Register_SVDF_INT8() { return Register_SVDF(); }
96
96
 
97
97
  #endif
98
- } // namespace tflite
98
+ } // namespace tflite_micro
99
99
 
100
100
  #endif // TENSORFLOW_LITE_MICRO_KERNELS_SVDF_H_
@@ -18,7 +18,7 @@ limitations under the License.
18
18
 
19
19
  #include "tensorflow/lite/c/common.h"
20
20
 
21
- namespace tflite {
21
+ namespace tflite_micro {
22
22
  extern const int8_t kConvInput1x32x32x3[];
23
23
  extern const int8_t kConvFilter8x3x3x3[];
24
24
  extern const int32_t kConvBiasQuantized8[];
@@ -32,6 +32,6 @@ extern const int32_t kConvZeroBias[];
32
32
  extern const int8_t kConvGoldenOutput4x4InputPaddingSame2x2[];
33
33
  extern const int8_t kConvGoldenOutput5x5InputPaddingSame3x3[];
34
34
 
35
- } // namespace tflite
35
+ } // namespace tflite_micro
36
36
 
37
37
  #endif // TENSORFLOW_LITE_MICRO_KERNELS_CONV_TEST_DATA_H_
@@ -22,7 +22,7 @@ limitations under the License.
22
22
  #include "tensorflow/lite/micro/kernels/lstm_shared.h"
23
23
  #include "tensorflow/lite/micro/test_helpers.h"
24
24
 
25
- namespace tflite {
25
+ namespace tflite_micro {
26
26
  namespace testing {
27
27
  // Data structure to store all the data used to check output of internal gates
28
28
  // of one time step
@@ -421,15 +421,15 @@ CreateQuantizedGateData(
421
421
  const bool fold_zero_point) {
422
422
  GateData<WeightType, BiasType, input_dimension, state_dimension>
423
423
  quantized_gate_params;
424
- tflite::SymmetricQuantize(gate_parameters.activation_weight,
424
+ tflite_micro::SymmetricQuantize(gate_parameters.activation_weight,
425
425
  quantized_gate_params.activation_weight,
426
426
  state_dimension * input_dimension,
427
427
  gate_quantization_params.activation_weight.scale);
428
- tflite::SymmetricQuantize(gate_parameters.recurrent_weight,
428
+ tflite_micro::SymmetricQuantize(gate_parameters.recurrent_weight,
429
429
  quantized_gate_params.recurrent_weight,
430
430
  state_dimension * state_dimension,
431
431
  gate_quantization_params.recurrent_weight.scale);
432
- tflite::SymmetricQuantize(gate_parameters.fused_bias,
432
+ tflite_micro::SymmetricQuantize(gate_parameters.fused_bias,
433
433
  quantized_gate_params.fused_bias, state_dimension,
434
434
  gate_quantization_params.bias.scale);
435
435
  // Note: steps below are not required for the generalized LSTM evaluation
@@ -441,7 +441,7 @@ CreateQuantizedGateData(
441
441
  std::memcpy(quantized_gate_params.activation_zp_folded_bias,
442
442
  quantized_gate_params.fused_bias, 2 * sizeof(int32_t));
443
443
  // Pre-calculate bias - zero_point * weight (a constant).
444
- tflite::tensor_utils::MatrixScalarMultiplyAccumulate(
444
+ tflite_micro::tensor_utils::MatrixScalarMultiplyAccumulate(
445
445
  quantized_gate_params.activation_weight,
446
446
  -1 * input_quantization_params.zero_point, 2, 2,
447
447
  quantized_gate_params.activation_zp_folded_bias);
@@ -451,7 +451,7 @@ CreateQuantizedGateData(
451
451
  quantized_gate_params.recurrent_zp_folded_bias[i] = 0;
452
452
  }
453
453
  // Calculate : -zero_point * weight since it is a constant
454
- tflite::tensor_utils::MatrixScalarMultiplyAccumulate(
454
+ tflite_micro::tensor_utils::MatrixScalarMultiplyAccumulate(
455
455
  quantized_gate_params.recurrent_weight,
456
456
  -1 * output_quantization_params.zero_point, 2, 2,
457
457
  quantized_gate_params.recurrent_zp_folded_bias);
@@ -574,6 +574,6 @@ Create2x3x2X2Int16NodeContents(const float* input_data = nullptr,
574
574
  const float* cell_state = nullptr);
575
575
 
576
576
  } // namespace testing
577
- } // namespace tflite
577
+ } // namespace tflite_micro
578
578
 
579
579
  #endif // TENSORFLOW_LITE_MICRO_KERNELS_TESTDATA_LSTM_TEST_DATA_H_
@@ -22,7 +22,7 @@ limitations under the License.
22
22
  #include "tensorflow/lite/c/common.h"
23
23
  #include "tensorflow/lite/kernels/internal/types.h"
24
24
 
25
- namespace tflite {
25
+ namespace tflite_micro {
26
26
 
27
27
  // This is the most generic TFLMRegistration. The actual supported types
28
28
  // may still be target dependent. The only requirement is that every
@@ -42,6 +42,6 @@ inline TFLMRegistration Register_UNIDIRECTIONAL_SEQUENCE_LSTM_INT8() {
42
42
  }
43
43
  #endif
44
44
 
45
- } // namespace tflite
45
+ } // namespace tflite_micro
46
46
 
47
47
  #endif // TENSORFLOW_LITE_MICRO_KERNELS_UNIDIRECTIONAL_SEQUENCE_LSTM_H_
@@ -26,7 +26,7 @@ limitations under the License.
26
26
  #include "tensorflow/lite/kernels/internal/compatibility.h"
27
27
  #include "tensorflow/lite/micro/kernels/xtensa/xtensa.h"
28
28
 
29
- namespace tflite {
29
+ namespace tflite_micro {
30
30
 
31
31
  // INT24 MIN/MAX
32
32
  #define INT24_MIN -8388608
@@ -134,6 +134,6 @@ inline int CreateQConstantForInt24(int integer_bits, float f) {
134
134
  return static_cast<int>(raw);
135
135
  }
136
136
 
137
- } // namespace tflite
137
+ } // namespace tflite_micro
138
138
  #endif // defined(HIFIMINI)
139
139
  #endif // TENSORFLOW_LITE_MICRO_KERNELS_XTENSA_HIFIMINI_FIXEDPOINT_UTILS_H_
@@ -24,7 +24,7 @@ limitations under the License.
24
24
  #include "tensorflow/lite/kernels/internal/reference/portable_tensor_utils_impl.h"
25
25
  #include "tensorflow/lite/micro/kernels/kernel_util.h"
26
26
 
27
- namespace tflite {
27
+ namespace tflite_micro {
28
28
  namespace ops {
29
29
  namespace micro {
30
30
  namespace lstm_eval {
@@ -212,5 +212,5 @@ TfLiteStatus EvalInteger8x8_8(
212
212
  } // namespace lstm_eval
213
213
  } // namespace micro
214
214
  } // namespace ops
215
- } // namespace tflite
215
+ } // namespace tflite_micro
216
216
  #endif // TENSORFLOW_LITE_KERNELS_LSTM_EVAL_H_
@@ -15,7 +15,7 @@ limitations under the License.
15
15
  #ifndef TENSORFLOW_LITE_KERNELS_LSTM_SHARED_H_
16
16
  #define TENSORFLOW_LITE_KERNELS_LSTM_SHARED_H_
17
17
 
18
- namespace tflite {
18
+ namespace tflite_micro {
19
19
  namespace ops {
20
20
  namespace micro {
21
21
  namespace lstm {
@@ -74,5 +74,5 @@ constexpr int kOutputTensor = 0;
74
74
  } // namespace lstm
75
75
  } // namespace micro
76
76
  } // namespace ops
77
- } // namespace tflite
77
+ } // namespace tflite_micro
78
78
  #endif // TENSORFLOW_LITE_KERNELS_LSTM_SHARED_H_
@@ -21,7 +21,7 @@ limitations under the License.
21
21
  #include "tensorflow/lite/c/builtin_op_data.h"
22
22
  #include "tensorflow/lite/c/common.h"
23
23
  #include "tensorflow/lite/micro/kernels/add.h"
24
- namespace tflite {
24
+ namespace tflite_micro {
25
25
 
26
26
  struct XtensaAddOpData {
27
27
  OpDataAdd reference_op_data;
@@ -43,6 +43,6 @@ TfLiteStatus AddEvalQuantizedVision(TfLiteContext* context, TfLiteNode* node,
43
43
 
44
44
  #endif // VISION_P6
45
45
 
46
- } // namespace tflite
46
+ } // namespace tflite_micro
47
47
 
48
48
  #endif // TENSORFLOW_LITE_MICRO_KERNELS_XTENSA_XTENSA_ADD_H_
@@ -21,7 +21,7 @@ limitations under the License.
21
21
  #include "tensorflow/lite/kernels/internal/types.h"
22
22
  #include "tensorflow/lite/micro/kernels/conv.h"
23
23
 
24
- namespace tflite {
24
+ namespace tflite_micro {
25
25
  struct XtensaConvOpData {
26
26
  OpDataConv reference_op_data;
27
27
 
@@ -84,6 +84,6 @@ TfLiteStatus ConvReferenceEvalInt16(TfLiteContext* context, TfLiteNode* node);
84
84
  void* ConvInitXtensa(TfLiteContext* context, const char* buffer, size_t length);
85
85
  TfLiteStatus ConvPrepareXtensa(TfLiteContext* context, TfLiteNode* node);
86
86
 
87
- } // namespace tflite
87
+ } // namespace tflite_micro
88
88
 
89
89
  #endif // TENSORFLOW_LITE_MICRO_KERNELS_XTENSA_XTENSA_CONV_H_
@@ -21,7 +21,7 @@ limitations under the License.
21
21
  #include "tensorflow/lite/kernels/internal/types.h"
22
22
  #include "tensorflow/lite/micro/kernels/depthwise_conv.h"
23
23
 
24
- namespace tflite {
24
+ namespace tflite_micro {
25
25
  struct XtensaDepthwiseConvOpData {
26
26
  OpDataConv reference_op_data;
27
27
 
@@ -69,6 +69,6 @@ TfLiteStatus DepthwiseConvEvalVision(TfLiteContext* context, TfLiteNode* node,
69
69
 
70
70
  #endif // VISION_P6
71
71
 
72
- } // namespace tflite
72
+ } // namespace tflite_micro
73
73
 
74
74
  #endif // TENSORFLOW_LITE_MICRO_KERNELS_XTENSA_XTENSA_DEPTHWISE_CONV_H_
@@ -22,7 +22,7 @@ limitations under the License.
22
22
  #include "tensorflow/lite/micro/kernels/fully_connected.h"
23
23
  #include "tensorflow/lite/micro/micro_log.h"
24
24
 
25
- namespace tflite {
25
+ namespace tflite_micro {
26
26
  struct XtensaFullyConnectedOpData {
27
27
  OpDataFullyConnected reference_op_data;
28
28
 
@@ -73,6 +73,6 @@ TfLiteStatus XtensaCalculateOpDataFullyConnected(
73
73
  TfLiteStatus XtensaPrepareFullyConnected(TfLiteContext* context,
74
74
  TfLiteNode* node);
75
75
 
76
- } // namespace tflite
76
+ } // namespace tflite_micro
77
77
 
78
78
  #endif // TENSORFLOW_LITE_MICRO_KERNELS_XTENSA_XTENSA_FULLY_CONNECTED_H_
@@ -19,7 +19,7 @@ limitations under the License.
19
19
 
20
20
  #include "tensorflow/lite/c/common.h"
21
21
  #include "tensorflow/lite/kernels/internal/types.h"
22
- namespace tflite {
22
+ namespace tflite_micro {
23
23
 
24
24
  struct OpDataPad {
25
25
  PadParams params;
@@ -44,6 +44,6 @@ TfLiteStatus PadEvalVision(const XtensaPadData& data,
44
44
  TfLiteEvalTensor* output);
45
45
  #endif // VISION_P6
46
46
 
47
- } // namespace tflite
47
+ } // namespace tflite_micro
48
48
 
49
49
  #endif // TENSORFLOW_LITE_MICRO_KERNELS_XTENSA_XTENSA_PAD_H_
@@ -21,7 +21,7 @@ limitations under the License.
21
21
  #include "tensorflow/lite/c/builtin_op_data.h"
22
22
  #include "tensorflow/lite/c/common.h"
23
23
  #include "tensorflow/lite/micro/kernels/pooling.h"
24
- namespace tflite {
24
+ namespace tflite_micro {
25
25
 
26
26
  struct XtensaOpDataPooling {
27
27
  OpDataPooling reference_op_data;
@@ -71,6 +71,6 @@ TfLiteStatus MaxEvalQuantizedHifi(TfLiteContext* context, TfLiteNode* node,
71
71
  void* XtensaPoolingInit(TfLiteContext* context, const char* buffer,
72
72
  size_t length);
73
73
 
74
- } // namespace tflite
74
+ } // namespace tflite_micro
75
75
 
76
76
  #endif // TENSORFLOW_LITE_MICRO_KERNELS_XTENSA_XTENSA_POOLING_H_
@@ -21,7 +21,7 @@ limitations under the License.
21
21
  #include "tensorflow/lite/kernels/internal/types.h"
22
22
  #include "tensorflow/lite/micro/kernels/reduce.h"
23
23
 
24
- namespace tflite {
24
+ namespace tflite_micro {
25
25
 
26
26
  struct XtensaReduceOpData {
27
27
  OpDataReduce reference_op_data;
@@ -42,6 +42,6 @@ TfLiteStatus ReduceEvalVision(const XtensaReduceOpData& data,
42
42
 
43
43
  #endif // VISION_P6
44
44
 
45
- } // namespace tflite
45
+ } // namespace tflite_micro
46
46
 
47
47
  #endif // TENSORFLOW_LITE_MICRO_KERNELS_XTENSA_XTENSA_REDUCE_H_
@@ -20,7 +20,7 @@ limitations under the License.
20
20
  #include "tensorflow/lite/c/common.h"
21
21
  #include "tensorflow/lite/kernels/internal/types.h"
22
22
 
23
- namespace tflite {
23
+ namespace tflite_micro {
24
24
 
25
25
  #if defined(VISION_P6)
26
26
 
@@ -39,6 +39,6 @@ TfLiteStatus ReshapeEvalVision(const XtensaReshapeData& data,
39
39
  TfLiteEvalTensor* output);
40
40
  #endif // VISION_P6
41
41
 
42
- } // namespace tflite
42
+ } // namespace tflite_micro
43
43
 
44
44
  #endif // TENSORFLOW_LITE_MICRO_KERNELS_XTENSA_XTENSA_RESHAPE_H_
@@ -20,7 +20,7 @@ limitations under the License.
20
20
  #include "tensorflow/lite/kernels/internal/types.h"
21
21
  #include "tensorflow/lite/micro/kernels/softmax.h"
22
22
 
23
- namespace tflite {
23
+ namespace tflite_micro {
24
24
 
25
25
  #if defined(HIFI3) || defined(HIFI4) || defined(HIFI5)
26
26
  struct XtensaSoftmaxOpData {
@@ -53,6 +53,6 @@ TfLiteStatus SoftmaxEvalVision(TfLiteContext* context, TfLiteNode* node,
53
53
  TfLiteEvalTensor* output);
54
54
  #endif // defined(VISION_P6)
55
55
 
56
- } // namespace tflite
56
+ } // namespace tflite_micro
57
57
 
58
58
  #endif // TENSORFLOW_LITE_MICRO_KERNELS_XTENSA_XTENSA_SOFTMAX_H_
@@ -22,7 +22,7 @@ limitations under the License.
22
22
  #include "tensorflow/lite/kernels/internal/types.h"
23
23
  #include "tensorflow/lite/micro/kernels/svdf.h"
24
24
 
25
- namespace tflite {
25
+ namespace tflite_micro {
26
26
  #if defined(HIFIMINI)
27
27
  TfLiteStatus EvalIntegerSvdfHifimini(
28
28
  TfLiteContext* context, TfLiteNode* node,
@@ -34,6 +34,6 @@ TfLiteStatus EvalIntegerSvdfHifimini(
34
34
  OpDataSvdf data);
35
35
  #endif // HIFIMINI
36
36
 
37
- } // namespace tflite
37
+ } // namespace tflite_micro
38
38
 
39
39
  #endif // TENSORFLOW_LITE_MICRO_KERNELS_XTENSA_XTENSA_SVDF_H_
@@ -21,7 +21,7 @@ limitations under the License.
21
21
  #include "tensorflow/lite/c/common.h"
22
22
  #include "tensorflow/lite/schema/schema_generated.h"
23
23
 
24
- namespace tflite {
24
+ namespace tflite_micro {
25
25
 
26
26
  // Returns the next pointer address aligned to the given alignment.
27
27
  uint8_t* AlignPointerUp(uint8_t* data, size_t alignment);
@@ -43,7 +43,7 @@ size_t AlignSizeUp(size_t count = 1) {
43
43
  TfLiteStatus TfLiteTypeSizeOf(TfLiteType type, size_t* size);
44
44
 
45
45
  // How many bytes are needed to hold a tensor's contents.
46
- TfLiteStatus BytesRequiredForTensor(const tflite::Tensor& flatbuffer_tensor,
46
+ TfLiteStatus BytesRequiredForTensor(const tflite_micro::Tensor& flatbuffer_tensor,
47
47
  size_t* bytes, size_t* type_size);
48
48
 
49
49
  // How many bytes are used in a TfLiteEvalTensor instance. The byte length is
@@ -59,6 +59,6 @@ TfLiteStatus AllocateOutputDimensionsFromInput(TfLiteContext* context,
59
59
  const TfLiteTensor* input2,
60
60
  TfLiteTensor* output);
61
61
 
62
- } // namespace tflite
62
+ } // namespace tflite_micro
63
63
 
64
64
  #endif // TENSORFLOW_LITE_MICRO_MEMORY_HELPERS_H_
@@ -19,7 +19,7 @@ limitations under the License.
19
19
  #include "tensorflow/lite/micro/compatibility.h"
20
20
  #include "tensorflow/lite/micro/memory_planner/micro_memory_planner.h"
21
21
 
22
- namespace tflite {
22
+ namespace tflite_micro {
23
23
 
24
24
  constexpr int kOnlinePlannedBuffer = -1;
25
25
 
@@ -165,6 +165,6 @@ class GreedyMemoryPlanner : public MicroMemoryPlanner {
165
165
  bool need_to_calculate_offsets_;
166
166
  };
167
167
 
168
- } // namespace tflite
168
+ } // namespace tflite_micro
169
169
 
170
170
  #endif // TENSORFLOW_LITE_MICRO_MEMORY_PLANNER_GREEDY_MEMORY_PLANNER_H_
@@ -19,7 +19,7 @@ limitations under the License.
19
19
  #include "tensorflow/lite/micro/compatibility.h"
20
20
  #include "tensorflow/lite/micro/memory_planner/micro_memory_planner.h"
21
21
 
22
- namespace tflite {
22
+ namespace tflite_micro {
23
23
 
24
24
  // The simplest possible memory planner that just lays out all buffers at
25
25
  // increasing offsets without trying to reuse memory.
@@ -48,6 +48,6 @@ class LinearMemoryPlanner : public MicroMemoryPlanner {
48
48
  TF_LITE_REMOVE_VIRTUAL_DELETE
49
49
  };
50
50
 
51
- } // namespace tflite
51
+ } // namespace tflite_micro
52
52
 
53
53
  #endif // TENSORFLOW_LITE_MICRO_MEMORY_PLANNER_LINEAR_MEMORY_PLANNER_H_
@@ -21,7 +21,7 @@ limitations under the License.
21
21
 
22
22
  #include "tensorflow/lite/micro/micro_utils.h"
23
23
 
24
- namespace tflite {
24
+ namespace tflite_micro {
25
25
 
26
26
  // This is an experimental feature and subjected to change.
27
27
  // More description is available at
@@ -68,6 +68,6 @@ constexpr size_t SizeOfBufferPlan(int32_t buffer_count) {
68
68
  sizeof(BufferDescriptor) * Max(buffer_count - 1, 0);
69
69
  }
70
70
 
71
- } // namespace tflite
71
+ } // namespace tflite_micro
72
72
 
73
73
  #endif // TENSORFLOW_LITE_MICRO_MEMORY_PLANNER_MEMORY_PLAN_STRUCT_H_
@@ -18,7 +18,7 @@ limitations under the License.
18
18
 
19
19
  #include "tensorflow/lite/c/common.h"
20
20
 
21
- namespace tflite {
21
+ namespace tflite_micro {
22
22
 
23
23
  // Interface class for planning the layout of memory buffers during the
24
24
  // execution of a graph.
@@ -90,6 +90,6 @@ class MicroMemoryPlanner {
90
90
  }
91
91
  };
92
92
 
93
- } // namespace tflite
93
+ } // namespace tflite_micro
94
94
 
95
95
  #endif // TENSORFLOW_LITE_MICRO_MICRO_MEMORY_PLANNER_MEMORY_PLANNER_H_
@@ -20,7 +20,7 @@ limitations under the License.
20
20
  #include "tensorflow/lite/micro/memory_planner/memory_plan_struct.h"
21
21
  #include "tensorflow/lite/micro/memory_planner/micro_memory_planner.h"
22
22
 
23
- namespace tflite {
23
+ namespace tflite_micro {
24
24
 
25
25
  /* This is an experimental feature and subjected to change.
26
26
  *
@@ -128,6 +128,6 @@ class NonPersistentMemoryPlannerShim : public MicroMemoryPlanner {
128
128
  TF_LITE_REMOVE_VIRTUAL_DELETE
129
129
  };
130
130
 
131
- } // namespace tflite
131
+ } // namespace tflite_micro
132
132
 
133
133
  #endif // TENSORFLOW_LITE_MICRO_MEMORY_PLANNER_NON_PERSISTENT_MEMORY_PLANNER_SHIM_H__
@@ -21,7 +21,7 @@ limitations under the License.
21
21
  #include "tensorflow/lite/micro/micro_allocator.h"
22
22
  #include "tensorflow/lite/schema/schema_generated.h"
23
23
 
24
- namespace tflite {
24
+ namespace tflite_micro {
25
25
 
26
26
  // Used to hold information used during allocation calculations.
27
27
  struct AllocationInfo {
@@ -127,12 +127,12 @@ class AllocationInfoBuilder {
127
127
  TfLiteStatus ValidateSubgraph(const SubGraph* subgraph,
128
128
  TfLiteEvalTensor* eval_tensors);
129
129
 
130
- const tflite::Model* model_ = nullptr;
130
+ const tflite_micro::Model* model_ = nullptr;
131
131
  INonPersistentBufferAllocator* non_persistent_allocator_ = nullptr;
132
132
  GraphAllocationInfo info_;
133
133
  int allocation_scope_count_ = 0;
134
134
  };
135
135
 
136
- } // namespace tflite
136
+ } // namespace tflite_micro
137
137
 
138
138
  #endif // TENSORFLOW_LITE_MICRO_MICRO_ALLOCATION_INFO_H_
@@ -26,7 +26,7 @@ limitations under the License.
26
26
  #include "tensorflow/lite/micro/tflite_bridge/flatbuffer_conversions_bridge.h"
27
27
  #include "tensorflow/lite/schema/schema_generated.h"
28
28
 
29
- namespace tflite {
29
+ namespace tflite_micro {
30
30
 
31
31
  // TODO(b/199402574): rename to tflite_internal or just remove internal
32
32
  // namespace.
@@ -39,7 +39,7 @@ namespace internal {
39
39
  TfLiteStatus InitializeTfLiteTensorFromFlatbuffer(
40
40
  IPersistentBufferAllocator* persistent_buffer_allocator,
41
41
  INonPersistentBufferAllocator* non_persistent_buffer_allocator,
42
- bool allocate_temp, const tflite::Tensor& flatbuffer_tensor,
42
+ bool allocate_temp, const tflite_micro::Tensor& flatbuffer_tensor,
43
43
  const flatbuffers::Vector<flatbuffers::Offset<Buffer>>* buffers,
44
44
  TfLiteTensor* result);
45
45
 
@@ -347,5 +347,5 @@ class MicroAllocator {
347
347
  TF_LITE_REMOVE_VIRTUAL_DELETE
348
348
  };
349
349
 
350
- } // namespace tflite
350
+ } // namespace tflite_micro
351
351
  #endif // TENSORFLOW_LITE_MICRO_MICRO_ALLOCATOR_H_
@@ -16,13 +16,13 @@ limitations under the License.
16
16
  #ifndef TENSORFLOW_LITE_MICRO_MICRO_ARENA_CONSTANTS_H_
17
17
  #define TENSORFLOW_LITE_MICRO_MICRO_ARENA_CONSTANTS_H_
18
18
 
19
- namespace tflite {
19
+ namespace tflite_micro {
20
20
 
21
21
  // The default buffer alignment requirement.
22
22
  // We align tensor buffers to 16-byte boundaries, since this is a common
23
23
  // requirement for SIMD extensions.
24
24
  constexpr int MicroArenaBufferAlignment() { return 16; }
25
25
 
26
- } // namespace tflite
26
+ } // namespace tflite_micro
27
27
 
28
28
  #endif // TENSORFLOW_LITE_MICRO_MICRO_ARENA_CONSTANTS_H_
@@ -23,7 +23,7 @@ limitations under the License.
23
23
 
24
24
  #ifdef NO_INTERPRETER
25
25
 
26
- namespace tflite {
26
+ namespace tflite_micro {
27
27
  const TfLiteStatus kTfLiteAbort = static_cast<TfLiteStatus>(15);
28
28
 
29
29
  struct MicroContext{
@@ -40,7 +40,7 @@ namespace tflite {
40
40
 
41
41
  #else
42
42
 
43
- namespace tflite {
43
+ namespace tflite_micro {
44
44
  // TODO(b/149795762): kTfLiteAbort cannot be part of the tflite TfLiteStatus.
45
45
  const TfLiteStatus kTfLiteAbort = static_cast<TfLiteStatus>(15);
46
46
 
@@ -169,7 +169,7 @@ inline TfLiteExternalContext* MicroContextGetExternalContext(
169
169
  void MicroContextReportOpError(struct TfLiteContext* context,
170
170
  const char* format, ...);
171
171
 
172
- } // namespace tflite
172
+ } // namespace tflite_micro
173
173
 
174
174
  #endif // NO_INTERPRETER
175
175
 
@@ -22,7 +22,7 @@ limitations under the License.
22
22
 
23
23
  #ifdef NO_INTERPRETER
24
24
 
25
- namespace tflite {
25
+ namespace tflite_micro {
26
26
  struct MicroGraph{
27
27
  int (*NumSubgraphs)();
28
28
  size_t (*NumSubgraphInputs)(int subgraph_idx);
@@ -35,11 +35,11 @@ namespace tflite {
35
35
 
36
36
  #else
37
37
 
38
- namespace tflite {
38
+ namespace tflite_micro {
39
39
 
40
40
  // Abstracts the details of interacting with the graph from the kernels
41
41
  //
42
- // Provides methods to invoke any subgraph in the tflite::Graph.
42
+ // Provides methods to invoke any subgraph in the tflite_micro::Graph.
43
43
  class MicroGraph {
44
44
  public:
45
45
  virtual ~MicroGraph() = default;
@@ -72,7 +72,7 @@ class MicroGraph {
72
72
  TF_LITE_REMOVE_VIRTUAL_DELETE
73
73
  };
74
74
 
75
- } // namespace tflite
75
+ } // namespace tflite_micro
76
76
 
77
77
  #endif // NO_INTERPRETER
78
78
 
@@ -35,7 +35,7 @@ limitations under the License.
35
35
  // tensorflow/core.
36
36
  #define TFLITE_SCHEMA_VERSION (3)
37
37
 
38
- namespace tflite {
38
+ namespace tflite_micro {
39
39
 
40
40
  class MicroInterpreter {
41
41
  public:
@@ -184,6 +184,6 @@ class MicroInterpreter {
184
184
  MicroInterpreterContext micro_context_;
185
185
  };
186
186
 
187
- } // namespace tflite
187
+ } // namespace tflite_micro
188
188
 
189
189
  #endif // TENSORFLOW_LITE_MICRO_MICRO_INTERPRETER_H_
@@ -22,7 +22,7 @@ limitations under the License.
22
22
  #include "tensorflow/lite/micro/micro_interpreter_graph.h"
23
23
  #include "tensorflow/lite/micro/micro_log.h"
24
24
 
25
- namespace tflite {
25
+ namespace tflite_micro {
26
26
 
27
27
  // A full implementation of the MicroContext, to be used by the
28
28
  // MicroInterpreter. Kernels should not depend on this directly. Instead they
@@ -120,6 +120,6 @@ class MicroInterpreterContext : public MicroContext {
120
120
  TF_LITE_REMOVE_VIRTUAL_DELETE
121
121
  };
122
122
 
123
- } // namespace tflite
123
+ } // namespace tflite_micro
124
124
 
125
125
  #endif // TENSORFLOW_LITE_MICRO_MICRO_INTERPRETER_CONTEXT_H_
@@ -22,12 +22,12 @@ limitations under the License.
22
22
  #include "tensorflow/lite/micro/micro_resource_variable.h"
23
23
  #include "tensorflow/lite/schema/schema_generated.h"
24
24
 
25
- namespace tflite {
25
+ namespace tflite_micro {
26
26
 
27
- // Abstracts the details of interacting with the tflite::Model.
27
+ // Abstracts the details of interacting with the tflite_micro::Model.
28
28
  //
29
29
  // Provides methods to access, initialize, prepare, invoke and free any
30
- // subgraph in the tflite::Graph.
30
+ // subgraph in the tflite_micro::Graph.
31
31
  class MicroInterpreterGraph : public MicroGraph {
32
32
  public:
33
33
  // The lifetime of the context, model, allocator and resource_variables must
@@ -105,6 +105,6 @@ class MicroInterpreterGraph : public MicroGraph {
105
105
  TF_LITE_REMOVE_VIRTUAL_DELETE
106
106
  };
107
107
 
108
- } // namespace tflite
108
+ } // namespace tflite_micro
109
109
 
110
110
  #endif // TENSORFLOW_LITE_MICRO_MICRO_INTERPRETER_GRAPH_H_