xmos-ai-tools 1.3.2.dev19__py3-none-macosx_11_0_arm64.whl → 1.3.2.dev37__py3-none-macosx_11_0_arm64.whl
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- xmos_ai_tools/runtime/include/lib_tflite_micro/api/inference_engine.h +9 -9
- xmos_ai_tools/runtime/include/lib_tflite_micro/src/tflite-xcore-kernels/xcore_common.h +2 -2
- xmos_ai_tools/runtime/include/lib_tflite_micro/src/tflite-xcore-kernels/xcore_custom_options.h +2 -2
- xmos_ai_tools/runtime/include/lib_tflite_micro/src/tflite-xcore-kernels/xcore_error_reporter.h +3 -3
- xmos_ai_tools/runtime/include/lib_tflite_micro/src/tflite-xcore-kernels/xcore_interpreter.h +8 -8
- xmos_ai_tools/runtime/include/lib_tflite_micro/src/tflite-xcore-kernels/xcore_ops.h +3 -3
- xmos_ai_tools/runtime/include/lib_tflite_micro/src/tflite-xcore-kernels/xcore_profiler.h +4 -4
- xmos_ai_tools/runtime/include/lib_tflite_micro/src/tflite-xcore-kernels/xcore_utils.h +5 -5
- xmos_ai_tools/runtime/include/lib_xud/lib_xud/api/xud.h +3 -3
- xmos_ai_tools/runtime/include/tensorflow/lite/array.h +4 -4
- xmos_ai_tools/runtime/include/tensorflow/lite/context_util.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/core/api/error_reporter.h +3 -3
- xmos_ai_tools/runtime/include/tensorflow/lite/core/api/flatbuffer_conversions.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/core/api/tensor_utils.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/core/c/c_api_types.h +3 -3
- xmos_ai_tools/runtime/include/tensorflow/lite/core/c/common.h +17 -17
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/common.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/cppmath.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/max.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/min.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/portable_tensor.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/portable_tensor_utils.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/quantization_util.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/add.h +3 -3
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/add_n.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/arg_min_max.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/batch_matmul.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/batch_to_space_nd.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/binary_function.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/broadcast_args.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/broadcast_to.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/ceil.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/comparisons.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/concatenation.h +3 -3
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/conv.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/cumsum.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/depth_to_space.h +3 -3
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/depthwiseconv_float.h +1 -1
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/depthwiseconv_uint8.h +1 -1
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/dequantize.h +4 -4
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/div.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/elu.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/exp.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/fill.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/floor.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/floor_div.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/floor_mod.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/fully_connected.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/hard_swish.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/integer_ops/add.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/integer_ops/conv.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/integer_ops/depthwise_conv.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/integer_ops/fully_connected.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/integer_ops/l2normalization.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/integer_ops/logistic.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/integer_ops/mul.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/integer_ops/pooling.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/integer_ops/tanh.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/integer_ops/transpose_conv.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/l2normalization.h +4 -4
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/leaky_relu.h +3 -3
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/log_softmax.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/logistic.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/lstm_cell.h +5 -5
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/maximum_minimum.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/mul.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/neg.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/pad.h +8 -8
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/pooling.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/portable_tensor_utils.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/portable_tensor_utils_impl.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/prelu.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/process_broadcast_shapes.h +3 -3
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/quantize.h +4 -4
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/reduce.h +3 -3
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/requantize.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/resize_bilinear.h +4 -4
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/resize_nearest_neighbor.h +3 -3
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/round.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/select.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/slice.h +5 -5
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/softmax.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/space_to_batch_nd.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/space_to_depth.h +3 -3
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/strided_slice.h +6 -6
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/sub.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/tanh.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/transpose.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/reference/transpose_conv.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/runtime_shape.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/strided_slice_logic.h +9 -9
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/tensor_ctypes.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/internal/types.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/kernel_util.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/kernels/padding.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/arena_allocator/ibuffer_allocator.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/arena_allocator/non_persistent_arena_buffer_allocator.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/arena_allocator/persistent_arena_buffer_allocator.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/arena_allocator/recording_single_arena_buffer_allocator.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/arena_allocator/single_arena_buffer_allocator.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/benchmarks/micro_benchmark.h +7 -7
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/fake_micro_context.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/flatbuffer_utils.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/activation_utils.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/activations.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/add.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/arc_mli/mli_function_specializations.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/arc_mli/mli_interface.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/arc_mli/mli_slicers.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/arc_mli/mli_tf_utils.h +4 -4
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/arc_mli/scratch_buf_mgr.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/arc_mli/scratch_buffers.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/ceva/types.h +6 -6
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/circular_buffer.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/conv.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/conv_test.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/depthwise_conv.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/dequantize.h +3 -3
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/ethosu.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/fully_connected.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/hard_swish.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/kernel_runner.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/kernel_util.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/leaky_relu.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/logical.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/logistic.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/lstm_eval.h +48 -48
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/lstm_eval_test.h +57 -57
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/lstm_shared.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/micro_ops.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/micro_tensor_utils.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/mul.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/pad.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/pooling.h +15 -15
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/prelu.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/quantize.h +3 -3
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/reduce.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/reshape.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/softmax.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/strided_slice.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/sub.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/svdf.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/testdata/conv_test_data.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/testdata/lstm_test_data.h +7 -7
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/unidirectional_sequence_lstm.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/xtensa/hifimini/fixedpoint_utils.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/xtensa/lstm_eval.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/xtensa/lstm_shared.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/xtensa/xtensa_add.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/xtensa/xtensa_conv.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/xtensa/xtensa_depthwise_conv.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/xtensa/xtensa_fully_connected.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/xtensa/xtensa_pad.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/xtensa/xtensa_pooling.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/xtensa/xtensa_reduce.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/xtensa/xtensa_reshape.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/xtensa/xtensa_softmax.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/xtensa/xtensa_svdf.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/memory_helpers.h +3 -3
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/memory_planner/greedy_memory_planner.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/memory_planner/linear_memory_planner.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/memory_planner/memory_plan_struct.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/memory_planner/micro_memory_planner.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/memory_planner/non_persistent_buffer_planner_shim.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_allocation_info.h +3 -3
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_allocator.h +3 -3
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_arena_constants.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_context.h +3 -3
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_graph.h +4 -4
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_interpreter.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_interpreter_context.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_interpreter_graph.h +4 -4
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_log.h +4 -4
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_mutable_op_resolver.h +50 -50
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_op_resolver.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_profiler.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_profiler_interface.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_resource_variable.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_time.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/micro_utils.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/mock_micro_graph.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/python/tflite_size/src/flatbuffer_size.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/python/tflite_size/src/flatbuffer_size_wrapper.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/recording_micro_allocator.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/recording_micro_interpreter.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/system_setup.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/test_helper_custom_ops.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/test_helpers.h +6 -6
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/testing/micro_test.h +4 -4
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/tflite_bridge/flatbuffer_conversions_bridge.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/tflite_bridge/micro_error_reporter.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/tools/benchmarking/log_utils.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/tools/benchmarking/metrics.h +3 -3
- xmos_ai_tools/runtime/include/tensorflow/lite/micro/tools/benchmarking/op_resolver.h +3 -3
- xmos_ai_tools/runtime/include/tensorflow/lite/portable_type_to_tflitetype.h +2 -2
- xmos_ai_tools/runtime/include/tensorflow/lite/schema/schema_generated.h +2731 -2731
- xmos_ai_tools/runtime/include/tensorflow/lite/schema/schema_utils.h +2 -2
- xmos_ai_tools/runtime/lib/libhost_xtflitemicro.a +0 -0
- xmos_ai_tools/runtime/lib/libxtflitemicro.a +0 -0
- xmos_ai_tools/xinterpreters/libs/macos/xtflm_python.1.0.1.dylib +0 -0
- xmos_ai_tools/xinterpreters/libs/macos/xtflm_python.dylib +0 -0
- {xmos_ai_tools-1.3.2.dev19.data → xmos_ai_tools-1.3.2.dev37.data}/data/bin/xcore-opt +0 -0
- {xmos_ai_tools-1.3.2.dev19.dist-info → xmos_ai_tools-1.3.2.dev37.dist-info}/METADATA +2 -2
- {xmos_ai_tools-1.3.2.dev19.dist-info → xmos_ai_tools-1.3.2.dev37.dist-info}/RECORD +206 -206
- {xmos_ai_tools-1.3.2.dev19.dist-info → xmos_ai_tools-1.3.2.dev37.dist-info}/WHEEL +0 -0
- {xmos_ai_tools-1.3.2.dev19.dist-info → xmos_ai_tools-1.3.2.dev37.dist-info}/top_level.txt +0 -0
@@ -18,7 +18,7 @@ limitations under the License.
|
|
18
18
|
#include "tensorflow/lite/c/builtin_op_data.h"
|
19
19
|
#include "tensorflow/lite/micro/micro_common.h"
|
20
20
|
|
21
|
-
namespace
|
21
|
+
namespace tflite_micro {
|
22
22
|
|
23
23
|
struct OpDataSvdf {
|
24
24
|
int32_t effective_scale_1_a;
|
@@ -95,6 +95,6 @@ TFLMRegistration Register_SVDF_INT8();
|
|
95
95
|
inline TFLMRegistration Register_SVDF_INT8() { return Register_SVDF(); }
|
96
96
|
|
97
97
|
#endif
|
98
|
-
} // namespace
|
98
|
+
} // namespace tflite_micro
|
99
99
|
|
100
100
|
#endif // TENSORFLOW_LITE_MICRO_KERNELS_SVDF_H_
|
@@ -18,7 +18,7 @@ limitations under the License.
|
|
18
18
|
|
19
19
|
#include "tensorflow/lite/c/common.h"
|
20
20
|
|
21
|
-
namespace
|
21
|
+
namespace tflite_micro {
|
22
22
|
extern const int8_t kConvInput1x32x32x3[];
|
23
23
|
extern const int8_t kConvFilter8x3x3x3[];
|
24
24
|
extern const int32_t kConvBiasQuantized8[];
|
@@ -32,6 +32,6 @@ extern const int32_t kConvZeroBias[];
|
|
32
32
|
extern const int8_t kConvGoldenOutput4x4InputPaddingSame2x2[];
|
33
33
|
extern const int8_t kConvGoldenOutput5x5InputPaddingSame3x3[];
|
34
34
|
|
35
|
-
} // namespace
|
35
|
+
} // namespace tflite_micro
|
36
36
|
|
37
37
|
#endif // TENSORFLOW_LITE_MICRO_KERNELS_CONV_TEST_DATA_H_
|
@@ -22,7 +22,7 @@ limitations under the License.
|
|
22
22
|
#include "tensorflow/lite/micro/kernels/lstm_shared.h"
|
23
23
|
#include "tensorflow/lite/micro/test_helpers.h"
|
24
24
|
|
25
|
-
namespace
|
25
|
+
namespace tflite_micro {
|
26
26
|
namespace testing {
|
27
27
|
// Data structure to store all the data used to check output of internal gates
|
28
28
|
// of one time step
|
@@ -421,15 +421,15 @@ CreateQuantizedGateData(
|
|
421
421
|
const bool fold_zero_point) {
|
422
422
|
GateData<WeightType, BiasType, input_dimension, state_dimension>
|
423
423
|
quantized_gate_params;
|
424
|
-
|
424
|
+
tflite_micro::SymmetricQuantize(gate_parameters.activation_weight,
|
425
425
|
quantized_gate_params.activation_weight,
|
426
426
|
state_dimension * input_dimension,
|
427
427
|
gate_quantization_params.activation_weight.scale);
|
428
|
-
|
428
|
+
tflite_micro::SymmetricQuantize(gate_parameters.recurrent_weight,
|
429
429
|
quantized_gate_params.recurrent_weight,
|
430
430
|
state_dimension * state_dimension,
|
431
431
|
gate_quantization_params.recurrent_weight.scale);
|
432
|
-
|
432
|
+
tflite_micro::SymmetricQuantize(gate_parameters.fused_bias,
|
433
433
|
quantized_gate_params.fused_bias, state_dimension,
|
434
434
|
gate_quantization_params.bias.scale);
|
435
435
|
// Note: steps below are not required for the generalized LSTM evaluation
|
@@ -441,7 +441,7 @@ CreateQuantizedGateData(
|
|
441
441
|
std::memcpy(quantized_gate_params.activation_zp_folded_bias,
|
442
442
|
quantized_gate_params.fused_bias, 2 * sizeof(int32_t));
|
443
443
|
// Pre-calculate bias - zero_point * weight (a constant).
|
444
|
-
|
444
|
+
tflite_micro::tensor_utils::MatrixScalarMultiplyAccumulate(
|
445
445
|
quantized_gate_params.activation_weight,
|
446
446
|
-1 * input_quantization_params.zero_point, 2, 2,
|
447
447
|
quantized_gate_params.activation_zp_folded_bias);
|
@@ -451,7 +451,7 @@ CreateQuantizedGateData(
|
|
451
451
|
quantized_gate_params.recurrent_zp_folded_bias[i] = 0;
|
452
452
|
}
|
453
453
|
// Calculate : -zero_point * weight since it is a constant
|
454
|
-
|
454
|
+
tflite_micro::tensor_utils::MatrixScalarMultiplyAccumulate(
|
455
455
|
quantized_gate_params.recurrent_weight,
|
456
456
|
-1 * output_quantization_params.zero_point, 2, 2,
|
457
457
|
quantized_gate_params.recurrent_zp_folded_bias);
|
@@ -574,6 +574,6 @@ Create2x3x2X2Int16NodeContents(const float* input_data = nullptr,
|
|
574
574
|
const float* cell_state = nullptr);
|
575
575
|
|
576
576
|
} // namespace testing
|
577
|
-
} // namespace
|
577
|
+
} // namespace tflite_micro
|
578
578
|
|
579
579
|
#endif // TENSORFLOW_LITE_MICRO_KERNELS_TESTDATA_LSTM_TEST_DATA_H_
|
@@ -22,7 +22,7 @@ limitations under the License.
|
|
22
22
|
#include "tensorflow/lite/c/common.h"
|
23
23
|
#include "tensorflow/lite/kernels/internal/types.h"
|
24
24
|
|
25
|
-
namespace
|
25
|
+
namespace tflite_micro {
|
26
26
|
|
27
27
|
// This is the most generic TFLMRegistration. The actual supported types
|
28
28
|
// may still be target dependent. The only requirement is that every
|
@@ -42,6 +42,6 @@ inline TFLMRegistration Register_UNIDIRECTIONAL_SEQUENCE_LSTM_INT8() {
|
|
42
42
|
}
|
43
43
|
#endif
|
44
44
|
|
45
|
-
} // namespace
|
45
|
+
} // namespace tflite_micro
|
46
46
|
|
47
47
|
#endif // TENSORFLOW_LITE_MICRO_KERNELS_UNIDIRECTIONAL_SEQUENCE_LSTM_H_
|
xmos_ai_tools/runtime/include/tensorflow/lite/micro/kernels/xtensa/hifimini/fixedpoint_utils.h
CHANGED
@@ -26,7 +26,7 @@ limitations under the License.
|
|
26
26
|
#include "tensorflow/lite/kernels/internal/compatibility.h"
|
27
27
|
#include "tensorflow/lite/micro/kernels/xtensa/xtensa.h"
|
28
28
|
|
29
|
-
namespace
|
29
|
+
namespace tflite_micro {
|
30
30
|
|
31
31
|
// INT24 MIN/MAX
|
32
32
|
#define INT24_MIN -8388608
|
@@ -134,6 +134,6 @@ inline int CreateQConstantForInt24(int integer_bits, float f) {
|
|
134
134
|
return static_cast<int>(raw);
|
135
135
|
}
|
136
136
|
|
137
|
-
} // namespace
|
137
|
+
} // namespace tflite_micro
|
138
138
|
#endif // defined(HIFIMINI)
|
139
139
|
#endif // TENSORFLOW_LITE_MICRO_KERNELS_XTENSA_HIFIMINI_FIXEDPOINT_UTILS_H_
|
@@ -24,7 +24,7 @@ limitations under the License.
|
|
24
24
|
#include "tensorflow/lite/kernels/internal/reference/portable_tensor_utils_impl.h"
|
25
25
|
#include "tensorflow/lite/micro/kernels/kernel_util.h"
|
26
26
|
|
27
|
-
namespace
|
27
|
+
namespace tflite_micro {
|
28
28
|
namespace ops {
|
29
29
|
namespace micro {
|
30
30
|
namespace lstm_eval {
|
@@ -212,5 +212,5 @@ TfLiteStatus EvalInteger8x8_8(
|
|
212
212
|
} // namespace lstm_eval
|
213
213
|
} // namespace micro
|
214
214
|
} // namespace ops
|
215
|
-
} // namespace
|
215
|
+
} // namespace tflite_micro
|
216
216
|
#endif // TENSORFLOW_LITE_KERNELS_LSTM_EVAL_H_
|
@@ -15,7 +15,7 @@ limitations under the License.
|
|
15
15
|
#ifndef TENSORFLOW_LITE_KERNELS_LSTM_SHARED_H_
|
16
16
|
#define TENSORFLOW_LITE_KERNELS_LSTM_SHARED_H_
|
17
17
|
|
18
|
-
namespace
|
18
|
+
namespace tflite_micro {
|
19
19
|
namespace ops {
|
20
20
|
namespace micro {
|
21
21
|
namespace lstm {
|
@@ -74,5 +74,5 @@ constexpr int kOutputTensor = 0;
|
|
74
74
|
} // namespace lstm
|
75
75
|
} // namespace micro
|
76
76
|
} // namespace ops
|
77
|
-
} // namespace
|
77
|
+
} // namespace tflite_micro
|
78
78
|
#endif // TENSORFLOW_LITE_KERNELS_LSTM_SHARED_H_
|
@@ -21,7 +21,7 @@ limitations under the License.
|
|
21
21
|
#include "tensorflow/lite/c/builtin_op_data.h"
|
22
22
|
#include "tensorflow/lite/c/common.h"
|
23
23
|
#include "tensorflow/lite/micro/kernels/add.h"
|
24
|
-
namespace
|
24
|
+
namespace tflite_micro {
|
25
25
|
|
26
26
|
struct XtensaAddOpData {
|
27
27
|
OpDataAdd reference_op_data;
|
@@ -43,6 +43,6 @@ TfLiteStatus AddEvalQuantizedVision(TfLiteContext* context, TfLiteNode* node,
|
|
43
43
|
|
44
44
|
#endif // VISION_P6
|
45
45
|
|
46
|
-
} // namespace
|
46
|
+
} // namespace tflite_micro
|
47
47
|
|
48
48
|
#endif // TENSORFLOW_LITE_MICRO_KERNELS_XTENSA_XTENSA_ADD_H_
|
@@ -21,7 +21,7 @@ limitations under the License.
|
|
21
21
|
#include "tensorflow/lite/kernels/internal/types.h"
|
22
22
|
#include "tensorflow/lite/micro/kernels/conv.h"
|
23
23
|
|
24
|
-
namespace
|
24
|
+
namespace tflite_micro {
|
25
25
|
struct XtensaConvOpData {
|
26
26
|
OpDataConv reference_op_data;
|
27
27
|
|
@@ -84,6 +84,6 @@ TfLiteStatus ConvReferenceEvalInt16(TfLiteContext* context, TfLiteNode* node);
|
|
84
84
|
void* ConvInitXtensa(TfLiteContext* context, const char* buffer, size_t length);
|
85
85
|
TfLiteStatus ConvPrepareXtensa(TfLiteContext* context, TfLiteNode* node);
|
86
86
|
|
87
|
-
} // namespace
|
87
|
+
} // namespace tflite_micro
|
88
88
|
|
89
89
|
#endif // TENSORFLOW_LITE_MICRO_KERNELS_XTENSA_XTENSA_CONV_H_
|
@@ -21,7 +21,7 @@ limitations under the License.
|
|
21
21
|
#include "tensorflow/lite/kernels/internal/types.h"
|
22
22
|
#include "tensorflow/lite/micro/kernels/depthwise_conv.h"
|
23
23
|
|
24
|
-
namespace
|
24
|
+
namespace tflite_micro {
|
25
25
|
struct XtensaDepthwiseConvOpData {
|
26
26
|
OpDataConv reference_op_data;
|
27
27
|
|
@@ -69,6 +69,6 @@ TfLiteStatus DepthwiseConvEvalVision(TfLiteContext* context, TfLiteNode* node,
|
|
69
69
|
|
70
70
|
#endif // VISION_P6
|
71
71
|
|
72
|
-
} // namespace
|
72
|
+
} // namespace tflite_micro
|
73
73
|
|
74
74
|
#endif // TENSORFLOW_LITE_MICRO_KERNELS_XTENSA_XTENSA_DEPTHWISE_CONV_H_
|
@@ -22,7 +22,7 @@ limitations under the License.
|
|
22
22
|
#include "tensorflow/lite/micro/kernels/fully_connected.h"
|
23
23
|
#include "tensorflow/lite/micro/micro_log.h"
|
24
24
|
|
25
|
-
namespace
|
25
|
+
namespace tflite_micro {
|
26
26
|
struct XtensaFullyConnectedOpData {
|
27
27
|
OpDataFullyConnected reference_op_data;
|
28
28
|
|
@@ -73,6 +73,6 @@ TfLiteStatus XtensaCalculateOpDataFullyConnected(
|
|
73
73
|
TfLiteStatus XtensaPrepareFullyConnected(TfLiteContext* context,
|
74
74
|
TfLiteNode* node);
|
75
75
|
|
76
|
-
} // namespace
|
76
|
+
} // namespace tflite_micro
|
77
77
|
|
78
78
|
#endif // TENSORFLOW_LITE_MICRO_KERNELS_XTENSA_XTENSA_FULLY_CONNECTED_H_
|
@@ -19,7 +19,7 @@ limitations under the License.
|
|
19
19
|
|
20
20
|
#include "tensorflow/lite/c/common.h"
|
21
21
|
#include "tensorflow/lite/kernels/internal/types.h"
|
22
|
-
namespace
|
22
|
+
namespace tflite_micro {
|
23
23
|
|
24
24
|
struct OpDataPad {
|
25
25
|
PadParams params;
|
@@ -44,6 +44,6 @@ TfLiteStatus PadEvalVision(const XtensaPadData& data,
|
|
44
44
|
TfLiteEvalTensor* output);
|
45
45
|
#endif // VISION_P6
|
46
46
|
|
47
|
-
} // namespace
|
47
|
+
} // namespace tflite_micro
|
48
48
|
|
49
49
|
#endif // TENSORFLOW_LITE_MICRO_KERNELS_XTENSA_XTENSA_PAD_H_
|
@@ -21,7 +21,7 @@ limitations under the License.
|
|
21
21
|
#include "tensorflow/lite/c/builtin_op_data.h"
|
22
22
|
#include "tensorflow/lite/c/common.h"
|
23
23
|
#include "tensorflow/lite/micro/kernels/pooling.h"
|
24
|
-
namespace
|
24
|
+
namespace tflite_micro {
|
25
25
|
|
26
26
|
struct XtensaOpDataPooling {
|
27
27
|
OpDataPooling reference_op_data;
|
@@ -71,6 +71,6 @@ TfLiteStatus MaxEvalQuantizedHifi(TfLiteContext* context, TfLiteNode* node,
|
|
71
71
|
void* XtensaPoolingInit(TfLiteContext* context, const char* buffer,
|
72
72
|
size_t length);
|
73
73
|
|
74
|
-
} // namespace
|
74
|
+
} // namespace tflite_micro
|
75
75
|
|
76
76
|
#endif // TENSORFLOW_LITE_MICRO_KERNELS_XTENSA_XTENSA_POOLING_H_
|
@@ -21,7 +21,7 @@ limitations under the License.
|
|
21
21
|
#include "tensorflow/lite/kernels/internal/types.h"
|
22
22
|
#include "tensorflow/lite/micro/kernels/reduce.h"
|
23
23
|
|
24
|
-
namespace
|
24
|
+
namespace tflite_micro {
|
25
25
|
|
26
26
|
struct XtensaReduceOpData {
|
27
27
|
OpDataReduce reference_op_data;
|
@@ -42,6 +42,6 @@ TfLiteStatus ReduceEvalVision(const XtensaReduceOpData& data,
|
|
42
42
|
|
43
43
|
#endif // VISION_P6
|
44
44
|
|
45
|
-
} // namespace
|
45
|
+
} // namespace tflite_micro
|
46
46
|
|
47
47
|
#endif // TENSORFLOW_LITE_MICRO_KERNELS_XTENSA_XTENSA_REDUCE_H_
|
@@ -20,7 +20,7 @@ limitations under the License.
|
|
20
20
|
#include "tensorflow/lite/c/common.h"
|
21
21
|
#include "tensorflow/lite/kernels/internal/types.h"
|
22
22
|
|
23
|
-
namespace
|
23
|
+
namespace tflite_micro {
|
24
24
|
|
25
25
|
#if defined(VISION_P6)
|
26
26
|
|
@@ -39,6 +39,6 @@ TfLiteStatus ReshapeEvalVision(const XtensaReshapeData& data,
|
|
39
39
|
TfLiteEvalTensor* output);
|
40
40
|
#endif // VISION_P6
|
41
41
|
|
42
|
-
} // namespace
|
42
|
+
} // namespace tflite_micro
|
43
43
|
|
44
44
|
#endif // TENSORFLOW_LITE_MICRO_KERNELS_XTENSA_XTENSA_RESHAPE_H_
|
@@ -20,7 +20,7 @@ limitations under the License.
|
|
20
20
|
#include "tensorflow/lite/kernels/internal/types.h"
|
21
21
|
#include "tensorflow/lite/micro/kernels/softmax.h"
|
22
22
|
|
23
|
-
namespace
|
23
|
+
namespace tflite_micro {
|
24
24
|
|
25
25
|
#if defined(HIFI3) || defined(HIFI4) || defined(HIFI5)
|
26
26
|
struct XtensaSoftmaxOpData {
|
@@ -53,6 +53,6 @@ TfLiteStatus SoftmaxEvalVision(TfLiteContext* context, TfLiteNode* node,
|
|
53
53
|
TfLiteEvalTensor* output);
|
54
54
|
#endif // defined(VISION_P6)
|
55
55
|
|
56
|
-
} // namespace
|
56
|
+
} // namespace tflite_micro
|
57
57
|
|
58
58
|
#endif // TENSORFLOW_LITE_MICRO_KERNELS_XTENSA_XTENSA_SOFTMAX_H_
|
@@ -22,7 +22,7 @@ limitations under the License.
|
|
22
22
|
#include "tensorflow/lite/kernels/internal/types.h"
|
23
23
|
#include "tensorflow/lite/micro/kernels/svdf.h"
|
24
24
|
|
25
|
-
namespace
|
25
|
+
namespace tflite_micro {
|
26
26
|
#if defined(HIFIMINI)
|
27
27
|
TfLiteStatus EvalIntegerSvdfHifimini(
|
28
28
|
TfLiteContext* context, TfLiteNode* node,
|
@@ -34,6 +34,6 @@ TfLiteStatus EvalIntegerSvdfHifimini(
|
|
34
34
|
OpDataSvdf data);
|
35
35
|
#endif // HIFIMINI
|
36
36
|
|
37
|
-
} // namespace
|
37
|
+
} // namespace tflite_micro
|
38
38
|
|
39
39
|
#endif // TENSORFLOW_LITE_MICRO_KERNELS_XTENSA_XTENSA_SVDF_H_
|
@@ -21,7 +21,7 @@ limitations under the License.
|
|
21
21
|
#include "tensorflow/lite/c/common.h"
|
22
22
|
#include "tensorflow/lite/schema/schema_generated.h"
|
23
23
|
|
24
|
-
namespace
|
24
|
+
namespace tflite_micro {
|
25
25
|
|
26
26
|
// Returns the next pointer address aligned to the given alignment.
|
27
27
|
uint8_t* AlignPointerUp(uint8_t* data, size_t alignment);
|
@@ -43,7 +43,7 @@ size_t AlignSizeUp(size_t count = 1) {
|
|
43
43
|
TfLiteStatus TfLiteTypeSizeOf(TfLiteType type, size_t* size);
|
44
44
|
|
45
45
|
// How many bytes are needed to hold a tensor's contents.
|
46
|
-
TfLiteStatus BytesRequiredForTensor(const
|
46
|
+
TfLiteStatus BytesRequiredForTensor(const tflite_micro::Tensor& flatbuffer_tensor,
|
47
47
|
size_t* bytes, size_t* type_size);
|
48
48
|
|
49
49
|
// How many bytes are used in a TfLiteEvalTensor instance. The byte length is
|
@@ -59,6 +59,6 @@ TfLiteStatus AllocateOutputDimensionsFromInput(TfLiteContext* context,
|
|
59
59
|
const TfLiteTensor* input2,
|
60
60
|
TfLiteTensor* output);
|
61
61
|
|
62
|
-
} // namespace
|
62
|
+
} // namespace tflite_micro
|
63
63
|
|
64
64
|
#endif // TENSORFLOW_LITE_MICRO_MEMORY_HELPERS_H_
|
@@ -19,7 +19,7 @@ limitations under the License.
|
|
19
19
|
#include "tensorflow/lite/micro/compatibility.h"
|
20
20
|
#include "tensorflow/lite/micro/memory_planner/micro_memory_planner.h"
|
21
21
|
|
22
|
-
namespace
|
22
|
+
namespace tflite_micro {
|
23
23
|
|
24
24
|
constexpr int kOnlinePlannedBuffer = -1;
|
25
25
|
|
@@ -165,6 +165,6 @@ class GreedyMemoryPlanner : public MicroMemoryPlanner {
|
|
165
165
|
bool need_to_calculate_offsets_;
|
166
166
|
};
|
167
167
|
|
168
|
-
} // namespace
|
168
|
+
} // namespace tflite_micro
|
169
169
|
|
170
170
|
#endif // TENSORFLOW_LITE_MICRO_MEMORY_PLANNER_GREEDY_MEMORY_PLANNER_H_
|
@@ -19,7 +19,7 @@ limitations under the License.
|
|
19
19
|
#include "tensorflow/lite/micro/compatibility.h"
|
20
20
|
#include "tensorflow/lite/micro/memory_planner/micro_memory_planner.h"
|
21
21
|
|
22
|
-
namespace
|
22
|
+
namespace tflite_micro {
|
23
23
|
|
24
24
|
// The simplest possible memory planner that just lays out all buffers at
|
25
25
|
// increasing offsets without trying to reuse memory.
|
@@ -48,6 +48,6 @@ class LinearMemoryPlanner : public MicroMemoryPlanner {
|
|
48
48
|
TF_LITE_REMOVE_VIRTUAL_DELETE
|
49
49
|
};
|
50
50
|
|
51
|
-
} // namespace
|
51
|
+
} // namespace tflite_micro
|
52
52
|
|
53
53
|
#endif // TENSORFLOW_LITE_MICRO_MEMORY_PLANNER_LINEAR_MEMORY_PLANNER_H_
|
@@ -21,7 +21,7 @@ limitations under the License.
|
|
21
21
|
|
22
22
|
#include "tensorflow/lite/micro/micro_utils.h"
|
23
23
|
|
24
|
-
namespace
|
24
|
+
namespace tflite_micro {
|
25
25
|
|
26
26
|
// This is an experimental feature and subjected to change.
|
27
27
|
// More description is available at
|
@@ -68,6 +68,6 @@ constexpr size_t SizeOfBufferPlan(int32_t buffer_count) {
|
|
68
68
|
sizeof(BufferDescriptor) * Max(buffer_count - 1, 0);
|
69
69
|
}
|
70
70
|
|
71
|
-
} // namespace
|
71
|
+
} // namespace tflite_micro
|
72
72
|
|
73
73
|
#endif // TENSORFLOW_LITE_MICRO_MEMORY_PLANNER_MEMORY_PLAN_STRUCT_H_
|
@@ -18,7 +18,7 @@ limitations under the License.
|
|
18
18
|
|
19
19
|
#include "tensorflow/lite/c/common.h"
|
20
20
|
|
21
|
-
namespace
|
21
|
+
namespace tflite_micro {
|
22
22
|
|
23
23
|
// Interface class for planning the layout of memory buffers during the
|
24
24
|
// execution of a graph.
|
@@ -90,6 +90,6 @@ class MicroMemoryPlanner {
|
|
90
90
|
}
|
91
91
|
};
|
92
92
|
|
93
|
-
} // namespace
|
93
|
+
} // namespace tflite_micro
|
94
94
|
|
95
95
|
#endif // TENSORFLOW_LITE_MICRO_MICRO_MEMORY_PLANNER_MEMORY_PLANNER_H_
|
@@ -20,7 +20,7 @@ limitations under the License.
|
|
20
20
|
#include "tensorflow/lite/micro/memory_planner/memory_plan_struct.h"
|
21
21
|
#include "tensorflow/lite/micro/memory_planner/micro_memory_planner.h"
|
22
22
|
|
23
|
-
namespace
|
23
|
+
namespace tflite_micro {
|
24
24
|
|
25
25
|
/* This is an experimental feature and subjected to change.
|
26
26
|
*
|
@@ -128,6 +128,6 @@ class NonPersistentMemoryPlannerShim : public MicroMemoryPlanner {
|
|
128
128
|
TF_LITE_REMOVE_VIRTUAL_DELETE
|
129
129
|
};
|
130
130
|
|
131
|
-
} // namespace
|
131
|
+
} // namespace tflite_micro
|
132
132
|
|
133
133
|
#endif // TENSORFLOW_LITE_MICRO_MEMORY_PLANNER_NON_PERSISTENT_MEMORY_PLANNER_SHIM_H__
|
@@ -21,7 +21,7 @@ limitations under the License.
|
|
21
21
|
#include "tensorflow/lite/micro/micro_allocator.h"
|
22
22
|
#include "tensorflow/lite/schema/schema_generated.h"
|
23
23
|
|
24
|
-
namespace
|
24
|
+
namespace tflite_micro {
|
25
25
|
|
26
26
|
// Used to hold information used during allocation calculations.
|
27
27
|
struct AllocationInfo {
|
@@ -127,12 +127,12 @@ class AllocationInfoBuilder {
|
|
127
127
|
TfLiteStatus ValidateSubgraph(const SubGraph* subgraph,
|
128
128
|
TfLiteEvalTensor* eval_tensors);
|
129
129
|
|
130
|
-
const
|
130
|
+
const tflite_micro::Model* model_ = nullptr;
|
131
131
|
INonPersistentBufferAllocator* non_persistent_allocator_ = nullptr;
|
132
132
|
GraphAllocationInfo info_;
|
133
133
|
int allocation_scope_count_ = 0;
|
134
134
|
};
|
135
135
|
|
136
|
-
} // namespace
|
136
|
+
} // namespace tflite_micro
|
137
137
|
|
138
138
|
#endif // TENSORFLOW_LITE_MICRO_MICRO_ALLOCATION_INFO_H_
|
@@ -26,7 +26,7 @@ limitations under the License.
|
|
26
26
|
#include "tensorflow/lite/micro/tflite_bridge/flatbuffer_conversions_bridge.h"
|
27
27
|
#include "tensorflow/lite/schema/schema_generated.h"
|
28
28
|
|
29
|
-
namespace
|
29
|
+
namespace tflite_micro {
|
30
30
|
|
31
31
|
// TODO(b/199402574): rename to tflite_internal or just remove internal
|
32
32
|
// namespace.
|
@@ -39,7 +39,7 @@ namespace internal {
|
|
39
39
|
TfLiteStatus InitializeTfLiteTensorFromFlatbuffer(
|
40
40
|
IPersistentBufferAllocator* persistent_buffer_allocator,
|
41
41
|
INonPersistentBufferAllocator* non_persistent_buffer_allocator,
|
42
|
-
bool allocate_temp, const
|
42
|
+
bool allocate_temp, const tflite_micro::Tensor& flatbuffer_tensor,
|
43
43
|
const flatbuffers::Vector<flatbuffers::Offset<Buffer>>* buffers,
|
44
44
|
TfLiteTensor* result);
|
45
45
|
|
@@ -347,5 +347,5 @@ class MicroAllocator {
|
|
347
347
|
TF_LITE_REMOVE_VIRTUAL_DELETE
|
348
348
|
};
|
349
349
|
|
350
|
-
} // namespace
|
350
|
+
} // namespace tflite_micro
|
351
351
|
#endif // TENSORFLOW_LITE_MICRO_MICRO_ALLOCATOR_H_
|
@@ -16,13 +16,13 @@ limitations under the License.
|
|
16
16
|
#ifndef TENSORFLOW_LITE_MICRO_MICRO_ARENA_CONSTANTS_H_
|
17
17
|
#define TENSORFLOW_LITE_MICRO_MICRO_ARENA_CONSTANTS_H_
|
18
18
|
|
19
|
-
namespace
|
19
|
+
namespace tflite_micro {
|
20
20
|
|
21
21
|
// The default buffer alignment requirement.
|
22
22
|
// We align tensor buffers to 16-byte boundaries, since this is a common
|
23
23
|
// requirement for SIMD extensions.
|
24
24
|
constexpr int MicroArenaBufferAlignment() { return 16; }
|
25
25
|
|
26
|
-
} // namespace
|
26
|
+
} // namespace tflite_micro
|
27
27
|
|
28
28
|
#endif // TENSORFLOW_LITE_MICRO_MICRO_ARENA_CONSTANTS_H_
|
@@ -23,7 +23,7 @@ limitations under the License.
|
|
23
23
|
|
24
24
|
#ifdef NO_INTERPRETER
|
25
25
|
|
26
|
-
namespace
|
26
|
+
namespace tflite_micro {
|
27
27
|
const TfLiteStatus kTfLiteAbort = static_cast<TfLiteStatus>(15);
|
28
28
|
|
29
29
|
struct MicroContext{
|
@@ -40,7 +40,7 @@ namespace tflite {
|
|
40
40
|
|
41
41
|
#else
|
42
42
|
|
43
|
-
namespace
|
43
|
+
namespace tflite_micro {
|
44
44
|
// TODO(b/149795762): kTfLiteAbort cannot be part of the tflite TfLiteStatus.
|
45
45
|
const TfLiteStatus kTfLiteAbort = static_cast<TfLiteStatus>(15);
|
46
46
|
|
@@ -169,7 +169,7 @@ inline TfLiteExternalContext* MicroContextGetExternalContext(
|
|
169
169
|
void MicroContextReportOpError(struct TfLiteContext* context,
|
170
170
|
const char* format, ...);
|
171
171
|
|
172
|
-
} // namespace
|
172
|
+
} // namespace tflite_micro
|
173
173
|
|
174
174
|
#endif // NO_INTERPRETER
|
175
175
|
|
@@ -22,7 +22,7 @@ limitations under the License.
|
|
22
22
|
|
23
23
|
#ifdef NO_INTERPRETER
|
24
24
|
|
25
|
-
namespace
|
25
|
+
namespace tflite_micro {
|
26
26
|
struct MicroGraph{
|
27
27
|
int (*NumSubgraphs)();
|
28
28
|
size_t (*NumSubgraphInputs)(int subgraph_idx);
|
@@ -35,11 +35,11 @@ namespace tflite {
|
|
35
35
|
|
36
36
|
#else
|
37
37
|
|
38
|
-
namespace
|
38
|
+
namespace tflite_micro {
|
39
39
|
|
40
40
|
// Abstracts the details of interacting with the graph from the kernels
|
41
41
|
//
|
42
|
-
// Provides methods to invoke any subgraph in the
|
42
|
+
// Provides methods to invoke any subgraph in the tflite_micro::Graph.
|
43
43
|
class MicroGraph {
|
44
44
|
public:
|
45
45
|
virtual ~MicroGraph() = default;
|
@@ -72,7 +72,7 @@ class MicroGraph {
|
|
72
72
|
TF_LITE_REMOVE_VIRTUAL_DELETE
|
73
73
|
};
|
74
74
|
|
75
|
-
} // namespace
|
75
|
+
} // namespace tflite_micro
|
76
76
|
|
77
77
|
#endif // NO_INTERPRETER
|
78
78
|
|
@@ -35,7 +35,7 @@ limitations under the License.
|
|
35
35
|
// tensorflow/core.
|
36
36
|
#define TFLITE_SCHEMA_VERSION (3)
|
37
37
|
|
38
|
-
namespace
|
38
|
+
namespace tflite_micro {
|
39
39
|
|
40
40
|
class MicroInterpreter {
|
41
41
|
public:
|
@@ -184,6 +184,6 @@ class MicroInterpreter {
|
|
184
184
|
MicroInterpreterContext micro_context_;
|
185
185
|
};
|
186
186
|
|
187
|
-
} // namespace
|
187
|
+
} // namespace tflite_micro
|
188
188
|
|
189
189
|
#endif // TENSORFLOW_LITE_MICRO_MICRO_INTERPRETER_H_
|
@@ -22,7 +22,7 @@ limitations under the License.
|
|
22
22
|
#include "tensorflow/lite/micro/micro_interpreter_graph.h"
|
23
23
|
#include "tensorflow/lite/micro/micro_log.h"
|
24
24
|
|
25
|
-
namespace
|
25
|
+
namespace tflite_micro {
|
26
26
|
|
27
27
|
// A full implementation of the MicroContext, to be used by the
|
28
28
|
// MicroInterpreter. Kernels should not depend on this directly. Instead they
|
@@ -120,6 +120,6 @@ class MicroInterpreterContext : public MicroContext {
|
|
120
120
|
TF_LITE_REMOVE_VIRTUAL_DELETE
|
121
121
|
};
|
122
122
|
|
123
|
-
} // namespace
|
123
|
+
} // namespace tflite_micro
|
124
124
|
|
125
125
|
#endif // TENSORFLOW_LITE_MICRO_MICRO_INTERPRETER_CONTEXT_H_
|
@@ -22,12 +22,12 @@ limitations under the License.
|
|
22
22
|
#include "tensorflow/lite/micro/micro_resource_variable.h"
|
23
23
|
#include "tensorflow/lite/schema/schema_generated.h"
|
24
24
|
|
25
|
-
namespace
|
25
|
+
namespace tflite_micro {
|
26
26
|
|
27
|
-
// Abstracts the details of interacting with the
|
27
|
+
// Abstracts the details of interacting with the tflite_micro::Model.
|
28
28
|
//
|
29
29
|
// Provides methods to access, initialize, prepare, invoke and free any
|
30
|
-
// subgraph in the
|
30
|
+
// subgraph in the tflite_micro::Graph.
|
31
31
|
class MicroInterpreterGraph : public MicroGraph {
|
32
32
|
public:
|
33
33
|
// The lifetime of the context, model, allocator and resource_variables must
|
@@ -105,6 +105,6 @@ class MicroInterpreterGraph : public MicroGraph {
|
|
105
105
|
TF_LITE_REMOVE_VIRTUAL_DELETE
|
106
106
|
};
|
107
107
|
|
108
|
-
} // namespace
|
108
|
+
} // namespace tflite_micro
|
109
109
|
|
110
110
|
#endif // TENSORFLOW_LITE_MICRO_MICRO_INTERPRETER_GRAPH_H_
|